Get the latest tech news

Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens


Chain-of-thought AI “degrades significantly” when asked to generalize beyond training.

As the researchers hypothesized, these basic models started to fail catastrophically when asked to generalize novel sets of transformations that were not directly demonstrated in the training data. In other cases, the LLM would sometimes stumble onto correct answers paired with "unfaithful reasoning paths" that didn't follow logically. "Rather than demonstrating a true understanding of text, CoT reasoning under task transformations appears to reflect a replication of patterns learned during training," the researchers write.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of LLMs

LLMs

Photo of abilities

abilities

Photo of brittle mirage

brittle mirage

Related news:

News photo

LLMs' 'Simulated Reasoning' Abilities Are a 'Brittle Mirage,' Researchers Find

News photo

LLMs aren't world models

News photo

I clustered four Framework Mainboards to test LLMs