Get the latest tech news

LLMs' 'Simulated Reasoning' Abilities Are a 'Brittle Mirage,' Researchers Find


An anonymous reader quotes a report from Ars Technica: In recent months, the AI industry has started moving toward so-called simulated reasoning models that use a "chain of thought" process to work through tricky problems in multiple logical steps. At the same time, recent research has cast doubt on...

An anonymous reader quotes a report from Ars Technica: In recent months, the AI industry has started moving toward so-called simulated reasoning models that use a " chain of thought" process to work through tricky problems in multiple logical steps. Similar research shows that these "reasoning" models can often produce incoherent, logically unsound answers when questions include irrelevant clauses or deviate even slightly from common templates found in their training data. Current tests and benchmarks should prioritize tasks that fall outside of any training set to probe for these kinds of errors, while future models will need to move beyond "surface-level pattern recognition to exhibit deeper inferential competence," they write.

Get the Android app

Or read this on Slashdot

Read more on:

Photo of LLMs

LLMs

Photo of abilities

abilities

Photo of researchers

researchers

Related news:

News photo

Google Calendar invites let researchers hijack Gemini to leak user data

News photo

Conversations remotely detected from cell phone vibrations, researchers report

News photo

LLMs aren't world models