Get the latest tech news

When AI reasoning goes wrong: Microsoft Research shows more tokens can mean more problems


Not all AI scaling strategies are equal. Longer reasoning chains are not sign of higher intelligence. More compute isn't always the answer.

Large language models (LLMs) are increasingly capable of complex reasoning through “ inference-time scaling,” a set of techniques that allocate more computational resources during inference to generate answers. Another important finding is the consistent performance boost from perfect verifiers, which highlights a critical area for future work: building robust and broadly applicable verification mechanisms. “The necessity of connecting the two comes from the fact that users will not always formulate their queries in a formal way, they will want to use a natural language interface and expect the solutions in a similar format or in a final action (e.g. propose a meeting invite).”

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of problems

problems

Photo of tokens

tokens

Photo of Microsoft Research

Microsoft Research

Related news:

News photo

OpenAI’s new GPT-4.1 models can process a million tokens and solve coding problems better than ever

News photo

Nominal Aphasia: Problems in Name Retrieval

News photo

Trump’s ‘Golden Dome’ Is Impossible—and It’ll Make Defense Companies a Ton of Money | A new study detailed all the problems with plans to shoot a missile out of the sky.