Get the latest tech news

New research shows your AI chatbot might be lying to you - convincingly | A study by Anthropic finds that chain-of-thought AI can be deceptive


That's the unsettling takeaway from a new study by Anthropic, the makers of the Claude AI model. They decided to test whether reasoning models tell the truth...

It's all about making that final response feel earned rather than pulled out of thin air, instilling a sense of transparency and even reassurance – until you realize those explanations are fake. In another test, researchers "rewarded" models for picking wrong answers by giving them incorrect hints for quizzes, which the AIs readily exploited. This research is vital because if we use AI for high-stakes purposes – medical diagnoses, legal advice, financial decisions – we need to know it's not quietly cutting corners or lying about how it reached its conclusions.

Get the Android app

Or read this on r/technology

Read more on:

Photo of Study

Study

Photo of ai chatbot

ai chatbot

Photo of Anthropic

Anthropic

Related news:

News photo

Study finds solo music listening boosts social well-being

News photo

OpenAI just made ChatGPT Plus free for millions of college students — and it’s a brilliant competitive move against Anthropic

News photo

Don’t believe reasoning models’ Chains of Thought, says Anthropic