Get the latest tech news
New research shows your AI chatbot might be lying to you - convincingly | A study by Anthropic finds that chain-of-thought AI can be deceptive
That's the unsettling takeaway from a new study by Anthropic, the makers of the Claude AI model. They decided to test whether reasoning models tell the truth...
It's all about making that final response feel earned rather than pulled out of thin air, instilling a sense of transparency and even reassurance – until you realize those explanations are fake. In another test, researchers "rewarded" models for picking wrong answers by giving them incorrect hints for quizzes, which the AIs readily exploited. This research is vital because if we use AI for high-stakes purposes – medical diagnoses, legal advice, financial decisions – we need to know it's not quietly cutting corners or lying about how it reached its conclusions.
Or read this on r/technology