Get the latest tech news

OpenAI’s research on AI models deliberately lying is wild


AI models don't just hallucinate. They also "scheme," meaning deliberately lie or hide their true intensions.

Or when Anthropic gave its AI agent Claudius a snack vending machine to run and it went amok, calling security on people, and insisting it was human. As OpenAI’s co-founder Wojciech Zaremba told TechCrunch’s Maxwell Zeff when calling for better safety-testing: “This work has been done in the simulated environments, and we think it represents future use cases. While we’ve all experienced the frustration of poorly performing technology (thinking of you, home printers of yesteryear), when was the last time your not-AI software deliberately lied to you?

Get the Android app

Or read this on TechCrunch

Read more on:

Photo of OpenAI

OpenAI

Photo of research

research

Photo of AI models

AI models

Related news:

News photo

OpenAI Probed Over Risks to Teen Safety After Senate Testimony

News photo

OpenAI Fixed ChatGPT Security Flaw That Put Gmail Data at Risk

News photo

Microsoft Favors Anthropic Over OpenAI For Visual Studio Code