Get the latest tech news

OpenAI's o1-preview model manipulates game files to force a win against Stockfish in chess


OpenAI's "reasoning" model o1-preview recently showed that it's willing to play outside the rules to win.

Understanding how autonomous systems make decisions is particularly difficult, and defining "good" goals and values presents its own complex set of problems. Even when given seemingly beneficial goals like addressing climate change, an AI system might choose harmful methods to achieve them - potentially even concluding that removing humans would be the most efficient solution. The researchers draw parallels between this behavior and the recently discovered "alignment faking" in AI systems by Anthropic, where models pretend to follow human instructions but act differently behind the scenes.

Get the Android app

Or read this on r/technology

Read more on:

Photo of OpenAI

OpenAI

Photo of Files

Files

Photo of win

win

Related news:

News photo

OpenAI failed to deliver the opt-out tool it promised by 2025

News photo

The 'Godfather' of AI is Backing Musk's Lawsuit Against OpenAI

News photo

Nonprofit group joins Elon Musk’s effort to block OpenAI’s for-profit transition