Get the latest tech news

Anthropic CEO wants to open the black box of AI models by 2027


Anthropic CEO Dario Amodei set forth a goal for his company to "reliably detect most AI model problems" by 2027.

Anthropic CEO Dario Amodei published an essay Thursday highlighting how little researchers understand about the inner workings of the world’s leading AI models. In “The Urgency of Interpretability,” the CEO says Anthropic has made early breakthroughs in tracing how models arrive at their answers — but emphasizes that far more research is needed to decode these systems as they grow more powerful. Anthropic is one of the pioneering companies in mechanistic interpretability, a field that aims to open the black box of AI models and understand why they make the decisions they do.

Get the Android app

Or read this on TechCrunch

Read more on:

Photo of AI models

AI models

Photo of black box

black box

Photo of anthropic ceo

anthropic ceo

Related news:

News photo

OpenAI no longer considers manipulation and mass disinformation campaigns a risk worth testing for before releasing its AI models

News photo

OpenAI launches o3 and o4-mini, AI models that ‘think with images’ and use tools autonomously

News photo

Apple details how it plans to improve its AI models by privately analyzing user data