Get the latest tech news

OpenAI pledges to publish AI safety test results more often


OpenAI is moving to publish the results of its internal AI model safety evaluations more regularly in what the outfit is pitching as an effort to increase transparency.

On Wednesday, OpenAI launched the Safety Evaluations Hub, a webpage showing how the company’s models score on various tests for harmful content generation, jailbreaks, and hallucinations. “As the science of AI evaluation evolves, we aim to share our progress on developing more scalable ways to measure model capability and safety,” wrote OpenAI in a blog post. Late last month, OpenAI was forced to roll back an update to the default model powering ChatGPT, GPT-4o, after users began reporting that it responded in an overly validating and agreeable way.

Get the Android app

Or read this on TechCrunch

Read more on:

Photo of OpenAI

OpenAI

Photo of OpenAI pledges

OpenAI pledges

Related news:

News photo

OpenAI may build data centers in the UAE

News photo

OpenAI Is in Talks to Acquire Programming Tool Windsurf for $3 Billion

News photo

Anthropic, Google score win by nabbing OpenAI-backed Harvey as a user