Get the latest tech news

A New Benchmark for the Risks of AI


MLCommons provides benchmarks that test the abilities of AI systems. It wants to measure the bad side of AI next.

MLCommons, a nonprofit that helps companies measure the performance of their artificial intelligence systems, is launching a new benchmark to gauge AI’s bad side too. The new benchmark, called AILuminate, assesses the responses of large language models to more than 12,000 test prompts in 12 categories including inciting violent crime, child sexual exploitation, hate speech, promoting self-harm, and intellectual property infringement. Peter Mattson, founder and president of MLCommons and a senior staff engineer at Google, says that measuring the potential harms of AI models is technically difficult, leading to inconsistencies across the industry.

Get the Android app

Or read this on Wired

Read more on:

Photo of Risks

Risks

Photo of new benchmark

new benchmark

Related news:

News photo

UK Cyber Chief Warns Country 'Widely Underestimating' Risks From Cyberattacks

News photo

BAE Systems Falls as BofA Sees Risk of Musk-Led Defense Cuts

News photo

Identity management in 2025: 4 ways security teams can address gaps and risks