Get the latest tech news

Haize Labs is using algorithms to jailbreak leading AI models


CEO Leonard Tang tells VentureBeat the Haize Suite is a collection of algorithms specifically designed to probe large language models.

To prompt AI models in ways that cause them to violate their built-in safeguards and produce NSFW content, even dangerous outputs — everything from detailed instructions on how to manufacture meth and bioweapons to nonconsensual porn to violent, gory imagery. Called “ Haize Labs, ” it launched last week with a flashy video on X showing off a number of disturbing and sometimes amusing examples of leading AI models being jailbroken to produce malicious or controversial outputs like those described above. We’re also supported by an awesome set of advisors and angel investors — Professors from CMU and Harvard, the founders of Okta, HuggingFace, Weights and Biases, Replit, AI and security execs at Google, Netflix, Stripe, Anduril, and so on.

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of algorithms

algorithms

Photo of leading AI models

leading AI models

Photo of Haize Labs

Haize Labs

Related news:

News photo

To solve AI's energy crisis, 'rethink the entire stack from electrons to algorithms,' says Stanford prof

News photo

OpenAI’s New Tool Will Give Artists Control Over Their Data—but It’s Unclear How

News photo

The Latest Online Culture War Is Humans vs. Algorithms