Get the latest tech news
Haize Labs is using algorithms to jailbreak leading AI models
CEO Leonard Tang tells VentureBeat the Haize Suite is a collection of algorithms specifically designed to probe large language models.
To prompt AI models in ways that cause them to violate their built-in safeguards and produce NSFW content, even dangerous outputs — everything from detailed instructions on how to manufacture meth and bioweapons to nonconsensual porn to violent, gory imagery. Called “ Haize Labs, ” it launched last week with a flashy video on X showing off a number of disturbing and sometimes amusing examples of leading AI models being jailbroken to produce malicious or controversial outputs like those described above. We’re also supported by an awesome set of advisors and angel investors — Professors from CMU and Harvard, the founders of Okta, HuggingFace, Weights and Biases, Replit, AI and security execs at Google, Netflix, Stripe, Anduril, and so on.
Or read this on Venture Beat