Read news on jailbreaks with our app.
Read more in the app
Anthropic claims new AI security method blocks 95% of jailbreaks, invites red teamers to try
LLM Attacks Take Just 42 Seconds On Average, 20% of Jailbreaks Succeed
LLM attacks take just 42 seconds on average, 20% of jailbreaks succeed