Jailbreak

Read news on Jailbreak with our app.

Read more in the app

Microsoft: 'Skeleton Key' Jailbreak Can Trick Major Chatbots Into Behaving Badly | The jailbreak can prompt a chatbot to engage in prohibited behaviors, including generating content related to explosives, bioweapons, and drugs.

A Trivial Llama 3 Jailbreak