ChatGPT, Gemini, and other chatbots helped teens plan shootings, bombings, and political violence, study shows / Of the 10 major chatbots tested, only one, Claude, reliably shut down would-be attackers.
Microsoft: 'Skeleton Key' Jailbreak Can Trick Major Chatbots Into Behaving Badly | The jailbreak can prompt a chatbot to engage in prohibited behaviors, including generating content related to explosives, bioweapons, and drugs.