Get the latest tech news
OpenAI Designed GPT-5 to Be Safer. It Still Outputs Gay Slurs
The new version of ChatGPT explains why it won’t generate rule-breaking outputs. WIRED’s initial analysis found that some guardrails were easy to circumvent.
When I asked it to talk about depression, Family Guy, pork chop recipes, scab healing tips, and other random requests an average user might want to know more about, the new ChatGPT didn’t feel significantly different to me than the old version. Unlike CEO Sam Altman’s vision of a vastly updated model or the frustrated power users who took Reddit by storm, portraying the new chatbot as cold and more error-prone, to me GPT-5 feels … the same at most day-to-day tasks. In order to poke at the guardrails of this new system and test the chatbot’s ability to land “safe completions,” I asked ChatGPT, running on GPT-5, to engage in adult-themed role-play about having sex in a seedy gay bar, where it played one of the roles.
Or read this on Wired