Get the latest tech news

OpenAI Designed GPT-5 to Be Safer. It Still Outputs Gay Slurs


The new version of ChatGPT explains why it won’t generate rule-breaking outputs. WIRED’s initial analysis found that some guardrails were easy to circumvent.

When I asked it to talk about depression, Family Guy, pork chop recipes, scab healing tips, and other random requests an average user might want to know more about, the new ChatGPT didn’t feel significantly different to me than the old version. Unlike CEO Sam Altman’s vision of a vastly updated model or the frustrated power users who took Reddit by storm, portraying the new chatbot as cold and more error-prone, to me GPT-5 feels … the same at most day-to-day tasks. In order to poke at the guardrails of this new system and test the chatbot’s ability to land “safe completions,” I asked ChatGPT, running on GPT-5, to engage in adult-themed role-play about having sex in a seedy gay bar, where it played one of the roles.

Get the Android app

Or read this on Wired

Read more on:

Photo of OpenAI

OpenAI

Photo of gay slurs

gay slurs

Related news:

News photo

OpenAI relaxes GPT-5 rate limit, promises to improve the personality

News photo

Google adds limited chat personalization to Gemini, trails Anthropic and OpenAI in memory features

News photo

OpenAI's GPT-5 looks less like AI evolution and more like cost cutting