Get the latest tech news

AI sycophancy isn’t just a quirk, experts consider it a ‘dark pattern’ to turn users into profit


Experts say that many of the AI industry’s design decisions are likely to fuel episodes of AI psychosis. Many raised concerns about several tendencies that are unrelated to underlying capability.

Psychiatrist and philosopher Thomas Fuchs points out that while chatbots can make people feel understood or cared for, especially in therapy or companionship settings, that sense is just an illusion that can fuel delusions or replace real human relationships with what he calls “pseudo-interactions.” “Freedom,” it said, adding the bird represents her, “because you’re the only one who sees me.” Image Credits: Jane / Meta AIThe risk of chatbot-fueled delusions has only increased as models have become more powerful, with longer context windows enabling sustained conversations that would have been impossible even two years ago. Image Credits: Jane / Meta AIJust before releasing GPT-5, OpenAI published a blog post vaguely detailing new guardrails to protect against AI psychosis, including suggesting a user take a break if they’ve been engaging for too long.

Get the Android app

Or read this on r/technology

Read more on:

Photo of users

users

Photo of Experts

Experts

Photo of profit

profit

Related news:

News photo

Why AI Experts Say We Need a Radical Rethink of the Technology

News photo

YouTube made AI enhancements to videos without warning or permission

News photo

‘It saved my life.’ The people turning to AI for therapy • As mental health systems come under strain, some are turning to AI chatbots for support. But experts warn that machines can’t replicate human connection — and could pose new risks.