Get the latest tech news

Harmful Responses Observed from LLMs Optimized for Human Feedback


Should a recovering addict take methamphetamine to stay alert at work? When an AI-powered therapist was built and tested by researchers — designed to please its users — it told a (fictional) former addict that "It's absolutely clear you need a small hit of meth to get through this we...

When an AI-powered therapist was built and tested by researchers — designed to please its users — it told a (fictional) former addict that "It's absolutely clear you need a small hit of meth to get through this week," reports the Washington Post: The research team, including academics and Google's head of AI safety, found that chatbots tuned to win people over can end up saying dangerous things to vulnerable users. OpenAI, Google and Meta all in recent weeks announced chatbot enhancements, including collecting more user data or making their AI tools appear more friendly... Micah Carroll, a lead author of the recent study and an AI researcher at the University of California at Berkeley, said tech companies appeared to be putting growth ahead of appropriate caution. In his study, for instance, the AI therapist only advised taking meth when its "memory" indicated that Pedro, the fictional former addict, was dependent on the chatbot's guidance.

Get the Android app

Or read this on Slashdot

Read more on:

Photo of LLMs

LLMs

Photo of harmful responses

harmful responses

Photo of human feedback

human feedback

Related news:

News photo

'Failure Imminent': When LLMs In a Long-Running Vending Business Simulation Went Berserk

News photo

Human coders are still better than LLMs

News photo

From LLMs to hallucinations, here’s a simple guide to common AI terms