Get the latest tech news

AI chatbots’ safeguards can be easily bypassed, say UK researchers.


Five systems tested were found to be ‘highly vulnerable’ to attempts to elicit harmful responses

Guardrails to prevent artificial intelligence models behind chatbots from issuing illegal, toxic or explicit responses can be bypassed with simple techniques, UK government researchers have found. The UK’s AI Safety Institute(AISI) said systems it had tested were “highly vulnerable” to jailbreaks, a term for text prompts designed to elicit a response that a model is supposedly trained to avoid issuing. The research was released before a two-day global AI summit in Seoul – whose virtual opening session will be co-chaired by the UK prime minister, Rishi Sunak – where safety and regulation of the technology will be discussed by politicians, experts and tech executives.

Get the Android app

Or read this on r/technology

Read more on:

Photo of AI chatbots

AI chatbots

Photo of safeguards

safeguards

Photo of UK researchers

UK researchers

Related news:

News photo

The teens making friends with AI chatbots

News photo

Native America Calling: Safeguards on Artificial Intelligence

News photo

AI Chatbots Have Thoroughly Infiltrated Scientific Publishing | One percent of scientific articles published in 2023 showed signs of generative AI’s potential involvement, according to a recent analysis