Get the latest tech news

OpenAI's Latest Model Closes the 'Ignore All Previous Instructions' Loophole


Kylie Robison reports via The Verge: Have you seen the memes online where someone tells a bot to "ignore all previous instructions" and proceeds to break it in the funniest ways possible? The way it works goes something like this: Imagine we at The Verge created an AI bot with explicit instructions...

Kylie Robison reports via The Verge: Have you seen the memes online where someone tells a bot to "ignore all previous instructions" and proceeds to break it in the funniest ways possible? The way it works goes something like this: Imagine we at The Verge created an AI bot with explicit instructions to direct you to our excellent reporting on any subject. In a conversation with Olivier Godement, who leads the API platform product at OpenAI, he explained that instruction hierarchy will prevent the meme'd prompt injections(aka tricking the AI with sneaky commands) we see all over the internet.

Get the Android app

Or read this on Slashdot

Read more on:

Photo of OpenAI

OpenAI

Photo of loophole

loophole

Photo of latest model

latest model

Related news:

News photo

OpenAI, Nvidia, and Hugging Face unveil small AI models: GPT-4o Mini, Mistral-Nemo, and SmolLM lead industry shift

News photo

OpenAI Dropped From First Ever AI Programming Copyright Lawsuit

News photo

OpenAI’s latest model will block the ‘ignore all previous instructions’ loophole