Get the latest tech news

Researchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious queries | ArtPrompt bypassed safety measures in ChatGPT, Gemini, Clause, and Llama2.


ArtPrompt bypassed safety measures in ChatGPT, Gemini, Clause, and Llama2.

Researchers based in Washington and Chicago have developed ArtPrompt, a new way to circumvent the safety measures built into large language models(LLMs). It is a simple and effective attack, and the paper provides examples of the ArtPrompt-induced chatbots advising on how to build bombs and make counterfeit money. Moreover, in a kind of technological game of whack-a-mole, the major AI players have spent plenty of time plugging linguistic and semantic holes to prevent people from wandering outside the guardrails.

Get the Android app

Or read this on r/technology

Read more on:

Photo of ChatGPT

ChatGPT

Photo of researchers

researchers

Photo of AI chatbots

AI chatbots

Related news:

News photo

Why people are falling in love with AI chatbots

News photo

Smarter than GPT-4: Claude 3 AI catches researchers testing it

News photo

Researchers tested leading AI models for copyright infringement using popular books, and GPT-4 performed worst