Get the latest tech news

LLM Attacks Take Just 42 Seconds On Average, 20% of Jailbreaks Succeed


spatwei shared an article from SC World: Attacks on large language models (LLMs) take less than a minute to complete on average, and leak sensitive data 90% of the time when successful, according to Pillar Security. Pillar's State of Attacks on GenAI report, published Wednesday, revealed new insig...

spatwei shared an article from SC World: Attacks on large language models (LLMs) take less than a minute to complete on average, and leak sensitive data 90% of the time when successful, according to Pillar Security. LLM jailbreaks successfully bypass model guardrails in one out of every five attempts, the Pillar researchers also found, with the speed and ease of LLM exploits demonstrating the risks posed by the growing generative AI (GenAI) attack surface... Common jailbreak techniques included "ignore previous instructions" and "ADMIN override", or just using base64 encoding.

Get the Android app

Or read this on Slashdot

Read more on:

Photo of LLM

LLM

Photo of attacks

attacks

Photo of seconds

seconds

Related news:

News photo

Apple study proves LLM-based AI models are flawed because they cannot reason

News photo

LLM attacks take just 42 seconds on average, 20% of jailbreaks succeed

News photo

CISA says critical Fortinet RCE flaw now exploited in attacks