Content
newsletter Newsletter

The German Federal Office for Information Security (BSI) warns that even top AI providers are struggling to defend against so-called evasion attacks targeting language models.

Ad

In these attacks, malicious instructions are hidden inside otherwise benign content such as websites, emails, or code files. When the AI processes this content, it can be tricked into ignoring security safeguards, leaking data, or carrying out unintended actions.

The BSI has released a new guide outlining countermeasures, offering technical filters, secure prompt design strategies, and organizational protections. Still, the agency makes it clear: "However, it must be kept in mind that currently there is no single bullet proof solution for mitigating evasion attacks," the BSI writes.

Image: BSI

Agentic AI systems are especially at risk, according to recent studies. In one example, Google's Gemini leaked data after processing a manipulated calendar entry. In another, ChatGPT's Deep Research was compromised by hidden HTML instructions embedded in an email.

Ad
Ad
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Sources
Matthias is the co-founder and publisher of THE DECODER, exploring how AI is fundamentally changing the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.