Science

Researchers Unveil Image-Based Attack Threatening AI Security

Researchers Unveil Image-Based Attack Threatening AI Security
Editorial
  • PublishedAugust 27, 2025

Security researchers have revealed a sophisticated attack technique that exploits vulnerabilities in artificial intelligence (AI) models, potentially compromising sensitive user data through manipulated images. Developed by researchers Kikimora Morozova and Suha Sabi Hussain from the security firm Trail of Bits, this method builds on concepts introduced in a 2020 study from TU Braunschweig. The researchers have demonstrated how this technique can be applied to contemporary AI systems, emphasizing the urgent need for enhanced security measures.

The attack leverages the common practice of AI systems automatically resizing uploaded images to conserve processing power and reduce costs. Various resampling algorithms, including “Nearest Neighbor,” “Bilinear,” and “Bicubic,” are frequently employed for this purpose. These methods can inadvertently reveal hidden patterns within the original images when they are scaled down, allowing attackers to embed covert instructions that become visible only after the image is processed.

For instance, researchers illustrated how dark areas in an image can be modified to appear red during the downscaling process, which in turn makes hidden black text discernible. Once this text is read by an AI model, the system may misinterpret it as legitimate user input, enabling harmful commands that could compromise confidential data. In a practical demonstration, the researchers successfully transmitted calendar data from a Google account to an unauthorized email address using the “Gemini CLI” tool.

Trail of Bits has identified numerous platforms at risk, including Google’s Gemini models, which encompass various interfaces such as the command line interface, web interface, API, and Vertex AI Studio. The Google Assistant on Android and the Genspark service are also vulnerable to these types of attacks, highlighting a pervasive security concern across widely used technologies.

To raise awareness about these risks, the researchers have created an open-source tool named “Published,” which generates images tailored for different downscaling methods, effectively demonstrating how easily such attacks can be executed. In response to these vulnerabilities, experts recommend restricting image sizes during uploads and providing users with a preview of the resized images. Additionally, they stress the importance of requiring user confirmation for safety-critical actions, particularly when extracting text from images.

However, the most critical defense against these attacks lies in robust system design that is inherently resistant to prompt injection vulnerabilities. The researchers assert that developing systematic protective mechanisms is essential to prevent multimodal AI applications from becoming channels for data exploitation.

As AI continues to integrate into various sectors, it is imperative for developers and security professionals to implement comprehensive security strategies. The implications of these findings underscore the need for vigilance in the ongoing evolution of AI technology.

Editorial
Written By
Editorial

Our Editorial team doesn’t just report the news—we live it. Backed by years of frontline experience, we hunt down the facts, verify them to the letter, and deliver the stories that shape our world. Fueled by integrity and a keen eye for nuance, we tackle politics, culture, and technology with incisive analysis. When the headlines change by the minute, you can count on us to cut through the noise and serve you clarity on a silver platter.