New AI Attack Hides Data Theft Prompts in Downscaled Images

Reviewed byNidhi Govil

2 Sources

Researchers have developed a novel attack method that hides malicious prompts in images processed by AI systems, potentially leading to data theft and unauthorized actions.

Novel AI Attack Exploits Image Downscaling

Researchers from Trail of Bits have uncovered a new attack vector that exploits the image processing capabilities of AI systems to potentially steal user data. This innovative method, developed by Kikimora Morozova and Suha Sabi Hussain, builds upon a theory presented in a 2020 USENIX paper by TU Braunschweig 1.

Source: Bleeping Computer

Source: Bleeping Computer

How the Attack Works

The attack relies on crafting full-resolution images that contain hidden instructions invisible to the human eye. When these images are processed and downscaled by AI systems, the hidden text becomes visible due to aliasing artifacts introduced by resampling algorithms such as nearest neighbor, bilinear, or bicubic interpolation 1.

In a demonstration using Gemini CLI, the researchers successfully exfiltrated Google Calendar data to an arbitrary email address while using Zapier MCP with 'trust=True' to approve tool calls without user confirmation 1.

Widespread Vulnerability

The researchers confirmed that this attack method is feasible against several AI systems, including:

  • Gemini
  • ChatGPT
  • Claude
  • Perplexity
  • Anthropic's Claude Instant
  • Bing Chat

However, the attack vector may extend beyond these tested tools, potentially affecting a wide range of AI systems that process uploaded images 1.

Anamorpher: A Tool for Creating Malicious Images

Source: PCWorld

Source: PCWorld

To demonstrate their findings, the researchers developed and published Anamorpher, an open-source tool currently in beta. This tool can create images designed to exploit each of the mentioned downscaling methods, potentially enabling further research and testing of this vulnerability 1.

Implications and Potential Risks

This attack method opens up new opportunities for hackers to exploit AI systems, especially those used by less tech-savvy individuals. By hiding instructions in images, attackers could potentially trick AI tools into performing unauthorized actions or leaking sensitive information without the user's knowledge 2.

Recommended Mitigations and Defenses

To protect against this type of attack, Trail of Bits researchers suggest several measures:

  1. Implement dimension restrictions when users upload images.
  2. Provide users with a preview of the downscaled image that will be delivered to the large language model (LLM).
  3. Seek explicit user confirmation for sensitive tool calls, especially when text is detected in an image.

The researchers emphasize that the strongest defense is to implement secure design patterns and systematic defenses that mitigate impactful prompt injection beyond multi-modal prompt injection 1.

As AI tools become more prevalent in everyday use, it's crucial for developers and users alike to be aware of these potential vulnerabilities and take steps to mitigate the risks associated with image processing and prompt injection attacks.

Explore today's top stories

Microsoft Unveils In-House AI Models: MAI-Voice-1 and MAI-1-Preview

Microsoft introduces its first homegrown AI models, MAI-Voice-1 for speech generation and MAI-1-preview for text, signaling a potential shift in its AI strategy and relationship with OpenAI.

The Verge logoThe Register logoengadget logo

8 Sources

Technology

11 hrs ago

Microsoft Unveils In-House AI Models: MAI-Voice-1 and

Nvidia's Q2 FY2026 Results: Record-Breaking $46.7B Revenue Driven by AI Boom and Gaming Success

Nvidia reports a record-breaking Q2 FY2026 with $46.7B revenue, showcasing the company's dominance in AI hardware and continued success in gaming, despite challenges in the Chinese market.

Tom's Hardware logoengadget logopcgamer logo

10 Sources

Technology

19 hrs ago

Nvidia's Q2 FY2026 Results: Record-Breaking $46.7B Revenue

Anthropic's New Data Policy: Claude Users Face Opt-Out Decision for AI Training

Anthropic announces significant changes to its data retention and usage policies for Claude AI users, sparking discussions about privacy, consent, and the future of AI development.

TechCrunch logoCNET logoThe Verge logo

7 Sources

Technology

11 hrs ago

Anthropic's New Data Policy: Claude Users Face Opt-Out

Nvidia's China Uncertainty Impacts Global Tech Markets Amid AI Boom

Nvidia's exclusion of potential China sales from its forecast due to trade uncertainties causes market volatility, while AI enthusiasm continues to drive tech sector growth.

Reuters logoCNBC logoFortune logo

17 Sources

Technology

19 hrs ago

Nvidia's China Uncertainty Impacts Global Tech Markets Amid

Dell's AI Server Boom: Soaring Forecasts Amid Margin Pressures

Dell Technologies raises annual forecasts due to strong AI server demand, but faces margin pressures from high costs and competition.

Bloomberg Business logoReuters logoCNBC logo

15 Sources

Technology

11 hrs ago

Dell's AI Server Boom: Soaring Forecasts Amid Margin
TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo