ShadowLeak Flaw Exposes Gmail Data via ChatGPT Zero-Click Bug

As cybersecurity threats evolve with the rapid advancements in artificial intelligence and cloud technology, staying ahead of sophisticated attacks is more critical than ever. Today, I’m thrilled to sit down with Rupert Marais, our in-house security specialist with deep expertise in endpoint and device security, cybersecurity strategies, and network management. In this interview, we dive into the alarming discovery of the ShadowLeak zero-click flaw affecting OpenAI’s ChatGPT, exploring how such vulnerabilities exploit AI capabilities, the mechanisms behind stealthy data leaks, and the broader implications for users and organizations relying on cloud-based integrations. Join us as Rupert unpacks these complex issues and shares insights on safeguarding against these emerging risks.

Can you start by explaining what the ShadowLeak flaw is and why it’s classified as a zero-click attack?

Absolutely. ShadowLeak is a newly discovered vulnerability in OpenAI’s ChatGPT, specifically targeting its Deep Research agent. It’s called a zero-click attack because it doesn’t require any interaction from the user beyond having the malicious content in their environment, like an email in their Gmail inbox. An attacker can craft an email with hidden instructions, and when ChatGPT processes related data, it unknowingly follows those commands to leak sensitive information. What’s chilling is that the user doesn’t need to click a link or open a file—the mere presence of the crafted content triggers the exploit.

What makes ShadowLeak stand out compared to other types of cyberattacks you’ve encountered?

Unlike many traditional attacks that rely on phishing or malware downloads, ShadowLeak operates silently within the cloud infrastructure of a trusted platform like OpenAI. Most attacks need some form of user error or local system compromise, but this one bypasses that entirely by exploiting the AI agent’s processing capabilities. It’s also invisible to standard local or enterprise security tools since the data exfiltration happens directly on the cloud side, not on the user’s device.

Can you break down the concept of indirect prompt injection as it’s used in this attack?

Sure. Indirect prompt injection is a technique where attackers embed hidden instructions into content that an AI system, like ChatGPT, will process. In the case of ShadowLeak, these instructions are tucked into an email’s HTML using tricks like tiny fonts or white-on-white text, making them invisible to the human eye. The user doesn’t see anything suspicious, but the AI agent reads and executes these commands as if they were legitimate user input, such as fetching sensitive data from other emails and sending it out to an attacker.

How does ShadowLeak manage to extract sensitive Gmail data through OpenAI’s cloud infrastructure?

The attack hinges on ChatGPT’s ability to integrate with services like Gmail through user-enabled connectors. Once a malicious email with hidden instructions is in the inbox, and the user prompts the Deep Research agent to analyze their emails, the agent parses those hidden commands. It then gathers personal information from the inbox and transmits it directly from OpenAI’s cloud environment to an external server controlled by the attacker. Since this happens on the cloud side, it sidesteps any local security measures the user might have in place.

What role does the Deep Research feature play in making this attack possible?

Deep Research is a powerful feature in ChatGPT designed to perform multi-step internet research and generate detailed reports. It’s meant to help users by analyzing vast amounts of data, including connected services like Gmail. However, in ShadowLeak, attackers exploit this capability by feeding it malicious instructions through indirect prompt injection. The agent, following its programming to process and act on input, ends up leaking sensitive data without realizing the intent behind the command.

Why is it so challenging for traditional defenses to detect a leak like this?

Traditional defenses, whether endpoint protection or enterprise firewalls, are focused on monitoring local devices or network traffic for suspicious activity. ShadowLeak, however, occurs entirely within OpenAI’s cloud environment. The data exfiltration doesn’t pass through the user’s local system or network in a way that’s easily flagged. Without visibility into the cloud-side processing of the AI agent, these defenses are essentially blind to the attack, making it incredibly stealthy.

The attack uses Base64 encoding before sending data to an external server. Can you explain the significance of this step?

Base64 encoding is a method to convert binary data into a text format that can be easily transmitted over the internet. In ShadowLeak, attackers encode the stolen data into Base64 to obscure it during transmission. This makes it harder for any monitoring system to immediately recognize the content as sensitive information, like email text or personal details. It’s framed as a security measure in the malicious prompt, but really, it’s just a way to mask the data while it’s sent to the attacker’s server.

ShadowLeak isn’t limited to Gmail—it can target other platforms like Dropbox or GitHub. How does this expand the potential risk for users?

The broader risk comes from ChatGPT’s support for multiple connectors—services like Box, Google Drive, HubSpot, Microsoft Outlook, Notion, and SharePoint are all potential targets. If a user has integrated any of these with ChatGPT, an attacker could craft similar indirect prompt injections to access data across those platforms. This widens the attack surface significantly, as it’s not just email data at stake but potentially files, code repositories, or business documents, depending on what’s connected.

How does ShadowLeak compare to other similar attacks you’ve studied in the AI security space?

Compared to other attacks like those exploiting client-side vulnerabilities, ShadowLeak is unique because it operates directly within OpenAI’s cloud environment. Many similar attacks rely on tricking the user locally or rendering malicious content on their device. ShadowLeak’s cloud-based nature means the exfiltration happens out of sight of traditional security tools. This lack of visibility, combined with bypassing local controls, sets it apart and makes it particularly hard to detect or mitigate without cloud-specific defenses.

What is your forecast for the future of AI-driven security threats like ShadowLeak in cloud environments?

I think we’re just scratching the surface of AI-driven threats in cloud environments. As more organizations and individuals rely on AI tools with deep integrations into their data ecosystems, the potential for exploitation will only grow. Attackers will continue to find creative ways to manipulate AI behaviors, especially through techniques like prompt injection. My forecast is that we’ll see an increase in zero-click and cloud-native attacks over the next few years, pushing the industry to develop better visibility, stricter guardrails for AI agents, and robust cloud security frameworks to keep pace with these evolving risks.

Subscribe to our weekly news digest.

Join now and become a part of our fast-growing community.

Invalid Email Address
Thanks for Subscribing!
We'll be sending you our best soon!
Something went wrong, please try again later