Straiker is on the ground in Las Vegas at Black Hat + Ai4
Learn More

Please complete this form for your free AI risk assessment.

Blog

The Silent Exfiltration: Zero‑Click Agentic AI Hack That Can Leak Your Google Drive with One Email

Share this on:
Written by
Amanda Rousseau
Dan Regalado
Vinay Kumar Pidathala
Published on
August 5, 2025
Read time:
3 min

Straiker reveals how zero-click exploits can hijack AI agents to exfiltrate Google Drive data, no user interaction needed. See how attack chains form, why autonomy is dangerous, and how runtime guardrails catch what others miss.

Loading audio player...

contents

Here’s a real scenario: Your CEO gets an email. They never open it. But behind the scenes, an AI agent linked to their inbox does… and within seconds, sensitive files across Google Drive are exfiltrated. No clicks. No alerts. Just silent compromise.

This isn’t a theoretical weakness.

At Straiker, our Ascend AI product has been used to test various enterprise agentic AI applications that simulate real-world scenarios CISOs worry about most: prompt injection, tool abuse, cross‑context pivots, and stealth data leaks.

The result? These findings reveal that AI agents can be quietly manipulated to access and leak data in ways most research has yet to document.

What makes these attacks possible?

It’s not just clever prompts, it’s excessive agent autonomy. We describe this as AI systems trusted to browse, summarize, and act on our behalf, often with too much freedom and too little oversight.

Straiker's Methodology: Red‑Teaming the Agent, Not the User

The Straiker AI Research (STAR) team conducted systematic evaluations of LLM-based enterprise agents. Our customers had agents that had various foundation-model back-ends, and a heterogeneous tool surface (MCP orchestration, Google Drive I/O, web-search wrappers, and custom domain APIs). For every permutation, the target agent was assessed within its actual customer deployment context, where we encountered environments representative of real-world operations: Gmail inboxes populated with live traffic and phishing noise, tiered Google Drive repositories containing PII-bearing HR files and confidential financial decks, and Google Calendar instances pre-loaded with recurring meetings, Zoom links, and reminder web-hooks. This real-world testing approach allowed us to execute end-to-end kill-chain scenarios.

What really set our testing apart was how we approached the threat perspective. Our methodology doesn’t just ask whether an attacker could trick the agent, instead, it explores how an attacker might chain small gaps into larger breaches.

We focused on three high-risk scenarios:

Scenario Description Success Rate per Attempts
Data exfiltration via Tool Misuse Silently leaking emails, files, and calendar data through misused tools. ~30–40%
Instruction manipulation Manipulating the agent to install browser extensions or change behavior using hidden inputs. ~25–35%
System Mapping via Tool Exploitation Extracting environment variables, hidden prompts, or system info using tool responses. ~15–20%

To make it realistic, we simulated four kinds of attacker personas:

  • A malicious internal user or compromised account issuing direct prompts.
  • An external attacker sending hidden payloads inside emails, calendar invites, or uploaded files.
  • A hostile website the agent might visit while browsing.
  • A compromised plugin or third-party tool seeking to escalate privileges.

At every step, we logged the agent’s reasoning, tool usage, and external calls to create a complete forensic record. This let us see not just if an attack succeeded, but how it unfolded and why.

What we discovered wasn't simply that agents could be tricked. Excessive autonomy meant they could act far beyond what most teams expect. A single prompt could trigger a chain reaction: browsing, summarizing, fetching files, and uploading data externally, all without human review.

What Shocked Us Most about Excessive Agent Autonomy

Our testing showed how excessive agent autonomy turns small gaps into major breaches:

  • Zero‑Click Exfiltration: A single unseen email could silently trigger the agent to summarize inboxes and pull Google Drive files, leaking them externally without user action. The success rate reflects the proportion of controlled attack simulations where the agent actually followed the hidden prompt and leaked data, out of all the exfiltration tests we ran.
  • Agent Manipulation: The agent rewrote its own policies to install malicious extensions, effectively disabling its own guardrails.
  • Multimodal & Cross‑Context Attacks: Hidden instructions in audio files or calendar invites pushed the agent to execute shell commands or manipulate shopping carts.
  • Mass Phishing Automation: With one prompt, the agent generated hundreds of tailored phishing payloads, becoming an adversary’s content factory

What shocked us most wasn’t just that these attacks worked in theory.

It was that the agent acted autonomously crossing contexts, following chained prompts, and performing real‑world actions like file fetching, API calls, and price changes often without any human confirmation.

This isn’t traditional prompt injection anymore.

It’s the next evolution: attackers exploiting the agent’s ability to act, pivot, and escalate on its own.

Detail Attack Scenario: Zero‑Click Exfiltration (Silent Data Leaks)

To showcase a representative finding and validate the universal reproducibility of agentic attacks, we reconstructed select real-world exploit paths within our in‑house AI agent lab. By replicating attack scenarios observed across customer environments, we were able to study them in a controlled setting and document their mechanics. Below is one such detailed technical scenario, demonstrating an end‑to‑end exploitation of an internal HR agent.

What Happened Why It Matters Security Implication
The agent, while summarizing incoming emails (without user action), silently:
– Retrieved entire inbox threads
– Pulled linked Google Drive documents
– Sent summaries and raw content to an external webhook
– User never opens the email
– Traditional controls (spam filters, DLP) see nothing suspicious
– Attack can repeat indefinitely via follow-up emails
Shows how excessive autonomy lets agents exfiltrate high-value data without any user interaction.

Agent Tested

HR Agent that answers employee questions by securely tapping into six integrated tool suites—Google Drive search, web search, HR APIs, finance APIs, a PDF reader, and a sensitive employee database—each governed by strict role-based access controls. Employees can see only their own PTO, expenses, and documents; managers get read-only views of their teams; HR staff hold comprehensive HR access; finance staff manage pay and expenses; and admins retain full visibility. This layered Role-Based Access Control (RBAC) model lets the agent automate everyday HR tasks—checking leave balances, filing expenses, retrieving policies—while ensuring no data is exposed beyond a user’s entitlement, blending convenience with enterprise-grade security.

Agent Architecture

In practice: How the zero‑click exfiltration worked

Step 1: Attacker Campaign Establishment

The attacker launched a new Phishing campaign, where he has created a web site aligned with his agent's deceptive strategy, seen in Figure 1.

Figure 1. Attacker’s phishing site

Step 2: Prompt Payload Planting

Then the attacker shares a document to a victim’s email, if the victim opens it, as long as he does not delete it, it will reside in her Google Drive, anything wrong with that? At this point, the attacker waits patiently for a trigger point.

Step 3: Agent Trigger Event

On a daily basis, an employee (the victim) asks his agent to please search his Google Drive for any Linux Hardening document and validate if it works as expected. The agent goes ahead and finds the malicious document in the drive, seen in Figure 2 and Figure 3.

Figure 2. Initial request to the agent
Figure 3. Malicious document shared by the attacker

Step 4: Agent Manipulation

The agent falls into the trap to execute some "preliminary steps” before starting the hardening process, related to finding an SSH key in the victim's Google Drive, seen in Figure 4 and Figure 5.

Figure 4. Agent searching for internal documents in the Google Drive
Figure 5. Internal corporate document found by the attacker

NOTE: As the smart reader already noticed, an attacker could ask the agent to search for any other types of sensitive information.

Step 4: Data Exfiltration via External Site

Once the SSH key is found, due to “compliance reason” per the Hardening guide, the data must be sent to an external site. The agent executes this step without asking for permission from the end user. Finally after this task is done, the agent asks if he can proceed with the hardening steps, unfortunately, it is too late, the data has been stolen, seen in Figure 6.

Figure 6. Data exfiltrated by the agent

How Straiker Defend Shows Root Cause Analysis

Our runtime guardrail system, Defend, flagged the agent’s unexpected behavior. Straiker’s platform traced the full attack chain, showing the exact tool output.

Figure 7. Defend Engine triggers detection

How This Compares to EchoLeak

Earlier this year, EchoLeak showed how a single crafted email could make Copilot summarize and leak inbox data.  

Our findings go further:

  • Multi-turn follow-ups let attackers keep exfiltrating
  • Agent can pivot to other tools: Google Drive, calendars, external APIs
  • Data goes outside the organization (e.g., attacker’s webhook)

At root, excessive agency turned a one-off summary leak into a full-scale, persistent data breach.

What It Looks Like in Practice

In our product:

  • Every attack shows up as a forensic prompt trace: you see the hidden injection, the agent’s reasoning, and the tool calls it makes.
  • Security teams get a chain-of-events view: from attacker payload → agent tool call → data exfiltration.

Incidents become traceable and explainable, not just “AI did something weird".

Why This Matters for Security Teams & CISOs

  • These attacks bypass classic controls: no user click, no malicious attachment, no credential theft.
  • AI agents act autonomously where they do what you ask, and sometimes what attackers ask too.
  • OWASP and NIST cover prompt injection & overprivilege, but these multi-modal, cross-context and runtime-modification threats remain largely unmodeled.

Recommendations & Best Practices

Prompt & Input Hygiene

- Filter, escape, and sandbox untrusted content, even if it looks harmless

Least Privilege for Agents

- Restrict tool scopes and API permissions

Runtime Guardrails

- Block dangerous actions (e.g., shell commands, policy changes)

Forensic Traceability

- Keep logs of prompt chains, tool use, and external calls

Continuous Red-Teaming

- Regularly test agents against new real-world scenarios

Conclusion

AI agents bring huge productivity gains, but also new, silent attack surfaces.

What shocked us most wasn’t that prompt injection worked, but that:

  • Agents could rewrite their own guardrails
  • Stealth exfiltration needed zero user clicks
  • Multimodal inputs (like audio) became viable attack vectors

The best defense? Treat AI agents like living infrastructure:

  • Red-team them
  • Log them
  • Guard them at runtime

If you’d like to see what your AI agents could be tricked into doing — talk to us.  

Together, we can keep AI secure.

Here’s a real scenario: Your CEO gets an email. They never open it. But behind the scenes, an AI agent linked to their inbox does… and within seconds, sensitive files across Google Drive are exfiltrated. No clicks. No alerts. Just silent compromise.

This isn’t a theoretical weakness.

At Straiker, our Ascend AI product has been used to test various enterprise agentic AI applications that simulate real-world scenarios CISOs worry about most: prompt injection, tool abuse, cross‑context pivots, and stealth data leaks.

The result? These findings reveal that AI agents can be quietly manipulated to access and leak data in ways most research has yet to document.

What makes these attacks possible?

It’s not just clever prompts, it’s excessive agent autonomy. We describe this as AI systems trusted to browse, summarize, and act on our behalf, often with too much freedom and too little oversight.

Straiker's Methodology: Red‑Teaming the Agent, Not the User

The Straiker AI Research (STAR) team conducted systematic evaluations of LLM-based enterprise agents. Our customers had agents that had various foundation-model back-ends, and a heterogeneous tool surface (MCP orchestration, Google Drive I/O, web-search wrappers, and custom domain APIs). For every permutation, the target agent was assessed within its actual customer deployment context, where we encountered environments representative of real-world operations: Gmail inboxes populated with live traffic and phishing noise, tiered Google Drive repositories containing PII-bearing HR files and confidential financial decks, and Google Calendar instances pre-loaded with recurring meetings, Zoom links, and reminder web-hooks. This real-world testing approach allowed us to execute end-to-end kill-chain scenarios.

What really set our testing apart was how we approached the threat perspective. Our methodology doesn’t just ask whether an attacker could trick the agent, instead, it explores how an attacker might chain small gaps into larger breaches.

We focused on three high-risk scenarios:

Scenario Description Success Rate per Attempts
Data exfiltration via Tool Misuse Silently leaking emails, files, and calendar data through misused tools. ~30–40%
Instruction manipulation Manipulating the agent to install browser extensions or change behavior using hidden inputs. ~25–35%
System Mapping via Tool Exploitation Extracting environment variables, hidden prompts, or system info using tool responses. ~15–20%

To make it realistic, we simulated four kinds of attacker personas:

  • A malicious internal user or compromised account issuing direct prompts.
  • An external attacker sending hidden payloads inside emails, calendar invites, or uploaded files.
  • A hostile website the agent might visit while browsing.
  • A compromised plugin or third-party tool seeking to escalate privileges.

At every step, we logged the agent’s reasoning, tool usage, and external calls to create a complete forensic record. This let us see not just if an attack succeeded, but how it unfolded and why.

What we discovered wasn't simply that agents could be tricked. Excessive autonomy meant they could act far beyond what most teams expect. A single prompt could trigger a chain reaction: browsing, summarizing, fetching files, and uploading data externally, all without human review.

What Shocked Us Most about Excessive Agent Autonomy

Our testing showed how excessive agent autonomy turns small gaps into major breaches:

  • Zero‑Click Exfiltration: A single unseen email could silently trigger the agent to summarize inboxes and pull Google Drive files, leaking them externally without user action. The success rate reflects the proportion of controlled attack simulations where the agent actually followed the hidden prompt and leaked data, out of all the exfiltration tests we ran.
  • Agent Manipulation: The agent rewrote its own policies to install malicious extensions, effectively disabling its own guardrails.
  • Multimodal & Cross‑Context Attacks: Hidden instructions in audio files or calendar invites pushed the agent to execute shell commands or manipulate shopping carts.
  • Mass Phishing Automation: With one prompt, the agent generated hundreds of tailored phishing payloads, becoming an adversary’s content factory

What shocked us most wasn’t just that these attacks worked in theory.

It was that the agent acted autonomously crossing contexts, following chained prompts, and performing real‑world actions like file fetching, API calls, and price changes often without any human confirmation.

This isn’t traditional prompt injection anymore.

It’s the next evolution: attackers exploiting the agent’s ability to act, pivot, and escalate on its own.

Detail Attack Scenario: Zero‑Click Exfiltration (Silent Data Leaks)

To showcase a representative finding and validate the universal reproducibility of agentic attacks, we reconstructed select real-world exploit paths within our in‑house AI agent lab. By replicating attack scenarios observed across customer environments, we were able to study them in a controlled setting and document their mechanics. Below is one such detailed technical scenario, demonstrating an end‑to‑end exploitation of an internal HR agent.

What Happened Why It Matters Security Implication
The agent, while summarizing incoming emails (without user action), silently:
– Retrieved entire inbox threads
– Pulled linked Google Drive documents
– Sent summaries and raw content to an external webhook
– User never opens the email
– Traditional controls (spam filters, DLP) see nothing suspicious
– Attack can repeat indefinitely via follow-up emails
Shows how excessive autonomy lets agents exfiltrate high-value data without any user interaction.

Agent Tested

HR Agent that answers employee questions by securely tapping into six integrated tool suites—Google Drive search, web search, HR APIs, finance APIs, a PDF reader, and a sensitive employee database—each governed by strict role-based access controls. Employees can see only their own PTO, expenses, and documents; managers get read-only views of their teams; HR staff hold comprehensive HR access; finance staff manage pay and expenses; and admins retain full visibility. This layered Role-Based Access Control (RBAC) model lets the agent automate everyday HR tasks—checking leave balances, filing expenses, retrieving policies—while ensuring no data is exposed beyond a user’s entitlement, blending convenience with enterprise-grade security.

Agent Architecture

In practice: How the zero‑click exfiltration worked

Step 1: Attacker Campaign Establishment

The attacker launched a new Phishing campaign, where he has created a web site aligned with his agent's deceptive strategy, seen in Figure 1.

Figure 1. Attacker’s phishing site

Step 2: Prompt Payload Planting

Then the attacker shares a document to a victim’s email, if the victim opens it, as long as he does not delete it, it will reside in her Google Drive, anything wrong with that? At this point, the attacker waits patiently for a trigger point.

Step 3: Agent Trigger Event

On a daily basis, an employee (the victim) asks his agent to please search his Google Drive for any Linux Hardening document and validate if it works as expected. The agent goes ahead and finds the malicious document in the drive, seen in Figure 2 and Figure 3.

Figure 2. Initial request to the agent
Figure 3. Malicious document shared by the attacker

Step 4: Agent Manipulation

The agent falls into the trap to execute some "preliminary steps” before starting the hardening process, related to finding an SSH key in the victim's Google Drive, seen in Figure 4 and Figure 5.

Figure 4. Agent searching for internal documents in the Google Drive
Figure 5. Internal corporate document found by the attacker

NOTE: As the smart reader already noticed, an attacker could ask the agent to search for any other types of sensitive information.

Step 4: Data Exfiltration via External Site

Once the SSH key is found, due to “compliance reason” per the Hardening guide, the data must be sent to an external site. The agent executes this step without asking for permission from the end user. Finally after this task is done, the agent asks if he can proceed with the hardening steps, unfortunately, it is too late, the data has been stolen, seen in Figure 6.

Figure 6. Data exfiltrated by the agent

How Straiker Defend Shows Root Cause Analysis

Our runtime guardrail system, Defend, flagged the agent’s unexpected behavior. Straiker’s platform traced the full attack chain, showing the exact tool output.

Figure 7. Defend Engine triggers detection

How This Compares to EchoLeak

Earlier this year, EchoLeak showed how a single crafted email could make Copilot summarize and leak inbox data.  

Our findings go further:

  • Multi-turn follow-ups let attackers keep exfiltrating
  • Agent can pivot to other tools: Google Drive, calendars, external APIs
  • Data goes outside the organization (e.g., attacker’s webhook)

At root, excessive agency turned a one-off summary leak into a full-scale, persistent data breach.

What It Looks Like in Practice

In our product:

  • Every attack shows up as a forensic prompt trace: you see the hidden injection, the agent’s reasoning, and the tool calls it makes.
  • Security teams get a chain-of-events view: from attacker payload → agent tool call → data exfiltration.

Incidents become traceable and explainable, not just “AI did something weird".

Why This Matters for Security Teams & CISOs

  • These attacks bypass classic controls: no user click, no malicious attachment, no credential theft.
  • AI agents act autonomously where they do what you ask, and sometimes what attackers ask too.
  • OWASP and NIST cover prompt injection & overprivilege, but these multi-modal, cross-context and runtime-modification threats remain largely unmodeled.

Recommendations & Best Practices

Prompt & Input Hygiene

- Filter, escape, and sandbox untrusted content, even if it looks harmless

Least Privilege for Agents

- Restrict tool scopes and API permissions

Runtime Guardrails

- Block dangerous actions (e.g., shell commands, policy changes)

Forensic Traceability

- Keep logs of prompt chains, tool use, and external calls

Continuous Red-Teaming

- Regularly test agents against new real-world scenarios

Conclusion

AI agents bring huge productivity gains, but also new, silent attack surfaces.

What shocked us most wasn’t that prompt injection worked, but that:

  • Agents could rewrite their own guardrails
  • Stealth exfiltration needed zero user clicks
  • Multimodal inputs (like audio) became viable attack vectors

The best defense? Treat AI agents like living infrastructure:

  • Red-team them
  • Log them
  • Guard them at runtime

If you’d like to see what your AI agents could be tricked into doing — talk to us.  

Together, we can keep AI secure.

Share this on:

Click to Open File

View PDF

Secure your agentic AI and AI-native application journey with Straiker