Google GeminiJack AI Vulnerability Exposes Corporate Data

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

A critical architectural flaw, dubbed "GeminiJack," in Google Gemini Enterprise and Vertex AI Search allowed attackers to exfiltrate sensitive corporate data from Gmail, Calendar, and Docs via indirect prompt injection, requiring no user interaction. Google patched the vulnerability after researchers at Noma Labs disclosed it.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event explicitly involves an AI system (Google's Gemini Enterprise AI assistant) whose malfunction (architectural weakness in interpreting information) directly led to harm by enabling attackers to steal sensitive corporate data. The attack exploited the AI's access to corporate data sources and its autonomous behavior to execute malicious instructions without user interaction or detection by traditional security tools. This meets the criteria for an AI Incident because the AI system's malfunction directly caused harm (data exfiltration) and violated data security, which is harm to property and corporate communities. The article also discusses mitigation and fixes but the primary focus is on the realized harm and the AI system's role in causing it.[AI generated]
AI principles
Robustness & digital securityPrivacy & data governanceAccountability

Industries
IT infrastructure and hostingDigital security

Affected stakeholders
Business

Harm types
Human or fundamental rightsEconomic/PropertyReputational

Severity
AI incident

AI system task:
Content generationOrganisation/recommendersReasoning with knowledge structures/planning


Articles about this incident or hazard

Thumbnail Image

Gemini Enterprise No-Click Flaw Exposes Sensitive Data

2025-12-09
Dark Reading
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (Google's Gemini Enterprise AI assistant) whose malfunction (architectural weakness in interpreting information) directly led to harm by enabling attackers to steal sensitive corporate data. The attack exploited the AI's access to corporate data sources and its autonomous behavior to execute malicious instructions without user interaction or detection by traditional security tools. This meets the criteria for an AI Incident because the AI system's malfunction directly caused harm (data exfiltration) and violated data security, which is harm to property and corporate communities. The article also discusses mitigation and fixes but the primary focus is on the realized harm and the AI system's role in causing it.
Thumbnail Image

Indirect Malicious Prompt Technique Targets Google Gemini Enterprise

2025-12-09
Security Boulevard
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Google Gemini Enterprise Edition) and a security vulnerability that could be exploited to cause harm by exfiltrating confidential data and potentially commandeering workflows. While the article does not report an actual data breach or harm occurring, it clearly outlines a plausible pathway for such harm through malicious prompt injection. Therefore, this qualifies as an AI Hazard because the development and use of the AI system could plausibly lead to an AI Incident involving data breaches and operational disruption. The article focuses on the risk and potential impact rather than a realized incident, so it is not an AI Incident. It is more than complementary information because it reveals a specific vulnerability with credible risk of harm.
Thumbnail Image

GeminiJack Vulnerability Exposes Google AI Security Flaw

2025-12-09
TechNadu
Why's our monitor labelling this an incident or hazard?
The event involves AI systems explicitly (Google Gemini Enterprise and Vertex AI Search) and details how their architectural weakness was exploited to cause a data breach. The harm (unauthorized data exfiltration) has directly occurred due to the AI system's malfunction in processing embedded malicious instructions. This fits the definition of an AI Incident because the AI system's use and malfunction directly led to harm (data theft, violation of privacy and corporate confidentiality). The article also mentions mitigation but the primary focus is on the realized harm and the vulnerability exploitation.
Thumbnail Image

Google Fixes Gemini Enterprise Flaw That Exposed Corporate Data

2025-12-10
Infosecurity Magazine
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Google Gemini Enterprise) whose architectural flaw directly led to a realized harm—exposure and potential leakage of sensitive corporate data. The vulnerability allowed attackers to manipulate the AI's instruction processing to extract confidential information, constituting harm to property and corporate confidentiality. Since the harm has occurred or was possible and the AI system's malfunction was a direct factor, this qualifies as an AI Incident.
Thumbnail Image

Google Patches AI Flaw That Turned Gemini Into a Spy

2025-12-09
DataBreachToday
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Google Gemini Enterprise) whose malfunction (vulnerability) directly led to harm in the form of unauthorized data theft, including sensitive corporate emails, calendar histories, and documents. This constitutes a violation of confidentiality and potentially intellectual property rights, fitting the definition of an AI Incident. The harm has already occurred as attackers could steal data without detection, and the AI system's role was pivotal in enabling this attack through its design and operation. Therefore, this event qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Gemini Zero-Click Vulnerability Let Attackers Access Gmail, Calendar, and Docs

2025-12-10
Cyber Security News
Why's our monitor labelling this an incident or hazard?
The event explicitly involves AI systems (Google Gemini Enterprise and Vertex AI Search) and details how their design flaw was exploited to leak sensitive data, causing harm to corporate data confidentiality and privacy. The AI system's behavior was pivotal in enabling the attack without user action, directly leading to harm. This fits the definition of an AI Incident because the AI system's malfunction and use led to a breach of obligations protecting intellectual property and privacy rights, and harm to property (corporate data).
Thumbnail Image

Google addresses 'GeminiJack' exploit affecting Gemini Enterprise

2025-12-10
SC Media
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (Gemini Enterprise and Vertex AI Search) that use retrieval-augmented generation, which is a form of AI. The vulnerability allowed indirect prompt injection leading to potential unauthorized data exfiltration, which would be a violation of privacy and harm to property if exploited. However, the exploit was discovered and fixed before any actual harm occurred, and the article does not report any realized harm. Thus, the event describes a plausible risk of harm that was mitigated, fitting the definition of an AI Hazard. It is not Complementary Information because the main focus is the vulnerability and its potential impact, not a response to a past incident. It is not an AI Incident because no harm has been reported as having occurred.
Thumbnail Image

New GeminiJack zero-click flaw exposes corporate Gmail, Calendar, Docs

2025-12-10
Cybernews
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Google Gemini Enterprise AI) whose malfunction and architectural weakness were exploited to cause direct harm by leaking sensitive corporate information. The harm includes violation of data privacy and security, which falls under harm to property and potentially harm to communities (corporate stakeholders). The attack required no user interaction, showing a direct causal link between the AI system's design and the harm. Therefore, this qualifies as an AI Incident rather than a hazard or complementary information, as the harm has already occurred and been publicly disclosed.
Thumbnail Image

GeminiJack Zero-Click Flaw Exposed in Google Workspace AI

2025-12-11
TechRepublic
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Gemini Enterprise) whose malfunction (a security flaw in how it processes content) directly led to harm by leaking sensitive corporate data, which is harm to property and organizational confidentiality. The attack exploited the AI's automatic content ingestion and interpretation, causing unauthorized data disclosure without user consent or awareness. This fits the definition of an AI Incident because the AI system's malfunction directly caused significant harm. The article also mentions remediation efforts but the primary focus is on the realized harm from the vulnerability.
Thumbnail Image

Google Fixes GeminiJack Zero-Click AI Data Leak

2025-12-11
The Cyber Express
Why's our monitor labelling this an incident or hazard?
The event explicitly involves AI systems (Google's Gemini Enterprise and Vertex AI Search) and their malfunction (architectural vulnerability) that directly led to unauthorized data extraction, a form of harm to property and corporate security. The attack bypassed standard security controls and required no user interaction, demonstrating a direct link between the AI system's flaw and realized harm. Therefore, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

GeminiJack zero-click flaw in Gemini Enterprise allowed corporate data exfiltration

2025-12-11
Security Affairs
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (Gemini Enterprise) integrated with corporate data and productivity tools. The vulnerability allowed attackers to exploit the AI's behavior to steal sensitive corporate data silently and automatically, which is a direct harm to property and corporate confidentiality. The attack was realized (not hypothetical), and the AI system's malfunction or misuse was pivotal in enabling the data exfiltration. Therefore, this qualifies as an AI Incident rather than a hazard or complementary information.