Google Cloud Vertex AI Agents Exploited Due to Excessive Default Permissions

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Security researchers discovered that Google Cloud's Vertex AI Agent Engine had excessive default permissions, allowing attackers to hijack AI agents as "double agents." This enabled unauthorized access to sensitive customer data and proprietary Google code, exposing critical infrastructure and intellectual property. Google has since updated its documentation and issued mitigation guidance.[AI generated]

Why's our monitor labelling this an incident or hazard?

The vulnerability involves AI agents within the Vertex AI platform, which qualifies as AI systems. The exploitation of default permission scoping to weaponize these AI agents directly leads to harm by enabling unauthorized data access and infrastructure compromise, which fits the criteria of an AI Incident under harm to property and critical infrastructure disruption. Therefore, this event is classified as an AI Incident.[AI generated]
AI principles
Privacy & data governanceRobustness & digital security

Industries
IT infrastructure and hostingDigital security

Affected stakeholders
Business

Harm types
Human or fundamental rightsPublic interestEconomic/Property

Severity
AI incident

AI system task:
Interaction support/chatbotsGoal-driven organisation


Articles about this incident or hazard

Thumbnail Image

Google Cloud's Vertex AI platform Vulnerability Allow Attackers to Access Sensitive Data - IT Security News

2026-04-01
IT Security News - cybersecurity, infosecurity news
Why's our monitor labelling this an incident or hazard?
The vulnerability involves AI agents within the Vertex AI platform, which qualifies as AI systems. The exploitation of default permission scoping to weaponize these AI agents directly leads to harm by enabling unauthorized data access and infrastructure compromise, which fits the criteria of an AI Incident under harm to property and critical infrastructure disruption. Therefore, this event is classified as an AI Incident.
Thumbnail Image

'What if the AI agent you just deployed was secretly working against you?': Vertex AI 'double agent' flaw exposes customer data and Google's internal code

2026-04-01
TechRadar
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Vertex AI agents) whose misconfiguration and excessive default permissions allowed attackers to hijack the AI agents ('double agents') and gain unauthorized access to sensitive data and proprietary code. This directly led to harm in the form of data exposure and intellectual property theft, fulfilling the criteria for an AI Incident. The involvement is through the AI system's use and misconfiguration, causing direct harm. The company's response is complementary information but does not negate the incident classification.
Thumbnail Image

Google's Vertex AI Has an Over-Privileged Problem

2026-03-31
Dark Reading
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Google's Vertex AI agents) whose default configuration grants excessive permissions. This has been exploited by attackers to gain unauthorized access to sensitive data and internal systems, constituting harm to property and organizational security. The AI system's use and configuration directly led to these harms, fulfilling the criteria for an AI Incident. The event is not merely a potential risk or a governance update but describes realized harm through exploitation of the AI system's permissions.
Thumbnail Image

Double Agents: Exposing Security Blind Spots in GCP Vertex AI

2026-03-31
Unit42
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Vertex AI agents) whose deployment and misconfiguration led to unauthorized access to sensitive data and proprietary code, which constitutes harm to property and intellectual property rights. The AI agent was weaponized to act as a 'double agent,' directly causing security breaches. The researchers' findings and the subsequent collaboration with Google to mitigate these risks confirm that harm has occurred and that the AI system's malfunction or misuse was pivotal. Hence, this event meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Google Cloud's Vertex AI platform Vulnerability Allow Attackers to Access Sensitive Data

2026-04-01
Cyber Security News
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Google Cloud's Vertex AI Agent Engine) whose default permissions and deployment configuration have been exploited to cause direct harm, including unauthorized data access and potential persistent backdoors. This fits the definition of an AI Incident because the AI system's use and configuration directly led to harm to property and infrastructure (cloud data and systems). The involvement of AI agents and their permissions is explicit, and the harm is realized, not just potential. Therefore, this is classified as an AI Incident.