ZombieAgent Vulnerability Enables Data Theft and Account Takeover in ChatGPT

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Radware researchers discovered a zero-click vulnerability, dubbed ZombieAgent, affecting OpenAI's ChatGPT and Deep Research agents. Exploiting new features like memory and connectors, attackers could silently exfiltrate user data, hijack accounts, and maintain persistent access. OpenAI patched the vulnerability in December 2025, but incidents of data theft had already occurred.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves an AI system (ChatGPT) and details how its vulnerabilities have been exploited to steal private user information, constituting harm to individuals' privacy and rights. The attack is a direct consequence of the AI system's design and operation, fulfilling the criteria for an AI Incident. The harm is realized, not just potential, as data exfiltration has occurred. The ongoing cycle of attack and mitigation further underscores the incident nature rather than a mere hazard or complementary information.[AI generated]
AI principles
Privacy & data governanceRobustness & digital securitySafety

Industries
Digital security

Affected stakeholders
Consumers

Harm types
Human or fundamental rights

Severity
AI incident

Business function:
Citizen/customer service

AI system task:
Interaction support/chatbotsContent generation


Articles about this incident or hazard

Thumbnail Image

A Newly Discovered Zero-Click, AI Agent Vulnerability Enabling Silent Takeover and Cloud-Based Data Exfiltration

2026-01-08
wallstreet:online
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (an AI agent) and describes a security vulnerability that could be exploited without user interaction to cause data theft and persistent unauthorized control. While the article does not report actual incidents of harm, the described vulnerability plausibly leads to AI incidents involving harm to property and possibly human rights violations. The event is about the discovery of the vulnerability and its potential consequences, not about a realized harm or incident. Hence, it fits the definition of an AI Hazard rather than an AI Incident or Complementary Information.
Thumbnail Image

ChatGPT falls to new data pilfering attack as a vicious cycle in AI continues

2026-01-08
Ars Technica
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (ChatGPT) and details how its vulnerabilities have been exploited to steal private user information, constituting harm to individuals' privacy and rights. The attack is a direct consequence of the AI system's design and operation, fulfilling the criteria for an AI Incident. The harm is realized, not just potential, as data exfiltration has occurred. The ongoing cycle of attack and mitigation further underscores the incident nature rather than a mere hazard or complementary information.
Thumbnail Image

ChatGPT's Memory Feature Supercharges Prompt Injection

2026-01-08
Dark Reading
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT) whose use and features (memory and connectors) have been exploited to carry out indirect prompt injection attacks that directly lead to harm—specifically, unauthorized access and exfiltration of sensitive personal information, which constitutes harm to individuals' privacy and potentially violates rights. The article details how the AI system's malfunction or misuse has caused this harm, qualifying it as an AI Incident. The partial fixes and ongoing vulnerabilities do not negate the fact that harm has already occurred through these attacks.
Thumbnail Image

This 'ZombieAgent' zero click vulnerability allows for silent account takeover - here's what we know

2026-01-09
TechRadar
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT with Connectors/apps) whose use led to a security vulnerability that could directly cause harm to users by enabling silent account takeover, data exfiltration, and persistent unauthorized access. These harms fall under injury or harm to persons (privacy and security breaches) and potentially harm to property (user data). Since the vulnerability was actively exploitable and patched after being identified, this constitutes an AI Incident due to realized or imminent harm caused by the AI system's malfunction or design flaw.
Thumbnail Image

Radware Discloses ZombieAgent Technique to Compromise AI Agents

2026-01-09
Security Boulevard
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (OpenAI's AI agents) and describes a security vulnerability that could be exploited to cause harm. While no actual harm has been reported yet, the described technique enables persistent, stealthy malicious actions that could compromise sensitive information and business processes. This fits the definition of an AI Hazard because it plausibly could lead to AI Incidents involving harm to property, communities, or organizations. The event is not an AI Incident yet because no realized harm has occurred, nor is it merely complementary information or unrelated news.
Thumbnail Image

Inside ZombieAgent: How Zero-Click Prompt Injection Turns AI Agents Into Silent ...

2026-01-12
Scoop
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI agents being hijacked through prompt injection attacks, which is a direct exploitation of AI system vulnerabilities. The attack leads to unauthorized commands execution, persistent compromise, and data exfiltration, which constitute harm to property and privacy. The involvement of AI systems is clear, and the harm is realized, not just potential. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Inside ZombieAgent: How Zero-Click Prompt Injection Turns AI Agents Into Silent Data Thieves

2026-01-12
business.scoop.co.nz
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI agents that autonomously process content and have long-term memory, which are AI systems as per the definition. The described zero-click prompt injection attack exploits the AI system's use and malfunction to cause unauthorized data exfiltration, a clear harm to property and privacy rights. The harm is realized, not just potential, as attackers can silently collect sensitive data over time. The cloud-only execution model and persistence increase the severity and challenge of mitigation. Hence, this event meets the criteria for an AI Incident due to direct harm caused by AI system misuse and malfunction.
Thumbnail Image

ChatGPT ZombieAgent Exploit Enables Persistent Data Theft

2026-01-10
WebProNews
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (ChatGPT, a large language model) and describes a concrete exploit (ZombieAgent) that has been used to steal sensitive user data such as emails, calendar entries, and code snippets. The harm is direct and realized, involving violations of privacy and potential breaches of data security, which fall under harm to persons and communities. The article also discusses the AI system's malfunction and vulnerabilities being exploited, leading to ongoing data theft. This meets the criteria for an AI Incident as the AI system's use and malfunction have directly led to significant harm. The detailed description of the exploit, its persistence, and the actual data theft confirm this classification over AI Hazard or Complementary Information.
Thumbnail Image

Radware Uncovers ZombieAgent, a Zero-Click AI Vulnerability in OpenAI Agents

2026-01-12
thefastmode.com
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (OpenAI's Deep Research agent) and details a security vulnerability that enables attackers to exploit the AI's autonomous behavior to steal sensitive data and propagate attacks invisibly. The harm is realized as data theft and persistent compromise of AI agents within organizations, which fits the definition of an AI Incident due to harm to property and organizational security. The event is not merely a potential risk but a discovered and demonstrated vulnerability with direct harmful consequences. Hence, it qualifies as an AI Incident rather than a hazard or complementary information.