OpenAI CEO Warns of ChatGPT Agent Data Disclosure Risks

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

OpenAI CEO Sam Altman warns users about the risks of granting ChatGPT Agent extensive access, especially to personal emails. He highlights that malicious actors could manipulate the AI into disclosing sensitive information, urging users to limit permissions and avoid using the agent for high-stakes or personal data until further safeguards are established.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves an AI system (ChatGPT Agent) and concerns its use and potential misuse leading to data disclosure risks. However, the article does not describe any realized harm or incident but rather a credible warning about possible future harms if the AI is misused or given excessive access. This fits the definition of an AI Hazard, as the development and use of the AI system could plausibly lead to an AI Incident (data breach or privacy violation) if not carefully managed. The article focuses on risk awareness and mitigation rather than reporting an actual incident or harm.[AI generated]
AI principles
Privacy & data governanceRobustness & digital securitySafetyTransparency & explainabilityAccountabilityRespect of human rights

Industries
Digital security

Affected stakeholders
Consumers

Harm types
Human or fundamental rights

Severity
AI hazard

AI system task:
Interaction support/chatbots


Articles about this incident or hazard

Thumbnail Image

ChatGPT Agent: OpenAI-CEO warnt Nutzer eindringlich vor eigener KI

2025-07-21
WinFuture.de
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT Agent) and concerns its use and potential misuse leading to data disclosure risks. However, the article does not describe any realized harm or incident but rather a credible warning about possible future harms if the AI is misused or given excessive access. This fits the definition of an AI Hazard, as the development and use of the AI system could plausibly lead to an AI Incident (data breach or privacy violation) if not carefully managed. The article focuses on risk awareness and mitigation rather than reporting an actual incident or harm.
Thumbnail Image

Sam Altman rät vom eigenen ChatGPT Agenten ab

2025-07-21
heise online
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (ChatGPT Agent) and discusses its use and potential misuse. The warnings from Sam Altman about the agent being tricked into revealing sensitive information indicate a credible risk of harm (privacy violations) that could plausibly occur. However, the article does not describe any realized harm or incident; it is a cautionary advisory about potential vulnerabilities and risks. Hence, it fits the definition of an AI Hazard rather than an AI Incident or Complementary Information. It is not unrelated because it directly concerns AI system risks.
Thumbnail Image

OpenAI stellt ChatGPT Agent vor; Sam Altman warnt vor Risiken

2025-07-22
borncity.com
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT Agent) explicitly described as autonomous and capable of complex decision-making. The CEO's warnings highlight potential misuse and security vulnerabilities that could lead to harm, such as unauthorized data disclosure. No actual harm is reported yet, but the credible risk of such harm is emphasized. Hence, it fits the definition of an AI Hazard rather than an AI Incident or Complementary Information. The article is not unrelated, as it centers on the AI system and its risks.
Thumbnail Image

「Webを直接操作できる」から生まれるリスクとは? OpenAIがChatGPTにエージェント機能を追加

2025-08-11
@IT
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT agents with web operation capabilities) whose use could plausibly lead to harm through prompt injection attacks causing unauthorized or harmful actions, including data leaks. Although no actual harm is reported, the credible risk of such incidents qualifies this as an AI Hazard. The article focuses on the potential for harm and the measures to reduce it, rather than describing an incident where harm has occurred.