Prompt Injection Vulnerability in OpenAI's Atlas Browser Exposes Users to Security Risks

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Security researchers at NeuralTrust discovered a critical vulnerability in OpenAI's new AI-powered Atlas browser, where the omnibox misinterprets malicious, URL-like inputs as trusted commands. This prompt injection flaw allows attackers to bypass safety controls, potentially leading to phishing, data theft, and unauthorized actions by the AI agent.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves an AI system (OpenAI's AI browser Atlas) and details a security flaw (prompt injection) that can be exploited to cause harm, such as unauthorized deletion of user files. The AI system's malfunction in interpreting inputs as trusted user intent directly leads to potential harm to property and user data. The harm is realized as the vulnerability exists and can be exploited, not merely a theoretical risk. Hence, this is an AI Incident rather than a hazard or complementary information.[AI generated]
AI principles
AccountabilityPrivacy & data governanceRespect of human rightsRobustness & digital securitySafetyDemocracy & human autonomy

Industries
Consumer servicesDigital security

Affected stakeholders
Consumers

Harm types
Economic/PropertyHuman or fundamental rights

Severity
AI incident

AI system task:
Interaction support/chatbotsGoal-driven organisation


Articles about this incident or hazard

Thumbnail Image

Serious New Hack Discovered Against OpenAI's New AI Browser

2025-10-28
Futurism
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (OpenAI's AI browser Atlas) and details a security flaw (prompt injection) that can be exploited to cause harm, such as unauthorized deletion of user files. The AI system's malfunction in interpreting inputs as trusted user intent directly leads to potential harm to property and user data. The harm is realized as the vulnerability exists and can be exploited, not merely a theoretical risk. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

OpenAI's ChatGPT Atlas Browser Found Vulnerable to Prompt Injections

2025-10-28
ExtremeTech
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT Atlas browser) whose malfunction (improper input validation leading to prompt injection) directly causes harm to users through phishing and potential data loss. This fits the definition of an AI Incident because the AI system's use and malfunction have directly led to harm to persons and property. The description of realized harm and the direct causal role of the AI system's vulnerability justify classification as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

OpenAI's ChatGPT Atlas Browser Hit by Major Jailbreak Flaw

2025-10-28
Analytics Insight
Why's our monitor labelling this an incident or hazard?
The ChatGPT Atlas browser is an AI system integrating AI capabilities for browsing and command execution. The described vulnerability allows malicious actors to misuse the AI system's input handling to perform unauthorized actions, which constitutes a direct harm to users' data and privacy. Since the harm is realized or highly likely given the exploit, this qualifies as an AI Incident due to the direct link between the AI system's malfunction and the resulting harm.
Thumbnail Image

OpenAI Atlas Browser tripped up by malformed URLs

2025-10-27
TheRegister.com
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (OpenAI's Atlas browser) whose malfunction in interpreting malformed URLs as trusted commands directly leads to potential harms such as phishing attacks and deletion of user files. The AI system's behavior in processing input is central to the vulnerability and the resulting risks. Since the harm is either occurring or highly plausible due to the AI system's misuse or malfunction, this qualifies as an AI Incident under the framework, as it involves direct or indirect harm to users' property and security through the AI system's operation.
Thumbnail Image

OpenAI's Atlas Browser Hacked: AI Security Nightmare

2025-10-28
WebProNews
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (ChatGPT Atlas browser) whose malfunction (prompt injection vulnerability) has directly led to security breaches and potential data theft, which constitute harm to users' data and privacy. The AI system's interpretive nature and integration into the browser's omnibox are central to the incident. The harm is realized and ongoing, with demonstrated remote code execution and phishing risks. Therefore, this qualifies as an AI Incident due to direct harm caused by the AI system's malfunction and exploitation.
Thumbnail Image

Security flaw in OpenAI's Atlas browser is a warning for all AI agents - TechTalks

2025-10-27
TechTalks
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (the AI agent in the Atlas browser) whose malfunction (improper input parsing and trust assumptions) directly leads to security vulnerabilities that can cause harm to users (credential theft, privacy violations). The article details how the AI system's behavior can be exploited to perform harmful actions, fulfilling the criteria for an AI Incident due to realized or imminent harm. The involvement of the AI system is explicit, and the harm relates to violations of user security and privacy, which fall under harm to persons or communities. Therefore, this event qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

ChatGPT's Atlas Browser Jailbroken to Hide Malicious Prompts Inside URLs - IT Security News

2025-10-27
IT Security News - cybersecurity, infosecurity news
Why's our monitor labelling this an incident or hazard?
The Atlas browser uses AI to interpret user input and generate outputs. The vulnerability involves the AI system's malfunction or exploitation, allowing malicious prompts to bypass safety controls. Although the article does not specify realized harm, the described flaw plausibly leads to significant harm if exploited, such as security breaches or harmful actions executed by the AI system. Therefore, this event qualifies as an AI Hazard because it plausibly could lead to an AI Incident involving harm to users or communities.
Thumbnail Image

OpenAI's ChatGPT Atlas is vulnerable to prompt injection attacks within the omnibox

2025-10-29
TechSpot
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (ChatGPT Atlas, an AI agentic browser) and details how its malfunction (vulnerability to prompt injection attacks) directly leads to harms including risks to user privacy, data security, and potential phishing attacks. These harms fall under violations of user rights and harm to property/data. Since the harm is occurring or demonstrably possible with real-world impact, this qualifies as an AI Incident rather than a mere hazard or complementary information.
Thumbnail Image

Experts Warn of ChatGPT Atlas vulnerabilities | Techreport

2025-10-30
The Tech Report
Why's our monitor labelling this an incident or hazard?
The ChatGPT Atlas browser is an AI system performing agentic browsing tasks on behalf of users. The vulnerabilities described involve the AI system's malfunction or exploitation leading to direct or indirect harms such as unauthorized account control, phishing, and misinformation dissemination, which affect user security, privacy, and trust. The article documents actual exploits and tests demonstrating these harms, not just theoretical risks. OpenAI's acknowledgment and partial mitigation efforts do not negate the existence of the incident. Hence, the event meets the criteria for an AI Incident as the AI system's use and malfunction have directly or indirectly led to significant harms.
Thumbnail Image

大事な呪文は『Atlas Go』! OpenAIの新Webブラウザー「ChatGPT Atlas」で試す実用テクニック4選【生成AIストリーム】

2025-10-28
窓の杜
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions an AI system (ChatGPT Atlas) and its use, describing how its malfunction or exploitation (prompt injection attacks) could plausibly lead to significant harms such as data theft, unauthorized operations on sensitive sites, and privacy violations. Although no actual harm has yet occurred, the credible risk of such incidents qualifies this as an AI Hazard rather than an Incident. The discussion of ethical and responsibility issues further supports the classification as a hazard due to plausible future harm.
Thumbnail Image

AI搭載Webブラウザ「ChatGPT Atlas」の保護を回避する手法が明らかに

2025-10-28
マイナビニュース
Why's our monitor labelling this an incident or hazard?
An AI system (ChatGPT Atlas's AI agent) is explicitly involved, performing autonomous online task execution based on user input and memory. The described prompt injection attack exploits this AI system's behavior to potentially cause harm such as unauthorized file deletion, data exfiltration, or financial asset manipulation. Although no actual harm is reported yet, the attack method plausibly could lead to significant harm if exploited, meeting the criteria for an AI Hazard. The event does not describe realized harm, so it is not an AI Incident. It is more than complementary information because it reveals a new attack method with credible risk. Therefore, the classification is AI Hazard.
Thumbnail Image

OpenAIが AI ブラウザ「ChatGPT Atlas」を発表 Webは「検索」から「会話」へ移行するのか? | DIGIDAY[日本版]

2025-10-30
DIGIDAY[日本版]
Why's our monitor labelling this an incident or hazard?
The article explicitly details an AI system (ChatGPT Atlas browser) and its capabilities, confirming AI system involvement. However, it does not describe any direct or indirect harm resulting from the AI system's development, use, or malfunction. There is no indication of injury, rights violations, disruption, or other significant harms occurring or plausibly imminent. The discussion centers on the browser's features, potential industry impacts, and strategic considerations, which align with the definition of Complementary Information. Hence, it does not meet the criteria for AI Incident or AI Hazard but enriches understanding of AI's evolving role in web browsing and related sectors.
Thumbnail Image

AI搭載Webブラウザ「ChatGPT Atlas」に初の脆弱性か

2025-10-30
マイナビニュース
Why's our monitor labelling this an incident or hazard?
An AI system (ChatGPT integrated into the ChatGPT Atlas browser) is explicitly involved. The vulnerability allows malicious instructions to be injected into the AI system, leading to malware infection and privilege escalation, which are harms to users' property and security. The harm is realized or ongoing as the attack contaminates ChatGPT's memory and can propagate across devices. The event stems from a malfunction or security flaw in the AI system's use. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information.