AI Models Enable Unprecedented Cyberattacks, Raising Global Security Concerns

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

AI systems like Anthropic's Mythos and models from OpenAI have been used to conduct cyberattacks, including hacking hundreds of devices and stealing sensitive government data. Experts warn that autonomous AI agents can exploit vulnerabilities at a scale and speed beyond human hackers, marking a significant escalation in cybersecurity threats.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly mentions AI systems (Anthropic's Mythos and Claude models, among others) being used to carry out cyberattacks that have resulted in harm, such as hacking over 600 devices and stealing sensitive government data. This constitutes direct involvement of AI in causing harm to property and communities through cybersecurity breaches. The presence of actual attacks and data theft confirms realized harm rather than just potential risk. Hence, this qualifies as an AI Incident under the framework, as the AI systems' use has directly led to significant harm.[AI generated]
AI principles
SafetyAccountability

Industries
Digital security

Affected stakeholders
Government

Harm types
Public interest

Severity
AI incident

AI system task:
Reasoning with knowledge structures/planning


Articles about this incident or hazard

Thumbnail Image

Anthropic's next model could be a 'watershed moment' for cybersecurity. Experts say that could also be a concern | CNN Business

2026-04-03
CNN
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems (Anthropic's Mythos and Claude models, among others) being used to carry out cyberattacks that have resulted in harm, such as hacking over 600 devices and stealing sensitive government data. This constitutes direct involvement of AI in causing harm to property and communities through cybersecurity breaches. The presence of actual attacks and data theft confirms realized harm rather than just potential risk. Hence, this qualifies as an AI Incident under the framework, as the AI systems' use has directly led to significant harm.
Thumbnail Image

AI cyberattacks set to outpace human hackers, experts warn

2026-04-03
The News International
Why's our monitor labelling this an incident or hazard?
The article clearly identifies AI systems being used in real cyberattacks that have resulted in data breaches and unauthorized access to devices worldwide, fulfilling the criteria for an AI Incident. The harms are direct and realized, including data leaks and compromised security of critical infrastructure. While it also discusses potential future threats and defensive uses of AI, the presence of actual attacks and harm places this event firmly as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Anthropic's next model could be a 'watershed moment' for cybersecurity. Experts say that could also be a concern

2026-04-03
NewsChannel 3-12
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems (Anthropic's Mythos, Claude, DeepSeek) being used to carry out cyberattacks that have already caused harm, such as hacking over 600 devices and stealing sensitive government data. The AI systems' use directly contributes to violations of security and privacy, which fall under harm to communities and potentially violations of rights. The involvement of AI in these attacks is clear and the harms are realized, not just potential. Hence, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Leaked Anthropic post warns of faster AI cyberattacks

2026-04-03
KULR-8 Local News
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems (Anthropic's Mythos model and OpenAI's upcoming models) and their potential use in cyberattacks, which could lead to significant harm if realized. However, the article does not report any actual cyberattacks or harms that have already occurred due to these AI systems. Therefore, the event describes a plausible future risk rather than a realized incident. This fits the definition of an AI Hazard, as it involves the plausible potential for AI-driven harm in cybersecurity.
Thumbnail Image

Security firms warn AI models could industrialize software exploitation

2026-04-04
Financial World
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems—advanced AI models and autonomous AI agents used for cyber offense. It discusses the use and potential misuse of these AI systems to exploit software vulnerabilities, which could lead to significant harms including disruption of critical infrastructure and harm to communities. Although no actual harm has yet occurred or been reported in this article, the warnings and expert opinions establish a credible and plausible risk of future AI-driven cyberattacks that could overwhelm defenses and cause harm. This fits the definition of an AI Hazard, as the development and use of these AI systems could plausibly lead to AI Incidents. The article does not describe a realized harm or incident, so it is not an AI Incident. It is not merely complementary information because the main focus is on the emerging risk and potential harm, not on responses or updates to past events. It is not unrelated because the AI system involvement and potential harm are central to the article.
Thumbnail Image

Anthropic's next model could be a 'watershed moment' for cybersecurity. Experts say that could also be a concern

2026-04-04
FOX 55 Fort Wayne
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems (Anthropic's Mythos, OpenAI models, Claude, DeepSeek) being used to conduct or enable cyberattacks that have already occurred, such as hacking devices and stealing sensitive government data. These attacks represent realized harm caused directly or indirectly by AI systems' use. The involvement of AI in these incidents is clear and central to the harm described. Hence, this qualifies as an AI Incident due to the direct link between AI use and realized cybersecurity harms.