Google Detects First AI-Generated Zero-Day Attack Code Used by State-Backed Hackers

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Google's Threat Intelligence Group identified the first known case of AI-generated zero-day exploit code, used in attempted cyberattacks by state-backed groups from North Korea, China, and Russia. AI systems autonomously developed and tested attack scripts, increasing the scale and sophistication of cyber threats, though the specific attack was thwarted.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly mentions AI systems being used to develop zero-day exploit codes and conduct cyberattacks, which are harmful acts targeting software vulnerabilities and potentially critical infrastructure. The AI involvement is direct in the development and attempted use of attack codes, which is a clear case of AI use leading to harm or attempted harm. Although the attack was not successful, the direct link between AI use and the creation of harmful exploit code qualifies this as an AI Incident rather than a hazard. The harms involved include cybersecurity breaches and potential disruption or damage to property and communities. Therefore, the event meets the criteria for an AI Incident.[AI generated]
AI principles
SafetyRobustness & digital security

Industries
Digital security

Severity
AI incident

AI system task:
Content generation


Articles about this incident or hazard

Thumbnail Image

구글 "AI 활용한 대규모 제로데이 공격 시도 첫 포착"

2026-05-12
기술로 세상을 바꾸는 사람들의 놀이터
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems being used to develop zero-day exploit codes and conduct cyberattacks, which are harmful acts targeting software vulnerabilities and potentially critical infrastructure. The AI involvement is direct in the development and attempted use of attack codes, which is a clear case of AI use leading to harm or attempted harm. Although the attack was not successful, the direct link between AI use and the creation of harmful exploit code qualifies this as an AI Incident rather than a hazard. The harms involved include cybersecurity breaches and potential disruption or damage to property and communities. Therefore, the event meets the criteria for an AI Incident.
Thumbnail Image

AI 위협 추적 보고서..."AI 기반 제로데이 공격 최초 탐지"

2026-05-12
아시아경제
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems being used by threat actors to develop zero-day attack codes and conduct autonomous vulnerability exploration, which directly leads to cyberattacks and associated harms. The AI involvement is in the use of AI to develop and execute malicious hacking tools, which is a direct cause of harm to property, organizations, and potentially critical infrastructure. Therefore, this event meets the criteria for an AI Incident as the AI system's use has directly led to realized harm through cyberattacks.
Thumbnail Image

구글, AI 이용해 생성된 '제로데이' 취약점 최초 발견

2026-05-12
kbench.com
Why's our monitor labelling this an incident or hazard?
The article explicitly states that AI, particularly large language models, was used to generate attack code that bypassed two-factor authentication in a web management tool, representing a direct use of AI in malicious activity. The AI system's role in automating and enhancing cyberattacks that exploit zero-day vulnerabilities directly leads to harm in terms of cybersecurity breaches, which can disrupt systems and compromise data. The mention of specific hacking groups using AI tools further supports the direct involvement of AI in causing harm. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

"AI가 신규 취약점 찾아 해킹 시도"...구글, 인공지능이 만든 '첫 공격 코드' 포착

2026-05-12
와이드경제
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems autonomously generating zero-day attack code and being used by state-backed hacking groups to conduct cyberattacks, espionage, and psychological operations. These activities have directly led to harm, including security breaches, development of malware, and disinformation campaigns, which fall under violations of rights and harm to communities. The AI's role is pivotal in enabling these attacks and increasing their scale and sophistication. Hence, this event meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

AI 스스로 취약점 찾아 해킹... '사이버 무기' 시대

2026-05-12
서울신문
Why's our monitor labelling this an incident or hazard?
The article explicitly describes AI systems autonomously finding zero-day vulnerabilities and generating attack code used in real cyberattacks by state-backed groups (North Korea's APT45, China-linked actors, Russia). This use of AI has directly led to harm in the form of cyberattacks, which disrupt digital infrastructure and threaten information security, fitting harm categories (b) and (d). The involvement of AI in both offensive and defensive cyber operations is clear, with AI systems playing a pivotal role in causing and mitigating harm. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

구글 "AI로 만든 제로데이 해킹 코드 첫 포착"

2026-05-12
포인트데일리
Why's our monitor labelling this an incident or hazard?
The article explicitly describes AI systems being used by malicious actors to develop and deploy zero-day exploit codes and AI-enhanced cyberattacks, including deepfake media for information operations. These activities have already led to or pose direct harm to cybersecurity, information integrity, and potentially national security, which are harms to communities and critical infrastructure. The AI involvement is clear and central to the incident, and the harms are realized or ongoing. Hence, this qualifies as an AI Incident rather than a hazard or complementary information.