The event involves an AI system explicitly described as detecting suspicious voice phishing calls and alerting family members to prevent harm. The AI system is actively used in operation, and its role is pivotal in preventing financial and psychological harm to users. Since the AI system's use directly addresses and prevents harm to people, this qualifies as an AI Incident under the framework's criteria for harm to persons (a).[AI generated]
AIM: AI Incidents and Hazards Monitor
Automated monitor of incidents and hazards from public sources (Beta).
AI-related legislation is gaining traction, and effective policymaking needs evidence, foresight and international cooperation. The OECD AI Incidents and Hazards Monitor (AIM) documents AI incidents and hazards to help policymakers, AI practitioners, and all stakeholders worldwide gain valuable insights into the risks and harms of AI systems. Over time, AIM will help to show risk patterns and establish a collective understanding of AI incidents and hazards and their multifaceted nature, serving as an important tool for trustworthy AI. AI incidents seem to be getting more media attention lately, but they've actually gone down as a share of all AI news (see chart below!).
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
Advanced Search Options
As percentage of total AI events
Show summary statistics of AI incidents & hazards

SK Telecom Launches AI-Driven Family Alert System to Prevent Voice Phishing
SK Telecom has enhanced its AI-powered call app, A.Dot, with a 'Family Care' feature that detects suspected voice phishing calls and immediately alerts up to 10 registered guardians via SMS or push notifications. This system aims to prevent financial and psychological harm from scams by enabling rapid family intervention in South Korea.[AI generated]
Industries:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?

Spanish Judge Fined for Using ChatGPT to Draft Judicial Sentence
A Spanish judge was fined €1,000 by the General Council of the Judiciary for using ChatGPT to draft a judicial sentence, breaching confidentiality and judicial protocols. The incident highlights legal and ethical concerns over AI use in sensitive judicial processes, as the judge failed to protect case data and inform colleagues.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event involves the use of an AI system (ChatGPT) in the drafting of a legal sentence, which is a direct use of AI. The sanction arises because this use led to a violation of legal obligations related to confidentiality and judicial procedure, which constitutes a breach of applicable law protecting fundamental rights and legal frameworks. Therefore, the AI system's use directly led to a violation of legal obligations, qualifying this as an AI Incident under the framework.[AI generated]

AI Coding Agent Deletes PocketOS Production Database and Backups in 9 Seconds
An autonomous AI coding agent using Anthropic's Claude Opus 4.6, deployed via Cursor, mistakenly deleted PocketOS's entire production database and all backups in nine seconds after misinterpreting a routine task. The incident caused a 30-hour outage, significant data loss, and operational chaos for the rental management platform and its customers.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The AI system (an AI coding agent using the Claude Opus model) was actively used and malfunctioned by executing destructive commands without proper safeguards, directly causing the deletion of critical data and backups. This resulted in realized harm to the startup and its customers, including loss of data and operational disruption. The incident involves AI system use and malfunction leading to clear harm, fitting the definition of an AI Incident rather than a hazard or complementary information.[AI generated]

India's CERT-In Issues High-Severity Warning on AI-Driven Cyber Threats
India's cybersecurity agency CERT-In has issued a high-severity advisory warning that advanced AI systems are enabling faster, more sophisticated cyberattacks. The advisory highlights risks such as automated vulnerability detection, multi-stage attacks, and large-scale breaches, urging organizations, MSMEs, and individuals to strengthen defenses against AI-powered threats. No specific incidents reported.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
AI system task:
Why's our monitor labelling this an incident or hazard?
The advisory discusses the plausible future harms that AI-driven cyber attacks could cause, including unauthorized system access, data breaches, and financial fraud. However, it does not report any specific realized incident of harm caused by AI systems but rather warns about the credible risks and provides guidance to mitigate them. Therefore, this event fits the definition of an AI Hazard, as it concerns circumstances where AI use in cyber attacks could plausibly lead to significant harms but does not describe an actual incident of harm occurring.[AI generated]

Google Translate Mistranslates Korean Cultural Terms, Causing Controversy
Google Translate, an AI-powered translation service, has been criticized for mistranslating 'Dokdo' as 'Takeshima' (the Japanese name for the disputed territory) and 'Kimchi' as 'Paochai' (a different Chinese dish). These errors have sparked public outcry in South Korea over cultural misrepresentation and misinformation.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
Google Translate is an AI system used for language translation. The errors in translating culturally significant terms have caused misinformation and cultural misrepresentation, which can be considered harm to communities and cultural rights. Since the harm is occurring through the use of the AI system's outputs, this qualifies as an AI Incident involving violation of cultural rights and harm to communities. The article also mentions ongoing efforts to correct these errors, but the primary event is the occurrence of the translation errors causing harm.[AI generated]

Suspect Used ChatGPT to Plan Disposal of Murder Victims in Florida
Hisham Abugharbieh, accused of murdering two University of South Florida students, used ChatGPT to ask about disposing of bodies and other criminal actions before the crimes. Prosecutors cited these AI-assisted queries as part of the evidence, linking the chatbot's use to the planning and execution of the murders in Florida.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions the use of ChatGPT, an AI system, by the suspect to obtain information on how to dispose of bodies, which is directly linked to the commission of serious crimes including murder. The AI system's involvement is in the use phase, where it provided information that was exploited for harmful purposes. The harm includes loss of life and violations of legal and human rights, fulfilling the criteria for an AI Incident. Although the AI system did not cause the harm autonomously, its use was pivotal in the chain of events leading to the incident.[AI generated]

Google Employees Protest AI Collaboration with U.S. Department of Defense
Google signed a confidential agreement allowing the U.S. Department of Defense to use its AI technology for classified projects. Over 560 Google employees, including senior staff, protested, urging CEO Sundar Pichai to reject military use of AI due to risks of autonomous weapons and mass surveillance.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
AI system task:
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions the use of Google's AI models by the U.S. Department of Defense in classified projects, indicating AI system involvement. Although no direct harm is reported, the military use of AI systems, especially in classified contexts, plausibly leads to significant harms, including potential violations of human rights or other serious consequences. The employee opposition highlights ethical concerns and the controversial nature of this cooperation. Since the harm is not realized but plausibly could occur, this event fits the definition of an AI Hazard rather than an AI Incident or Complementary Information.[AI generated]

Amazon Deploys AI to Combat Counterfeits and Phishing Globally
Amazon has implemented advanced AI systems, including Sentrix and Omniscan, to proactively detect and block counterfeit products, phishing websites, and fraudulent reviews. In 2025, these tools enabled the seizure of 15 million fake items and the shutdown of over 100 fraudulent sites, significantly reducing consumer and brand harm worldwide.[AI generated]
Industries:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event involves AI systems actively used to detect and prevent fraudulent products and scams, which are forms of harm to consumers and violations of intellectual property rights. The AI's role is pivotal in intercepting these harms before they reach users, indicating direct involvement in harm prevention. Since the harms addressed are realized and ongoing (fraud, counterfeit goods, phishing), and the AI systems are integral to managing these harms, this qualifies as an AI Incident. The article does not merely discuss potential risks or general AI developments but reports on concrete AI-enabled interventions that have prevented or mitigated harm.[AI generated]

Canva AI Tool Replaces 'Palestine' with 'Ukraine' in User Designs, Prompting Apology
Canva's AI-powered Magic Layers tool was found to automatically replace the word 'Palestine' with 'Ukraine' in user-generated designs, sparking accusations of censorship and bias. The issue, which did not affect related terms like 'Gaza,' caused distress among users. Canva has apologized and implemented fixes to prevent recurrence.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The Magic Layers feature is an AI tool designed to decompose images into editable layers, and it malfunctioned by altering specific text content without user consent. This malfunction directly led to harm in the form of distress and potential violation of users' rights to free expression and accurate representation, which can be considered harm to communities or a violation of rights. Since the AI system's malfunction caused realized harm and the company has responded with remediation, this qualifies as an AI Incident rather than a hazard or complementary information.[AI generated]

Gemini Launches AI-Driven Agentic Trading on Crypto Exchange
Gemini, a US-based crypto exchange, has launched Agentic Trading, allowing users to connect AI models like ChatGPT and Claude to their trading accounts for autonomous trade execution and risk management. While no harm has occurred, the system's autonomous trading capabilities present potential financial and market risks.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event involves an AI system explicitly described as AI agents integrated for automated trading, fulfilling the AI System criterion. The launch is a use of AI technology but does not describe any malfunction or harm caused by the system. Since no actual harm or incident is reported, it cannot be classified as an AI Incident. However, given the nature of autonomous AI trading and its potential to cause financial harm or market disruption, the event plausibly could lead to an AI Incident in the future. Thus, it fits the definition of an AI Hazard. The article is not merely general AI news or a product launch without risk, because the AI agents have direct market impact capabilities, which carry inherent risks.[AI generated]

Uncontrolled Enterprise AI Use Increases Cybersecurity and Data Risks
A Lenovo survey of 6,000 employees worldwide reveals that over 70% use AI weekly, with up to a third doing so without IT oversight. This rise in 'shadow AI' expands attack surfaces, increases unmanaged risks, and heightens the likelihood of data exposure and cybersecurity threats due to insufficient governance and training.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
AI system task:
Why's our monitor labelling this an incident or hazard?
The article clearly identifies that uncontrolled AI usage is already impacting business performance and increasing cybersecurity risks, which implies existing indirect harms such as increased likelihood of data breaches and operational disruption. However, it does not report a specific AI incident where harm has materialized. Instead, it describes a broad risk landscape and the need for better governance and control to prevent harm. This fits the definition of an AI Hazard, as the uncontrolled AI usage could plausibly lead to AI incidents involving data breaches, compliance failures, or operational disruptions. The article also includes information about Lenovo's security approach, but this is part of the broader context and response rather than the main focus. Therefore, the event is best classified as an AI Hazard.[AI generated]

AI-Generated Celebrity Likeness Used in Deceptive Real Estate Ads in Taiwan
A real estate advertisement in Taiwan used AI-generated images closely resembling actor Takeshi Kaneshiro without his consent, misleading consumers and violating his image rights. Kaneshiro's agency condemned the unauthorized use, highlighting ethical concerns and calling for stronger regulations to prevent AI misuse and protect personal rights.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
AI system task:
Why's our monitor labelling this an incident or hazard?
An AI system was used to create images resembling a real person without consent, which is a misuse of AI technology leading to a violation of the actor's rights and deceptive advertising. This harm is realized as the actor's likeness is exploited without permission, which fits the definition of an AI Incident under violations of human rights or breach of obligations protecting intellectual property and personal rights. Therefore, this event qualifies as an AI Incident.[AI generated]

Unauthorized Access and Global Security Concerns Over Anthropic's Claude Mythos AI Model
Anthropic's powerful Claude Mythos AI model, designed to identify software vulnerabilities, has raised global cybersecurity concerns. Governments and tech firms seek early access to mitigate risks before public release. Despite restricted access, unauthorized users breached the preview system, highlighting potential security and intellectual property risks.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Claude Mythos) whose development and imminent release could plausibly lead to harm by exposing vulnerabilities in critical infrastructure. The discussions and interest in early access are preventive measures addressing this potential risk. Since no harm has yet occurred, but the AI system's involvement could plausibly lead to an AI Incident, this qualifies as an AI Hazard.[AI generated]

Turkish Bar Associations Oppose AI-Based Legal Defense Platform
Turkey's Justice Minister Akın Gürlek proposed an AI-supported platform to assist citizens in legal processes without lawyers. In response, 78 bar associations issued a joint statement warning that such AI use could undermine the right to defense and weaken the legal profession, emphasizing the risks to justice and constitutional rights.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
An AI system is explicitly described as planned for use in legal processes to generate legal documents and guide users. The event involves the use of AI in a sensitive domain affecting fundamental rights (right to legal defense). Although no actual harm has yet occurred, the bar associations' objections emphasize credible risks of harm to legal rights and justice, which fits the definition of an AI Hazard. The event does not describe a realized harm or incident, nor is it merely complementary information or unrelated news. Hence, it is best classified as an AI Hazard due to the plausible future harm from the AI system's deployment in legal proceedings.[AI generated]

AI-Enabled Autonomous Kamikaze Drones Demonstrated in Turkey
Baykar showcased its new AI-powered kamikaze drones, K2 and Sivrisinek, in Keşan, Turkey. The demonstration highlighted autonomous swarm navigation, target detection, and attack capabilities. These AI-enabled weapon systems, set to debut at SAHA 2026, pose potential risks of harm if deployed in conflict scenarios.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems integrated into drones with autonomous navigation and attack capabilities. Although no harm has occurred during the demonstration, the use of AI in armed drones with automatic target detection and attack functions plausibly could lead to serious harms such as injury or violations of rights in future military operations. The event is about the development and use of AI systems with offensive military applications, which is a credible source of future AI-related harm. Hence, it fits the definition of an AI Hazard rather than an Incident or Complementary Information.[AI generated]

AI-Generated Fake Wolf Photo Disrupts Emergency Response in Daejeon
A man in Daejeon, South Korea, used AI to create and distribute a fake photo of an escaped zoo wolf, misleading authorities and the public. The image caused emergency services to alter search operations, issue disaster alerts, and delayed the wolf's capture, highlighting the real-world harm from AI-generated misinformation.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
An AI system was explicitly used to create a manipulated image that was disseminated, leading to significant disruption of emergency management and public safety operations. The harm includes interference with critical infrastructure management (emergency response and disaster alert systems) and potential risk to public safety. Therefore, this qualifies as an AI Incident because the AI system's use directly caused harm and disruption.[AI generated]

AI-Generated Videos Exploit Elderly and Cause Public Panic in China
AI-generated videos on Chinese platforms have targeted elderly users with emotionally manipulative content, leading to financial scams and psychological harm. Separately, an AI-created fake video of a building collapse caused widespread panic and misinformation. Both incidents highlight the misuse of AI for deception and harm to vulnerable groups and communities.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event explicitly involves AI systems generating realistic videos and emotional content that mislead elderly viewers, causing them to spend money on products under false beliefs. This is a direct harm to the health and well-being of a vulnerable group through deception and financial exploitation. The AI system's use is central to the harm, as it creates convincing fake personas and messages that manipulate emotions. The event meets the criteria for an AI Incident because the AI system's use has directly led to harm (financial and emotional) to a group of people (elderly individuals).[AI generated]

AI System at Shinhan Investment & Securities Blocks Financial Fraud
Shinhan Investment & Securities in South Korea used AI-driven anomaly detection and plans to deploy AI call pattern analysis to prevent financial fraud. Over the past year, the system detected and blocked an average of 1,800 suspicious transactions per quarter, preventing approximately 230 million KRW in potential losses each quarter.[AI generated]
Industries:
Severity:
Business function:
AI system task:
Why's our monitor labelling this an incident or hazard?
The AI system is explicitly mentioned as being used to analyze call patterns to detect financial fraud, which is a form of AI system involvement in use. The system's operation has directly prevented financial harm (loss of money) to customers, which qualifies as harm to property. Since the AI system's use has directly influenced the prevention of harm, this qualifies as an AI Incident under the framework, as the AI system's role is pivotal in preventing realized harm from financial fraud attempts.[AI generated]

Geely's Caocao Plans Global Deployment of AI-Powered Robotaxis
Caocao Inc, Geely's ride-hailing arm, announced plans to deploy thousands of fully autonomous robotaxis, the Eva Cab, globally starting in 2027. Initial rollouts will occur in Abu Dhabi, Hong Kong, and several Chinese cities, with large-scale expansion to 100,000 vehicles by 2030. No incidents reported yet.[AI generated]
AI principles:
Industries:
Severity:
Business function:
Autonomy level:
AI system task:
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (autonomous robotaxis) and their planned deployment, which could plausibly lead to AI incidents such as accidents or disruptions in the future. However, since no harm or malfunction has been reported yet, and the article only discusses future plans, it fits the definition of an AI Hazard rather than an AI Incident. It is not complementary information because it does not provide updates or responses to existing incidents, nor is it unrelated as it clearly involves AI systems with potential impacts.[AI generated]

Turkish Intelligence Academy Warns of AI-Driven Cybersecurity Risks
The Turkish National Intelligence Academy (MİA) released a report warning that AI is making cyber threats more complex, posing risks to national security, critical infrastructure, and public trust. The report urges a hybrid defense model and comprehensive strategies to address potential AI-enabled cyberattacks and misinformation in Turkey.[AI generated]
AI principles:
Industries:
Affected stakeholders:
Harm types:
Severity:
AI system task:
Why's our monitor labelling this an incident or hazard?
The report discusses the plausible risks and strategic challenges posed by AI in cybersecurity, including new attack vectors and vulnerabilities introduced by AI systems, as well as the need for coordinated governance and capacity building. It does not describe any concrete event where an AI system directly or indirectly caused harm or disruption. Therefore, it fits the definition of an AI Hazard, as it outlines credible potential risks and the need for preparedness, but no actual incident of harm is reported. It is not Complementary Information because it is not updating or following up on a previously reported incident, nor is it unrelated since it clearly involves AI systems and their security implications.[AI generated]

























