Persistent AI Hallucinations Highlight Risks in Critical Applications

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Recent research and expert warnings highlight that hallucinations—false outputs generated by large language models (LLMs)—are unavoidable and increase with input size. These inaccuracies pose significant risks in high-stakes fields like law and accounting, challenging the reliability of AI for critical tasks.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves AI systems (LLMs) and their use, specifically their tendency to hallucinate false outputs. Although no direct harm is described as having occurred, the article clearly outlines the potential for these hallucinations to cause significant harm in critical domains. Therefore, this situation fits the definition of an AI Hazard, as the development and use of these AI systems could plausibly lead to an AI Incident involving harm to persons, organizations, or communities relying on accurate outputs.[AI generated]
AI principles
Robustness & digital securitySafety

Industries
Financial and insurance servicesGovernment, security, and defence

Affected stakeholders
ConsumersBusiness

Harm types
Economic/PropertyReputational

Severity
AI hazard

Business function:
Compliance and justice

AI system task:
Content generation


Articles about this incident or hazard

Thumbnail Image

The fatal flaw of AI-driven business models

2026-04-01
Bangkok Post
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (LLMs) and discusses their development and use, focusing on their probabilistic nature and hallucination issues. While it references potential harms (legal and financial risks) and cites an investigation showing serious errors in tax form completion by LLMs, it does not describe a specific incident of harm occurring due to AI use, nor does it describe a particular event where harm was narrowly avoided. Instead, it presents research findings and expert analysis about the fundamental limitations and risks of LLMs, which informs understanding of AI's impact on business and investment. This aligns with the definition of Complementary Information, as it enhances understanding of AI harms and risks without reporting a new AI Incident or AI Hazard.
Thumbnail Image

Does the AI business model have a fatal flaw?: Joachim Klement

2026-04-02
Zawya.com
Why's our monitor labelling this an incident or hazard?
The article focuses on the inherent limitations and risks of LLMs, supported by recent research, and discusses the implications for investors and businesses. It does not report a realized harm or a specific incident involving an AI system causing injury, rights violations, or other harms. Nor does it describe a particular event where an AI system's malfunction or misuse plausibly led to harm. Therefore, it does not meet the criteria for an AI Incident or AI Hazard. Instead, it provides contextual and analytical information about AI system capabilities and risks, which fits the definition of Complementary Information.
Thumbnail Image

AI Hallucinations: A Challenge Too Costly to Ignore | Technology

2026-04-01
Devdiscourse
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (LLMs) and their use, specifically their tendency to hallucinate false outputs. Although no direct harm is described as having occurred, the article clearly outlines the potential for these hallucinations to cause significant harm in critical domains. Therefore, this situation fits the definition of an AI Hazard, as the development and use of these AI systems could plausibly lead to an AI Incident involving harm to persons, organizations, or communities relying on accurate outputs.
Thumbnail Image

Researcher Warns That AI Hallucinations Are Unavoidable

2026-03-30
The Hoya
Why's our monitor labelling this an incident or hazard?
The article centers on a research seminar presentation warning about the persistent and unavoidable nature of hallucinations in LLMs. While it acknowledges the potential risks and challenges these hallucinations pose, it does not describe any actual harm or incident caused by AI hallucinations. The content is primarily informative and cautionary, discussing plausible future risks and the importance of user education to mitigate errors. Therefore, it fits the definition of an AI Hazard, as it highlights a credible risk that AI hallucinations could lead to harm, but no harm has yet occurred or been reported in this context.