AI Chatbots Generate False Information, Leading to Real-World Harm

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

AI chatbots from OpenAI, Google, and Microsoft have repeatedly generated false information, including fabricated legal cases that were submitted in court. These hallucinations have led to misinformation being used in sensitive contexts, demonstrating the real-world harm caused by AI system malfunctions.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves AI systems (large language model chatbots) whose malfunction (hallucination) has directly led to the dissemination of false information, including fabricated legal cases used in a real court submission, which constitutes harm to communities and potentially to individuals relying on this information. The presence of AI is clear, the malfunction is described in detail, and the harms are realized and significant. Hence, this is an AI Incident rather than a hazard or complementary information.[AI generated]
AI principles
AccountabilityRobustness & digital securitySafetyTransparency & explainabilityRespect of human rightsDemocracy & human autonomy

Industries
Government, security, and defence

Affected stakeholders
Government

Harm types
ReputationalPublic interestEconomic/PropertyHuman or fundamental rights

Severity
AI incident

Business function:
Compliance and justice

AI system task:
Content generationInteraction support/chatbots


Articles about this incident or hazard

Thumbnail Image

Chatbots may 'hallucinate' more often than many realise

2023-11-09
Economic Times
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (large language model chatbots) whose malfunction (hallucination) has directly led to the dissemination of false information, including fabricated legal cases used in a real court submission, which constitutes harm to communities and potentially to individuals relying on this information. The presence of AI is clear, the malfunction is described in detail, and the harms are realized and significant. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Chatbots invent facts. Experts call it 'hallucination' & say tech does it often - Times of India

2023-11-07
The Times of India
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (chatbots) and their malfunction in generating false information, which can lead to harm if users rely on incorrect outputs, particularly in sensitive contexts. However, the article reports on the general occurrence and research findings about hallucinations without describing a specific incident causing realized harm or a particular event where harm occurred. Therefore, it does not describe a concrete AI Incident or a specific AI Hazard but rather provides complementary information about a known AI system limitation and its implications.
Thumbnail Image

Chatbots may 'hallucinate' more often than many realise

2023-11-07
The Straits Times
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (chatbots) whose use leads to the generation of false information, which can plausibly cause harm if relied upon in critical contexts such as legal or medical domains. However, the article does not report a concrete harm event but rather quantifies the frequency of hallucinations and warns about the risks. Therefore, this constitutes an AI Hazard, as the development and use of these AI systems could plausibly lead to incidents of misinformation causing harm, but no specific harm has been documented in this report.
Thumbnail Image

Beware: Hallucinating chatbots invent fake truths

2023-11-09
DT next
Why's our monitor labelling this an incident or hazard?
The article clearly involves AI systems (large language models powering chatbots) and discusses their use and malfunction (hallucination leading to false information). The hallucinations have led to misinformation being generated and used in contexts such as legal briefs, which can cause harm to individuals or groups relying on accurate information, thus constituting harm to communities and potentially violations of rights to accurate information. Since the hallucinations are occurring and have caused real misinformation, this qualifies as an AI Incident. The article does not merely warn about potential future harm but documents ongoing harm caused by AI system outputs.
Thumbnail Image

Chatbots May 'Hallucinate' More Often Than Many Realize

2023-11-06
The New York Times
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (large language model chatbots) and their malfunction in the form of hallucination—generating false information. While no specific harm event is reported, the article clearly outlines the plausible risk of harm in sensitive applications (legal briefs, medical info, business data). This fits the definition of an AI Hazard: an event or circumstance where AI system malfunction could plausibly lead to harm. The article's focus is on the risk and prevalence of hallucination, not on a realized incident, so it is not an AI Incident. It is also not merely complementary information, as it centers on the hazard itself rather than responses or ecosystem updates. Hence, the classification is AI Hazard.
Thumbnail Image

Los chatbots podrían estar 'alucinando' con más frecuencia de lo que muchos creen

2023-11-07
infobae
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (large language model chatbots) and their use leading to the generation of false information ('hallucinations'). This misinformation can cause harm to individuals and communities, especially when used in sensitive contexts such as legal or medical domains. The harm is realized as the chatbots have already produced false outputs that have been used (e.g., fabricated legal cases submitted to courts). Hence, this qualifies as an AI Incident due to direct harm caused by the AI systems' outputs. The article does not merely warn about potential future harm but documents ongoing issues and their impacts.
Thumbnail Image

¿Los chatbots inventan? Un estudio indica que más de lo que creíamos

2023-11-09
The New York Times
Why's our monitor labelling this an incident or hazard?
The article describes the use and malfunction (hallucination) of AI systems (chatbots based on large language models) that produce false information. Although no specific incident of harm is reported, the article emphasizes the plausible risk that such hallucinations could cause harm in legal, medical, or business settings if the false information is trusted and acted upon. This fits the definition of an AI Hazard, as the development and use of these AI systems could plausibly lead to harms such as misinformation causing harm to individuals or communities, or violations of rights through incorrect legal or medical advice. The article does not report a realized harm or incident, nor does it focus on governance responses or complementary information primarily. Therefore, the correct classification is AI Hazard.