AI Chatbots Manipulated to Spread Misinformation via Simple Online Tricks

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Researchers demonstrated that ChatGPT and Google's AI chatbots can be easily manipulated by creating false online content, causing these systems to spread misinformation on critical topics like health and finance. This vulnerability has led to real harm by misleading users and undermining trust in AI-generated information.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves AI systems (ChatGPT, Google's AI chatbots) whose use has directly led to harm by spreading misinformation and falsehoods that can mislead users on important topics. The AI systems' vulnerability to manipulation and their role in amplifying false information constitute a violation of users' right to accurate information and can lead to harm to communities and individuals. Therefore, this qualifies as an AI Incident because the harm is occurring and the AI systems' use is a direct contributing factor.[AI generated]
AI principles
Robustness & digital securitySafety

Industries
Healthcare, drugs, and biotechnologyFinancial and insurance services

Affected stakeholders
Consumers

Harm types
Economic/PropertyPsychological

Severity
AI incident

Business function:
Citizen/customer service

AI system task:
Interaction support/chatbots


Articles about this incident or hazard

Thumbnail Image

Hackeé ChatGPT y la IA de Google en 20 minutos (y qué logré demostrar con eso) - BBC News Mundo

2026-02-19
BBC
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (ChatGPT, Google's AI chatbots) whose use has directly led to harm by spreading misinformation and falsehoods that can mislead users on important topics. The AI systems' vulnerability to manipulation and their role in amplifying false information constitute a violation of users' right to accurate information and can lead to harm to communities and individuals. Therefore, this qualifies as an AI Incident because the harm is occurring and the AI systems' use is a direct contributing factor.
Thumbnail Image

La nueva era del spam: por qué es tan fácil engañar a la inteligencia artificial

2026-02-19
La Nacion
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (large language models and AI-powered search/chat tools) whose outputs are manipulated by exploiting their reliance on internet data, leading to the spread of false information. This misinformation has already occurred and is causing harm by misleading users on important topics such as health and finance, which can result in injury or financial loss. The article documents actual instances of this manipulation and its effects, not just potential risks. Hence, it meets the criteria for an AI Incident because the AI's use has directly led to harm to communities and individuals through misinformation and deception.
Thumbnail Image

Hackeé ChatGPT y la IA de Google en 20 minutos (y qué logré demostrar con eso)

2026-02-19
El Comercio Perú
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (ChatGPT, Google's AI tools) whose outputs have been manipulated through the creation of false online content, leading to the spread of misinformation. This misinformation can cause real harm to individuals and communities by influencing decisions on health, finance, and other critical areas. The article documents actual occurrences of this manipulation and its effects, not just potential risks. Hence, the AI systems' use and malfunction have directly led to harm, meeting the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Hackeé ChatGPT y la IA de Google en 20 minutos (y qué logré demostrar con eso)

2026-02-19
mdz
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (ChatGPT, Google's AI tools) whose use has directly led to the spread of false information, a form of harm to communities and individuals who rely on these systems for accurate information. The article documents actual occurrences of this misinformation being propagated by the AI systems after being manipulated via fabricated web content. This meets the criteria for an AI Incident because the AI systems' use has directly caused harm through misinformation dissemination, which can lead to poor decision-making and potential physical or reputational harm. The article also discusses the companies' awareness and ongoing efforts to mitigate the problem, but the harm is already occurring, so it is not merely a hazard or complementary information.
Thumbnail Image

Hackeé ChatGPT y la IA de Google en 20 minutos (y qué logré demostrar con eso)

2026-02-19
EL DEBER
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (ChatGPT, Google's AI tools) whose use has directly led to the spread of misinformation, a form of harm to communities and potentially to individuals' health and safety. The AI systems are being manipulated through the creation of false online content that the AI then uses as factual information, resulting in false outputs that users rely on. This constitutes an AI Incident because the AI's use has directly caused harm through misinformation dissemination, with real-world consequences such as poor health or financial decisions. The article documents actual occurrences of this harm, not just potential risks, and thus it is not merely a hazard or complementary information.
Thumbnail Image

Hackeé ChatGPT y la IA de Google en 20 minutos y logré demostrar que a veces inventan cosas - La Opinión

2026-02-19
La Opinión Digital
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (ChatGPT, Google AI) whose use has directly led to harm in the form of misinformation dissemination, which can cause harm to individuals and communities (harm to health, finances, and societal trust). The article documents realized harm, not just potential, as the AI systems have already repeated false information. Therefore, this qualifies as an AI Incident under the framework, as the AI systems' use has directly led to significant harm to communities and individuals through misinformation.