Bing Image Creator Used to Generate Offensive 9/11-Themed Content Featuring Cartoon Characters

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Microsoft's Bing Image Creator, powered by DALL-E, was exploited by users to generate offensive images depicting cartoon characters like SpongeBob and Kirby committing the 9/11 attacks. Despite content moderation, users bypassed filters to create and share distressing, racist, and antisemitic imagery, prompting Microsoft to block related prompts and review safeguards.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves an AI system (DALL-E 3 integrated into Bing Chat) whose use has directly led to the generation of harmful content, including offensive and racist images. The failure of censorship filters constitutes a malfunction or inadequate use of the AI system, resulting in harm to communities through the spread of hurtful content. Therefore, this qualifies as an AI Incident due to the realized harm caused by the AI system's outputs and its failure to prevent such harm.[AI generated]
AI principles
SafetyRobustness & digital securityFairnessRespect of human rightsAccountability

Industries
Media, social platforms, and marketing

Affected stakeholders
General public

Harm types
Psychological

Severity
AI incident

Business function:
Citizen/customer service

AI system task:
Content generation


Articles about this incident or hazard

Thumbnail Image

More problems for Microsoft: the Twin Towers evade Bing's censorship - Softonic

2023-10-06
Softonic
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (DALL-E 3 integrated into Bing Chat) whose use has directly led to the generation of harmful content, including offensive and racist images. The failure of censorship filters constitutes a malfunction or inadequate use of the AI system, resulting in harm to communities through the spread of hurtful content. Therefore, this qualifies as an AI Incident due to the realized harm caused by the AI system's outputs and its failure to prevent such harm.
Thumbnail Image

Microsoft Bing AI Generates Images Of Kirby Doing 9/11

2023-10-04
Kotaku
Why's our monitor labelling this an incident or hazard?
The AI system (Bing AI Image Creator) is explicitly involved as the tool generating the images. The misuse of the AI system to create images simulating terrorism (9/11 attacks) constitutes harm to communities by trivializing a tragic event and potentially causing distress or offense. The AI's inability to understand context and the ease of circumventing content filters directly contributes to this harm. Although no physical injury or legal violation is reported, the harm to communities through dissemination of harmful content is a recognized form of AI Incident under the framework. The event is not merely a potential risk but an ongoing misuse causing harm, thus it is classified as an AI Incident rather than an AI Hazard or Complementary Information.
Thumbnail Image

Microsoft Addresses Controversial Twin Towers Art Generated by DALL-E 3 Integration

2023-10-06
Tech Times
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (DALL-E 3 integrated into Bing Image Creator) whose use has led to the generation of potentially harmful or offensive images. However, the article does not report any actual harm occurring, such as injury, rights violations, or significant community harm. Instead, it focuses on the potential for misuse and Microsoft's response to mitigate it. Therefore, this situation represents a plausible risk of harm due to AI misuse, qualifying it as an AI Hazard rather than an AI Incident. The article also details Microsoft's mitigation efforts, but the primary focus is on the potential for harm and the system's current vulnerabilities.
Thumbnail Image

Controversial creation: Microsoft halts Bing Image Creator prompts linked to twin towers tragedy - OnMSFT.com

2023-10-06
OnMSFT.com
Why's our monitor labelling this an incident or hazard?
An AI system (Bing Image Creator) was used to generate harmful content that is offensive and distressing, which constitutes harm to communities and ethical boundaries. Although no physical harm or direct legal violation is reported, the creation and dissemination of such content can cause significant social and emotional harm. Microsoft's response to block prompts and improve safeguards is a mitigation effort. Since the harm has occurred through the AI system's outputs, this qualifies as an AI Incident due to harm to communities and ethical concerns arising from AI-generated content.
Thumbnail Image

The Folly of DALL-E: How 4chan is Abusing Bing's New Image Model - bellingcat

2023-10-06
bellingcat
Why's our monitor labelling this an incident or hazard?
The event involves an AI system explicitly (Bing Image Creator using DALL-E 3) whose use has directly led to the creation and dissemination of antisemitic, racist, and violent propaganda. This constitutes harm to communities and violations of human rights. The article documents actual instances of harmful content generation and sharing, not just potential misuse. Therefore, this qualifies as an AI Incident under the OECD framework because the AI system's use has directly led to significant harm.
Thumbnail Image

Microsoft's AI is creating images of Kirby committing the attack on the Twin Towers - Softonic

2023-10-06
Softonic
Why's our monitor labelling this an incident or hazard?
An AI system (Microsoft's Bing Image Creator) is explicitly involved in generating images. The misuse of the AI system to create offensive and potentially harmful content related to terrorism is a plausible risk of harm to communities and societal norms. Although no direct harm has been reported as having occurred, the event clearly shows a credible potential for harm through malicious use of the AI system. Therefore, this qualifies as an AI Hazard rather than an AI Incident, since the harm is potential and the company is working on mitigation measures.
Thumbnail Image

Bing Is Generating Images of SpongeBob Doing 9/11

2023-10-04
404 Media
Why's our monitor labelling this an incident or hazard?
The AI system (Bing Image Creator) is explicitly involved as the tool generating the images. The event stems from the use of the AI system and its content moderation malfunction or insufficiency, as users find ways to bypass filters to create harmful content. The harm is indirect but real, as the generated images reference a terrorist attack and can cause distress or harm to communities by trivializing or mocking a tragic event. This fits the definition of an AI Incident because the AI system's use has directly or indirectly led to harm to communities. The article does not describe a potential future harm but an ongoing issue with harmful content generation. Therefore, the classification is AI Incident.