AI Moderation Failure on X (Twitter) Enables Hate Speech After Musk Takeover

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

After Elon Musk's takeover of X (formerly Twitter), AI-driven content moderation has drastically declined, with 98% of hate-filled posts—including calls for violence and Holocaust denial—remaining online. Researchers found that the platform failed to remove harmful content, enabling ongoing hate speech and incitement.[AI generated]

Why's our monitor labelling this an incident or hazard?

The social media platform X uses AI systems for content recommendation and moderation. The article describes a significant increase in hate speech and antisemitic content on the platform, which has led to real-world harm including violence and community harm. The AI system's role in amplifying and promoting such content, combined with the owner's public endorsement of antisemitic posts, directly or indirectly leads to violations of rights and harm to communities. Therefore, this qualifies as an AI Incident under the framework, as the AI system's use has directly or indirectly led to harm (violations of rights and harm to communities).[AI generated]
AI principles
AccountabilitySafetyRespect of human rightsRobustness & digital securityTransparency & explainabilityHuman wellbeing

Industries
Media, social platforms, and marketing

Affected stakeholders
General public

Harm types
PsychologicalReputationalPublic interestHuman or fundamental rights

Severity
AI incident

Business function:
Monitoring and quality control

AI system task:
Recognition/object detection


Articles about this incident or hazard

Thumbnail Image

Casa Branca condena 'promoção abominável' de antissemitismo por Musk no X

2023-11-17
O Globo
Why's our monitor labelling this an incident or hazard?
The social media platform X uses AI systems for content recommendation and moderation. The article describes a significant increase in hate speech and antisemitic content on the platform, which has led to real-world harm including violence and community harm. The AI system's role in amplifying and promoting such content, combined with the owner's public endorsement of antisemitic posts, directly or indirectly leads to violations of rights and harm to communities. Therefore, this qualifies as an AI Incident under the framework, as the AI system's use has directly or indirectly led to harm (violations of rights and harm to communities).
Thumbnail Image

Neue Kritik an Musks X-Plattform von verklagten Forschern - WELT

2023-11-15
DIE WELT
Why's our monitor labelling this an incident or hazard?
The article involves AI systems implicitly, as content moderation and detection of manipulated content typically rely on AI. However, the main focus is on the dispute between the platform and researchers, enforcement actions, and compliance with legal frameworks. There is no direct or indirect harm caused by AI systems themselves described, nor a credible plausible future harm solely from AI. The event is best classified as Complementary Information because it provides context on societal and governance responses to AI-related content moderation issues rather than reporting a new AI Incident or AI Hazard.
Thumbnail Image

Hass und Hetze auf Twitter: "Musk hat einen Safe Space für Rassisten geschaffen" - netzpolitik.org

2023-11-15
netzpolitik.org
Why's our monitor labelling this an incident or hazard?
The event involves an AI system insofar as content moderation on large social media platforms typically relies on AI systems for detecting and removing hateful content. The article highlights that after Musk's takeover, moderation has been drastically reduced, with many hateful posts remaining online. This indicates a failure or change in the AI moderation system's use or effectiveness, leading to ongoing harm (hate speech, threats, violence incitement) to communities and violations of rights. The harm is realized and ongoing, not merely potential. Hence, it meets the criteria for an AI Incident, as the AI system's use or malfunction has directly or indirectly led to harm.
Thumbnail Image

Ισραήλ: Ο Ελον Μασκ συναντά Νετανιάχου και Χέρτζογκ εν μέσω κατηγοριών για ενίσχυση του αντιεβραϊκού μίσους | in.gr

2023-11-27
in.gr
Why's our monitor labelling this an incident or hazard?
The article explicitly links Elon Musk's ownership and management of the social media platform X to the amplification of antisemitic hate speech, which is a violation of human rights and causes harm to communities. The platform uses AI systems for content moderation and recommendation, which have directly or indirectly led to the spread of harmful content. The harm is realized, as evidenced by public condemnation, advertiser boycotts, and official complaints. Thus, the event meets the criteria for an AI Incident involving violations of rights and harm to communities caused by the AI system's use.
Thumbnail Image

Χέρτζογκ σε Μασκ: Το "X" είναι γεμάτο αντισημιτισμό

2023-11-27
Liberal.gr
Why's our monitor labelling this an incident or hazard?
The social media platform 'X' is an AI system that uses algorithms to moderate, recommend, and display content. The article reports that antisemitic content is widespread on the platform, causing harm to communities and violating rights. This harm is directly linked to the use of the AI system in content curation and moderation, which has failed to prevent or has indirectly enabled the spread of hate speech. Hence, this qualifies as an AI Incident due to violations of human rights and harm to communities caused by the AI system's use.
Thumbnail Image

Ίλον Μασκ: Η επίσκεψη στο Ισραήλ, η κατακραυγή και η υπόσχεση για "πάταξη" του αντισημιτισμού

2023-11-28
NEWS 24/7
Why's our monitor labelling this an incident or hazard?
The social media platform X, owned by Elon Musk, is an AI system that uses algorithms to moderate and recommend content. The spread and support of an antisemitic post on this platform have led to harm in the form of antisemitism, which is a violation of human rights and harms communities. The event describes realized harm caused or facilitated by the AI system's use (or misuse), fulfilling the criteria for an AI Incident. The visit and promises to combat hate speech are complementary responses but do not negate the incident classification.
Thumbnail Image

Επίσκεψη Έλον Μασκ στο Ισραήλ: Θα συναντηθεί με Νετανιάχου και Χέρτζογκ

2023-11-26
enikos.gr
Why's our monitor labelling this an incident or hazard?
The event centers on the consequences of AI-driven content moderation and recommendation on the platform X, which has been linked to the spread of antisemitic hate speech, a violation of human rights. The involvement of AI systems in the platform's operation and the resulting harm to communities through hate speech meets the criteria for an AI Incident. Although the article focuses on a meeting and discussion rather than a new incident, the described harms have already occurred and are directly linked to the AI system's use in the platform.
Thumbnail Image

Έλον Μασκ: Συνάντηση με τον πρόεδρο του Ισραήλ και τον πρωθυπουργό Νετανιάχου

2023-11-27
insider.gr
Why's our monitor labelling this an incident or hazard?
The platform X uses AI systems for content curation and moderation. The article describes how the platform has been accused of enabling antisemitic hate speech, which constitutes harm to communities and violations of human rights. The AI system's role in amplifying or failing to adequately moderate such content is an indirect cause of this harm. The meeting with political leaders underscores the seriousness of these harms. Hence, this qualifies as an AI Incident under the framework, as the AI system's use has directly or indirectly led to violations of rights and harm to communities.