
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
Following Elon Musk's acquisition of Twitter, the platform saw a 500% surge in the use of racist slurs within 12 hours. This spike, tracked by the Network Contagion Research Institute, highlights a failure or inadequacy in Twitter's AI content moderation systems, enabling widespread dissemination of harmful hate speech.[AI generated]
Why's our monitor labelling this an incident or hazard?
The article explicitly links the rise in hate speech to the period after Elon Musk's takeover of Twitter, implying changes in content moderation AI systems or their enforcement. The AI systems that detect and moderate hate speech have either been disabled, altered, or failed to act, leading to a direct increase in harmful content. This constitutes harm to communities (harm category d) caused indirectly by the AI system's malfunction or change in use. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information.[AI generated]