
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
Instagram's AI-based content moderation system failed to remove accounts posting sexually exploitative images of children, even after user reports. Despite Meta's claims of zero tolerance, the automated system allowed harmful content to remain accessible, exposing children to exploitation and violating their rights.[AI generated]
Why's our monitor labelling this an incident or hazard?
The event involves the use of AI-based automated content moderation systems on Instagram that failed to remove harmful accounts exploiting children sexually. This failure constitutes a malfunction of the AI system, which has indirectly led to harm by allowing the continued presence and dissemination of exploitative content involving children. The harm includes violation of children's rights and exposure to sexual exploitation, which fits the definition of an AI Incident. Therefore, this event is classified as an AI Incident due to the direct link between the AI system's malfunction and the harm caused.[AI generated]