Instagram's AI Moderation Fails to Remove Child Exploitation Accounts

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Instagram's AI-based content moderation system failed to remove accounts posting sexually exploitative images of children, even after user reports. Despite Meta's claims of zero tolerance, the automated system allowed harmful content to remain accessible, exposing children to exploitation and violating their rights.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves the use of AI-based automated content moderation systems on Instagram that failed to remove harmful accounts exploiting children sexually. This failure constitutes a malfunction of the AI system, which has indirectly led to harm by allowing the continued presence and dissemination of exploitative content involving children. The harm includes violation of children's rights and exposure to sexual exploitation, which fits the definition of an AI Incident. Therefore, this event is classified as an AI Incident due to the direct link between the AI system's malfunction and the harm caused.[AI generated]
AI principles
AccountabilitySafetyRespect of human rightsRobustness & digital securityTransparency & explainabilityHuman wellbeing

Industries
Media, social platforms, and marketing

Affected stakeholders
Children

Harm types
Human or fundamental rightsPsychologicalPublic interestReputational

Severity
AI incident

Business function:
Monitoring and quality controlCompliance and justice

AI system task:
Recognition/object detectionOrganisation/recommenders


Articles about this incident or hazard

Thumbnail Image

هل يستغل إنستغرام الأطفال جنسيا؟

2022-04-18
أخبار الآن
Why's our monitor labelling this an incident or hazard?
The event involves the use of AI-based automated content moderation systems on Instagram that failed to remove harmful accounts exploiting children sexually. This failure constitutes a malfunction of the AI system, which has indirectly led to harm by allowing the continued presence and dissemination of exploitative content involving children. The harm includes violation of children's rights and exposure to sexual exploitation, which fits the definition of an AI Incident. Therefore, this event is classified as an AI Incident due to the direct link between the AI system's malfunction and the harm caused.
Thumbnail Image

"نافذة متجر" على إنستغرام.. اتهامات بشأن حسابات تستغل الأطفال جنسيا

2022-04-17
الحرة
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions the use of automated moderation technology (AI systems) that failed to remove accounts posting sexually exploitative content involving children. This failure allowed harmful content to remain accessible, leading to direct harm to children and communities, and violating fundamental rights. The AI system's malfunction in content moderation is a contributing factor to the harm. Therefore, this event qualifies as an AI Incident due to the realized harm linked to the AI system's use and malfunction.
Thumbnail Image

هل يسهّل "إنستغرام" حسابات تستغل الأطفال جنسياً؟

2022-04-18
almodon
Why's our monitor labelling this an incident or hazard?
The event involves AI systems used for automated content moderation on social media platforms. The AI's malfunction or inadequacy in detecting and removing child sexual exploitation content has indirectly led to harm to children, violating their rights and exposing them to exploitation. This fits the definition of an AI Incident because the AI system's malfunction (failure to act effectively) has directly contributed to harm (violation of rights and exposure to exploitation).
Thumbnail Image

"نافذة متجر" على إنستغرام.. اتهامات بشأن حسابات تستغل الأطفال جنسياً

2022-04-19
baghdadtoday.news
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions the use of automated AI-based moderation tools in Instagram's content review process, which failed to remove accounts posting sexually exploitative images of children. This failure directly led to harm by allowing the continued presence of harmful content and exposure of children to exploitation risks. The AI system's malfunction or insufficiency in detecting and removing such content is central to the incident. Therefore, this qualifies as an AI Incident involving harm to people (children) and communities through violations of rights and exposure to exploitation.
Thumbnail Image

"نافذة متجر" على إنستغرام.. اتهامات بشأن حسابات تستغل الأطفال جنسياً - عاجل

2022-04-19
موسوعه العراق
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Instagram's automated content moderation and reporting system) that is explicitly mentioned as failing to remove harmful content exploiting children. The AI system's malfunction or inadequacy has directly led to harm by allowing the continued presence of exploitative accounts and content, which harms children and violates their rights. This fits the definition of an AI Incident because the AI system's use and malfunction have directly led to violations of human rights and harm to vulnerable groups (children).