X (Twitter) Fails to Remove AI-Generated Deepfake Nudes Reported as Nonconsensual Content

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Researchers found that X (formerly Twitter) swiftly removes AI-generated deepfake nude images when reported as copyright violations, but largely ignores reports of nonconsensual nudity. This highlights a significant failure in the platform's AI-assisted moderation, allowing harmful synthetic content to remain accessible and exposing gaps in user protection.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves AI-generated deepfake images (an AI system generating synthetic content) and the platform's use of AI or algorithmic moderation systems to manage content reports. The failure to remove nonconsensual deepfake nude images despite reports leads to ongoing harm to individuals' rights and communities, fulfilling the criteria for an AI Incident. The AI system's role is pivotal as it generates the harmful content, and the platform's moderation system's failure to act constitutes a breach of obligations to protect rights.[AI generated]
AI principles
AccountabilityFairnessPrivacy & data governanceRespect of human rightsRobustness & digital securitySafetyTransparency & explainabilityHuman wellbeing

Industries
Media, social platforms, and marketingDigital security

Affected stakeholders
Consumers

Harm types
Human or fundamental rightsPsychologicalReputational

Severity
AI incident

Business function:
Monitoring and quality controlCitizen/customer service

AI system task:
Recognition/object detection


Articles about this incident or hazard

Thumbnail Image

Copyright is the only functional law of the internet, deepfake nudes edition.

2024-10-08
The Verge
Why's our monitor labelling this an incident or hazard?
The article discusses the effectiveness of copyright law in removing nonconsensual sexual imagery, including deepfake nudes, from a platform. While AI-generated content (deepfakes) is involved, the event centers on legal enforcement and platform moderation outcomes rather than an AI system causing harm through its development, use, or malfunction. There is no direct or indirect harm caused by an AI system itself, nor is there a plausible future harm from AI system malfunction or misuse described. Therefore, this is complementary information about the societal and legal context of AI-generated content rather than an AI Incident or Hazard.
Thumbnail Image

Study: Reports of nonconsensual nude images are ignored on X

2024-10-10
Mashable
Why's our monitor labelling this an incident or hazard?
The event involves AI-generated deepfake images (an AI system generating synthetic content) and the platform's use of AI or algorithmic moderation systems to manage content reports. The failure to remove nonconsensual deepfake nude images despite reports leads to ongoing harm to individuals' rights and communities, fulfilling the criteria for an AI Incident. The AI system's role is pivotal as it generates the harmful content, and the platform's moderation system's failure to act constitutes a breach of obligations to protect rights.
Thumbnail Image

Study: Reports of nonconsensual nude images are ignored on X

2024-10-10
Mashable SEA
Why's our monitor labelling this an incident or hazard?
The event describes the creation and dissemination of AI-generated deepfake nude images without consent, which were reported but not removed by the platform under its nonconsensual nudity policy. The AI system's involvement is explicit in the generation of deepfake images. The harm is realized as the images remain accessible, causing violation of rights and harm to communities. The platform's failure to remove such content despite reports indicates a malfunction or failure in the AI-assisted content moderation system. Hence, this is an AI Incident involving harm to rights and communities due to the AI system's outputs and the platform's inadequate mitigation.
Thumbnail Image

Twitter Acts Fast on Nonconsensual Nudity If It Thinks It's a Copyright Violation

2024-10-08
404 Media
Why's our monitor labelling this an incident or hazard?
The article involves AI-generated nude images (AI system involvement) and discusses the harm of nonconsensual intimate media, which is a violation of rights. However, the event is a research study testing platform responses and highlighting policy and legal gaps rather than reporting an actual AI Incident where harm has directly or indirectly occurred. It also does not describe a plausible future harm event but rather the systemic issue and need for legal reform. Thus, it fits the definition of Complementary Information, providing supporting data and context to understand AI-related harms and governance challenges.