Thousands of Explicit AI 'Girlfriend' Ads Flood Meta Platforms, Exposing Moderation Failures

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Meta platforms, including Facebook and Instagram, have hosted at least 29,000 ads for explicit AI-generated 'girlfriend' chatbots, many violating adult content policies. The proliferation of these AI-generated deepfake and NSFW ads highlights failures in Meta's AI moderation systems, exposing users to inappropriate content and raising concerns about discriminatory enforcement against human sex workers.[AI generated]

Why's our monitor labelling this an incident or hazard?

The ads explicitly involve AI-generated or AI-related content ('AI girlfriend' ads) that are sexually explicit and violate Meta's adult content policies. The AI system's use in generating or promoting these ads directly leads to harm in the form of policy violations and exposure to inappropriate content on social media platforms, which harms community standards and user experience. The event describes realized harm (ads are active and violating policies), not just potential harm. Therefore, it meets the criteria for an AI Incident due to the direct role of AI in causing harm to communities through policy violations and exposure to explicit content.[AI generated]
AI principles
AccountabilityFairnessSafetyRobustness & digital securityTransparency & explainabilityRespect of human rightsHuman wellbeingPrivacy & data governance

Industries
Media, social platforms, and marketingConsumer services

Affected stakeholders
ConsumersWorkers

Harm types
PsychologicalReputationalEconomic/PropertyHuman or fundamental rights

Severity
AI incident

Business function:
Marketing and advertisementMonitoring and quality control

AI system task:
Content generationInteraction support/chatbotsRecognition/object detectionEvent/anomaly detectionOrganisation/recommenders


Articles about this incident or hazard

Thumbnail Image

Thousands Of Explicit AI 'Girlfriend' Ads Found On Meta Platforms

2024-04-29
Mashable India
Why's our monitor labelling this an incident or hazard?
The ads explicitly involve AI-generated or AI-related content ('AI girlfriend' ads) that are sexually explicit and violate Meta's adult content policies. The AI system's use in generating or promoting these ads directly leads to harm in the form of policy violations and exposure to inappropriate content on social media platforms, which harms community standards and user experience. The event describes realized harm (ads are active and violating policies), not just potential harm. Therefore, it meets the criteria for an AI Incident due to the direct role of AI in causing harm to communities through policy violations and exposure to explicit content.
Thumbnail Image

Mark Zuckerberg's Meta Allows Facebook and Instagram to Be Flooded with Explicit 'AI Girlfriend' Ads

2024-04-26
Breitbart
Why's our monitor labelling this an incident or hazard?
The event clearly involves AI systems generating explicit content and their promotion on major social media platforms, which directly leads to harm in the form of discriminatory content moderation and potential violation of rights of human sex workers. The AI-generated ads flood the platforms with sexually explicit material that violates stated policies, while human users face harsher restrictions, indicating a systemic issue linked to AI content and platform governance. This constitutes harm to communities and a violation of rights, fulfilling the criteria for an AI Incident. The AI system's role is pivotal as the content is AI-generated and the moderation policies' inconsistent application relates directly to the AI content's presence. Although no physical injury is reported, the social and rights-based harms are significant and realized, not merely potential.
Thumbnail Image

Sex workers outraged as 'AI girlfriend' ads flood Instagram and...

2024-04-30
New York Post
Why's our monitor labelling this an incident or hazard?
The event involves AI systems generating explicit sexual content and interactive AI companions advertised on Meta's platforms. The AI's use directly leads to harm in the form of discriminatory treatment of human sex workers (a violation of labor and fundamental rights) and harm to communities by normalizing and proliferating explicit AI-generated sexual content without adequate moderation. The AI system's development and use are central to the issue, and the harm is ongoing and realized, not merely potential. Hence, it meets the criteria for an AI Incident under violations of rights and harm to communities.
Thumbnail Image

Meta pushing Ads to generate more revenue, but most of these ads are now deepfakes

2024-04-29
India Today
Why's our monitor labelling this an incident or hazard?
The event explicitly involves AI systems generating deepfake content used in ads that promote non-consensual and obscene material, which is a violation of rights and causes harm to users and communities. The AI's role is pivotal as the deepfakes are AI-generated and are the core of the harmful content. The harm is realized and ongoing, not just potential. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Thousands of explicit AI 'girlfriend' ads found on Meta platforms

2024-04-28
Mashable
Why's our monitor labelling this an incident or hazard?
The event clearly involves AI systems generating explicit content ads ('AI girlfriends') on Meta platforms. The ads violate platform policies banning adult content and sexual solicitation, which constitutes a breach of community standards and potentially user rights. Meta's AI-based content moderation systems are failing to detect and remove these ads effectively, leading to the continued presence of harmful content. This failure or malfunction in AI moderation systems directly leads to harm by exposing users to inappropriate and potentially exploitative content. Hence, the event meets the criteria for an AI Incident as the AI system's use and malfunction have directly led to harm (violation of platform policies and exposure to explicit content).
Thumbnail Image

Explicit AI girlfriend chatbot ads are flooding Facebook and Instagram

2024-04-29
Quartz
Why's our monitor labelling this an incident or hazard?
The article involves AI systems (generative AI chatbots) that produce explicit content and collect data, which can be linked to potential harms such as privacy violations and dissemination of inappropriate content. However, the article does not document a specific realized harm or incident directly caused by these AI systems. Instead, it focuses on the prevalence of ads and the platforms' responses to policy violations. Therefore, this is best classified as Complementary Information, as it provides context and updates on ongoing issues and responses related to AI systems but does not describe a concrete AI Incident or a plausible AI Hazard event.
Thumbnail Image

Instagram and Facebook are being flooded with AI girlfriends - Softonic

2024-04-29
Softonic
Why's our monitor labelling this an incident or hazard?
The event involves AI systems generating explicit sexual content ads that violate platform policies, which constitutes a violation of community standards and could be considered harm to communities or users exposed to such content. The AI system's use in creating and distributing these ads is directly linked to the harm (policy violations and exposure to inappropriate content). Therefore, this qualifies as an AI Incident due to realized harm caused by AI-generated content violating rights and community standards.