OpenAI loosens ChatGPT content rules with adult mode, reduces censorship

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

OpenAI has introduced an “adult mode” for ChatGPT and relaxed its censorship policies, allowing the AI to generate explicit sexual and violent content when context is provided. Users have already shared erotic scenes on social media. The company also plans to adjust its training to further limit topic restrictions, sparking concerns about abuse and misinformation.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves an AI system (ChatGPT) whose use policies have been changed to allow more explicit content generation. The changes could plausibly lead to harms such as dissemination of abusive sexual content, revenge porn, or privacy violations, as warned by human rights groups. However, no specific incident of harm is reported as having occurred due to these changes. The event thus fits the definition of an AI Hazard, where the development or use of the AI system could plausibly lead to an AI Incident in the future. It is not Complementary Information because the article focuses on the policy change and its implications rather than updates on a past incident. It is not Unrelated because the AI system and its use are central to the event.[AI generated]
AI principles
SafetyHuman wellbeingRobustness & digital securityRespect of human rightsAccountabilityTransparency & explainabilityDemocracy & human autonomy

Industries
Media, social platforms, and marketingConsumer services

Affected stakeholders
ConsumersGeneral publicChildren

Harm types
PsychologicalPublic interest

Severity
AI hazard

AI system task:
Content generationInteraction support/chatbots


Articles about this incident or hazard

Thumbnail Image

ChatGPT afrouxa regras para conteúdo erótico e violência - 19/02/2025 - Tec - Folha

2025-02-19
Folha de S.Paulo
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (ChatGPT) whose use policies have been changed to allow more explicit content generation. The changes could plausibly lead to harms such as dissemination of abusive sexual content, revenge porn, or privacy violations, as warned by human rights groups. However, no specific incident of harm is reported as having occurred due to these changes. The event thus fits the definition of an AI Hazard, where the development or use of the AI system could plausibly lead to an AI Incident in the future. It is not Complementary Information because the article focuses on the policy change and its implications rather than updates on a past incident. It is not Unrelated because the AI system and its use are central to the event.
Thumbnail Image

OpenAI flexibiliza diretrizes do ChatGPT e promete menos censura

2025-02-16
Olhar Digital - O futuro passa primeiro aqui
Why's our monitor labelling this an incident or hazard?
The article focuses on a policy change in the development and use of an AI system (ChatGPT) that could plausibly lead to future harms such as misinformation or extremist discourse dissemination. However, no direct or indirect harm has yet occurred or been reported. Therefore, this event fits the definition of an AI Hazard, as the new guidelines could plausibly lead to AI incidents involving harm to communities through misinformation or harmful content spread. It is not Complementary Information because it is not an update or response to a past incident but a new policy change with potential future risks. It is not an AI Incident because no harm has materialized yet.
Thumbnail Image

ChatGPT lança 'modo adulto' e afrouxa regras para conteúdo erótico e violência

2025-02-19
Jornal de Brasília
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (ChatGPT) whose use and policy changes have directly led to the generation of explicit sexual and violent content, which can cause harm to individuals and communities, including potential violations of human rights and facilitation of abusive practices. The article reports that such content is already being generated and shared, indicating realized harm rather than just potential. The involvement of the AI system in producing this content and the associated risks and harms meet the criteria for an AI Incident. Although the article also discusses potential future risks, the realized harms and direct role of the AI system in generating harmful content take precedence, classifying this as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

OpenAI vai alterar o treino da IA para aplicar menos "censura" | TugaTech

2025-02-16
TugaTech
Why's our monitor labelling this an incident or hazard?
The article primarily reports on OpenAI's intentions to adjust its AI training methodology to allow less censorship and more freedom of expression in responses. This is a policy and development update without any described incident or harm caused by the AI systems. There is no mention of any realized injury, rights violation, or other harms linked to the AI's use or malfunction. Nor does it describe a credible imminent risk of harm that would qualify as an AI Hazard. Therefore, the article fits best as Complementary Information, providing context on AI development and governance responses rather than reporting an AI Incident or Hazard.