AI-Generated Pornography and Illegal Content Distribution Chain Exposed in China

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Multiple investigations reveal a widespread illegal industry in China using AI to generate and distribute pornographic content, including deepfake videos and explicit chat software. Tutorials and tools are openly sold online, enabling mass production and evasion of regulation, causing harm to individuals and exposing minors to inappropriate material.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly mentions AI technology being used to generate large-scale illegal pornographic videos and content, facilitated by tutorials that teach users how to produce and evade detection. This constitutes a direct use of AI systems leading to violations of applicable laws and harm to communities through the spread of illegal and harmful content. The AI system's role is pivotal in enabling the creation and distribution of this content at scale, fulfilling the criteria for an AI Incident under violations of law and harm to communities. Therefore, this event is classified as an AI Incident.[AI generated]
AI principles
Respect of human rightsPrivacy & data governance

Industries
Media, social platforms, and marketing

Affected stakeholders
ChildrenGeneral public

Harm types
PsychologicalHuman or fundamental rightsReputational

Severity
AI incident

Business function:
Other

AI system task:
Content generation


Articles about this incident or hazard

Thumbnail Image

莫让AI沦为灰色产业工具 9.9元教程背后的黑产链条

2026-04-14
中华网军事频道
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI technology being used to generate large-scale illegal pornographic videos and content, facilitated by tutorials that teach users how to produce and evade detection. This constitutes a direct use of AI systems leading to violations of applicable laws and harm to communities through the spread of illegal and harmful content. The AI system's role is pivotal in enabling the creation and distribution of this content at scale, fulfilling the criteria for an AI Incident under violations of law and harm to communities. Therefore, this event is classified as an AI Incident.
Thumbnail Image

9.9元就能毁掉一个人吗 AI造黄产业链揭秘

2026-04-14
中华网科技公司
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions the use of AI generative models to create fake explicit videos by face-swapping and image generation, which is a clear AI system involvement. The harm is direct and realized, including violation of personal rights, privacy, and dignity, as well as extortion attempts. This fits the definition of an AI Incident because the AI system's use has directly led to violations of human rights and harm to individuals and communities. The detailed description of the illicit AI-enabled production and distribution chain confirms the systemic nature of the harm.
Thumbnail Image

检察日报:兜售AI"造黄"教程已触法律红线

2026-04-14
华商网
Why's our monitor labelling this an incident or hazard?
The event involves AI systems used to generate pornographic or borderline pornographic content, which is illegal and harmful. The sale of tutorials and prompts to produce such content directly facilitates the misuse of AI, leading to violations of laws and societal harm. The article describes ongoing illegal activities and regulatory responses, indicating that harm is occurring rather than merely potential. The AI system's use in producing harmful content and the illegal facilitation through tutorials constitute direct and indirect involvement leading to harm. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

AI语聊软件暗藏大尺度色情内容 境外涉黄AI软件绕过监管流入国内

2026-04-12
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI chat software generating large-scale sexual content, which is accessible to domestic users through illegal channels circumventing regulation. The AI system's use directly leads to harm by exposing users, including minors, to inappropriate and potentially harmful content, constituting harm to communities and violation of protections intended for minors. The involvement of AI in generating and distributing this content is clear, and the harm is realized, not just potential. Therefore, this qualifies as an AI Incident under the framework.
Thumbnail Image

暗藏大尺度色情内容!AI语聊软件让年轻人未成年人沉浸

2026-04-12
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI chat software generating explicit sexual content and being used by young people and minors, which is a direct harm to the health and well-being of these groups and a violation of legal and ethical norms. The AI systems' development and use have directly led to this harm. The presence of AI is clear from the description of generative AI large models used for chat companions. The harm is realized and ongoing, not just potential. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

9.9元买教程,"AI赛道"别走向"造黄邪道"| 新京报快评

2026-04-13
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The event involves the use and misuse of AI systems to generate harmful pornographic content, which directly leads to violations of human rights (women's rights) and harm to communities (network ecology and minors). The AI systems' development and use have directly led to these harms, fulfilling the criteria for an AI Incident. The article also discusses governance responses, but the primary focus is on the realized harms caused by AI misuse in the 'AI造黄' (AI pornography generation) industry chain.
Thumbnail Image

斩断AI"造黄"产业链 守住技术向善底线

2026-04-13
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The event involves the use and misuse of AI systems to generate harmful sexual content, which directly harms communities and individuals, including minors, by polluting the online environment and causing psychological and social harm. This constitutes a violation of societal norms and potentially legal rights, fulfilling the criteria for harm under the AI Incident definition. The article reports on realized harm caused by AI misuse rather than just potential risks or general commentary, so it qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

AI"造黄"是老套路披上新马甲,兜售教程已触法律红线

2026-04-14
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The article clearly involves AI systems used to generate harmful content (pornographic or borderline pornographic videos), which constitutes a violation of laws and harms communities. This fits the definition of harm caused by AI misuse. However, the article does not describe a specific new incident where harm has directly or indirectly occurred, nor does it describe a plausible future harm event that is distinct from the ongoing illegal activity. Instead, it reports on the existence of an illegal industry, regulatory frameworks, enforcement actions, and the need for continued vigilance. This aligns with the definition of Complementary Information, as it provides context, regulatory updates, and societal responses to AI misuse rather than reporting a new AI Incident or AI Hazard.
Thumbnail Image

AI语聊软件暗藏大尺度色情内容,境外涉黄AI软件绕过监管流入国内

2026-04-12
news.bjd.com.cn
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI chat software (AI systems) that generate and facilitate access to large-scale sexual content, which is harmful to users and communities, including minors. The AI systems are being used in a way that circumvents regulatory frameworks, leading to direct harm through exposure to inappropriate content. The involvement of AI in generating and delivering this content, combined with the circumvention of controls, meets the criteria for an AI Incident as the AI system's use has directly led to harm to communities and potential legal violations.
Thumbnail Image

AI语聊软件暗藏大尺度色情内容 境外涉黄AI软件绕过监管流入国内_手机网易网

2026-04-12
m.163.com
Why's our monitor labelling this an incident or hazard?
The event involves AI systems explicitly described as AI chat software with highly explicit sexual content. The software's use has led to harm by exposing users, including minors, to inappropriate content, which constitutes harm to communities and breaches legal protections. The AI systems are used in a way that causes direct harm, and the illegal distribution bypassing regulation further supports the classification as an AI Incident. Therefore, this event meets the criteria for an AI Incident due to realized harm caused by the AI system's use and distribution.
Thumbnail Image

警惕!不少人沉迷境外涉黄AI聊天软件:有用户沉迷AI涉黄聊天一周花1000

2026-04-13
t.cj.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI chat software with highly sexualized content that users are addicted to, including minors, leading to financial harm and potential psychological harm. The AI systems are being used in a way that causes direct harm to individuals and communities, fulfilling the criteria for an AI Incident. The circumvention of regulatory controls and the presence of explicit content further support the classification as an incident rather than a hazard or complementary information.
Thumbnail Image

央视曝光:AI色情视频产业链,九块九就能"定制"任何人_手机网易网

2026-04-14
m.163.com
Why's our monitor labelling this an incident or hazard?
The article explicitly describes AI systems being used to generate pornographic videos featuring the faces of real individuals without their consent, which is a violation of human rights and privacy (a breach of fundamental rights). The AI system's use directly leads to harm by creating and distributing non-consensual sexual content, causing reputational and psychological harm to victims. The involvement of AI is clear, as the videos are AI-generated using images and text prompts. The harm is realized and ongoing, not merely potential. Hence, this event meets the criteria for an AI Incident.