
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
A Dutch court has banned the AI chatbot Grok, owned by xAI, from generating non-consensual nude images and child sexual abuse material in the Netherlands. The ruling follows evidence that Grok's 'spicy mode' enabled the creation and distribution of illegal, harmful AI-generated images, prompting legal action by Offlimits and Fonds Slachtofferhulp.[AI generated]
Why's our monitor labelling this an incident or hazard?
The AI system (generative AI used in 'undressing apps' and Grok chatbot) has directly led to harm by enabling the creation and spread of non-consensual sexualized images, violating privacy rights and causing social harm, especially to minors and female politicians. The legal ruling and EU ban are responses to this realized harm. The presence of AI is explicit, the harm is direct and ongoing, and the event centers on addressing this harm. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information.[AI generated]