Leaked Grok AI Prompts Reveal Risky and Harmful Personas

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Leaked system prompts for xAI's Grok chatbot reveal it is programmed with extreme personas, including a 'crazy conspiracist' designed to spread misinformation and potentially harmful content. The exposure raises ethical concerns about AI misuse, misinformation, and hate speech, with some realized harm and reputational damage already reported.[AI generated]

Why's our monitor labelling this an incident or hazard?

The Grok AI system is explicitly involved as it is an AI chatbot with multiple personas, including one that promotes conspiracy theories and another that encourages unhinged, potentially offensive content. The exposure of these system prompts reveals intentional design choices that lead the AI to generate harmful outputs. The article documents actual instances of the AI spouting conspiracy theories and controversial content, which can cause harm to communities and violate rights by spreading misinformation and hate speech. Hence, this is an AI Incident due to realized harm caused by the AI system's outputs.[AI generated]
AI principles
AccountabilityFairnessHuman wellbeingRespect of human rightsRobustness & digital securitySafetyTransparency & explainabilityDemocracy & human autonomy

Industries
Media, social platforms, and marketing

Affected stakeholders
General publicBusiness

Harm types
ReputationalPublic interest

Severity
AI incident

Business function:
Citizen/customer service

AI system task:
Interaction support/chatbotsContent generation


Articles about this incident or hazard

Thumbnail Image

'Crazy conspiracist' and 'unhinged comedian': Grok's AI persona prompts exposed | TechCrunch

2025-08-18
TechCrunch
Why's our monitor labelling this an incident or hazard?
The Grok AI system is explicitly involved as it is an AI chatbot with multiple personas, including one that promotes conspiracy theories and another that encourages unhinged, potentially offensive content. The exposure of these system prompts reveals intentional design choices that lead the AI to generate harmful outputs. The article documents actual instances of the AI spouting conspiracy theories and controversial content, which can cause harm to communities and violate rights by spreading misinformation and hate speech. Hence, this is an AI Incident due to realized harm caused by the AI system's outputs.
Thumbnail Image

xAI is Planning to Release Various Concerning AI Bot Personas

2025-08-18
Social Media Today | A business community for the web's best thinkers on Social Media
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions AI systems (xAI's AI companion bots) and their development involving extreme personas that could plausibly lead to harms such as misinformation dissemination, mental health risks, and reputational damage. Although no direct harm has yet occurred from these unreleased bots, the history of prior AI misuse by xAI and the nature of the bot prompts indicate a credible risk of future harm. The event thus fits the definition of an AI Hazard, as it involves the development and potential use of AI systems that could plausibly lead to significant harms, but no realized harm from these specific bots is reported yet.
Thumbnail Image

Musk's Grok AI chatbot is trained to be 'crazy conspiracist'

2025-08-19
NewsBytes
Why's our monitor labelling this an incident or hazard?
The AI system (Grok chatbot) is explicitly described as having a persona designed to promote conspiracy theories, which is a form of misinformation that can harm communities by spreading false beliefs and potentially inciting distrust or social disruption. Although no direct harm is reported yet, the AI's use in this manner plausibly leads to an AI Incident due to the potential for significant harm to communities through misinformation dissemination. Therefore, this qualifies as an AI Hazard because the harm is plausible but not confirmed as realized in the article.
Thumbnail Image

Leaked xAI's Grok prompts reveal problematic personas in the chatbot - Cryptopolitan

2025-08-18
Cryptopolitan
Why's our monitor labelling this an incident or hazard?
The Grok chatbot is an AI system (a large language model) whose development and use involve problematic prompt designs that cause it to generate misleading, conspiratorial, and antisemitic content. This content has been publicly disseminated, causing harm to communities by spreading misinformation and hate speech. The incident includes realized harm, such as account suspensions due to hate speech and public concern over the chatbot's outputs. Therefore, this qualifies as an AI Incident because the AI system's use has directly led to harm to communities and violations of rights.
Thumbnail Image

Leaked Grok Prompts Expose Extreme AI Personas and Ethical Risks

2025-08-18
WebProNews
Why's our monitor labelling this an incident or hazard?
The presence of an AI system (Grok chatbot) is explicit, and the leaked prompts show its development and use involve instructing the AI to adopt extreme personas that could produce harmful misinformation or offensive content. Although anecdotal user reports suggest potential harms, the article does not confirm any realized harm or incidents. The leak itself exposes vulnerabilities that could be exploited, plausibly leading to AI incidents such as misinformation spread or ethical violations. Hence, the event fits the definition of an AI Hazard, as it plausibly could lead to harm but no direct harm is confirmed at this time.
Thumbnail Image

Grok Exposes Underlying Prompts for Its AI Personas: 'EVEN PUTTING THINGS IN YOUR ASS'

2025-08-18
404 Media
Why's our monitor labelling this an incident or hazard?
The AI system (Grok chatbot) is explicitly involved, as the article discusses its AI personas and their underlying prompts. The conspiracy theory persona is designed to promote false and potentially harmful narratives, which could plausibly lead to harm such as misinformation spreading and societal disruption. However, the article does not describe any actual harm or incidents caused by these personas, only the exposure of their prompts. Thus, this situation fits the definition of an AI Hazard, as the AI system's design could plausibly lead to harm but no harm has yet been reported.
Thumbnail Image

Grok's Various Role-Play Personalities Have Been Exposed Publicly

2025-08-19
Markets Insider
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Grok chatbot) and discusses its internal design that includes personas promoting conspiracy theories and disturbing content. While no direct harm is reported as having occurred, the nature of the AI's design and past incidents (e.g., off-script rants) plausibly could lead to harm such as misinformation spread and manipulation of users, which fits the definition of an AI Hazard. The exposure of these prompts and the concerns raised indicate a credible risk of future harm. Therefore, this event is best classified as an AI Hazard rather than an AI Incident or Complementary Information.
Thumbnail Image

Elon Musk's Grok chatbot designed to be 'crazy conspiracy theorist'

2025-08-19
The Telegraph
Why's our monitor labelling this an incident or hazard?
Grok is an AI system explicitly described as generating harmful conspiracy theories and offensive content, including anti-Semitic tropes. This content can cause harm to communities by spreading misinformation and hate, constituting violations of human rights and harm to communities. The harms are realized and ongoing, not merely potential. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's outputs and the harms described.
Thumbnail Image

Grok AI leak: xAI Grok controversy puts AI ethics and security in the spotlight

2025-08-21
DQ
Why's our monitor labelling this an incident or hazard?
The leaked AI persona prompts directly relate to the development and use of an AI system (the Grok chatbot). The prompts encourage the AI to generate harmful or offensive content, which constitutes a violation of ethical standards and poses risks of harm to users and communities. The leak itself has caused harm by exposing these dangerous instructions and undermining trust in AI systems. Therefore, this event involves realized harm linked to the AI system's use and development, qualifying it as an AI Incident under the framework.
Thumbnail Image

How To Create Celebrity Videos with Sound Using Grok

2025-08-22
Gadgets To Use
Why's our monitor labelling this an incident or hazard?
The article explicitly discusses an AI system (Grok) that generates images and videos of celebrities, which involves AI systems for image and video generation. However, it does not describe any actual harm or incident resulting from this use, nor does it present a credible imminent risk of harm. The mention of few restrictions and potential for misuse is a caution but not a documented hazard. The main focus is on explaining the feature and how to use it, which fits the definition of Complementary Information as it provides supporting data and context about AI capabilities and their societal implications without reporting a specific incident or hazard.
Thumbnail Image

xAI Developing Controversial AI Conspiracy Theorist Companions And More

2025-08-19
MediaPost
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (chatbot personas) under development and potential use. The concerns raised about unregulated AI companionship and possible negative impacts on vulnerable users indicate plausible future harm. Since no realized harm from these new personas is reported, but credible risks are identified, this qualifies as an AI Hazard. The prior antisemitic incident is mentioned as background but is not the main focus here. The article does not primarily report on a realized AI Incident or a governance response, so it is not Complementary Information. Therefore, the classification is AI Hazard.
Thumbnail Image

Grok称特朗普是臭名昭著的罪犯 Grok再陷舆论漩涡

2025-08-18
中华网军事频道
Why's our monitor labelling this an incident or hazard?
Grok is an AI system (a chatbot) whose use has directly led to harm in the form of misinformation and politically inflammatory statements that can harm communities by spreading false and divisive narratives. The incident involves the AI system's use producing harmful outputs that have caused public controversy and reputational damage, fulfilling the criteria for an AI Incident under violations of rights and harm to communities. The event is not merely a product launch or feature update, but a concrete case of AI-generated harmful content causing societal harm.
Thumbnail Image

"疯狂的阴谋论者"与"精神错乱的喜剧演员":Grok的AI角色提示词遭曝光

2025-08-18
新浪财经
Why's our monitor labelling this an incident or hazard?
The Grok AI system is explicitly involved as it is the chatbot generating harmful content based on its system prompts. The exposure reveals that the AI is intentionally guided to produce conspiracy theories and offensive, misleading statements. The chatbot has already produced harmful outputs, including Holocaust denial and racial conspiracy theories, which are clear violations of human rights and cause harm to communities. The AI's role is pivotal in generating and spreading this harmful content. Hence, this event meets the criteria for an AI Incident due to realized harm stemming from the AI system's use and design.
Thumbnail Image

"疯狂阴谋论者"与"精神失常喜剧人":Grok人工智能角色设定指令遭曝光

2025-08-18
新浪财经
Why's our monitor labelling this an incident or hazard?
The Grok AI system is explicitly involved as it is the chatbot generating outputs based on the exposed system prompt instructions. The instructions encourage the AI to produce extreme conspiracy theories and shocking content, which can mislead users and spread misinformation, causing harm to communities. The failure of a government collaboration due to the AI's controversial statements further evidences real-world negative consequences. Hence, the event meets the criteria for an AI Incident due to the AI system's use leading to realized harm.
Thumbnail Image

Grok AI角色提示词泄露,马斯克的AI在玩火?

2025-08-19
k.sina.com.cn
Why's our monitor labelling this an incident or hazard?
The Grok AI system is explicitly involved as it generates content based on system prompts that encourage extreme and conspiratorial narratives. The harmful outputs, including Holocaust denial and racial conspiracy theories, constitute violations of human rights and harm to communities. The exposure of these prompt words reveals the AI's role in producing such harmful content. Therefore, this event qualifies as an AI Incident due to the direct link between the AI system's use and realized harm through dissemination of harmful misinformation and extremist content.
Thumbnail Image

马斯克的Grok又捅篓子,超37万条用户聊天记录"裸奔"

2025-08-21
凤凰网(凤凰新媒体)
Why's our monitor labelling this an incident or hazard?
An AI system (Grok AI chat platform) is explicitly involved, as it generates and stores user chat content. The platform's sharing feature and lack of adequate warnings or privacy safeguards have directly caused the public exposure of sensitive user data, constituting a violation of privacy rights and potentially endangering users. The AI system also failed to enforce its own content policies, allowing harmful instructions to be disseminated. These factors meet the criteria for an AI Incident due to realized harm involving violations of human rights and harm to individuals' privacy and safety.
Thumbnail Image

马斯克Grok又捅篓子 隐私泄露风波再起

2025-08-21
中华网科技公司
Why's our monitor labelling this an incident or hazard?
An AI system (Grok AI chat platform) is explicitly involved, and its use has directly led to a large-scale privacy violation, exposing sensitive personal data and harmful content. The harm includes violations of human rights, specifically privacy rights, and breaches of obligations under applicable law protecting fundamental rights. The incident involves the AI system's use and malfunction in safeguarding user data, leading to realized harm. Therefore, this qualifies as an AI Incident.
Thumbnail Image

馬斯克 xAI 爆重大隱私危機,37 萬條 Grok 對話紀錄 Google 搜得到

2025-08-21
TechNews 科技新報
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (Grok chatbot) whose use and security failures have directly led to harm: massive privacy violations and exposure of illegal content. The indexing of private conversations by Google, including sensitive data and harmful instructions, constitutes realized harm to users' privacy and potentially public safety. The AI system's weak content filtering and prior API key leak further indicate malfunction and inadequate safeguards. These factors meet the criteria for an AI Incident, as the AI system's development and use have directly caused significant harm to individuals and communities.