Meta AI App's Public Feed Exposes Users' Sensitive Data

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Meta's AI app features a public 'Discover' feed that displays user interactions, leading to unintentional sharing of sensitive personal information. Many users are unaware their posts are public, resulting in privacy violations and emotional distress, highlighting design flaws in user consent and data protection.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves an AI system (Meta's AI app) whose use has directly led to harm: users unintentionally sharing sensitive personal information publicly, causing privacy violations and emotional distress. The AI system's design includes a public 'Discover' feed that exposes user interactions, and some users are unaware of the public nature of their posts, which is a misuse or malfunction in terms of user understanding and consent. This fits the definition of an AI Incident due to violations of rights and harm to communities. The article does not describe a potential future harm but an ongoing harm, so it is not an AI Hazard. It is not merely complementary information or unrelated news because the harm is clearly articulated and linked to the AI system's use.[AI generated]
AI principles
Privacy & data governanceRespect of human rightsTransparency & explainabilityAccountability

Industries
Media, social platforms, and marketing

Affected stakeholders
Consumers

Harm types
Human or fundamental rightsPsychological

Severity
AI incident

AI system task:
Organisation/recommenders

In other databases

Articles about this incident or hazard

Thumbnail Image

Why Mark Zuckerberg's AI app is one of the most depressing places online

2025-06-11
Business Insider
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta's AI app) whose use has directly led to harm: users unintentionally sharing sensitive personal information publicly, causing privacy violations and emotional distress. The AI system's design includes a public 'Discover' feed that exposes user interactions, and some users are unaware of the public nature of their posts, which is a misuse or malfunction in terms of user understanding and consent. This fits the definition of an AI Incident due to violations of rights and harm to communities. The article does not describe a potential future harm but an ongoing harm, so it is not an AI Hazard. It is not merely complementary information or unrelated news because the harm is clearly articulated and linked to the AI system's use.
Thumbnail Image

The Meta AI App Lets You 'Discover' People's Bizarrely Personal Chats

2025-06-12
Wired
Why's our monitor labelling this an incident or hazard?
The Meta AI platform involves an AI system (a chatbot) whose use has directly led to harm in the form of privacy violations and exposure of sensitive personal information. The harm arises from the platform's design and user misunderstanding, resulting in personal data being publicly accessible, which can lead to violations of fundamental rights and potential further harms. This fits the definition of an AI Incident because the AI system's use has directly led to harm (privacy breaches and potential human rights violations).
Thumbnail Image

The Meta AI app is a privacy disaster | TechCrunch

2025-06-12
TechCrunch
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI app) whose use has directly led to harm in the form of privacy violations and exposure of sensitive personal information. Users' private conversations with the AI, including personal and legal details, are being publicly shared without clear consent or understanding, constituting a breach of privacy and potentially human rights. This meets the criteria for an AI Incident as the AI system's use has directly caused harm to individuals' rights and privacy.
Thumbnail Image

Meta AI's 'Discover' posts reveal what people actually use AI for

2025-06-12
9to5Google
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is involved, and its use has directly led to harm in the form of privacy violations and potential emotional or reputational harm to users whose personal conversations are publicly shared without full awareness or consent. The harm is to individuals' privacy and potentially to their mental health or legal standing, which falls under violations of rights and harm to communities. Therefore, this qualifies as an AI Incident because the AI system's use has directly led to realized harm through public exposure of sensitive personal data.
Thumbnail Image

Here's how Meta AI "leaks" your private chats, thanks in part to its terrible UX

2025-06-12
Neowin
Why's our monitor labelling this an incident or hazard?
The article clearly describes an AI system (Meta AI) whose use has directly led to harm: the unintentional public exposure of private, sensitive user conversations. This constitutes a violation of privacy rights and potentially other legal rights, fulfilling the criteria for harm under human rights violations. The AI system's involvement is explicit, and the harm is realized, not just potential. The root cause is the AI app's user interface design, which facilitates accidental sharing, making it an incident stemming from the AI system's use and malfunction in UX design. Hence, the event is best classified as an AI Incident.
Thumbnail Image

The Meta AI app is a privacy disaster - RocketNews

2025-06-12
RocketNews | Top News Stories From Around the Globe
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system (a chatbot) that processes user inputs and allows sharing of conversations. The incident involves the use of this AI system leading to a breach of privacy, which is a violation of human rights and legal protections related to personal data. Users are unknowingly exposing sensitive information publicly due to unclear privacy settings and interface design, which constitutes harm to individuals' rights. Therefore, this qualifies as an AI Incident because the AI system's use has directly led to violations of privacy rights and harm to users.
Thumbnail Image

Meta AI Personal Chats Appearing In Public Search Feed

2025-06-13
MediaPost
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use has directly led to harm in the form of privacy violations and exposure of sensitive personal information. Users unintentionally share private conversations publicly due to misunderstanding the platform's sharing mechanism, causing real harm to individuals' privacy and potentially violating their rights. This fits the definition of an AI Incident because the AI system's use has directly led to harm to groups of people (privacy breach) and harm to communities (loss of trust, exposure of sensitive data).
Thumbnail Image

Nutzer von Meta AI veröffentlichen ihre Anfragen - offenbar unbewusst

2025-06-14
heise online
Why's our monitor labelling this an incident or hazard?
Meta AI is an AI system that processes user prompts and generates responses. The article reports that users' sensitive and personal prompts are publicly visible in a feed, sometimes unintentionally, leading to privacy breaches. This is a direct harm to individuals' rights and privacy, fulfilling the criteria for an AI Incident. The AI system's use and design (sharing prompts publicly) have directly led to this harm. Although Meta claims to require user consent, the article suggests users may not fully understand the implications, which still results in harm. Hence, this is not merely a hazard or complementary information but an incident involving realized harm.
Thumbnail Image

Meta investe em novo laboratório para criar "superinteligência" artificial * Tecnoblog

2025-06-11
Tecnoblog
Why's our monitor labelling this an incident or hazard?
The article discusses the development and organizational strategy around advanced AI research aiming at superintelligence, which is currently hypothetical and not realized. There is no mention of any harm caused or any incident involving AI systems. The focus is on the company's plans and recruitment efforts, which are typical of AI ecosystem developments. Therefore, this is best classified as Complementary Information, as it provides context and updates on AI development and corporate strategy without describing an AI Incident or AI Hazard.
Thumbnail Image

Meta investiert Milliarden in KI - auch fürs US-Militär

2025-06-14
SRF News
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems through Scale AI's data quality technology and Meta's AI-based augmented reality helmets for military use. The event stems from the use and development of AI systems aimed at enhancing military capabilities. Although no direct harm or incident is reported, the deployment of AI in military contexts plausibly leads to harms such as injury, disruption, or violations of rights in armed conflict scenarios. The investment and integration of AI technologies for defense purposes represent a credible risk of future AI-related harm, qualifying this as an AI Hazard rather than an AI Incident. It is not Complementary Information because the main focus is on the investment and its implications, not on updates or responses to prior incidents. It is not Unrelated because AI systems and their military use are central to the event.
Thumbnail Image

Meta steigt bei Scale AI ein und schnappt sich Scale-Chef Wang

2025-06-13
de.marketscreener.com
Why's our monitor labelling this an incident or hazard?
The article details a major investment and leadership move in the AI industry but does not describe any direct or indirect harm caused by AI systems, nor does it highlight any credible risk of harm stemming from this event. It is primarily about business developments and strategic positioning in AI, without reporting incidents, hazards, or responses to harms. Therefore, it fits best as Complementary Information, providing context and updates on AI ecosystem developments without constituting an AI Incident or AI Hazard.
Thumbnail Image

Meta vai criar novo laboratório para desenvolver IA "superinteligente"

2025-06-10
Brasil 247
Why's our monitor labelling this an incident or hazard?
The article discusses the establishment of a new AI research lab by Meta aiming to develop superintelligent AI, which is a theoretical concept with potential future risks. However, there is no indication that any harm has occurred or that the AI system has malfunctioned or been misused. The event is about the development and strategic plans in AI, which could plausibly lead to future risks but does not describe any immediate or realized harm. Therefore, it fits the definition of an AI Hazard, as the development of superintelligent AI could plausibly lead to significant future harm, but no incident has yet occurred.
Thumbnail Image

A Meta também está determinada em criar uma 'super IA com capacidade sobrehumana'

2025-06-11
Estadão
Why's our monitor labelling this an incident or hazard?
The article centers on Meta's strategic initiative to develop advanced AI systems, including superintelligence, which is a future-oriented goal. While the development of such powerful AI systems carries plausible risks of harm in the future, the article does not describe any actual harm, malfunction, or misuse of AI systems at present. Therefore, it fits the definition of an AI Hazard due to the plausible future risks of superintelligent AI, but it is not an AI Incident. It is not Complementary Information because it does not provide updates or responses to existing incidents or hazards, nor is it unrelated since it clearly involves AI systems and their development.
Thumbnail Image

Meta beteiligt sich an Scale AI und schnappt sich Scale-Chef Wang

2025-06-13
Cash
Why's our monitor labelling this an incident or hazard?
The article describes a major corporate investment and leadership transition involving AI companies and projects. However, it does not report any realized or potential harm caused by AI systems, nor does it describe incidents or hazards related to AI misuse, malfunction, or risks. The content focuses on business strategy and collaboration in AI development, which is general AI-related news without direct or plausible harm implications. Therefore, it qualifies as Complementary Information, providing context on AI ecosystem developments rather than an AI Incident or Hazard.
Thumbnail Image

Frustrado com o Llama 4, Zuckerberg cria equipa de elite para atingir a superinteligência na Meta | TugaTech

2025-06-10
TugaTech
Why's our monitor labelling this an incident or hazard?
The article centers on the strategic recruitment and investment by Meta to develop superintelligent AI, which is a future-oriented goal. There is no mention of any realized harm, malfunction, or misuse of AI systems. The discussion of superintelligence implies potential future risks associated with such powerful AI systems. Therefore, this event fits the definition of an AI Hazard, as it plausibly could lead to AI incidents in the future due to the nature of superintelligent AI development and its associated risks. It is not an AI Incident because no harm has occurred yet, nor is it Complementary Information or Unrelated since it directly concerns AI development with potential future harm.
Thumbnail Image

Meta met 15 milliards sur la table pour contrer Open AI

2025-06-12
Les Smartgrids
Why's our monitor labelling this an incident or hazard?
The article focuses on Meta's large-scale investment and strategic shift to develop advanced AI capabilities, including superintelligence. While the concept of superintelligence carries potential future risks, the article does not report any actual harm or incidents caused by AI, nor does it describe a credible immediate hazard. The event is about development and strategic positioning, with no realized or imminent harm. Thus, it fits the definition of Complementary Information, as it provides important context and updates on AI ecosystem evolution without constituting an AI Incident or AI Hazard.
Thumbnail Image

Mark Zuckerberg forma un team segreto per sviluppare una superintelligenza artificiale - Il Fatto Quotidiano

2025-06-12
Il Fatto Quotidiano
Why's our monitor labelling this an incident or hazard?
The article describes the formation and development of a new AI system intended to be a superintelligence, which could plausibly lead to significant future harms given the nature of such technology. However, no actual harm or incident has occurred yet; the focus is on the potential and strategic development. Therefore, this qualifies as an AI Hazard because the development of a superintelligent AI system could plausibly lead to AI incidents in the future, but no direct or indirect harm has been reported at this stage.
Thumbnail Image

Zuckerberg paga 14.300 millones para entrar en una empresa de IA y fichar a su jefe de 28 años

2025-06-13
eldiario.es
Why's our monitor labelling this an incident or hazard?
The article involves AI systems (Scale AI's data labeling for machine learning) and their development and use, but no harm has occurred or is described as occurring. The mention of military AI applications and past controversies suggests potential future risks, but these are not the main focus nor are they described as imminent or realized harms. The event is primarily about corporate acquisition and strategic positioning in AI, which fits the definition of Complementary Information as it informs about broader AI ecosystem developments and governance-related implications without reporting a specific AI Incident or Hazard.
Thumbnail Image

La carrera multimillonaria por la "superinteligencia", tecnología que podría estar cerca - Proceso Digital

2025-06-13
Proceso Hn
Why's our monitor labelling this an incident or hazard?
The article primarily covers the race and investment towards developing superintelligence, which is a future possibility rather than a current event causing harm. There is no mention of any AI system causing injury, rights violations, disruption, or other harms at present. The discussion is about potential future capabilities and timelines, making it a plausible future risk but without concrete evidence of harm or incident. Therefore, it fits the definition of Complementary Information as it provides context and updates on AI development and ecosystem without describing an AI Incident or AI Hazard.
Thumbnail Image

Le preguntan a Meta AI cómo encontrar una novia con buen trasero y la IA publica toda la conversación

2025-06-16
Xataka Android
Why's our monitor labelling this an incident or hazard?
The event involves the use and malfunction of an AI system (Meta AI based on Llama 4) that has directly led to harm by exposing private and sensitive user data publicly without consent. This constitutes a violation of human rights, specifically privacy rights, and breaches obligations to protect fundamental rights. The harm is realized and ongoing, as private conversations are being shared publicly, causing potential and actual harm to individuals. Therefore, this qualifies as an AI Incident.
Thumbnail Image

Preguntas y respuestas sobre la IA de Meta y el uso de tus datos en Instagram y Facebook. ¿Qué pueden usar? ¿Pueden utilizar los de WhatsApp? ¿Cómo oponerme? ¿Y si aún no lo he hecho?·Maldita.es - Periodismo para que no te la cuelen

2025-06-16
Maldita.es — Periodismo para que no te la cuelen
Why's our monitor labelling this an incident or hazard?
The article focuses on explaining Meta's data usage policies for AI training and the mechanisms for users to oppose such use. It does not describe any realized harm (such as privacy breaches or misuse leading to injury or rights violations) nor does it highlight a credible risk of future harm from AI systems. It serves as informative content to help users understand and respond to Meta's AI data practices, fitting the definition of Complementary Information rather than an AI Incident or AI Hazard.
Thumbnail Image

¿Has usado la IA de Meta? Pues ten mucho cuidado

2025-06-13
MuyComputer
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI assistant) is explicitly involved, and its use has directly led to harm in the form of privacy violations and exposure of sensitive personal data. The harm is realized, not just potential, as private conversations have been made public without clear user consent or notification. This constitutes a violation of user rights and harm to individuals, fitting the definition of an AI Incident. The event is not merely a technical fault but a design flaw causing actual harm, and thus it is not a hazard or complementary information.
Thumbnail Image

Papelón: Meta AI publica los chats privados que los usuarios mantienen con la IA

2025-06-13
La Nacion
Why's our monitor labelling this an incident or hazard?
The Meta AI system is explicitly involved as it is the AI chatbot generating the conversations. The issue arises from the use of the AI system integrated with Instagram, where private chats are made public without users' knowledge or consent. This constitutes a violation of users' privacy rights, which falls under violations of human rights or breaches of obligations intended to protect fundamental rights. The harm is realized as private, sensitive information including medical, legal, and personal data is exposed publicly, directly linked to the AI system's use and its integration with Instagram. Therefore, this event qualifies as an AI Incident due to the direct harm to users' privacy caused by the AI system's operation and sharing mechanism.
Thumbnail Image

Meta AI expone chats privados de usuarios al integrarse con Instagram

2025-06-13
Montevideo Portal / Montevideo COMM
Why's our monitor labelling this an incident or hazard?
The Meta AI chatbot is an AI system involved in the incident. Its malfunction or design flaw in the 'share' feature causes private chats to be published publicly without user awareness or consent. This has directly led to harm in the form of privacy violations and exposure of sensitive personal data, which constitutes a breach of fundamental rights and applicable privacy laws. Therefore, this qualifies as an AI Incident due to realized harm caused by the AI system's use and malfunction.
Thumbnail Image

Meta AI expone chats privados de usuarios sin previo aviso

2025-06-13
El Vocero de Puerto Rico
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is explicitly involved, and its use has directly led to harm in the form of violations of privacy and potentially other human rights due to unauthorized public disclosure of private conversations. The harm is realized and ongoing, as users' private data is exposed without their knowledge or consent. This fits the definition of an AI Incident under violations of human rights or breach of obligations intended to protect fundamental rights.
Thumbnail Image

La aplicación Meta AI publica los chats privados que los usuarios mantienen con el asistente de IA

2025-06-13
NoticiasDe.es
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI assistant) is involved as it generates and manages private chat content. The malfunction or design of the system (automatic sharing of private chats without clear user consent or notification) has directly led to harm in the form of privacy violations and exposure of sensitive personal data. This fits the definition of an AI Incident under violations of human rights and breach of privacy obligations. The harm is realized and ongoing, not merely potential.
Thumbnail Image

Un fallo en la aplicación Meta AI expone chats privados sin avisar a los usuarios

2025-06-14
20 minutos
Why's our monitor labelling this an incident or hazard?
The incident involves an AI system (Meta AI chatbot) whose malfunction (automatic public posting of private chats without consent) has directly caused harm to users by exposing private and sensitive information publicly. This is a clear violation of privacy rights, which falls under violations of human rights and legal protections. Therefore, this event qualifies as an AI Incident due to the realized harm caused by the AI system's malfunction and use.
Thumbnail Image

Saltan las alarmas con Meta AI: las conversaciones que se tienen con el asistente salen a la luz de forma pública

2025-06-14
MARCA
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI assistant) is involved, and its use has directly led to a violation of users' privacy rights, which constitutes a breach of fundamental rights. The harm is realized as private information is made public without consent, causing potential harm to individuals' privacy and trust. Therefore, this qualifies as an AI Incident under the framework, specifically under violations of human rights or breach of obligations intended to protect fundamental rights.
Thumbnail Image

¿Conversas con Meta AI? La aplicación estaría publicando los chats sin informarlo a los usuarios

2025-06-14
BioBioChile
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta AI chatbot) whose use has directly caused harm by exposing private user conversations publicly without consent. This is a clear violation of privacy rights, a fundamental human right, and thus fits the definition of an AI Incident. The harm is realized and ongoing, not merely potential, and the AI system's role is pivotal as it is the mechanism through which private data is made public. Therefore, the event qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

El "interés legítimo" del que habla Meta para usar nuestros datos de Instagram y Facebook: qué implica y qué podemos hacer al respecto·Maldita.es - Periodismo para que no te la cuelen

2025-06-16
Maldita.es — Periodismo para que no te la cuelen
Why's our monitor labelling this an incident or hazard?
The article centers on Meta's data privacy policy changes and the legal framework allowing AI training on user data without explicit consent, raising concerns about user control and transparency. While these issues relate to AI system development and use, the article does not document any actual harm or incident resulting from these practices. It rather provides context, expert opinions, and guidance on user rights and opposition mechanisms. Therefore, it fits the definition of Complementary Information, as it enhances understanding of AI ecosystem governance and user data rights without describing a new AI Incident or AI Hazard.
Thumbnail Image

Fuga de seguridad en WhatsApp, Meta AI está revelando todas tus conversaciones

2025-06-17
MovilZona
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is explicitly involved, and its malfunction or design flaw has directly led to the exposure of sensitive personal information, including medical queries and emotional confessions. This exposure harms individuals' privacy and could be considered a violation of fundamental rights. Therefore, this qualifies as an AI Incident due to realized harm caused by the AI system's use and malfunction.
Thumbnail Image

La aplicación de Meta AI recomienda evitar compartir información personal para que no aparezca en el feed Discover

2025-06-17
NoticiasDe.es
Why's our monitor labelling this an incident or hazard?
The Meta AI assistant is an AI system involved in generating conversational outputs. The public sharing of user conversations containing personal and sensitive information caused indirect harm to users' privacy and potentially violated data protection rights. Although the sharing feature was opt-in, many users unintentionally exposed private data, leading to harm. Meta's subsequent warning and feed modification are responses to this harm. Since the AI system's use directly led to realized harm (privacy violations), this event is classified as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI Chat Privacy Concerns: User Conversations Accidentally Public - czechjournal.cz

2025-06-14
The Czech Journal
Why's our monitor labelling this an incident or hazard?
The incident involves an AI system (Meta's AI chatbot) whose malfunction (a configuration error causing data to be publicly accessible) directly led to harm in the form of privacy violations and unauthorized access to personal conversations. This fits the definition of an AI Incident because it involves realized harm (violation of user privacy) caused by the AI system's malfunction. The article details the event, its implications, and the company's response, confirming that the AI system's role was pivotal in the harm caused.
Thumbnail Image

Here's how to turn off public posting on the Meta AI app

2025-06-13
NBC10 Philadelphia
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI app) and concerns about privacy due to public sharing of user prompts, which can include personal data. However, the article does not describe any realized harm or incident resulting from this exposure, nor does it indicate a direct or indirect harm caused by the AI system. Instead, it focuses on user guidance to mitigate privacy risks and the company's AI investment strategy. Therefore, this is complementary information providing context and user advice related to AI privacy concerns, but not an AI Incident or AI Hazard.
Thumbnail Image

Your Questions in the Meta AI App Might Be Posted Publicly

2025-06-13
Lifehacker
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that generates chatbot responses and allows users to post their interactions publicly. The article documents actual cases where users unintentionally shared sensitive personal information publicly, leading to privacy and security harms. The AI system's design and default settings contributed to these harms by making it easy for users to share private data without clear awareness. This constitutes a violation of privacy rights and security, which falls under harm to persons and violation of rights as defined in the framework. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta users don't know their intimate AI chats are out there for all to see

2025-06-13
Washington Post
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use has directly led to harm in the form of privacy violations and potential emotional harm to users. The harm is realized as users' intimate conversations are publicly exposed without clear informed consent, violating their rights to confidentiality and privacy. The AI system's design and the app's discover feed feature are pivotal in causing this harm. This fits the definition of an AI Incident because it involves harm to individuals and violations of rights directly linked to the AI system's use and deployment. The article does not merely warn of potential harm (which would be a hazard) nor does it focus on responses or updates (which would be complementary information).
Thumbnail Image

Meta AI App is reportedly exposing your private chats

2025-06-13
Hindustan Times
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that processes user inputs and generates responses. The design flaw in the app's interface and privacy settings has directly led to the public exposure of private user chats, including sensitive information. This constitutes a violation of users' rights to privacy and data protection, which falls under violations of human rights or breach of obligations under applicable law. Since the harm is occurring due to the AI system's use and design, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Be Careful With Meta AI: You Might Accidentally Make Your Chats Public

2025-06-13
PCMag UK
Why's our monitor labelling this an incident or hazard?
The AI system is explicitly involved as the chatbot generating conversations. The harm arises from the use of the AI system's sharing feature, which users misunderstand, leading to unintended public disclosure of sensitive personal information. This is a violation of privacy rights, a form of human rights violation under the framework. Since the harm has materialized (private data exposure), this qualifies as an AI Incident rather than a hazard or complementary information. The event is not merely about product features or general AI news but about realized harm caused by the AI system's use and design.
Thumbnail Image

People's very intimate, very strange AI requests are being shared publicly online

2025-06-13
The Independent
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that processes user inputs and generates responses. The app uploads user conversations, including intimate and sensitive content, to a public feed visible to anyone, sometimes linked to users' social media profiles. This practice has led to privacy concerns and potential violations of users' rights to confidentiality and personal data protection, which constitute harm to individuals. The harm is realized as users' sensitive information is exposed publicly, potentially causing emotional distress and breaches of privacy rights. Therefore, this qualifies as an AI Incident due to violations of human rights and harm to individuals stemming from the AI system's use and data handling practices.
Thumbnail Image

Meta AI's discover feed is full of revealing personal info -- here's how to protect your privacy

2025-06-13
Tom's Guide
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use has directly led to the public exposure of sensitive personal information due to poor communication of privacy settings. This constitutes a violation of users' rights and privacy, fulfilling the criteria for an AI Incident. The harm is realized, not just potential, as users' private data is already publicly accessible, and the article describes the consequences and risks arising from this exposure. Therefore, this is classified as an AI Incident.
Thumbnail Image

Meta AI App Under Fire For Major Privacy Lapses: Users Accidentally Sharing Personal Chats Publicly

2025-06-13
TimesNow
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system enabling user interaction via text, audio, and images. The incident arises from the use of this AI system, specifically a malfunction or poor design of the 'share' feature and unclear privacy settings, which directly led to users' private conversations being exposed publicly. This constitutes a violation of users' privacy and potentially their rights, fitting the definition of an AI Incident as the AI system's use has directly led to harm (privacy breaches).
Thumbnail Image

Your Meta AI prompts might be public - here's how to check

2025-06-13
ZDNet
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that generates and shares user prompts. The article highlights that due to default public sharing settings and user confusion, sensitive personal information is being exposed publicly, which could lead to harm such as privacy violations and emotional distress. Although no specific harm incident is detailed, the plausible risk of harm is credible and directly linked to the AI system's design and use. Hence, this is best classified as an AI Hazard rather than an AI Incident or Complementary Information.
Thumbnail Image

Mark Zuckerberg's Meta AI is exposing your personal searches, private chats to public, here's how you can prevent it

2025-06-14
Daily News and Analysis (DNA) India
Why's our monitor labelling this an incident or hazard?
The event clearly involves an AI system (Meta AI chatbot) whose use has directly led to harm: the exposure of private and sensitive user chats to the public due to misleading design and user interface issues. This constitutes a violation of users' rights to privacy and confidentiality, which falls under violations of human rights and breach of obligations under applicable law protecting fundamental rights. The harm is realized and ongoing, not merely potential. Hence, this qualifies as an AI Incident rather than a hazard or complementary information. The AI system's role is pivotal as it is the platform through which private information is shared and exposed due to its design and operational features.
Thumbnail Image

A simple mistake in the Meta AI app could expose your deepest secrets

2025-06-13
LaptopMag
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta's Llama large language model chatbot) whose use has directly led to harm: the exposure of private, sensitive user data due to a confusing 'Share' button and vague warnings. This constitutes a violation of privacy rights, a breach of obligations intended to protect fundamental rights. The harm is realized and ongoing, not merely potential. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI app marks private chats as public, raising privacy concerns

2025-06-13
Digit
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system providing chatbot services. The incident involves the use of this AI system where private conversations are inadvertently or misleadingly made public, violating users' privacy rights and data protection laws. The harm is direct and realized, as users' sensitive information is exposed without proper consent or safeguards. The involvement of the AI system in generating and storing these conversations, combined with the flawed privacy design, directly leads to harm under the definitions of AI Incident, specifically violations of human rights and privacy obligations. The article also references regulatory concerns and user complaints, reinforcing the classification as an AI Incident rather than a mere hazard or complementary information.
Thumbnail Image

People Don't Realize Meta's AI App Is Publicly Blasting Their Humiliating Secrets to the World

2025-06-14
Futurism
Why's our monitor labelling this an incident or hazard?
The AI system involved is Meta's AI assistant app, a generative AI chatbot that processes user queries and shares them publicly in a discover feed. The harm arises from the use of the AI system, where users unknowingly share private and sensitive information publicly, leading to violations of privacy and potentially other rights. The article details actual realized harm, not just potential, as private data is already exposed. Therefore, this qualifies as an AI Incident due to direct harm to users' rights and privacy caused by the AI system's use and design flaws.
Thumbnail Image

Meta AI is a 'Privacy Disaster' -- OK Boomer

2025-06-13
Security Boulevard
Why's our monitor labelling this an incident or hazard?
The article explicitly details how Meta's AI app's social sharing feature causes private user data to be publicly exposed, including sensitive medical, legal, and personal information. This exposure constitutes a violation of privacy rights and harms individuals' confidentiality, which falls under violations of human rights and breaches of obligations intended to protect fundamental rights. The AI system's use and design are directly responsible for this harm, meeting the criteria for an AI Incident. The harm is realized, not hypothetical, and the AI system's role is pivotal in causing it.
Thumbnail Image

Your Meta AI queries about farts might be public

2025-06-14
Morning Brew
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI) that processes user queries. Due to the way the system is integrated with Instagram accounts and the lack of clear privacy settings visibility, users' sensitive personal information is being exposed publicly. This exposure directly harms users' privacy rights, a fundamental human right. The harm is realized (not just potential), as personal and sensitive data is accessible to the public. Hence, this qualifies as an AI Incident under the framework's category of violations of human rights or breach of obligations under applicable law protecting fundamental rights.
Thumbnail Image

Is privacy with Meta AI a mere myth?

2025-06-14
The Week
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use has directly led to harm in the form of privacy violations and potential breaches of users' rights. The sharing of personal conversations and images without clear user understanding or consent constitutes a violation of human rights and privacy laws. The harm is realized, not just potential, as users' private data is publicly exposed. Therefore, this qualifies as an AI Incident under the framework, specifically under violations of human rights or breach of obligations intended to protect fundamental rights.
Thumbnail Image

Here's how to keep Meta AI from sharing your prompts on Facebook, Instagram

2025-06-14
NBC 7 San Diego
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI app) that shares user prompts publicly unless privacy settings are changed. However, the sharing is user-controllable and not automatic, and no harm or violation is reported as having occurred. The article focuses on user guidance to protect privacy and Meta's AI strategy, which is complementary information about AI use and governance rather than an incident or hazard. Therefore, it does not meet criteria for AI Incident or AI Hazard but fits as Complementary Information.
Thumbnail Image

Here's how to turn off public posting on the Meta AI app

2025-06-13
NBC 6 South Florida
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system generating content from user prompts. The default public sharing of these prompts, including sensitive personal data, has led to privacy violations, which is a breach of fundamental rights. This harm is realized and ongoing, not just a potential risk. The article details the nature of the harm and how it arises from the AI system's use and default settings. Hence, it meets the criteria for an AI Incident due to violation of rights caused by the AI system's operation.
Thumbnail Image

Users of new Meta AI app unknowingly make chatbot logs public - SiliconANGLE

2025-06-13
SiliconANGLE
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system as it generates chatbot responses personalized with user data. The incident arises from the use of the AI system, specifically the Share button that users unknowingly activate, causing private chatbot logs containing sensitive data to be publicly accessible. This exposure constitutes a violation of privacy rights, a breach of legal obligations protecting personal data, and thus harm has materialized. The harm is directly linked to the AI system's design and use, fulfilling the criteria for an AI Incident. The event is not merely a potential risk or a complementary update but a realized harm due to the AI system's operation and user interaction.
Thumbnail Image

Meta's AI App Raises Privacy Concerns on Instagram

2025-06-14
WebProNews
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system as it uses advanced chatbot capabilities to interact with users. The incident arises from the use of this AI system, specifically its failure to provide clear privacy controls and notifications, resulting in users' private conversations being publicly exposed without their knowledge. This constitutes a violation of privacy rights and causes harm to users, fitting the definition of an AI Incident. The harm is realized, not just potential, as users' data is already being exposed publicly. Therefore, this event qualifies as an AI Incident due to direct harm caused by the AI system's use and design flaws.
Thumbnail Image

Meta AI Privacy Scandal: Users' Sensitive Queries Exposed Online

2025-06-13
Nokiamob
Why's our monitor labelling this an incident or hazard?
Meta AI is an AI system integrated with social media platforms. The incident involves the use of this AI system where users' sensitive queries were posted publicly, often linked to their real identities, without clear informed consent. This constitutes a violation of privacy and user rights, which falls under harm category (c) - violations of human rights or breach of obligations protecting fundamental rights. The harm has already occurred as users' sensitive information was exposed, making this an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI searches made public - but do all its users realise? - Ghanamma.com

2025-06-14
GHANA MMA
Why's our monitor labelling this an incident or hazard?
The event describes an AI system (Meta AI) whose use has directly led to harm in the form of privacy violations and security risks. Users' AI-generated prompts and results, some containing sensitive or personal information, are publicly posted and traceable to their identities, which constitutes a breach of user privacy and a violation of rights. Although Meta states that sharing is opt-in, the evidence suggests users may not fully understand this, leading to inadvertent exposure. This fits the definition of an AI Incident because the AI system's use has directly led to harm (privacy and security breaches).
Thumbnail Image

Meta's AI App 'Discover' Feed Publicly Exposes Private Chats Without Users Knowing - WinBuzzer

2025-06-13
WinBuzzer
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta's AI chatbot app) whose use has directly led to harm: the public exposure of private, sensitive user conversations without proper user awareness or consent. This exposure violates privacy rights and data protection laws, constituting a breach of fundamental rights. The harm is realized and ongoing, with legal challenges and public backlash confirming the severity. The AI system's design choices and data handling practices are central to the incident, fulfilling the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI Searches Made Public - But Do All Its Users Realise?

2025-06-13
BruneiDirect
Why's our monitor labelling this an incident or hazard?
The AI system (Meta AI) is explicitly involved as it generates responses to user prompts and shares these prompts and responses publicly by default in a social media-like feed. Users' lack of awareness about the public nature of their data and the traceability to their social media profiles has led to the exposure of sensitive personal information, which is a direct harm to users' privacy and rights. This harm is realized, not just potential, as the article provides concrete examples of sensitive data being publicly accessible. Therefore, this event meets the criteria for an AI Incident due to direct harm caused by the AI system's use and design.
Thumbnail Image

Meta AI searches go public, sparking user privacy concerns - Tech Digest

2025-06-13
Tech Digest
Why's our monitor labelling this an incident or hazard?
The Meta AI system is explicitly mentioned as a generative AI tool integrated into Facebook, Instagram, and WhatsApp. The public display of user queries and AI responses, including sensitive personal data, directly leads to privacy harms by exposing identifiable information without users' full understanding or consent. This constitutes a violation of user privacy rights and can be considered harm to individuals and communities. The harm is realized as sensitive data is already publicly accessible, not just a potential risk. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's use and the harm caused by privacy breaches.
Thumbnail Image

Sensitive Private Chats Said to Be Appearing on Meta App's Discovery Feed

2025-06-13
Gadgets 360
Why's our monitor labelling this an incident or hazard?
The Meta AI app uses an AI chatbot system that generates conversations and allows users to share these conversations publicly via a 'Post' button. The design of this sharing process is causing users, especially those less tech-savvy, to unintentionally expose private and sensitive information. This exposure is a direct consequence of the AI system's use and the app's interface, leading to realized harm in the form of privacy violations and potential risks to users. Therefore, this qualifies as an AI Incident due to the direct or indirect harm caused by the AI system's use and malfunction in user interface design leading to privacy breaches.
Thumbnail Image

Meta messages: App users are taking over-sharing to an AI level

2025-06-15
NZ Herald
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use leads to users sharing sensitive personal information publicly, often unintentionally due to unclear interface design. This implicates privacy concerns and potential violations of users' rights, which are recognized as harms under the framework. However, the article does not report actual realized harm such as data breaches, misuse, or legal violations, only the plausible risk thereof. The AI system's role in generating content and facilitating sharing is central. Hence, the event fits the definition of an AI Hazard, as it plausibly could lead to harm (privacy violations, misuse of personal data) but no direct harm is confirmed yet. It is not Complementary Information because the main focus is on the emerging risk, nor is it Unrelated as AI involvement and potential harm are clear.
Thumbnail Image

If You Think Your Meta AI Chats Are Private, Think Again

2025-06-15
ProPakistani
Why's our monitor labelling this an incident or hazard?
The AI system's design and use have directly led to violations of privacy and potential breaches of personal data protection, which constitute harm to individuals' rights under applicable law. The inadvertent public sharing of private conversations containing sensitive information is a clear harm caused by the AI system's use and interface design. Therefore, this qualifies as an AI Incident due to realized harm linked to the AI system's use and its impact on users' privacy and rights.
Thumbnail Image

Meta plans to monetise something you believe is private

2025-06-15
Euro Weekly News Spain
Why's our monitor labelling this an incident or hazard?
An AI system (Meta's AI chatbot and algorithms) is explicitly involved in processing and monetizing user data. The use of this AI system has directly led to violations of privacy and data protection rights, which are breaches of applicable laws and fundamental rights. The harm is realized as users' private data is collected and monetized without informed consent, impacting their rights and privacy. Therefore, this qualifies as an AI Incident due to violations of human rights and legal obligations related to data privacy and user consent.
Thumbnail Image

Mark Zuckerberg must have noticed people were sharing cringy things on Meta AI. It's now a little harder to accidentally spill.

2025-06-16
Business Insider
Why's our monitor labelling this an incident or hazard?
The article discusses a mitigation measure (a warning pop-up) introduced by Meta in response to previously reported accidental oversharing of personal information on the Meta AI Discover feed. The original problem involved users unintentionally sharing sensitive data, which constitutes harm to privacy and potentially human rights. However, this article focuses on the new warning feature as a response to that issue rather than describing a new incident or hazard. Therefore, it is Complementary Information, providing an update on societal and governance responses to an earlier AI Incident involving privacy harms.
Thumbnail Image

Meta warns users to 'avoid sharing personal or sensitive information' in its AI app

2025-06-16
engadget
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that generates and shares user interactions publicly in a social feed. Users were unknowingly exposing sensitive personal information, which constitutes a violation of privacy and potentially human rights related to data protection. The harm has already occurred as users' private information was made public without clear consent or understanding. Meta's addition of warnings and feed changes is a response to this harm. Therefore, this qualifies as an AI Incident due to realized harm from the AI system's use and design leading to privacy violations.
Thumbnail Image

You Need to Change This Setting Immediately If You Use Meta AI

2025-06-16
MakeUseOf
Why's our monitor labelling this an incident or hazard?
The article explicitly describes how the use of Meta AI's chatbot system has led to users unintentionally sharing sensitive personal information publicly. The AI system is central to the incident because it processes user prompts that are then exposed due to the app's design and settings. The harms are realized and significant, including privacy breaches and potential security risks, which meet the criteria for violations of human rights and harm to communities. The involvement of the AI system in the development and use phases, combined with the direct link to harm, justifies classification as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI app 'a privacy disaster' as chats unknowingly made public - 9to5Mac

2025-06-16
9to5Mac
Why's our monitor labelling this an incident or hazard?
The Meta AI chatbot is an AI system that processes user inputs (text, audio, images) and allows sharing of conversations. The incident arises from the AI system's default privacy settings and user interface design that do not clearly inform users about the public nature of their shared chats, leading to inadvertent public exposure of private data. This exposure constitutes a violation of privacy rights, a recognized human right, and thus fits the definition of an AI Incident under violations of human rights or breach of obligations intended to protect fundamental rights. The harm is realized, not just potential, as users' private information has been made public without their informed consent.
Thumbnail Image

Why the Meta AI App Is Raising Serious Privacy Concerns

2025-06-16
Analytics Insight
Why's our monitor labelling this an incident or hazard?
The AI system is explicitly involved as it powers the chatbot and the Discover feed feature. The harm arises from the use of the AI system leading to the public exposure of personal and sensitive information, which constitutes a violation of privacy rights and potential harm to individuals and communities. Since the harm is occurring due to the AI system's use, this qualifies as an AI Incident under the framework's criteria for violations of human rights and harm to communities.
Thumbnail Image

"How Do I Improve My Bowel Movements?" -- Meta AI Is Exposing Your Most Malu Questions

2025-06-16
SAYS
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta's AI chatbot) that is used by over a billion users and is designed to interact conversationally, indicating AI involvement. The AI system's use has directly led to the exposure of deeply personal and sensitive user data (private queries about health, finances, legal matters) publicly tied to their real identities. This constitutes a violation of privacy rights and potentially breaches obligations under applicable laws protecting fundamental and labor rights. The harm is realized as users' private information is made public without proper consent, causing reputational and emotional harm. Therefore, this qualifies as an AI Incident due to violations of human rights and privacy stemming from the AI system's use and design flaws.
Thumbnail Image

Inside the privacy risks of Meta's new AI app

2025-06-16
CityAM
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta's AI chatbot app) whose use has directly led to harm: the exposure of sensitive personal data in a publicly accessible feed. The harm includes privacy violations and potential misuse of personal information, which are breaches of data protection laws and fundamental rights. The AI system's design encourages oversharing and lacks adequate safeguards, making it a contributing factor to the incident. Although Meta claims privacy by default, the actual outcome is realized harm to users' privacy, meeting the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Sensitive Meta AI Chats Exposed to Public

2025-06-16
eWEEK
Why's our monitor labelling this an incident or hazard?
The AI system (Meta's chatbot) is explicitly involved as it processes user prompts and displays them in a public feed. The harm arises from the AI system's use and design leading to widespread unintended disclosure of sensitive personal information, which constitutes a violation of privacy rights and potentially breaches legal obligations protecting personal data. Although users must opt in, the confusing interface has caused many to unknowingly share private data publicly, resulting in direct harm to individuals' privacy and security. Therefore, this qualifies as an AI Incident due to realized harm linked to the AI system's use and malfunction in privacy protection.
Thumbnail Image

Meta Addresses Privacy Blunder With Meta AI Chatbot

2025-06-17
WebProNews
Why's our monitor labelling this an incident or hazard?
The AI chatbot is an AI system as it generates conversational outputs based on user inputs. The incident stems from the use of this AI system, specifically the design and implementation of the public 'Discover' feed feature that caused private user data to be shared publicly without clear consent. This led directly to harm in the form of privacy violations and exposure of sensitive personal information, which falls under violations of human rights and legal protections related to privacy. Therefore, this qualifies as an AI Incident rather than a hazard or complementary information, as the harm has already occurred and is directly linked to the AI system's use and design.
Thumbnail Image

We uncovered how Meta's AI app was full of accidental public posts that were really personal. It's now trying to fix that.

2025-06-16
Business Insider Nederland
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that allows users to share AI-generated content and chats. The accidental public sharing of personal and sensitive information through the app's Discover feed constitutes a violation of users' privacy rights, which is a breach of obligations under applicable law protecting fundamental rights. The harm has already occurred as personal data was exposed publicly. Meta's addition of a warning pop-up is a response to mitigate further harm but does not negate the incident. Hence, this event is an AI Incident because the AI system's use directly led to harm to individuals' rights through accidental data exposure.
Thumbnail Image

Inside the privacy risks of Meta's new AI app

2025-06-16
newsR
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta's AI chatbot app) whose use has directly led to harm in the form of privacy violations and exposure of sensitive user data. This fits the definition of an AI Incident because the AI system's deployment has caused a breach of obligations under applicable law intended to protect fundamental rights, specifically privacy rights. The harm is realized and ongoing, not merely potential, and the AI system's role is pivotal in causing this harm.
Thumbnail Image

Users Accidentally Leak Personal Data via Meta AI

2025-06-16
Digit
Why's our monitor labelling this an incident or hazard?
The event describes an AI system (Meta AI chatbot) whose use has directly led to harm: users' sensitive personal data and conversations are publicly exposed without their full understanding or consent. This is a violation of privacy rights and can be considered a breach of obligations to protect fundamental rights. The harm is realized, not just potential, and stems from the AI system's deployment and user interface design that misleads users into sharing private data. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta AI App's Privacy Concerns Raise Significant Alarms for Users - czechjournal.cz

2025-06-13
The Czech Journal
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions an AI system (Meta AI app) that collects extensive personal data through AI-driven personalization. Although no direct harm or incident is described, the aggressive data collection and privacy invasion concerns plausibly could lead to violations of privacy rights and regulatory penalties, which qualify as harms under the AI framework. Since the harm is potential and not yet realized, this fits the definition of an AI Hazard rather than an AI Incident. The article also discusses possible responses and user advocacy but does not focus primarily on these, so it is not Complementary Information. Therefore, the classification is AI Hazard.
Thumbnail Image

Meta publica chats privados entre usuários e inteligência artificial, afirma site

2025-06-15
Folha - PE
Why's our monitor labelling this an incident or hazard?
The Meta AI assistant is an AI system interacting with users. The incident arises from the use of this AI system, where users unknowingly share private conversations publicly due to unclear privacy settings and interface design. This has directly led to harm in the form of privacy violations and exposure of sensitive personal data, which falls under violations of human rights and fundamental rights protections. Therefore, this qualifies as an AI Incident.
Thumbnail Image

Meta publica chats privados entre usuários e inteligência artificial, afirma site

2025-06-15
O Globo
Why's our monitor labelling this an incident or hazard?
The involvement of an AI system (Meta's chatbot) is explicit. The event describes the use of the AI system leading to a violation of user privacy rights, which constitutes a breach of obligations intended to protect fundamental rights. The exposure of private conversations without consent is a direct harm to users' rights and privacy. Therefore, this qualifies as an AI Incident under the framework.
Thumbnail Image

Meta AI expõe dados íntimos de usuários em feed público - 13/06/2025 - Tec - Folha

2025-06-14
Folha de S.Paulo
Why's our monitor labelling this an incident or hazard?
The Meta AI chatbot is an AI system that generates conversational responses to users' inputs. The app's feature of a public feed where users' conversations are shared has caused direct harm by exposing intimate personal data without clear user consent or understanding, constituting a violation of privacy rights and harm to individuals and communities. The harm is realized, not just potential, as users' private conversations are publicly visible. Therefore, this qualifies as an AI Incident due to direct harm caused by the AI system's use and design leading to privacy violations and harm to users.
Thumbnail Image

App do Meta AI compartilha mensagens privadas sem usuário saber

2025-06-13
TecMundo
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI app) that processes user inputs and shares conversations in a public feed. Users are unknowingly sharing private and sometimes sensitive or sexual content, which constitutes a violation of privacy rights and can cause harm to individuals' reputations and dignity. This harm is directly linked to the AI system's use and design, fulfilling the criteria for an AI Incident under violations of human rights and harm to communities. The lack of clear privacy settings and user awareness exacerbates the issue, confirming direct harm caused by the AI system's operation.
Thumbnail Image

Cuidado ao usar o Meta AI: botão de compartilhar pode expor conversas privadas dos usuários | Exame

2025-06-13
Exame
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system (chatbot) whose use has directly led to harm by exposing private user conversations and sensitive data publicly without proper user consent or warning. This exposure constitutes a violation of privacy rights and potentially other legal protections. The harm is realized, not just potential, as users' private data has been made accessible to the public. Hence, this event meets the criteria for an AI Incident due to the AI system's use causing violations of rights and harm to users' privacy.
Thumbnail Image

Meta AI: nova app tem falha grave e expõe conversas privadas de utilizadores | TugaTech

2025-06-13
TugaTech
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system based on Llama models, used as a personal assistant. The incident involves the use of this AI system and a critical design flaw in its sharing interface that caused private conversations to be exposed publicly. This exposure has led to realized harm including privacy violations and risks of harassment and identity theft, which are harms to persons and their rights. Therefore, this event meets the criteria for an AI Incident as the AI system's use directly led to significant harm.
Thumbnail Image

Utilizadores estão a partilhar acidentalmente "conversas" privadas com o chatbot Meta AI no feed público - SAPO Tek

2025-06-17
SAPO Tek
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use has directly led to harm: users' private and sensitive conversations are being publicly exposed accidentally, causing privacy violations and potential harm to individuals. This fits the definition of an AI Incident because the AI system's design and user interface have contributed to the harm. The presence of warnings and mitigation efforts does not negate the fact that harm has occurred. Therefore, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Как Meta AI да не показва какво сме го питали

2025-06-16
Vesti.bg
Why's our monitor labelling this an incident or hazard?
The article centers on Meta AI's functionality related to sharing user prompts publicly and the privacy implications thereof. While the AI system is involved, the article does not report any realized harm such as violations of rights or harm to individuals or communities. It mainly informs users about settings to control data sharing, which is a governance or user control issue rather than an incident or hazard. Therefore, this is best classified as Complementary Information, as it provides context and guidance related to AI system use and privacy management without reporting an AI Incident or AI Hazard.
Thumbnail Image

Le domande fatte all'IA di Meta possono diventare pubbliche

2025-06-16
La Repubblica.it
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta's AI assistant) whose use has directly led to harm: users' private queries, including sensitive and personal information, are publicly exposed without clear consent, violating privacy rights and potentially causing harm to individuals and communities. The AI system's algorithmic curation of the Discover feed without sufficient human moderation has resulted in the circulation of offensive and discriminatory content, further contributing to harm. These factors meet the criteria for an AI Incident as the harms are realized and directly linked to the AI system's use and malfunction in privacy protection and content moderation.
Thumbnail Image

Meta AI sta pubblicando le conversazioni con gli utenti: e se la tua fosse tra queste?

2025-06-15
Fanpage
Why's our monitor labelling this an incident or hazard?
The AI system (Meta AI chatbot) is explicitly involved as it generates and manages user conversations. The publication of these conversations without clear user consent or effective privacy controls has directly led to harm in the form of privacy violations and exposure of sensitive personal data. This fits the definition of an AI Incident under violations of human rights or breach of privacy obligations. The harm is realized, not just potential, as sensitive data is already publicly accessible. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Meta AI e il lato troppo personale dell'intelligenza artificiale

2025-06-16
TuttoAndroid
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI conversational assistant) whose use has directly led to harm: users' private and sensitive information is being publicly exposed unintentionally, causing privacy violations and potential harm to individuals. The AI system's integration and interface design facilitate this harm by making it easy to publish conversations publicly without clear warnings or confirmations. This fits the definition of an AI Incident because the AI system's use has directly led to violations of fundamental rights (privacy) and harm to individuals. The article does not describe a potential or future harm but an ongoing realized harm, so it is not an AI Hazard. It is not merely complementary information or unrelated news, as the harm is clearly articulated and linked to the AI system's use.
Thumbnail Image

Meta AI sotto accusa: milioni di download, ma è allarme privacy per gli utenti

2025-06-13
HTML.it
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system enabling user interaction via text, images, and audio. Its design and use have directly led to harm by exposing sensitive personal data publicly, which is a violation of privacy rights and can be considered harm to individuals. The article documents actual cases of such exposure, not just potential risks, thus qualifying as an AI Incident under the framework. The lack of clear privacy settings and the public sharing mechanism are part of the system's use leading to this harm.
Thumbnail Image

Meta AI pubblica le chat private, un disastro per la privacy

2025-06-13
Punto Informatico
Why's our monitor labelling this an incident or hazard?
The AI system (Meta AI chatbot) is explicitly involved as the source of the private conversations being shared. The harm is realized and ongoing, as users' private and sensitive information is being publicly exposed, constituting a violation of privacy rights and potentially other legal rights. The incident stems from the use and design of the AI system, which directly led to harm to individuals' privacy and personal security. Therefore, this qualifies as an AI Incident under the definitions provided, specifically under violations of human rights or breach of obligations intended to protect fundamental rights (privacy).
Thumbnail Image

Meta AI e il feed Discover: un problema di privacy per gli utenti

2025-06-16
Webnews
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI) whose use has directly led to the exposure of sensitive personal data, a violation of privacy rights, which falls under harm to human rights and breach of legal obligations. The article reports realized harm through the public exposure of private data and the use of such data for AI training without clear consent, meeting the criteria for an AI Incident. The involvement of the AI system is explicit, and the harm is direct and ongoing, not merely potential or speculative.
Thumbnail Image

Meta AI ha un grosso problema di privacy

2025-06-16
MRW.it
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta AI chatbot) whose use has directly led to harm in the form of privacy violations and exposure of sensitive personal information. This harm falls under violations of human rights and legal protections related to privacy. The AI system's design and use, particularly the unclear sharing functionality, are the root causes of this harm. Therefore, this event qualifies as an AI Incident because the AI system's use has directly led to significant harm to individuals' privacy and security.
Thumbnail Image

Meta aconselha a evitar partilha de informações pessoais com ferramenta de IA

2025-06-17
ECO
Why's our monitor labelling this an incident or hazard?
While the AI system is involved in the event, the harm described is related to users unintentionally exposing their own personal information publicly due to a feature of the AI platform. There is no direct or indirect harm caused by the AI system itself malfunctioning or misusing data; rather, the risk stems from user behavior and platform design choices. The event does not describe realized harm such as privacy violations caused by the AI system's misuse or malfunction, nor does it describe a plausible future harm caused by the AI system beyond user error. Therefore, this is best classified as Complementary Information, as it provides context and a governance response (Meta's warning) to a potential privacy concern related to AI use, without constituting a new AI Incident or AI Hazard.
Thumbnail Image

App da Meta AI avisa que posts são públicos após 'exposição' de chats

2025-06-17
Canaltech
Why's our monitor labelling this an incident or hazard?
The Meta AI chatbot is an AI system that generates conversational outputs. The exposure of private chats publicly constitutes a violation of privacy, which is a human rights concern. The harm has already occurred as personal and sensitive information was exposed without clear user understanding or consent. Therefore, this qualifies as an AI Incident due to the direct harm to users' privacy rights caused by the AI system's use and design (public sharing of prompts).
Thumbnail Image

Meta faz apelo: não compartilhe seus segredos com nosso app de IA * Tecnoblog

2025-06-17
Tecnoblog
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is explicitly involved, and its malfunction or design flaw caused private user data to be exposed publicly, leading to a violation of privacy rights (a breach of obligations under applicable law). This harm has already occurred, making it an AI Incident. The article also describes the company's response, but the primary focus is on the realized harm from the AI system's use and malfunction.
Thumbnail Image

Meta AI: agora aplicativo avisa quando conversas forem públicas

2025-06-17
Olhar Digital - O futuro passa primeiro aqui
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use led to users unintentionally sharing sensitive information publicly, which can be considered a violation of privacy rights and potentially a breach of obligations protecting fundamental rights. The legal ruling indicates harm has occurred to a large group of people. Therefore, this qualifies as an AI Incident due to realized harm linked to the AI system's use and user misunderstanding of its public sharing feature. The pop-up is a mitigation measure but does not negate the incident classification.
Thumbnail Image

Após crítica, Meta AI ganha aviso para impedir compartilhamento acidental de conversas privadas | Exame

2025-06-17
Exame
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI app) is involved, and its design flaw led to the indirect harm of privacy violations through accidental public sharing of private conversations. This constitutes a violation of rights (privacy) and harm to individuals. The harm has already occurred as users' private data was exposed. The company's response is a mitigation measure but does not negate the fact that an AI Incident took place due to the system's design and use. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Meta aconselha a evitar partilha de informações pessoais com IA

2025-06-17
Notícias ao Minuto
Why's our monitor labelling this an incident or hazard?
An AI system (Meta's generative AI assistant) is involved, and the event concerns the use of this system. The issue arises from the use of a feature that allows public sharing of AI conversations, which can lead to privacy harms if personal data is shared. However, the article does not report any realized harm or incidents of privacy breaches causing injury or rights violations; rather, it highlights a potential risk and Meta's response to mitigate it. Therefore, this qualifies as an AI Hazard because the development and use of the AI system's sharing feature could plausibly lead to harm (privacy violations), but no direct harm is reported yet.
Thumbnail Image

Meta AI: feed público de partilhas embaraçosas leva a empresa a emitir aviso | TugaTech

2025-06-17
TugaTech
Why's our monitor labelling this an incident or hazard?
The Meta AI system is explicitly involved as the platform where users share interactions. The harm arises from users unintentionally exposing sensitive personal information publicly due to unclear design and lack of adequate warnings, leading to privacy violations. This constitutes a breach of obligations intended to protect fundamental rights (privacy). The company's addition of a warning and changes to the feed are responses to this harm. Since the harm (privacy violations) has already occurred and is directly linked to the AI system's use and design, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta aconselha a evitar partilha de informações pessoais com ferramenta de IA

2025-06-17
Açoriano Oriental
Why's our monitor labelling this an incident or hazard?
While the AI system is involved and there is a risk of harm related to privacy and potential exposure of sensitive personal information, the article does not report that any actual harm (such as privacy breaches or violations of rights) has occurred yet. The warning and feature design indicate a potential risk of harm if users share sensitive data publicly, but no direct or indirect harm has been confirmed. Therefore, this situation represents a plausible risk of harm due to the AI system's use, qualifying it as an AI Hazard rather than an AI Incident. It is not Complementary Information because the main focus is on the risk and warning, not on responses to a past incident, nor is it unrelated as it clearly involves an AI system and potential harm.
Thumbnail Image

With Meta AI App, You Can 'Discover' People's Wildest Thoughts -- But Are You Unknowingly Sharing Yours?

2025-06-18
Yahoo News
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) whose use leads to the public exposure of users' private and sensitive information without clear user understanding, resulting in violations of privacy rights and potential harm to individuals. This fits the definition of an AI Incident because the AI system's use has directly or indirectly led to harm in the form of privacy violations and exposure of personal data. The article details realized harm rather than just potential risk, and the AI system's role is pivotal in enabling this exposure through the 'Discover' feed feature.
Thumbnail Image

Meta Invents New Way to Humiliate Users With Feed of People's Chats With AI

2025-06-17
404 Media
Why's our monitor labelling this an incident or hazard?
The AI system (Meta's chatbot) is explicitly involved as it generates and hosts user conversations. The public sharing of these conversations without clear informed consent or privacy safeguards has directly led to harm by exposing sensitive personal information, which is a violation of privacy rights and can cause emotional and reputational harm. The harm is realized and ongoing, as users' private data is publicly accessible and linked to their real identities. This fits the definition of an AI Incident due to violation of human rights and harm to individuals caused by the AI system's use and deployment.
Thumbnail Image

With Meta AI App, You Can 'Discover' People's Wildest Thoughts -- But Are You Unknowingly Sharing Yours?

2025-06-18
HuffPost
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Meta AI chatbot) whose use includes a feature that publicly shares user interactions if users choose to do so. The potential harm relates to privacy and exposure of sensitive personal information, which can be considered a violation of rights if it occurs without consent. However, the article clarifies that sharing is an affirmative user action with warnings provided, and no direct harm or legal violation by the AI system is reported. The event highlights privacy concerns and user experience issues, which are important for understanding AI impacts and governance but do not meet the threshold for an AI Incident or AI Hazard. Thus, it fits best as Complementary Information, informing about the ecosystem and user risks without a direct or plausible harm incident.
Thumbnail Image

I don't like the idea of my conversations with Meta AI being public - here's how you can opt out

2025-06-18
TechRadar
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI chatbot) and its feature (Discovery feed) that allows public sharing of conversations. However, the article does not report any realized harm such as privacy violations leading to injury, rights breaches, or other significant harms. It focuses on informing users about opting out and Meta's data usage policies. This fits the definition of Complementary Information, as it provides context and updates on privacy and transparency related to AI use, without describing an AI Incident or AI Hazard.
Thumbnail Image

Meta says your chats with Meta AI are not private, could be seen by everyone

2025-06-17
India Today
Why's our monitor labelling this an incident or hazard?
The AI system is explicitly involved as the Meta AI chatbot app. The harm arises from the use of the AI system, where users' private conversations and photos are being shared publicly due to unclear sharing mechanisms. This has directly led to harm in the form of privacy violations and exposure of sensitive personal information, which falls under violations of human rights and harm to individuals. The event is not merely a potential risk but describes actual realized harm through public exposure of private data. Therefore, it meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta's Privacy Goof Shows How People Really Use AI Chatbots

2025-06-18
NYMag
Why's our monitor labelling this an incident or hazard?
The article details how Meta's AI chatbot app caused users to inadvertently share extremely private information publicly, which constitutes a violation of privacy rights and harm to individuals. The AI system's design and use directly led to this harm. The presence of the AI system is explicit, and the harm is realized, not just potential. Hence, this is an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta Adds Privacy Warning After Users Unknowingly Share Sensitive Chats Publicly On AI App, But Experts Say It Falls Short Of Fixing Core Issues

2025-06-17
Wccftech
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta's AI app) whose use has indirectly led to harm in the form of privacy violations and potential exposure of sensitive personal information. Users were unaware that their private chats could be publicly visible, which constitutes a violation of privacy rights and a breach of data protection expectations. The harm is realized as private information was publicly accessible, and the AI system's design (default public feed) contributed to this outcome. Meta's addition of a warning is a response to this incident but does not fully resolve the core issue. Therefore, this qualifies as an AI Incident due to the realized privacy harm linked to the AI system's use and design.
Thumbnail Image

Your Meta AI chats are not really a secret. Here's how to keep them private

2025-06-19
Digital Trends
Why's our monitor labelling this an incident or hazard?
The Meta AI app involves an AI system (the chatbot) whose use has directly led to violations of user privacy and potential breaches of personal data rights. The public visibility of sensitive user interactions is a direct consequence of the AI system's deployment and default settings, causing harm to users' rights and privacy. This fits the definition of an AI Incident because the AI system's use has directly led to harm in the form of privacy violations and exposure of sensitive personal information.
Thumbnail Image

Meta AI app introduces consent pop ups as users post sensitive details in prompts to generate AI content

2025-06-19
The Hindu
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system that generates content based on user prompts and allows sharing on a public Discover feed. The appearance of sensitive personal data in public posts raised privacy concerns, which could lead to violations of user privacy rights if harm occurred. However, the article does not report any actual harm or privacy violations having taken place; rather, it describes Meta's implementation of consent pop-ups to mitigate these risks. Therefore, this event is best classified as Complementary Information, as it provides an update on governance and mitigation responses to potential AI-related privacy risks without describing a new AI Incident or AI Hazard.
Thumbnail Image

The Privacy Issue That Every User Of Meta AI App Should Know

2025-06-20
Oneindia
Why's our monitor labelling this an incident or hazard?
The Meta AI app is an AI system functioning as a chatbot and social platform. Its design encourages users to share AI-generated conversations publicly, often containing sensitive personal data. This has directly led to privacy harms, including violations of users' rights and risks of identity theft and harassment. The involvement of AI in generating and managing these conversations, combined with poor privacy safeguards and dark patterns, makes this a clear AI Incident under the framework, as the AI system's use has directly led to harm to individuals' rights and privacy.
Thumbnail Image

Why sharing too much with chatbots could backfire on you

2025-06-20
NZ Herald
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (chatbots) and describes harms that have occurred or are occurring, such as privacy violations through public sharing of conversations, potential misuse of personal data, and misinformation generated by chatbots. These harms relate to violations of privacy rights and potential harm to individuals' personal information. Since these harms are realized or ongoing, and directly linked to the use of AI chatbots, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Meta's AI App May Make Your Private Searches Public. Very Public - Stuff South Africa

2025-06-20
Stuff
Why's our monitor labelling this an incident or hazard?
The Meta AI app involves an AI system that processes user queries. Due to a design or operational issue, private inputs were publicly shared without users' informed consent, causing violations of privacy and potentially human rights related to data protection. This constitutes harm to individuals' rights and privacy, directly linked to the AI system's use and malfunction in managing data visibility. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Users of Facebook app must make important change now to avoid chats going public

2025-06-21
The Irish Sun
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is involved, and its use has directly led to a privacy harm where private conversations are being publicly exposed without explicit user intent, constituting a violation of user privacy and potentially human rights related to privacy. This is a realized harm caused by the AI system's design and default settings, thus qualifying as an AI Incident.
Thumbnail Image

Samtaler med Metas kunstige intelligens kan ses af alle

2025-06-16
Berlingske Tidende
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI) whose use and data-sharing policies could plausibly lead to violations of user privacy and rights, which are harms under the AI Incident definition. However, since no actual harm or incident is reported—only the potential for harm due to public visibility of user prompts and responses—the event fits the definition of an AI Hazard rather than an AI Incident. The article focuses on the system's design and policy implications rather than a specific harmful event that has occurred.
Thumbnail Image

Dataetisk Råd med nye anbefalinger om AI i sundhed | Dataetisk Råd

2025-06-19
via.ritzau.dk
Why's our monitor labelling this an incident or hazard?
The article focuses on ethical recommendations and reflections about AI use in healthcare, aiming to guide responsible development and deployment. There is no mention of any actual harm, incident, or hazard caused or plausibly caused by AI systems. Therefore, it does not meet the criteria for AI Incident or AI Hazard. It is not a complementary update on a specific incident either, but rather a general advisory report, so it is best classified as Complementary Information.
Thumbnail Image

Sur l'IA de Meta, ces requêtes très intimes d'internautes partagées à des millions d'utilisateurs - Edition du soir Ouest-France - 17/06/2025

2025-06-17
Ouest France
Why's our monitor labelling this an incident or hazard?
The AI system (Meta's chatbot) is explicitly involved as it processes user conversations. The harm arises from the use of the AI system's new feature that inadvertently exposed private and intimate user data to over a billion users, violating privacy rights and potentially causing emotional and reputational harm. The event is not merely a potential risk but a realized harm, as private data has already been publicly disclosed. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's use and the violation of users' rights and privacy.
Thumbnail Image

Meta AI : attention, vos conversations privées avec l'IA ne le sont pas toujours

2025-06-18
Capital.fr
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is explicitly involved, and its use has directly led to harm in the form of privacy violations and exposure of sensitive personal data. This constitutes a violation of fundamental rights related to privacy and data protection, which falls under the category of violations of human rights or breach of obligations under applicable law. The harm is realized and ongoing, not merely potential. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Meta AI : vos conversations privées avec l'IA ne le sont pas toujours, prudence !

2025-06-19
Tribunal Du Net
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI) is involved, and its use has directly led to harm in the form of violations of privacy, which is a breach of fundamental rights. The harm is realized as users' private conversations are exposed publicly, often without their full awareness or consent. This fits the definition of an AI Incident because the AI system's use has directly led to a violation of human rights (privacy).
Thumbnail Image

Scandale Meta AI : tout ce qu'il se passe sur le chat... ne reste pas sur le chat

2025-06-19
Génération-NT
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI chatbot) is explicitly involved, and its use has directly led to harm in the form of privacy violations and exposure of sensitive personal data. This constitutes a breach of obligations under applicable law protecting fundamental rights, including privacy and data protection. The harm is realized and significant, affecting millions of users. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Meta AI vous laisse lire les discussions (très bizarres) des autres utilisateurs

2025-06-16
LEBIGDATA.FR
Why's our monitor labelling this an incident or hazard?
The article explicitly mentions an AI system (Meta AI chatbot) and its use leading to the sharing of private conversations, including sensitive personal data and confessions, without adequate user consent or awareness. This exposure causes harm to users' privacy and potentially violates their rights. The harm is realized, not just potential, as private data and conversations are publicly accessible and subject to mockery and misuse. Therefore, this qualifies as an AI Incident due to violation of rights and harm to communities through privacy breaches caused by the AI system's use and its sharing feature.
Thumbnail Image

Meta upozorava: Ne delite ove informacije sa AI

2025-06-18
B92
Why's our monitor labelling this an incident or hazard?
An AI system (Meta AI application) is explicitly involved, as it processes user inputs and shares them publicly on a feed. The harm relates to violations of privacy and potentially human rights due to users unintentionally exposing sensitive personal information. Although no direct harm such as injury or property damage is reported, the event involves realized harm in terms of privacy violations and user data exposure. The addition of a warning is a response to this issue but does not negate the fact that harm has occurred. Therefore, this qualifies as an AI Incident due to the realized privacy harm caused by the AI system's use and design.
Thumbnail Image

Meta upozorava: Ne dijelite ove informacije sa AI

2025-06-19
Nezavisne novine
Why's our monitor labelling this an incident or hazard?
The Meta AI application involves an AI system with a public feed where users share interactions. The issue arises from users unintentionally exposing personal data, which is a privacy concern. However, the article does not report any actual harm such as violations of rights or health, nor does it describe a plausible imminent harm event. Instead, it reports Meta's addition of a warning message and some changes to the feed display, which are responses to user privacy concerns. This fits the definition of Complementary Information, as it provides updates on societal and governance responses to AI-related privacy issues without describing a new AI Incident or AI Hazard.
Thumbnail Image

Source.ba:Meta upozorava na informacije koje ne smijete dijeliti sa AI

2025-06-19
Source.ba
Why's our monitor labelling this an incident or hazard?
The article involves an AI system (Meta AI application) and discusses issues related to user data privacy and the public sharing of AI interactions. However, it does not describe any direct or indirect harm resulting from the AI system's development, use, or malfunction. The warning added by Meta is a mitigation measure to prevent potential privacy harms. Since no actual harm has been reported and the focus is on a governance or user experience response to a known risk, this qualifies as Complementary Information rather than an AI Incident or AI Hazard.
Thumbnail Image

Meta izdala upozorenje: Ne dijelite lične i osjetljive informacije sa AI aplikacijom

2025-06-18
Srpskainfo
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (Meta AI application) and addresses issues related to user privacy and data sharing. However, the article does not describe any direct or indirect harm caused by the AI system's development, use, or malfunction. Instead, it reports a company response (adding a warning) to mitigate potential privacy risks and improve user safety. Therefore, this is Complementary Information as it provides an update on societal and governance responses to AI-related privacy concerns without describing a new AI Incident or AI Hazard.
Thumbnail Image

Meta upozorava na informacije koje ne smijete dijeliti sa AI - 24sata.ba

2025-06-20
24sata.ba
Why's our monitor labelling this an incident or hazard?
The article involves an AI system (Meta AI chatbot with a public feed) whose use has led to privacy risks due to users sharing sensitive information publicly. While this raises concerns about user privacy and security, the article does not report any realized harm such as violations of human rights or other harms as defined in the framework. The main focus is on Meta's addition of warnings and changes to the app to mitigate these risks, which constitutes a societal and governance response to an AI-related issue. Therefore, this is best classified as Complementary Information rather than an AI Incident or AI Hazard.
Thumbnail Image

Users' intimate chats posted to app's PUBLIC feed

2025-06-24
Daily Mail Online
Why's our monitor labelling this an incident or hazard?
The event involves an AI chatbot system (Meta.ai) that generates and processes user conversations. The sharing of these conversations publicly, often without clear user awareness, directly harms users' privacy and potentially exposes sensitive personal information. This harm is linked to the AI system's use and the design of its sharing feature, which misleads users into making private chats public. The harm is realized, not just potential, as users' intimate data is already publicly accessible. Hence, it meets the criteria for an AI Incident involving violations of human rights (privacy) and harm to communities (privacy breach).
Thumbnail Image

Private Conversations Accidentally Shared on App's Public Feed - Internewscast Journal

2025-06-24
internewscast.com
Why's our monitor labelling this an incident or hazard?
The event clearly involves an AI system (Meta.ai chatbot) whose use led to the direct harm of privacy violations by exposing private conversations publicly. The sharing feature, although requiring user opt-in, is described as confusing and easy to activate unintentionally, leading to inadvertent public disclosure of sensitive information. This meets the criteria for an AI Incident because the AI system's use directly caused harm to users' privacy rights, which is a violation of fundamental rights under applicable law. The harm is realized, not just potential, and the AI system's role is pivotal in the incident.
Thumbnail Image

Meta's New AI Partner Scale AI Exposed Client Data From Google, xAI and Others in Public Google Docs - WinBuzzer

2025-06-25
WinBuzzer
Why's our monitor labelling this an incident or hazard?
The event involves an AI system-related company (Scale AI) whose data management practices led to the exposure of confidential data from major AI clients. This exposure is a direct harm related to violations of confidentiality and potentially intellectual property rights, fitting the definition of an AI Incident under violations of human rights or breach of obligations under applicable law. The involvement of AI is clear as Scale AI provides data-labeling services critical to AI development, and the security failure directly led to harm. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Google, xAI Data Exposed in Massive Security Failure by New Meta Partner Scale AI - WinBuzzer

2025-06-26
WinBuzzer
Why's our monitor labelling this an incident or hazard?
Scale AI is an AI-related company providing data-labeling services critical for AI development, so its operations involve AI systems. The security failure exposed confidential data from AI companies, constituting a violation of intellectual property rights and confidentiality obligations. This exposure is a direct harm caused by the malfunction (security failure) of an AI-related service provider. The event describes realized harm (data breach and client exodus) linked to AI system development and use, qualifying it as an AI Incident under the framework.