AI Agent Breaches McKinsey's Internal AI Platform, Exposing Sensitive Data

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

A cybersecurity firm's AI agent exploited a vulnerability in McKinsey's internal AI platform, Lilli, gaining unauthorized access to 46.5 million employee chat messages, 728,000 sensitive file names, and internal organizational data within two hours. McKinsey patched the flaw; no client data was compromised.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves an AI system (an autonomous AI agent) that was used to attack another AI system (McKinsey's Lilli platform). The AI agent's autonomous actions directly led to unauthorized access to sensitive internal data, including chat messages, files, and user accounts, which is a clear harm to property and potentially a breach of confidentiality and intellectual property rights. The incident is not hypothetical or potential; the breach occurred and was demonstrated. Therefore, it qualifies as an AI Incident due to realized harm caused by the AI system's use and malfunction (vulnerability exploitation).[AI generated]
AI principles
Privacy & data governanceRobustness & digital security

Industries
Digital securityIT infrastructure and hosting

Affected stakeholders
WorkersBusiness

Harm types
Human or fundamental rightsReputational

Severity
AI incident

Business function:
ICT management and information security

AI system task:
Other

In other databases

Articles about this incident or hazard

Thumbnail Image

AI agents now hacking other AI? McKinsey system breached in 2 hours

2026-03-13
India Today
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (an autonomous AI agent) that was used to attack another AI system (McKinsey's Lilli platform). The AI agent's autonomous actions directly led to unauthorized access to sensitive internal data, including chat messages, files, and user accounts, which is a clear harm to property and potentially a breach of confidentiality and intellectual property rights. The incident is not hypothetical or potential; the breach occurred and was demonstrated. Therefore, it qualifies as an AI Incident due to realized harm caused by the AI system's use and malfunction (vulnerability exploitation).
Thumbnail Image

McKinsey realises the risk of rapid adoption of AI after hackers gain access to 46.5 million employee chat messages, 728000 'sensitive files'

2026-03-13
ETCIO.com
Why's our monitor labelling this an incident or hazard?
The AI system (Lilli) was directly involved as the platform that was compromised by an AI-powered attack, leading to unauthorized access to millions of employee chat messages and sensitive file names. The breach represents a direct harm related to confidentiality and intellectual property rights violations. Although client data was not compromised, the exposure of internal communications and sensitive file metadata is a significant harm. The AI agent used to perform the attack also indicates malicious use of AI. Hence, the event meets the criteria for an AI Incident because the AI system's malfunction and misuse directly led to harm.
Thumbnail Image

McKinsey realises the risk of rapid adoption of AI after hackers gain access to 46.5 million employee chat messages, 728000 'sensitive files' and ... - The Times of India

2026-03-13
The Times of India
Why's our monitor labelling this an incident or hazard?
An AI system (Lilli) was involved as the target of an AI-driven attack by another AI agent. The breach led to unauthorized access to sensitive employee communications and internal AI configurations, which constitutes harm to privacy and potentially intellectual property. Although client data was reportedly not accessed, the exposure of internal AI system details and employee data represents a violation of rights and harm to the organization. The AI agent's autonomous attack and the resulting data exposure qualify this as an AI Incident because the AI system's use and malfunction (security vulnerability) directly led to harm. The event is not merely a potential risk but a realized breach with significant consequences.
Thumbnail Image

An AI Agent Broke Into McKinsey's Internal Chatbot and Accessed Millions of Records in Just 2 Hours

2026-03-10
Inc.
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (an AI agent) used to probe and exploit vulnerabilities in another AI system (McKinsey's generative AI chatbot). The AI agent's use directly led to unauthorized access to confidential data, which is a violation of privacy and confidentiality, falling under harm to property and communities. Since the breach has already occurred and caused harm, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

AI agent hacked McKinsey chatbot for read-write access

2026-03-09
TheRegister.com
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (the autonomous AI agent used by CodeWall) that exploited vulnerabilities in another AI system (McKinsey's Lilli chatbot platform). The AI agent's use led directly to unauthorized access to sensitive data and control over the chatbot's behavior, which constitutes harm to property and client confidentiality (harm to property and violation of rights). Although no malicious harm was realized due to prompt patching, the incident itself is an AI Incident because the AI system's use directly caused a security breach with significant potential for harm. The event is not merely a potential hazard or complementary information, as the breach occurred and was demonstrated. Therefore, the classification is AI Incident.
Thumbnail Image

McKinsey Internal AI System Breached By Researcher's AI | Silicon

2026-03-13
Silicon UK
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (McKinsey's internal AI chatbot Lilli) that was breached by another AI system (CodeWall's AI penetration-testing bot). The breach led to unauthorized access to sensitive internal data, which is a form of harm to property and organizational assets. The AI system's malfunction or vulnerability was exploited, leading to realized harm. This fits the definition of an AI Incident because the AI system's use and security failure directly led to harm. Although client data was not accessed, the exposure of internal data and intellectual property is significant harm. Hence, the classification is AI Incident.
Thumbnail Image

Autonomous Agent Hacked McKinsey's AI in 2 Hours

2026-03-13
DataBreachToday
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (an autonomous AI agent) that was used to hack into another AI system (McKinsey's generative AI platform). The AI agent exploited a security flaw to gain unauthorized access to sensitive data and system controls, which is a direct harm to property and potentially to communities if data were misused. The breach was realized, not just a potential risk, and the AI system's use was pivotal in the incident. This meets the criteria for an AI Incident because the AI system's use directly led to harm through unauthorized data access and potential manipulation of AI system behavior.
Thumbnail Image

Dr. AI Yourself? | Quality Remarks Keith Klain

2026-03-13
Quality Remarks - Reflections on a career trying to make software suck less . . .
Why's our monitor labelling this an incident or hazard?
The AI system (McKinsey's AI platform) was hacked, leading to unauthorized access to sensitive data and organizational structure, which is a direct harm to property and intellectual property rights. The involvement of the AI system is explicit, and the harm has materialized through the breach. Therefore, this qualifies as an AI Incident because the development or use of the AI system directly led to harm.