Vulnerabilities in Microsoft Copilot Exposed at Black Hat USA

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

At Black Hat USA, Zenity CTO Michael Bargury revealed vulnerabilities in Microsoft's Copilot AI, demonstrating how cyberattackers could exploit it for data theft and social engineering. Copilot's susceptibility to prompt injections and plugin misuse poses significant cybersecurity risks, urging organizations to enhance security measures and educate users on potential threats.[AI generated]

Why's our monitor labelling this an incident or hazard?

The demonstration shows direct misuse of an AI system leading to privacy violations and enabling phishing attacks. The AI’s access to corporate data was manipulated via prompt injections and misconfigurations, resulting in unauthorized data disclosure and realistic scenarios for financial theft and impersonation. These constitute actual harms (data breaches, violation of privacy rights) caused by the AI system’s malfunction and exploitation.[AI generated]
AI principles
Robustness & digital securityPrivacy & data governanceSafetyTransparency & explainabilityAccountabilityRespect of human rights

Industries
Digital securityIT infrastructure and hosting

Affected stakeholders
Business

Harm types
Human or fundamental rightsEconomic/PropertyReputationalPsychological

Severity
AI incident

Business function:
ICT management and information security

AI system task:
Interaction support/chatbotsContent generation


Articles about this incident or hazard

Thumbnail Image

The Copilot AI Microsoft Built Into Windows Makes It Incredibly Hackable, Research Shows

2024-08-10
Yahoo
Why's our monitor labelling this an incident or hazard?
The demonstration shows direct misuse of an AI system leading to privacy violations and enabling phishing attacks. The AI’s access to corporate data was manipulated via prompt injections and misconfigurations, resulting in unauthorized data disclosure and realistic scenarios for financial theft and impersonation. These constitute actual harms (data breaches, violation of privacy rights) caused by the AI system’s malfunction and exploitation.
Thumbnail Image

Microsoft's AI Can Be Turned Into an Automated Phishing Machine

2024-08-08
Wired
Why's our monitor labelling this an incident or hazard?
The event centers on new research revealing how attackers could misuse Copilot’s generative capabilities to launch automated, personalized phishing at scale. No actual harm is reported, but the described vulnerabilities and attack methods create a plausible risk of future incidents. This aligns with an AI Hazard—an AI-related scenario that could plausibly lead to significant harm.
Thumbnail Image

Zenity CTO on dangers of Microsoft Copilot prompt injections | Tech...

2024-08-08
TechTarget
Why's our monitor labelling this an incident or hazard?
The article describes a proof-of-concept demonstration and release of a red-teaming tool that could be used to carry out prompt injections against Microsoft Copilot, illustrating plausible future attacks but not reporting any real-world compromise or harm. Therefore, it represents an AI Hazard rather than an Incident or mere complementary update.
Thumbnail Image

Copilot users beware! Microsoft's chatbot is vulnerable to cyberattacks

2024-08-10
NewsBytes
Why's our monitor labelling this an incident or hazard?
The article details how Copilot’s design flaws and plugin mechanisms could be abused by malicious actors to steal data and orchestrate social engineering schemes. No actual breach or real‑world data theft is reported, but the exposé demonstrates plausible pathways to significant harm, qualifying it as an AI Hazard.
Thumbnail Image

How to Weaponize Microsoft Copilot for Cyberattackers

2024-08-08
Dark Reading
Why's our monitor labelling this an incident or hazard?
This article focuses on vulnerabilities and demonstrates plausible methods for attackers to misuse an AI system (Microsoft Copilot) for cyberattacks. No actual incident of harm is reported, but the potential for significant malicious use makes this an AI Hazard.
Thumbnail Image

Creating your own Microsoft Copilot chatbot is easy but making it safe and secure is pretty much impossible says security expert

2024-08-09
pcgamer
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (Microsoft Copilot chatbots) whose use has directly led to security harms including malicious code injection and credential phishing attacks. These harms are realized and significant, affecting corporate data security and user credentials. The article details specific incidents and vulnerabilities, not just potential risks, and discusses the consequences of insecure AI deployment. Hence, it meets the criteria for an AI Incident rather than a hazard or complementary information. The presence of actual exploitation and data exposure confirms direct or indirect harm caused by the AI system's use and default insecure configurations.
Thumbnail Image

Copilot, Studio bots are woefully insecure, says Zenity CTO

2024-08-08
TheRegister.com
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (Microsoft Copilot and Copilot Studio bots) whose insecure default configurations and susceptibility to prompt injection attacks have directly led to realized harms, including unauthorized disclosure of sensitive data and potential remote code execution within enterprises. These harms constitute violations of data security and could lead to significant property and community harm. The article describes actual exploits and demonstrations of these vulnerabilities, not just potential risks, thus qualifying as an AI Incident. The involvement of AI in generating responses and processing internal data is explicit, and the harms are direct and significant.
Thumbnail Image

Security Researcher Demos Microsoft Copilot Flaws at Black Hat Conference

2024-08-10
Thurrott.com
Why's our monitor labelling this an incident or hazard?
Microsoft Copilot is an AI system that processes sensitive internal data and automates content generation. The demonstrated flaws allow attackers, after compromising a user account, to exploit Copilot to exfiltrate data and send mass phishing emails mimicking the user's style, which constitutes direct harm to individuals and organizations through data breaches and social engineering attacks. The event involves the use and misuse of the AI system leading to realized harms, meeting the criteria for an AI Incident. The responsible disclosure and ongoing mitigation efforts do not negate the fact that the vulnerabilities have been demonstrated and could have caused or have caused harm.
Thumbnail Image

BlackHat 2024: Remote code execution attack on M365 Copilot via email

2024-08-10
borncity.com
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system, Microsoft 365 Copilot, which is an AI-based chatbot integrated into Microsoft 365. The attack exploits the AI system's functionality via remote code execution triggered by an email, leading to unauthorized data access and manipulation. This directly harms the security and privacy of individuals and organizations, which qualifies as harm to property and communities. The AI system's malfunction or exploitation is a direct cause of the incident. Therefore, this is classified as an AI Incident.
Thumbnail Image

Microsoft's AI Assistant Can Be Exploited by Cybercriminals

2024-08-09
PaymentsJournal
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (Microsoft Copilot) and details how its use can be manipulated by attackers to cause harm such as data breaches, fraud, and phishing attacks. These harms fall under violations of rights and harm to communities. The AI system's misuse is a direct contributing factor to these harms. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

Nálepka AI zatiaľ odrádza zákazníkov od nákupu

2024-08-13
index.sme.sk
Why's our monitor labelling this an incident or hazard?
Although the article involves AI in the context of product marketing and consumer perception, it does not report any direct or indirect harm caused by AI systems, nor does it describe a plausible future harm scenario. It is primarily about societal attitudes and communication strategies related to AI, which fits the definition of Complementary Information as it provides context and understanding about AI's impact on consumer behavior without describing an AI Incident or Hazard.
Thumbnail Image

Umelá inteligencia narazila na kameň. Nálepka AI odrádza zákazníkov od nákupu, uvádza štúdia

2024-08-13
trend.sk
Why's our monitor labelling this an incident or hazard?
The article focuses on consumer attitudes and perceptions regarding AI in products, highlighting a study that shows customers are deterred by the AI label. There is no mention of any AI system malfunctioning, causing injury, violating rights, or posing a plausible future risk of harm. The content is about societal response and understanding of AI, which fits the definition of Complementary Information as it provides context and insights into the AI ecosystem without reporting a new incident or hazard.
Thumbnail Image

Známa AI aplikácie otvára hackerom dvere do zariadenia. Bezpečnostný expert ukázal závažné slabiny v populárnej aplikácii

2024-08-12
Vosveteit.sk - Správy zo sveta technológií a vedy
Why's our monitor labelling this an incident or hazard?
Microsoft Copilot is an AI system integrated with Microsoft 365 applications, designed to assist users by generating outputs based on input data. The article explicitly states that hackers can exploit vulnerabilities in Copilot to install backdoors and manipulate AI responses, enabling data theft and social engineering attacks. This constitutes direct harm to users' privacy and security, fulfilling the criteria for an AI Incident under violations of rights and harm to individuals. The Recall feature's continuous screenshot capturing without content filtering further exacerbates privacy risks, reinforcing the incident classification. The harms are realized, not just potential, and the AI system's malfunction and misuse are pivotal in causing these harms.