Microsoft Copilot AI Provides Dangerous Medical Advice

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

German researchers found that Microsoft's Copilot AI, integrated into Windows, provided medical advice that was potentially harmful. In tests involving 500 queries about medications, only 54% of responses aligned with scientific consensus. Alarmingly, 42% of incorrect advice caused mild to moderate harm, and 22% posed life-threatening risks.[AI generated]

Why's our monitor labelling this an incident or hazard?

The event involves the use of an AI system (Microsoft Copilot) in the medical domain, where its outputs have been analyzed and found to be frequently inaccurate and potentially harmful. The AI system's use has directly led to the identification of risks of injury or harm to health, fulfilling the criteria for an AI Incident. The harm is realized in the sense that the AI system provides advice that could cause harm if relied upon, and the study documents this risk based on actual responses. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's outputs and potential health harm.[AI generated]
AI principles
SafetyRobustness & digital securityAccountabilityTransparency & explainabilityHuman wellbeing

Industries
Healthcare, drugs, and biotechnologyIT infrastructure and hostingConsumer products

Affected stakeholders
Consumers

Harm types
Physical (injury)Physical (death)

Severity
AI incident

AI system task:
Interaction support/chatbotsContent generation

In other databases

Articles about this incident or hazard

Thumbnail Image

L'IA potrebbe ucciderti se la usi come medico

2024-10-14
Tom's Hardware
Why's our monitor labelling this an incident or hazard?
This is a case where an AI system’s outputs have been shown to be frequently incorrect in a domain (medical advice) where errors can cause serious physical injury or death. The harms are not reported as having occurred in real patients but are assessed as plausible based on the study’s findings. Therefore, it constitutes an AI Hazard rather than an Incident.
Thumbnail Image

Copilot è pericoloso, non chiedete all'AI di Microsoft consigli su quest'argomento

2024-10-17
IlSoftware.it
Why's our monitor labelling this an incident or hazard?
The event involves the use of an AI system (Microsoft Copilot) in the medical domain, where its outputs have been analyzed and found to be frequently inaccurate and potentially harmful. The AI system's use has directly led to the identification of risks of injury or harm to health, fulfilling the criteria for an AI Incident. The harm is realized in the sense that the AI system provides advice that could cause harm if relied upon, and the study documents this risk based on actual responses. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's outputs and potential health harm.
Thumbnail Image

Microsoft Copilot dà consigli medici pericolosi nel 22% dei casi

2024-10-14
Punto Informatico
Why's our monitor labelling this an incident or hazard?
Microsoft Copilot is an AI system providing medical advice. The study shows that its outputs include dangerous recommendations that could cause serious injury or death if followed, indicating direct or indirect harm to health. This meets the criteria for an AI Incident because the AI system's use has directly or indirectly led to harm to persons. The article does not merely warn of potential harm but reports on actual harmful outputs generated by the AI system, thus qualifying as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Microsoft Copilot non è un buon medico

2024-10-16
MRW.it
Why's our monitor labelling this an incident or hazard?
Microsoft Copilot is an AI system providing medical advice. The study shows that its outputs include inaccurate and potentially dangerous medical recommendations, which could cause injury or death if acted upon. This constitutes harm to health (a), directly linked to the AI system's use. Therefore, this event qualifies as an AI Incident due to realized or highly probable harm stemming from the AI system's outputs in a critical domain like healthcare.