
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
German researchers found that Microsoft's Copilot AI, integrated into Windows, provided medical advice that was potentially harmful. In tests involving 500 queries about medications, only 54% of responses aligned with scientific consensus. Alarmingly, 42% of incorrect advice caused mild to moderate harm, and 22% posed life-threatening risks.[AI generated]
Why's our monitor labelling this an incident or hazard?
The event involves the use of an AI system (Microsoft Copilot) in the medical domain, where its outputs have been analyzed and found to be frequently inaccurate and potentially harmful. The AI system's use has directly led to the identification of risks of injury or harm to health, fulfilling the criteria for an AI Incident. The harm is realized in the sense that the AI system provides advice that could cause harm if relied upon, and the study documents this risk based on actual responses. Therefore, this qualifies as an AI Incident due to the direct link between the AI system's outputs and potential health harm.[AI generated]