AI Symptom Checkers Found Inaccurate, Risking Public Health

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Australian research reveals that AI-powered online symptom checkers provide correct diagnoses only about a third of the time and accurate triage advice in less than half of cases. Their frequent inaccuracies risk misleading users, potentially causing unnecessary medical visits or dangerous delays in seeking proper care.[AI generated]

Why's our monitor labelling this an incident or hazard?

The online symptom checkers are AI systems that analyze user input to provide health diagnoses and recommendations. The study highlights that these AI systems often provide inaccurate or misleading information, which can indirectly lead to harm by causing users to misjudge the severity of their health conditions or delay seeking proper medical care. This constitutes indirect harm to health (a), as users may make poor health decisions based on faulty AI outputs. Therefore, this event qualifies as an AI Incident due to the realized harm stemming from the use of AI systems in health diagnosis.[AI generated]
AI principles
SafetyAccountabilityHuman wellbeing

Industries
Healthcare, drugs, and biotechnology

Affected stakeholders
ConsumersGeneral public

Harm types
Physical (injury)

Severity
AI incident

Business function:
Citizen/customer service

AI system task:
Forecasting/predictionInteraction support/chatbots


Articles about this incident or hazard

Thumbnail Image

Online health diagnosis tools only accurate around 36 pct of the time: Aussie study - Xinhua | English.news.cn - Xinhua

2020-05-18
新华网
Why's our monitor labelling this an incident or hazard?
The online symptom checkers are AI systems that analyze user input to provide health diagnoses and recommendations. The study highlights that these AI systems often provide inaccurate or misleading information, which can indirectly lead to harm by causing users to misjudge the severity of their health conditions or delay seeking proper medical care. This constitutes indirect harm to health (a), as users may make poor health decisions based on faulty AI outputs. Therefore, this event qualifies as an AI Incident due to the realized harm stemming from the use of AI systems in health diagnosis.
Thumbnail Image

Two-in-three people have been misdiagnosed after seeking help online

2020-05-18
Daily Mail Online
Why's our monitor labelling this an incident or hazard?
Online symptom checkers are AI systems that analyze symptoms to provide diagnostic suggestions and triage advice. The study reveals that these systems are often inaccurate, with only 36% correct diagnoses and inappropriate emergency advice in many cases. This misdiagnosis and poor triage can cause harm to individuals by delaying proper medical care or causing unnecessary emergency visits, thus impacting health outcomes. The harm is realized and directly linked to the AI systems' outputs. Hence, this qualifies as an AI Incident due to indirect harm to health caused by the AI systems' malfunction or limitations.
Thumbnail Image

Symptom Checker Apps Are Awful at Guessing the Right Illness, Study Finds

2020-05-18
Gizmodo
Why's our monitor labelling this an incident or hazard?
The symptom checker apps are AI systems as they use AI algorithms to infer diagnoses and triage recommendations from user input. The study shows these AI systems' use has led to inaccurate diagnoses and triage advice, which can indirectly cause harm to users' health by misleading them about the seriousness of their conditions or delaying necessary care. This constitutes indirect harm to health (a) due to the AI systems' use. Therefore, this event qualifies as an AI Incident because the AI systems' use has directly or indirectly led to harm risks that are realized or highly plausible, as supported by the study's findings and expert caution.
Thumbnail Image

Australian research finds Doctor Google is usually wrong

2020-05-18
The Express Tribune
Why's our monitor labelling this an incident or hazard?
Online symptom checkers are AI systems that generate medical predictions based on user symptoms. The research shows these systems frequently provide incorrect diagnoses and inaccurate advice on seeking medical help, which can lead to harm if users rely on them instead of professional medical consultation. This constitutes indirect harm to health due to the AI system's use, fitting the definition of an AI Incident.
Thumbnail Image

Why online symptom checkers often yield terrible advice

2020-05-18
Salon.com
Why's our monitor labelling this an incident or hazard?
The article explicitly describes symptom checkers as algorithm-based programs that provide diagnostic and triage advice, which fits the definition of AI systems. It reports that these systems often provide inaccurate or misleading advice, which can indirectly lead to harm to individuals' health if they rely on incorrect diagnoses or triage recommendations. Therefore, the event involves the use of AI systems leading to indirect harm to health, qualifying it as an AI Incident under the framework.
Thumbnail Image

New research finds 'Dr. Google' is almost always wrong

2020-05-17
Medical Xpress - Medical and Health News
Why's our monitor labelling this an incident or hazard?
The article explicitly discusses AI-based symptom checkers (AI systems) and their performance in diagnosis and triage advice. While no direct harm is reported, the inaccuracies and lack of quality control imply a credible risk of harm to users' health (harm category a). The AI systems' use could plausibly lead to incidents such as misdiagnosis or inappropriate healthcare seeking behavior. Hence, this qualifies as an AI Hazard rather than an AI Incident. It is not Complementary Information because it is not an update or response to a prior incident, nor is it unrelated as it clearly involves AI systems and their impact on health.
Thumbnail Image

'Dr Google' wrong most of the time, study finds

2020-05-18
Institution of Engineering and Technology
Why's our monitor labelling this an incident or hazard?
The symptom checkers are AI systems that infer diagnoses and triage advice from user symptoms. The study shows these systems produce correct diagnoses only about a third of the time and provide triage advice that is only accurate about half the time, with a tendency toward risk-averse recommendations that may cause unnecessary urgent care visits. This inaccurate output can directly or indirectly harm users' health by misleading them about the seriousness of their conditions or prompting unnecessary medical actions. The harm is realized as users rely on these AI systems for health decisions, fulfilling the criteria for an AI Incident under the framework.
Thumbnail Image

Symptom Checker Apps Are Awful At Guessing The Right Illness, Study Finds

2020-05-18
Gizmodo AU
Why's our monitor labelling this an incident or hazard?
The article involves AI systems (symptom checker apps using AI algorithms) and discusses their use and performance. While it identifies significant risks and potential for harm (misdiagnosis, incorrect triage advice), it does not report a concrete event where these apps have directly or indirectly caused injury, rights violations, or other harms. Therefore, it does not meet the threshold for an AI Incident. It also does not focus primarily on a future plausible harm scenario or a near miss, so it is not an AI Hazard. The article provides important contextual information about AI system performance and risks, which enhances understanding of the AI ecosystem and informs future risk assessment. Hence, it fits best as Complementary Information.
Thumbnail Image

Don't use 'Dr Google' to self-diagnose Covid-19 symptoms

2020-05-18
@dispatch_DD
Why's our monitor labelling this an incident or hazard?
Online symptom checkers are AI systems that analyze user input to provide medical diagnoses. The article states that these systems are inaccurate most of the time, which can indirectly lead to harm by causing misdiagnosis or delayed proper treatment. Although no specific incident of harm is described, the unreliability of these AI systems in medical diagnosis represents a plausible risk of harm to health. However, since the article focuses on the general unreliability and potential danger rather than a specific realized harm event, this qualifies as an AI Hazard rather than an AI Incident.