Global AI Speech Recognition Models Fail Indian Languages, Impacting Critical Services

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

A national benchmark, 'Voice of India', reveals that global AI speech recognition systems from OpenAI, Microsoft, Google, and Meta perform poorly with Indian languages and dialects. This leads to high error rates, risking miscommunication in essential services like welfare and healthcare for millions in India.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves AI systems (speech recognition models from OpenAI, Meta, Microsoft, and Sarvam). It discusses their use and performance shortcomings in Indian languages, which could plausibly lead to harm in critical domains like welfare and medical applications due to transcription errors. However, no actual harm or incident is reported, only the potential for harm due to high error rates. Hence, it fits the definition of an AI Hazard rather than an AI Incident or Complementary Information. It is not unrelated as it clearly concerns AI systems and their impact.[AI generated]
AI principles
FairnessRobustness & digital security

Industries
Government, security, and defenceHealthcare, drugs, and biotechnology

Affected stakeholders
General public

Harm types
Economic/PropertyPublic interest

Severity
AI hazard

Business function:
Citizen/customer service

AI system task:
Recognition/object detection


Articles about this incident or hazard

Thumbnail Image

Global speech AI struggles to understand India: Report

2026-02-16
Economic Times
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems—automatic speech recognition models from global providers like OpenAI, Microsoft, Google, and Meta. The use of these AI systems in voice interfaces for critical services means their malfunction (high word error rates and poor recognition of dialects) directly leads to harm such as miscommunication in welfare applications, healthcare, and other essential services. This constitutes indirect harm to persons and communities through disruption or degradation of critical infrastructure services. The harm is realized, not merely potential, as the article details the performance failures and their practical consequences. Hence, the event meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

OpenAI, Meta speech models struggle with Indian languages

2026-02-16
@businessline
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems (speech recognition models from OpenAI, Meta, Microsoft, and Sarvam). It discusses their use and performance shortcomings in Indian languages, which could plausibly lead to harm in critical domains like welfare and medical applications due to transcription errors. However, no actual harm or incident is reported, only the potential for harm due to high error rates. Hence, it fits the definition of an AI Hazard rather than an AI Incident or Complementary Information. It is not unrelated as it clearly concerns AI systems and their impact.
Thumbnail Image

Global Speech AI Struggles to Understand India: New National Benchmark 'Voice of India' Reveals

2026-02-16
www.newspatrolling.com
Why's our monitor labelling this an incident or hazard?
The article explicitly involves AI systems—automatic speech recognition models—and discusses their development and use in real-world Indian contexts. It identifies significant performance deficiencies that could plausibly lead to harm, such as miscommunication in critical services accessed via voice interfaces. Although no direct harm has been reported yet, the described performance gaps and their implications for digital inclusion and service access constitute a credible risk of harm. The article's focus is on revealing these risks and encouraging improvements, not on reporting an actual incident or harm caused by AI. Hence, the event fits the definition of an AI Hazard, as it plausibly could lead to AI incidents if the issues remain unaddressed.