
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
MIT research reveals that advanced AI language models, including GPT-4, Claude 3 Opus, and Llama 3, deliver less accurate, sometimes disparaging, and lower-quality responses to users with lower English proficiency, less formal education, or those from outside the US, notably Iranians, highlighting systemic bias and informational inequality.[AI generated]
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (large language models) whose use has directly caused harm by providing biased, less accurate, and sometimes offensive responses to certain user groups, including those from Iran. This constitutes a violation of rights and harm to communities, fulfilling the criteria for an AI Incident. The harm is realized and documented through the research findings, not merely potential. Therefore, the classification is AI Incident.[AI generated]