
The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.
Anthropic unveiled its advanced AI model, Claude Mythos, which demonstrated unprecedented ability to detect thousands of critical, previously unknown cybersecurity vulnerabilities. Due to concerns over potential misuse and the risk of cyberattacks, Anthropic is withholding public release, limiting access to a defensive industry consortium and launching Project Glasswing for secure deployment.[AI generated]
Why's our monitor labelling this an incident or hazard?
The article explicitly discusses an AI system (Claude Mythos Preview) with advanced capabilities in vulnerability detection and exploit development, which is a clear AI system involvement. The company acknowledges the dual-use risk, restricting access to prevent malicious use, indicating awareness of plausible future harms. No actual incidents of harm caused by the AI system are reported, only the potential for such harms if the system were to be misused. This fits the definition of an AI Hazard, as the AI system's development and use could plausibly lead to significant harms (e.g., cyberattacks exploiting vulnerabilities). The event is not an AI Incident because no realized harm is described, nor is it Complementary Information or Unrelated, as the focus is on the AI system's capabilities and associated risks.[AI generated]