GPT-4 Demonstrated Capable of Autonomously Exploiting Real-World Security Vulnerabilities

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

Researchers at the University of Illinois Urbana-Champaign showed that OpenAI's GPT-4 can autonomously exploit real-world software vulnerabilities by reading public advisories, achieving an 87% success rate. This capability, unmatched by other AI models or traditional tools, raises significant concerns about AI-enabled cyberattacks and system security.[AI generated]

Why's our monitor labelling this an incident or hazard?

The article explicitly involves GPT-4, an AI system, demonstrating autonomous exploitation of zero-day vulnerabilities, which are critical security flaws that can lead to serious harm if exploited. While no actual harm or attack is reported as having occurred, the research shows a credible risk that such AI capabilities could be used maliciously to cause cybersecurity incidents. This fits the definition of an AI Hazard, as the AI system's use could plausibly lead to incidents causing harm to property, organizations, or communities through cyberattacks. The event is not an AI Incident because no actual harm has been reported yet, nor is it merely Complementary Information or Unrelated, as the focus is on the AI system's potential to cause harm through autonomous exploitation of vulnerabilities.[AI generated]
AI principles
Robustness & digital securitySafetyAccountabilityPrivacy & data governanceRespect of human rightsTransparency & explainability

Industries
Digital securityIT infrastructure and hostingGovernment, security, and defence

Affected stakeholders
BusinessGovernmentGeneral public

Harm types
Economic/PropertyReputationalPublic interestHuman or fundamental rights

Severity
AI hazard

Business function:
ICT management and information securityResearch and development

AI system task:
Reasoning with knowledge structures/planningGoal-driven organisationContent generation


Articles about this incident or hazard

Thumbnail Image

GPT-4 es capaz de explotar vulnerabilidades con 87 % de éxito y sin intervención humana

2024-04-25
El Tiempo
Why's our monitor labelling this an incident or hazard?
GPT-4 is explicitly identified as an AI system. The event involves the AI system's use (its capability to autonomously exploit vulnerabilities) which directly relates to potential harm such as security breaches and associated damages. Since the article describes realized capability to exploit vulnerabilities autonomously, this constitutes an AI Incident due to the direct link between the AI system's use and potential or actual harm. The high success rate and autonomous nature of exploitation imply a direct role of the AI in causing harm or enabling harm, meeting the criteria for an AI Incident.
Thumbnail Image

GPT-4 es capaz de explotar vulnerabilidades con 87% de éxito y sin intervención humana

2024-04-25
El Comercio Perú
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) that autonomously exploits security vulnerabilities, which is a direct use of AI leading to potential harm (security breaches, exploitation of zero-day vulnerabilities). This constitutes an AI Incident because the AI's use has directly led to the capability to cause harm, even if the harm is demonstrated in a research setting rather than a real-world attack. The AI system's development and use are central to the event, and the harm category includes harm to property and communities through cybersecurity threats. Therefore, this event meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

Descubren que GPT-4 puede explotar vulnerabilidades de día cero al...

2024-04-25
europa press
Why's our monitor labelling this an incident or hazard?
The article explicitly involves GPT-4, an AI system, demonstrating autonomous exploitation of zero-day vulnerabilities, which are critical security flaws that can lead to serious harm if exploited. While no actual harm or attack is reported as having occurred, the research shows a credible risk that such AI capabilities could be used maliciously to cause cybersecurity incidents. This fits the definition of an AI Hazard, as the AI system's use could plausibly lead to incidents causing harm to property, organizations, or communities through cyberattacks. The event is not an AI Incident because no actual harm has been reported yet, nor is it merely Complementary Information or Unrelated, as the focus is on the AI system's potential to cause harm through autonomous exploitation of vulnerabilities.
Thumbnail Image

Un estudio demostró que los modelos de inteligencia artificial como GPT son capaces de efectuar ciberataques de manera autónoma

2024-04-23
Rosario3
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (LLMs such as GPT-4) being tested for their ability to autonomously exploit cybersecurity vulnerabilities. While no actual harm has been reported yet, the study shows that such AI systems could plausibly be used to carry out cyberattacks, which would constitute harm to property, data, and potentially critical infrastructure. Therefore, this event qualifies as an AI Hazard because it describes a credible risk of future AI-driven cyberattacks based on demonstrated autonomous capabilities, but no realized harm or incident is reported.
Thumbnail Image

Investigadores advierten nuevos riesgos de seguridad con el Chat GPT-4

2024-04-25
El Vocero de Puerto Rico
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) and its autonomous capability to exploit security vulnerabilities, which is a direct AI system use. Although no actual harm is reported as having occurred yet, the demonstrated ability to exploit zero-day vulnerabilities indicates a credible risk of future harm, including cybersecurity breaches and associated damages. This fits the definition of an AI Hazard, as the development and use of GPT-4 could plausibly lead to an AI Incident involving harm to property, systems, or communities. The article also discusses mitigation strategies, but the main focus is on the potential risk rather than an incident that has already happened.
Thumbnail Image

GPT-4 ya puede explotar las vulnerabilidades 0-day - MuyComputerPRO

2024-04-25
MuyComputerPRO
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) capable of autonomously exploiting zero-day vulnerabilities, which are critical security flaws that can cause harm to computer systems and data. The study shows this capability with a high success rate, indicating a real and present risk. While no actual cyberattacks are reported as having occurred due to this AI, the demonstrated ability and the concern about democratizing cybercrime tools constitute a plausible future harm. This fits the definition of an AI Hazard, as the AI's development and use could plausibly lead to an AI Incident involving harm to property, communities, or critical infrastructure through cyberattacks. The article does not report an actual incident of harm yet, so it is not an AI Incident. It is not merely complementary information because the main focus is on the potential threat posed by the AI's capabilities, not on responses or ecosystem updates.
Thumbnail Image

GPT-4 de OpenAI puede explotar de forma autónoma el 87% de las vulnerabilidades de un día - Notiulti

2024-04-26
Notiulti
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) that autonomously exploits software vulnerabilities, which is a direct use of AI leading to potential or actual harm. The harm includes increased risk of cyberattacks exploiting known vulnerabilities, which can cause financial damage, data breaches, or disruption of services—harms covered under the AI Incident definition (harm to property, communities, or critical infrastructure). The study shows the AI system's capability is real and effective, not hypothetical, and the researchers acknowledge the potential for malicious use. Thus, the event meets the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

ChatGPT can craft attacks based on chip vulnerabilities -- GPT-4 model tested by UIUC computer scientists

2024-04-18
Yahoo
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) performing complex tasks of understanding and exploiting cybersecurity vulnerabilities, which is a clear AI system involvement. The event stems from the AI system's use in exploiting vulnerabilities, demonstrating a capability that could be misused maliciously. Although no actual harm or cybersecurity breach has been reported as a result of this research, the demonstrated ability plausibly leads to future harms such as system breaches, data theft, or infrastructure disruption. Hence, it fits the definition of an AI Hazard rather than an AI Incident. The article does not describe any realized harm or incident, nor does it focus on responses or governance measures, so it is not Complementary Information. It is not unrelated because it clearly involves AI and potential harm.
Thumbnail Image

GPT-4 can exploit real vulnerabilities by reading advisories

2024-04-17
TheRegister.com
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4 integrated with automation frameworks) that has been demonstrated to autonomously exploit real-world security vulnerabilities, which constitutes a direct link to harm (security breaches, potential damage to systems and data). This meets the definition of an AI Incident because the AI system's use has directly led to a capability to cause harm. The harm is not hypothetical but demonstrated through successful exploitation of vulnerabilities. Therefore, this event qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

GPT-4 Can Exploit Most Vulns Just by Reading Threat Advisories

2024-04-18
Dark Reading
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (GPT-4 combined with a framework and tools) that can autonomously exploit software vulnerabilities, which is a clear AI system involvement. The use of the AI system to exploit vulnerabilities is a use case that could plausibly lead to harm, such as unauthorized access, data breaches, or disruption of critical infrastructure. Although the article does not report actual incidents of harm caused by this AI-enabled exploitation, the demonstrated capability and the potential for malicious actors to adopt this technology represent a credible threat. Hence, it is an AI Hazard rather than an AI Incident. The article also discusses the implications and the need for security best practices, reinforcing the assessment of plausible future harm.
Thumbnail Image

OpenAI GPT-4 Demonstrates High Success Rate in Exploiting Real-World System Flaws, Study Shows - WinBuzzer

2024-04-18
WinBuzzer
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) autonomously exploiting real-world system vulnerabilities, which is a direct use of AI leading to potential harm such as breaches of security and exploitation of critical systems. The AI's role is pivotal in the exploitation process, and the harms associated with such exploits (e.g., damage to systems, data breaches) fall under harm to property and communities. Although the article is a research study, it documents realized AI capabilities that have direct harmful implications, meeting the criteria for an AI Incident rather than a hazard or complementary information.
Thumbnail Image

ChatGPT can exploit computer security vulnerabilities very easily, here's how

2024-04-17
Gearrice
Why's our monitor labelling this an incident or hazard?
The event involves the use of an AI system (ChatGPT/GPT-4) to exploit computer security vulnerabilities, which directly leads to harm in the form of cyberattacks and potential damage to computer systems and data. The AI's capability to generate exploit code and ransomware instructions constitutes a direct link to harm (a form of harm to property and potentially to communities). Therefore, this qualifies as an AI Incident because the AI system's use has directly led to realized harms through enabling cyberattacks. The article also mentions that OpenAI requested the non-release of the automated script, indicating awareness of the harm potential. Hence, the classification is AI Incident.
Thumbnail Image

Could ChatGPT be the next big cybersecurity worry -- researchers say it can crack systems faster than ever before

2024-04-23
TechRadar
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) demonstrating the ability to exploit cybersecurity vulnerabilities, which is a direct use of the AI system. This capability could plausibly lead to AI Incidents involving harm to property, infrastructure, or communities through cyberattacks. Since the article reports research findings and raises concerns about potential threats rather than describing an actual realized harm event, it fits the definition of an AI Hazard. The AI system's role is pivotal in the plausible future harm scenario. Therefore, the event is best classified as an AI Hazard.
Thumbnail Image

ChatGPT Plus can exploit zero-day security vulnerabilities -- why this should concern you

2024-04-23
Mashable
Why's our monitor labelling this an incident or hazard?
The article explicitly involves GPT-4, a large language model AI system, demonstrating autonomous exploitation of zero-day vulnerabilities, which is a clear AI system involvement. The study shows the AI's use in security penetration, which could plausibly lead to cyberattacks (harm to property, communities, or systems). Since the article discusses potential risks and implications without reporting actual realized harm or incidents, it fits the definition of an AI Hazard. The event does not describe an actual AI Incident because no direct or indirect harm has yet occurred, only the demonstrated capability and risk of future harm. The article also discusses governance and mitigation suggestions but focuses mainly on the hazard posed by the AI's capabilities.
Thumbnail Image

GPT-4 can exploit zero-day security vulnerabilities all by itself, a new study finds

2024-04-22
TechSpot
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) that is demonstrated to autonomously exploit zero-day vulnerabilities, which are known to cause harm to computer systems and data. The AI's capability to perform such exploitation directly relates to harm (property and data harm) and cybercrime facilitation. The study's findings indicate realized harm potential through the AI's use, not just a hypothetical risk, as GPT-4 successfully exploited 87% of tested vulnerabilities. Therefore, this event meets the criteria for an AI Incident due to the direct link between the AI system's use and significant harm.
Thumbnail Image

GPT-4 can exploit security flaws on its own, study shows

2024-04-23
BGR
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (GPT-4) explicitly used to autonomously exploit cybersecurity vulnerabilities, which is a clear AI system involvement. The study shows the AI's use in a way that could lead to harm (cybersecurity breaches), but no actual harm or incident is reported as having occurred yet. The researchers emphasize the potential risks and the need for defensive measures, indicating plausible future harm. Hence, the event is best classified as an AI Hazard rather than an AI Incident or Complementary Information.
Thumbnail Image

GPT-4 capable of autonomously exploiting vulnerabilities in real world systems through security advisories, says study

2024-04-24
MediaNama
Why's our monitor labelling this an incident or hazard?
The study involves the use of GPT-4, an AI system, autonomously exploiting software vulnerabilities, which is a direct use of AI. While the experiments were conducted in a sandbox environment preventing real damage, the demonstrated capability indicates a credible potential for future misuse leading to AI Incidents such as cyberattacks causing harm to property or critical infrastructure. Therefore, this event qualifies as an AI Hazard because it plausibly could lead to significant harm through malicious exploitation of vulnerabilities by AI systems.
Thumbnail Image

LLM and its new role in cybersecurity

2024-04-21
Aspetuck News
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (LLMs like GPT-4) that autonomously exploits cybersecurity vulnerabilities, which is a direct use of AI leading to potential harm in the form of cybersecurity breaches. The research shows that GPT-4 can successfully exploit 1-day vulnerabilities, which if applied maliciously in real-world systems, would cause harm to property, systems, and potentially communities. Although the experiments were done in isolated environments to prevent actual harm, the demonstrated capability and discussion of misuse risks indicate a direct AI Incident because the AI system's use has led to realized harm in controlled settings and reveals a clear pathway to harm in real-world applications. The ethical considerations and calls for responsible use further support the classification as an AI Incident rather than a mere hazard or complementary information.
Thumbnail Image

Index - Tech-Science - According to a study, GPT-4 is an excellent hacker

2024-04-21
newsbeezer.com
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) and its use in exploiting cybersecurity vulnerabilities, which can lead to harm such as disruption of critical infrastructure or damage to property and communities. Since the study shows the AI's capability to autonomously create exploit programs, this represents a plausible future risk of AI-driven cyberattacks. No actual harm or incident is reported as having occurred yet, so this event fits the definition of an AI Hazard rather than an AI Incident. The AI system's development and use are central to the event, and the potential for harm is credible and significant.
Thumbnail Image

AI's use as a hacking tool has been overhyped

2024-04-22
ITPro
Why's our monitor labelling this an incident or hazard?
The event involves AI systems (LLMs) and their potential use in hacking, which is a security-related application. However, the article focuses on research findings about the current capabilities and limitations of these AI systems rather than describing any realized harm or malicious use causing injury, rights violations, or other harms. The article discusses plausible future risks but does not report any actual AI-driven incidents or harms. Therefore, this qualifies as an AI Hazard, as it highlights a credible potential for AI to be used maliciously in the future, but no incident has occurred yet.
Thumbnail Image

GPT-4 Is Capable Of Exploiting 87% Of One-Day Vulnerabilities

2024-04-22
Cyber Security News
Why's our monitor labelling this an incident or hazard?
The article explicitly involves an AI system (GPT-4) used to exploit known cybersecurity vulnerabilities, which is a direct use of AI leading to harm in the cybersecurity domain. The exploitation of vulnerabilities can cause harm to property (computer systems) and communities relying on those systems. The research shows GPT-4's high success rate in exploiting these vulnerabilities, indicating realized harm potential rather than mere hazard. Therefore, this qualifies as an AI Incident due to the direct link between AI use and harm potential in cybersecurity.