GitHub Copilot Accused of Copyright and License Violations by Emitting Unattributed Code

Thumbnail Image

The information displayed in the AIM should not be reported as representing the official views of the OECD or of its member countries.

GitHub Copilot, an AI code assistant developed by Microsoft and OpenAI, is accused by developers and legal experts of reproducing large segments of copyrighted and open source code without proper attribution or license compliance. This has led to legal investigations and potential lawsuits over intellectual property and licensing violations.[AI generated]

Why's our monitor labelling this an incident or hazard?

GitHub Copilot is an AI system that generates code suggestions based on training data from public repositories. The allegation is that its use of this code violates licensing agreements, specifically regarding attribution requirements. This constitutes a potential violation of intellectual property rights, which falls under harm category (c) in the AI Incident definition. Since the event describes an ongoing legal challenge based on actual use and alleged harm, it qualifies as an AI Incident rather than a hazard or complementary information. The harm is indirect but material, as the AI system's development and use have led to alleged legal rights violations.[AI generated]
AI principles
AccountabilityTransparency & explainability

Industries
IT infrastructure and hosting

Affected stakeholders
BusinessConsumers

Harm types
Economic/PropertyReputational

Severity
AI incident

Business function:
Research and development

AI system task:
Content generation


Articles about this incident or hazard

Thumbnail Image

GitHub faces potential class action for violating coders' rights

2022-10-20
Computing
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions based on training data from public repositories. The allegation is that its use of this code violates licensing agreements, specifically regarding attribution requirements. This constitutes a potential violation of intellectual property rights, which falls under harm category (c) in the AI Incident definition. Since the event describes an ongoing legal challenge based on actual use and alleged harm, it qualifies as an AI Incident rather than a hazard or complementary information. The harm is indirect but material, as the AI system's development and use have led to alleged legal rights violations.
Thumbnail Image

GitHub Users Want to Sue Microsoft For Training an AI Tool With Their Code

2022-10-18
VICE
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that uses a model trained on code scraped from public repositories. The lawsuit alleges that this training constitutes a violation of intellectual property rights, which falls under harm category (c) in the AI Incident definition. Since the event describes an actual legal action based on the AI system's development and use causing rights violations, it qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

GitHub Copilot Fears Copyright Lawsuits Among Developers - Research Snipers

2022-10-20
Research Snipers
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions by learning from large datasets of source code. The article reports that the AI has directly produced code snippets that are copyrighted and not authorized for reuse, leading to legal concerns and potential lawsuits. This is a clear case where the AI system's use has directly led to a violation of intellectual property rights, fulfilling the criteria for an AI Incident. The harm is realized (copyright infringement risk and legal action), not just potential, so it is not merely a hazard or complementary information.
Thumbnail Image

Developers warned: GitHub Copilot code may be licensed

2022-10-20
TechTarget
Why's our monitor labelling this an incident or hazard?
The article centers on the potential legal risks and lawsuits related to GitHub Copilot's use of copyrighted open source code, which involves an AI system generating code suggestions. While there is concern about copyright infringement and responsibility, no actual harm or legal violation has been confirmed or realized yet. The article mainly discusses the possibility of future legal consequences and advises caution, fitting the definition of an AI Hazard rather than an AI Incident or Complementary Information. It is not unrelated because it involves an AI system and potential harm related to intellectual property rights.
Thumbnail Image

GitHub Copilot may steer Microsoft into a copyright lawsuit

2022-10-19
TheRegister.com
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions based on training data that includes publicly available source code. The event describes realized harm in the form of potential copyright infringement and violation of open source licenses due to Copilot reproducing copyrighted code. This directly implicates violations of intellectual property rights, which fits the definition of an AI Incident. The involvement of the AI system's use (generation of code) has directly led to these harms, and legal investigations are underway. Therefore, this event qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

GitHub Copilot apparently violating open source licensing, says programmers - OnMSFT.com

2022-10-18
OnMSFT.com
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (GitHub Copilot) whose development and use have directly led to a violation of intellectual property rights by using licensed open source code without proper attribution. This constitutes a breach of obligations under applicable law protecting intellectual property rights, fitting the definition of an AI Incident.
Thumbnail Image

How GitHub Copilot Could Steer Microsoft Into a Copyright Storm (Slashdot)

2022-10-19
Tech Investor News
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions based on training from public source code. The event describes realized harm in the form of potential copyright infringement and violation of open source licenses due to Copilot reproducing copyrighted code verbatim or nearly so. This constitutes a breach of intellectual property rights, which is explicitly listed as a type of harm under AI Incidents. The involvement of legal investigations and complaints further supports that this is a materialized harm rather than a mere potential risk. Therefore, this event qualifies as an AI Incident.
Thumbnail Image

GitHub Copilot under fire as dev claims it emits 'large chunks of my copyrighted code' * DEVCLASS

2022-10-17
DEVCLASS
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions based on its training data. The developer's claim that Copilot emits large chunks of copyrighted code without attribution or proper licensing indicates a breach of intellectual property rights. The event involves the use of the AI system leading to realized harm in terms of potential copyright violations and license incompatibility. The discussion about the AI system's filtering mechanisms and the responses from GitHub engineers further confirm the AI system's role in the incident. Therefore, this event qualifies as an AI Incident due to the direct link between the AI system's outputs and the violation of intellectual property rights.
Thumbnail Image

Une avocate spécialisée dans l'open source donne son point de vue sur la plainte contre GitHub Copilot, l'assistant de programmation basé sur l'IA

2022-11-10
Developpez.com
Why's our monitor labelling this an incident or hazard?
The article explicitly discusses a legal complaint against GitHub Copilot, an AI system, alleging violations of open source licenses and privacy laws. These relate to intellectual property rights and legal obligations, which fall under the definition of harm (c) violations of human rights or breach of obligations under applicable law intended to protect intellectual property rights. However, the complaint is ongoing, and the article does not report that these violations have been legally established or that harm has concretely occurred due to the AI system's malfunction or misuse. Instead, it provides expert legal perspectives and analysis of the complaint's claims and implications. This fits the definition of Complementary Information, as it updates and contextualizes the AI ecosystem and legal responses without describing a new AI Incident or AI Hazard. There is no direct or indirect evidence of realized harm or plausible future harm beyond the legal dispute itself at this stage.
Thumbnail Image

Première action judiciaire contre GitHub Copilot - Le Monde Informatique

2022-11-07
Le Monde Informatique
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code recommendations based on training from public open source code repositories. The lawsuit claims that Copilot's use of this code violates open source licenses and copyright laws, constituting a breach of intellectual property rights. This is a direct harm linked to the AI system's development and use. The event involves realized harm (legal violations and alleged damages) caused by the AI system's outputs, fitting the definition of an AI Incident. The presence of the AI system, the nature of the harm (copyright infringement and legal violations), and the direct causal link justify classification as an AI Incident.
Thumbnail Image

Microsoft et GitHub accusés de violation de licences open-source

2022-11-08
InformatiqueNews.fr
Why's our monitor labelling this an incident or hazard?
The AI system Copilot is explicitly mentioned and is central to the event. The complaint alleges that Copilot's use and training on open-source code without proper attribution directly leads to violations of copyright and open-source licenses, which are breaches of intellectual property rights. This fits the definition of an AI Incident as the AI system's use has directly led to harm in the form of legal violations and potential financial damages. The event is not merely a potential risk or a complementary update but a concrete legal dispute over realized harms caused by the AI system.
Thumbnail Image

Microsoft is being sued over Github Copilot piracy (TechRadar)

2022-11-07
Tech Investor News
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code based on training data. The lawsuit claims that its training process violated copyright and open source license terms, which are legal protections of intellectual property rights. Since the AI system's development and use have directly led to alleged legal violations and potential damages, this qualifies as an AI Incident under the framework's category (c) violations of human rights or breach of obligations under applicable law protecting intellectual property rights.
Thumbnail Image

OpenAI, Microsoft, and GitHub hit with lawsuit over Copilot

2022-11-07
TheRegister.com
Why's our monitor labelling this an incident or hazard?
The lawsuit directly challenges the legality of the AI system's training and output, alleging infringement of copyright and open-source licenses, which is a breach of intellectual property rights (harm category c). This harm has already occurred as the AI system was used and deployed. Hence, it qualifies as an AI Incident. The removal of Hikvision cameras due to privacy and human rights concerns is a governance response and does not itself describe a new incident or hazard. Other AI product announcements do not describe harm or plausible harm and are unrelated.
Thumbnail Image

GitHub developer files class action lawsuit against Microsoft's open-source 'piracy' Copilot project - OnMSFT.com

2022-11-07
OnMSFT.com
Why's our monitor labelling this an incident or hazard?
The lawsuit directly concerns the development and use of an AI system (GitHub Copilot) that was trained on public open-source code repositories. The claim is that this use violates intellectual property rights and legal obligations, constituting a breach of applicable law protecting intellectual property rights. Since the AI system's development and use are alleged to have caused legal violations affecting a large group of creators, this qualifies as an AI Incident under the framework's definition of violations of intellectual property rights (harm category c).
Thumbnail Image

Joseph Saveri Law Firm and Matthew Butterick File Class-Action Lawsuit Against GitHub, Microsoft, and OpenAI Over Violations

2022-11-07
AiThority
Why's our monitor labelling this an incident or hazard?
The event involves an AI system (GitHub Copilot) whose development and use are alleged to have directly led to violations of intellectual property rights, a breach of applicable law protecting such rights. This fits the definition of an AI Incident because the AI system's use is directly linked to harm (violation of intellectual property rights) experienced by open-source programmers. The filing of a lawsuit indicates that the harm is realized and significant enough to warrant legal action.
Thumbnail Image

Microsoft's GitHub Copilot sued over "software piracy on an unprecedented scale" | IT PRO

2022-11-05
IT PRO
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code suggestions based on training data. The lawsuit claims that the AI was trained on code under open-source licenses requiring attribution, and that this training and the resulting outputs violate those licenses, constituting a breach of intellectual property rights. Since the lawsuit challenges the legality of the AI system's development and use due to alleged rights violations, this constitutes an AI Incident under the definition of violations of intellectual property rights caused by the AI system's development and use.
Thumbnail Image

US coder sues Microsoft and OpenAI for open-source piracy

2022-11-07
Silicon Republic
Why's our monitor labelling this an incident or hazard?
The event involves an AI system, GitHub Copilot, which is explicitly described as an AI coding assistant trained on public source code. The lawsuit claims that the development and use of this AI system have led to violations of intellectual property rights, specifically breaches of open-source licenses and copyright laws. This constitutes a violation of legal obligations intended to protect intellectual property rights, which fits the definition of an AI Incident under category (c). Therefore, this event qualifies as an AI Incident due to the direct link between the AI system's use and the alleged harm to intellectual property rights.
Thumbnail Image

Microsofts GitHub Copilot Sued Over Software Piracy on an Unprecedented Scale (Slashdot)

2022-11-05
Tech Investor News
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (GitHub Copilot powered by OpenAI Codex) whose development and use are alleged to have caused violations of intellectual property rights, a recognized category of harm under the AI Incident definition. The lawsuit claims that the AI system reproduces code without required attribution, violating open-source licenses and legal protections, which constitutes a breach of obligations under applicable law. This is a direct harm caused by the AI system's use and outputs. Therefore, this qualifies as an AI Incident rather than a hazard or complementary information.
Thumbnail Image

OpenAI and Microsoft hit with lawsuit over GitHub Copilot

2022-11-09
AI News
Why's our monitor labelling this an incident or hazard?
GitHub Copilot is an AI system that generates code based on training data consisting of public code repositories. The lawsuit alleges that the system's training and output infringe on copyright and attribution requirements, constituting a violation of intellectual property rights. This is a direct harm related to the AI system's development and use. Since the lawsuit is filed and the harm (violation of rights) is alleged to have occurred, this qualifies as an AI Incident under the framework, specifically under category (c) violations of human rights or breach of obligations under applicable law protecting intellectual property rights.
Thumbnail Image

Microsoft, GitHub and OpenAI Accused of Software Piracy, Sued for $9B in Damages |

2022-11-08
Spiceworks
Why's our monitor labelling this an incident or hazard?
The event explicitly involves an AI system (GitHub Copilot) whose development and use have directly led to alleged violations of intellectual property rights and copyright law, which are harms under the AI Incident definition (c). The lawsuit details how the AI system reproduces licensed code without proper attribution or compliance with license terms, constituting a breach of legal obligations protecting intellectual property rights. This is a direct harm caused by the AI system's use and training data practices. Hence, the event is classified as an AI Incident rather than a hazard or complementary information.