2022 – 2026
ChatGPT Lawsuits
The lawsuits filed against OpenAI over ChatGPT's training, operation, and corporate structure — case captions, courts, filing dates, status, key rulings, and the lay of the docket. NYT v. OpenAI is the precedent-setting fair-use case in the Southern District of New York; Authors Guild, the consolidated newspaper-publisher coalition, Doe v. GitHub Copilot, and Musk v. Altman each plead substantially different theories.
Sibling page: ChatGPT Versions — release timeline with the lawsuits surfaced inline where they shaped a release.
Background
The training-data copyright theory
The dominant theory against OpenAI through 2023 – 2026 has been the same one running against every frontier-model lab: training a large language model requires ingesting tens of billions of words, the cleanest sources of high-quality text are copyrighted books and articles, and copying those works into a training corpus — even temporarily — is reproduction within the meaning of the Copyright Act. Plaintiffs argue the training itself is therefore an infringing use; OpenAI responds that training is transformative fair use under Authors Guild v. Google (the Google Books decision) and Sony v. Universal (the Betamax decision).
The theory shows up in NYT v. OpenAI (newspapers), Authors Guild v. OpenAI (book authors), the Daily News / CIR / Intercept / Raw Story publisher coalition (regional and nonprofit news), and Doe v. GitHub Copilot (open-source code, with OpenAI as a co-defendant alongside GitHub and Microsoft). What's distinctive about the OpenAI docket relative to the Anthropic side is volume and venue: more cases, more plaintiff archetypes, and a concentration in the Southern District of New York rather than the Northern District of California.
The Alsup ruling and its read-across
On June 23, 2025, Judge William Alsup of the Northern District of California granted partial summary judgment in Bartz v. Anthropic. The opinion split the copyright question along a line that had been theoretical until the ruling landed and is now load-bearing: training versus acquisition. Training a model on lawfully-acquired text is fair use; acquiring pirated copies of those works to build the corpus is not. Bartz settled in August 2025 for $1.5 billion before the piracy-damages trial — the largest copyright settlement in U.S. history — but the fair-use ruling on training stood as precedent.
The OpenAI docket sits in a different posture. It is concentrated in the Second Circuit, not the Ninth, and it is mostly stuck at the motion-to-dismiss / discovery stage rather than at summary judgment. The Alsup ruling is therefore persuasive rather than binding for the OpenAI cases, and the NYT court (Judge Sidney Stein) has the first chance to either adopt the Alsup distinction or distinguish it.
The OpenAI cases also have a complication the Anthropic cases lacked: the output-reproduction evidence is more developed. The Times attached examples to its complaint of GPT-4 reproducing paywalled Times articles in response to crafted prompts; the music-publisher case against Anthropic used a similar output-side theory but with lyrics. Whether the Alsup line (train freely on what you've licensed; piracy is the violation) survives contact with serious output-reproduction evidence is the open question this docket will answer.
The publisher coalition
NYT v. OpenAI was the first major publisher case, but it is not alone. The S.D.N.Y. publisher docket also includes the Daily News coalition of eight regional newspapers (April 2024); the Center for Investigative Reporting (Mother Jones / Reveal) case (June 2024), the first nonprofit-news plaintiff in the docket; and the Intercept and Raw Story / AlterNet cases (February 2024), which lead with a DMCA Section 1202 (copyright-management-information) theory rather than direct infringement.
The S.D.N.Y. court dismissed the Raw Story Section 1202 claims in late 2024 for lack of Article III standing — holding that the plaintiff had not pleaded a concrete injury from the alleged CMI removal. Plaintiffs appealed to the Second Circuit; that appeal is the load-bearing piece for whether DMCA Section 1202 survives as an independent theory against AI training. If the Second Circuit reverses, the DMCA-1202 frame is back on the table for every plaintiff who has it pleaded; if it affirms, the direct-copyright theory is the only viable training-input claim left.
The publisher cluster matters separately from the Times case because it tests whether the same training-fair-use analysis travels across publisher size and business model. A ruling that turns on the Times's subscription business and global readership would not obviously help (or hurt) a regional newspaper or a nonprofit news outlet; a ruling that turns on the training process itself would apply uniformly. The cases are informally coordinated for discovery in S.D.N.Y. on that recognition.
Musk v. Altman and the founding-charter question
Musk v. Altman runs on a different track from the copyright cases. There is no infringement claim. The complaint pleads breach of contract on the alleged “founding agreement” that OpenAI would operate as a nonprofit committed to safe, broadly-shared AGI, plus breach of fiduciary duty against the cofounder defendants and (in the federal refiling) civil RICO. The animating allegation is that the for-profit conversion betrayed the founding promise.
Procedurally, the case has been unusually winding. Filed in California state court in February 2024, voluntarily dismissed by Musk in June 2024 on the eve of a demurrer hearing, refiled in the Northern District of California in August 2024 with expanded claims and defendants. Some counts have been dismissed; others survive. Musk's separate motion to preliminarily enjoin the for-profit conversion was denied in early 2025 while leaving the underlying breach claims in place.
The case matters separately from the copyright docket because it puts the OpenAI corporate-structure question into a federal courtroom directly. The November 2023 board episode, the 2024 leadership exodus, and the 2024 – 2025 for-profit-conversion fight all sit downstream of the same structural question; Musk v. Altman is the case that asks a court to answer it. The competitive backdrop — Musk's xAI is a direct competitor to OpenAI — is part of why the case is closely watched and part of why settlement plausibility is hard to read.
Doe v. GitHub Copilot — the first AI-training case
Doe v. GitHub Copilot was filed in November 2022 — before ChatGPT launched. It was the first major lawsuit specifically targeting generative-AI training, and the early dismissals shaped the pleading map for everything that came after.
The complaint stacked theories: direct copyright infringement, DMCA Section 1202 violations, open-source license breaches (MIT, Apache, GPL, BSD attribution requirements), unfair competition, and unjust enrichment. The court dismissed most of the layered claims early, leaving direct copyright on a narrower factual basis — tied to outputs that reproduce identifiable copyrighted code from named plaintiffs' repositories. The DMCA Section 1202 dismissal (for lack of plausible identicality between trained outputs and source code) is the early ruling that subsequent plaintiffs' counsel quietly trimmed their DMCA-1202 and unjust-enrichment theories in response to.
Why it matters separately from the publisher and book-author cases: Copilot is the test of whether open-source licenses are reachable through training. The answer so far is “not as pleaded.” That answer is structurally important — if AI labs can train on permissively-licensed open-source code without triggering the attribution and copyleft requirements those licenses impose on traditional redistribution, the open-source ecosystem looks different than its drafters intended. The narrowed direct-copyright theory is the path the case is now traveling on; whether it reaches class certification with anonymized plaintiffs and Copilot's evolved (post-Codex) model lineup is the next inflection point.
What this docket means for the broader AI bar
OpenAI's docket is the broadest of any AI company's. It runs across every major flavor of AI training-and-operation theory: copyright on training input (NYT, Authors Guild, the publisher coalition, Doe v. Copilot), copyright on training output (NYT's paywall-bypass examples), DMCA Section 1202 on training-time CMI stripping (Intercept, Raw Story), open-source licenses (Doe v. Copilot), and corporate-charter contract theories (Musk v. Altman). The Anthropic docket reaches a subset of these; the OpenAI docket is the full surface.
The broader effect, as of early 2026: the LLM-copyright bar is split between the Northern District of California (where the Alsup ruling and Doe v. Copilot sit) and the Southern District of New York (where the NYT and publisher coalition sit), with active discovery on both coasts and motion-to-dismiss outcomes shaping plaintiffs' counsel's pleading conventions across the docket. Fair-use defenses on properly-acquired training data are stronger after Alsup than before; piracy-sourced corpora are uniquely exposed; output-reproduction theories are alive but not yet tested at trial; the DMCA Section 1202 frame is on appeal in the Second Circuit; and the contract-and-charter theory in Musk v. Altman is its own track. Every other AI lab reads this docket for what survives.