Get AI summaries of any video or article — Sign up free
Anthropic AI Copyright Ruling is a BIG Deal: Fair Use Wins, Piracy Loses thumbnail

Anthropic AI Copyright Ruling is a BIG Deal: Fair Use Wins, Piracy Loses

5 min read

Based on AI News & Strategy Daily | Nate B Jones's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Judge William Alup’s ruling supports fair use for AI training when the process is genuinely transformative, but it rejects fair-use protection for training data obtained through piracy.

Briefing

A copyright ruling in the case Barts versus Anthropic draws a sharp line for AI training: model development can qualify as fair use when it is genuinely transformative, but downloading books from pirate sites does not earn the same protection. The decision, issued by Judge William Alup, matters because it offers a workable framework for future AI copyright disputes—one that turns on how training data is obtained, not just what the model does with it.

Alup’s reasoning treats AI training as “quintessentially transformative.” The judge likened the process to how humans read texts and then create new ones—absorbing information and generating outputs that are meaningfully different from the originals. That framing is important because it directly addresses a common legal objection: that training is just copying. By grounding fair-use analysis in transformation, the ruling gives AI companies a conceptual foundation for arguing that training changes the underlying work in a way that goes beyond mere reproduction.

Yet the ruling simultaneously condemns the data acquisition method. The court distinguished between training on books and the earlier step of obtaining those books through piracy. Anthropic had previously trained models using content sourced from Library Genesis and other dubious sites, including for earlier versions of Claude. That choice, the judge made clear, cannot be treated as a free pass.

The case’s practical turning point came in 2024, when Anthropic shifted toward legal procurement. The company hired Tom Turvy, formerly head of Google’s book scanning project, and tasked him with obtaining books through lawful channels. Anthropic then spent millions—described as a significant share of training costs for its new Sonnet and Opus models—buying physical books, often secondhand, and digitizing them by removing them from bindings and scanning them. Although the physical books were destroyed in the process, the court ruled that the resulting digital copies were legitimate fair use because the books were acquired lawfully.

This “Solomon’s choice” approach—fair use for transformative training, but liability for theft—signals how judges may evaluate future AI cases. The decision also notes Anthropic’s financial capability: if the company could afford to buy books later, it could have purchased them earlier. That observation may not eliminate liability, but it could affect the extent of statutory damages.

For authors, the ruling offers a limited but meaningful reassurance. Even if AI outputs are argued to be fair use, the court’s expectation that companies pay for access supports a more sustainable equilibrium for the creative economy. The message is not that authors will automatically win every argument about AI, but that scraping and stealing work without compensation is less likely to be tolerated.

The decision also lands amid other active lawsuits, including Kadre versus Meta over training Llama on Books 3, and potential disputes involving image generation models that may try to extend transformative-use reasoning from text. Broader uncertainty remains because courts differ across jurisdictions, with circuit splits affecting platform liability standards. Still, the Alup framework—transformative use plus scrutiny of lawful data acquisition—provides clearer guidance for where AI copyright fights may head next.

Cornell Notes

Judge William Alup’s ruling in Barts versus Anthropic treats AI training as “quintessentially transformative,” supporting fair use for the creation of new outputs from learned patterns. But the court draws a strict boundary: using pirated books to build models does not receive the same fair-use protection. Anthropic’s 2024 pivot is central—after hiring Tom Turvy and spending millions to buy and scan physical books for its Sonnet and Opus models, the digitized copies were considered legitimate fair use because the underlying books were acquired lawfully. The decision also suggests that a company’s ability to pay for lawful access can influence damages. For authors, the ruling strengthens the expectation that AI companies should compensate rights holders, even when training is argued to be transformative.

What is the core legal distinction the court draws in Barts versus Anthropic?

The ruling separates two issues: (1) whether AI training is transformative enough to qualify as fair use, and (2) whether the training data was obtained lawfully. Judge William Alup found AI training to be “quintessentially transformative,” comparing it to reading and then writing new texts. But the court refused to treat piracy as harmless—Anthropic’s earlier use of books from pirate sources (including Library Genesis) was not excused by the transformative nature of the training.

Why does the court’s “transformative” framing matter for AI companies?

It gives AI companies a conceptual foundation for arguing that training is not mere copying. The judge’s reasoning emphasizes that AI learns from texts and then produces new outputs in different ways, making the process more like human reading-to-writing than reproducing the original work. That approach can influence how future courts assess fair use in training cases, especially when outputs are meaningfully different from the source material.

What role did Anthropic’s 2024 data acquisition shift play in the ruling?

Anthropic moved from relying on pirated content to legally obtaining books. In 2024 it hired Tom Turvy (former head of Google’s book scanning project) with a mandate to legally obtain books. The company then spent millions—described as a significant portion of training costs for its new Sonnet and Opus models—purchasing physical books (often secondhand) and scanning them into digital form after removing them from bindings. The court treated the digitized copies as legitimate fair use because the books were acquired lawfully.

How does the decision address the idea that later lawful purchasing could erase earlier wrongdoing?

The judge said later lawful purchasing does not absolve liability for earlier theft. However, it may affect the extent of statutory damages. The court also noted Anthropic had the money all along, implying that the ability to purchase earlier can weigh against the company when assessing consequences.

What does the ruling imply for authors and the creative economy?

It supports an expectation that AI companies should pay for access to works, even if training is argued to be fair use. The decision offers authors a “glimmer of hope” by making it harder for companies to rely on scraping and stealing without compensation. That expectation is framed as a step toward a more sustainable equilibrium where authors can benefit if they choose to participate in AI tools.

How might this ruling influence other AI copyright cases beyond text training?

The ruling could be used as persuasive reasoning in other pending disputes. The transcript points to lawsuits like Kadre versus Meta (training Llama on Books 3) and suggests that image-generation companies may try to extend the transformative-use logic from text to images. Still, outcomes may vary because courts differ by jurisdiction and there are circuit splits that affect related liability standards for platforms.

Review Questions

  1. How does the ruling connect fair use to both transformation and lawful data acquisition, and why is that combination important?
  2. What specific actions by Anthropic in 2024 were treated as legally significant, and how did they relate to Sonnet and Opus training?
  3. Why might a company’s ability to pay for lawful access affect statutory damages even if training is transformative?

Key Points

  1. 1

    Judge William Alup’s ruling supports fair use for AI training when the process is genuinely transformative, but it rejects fair-use protection for training data obtained through piracy.

  2. 2

    The court’s “quintessentially transformative” framing treats AI training as analogous to reading and creating new texts, not simply reproducing originals.

  3. 3

    Anthropic’s 2024 pivot to lawful book acquisition—hiring Tom Turvy and buying physical books for scanning—was central to the court’s fair-use conclusion for the digitized copies.

  4. 4

    Later lawful purchasing does not erase earlier liability for theft, though it may influence the extent of statutory damages.

  5. 5

    The decision emphasizes that financial capability matters: if a company could have purchased earlier, that fact can weigh against it.

  6. 6

    For authors, the ruling strengthens the expectation that AI companies should compensate rights holders, supporting a more sustainable creative economy.

  7. 7

    Future outcomes may depend on jurisdiction and circuit splits, even as the Alup framework offers clearer guidance for AI copyright disputes.

Highlights

The ruling draws a “Solomon’s choice”: transformative AI training can be fair use, but piracy used to obtain training data is not forgiven.
Anthropic’s lawful procurement in 2024—buying and scanning physical books for Sonnet and Opus—helped the court treat the digitized copies as legitimate fair use.
The court said later lawful purchasing won’t absolve earlier theft, though it may reduce statutory damages.
The decision offers authors a practical lever: courts may expect payment for access, not just transformation-based defenses.

Topics

  • Copyright Fair Use
  • AI Training
  • Piracy vs Lawful Acquisition
  • Statutory Damages
  • Circuit Splits

Mentioned