
Until two weeks ago, no U.S. court had ruled on whether training generative AI models on copyrighted works could constitute a fair use, or if the simple act of training such models without a license would constitute copyright infringement. Two new summary judgment rulings out of the Northern District of California have now set the stage for how future courts may interpret fair use in the context of AI training—Richard Kadrey, et al., v. Meta Platforms, Inc. and Andrea Bartz, et al. v. Anthropic PBC. That stage poses significant dangers for technology companies.
In both cases, the court held that using copyrighted books to train large language models (LLMs), like Anthropic’s Claude and Meta’s LLaMA, may qualify as a fair use. Yet neither case constitutes an unalloyed victory for the defendants. Indeed, the two new decisions, while conflicting on important particulars, both suggest that some plaintiffs could be in a strong position to defeat a fair use defense—especially where the copyrighted content is not books. Both decisions offer key insights for rights holders, AI developers, and litigators navigating this unsettled frontier.
The Kadrey and Bartz fair use findings rest on some important commonalities. Both courts held that using copyrighted books to train LLMs is transformative, even when the full text of a work is ingested. Copying the full text of the plaintiffs’ copyrighted books was found to be both necessary and reasonable for LLM training, even though the courts acknowledged the books are creative works, a factor which traditionally weighs against fair use. The purpose of AI training—to enable models to learn statistical patterns of language to generate new content, rather than to reproduce or compete with original works—proved to be determinative in finding a transformative use under the first factor of the fair use analysis. This factor is the second most important factor in a fair use analysis.
While agreeing on the application of these foundational principles, the Kadrey and Bartz decisions sharply diverge in their treatment of data acquisition, the implications of using pirated materials, and market harm.
The most significant difference is how the courts framed their analyses of the fair use defense. In Anthropic’s case in Bartz, the Court divided the AI training process into two distinct stages: data collection and model training. Thus, when the Bartz Court found that training AI models was transformative and a fair use, this was only a partial defense to liability for Anthropic because the Bartz Court separately analyzed how Anthropic acquired its training data. Anthropic used two different approaches for this latter task, leading to different results for the fair use analysis. In the first approach, the Court concluded that permanently converting legally purchased hardbacks into a digitally-formatted “central library” was transformative and a fair use. In the second, the Court took great issue with Anthropic’s use of pirated copies of the Plaintiffs’ novels from online “shadow libraries.” For those copies, it ruled that building and retaining a "central library" of pirated books could not be a fair use. Notably, the Bartz Court referred repeatedly to Anthropic’s independent goal of building a library of “all the books in the world” and to its practice of retaining copies of pirated books, even after Anthropic had decided never to use them in training its models again. Ultimately, because of the two different sources of books used in training Claude, the Bartz Court granted Anthropic summary judgment as to the lawfully acquired and converted hardbacks, but it denied Anthropic’s motion with respect to its use of pirated online materials.
The Kadrey Court, by contrast, collapsed Bartz’s multi-stage analysis into a single step. While the plaintiffs in Kadrey, like in Bartz, alleged that the accused model had been trained on pirated data, the Kadrey Court focused only on how Meta used the data and whether its use in training LLaMA was transformative. The Court did not conduct a separate fair use analysis into how Meta acquired the works. Finding the model’s use of the books was transformative then carried the Court a long ways towards a fair use finding.
The second key distinction between the Anthropic and Meta cases rests in their analyses of market harm. The Kadrey Court emphasized that the harm an infringing work inflicts on the market, or potential market, for the copyrighted work is “the single most important element” in determining fair use and thus focused its most attention on this element. And its ruling was consistent with the concerns it had expressed at the motion hearing: the Court demonstrated substantial apprehension that a model trained with copyrighted books “might harm the market for those works…by helping to enable the rapid generation of countless works that compete with the originals, even if those [new] works aren’t themselves infringing.” But the Court was stymied in analyzing this market dilution theory and could not rule in the plaintiffs’ favor on summary judgment. The plaintiffs there had not pled, argued or offered any evidence to demonstrate a triable issue of fact as to market dilution causing harm to the markets for their works. The Meta Court thus found that the “most important” factor in fair use of market harm favored Meta, and, when combined with its finding that Meta’s use of the books was transformative, that Meta’s fair use defense applied. In short, while the Court dismissed the plaintiffs’ claims because of a lack of evidence, it left the door wide open for future plaintiffs to pursue claims with better developed theories and evidence substantiating market harm.
The Anthropic Court’s approach to the possibility of market dilution by AI-written books was starkly different. While the plaintiffs in that case did argue that AI models trained with their works would generate an “explosion” of competing works, the Bartz Court gave this argument short shrift. It viewed any works generated by AI models as merely being “competition” for the plaintiffs’ books—a type that the Copyright Act is not intended to prevent. The Meta Court, by contrast, viewed such competition as being (at least potentially) an “indirect substitution”, which is protected under the Copyright Act. It noted that the models were more effective because they had been trained on the creative expression in the copyrighted novels and highlighted that AI models, unlike infringing human authors, could “generate literally millions of secondary works, with a miniscule fraction of the time and creativity used to create the original works [on which] it was trained….” This difference in the scale of the threat of market substitution, disregarded by the Anthropic Court, weighed heavily with the Meta Court.
Taken together, the Anthropic and Meta decisions offer a glimpse into what to expect moving forward and into the dangers that technology companies face from copyright suits. While both cases affirm the possibility that AI training can qualify as fair use, they also reflect emerging fault lines along which liability could be found: between lawful and unlawful data acquisition, between training and storage, and between presumed and proven market harm.
In considering these opinions, one also cannot ignore that the plaintiffs in both cases were novelists. Other types of plaintiffs, such as newspapers, visual artists, or music publishers, may find it much easier to prove that the outputs of AI models infringe on their works, are less transformative, and that they dilute the demand for their works in a way that constitutes cognizable market harm. The prospects for a successful fair use defense in such circumstances seems uncertain at best. For litigators and rights holders, these decisions thus mark both a moment of clarity and a preview of the complexities to come.
Key Takeaways
- Use context matters. Courts may separate training from storage or reuse, and assess each under a different legal lens.
- Lawful acquisition counts, especially when a court analyzes acquisition separately from training. Rights holders should monitor whether platforms are sourcing data from authorized, licensed repositories. To discourage courts from analyzing data acquisition separately, companies that have collected unlicensed works should make them available only for the purpose of training their generative AI models. If the companies decide to stop using certain works (or data sets) for training purposes, then those works should be promptly deleted.
- Evidence of market harm is essential. Speculation about market harm or regurgitation is not enough. Rights holders must offer evidence of actual output overlap or market dilution. Rights holders should gather such evidence early in litigation, or even before, as defendants may move quickly for summary judgment in the hope that any market-based evidence that could support a dilution argument is not available to oppose it.
- Future risks and uncertainties persist. Output infringement remains untouched and future decisions focusing on LLM-generated content that mirrors training materials are more likely to find in favor of the copyright holders.
- Partner
Banu Naraghi is a Partner in the Litigation Department.
Banu’s practice focuses on corporate and intellectual property litigation in both state and federal court. She has represented a wide range of clients including content ...
- Partner
Jason Haas is a Partner in the Litigation Department.
Jason has spent over twenty years litigating a wide variety of complex business disputes for his clients in federal and state courts and in private arbitration. Jason focuses his ...
Subscribe
Recent Posts
- Employers May Now Obtain Equitable Relief for Untimely Arbitration Payments | By: Jared W. Slater
- California Appellate Courts Split on “Headless” PAGA Standing: CRST Expedited Permits Broad Claims, While Leeper Awaits Review | By: Jared W. Slater
- A New Era for Workplace Surveillance? | By: Jared W. Slater
- Assembly Bill 1018: A Bill for Employers to Keep an (A)Eye On | By: Jared W. Slater
- Surprise! Misleading Statements and Time Pressure May Render an Employer’s Arbitration Agreement Unenforceable | By: Jared W. Slater
- The Future of Copyright Enforcement: A Pivotal Supreme Court Case for the Digital Age | By: Banu Naraghi
- Contemporaneous Onboarding Documents May Invalidate An Employer’s “Squeaky Clean” Arbitration Agreement | By: Jared W. Slater
- A Temporary Victory: What the New Anthropic and Meta Rulings Actually Reveal About a Fair Use Defense for Companies Accused of Using Copyrighted Works to Train Generative AI | By: Jason L. Haas and Banu Naraghi
- Timbaland & Suno Highlight The Legal Risks of AI Music Tools | By: Banu Naraghi
- Receiver Selling Property Outside The Jurisdiction Of The Receivership Court | By: Peter A. Davidson
Blogs
Contributors
Archives
- August 2025
- July 2025
- June 2025
- May 2025
- April 2025
- March 2025
- February 2025
- January 2025
- December 2024
- November 2024
- October 2024
- September 2024
- August 2024
- July 2024
- June 2024
- May 2024
- April 2024
- March 2024
- February 2024
- January 2024
- December 2023
- November 2023
- October 2023
- September 2023
- August 2023
- July 2023
- June 2023
- May 2023
- April 2023
- March 2023
- February 2023
- January 2023
- December 2022
- November 2022
- October 2022
- September 2022
- August 2022
- July 2022
- June 2022
- May 2022
- April 2022
- March 2022
- February 2022
- January 2022
- December 2021
- November 2021
- October 2021
- September 2021
- August 2021
- July 2021
- June 2021
- May 2021
- April 2021
- March 2021
- February 2021
- January 2021
- December 2020
- November 2020
- October 2020
- September 2020
- August 2020
- July 2020
- June 2020
- May 2020
- April 2020
- March 2020
- February 2020
- January 2020
- December 2019
- November 2019
- October 2019
- September 2019
- August 2019
- July 2019
- June 2019
- May 2019
- March 2019
- February 2019
- January 2019
- November 2018
- October 2018
- September 2018
- August 2018
- July 2018
- June 2018
- May 2018
- April 2018
- February 2018
- January 2018
- December 2017
- November 2017
- October 2017
- September 2017
- August 2017
- July 2017
- June 2017
- May 2017
- April 2017
- March 2017
- February 2017
- January 2017
- December 2016
- November 2016
- October 2016
- September 2016
- August 2016
- July 2016
- June 2016
- May 2016
- April 2016
- January 2016
- December 2015
- November 2015
- October 2015
- September 2015
- August 2015
- July 2015
- June 2015
- May 2015
- April 2015
- March 2015
- February 2015
- January 2015
- December 2014
- November 2014
- October 2014
- September 2014
- August 2014
- July 2014
- June 2014
- May 2014
- April 2014
- March 2014
- February 2014
- January 2014