Five major publishers are suing Meta over Llama. They have evidence that the previous plaintiffs did not.
Five major publishers including Elsevier, Cengage, Hachette, Macmillan, and McGraw Hill have filed a class action lawsuit against Meta in Manhattan, accusing the company of using millions of their copyrighted works without permission to train its large language model, Llama. The lawsuit also includes author Scott Turow as a plaintiff. This legal action follows a June 2025 ruling by Judge Chhabria, which suggested that stronger evidence was needed to prove market harm caused by AI training practices. Plaintiffs like these publishers have now stepped forward with what they claim is clearer proof that Meta’s use of their content has a tangible negative effect on their businesses.
This lawsuit matters because it directly challenges how AI companies acquire and use training data. Large language models like Llama learn to generate text by analyzing huge amounts of written material. Publishers argue that using their copyrighted books without compensation harms their ability to sell that content, especially if AI-generated text competes with human-authored work. For businesses reliant on book sales and licensing, an unchecked use of content for AI training could threaten their revenue streams and undermine copyright law. For developers and the AI industry, the outcome may set important legal precedents concerning the boundaries of data scraping and fair use.
The conflict reflects a larger ongoing debate about intellectual property in the age of artificial intelligence. Many AI systems rely on web scraping or purchasing massive datasets to build their models. Publishers and creators have long raised concerns that their work is being exploited without proper credit or payment. The June 2025 ruling brought attention to the need for stronger evidence showing concrete market harm to these rights holders. Now, with this new lawsuit presenting clear market harm claims, we see a potential shift toward more rigorous legal scrutiny of AI training practices.
This lawsuit signals that the AI industry may face tighter restrictions on how training data is gathered, especially when it involves copyrighted material. Companies may need to negotiate licensing deals with content providers or risk expensive litigation. It may also encourage the development of alternative approaches to training AI models that respect copyright more carefully. People should watch for potential settlements or landmark court opinions that could reshape data sourcing norms. The case may influence not only Meta but other AI developers as well, enforcing a new standard that balances innovation with intellectual property rights.
— AI Quick Briefs Editorial Desk