- By JeffkomStory Team
- Published on
Adobe Faces Class-Action Lawsuit Over Alleged Use of Pirated Books in AI Training
Adobe, like many major tech companies, has aggressively embraced artificial intelligence in recent years. From AI-powered design tools to generative media platforms, the company has positioned itself as a leader in creative AI. But that push may now come at a legal cost.
A newly proposed class-action lawsuit accuses Adobe of using pirated books to train one of its AI language models. And raising fresh concerns about how AI systems are built and whose work is being used behind the scenes.
What the Lawsuit Claims
According to the lawsuit, Adobe trained its SlimLM language model using copyrighted books, including Lyon’s own works.
SlimLM is described by Adobe as a lightweight language model designed for document assistance tasks, particularly on mobile devices. According to Adobe, the model was pre-trained using SlimPajama-627B, an open-source dataset released by AI chipmaker Cerebras in June 2023.
However, Lyon’s lawsuit argues that SlimPajama itself is derived from another dataset called RedPajama. Which allegedly includes a controversial collection of pirated books known as Books3.
The Problem With Books3 and RedPajama
Books3 is a massive dataset containing around 191,000 books. It has become a recurring point of legal conflict in the AI industry, as many authors claim their copyrighted works were included without consent, credit, or compensation.
According to the lawsuit, SlimPajama was created by copying and modifying RedPajama, which in turn includes Books3. Because of this chain, the lawsuit argues that Adobe indirectly used copyrighted material when training SlimLM.
“The SlimPajama dataset was created by copying and manipulating the RedPajama dataset,” the complaint states, “and therefore contains the Books3 dataset, including the copyrighted works of the Plaintiff and Class members.”
Adobe Is Not Alone
Adobe is far from the only company facing these accusations.
-
Apple was sued in September over claims that its Apple Intelligence models were trained on copyrighted material without permission.
-
Salesforce faced a similar lawsuit in October, also tied to the RedPajama dataset.
-
Anthropic, the company behind Claude AI, agreed to pay $1.5 billion to authors earlier this year to settle claims that it used pirated books for training.
These cases highlight a growing legal backlash against how generative AI models are trained.
A Bigger Issue for the AI Industry
At the heart of these lawsuits is a fundamental question:
Can AI companies legally train models on copyrighted material without permission?
AI systems require enormous amounts of data to function effectively. But as authors, artists, and publishers push back, courts are increasingly being asked to define the boundaries between innovation and intellectual property rights.
The Anthropic settlement was seen by many as a potential turning point—suggesting that using pirated or unauthorized content may no longer be legally or financially sustainable.
What Comes Next for Adobe and AI Companies
Adobe has not yet publicly resolved the claims, and the lawsuit is still in its early stages. But the case adds to mounting pressure on AI developers to be more transparent about training data and to establish clearer licensing practices.
As generative AI continues to expand, these legal battles may shape the future of how models are built—and who gets paid for the knowledge that powers them.
One thing is clear: the era of “train first, ask later” may be coming to an end.
Here are some related articles you may find interesting:
AI Inference Startup Modal Labs in Talks to Raise at $2.5B Valuation
Modal Labs, an AI inference infrastructure startup, is reportedly in discussions with venture capital...
Amazon May Launch AI Content Marketplace for Media Publishers
Amazon may soon launch a new content marketplace. This platform would allow media companies to sell their...
Waymo Begins Driverless Robotaxi Testing in Nashville Ahead of 2026 Launch
Waymo has officially removed human safety drivers from its autonomous test vehicles in Nashville, marking...
a16z Warns Founders: Don’t Chase Hype-Driven ARR, Build Durable Growth Instead
The AI startup boom has reignited a familiar Silicon Valley pattern: massive venture capital flowing...
Google’s Gemini App Crosses 750 Million Monthly Users as AI Adoption Accelerates
Google’s AI chatbot Gemini has reached a major milestone, surpassing 750 million monthly active users...
Y Combinator Allows Startups to Receive Seed Funding in Stablecoins
Y Combinator is taking a big leap towards incorporating blockchain into the way they fund startups. And...
Apple Acquires Israeli AI Startup Q.ai to Strengthen Audio and Hardware Intelligence
Apple is one step further along in the high-stakes AI game. Tech giants like Apple, Meta, and Google...
Where’s My State Tax Refund? How to Check Your Status and Avoid Delays
Waiting for a tax refund can be frustrating, especially when it’s unclear who’s responsible for issuing...
Trump Administration Loosens Nuclear Safety Rules, Accelerating Reactor Development
US nuclear energy is charging into a new era of rapid growth, but controversy is in tow. With nuclear...
Everything You Need to Know About Viral Personal AI Assistant Clawdbot (Now Moltbot)
The latest wave of AI innovation has produced an unexpected breakout star: a lobster-themed personal...
Popular Posts

AI Inference Startup Modal Labs in Talks to Raise at $2.5B Valuation
JeffkomStory Team
Modal Labs, an AI inference

Amazon May Launch AI Content Marketplace for Media Publishers
JeffkomStory Team
Amazon may soon launch a

Waymo Begins Driverless Robotaxi Testing in Nashville Ahead of 2026 Launch
JeffkomStory Team
Waymo has officially removed human

a16z Warns Founders: Don’t Chase Hype-Driven ARR, Build Durable Growth Instead
JeffkomStory Team
The AI startup boom has
Join Our Newsletter
Start your day with impactful startup stories and concise news! All delivered in a quick five-minute read in your inbox.