
Explained: AI & copyright law
While the two judgments arrived at their conclusions differently, they are the first to address a central question around generative AI models: are these built on stolen creative work?
At a very basic level, AI models such as ChatGPT and Gemini identify patterns from massive amounts of data. Their ability to generate passages, scenes, videos, and songs in response to prompts depends on the quality of the data they have been trained on. This training data has thus far come from a wide range of sources, from books and articles to images and sounds, and other material available on the Internet.
There are at the moment at least 21 ongoing lawsuits in the US, filed by writers, music labels, and news agencies, among others, against tech companies for training AI models on copyrighted work. This, the petitioners have argued, amounts to 'theft'.
In their defence, tech companies say they are using the data to create 'transformative' AI models, which falls within the ambit of 'fair use' — a concept in law that permits use of copyrighted material in limited capacities for larger public interests (for instance, quoting a paragraph from a book for a review).
Here's what happened in the two cases, and why the judgments matter.
In August 2024, journalist-writers Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson filed a class action complaint — a case that represents a large group that could be/were similarly harmed — against Anthropic, the company behind the Claude family of Large Language Models (LLMs).
The petitioners argued that Anthropic downloaded pirated versions of their works, made copies of them, and 'fed these pirated copies into its models'. They said that Anthropic has 'not compensated the authors', and 'compromised their ability to make a living as the LLMs allow anyone to generate — automatically and freely (or very cheaply) — texts that writers would otherwise be paid to create and sell'.
Anthropic downloaded and used Books3 — an online shadow library of pirated books with about seven million copies — to train its models. That said, it also spent millions of dollars to purchase millions of printed books and scanned them digitally to create a general 'research library' or 'generalised data area'.
Judge William Alsup of the District Court in the Northern District of California ruled on June 23 that Anthropic's use of copyrighted data was 'fair use', centering his arguments around the 'transformative' potential of AI.
Alsup wrote: 'Like any reader aspiring to be a writer, Anthropic's LLMs trained upon works not to race ahead and replicate or supplant them — but to turn a hard corner and create something different. If this training process reasonably required making copies within the LLM or otherwise, those copies were engaged in a transformative use.'
Thirteen published authors, including comedian Sarah Silverman and Ta-Nehisi Coates of Black Panther fame, filed a class action suit against Meta, arguing they were 'entitled to statutory damages, actual damages, restitution of profits, and other remedies provided by law'.
The thrust of their reasoning was similar to what the petitioners in the Anthropic case had argued: Meta's Llama LLMs 'copied' massive amounts of text, with its responses only being derived from the training dataset comprising the authors' work.
Meta too trained its models on data from Books3, as well as on two other shadow libraries — Anna's Archive and Libgen. However, Meta argued in court that it had 'post-trained' its models to prevent them from 'memorising' and 'outputting certain text from their training data, including copyrighted material'. Calling these efforts 'mitigations', Meta said it 'could get no model to generate more than 50 words and punctuation marks…' from the books of the authors that had sued it.
In a ruling given on June 25, Judge Vince Chhabria of the Northern District of California noted that the plaintiffs were unable to prove that Llama's works diluted their markets. Explaining market dilution in this context, he cited the example of biographies.
If an LLM were to use copyrighted biographies to train itself, it could, in theory, generate an endless number of biographies which would severely harm the market for biographies. But this does not seem to be the case thus far.
However, while Chabbria agreed with Alsup that AI is groundbreaking technology, he also said that tech companies who have minted billions of dollars because of the AI boom should figure out a way to compensate copyright holders.
Significance of rulings
These judgments are a win for Anthropic and Meta. That said, both companies are not entirely scot-free: they still face questions regarding the legality of downloading content from pirated databases.
Anthropic also faces another suit from music publishers who say Claude was trained on their copyrighted lyrics. And there are many more such cases in the pipeline.
Twelve separate copyright lawsuits filed by authors, newspapers, and other publishers — including one high-profile lawsuit filed by The New York Times — against OpenAI and Microsoft are now clubbed into a single case. OpenAI is also being separately sued by publishing giant Ziff Davis.
A group of visual artists are suing image generating tools Stability AI, Runway AI, Deviant Art, and Midjourney for training their tools on their work. Stability AI is also being sued by Getty Images for violating its copyright by taking more than 12 million of its photographs.
In 2024, news agency ANI filed a case against OpenAI for unlawfully using Indian copyrighted material to train its AI models. The Digital News Publishers Association (DNPA), along with some of its members, which include The Indian Express, Hindustan Times, and NDTV, later joined the proceedings. Going forward, this is likely to be a major issue in India too.
Thus, while significant, the judgments last week do not settle questions surrounding AI and copyright — far from it.
And as AI models keep getting better, and spit out more and more content, there is also the larger question at hand: where does AI leave creators, their livelihoods, and more importantly, creativity itself?
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Mint
14 minutes ago
- Mint
ChatGPT, Claude and Gemini not helping? Here's how to fix your prompts for better output
AI chatbots like OpenAI's ChatGPT, Google's Gemini, and Anthropic's Claude are increasingly woven into the fabric of daily life, assisting with everything from writing and coding to travel planning and customer service. But if you are not impressed with the results, the problem may not lie with the bot. Instead, it could be your prompt. According to an AP report, industry experts and the chatbot creators themselves agree: the key to unlocking accurate, creative and useful responses lies in how you communicate with these tools. Whether you are seeking help drafting a speech or planning a holiday, refining your input can make all the difference. Unlike a traditional web search, AI chatbots do not respond well to keyword-heavy or ambiguous instructions. 'Ensure your prompts are clear, specific, and provide enough context,' OpenAI advises. Rather than asking it to simply 'design a logo,' users should include the business name, industry, colour preferences and design style. The more detailed the request, the more personalised the response. Chatbot interactions are not one-and-done. Think of them more as a conversation. OpenAI recommends tweaking the wording, simplifying complex requests, or adding fresh context to achieve better output. Google echoes this in its Gemini prompting guide, noting that users should be ready to experiment with different phrasing. Iteration is key. If the first response falls flat, adjust and try again. You can shape the tone and voice of an AI's output by explicitly stating the style you want: professional, casual, humorous or formal. Chatbots can also adjust their language depending on the intended audience, whether it is schoolchildren or industry experts. This helps create responses that are not just accurate, but appropriately pitched. Generic inputs yield generic answers. If you ask for a 'weeklong itinerary in London,' the bot might list the city's usual tourist hotspots. But add context, such as your interests, budget, travel dates and group details, and the result becomes a far more tailored and relevant plan. As Google's guidance notes, examples and background information help the AI understand exactly what you need. If you are overwhelmed by long-winded responses, you can take control by setting limits. Ask the chatbot to respond in a specific format or with a word count cap, for example, 'Explain in under 150 words,' or 'Summarise in five bullet points.' This kind of constraint pushes the AI to be concise and focused, helping users absorb complex topics more easily.


Time of India
17 minutes ago
- Time of India
Bill Gates' daughter taps ChatGPT to taste startup success: Here's all you need to know
Academy Empower your mind, elevate your skills Tech billionaire Bill Gates' daughter's fashion tech startup Phia , has made headlines as the founder, Phoebe Gates , revealed her marketing strategy woven around artificial intelligence (AI) Gates said AI tools like ChatGPT have been a creative partner for her startup, Phia, to track market trends and viral content on social media platforms like Instagram and TikTok The comments came during her interaction with cofounder Sophia Kianni on The Burnouts podcast, which was started by the duo to invite guests who candidly talk about their careers and experiences as startup founder goes on to explain the marketing plan. Phia collects data of top-performing videos from Instagram and TikTok, records the data in Excel, and does an in-depth analysis of them, noting the minute details like lighting, narrative, tone, in 2022 by Phoebe Gates, the startup works as an AI assistant that helps you shop smartly by offering the best brand deals teamed up with her former Stanford roommate and climate activist Sophia Kianni to launch Phia to transform online startup claims that it tracks over 40,000 sites to find the best price for the customer and that the end-to-end personal shopping assistant has partnered with thousands of brands to make it possible. Among its standout features are the "Should I Buy This?" button and personalised shopping preferences, both aimed at making shopping simpler and more also allows users to create shoppable moodboards. Through this, users can gain insights into stock and price app's vast database includes over 250 million secondhand items from platforms such as The RealReal, Vestiaire Collective, and Poshmark. It operates as an iOS app and a Chrome browser extension, catering to users on mobile and has already attracted investments from prominent figures, including businessperson Kris Jenner and entrepreneur Desiree Gruber.


Time of India
an hour ago
- Time of India
'Dude clears interviews': Indian-origin entrepreneur says truth of Soham Parekh is simple
As Soham Parekh emerged as the 'Me Too' moment for US tech world, a US-based Indian-origin entrepreneur, Pritika Mehta, broke down the mystery and decided that the truth of Soham Parekh was pretty simple. Tired of too many ads? go ad free now Paresk cleared interviews as he was good with that, then he had junior people working for him like a small dev shop. The curious case of Soham Parekh broke the Internet after a US-based AI founder, Suhail Doshi, outed him on X claiming that Parekh has been moonlighting for years, which means he has been working for several software companies at the same time, and he is not even in the US. His resume went viral though his credentials are now disputed as he has been accused of fraud. 'Dog-eat-dog world out there' Another Indian-origin CEO Mihir Deo said Soham Parekh;s story did not surprise him at all. "That's just how it goes with a lot of the Indian engineering market, straight up. Some people are shocked by this, but if there's anyone who knows the Indian engineering market, the Soham Parekh stuff is like: "yeah, of course they do that"," he wrote. "I explored starting a business in India in the past. But the lack of trust that I had in people dissuaded me. People overpromise (and aren't afraid of overpromising) and often stab you in the back to get their lunch. It's a dog eat dog world out there. I decided against starting a company in India because I determined I wasn't thick skinned enough (e.g. also too rigorously honest as well) to play that game," Mihir Deo wrote. While US tech workers have been complaining about Indians eating up their jobs, the news of one Indian taking so many jobs in the US raised questions over the hiring and the management of these startups. Tech influencer Deedy Das suggested Parekh is 'just the tip of the iceberg,' pointing to a broader trend of 'overemployment' in remote work, with a Reddit community discussing similar cases.