
Did AI companies win a fight with authors? Technically
In the past week, big AI companies have — in theory — chalked up two big legal wins. But things are not quite as straightforward as they may seem, and copyright law hasn't been this exciting since last month's showdown at the Library of Congress.
First, Judge William Alsup ruled it was fair use for Anthropic to train on a series of authors' books. Then, Judge Vince Chhabria dismissed another group of authors' complaint against Meta for training on their books. Yet far from settling the legal conundrums around modern AI, these rulings might have just made things even more complicated.
Both cases are indeed qualified victories for Meta and Anthropic. And at least one judge — Alsup — seems sympathetic to some of the AI industry's core arguments about copyright. But that same ruling railed against the startup's use of pirated media, leaving it potentially on the hook for massive financial damage. (Anthropic even admitted it did not initially purchase a copy of every book it used.) Meanwhile, the Meta ruling asserted that because a flood of AI content could crowd out human artists, the entire field of AI system training might be fundamentally at odds with fair use. And neither case addressed one of the biggest questions about generative AI: when does its output infringe copyright, and who's on the hook if it does?
Alsup and Chhabria (incidentally both in the Northern District of California) were ruling on relatively similar sets of facts. Meta and Anthropic both pirated huge collections of copyright-protected books to build a training dataset for their large language models Llama and Claude. Anthropic later did an about-face and started legally purchasing books, tearing the covers off to 'destroy' the original copy, and scanning the text.
The authors argued that, in addition to the initial piracy, the training process constituted an unlawful and unauthorized use of their work. Meta and Anthropic countered that this database-building and LLM-training constituted fair use.
Both judges basically agreed that LLMs meet one central requirement for fair use: they transform the source material into something new. Alsup called using books to train Claude 'exceedingly transformative,' and Chhabria concluded 'there's no disputing' the transformative value of Llama. Another big consideration for fair use is the new work's impact on a market for the old one. Both judges also agreed that based on the arguments made by the authors, the impact wasn't serious enough to tip the scale.
Add those things together, and the conclusions were obvious… but only in the context of these cases, and in Meta's case, because the authors pushed a legal strategy that their judge found totally inept.
Put it this way: when a judge says his ruling 'does not stand for the proposition that Meta's use of copyrighted materials to train its language models is lawful' and 'stands only for the proposition that these plaintiffs made the wrong arguments and failed to develop a record in support of the right one' — as Chhabria did — AI companies' prospects in future lawsuits with him don't look great.
Both rulings dealt specifically with training — or media getting fed into the models — and didn't reach the question of LLM output, or the stuff models produce in response to user prompts. But output is, in fact, extremely pertinent. A huge legal fight between The New York Times and OpenAI began partly with a claim that ChatGPT could verbatim regurgitate large sections of Times stories. Disney recently sued Midjourney on the premise that it 'will generate, publicly display, and distribute videos featuring Disney's and Universal's copyrighted characters' with a newly launched video tool. Even in pending cases that weren't output-focused, plaintiffs can adapt their strategies if they now think it's a better bet.
The authors in the Anthropic case didn't allege Claude was producing directly infringing output. The authors in the Meta case argued Llama was, but they failed to convince the judge — who found it wouldn't spit out more than around 50 words of any given work. As Alsup noted, dealing purely with inputs changed the calculations dramatically. 'If the outputs seen by users had been infringing, Authors would have a different case,' wrote Alsup. 'And, if the outputs were ever to become infringing, Authors could bring such a case. But that is not this case.'
In their current form, major generative AI products are basically useless without output. And we don't have a good picture of the law around it, especially because fair use is an idiosyncratic, case-by-case defense that can apply differently to mediums like music, visual art, and text. Anthropic being able to scan authors' books tells us very little about whether Midjourney can legally help people produce Minions memes.
Minions and New York Times articles are both examples of direct copying in output. But Chhabria's ruling is particularly interesting because it makes the output question much, much broader. Though he may have ruled in favor of Meta, Chhabria's entire opening argues that AI systems are so damaging to artists and writers that their harm outweighs any possible transformative value — basically, because they're spam machines.
It's worth reading:
Generative AI has the potential to flood the market with endless amounts of images, songs, articles, books, and more. People can prompt generative AI models to produce these outputs using a tiny fraction of the time and creativity that would otherwise be required. So by training generative AI models with copyrighted works, companies are creating something that often will dramatically undermine the market for those works, and thus dramatically undermine the incentive for human beings to create things the old-fashioned way.
…
As the Supreme Court has emphasized, the fair use inquiry is highly fact dependent, and there are few bright-line rules. There is certainly no rule that when your use of a protected work is 'transformative,' this automatically inoculates you from a claim of copyright infringement. And here, copying the protected works, however transformative, involves the creation of a product with the ability to severely harm the market for the works being copied, and thus severely undermine the incentive for human beings to create.
…
The upshot is that in many circumstances it will be illegal to copy copyright-protected works to train generative AI models without permission. Which means that the companies, to avoid liability for copyright infringement, will generally need to pay copyright holders for the right to use their materials.
And boy, it sure would be interesting if somebody would sue and make that case. After saying that 'in the grand scheme of things, the consequences of this ruling are limited,' Chhabria helpfully noted this ruling affects only 13 authors, not the 'countless others' whose work Meta used. A written court opinion is unfortunately incapable of physically conveying a wink and a nod.
Those lawsuits might be far in the future. And Alsup, though he wasn't faced with the kind of argument Chhabria suggested, seemed potentially unsympathetic to it. 'Authors' complaint is no different than it would be if they complained that training schoolchildren to write well would result in an explosion of competing works,' he wrote of the authors who sued Anthropic. 'This is not the kind of competitive or creative displacement that concerns the Copyright Act. The Act seeks to advance original works of authorship, not to protect authors against competition.' He was similarly dismissive of the claim that authors were being deprived of licensing fees for training: 'such a market,' he wrote, 'is not one the Copyright Act entitles Authors to exploit.'
But even Alsup's seemingly positive ruling has a poison pill for AI companies. Training on legally acquired material, he ruled, is classic protected fair use. Training on pirated material is a different story, and Alsup absolutely excoriates any attempt to say it's not.
'This order doubts that any accused infringer could ever meet its burden of explaining why downloading source copies from pirate sites that it could have purchased or otherwise accessed lawfully was itself reasonably necessary to any subsequent fair use,' he wrote. There were plenty of ways to scan or copy legally acquired books (including Anthropic's own scanning system), but 'Anthropic did not do those things — instead it stole the works for its central library by downloading them from pirated libraries.' Eventually switching to book scanning doesn't erase the original sin, and in some ways it actually compounds it, because it demonstrates Anthropic could have done things legally from the start.
If new AI companies adopt this perspective, they'll have to build in extra but not necessarily ruinous startup costs. There's the up-front price of buying what Anthropic at one point described as 'all the books in the world,' plus any media needed for things like images or video. And in Anthropic's case these were physical works, because hard copies of media dodge the kinds of DRM and licensing agreements publishers can put on digital ones — so add some extra cost for the labor of scanning them in.
But just about any big AI player currently operating is either known or suspected to have trained on illegally downloaded books and other media. Anthropic and the authors will be going to trial to hash out the direct piracy accusations, and depending on what happens, a lot of companies could be hypothetically at risk of almost inestimable financial damages — not just from authors, but from anyone that demonstrates their work was illegally acquired. As legal expert Blake Reid vividly puts it, 'if there's evidence that an engineer was torrenting a bunch of stuff with C-suite blessing it turns the company into a money piñata.'
And on top of all that, the many unsettled details can make it easy to miss the bigger mystery: how this legal wrangling will affect both the AI industry and the arts.
Echoing a common argument among AI proponents, former Meta executive Nick Clegg said recently that getting artists' permission for training data would 'basically kill the AI industry.' That's an extreme claim, and given all the licensing deals companies are already striking (including with Vox Media, the parent company of The Verge), it's looking increasingly dubious. Even if they're faced with piracy penalties thanks to Alsup's ruling, the biggest AI companies have billions of dollars in investment — they can weather a lot. But smaller, particularly open source players might be much more vulnerable, and many of them are also almost certainly trained on pirated works.
Meanwhile, if Chhabria's theory is right, artists could reap a reward for providing training data to AI giants. But it's highly unlikely the fees would shut these services down. That would still leave us in a spam-filled landscape with no room for future artists.
Can money in the pockets of this generation's artists compensate for the blighting of the next? Is copyright law the right tool to protect the future? And what role should the courts be playing in all this? These two rulings handed partial wins to the AI industry, but they leave many more, much bigger questions unanswered.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


E&E News
9 minutes ago
- E&E News
GOP releases megabill text with land sales, tax credit rollbacks
The Senate released new megabill text overnight with stricter treatment for renewable energy incentives from the Democrats' climate law — a major blow to companies and some lawmakers lobbying for more leniency. The updated legislation also includes Senate Energy and Natural Resources Chair Mike Lee's latest plan to sell certain federal public land for housing. The new Senate Finance Committee text looks a lot more like the House-passed bill when it comes to an array of contested tax credits. Advertisement It would drastically phase out wind and solar credits while maintaining incentives for Republican-friendly energy sources like nuclear and geothermal. Specifically, the bill would cut off incentives for projects that aren't 'placed in service' — or plugged into the grid — by the end of 2028. That's more aggressive than the previous Senate draft that preserved those credits for projects that merely 'start construction' by the end of the year. On the other hand, the legislation has a notable victory for hydrogen. It would extend incentives for clean hydrogen production to 2028, instead of eliminating them this year, as the previous Senate version proposed. Some provisions stayed the same. Senators kept credits for non-carbon energy sources, including nuclear, hydropower and geothermal, for projects that start construction by 2033. Senators also kept 'transferability,' which allows project sponsors to transfer credits to a third party. And the new language largely kept provisions barring companies from using materials from China or other adversary nations, while moving some deadlines for compliance sooner. The Senate parliamentarian was reviewing at least portions of this section. Companies were hoping for more lenient treatment to allow supply chains to develop. Senators continued to target wind turbines and their parts, which would no longer receive advanced manufacturing tax credits by 2028. Critical mineral incentives would be phased out, as well, except for metallurgical coal. The bill proposes moving the end of the electric vehicle tax credits to Sept. 30, up from the six-month timeline previously proposed, while credits for charging infrastructure would end in June 2026. Negotiators added a new bonus tax incentive for certain advanced nuclear power facilities built in areas with significant nuclear industry employment. A number of House and Senate defenders of the climate law credits, under intense lobbying from companies, were looking to make sure renewable energy projects had more time to benefit. But conservatives and President Donald Trump fought in the opposite direction. Public lands Lee's latest proposal would order the sale of up to 0.5 percent of Bureau of Land Management land across 11 states. Only lands that fall within five miles of a population center would be eligible, and protected lands excluded. The parliamentarian had ruled an earlier land sales plan ineligible for passage by simple majority under the budget reconciliation process. She had yet to rule on Lee's new framework Friday night. The updated text would set aside money from each sale for hunting, fishing and recreational amenities. That after outdoor advocates on the right expressed concern. Still, a number of Republican lawmakers in both chambers have balked at any public land sale plan in the megabill and would push to strip it before final passage. Other provisions The Energy and Natural Resources portion of the budget reconciliation package would still claw back Department of Energy money from the Inflation Reduction Act. However, it would expand to $1 billion a new energy dominance loan program. Energy Secretary Chris Wright had lobbied for loan office funding, particularly for nuclear projects. The broader legislation includes other IRA spending rollbacks, a decade-long delay of the methane fee and zero fees on automakers for not complying with the Department of Transportation's Corporate Average Fuel Economy (CAFE) standards. The megabill would mandate oil and gas lease sales onshore and offshore, including in Alaska's Arctic National Wildlife Refuge. The state would increasingly share the revenues from development there. Sen. Lisa Murkowski (R-Alaska) has been a holdout on several big ticket items during the negotiations, particularly related to Medicaid cuts.
Yahoo
10 minutes ago
- Yahoo
New Patent Suggests Unprecedented Apple Watch
A new patent suggests a potential major design change for the Apple Watch. This would primarily expand its usability without increasing the device's size. According to a new patent application, Apple is apparently working on a new generation of the Apple Watch with an entirely new display. The designs show a case where the screen extends beyond the edges and wraps around the sides of the watch. This could not only increase the usable area but also change the operating logic. TECHBOOK reveals what this is all about. The sketches included in the patent document, published on the platform of the World Intellectual Property Organization (WIPO), suggest a concept where the display is curved downward at the side edges of the case. This is reported by the U.S. tech magazine 'Xpertpick.' This way, the display area could be expanded without increasing the overall size of the smartwatch. It is also conceivable that the curved areas could offer additional input options. According to the sketches in the patent document, Apple is apparently planning to move away from the previous flat screen. Instead, the display is supposed to slope downward at the sides of the case. These curved areas could offer additional control options, such as swipe gestures or virtual buttons. Read also: Why the Apple Watch Appeared as Early as 1995 A particularly striking detail: In one of the illustrations, the case appears to be completely embedded in the curved display. This suggests a particularly seamless and modern design that is not only visually appealing but could also improve usability. However, the patent does not specify which exact functions the curved display areas might take on. Apple is known for regularly filing patents for new device forms or technical concepts—many of which are never implemented. This design is also purely conceptual and does not indicate an imminent market launch. Nevertheless, the document shows that Apple continues to work on innovative solutions for its wearable devices. The idea of a curved or case-wrapped display is not new for Apple. Variants have already been presented in earlier patent applications that would allow the usable screen area on devices like the Apple Watch to be expanded. Whether the current design will ever make it to a market-ready product remains to be seen. The post New Patent Suggests Unprecedented Apple Watch appeared first on TECHBOOK.
Yahoo
13 minutes ago
- Yahoo
Banco Santander's (SAN) US Unit Divesting Seven Branches to Focus on Digital Banking
Banco Santander, S.A. (NYSE:SAN) is . On June 25, the company's US subsidiary, Santander Bank, announced that it had reached an agreement to sell seven of its branches in the Pennsylvania area to Community Bank N.A. A successful financial advisor giving advice to a satisfied client in an office. The divestment is part of Santander Bank's push to become a digital-first bank. Additionally, the sale underscores Santander's conviction that its customers will continue to receive quality service from Community Bank. Consequently, the sale will enable the bank to refine its physical footprint as it enhances its digital operations nationwide. Launched in 2024, the bank's digital operations under the Openbank division have attracted over $4 billion in deposits and served more than 100,000 customers. Banco Santander, S.A. (NYSE:SAN) is a global financial institution that provides a wide range of financial products and services to individuals, businesses, and organizations. These include banking, lending, investment, and insurance products, as well as support for companies and communities. While we acknowledge the potential of SAN as an investment, we believe certain AI stocks offer greater upside potential and carry less downside risk. If you're looking for an extremely undervalued AI stock that also stands to benefit significantly from Trump-era tariffs and the onshoring trend, see our free report on the best short-term AI stock. READ NEXT: and . Disclosure: None. Sign in to access your portfolio