Article image
SEIKOURI Inc.

The Lyrics War

Why the Anthropic case could decide whether AI training counts as fair use

Markus Brinsa 27 Apr 29, 2026 8 8 min read Download Web Insights Edgefiles™ seikou.AI™

Sources

The fight moved upstream

For a while, the public version of the AI copyright debate stayed fixated on outputs. Can an AI-generated image be copyrighted. Can a generated article be owned. How much prompting or editing does a human need before the law starts seeing authorship again.

Those questions were always real, but they were also increasingly downstream of the larger conflict. The more consequential battle was never just about what came out of the model. It was about what went into it, how it got there, and whether the companies building these systems had the right to treat the cultural archive as raw material by default.

That is why Anthropic’s latest move in its lawsuit with major music publishers matters. The company is asking a federal court to rule that using copyrighted song lyrics to train Claude was fair use. The publishers want the court to reject that defense outright. On the surface, it is a dispute over lyrics. In reality, it is one of the clearest tests yet of whether AI companies can keep describing mass ingestion as transformative innovation while sidestepping the permission question.

This is not a narrow artist-rights skirmish. It goes to the legal and commercial foundation of model development.

Music is the hard case for AI

Books gave AI companies some room to argue abstraction. They could say the model was not functioning as a digital bookshelf, that it was extracting statistical relationships, that training was remote from conventional consumption, and that the end product was a new kind of general-purpose system rather than a substitute for any one title.

Music lyrics are harder terrain.

Lyrics are not sprawling databases of prose. They are compact, highly expressive works that sit close to the center of copyright protection. They are also commercially organized. There are established licensing structures, established enforcement behavior, and an industry that already knows how to monetize fragments of language. That makes it much harder to pretend the relevant market is vague, immature, or hypothetical.

Anthropic’s fair-use theory depends on the familiar claim that training is transformative because the model does not republish the work for its original expressive purpose. It uses copied text to learn language patterns and build a new system. That is the argument. But the publishers’ response cuts straight at the weak point. If the system can reproduce lyrics on demand, generate imitations that compete with licensed uses, and develop those capabilities from unlicensed ingestion in a market where licenses already exist, the “transformative” story starts looking incomplete.

That is why this case matters so much. It tests whether courts will keep focusing on training as a technical process in the abstract, or whether they will start treating training as a commercial act that sits inside real content markets with real rights holders and real lost bargaining power.

Anthropic wants a clean fair-use ruling

Anthropic is not just defending itself against infringement allegations. It is asking for a court win broad enough to establish that lyric training itself falls under fair use. That is an ambitious move because it tries to settle the central doctrinal question before a full trial.

The company’s position is strategically obvious. If it can persuade the court that copying lyrics into a training pipeline is transformative as a matter of law, it gains much more than a case victory. It gains a precedent that other AI firms can cite when they defend the ingestion of protected expressive works at scale.

That would be a meaningful prize, especially after the company already benefited from the books decision that treated training on lawfully acquired books as transformative. Anthropic wants to extend that logic into a far less forgiving category of content. It wants the court to say, in effect, that the legal frame does not change just because the material is lyrics.

But that extension is exactly what makes the motion so important. Music is not just another input class. It is a better stress test for the limits of the doctrine.

The publishers are attacking the fourth factor

The most important part of this dispute is not the usual rhetorical clash between innovation and protection. It is the market question.

Fair use lives or dies on specific facts, and increasingly, the facts that matter most are about substitution, market harm, and available licensing pathways. That is where the publishers are aiming. Their argument is that Claude does not merely learn from lyrics in some distant, academic sense. It can generate outputs that reproduce or derive from those works, compete with them, and dilute the market around them.

That matters because once the court starts looking seriously at market effects, the case becomes more dangerous for AI developers. The comforting narrative that models only “learn patterns” begins to collide with a more commercial reality. What matters is not only what happened inside the training process. What matters is what the resulting system can do in the marketplace, and whether that capability was built by bypassing a licensing regime that already exists.

This is where a lot of shallow commentary on AI copyright still goes wrong. It treats fair use as though the transformative label settles everything. It does not. It merely starts the argument. Once the court sees a functioning market, expressive source material, and evidence that outputs can move close enough to the originals to matter economically, the analysis gets much less comfortable for the model builder.

That is the real tension inside this lawsuit. Anthropic wants the court to look at training as machine learning. The publishers want the court to look at it as industrial-scale uncompensated use in a licensable market.

The books cases did not end this fight

One of the laziest readings of recent copyright litigation is that AI companies already won the central battle because courts found some training uses transformative. That is too broad, and this case is one reason why.

The books rulings helped AI developers, but they did not hand out blanket immunity. They suggested that courts may accept fair use where the use is genuinely transformative and the sourcing story is comparatively cleaner. They also left obvious room for harder cases involving piracy, stronger market-harm records, or content categories where the distance between training input and commercial output is less easy to romanticize.

Music lyrics fit that harder profile.

They are short enough to be memorized, recognizable enough to be detected, valuable enough to be licensed, and expressive enough to resist the argument that they are just generic ingredients in a broader linguistic soup. The case also sits in a context where the court has already had to deal with output guardrails and allegations that users could obtain protected lyrics from Claude. That weakens any effort to keep training and output neatly separated in practice.

In other words, this is not a repeat of the books fight in a different costume. It is the next pressure point. It asks whether the reasoning that helped AI companies with books can survive contact with a medium that is more commercially concentrated, more stylistically distinctive, and more obviously monetized through licensing.

Licensing is no longer the side story

For too long, licensing was treated as a secondary issue in AI copyright. Something optional. Something for cautious companies. Something to smooth relationships while the real war over fair use played out elsewhere.

That description no longer works.

Once a market for AI-related licenses begins to exist at visible scale, the legal environment changes. Licensing becomes evidence. It becomes harder for a defendant to say there was no practical market to enter, no channel to negotiate, no recognized economic interest to respect. The more developed those channels become, the less persuasive the “we had to take first and sort it out later” posture sounds.

That is especially true in music, where licensing is not a speculative future possibility. It is the industry’s operating system. AI companies may dislike the friction, cost, and transactional complexity that come with that fact, but courts do not have to rescue them from the existence of a functioning market.

This is one reason the Anthropic case matters beyond the legal doctrine itself. It is a business signal. It tells operators, investors, and acquirers that training-data rights are moving out of the category of abstract policy concern and into procurement, diligence, and deal structure.

That shift is bigger than one motion.

This is becoming an operating-model issue

A lot of AI firms still behave as if copyright exposure is mainly a litigation reserve problem. Budget for lawyers, keep the product moving, and hope doctrine develops in your favor before damages do.

That approach is starting to look dated.

If the court rejects Anthropic’s fair-use argument in the lyrics case, the practical consequence will not stop with one company or one medium. It will reinforce a broader market lesson that model quality cannot be separated from provenance quality forever. Where did the data come from. Was access lawful. Was copying authorized. Is there a licensing market. Can the company explain what it used and why it believed that use was defensible.

Those questions are no longer soft governance questions. They are operational questions.

Enterprise buyers will ask them because procurement risk increasingly includes provenance risk. Investors will ask them because unresolved data rights now affect future margins, settlement exposure, and exit value. Boards will ask them because “the model works” is no longer a complete answer when the legal basis of the training corpus is unstable.

This is the shift many AI companies resisted for as long as possible. They wanted data acquisition to be treated as a technical precondition, not as a rights-sensitive business function. That is getting harder to maintain.

What the court may really decide

The most important outcome here may not be a dramatic declaration that all AI training is lawful or unlawful. Courts usually move with more precision than that. What matters is how they draw the line.

If the court accepts Anthropic’s position, AI firms will argue that highly expressive content can still be copied for training without permission so long as the use is framed as building general-purpose capabilities. That would strengthen the industry’s preferred story and buy more room for unlicensed ingestion.

If the court sides with the publishers, the message will be very different. It will suggest that at least some categories of copyrighted material are too commercially and expressively specific to be absorbed into AI systems under a generic transformative-use theory, especially where licensing markets are real and outputs create plausible substitution pressure.

That would not end the broader AI copyright fight. But it would narrow the zone of comfort. It would make the next generation of cases less about grand innovation rhetoric and more about the facts companies would rather leave blurry.

And that is the point. The real copyright war is not about whether a chatbot output qualifies as authorship. It is about whether the industry had the right to ingest first and negotiate later.

Anthropic’s lyrics case puts that question where it belongs: at the legal center of the AI business model.

The deeper consequence

The first phase of generative AI rewarded scale, speed, and indifference to provenance. The second phase is starting to punish that indifference.

That is why this case matters. It is not just another rights-holder complaint. It is a test of whether courts will keep treating training as a mostly abstract computational act, or whether they will begin treating it as what it has always also been: a massive commercial appropriation problem with technical features.

If the answer moves toward the latter, the industry will have to adapt quickly. Not by issuing nicer policy statements, but by rebuilding parts of the stack around permissions, sourcing discipline, auditability, and licensing strategy.

The companies that understand this early will look more expensive in the short term and more durable in the long term. The ones that do not may discover that the real cost of AI was never compute alone.

About the Author

Markus Brinsa is the Founder & CEO of SEIKOURI Inc., an international strategy firm that gives enterprises and investors human-led access to pre-market AI—then converts first looks into rights and rollouts that scale. As an AI Risk & Governance Strategist, he created "Chatbots Behaving Badly," a platform and podcast that investigates AI’s failures, risks, and governance. With over 30 years of experience bridging technology, strategy, and cross-border growth in the U.S. and Europe, Markus partners with executives, investors, and founders to turn early signals into a durable advantage.

©2026 Copyright by Markus Brinsa | SEIKOURI Inc.