Model Diplomacy
Mirror, Mirror on the Wall!
In February 2026 Elon Musk posted on X that Anthropic had trained its models on unlicensed data at massive scale. He also referenced the company’s multi-billion-dollar settlements. The statement came right after Anthropic accused Chinese AI firms DeepSeek, Moonshot, and MiniMax of using more than 24,000 fake accounts. Those accounts generated millions of interactions with Claude. The goal was to distill its advanced reasoning, coding, and tool-use capabilities into their own systems.
Distillation enables a less powerful model to replicate the behavior of a stronger one by learning directly from its outputs. The method is rapid and cost-effective. When carried out without authorization it breaches Anthropic’s terms of service. According to the company it also constitutes a genuine national security threat.
Anthropic has never made Claude available in China. It continues to block access from entities linked to Chinese organizations. It strongly endorses stringent U.S. export controls on cutting-edge semiconductors. These controls aim to restrict both the scale of training and the ease of unauthorized replication.
Critics were swift to draw attention to the striking parallel. Virtually every leading foundation model was built on enormous datasets scraped from the public internet. This includes OpenAI’s GPT series, Google’s Gemini family, Meta’s Llama lineup, and xAI’s Grok. Most were collected without explicit permission from the copyright holders.
Although Musk acknowledged the similarity he nonetheless described Anthropic as super smug, sanctimonious, and hypocritical. He later forecasted the company’s downfall with the biting remark “Anthropic will be Misanthropic.” He framed the entire exchange as part of the fierce contest for lucrative defense and government contracts.
In August 2025 Anthropic reached a $1.5 billion settlement in Bartz v. Anthropic. It is one of the largest copyright resolutions ever recorded in the United States. Authors sued over the use of approximately 500,000 pirated books sourced from shadow libraries such as LibGen and PiLiMi.
A federal judge had previously determined that training on legally acquired copies could qualify as fair use because the transformation of the material into model weights serves a different purpose. The court found that deliberately retaining and storing clearly infringing copies crossed into impermissible territory. Rather than proceed to trial Anthropic agreed to destroy the offending dataset. It relinquished any claim to future licensing rights for that material. It also agreed to distribute roughly $3,000 per qualifying work with proceeds shared between authors and publishers.
Everyone is Toast!
This pattern of aggressive data acquisition has been widespread across the industry. Laboratories competed fiercely in the early days of large language models to assemble the most comprehensive training material possible. They often valued velocity and capability over immediate legal clarity.
Now that possession of frontier-level performance confers substantial geopolitical and economic advantage the very companies that once scraped with impunity are fortifying their positions. Others seek to follow the same path. The development has intensified pressure for structured licensing markets, collective rights-management systems, and mandatory disclosure of training datasets.
U.S. export controls classify advanced AI hardware as a critical strategic asset intended to constrain China’s ability to achieve comparable scale. Yet when Chinese developers employ distillation techniques to circumvent some of those hardware limitations American laboratories voice strong objections.
Many of their own foundational models were constructed using data whose copyright status was at the time of collection frequently ambiguous or unresolved. Observers have repeatedly observed that the intellectual-property rationale underpinning these export restrictions remains legally and practically unsettled within the domestic United States itself.
Only the Trial Lawyers will Survive AI.
Several plausible trajectories are now emerging for the future of this ecosystem. Licensing agreements between publishers, news organizations, stock-photography libraries, and AI developers could gradually become the norm. Training data would effectively transform into a regulated commodity market characterized by recurring royalties and negotiated terms.
High-profile settlements may evolve into de facto recurring industry fees that elevate barriers to entry for new competitors. Courts continue to refine the boundaries of fair use by distinguishing between broad general-purpose training and the generation of outputs that compete directly with original works such as images created in the distinctive style of a named artist.
Artists, writers, and other content owners have supplied a substantial portion of the raw material that powers the current AI boom. They frequently receive no direct compensation. The same technologies also broaden audience reach, lower barriers to creation, and generate novel revenue streams for those who choose to license their works for training purposes.
This fundamental tension continues to fuel growing calls for more robust opt-out mechanisms, standardized watermarking protocols, enforceable attribution rights, and strengthened protections. The movement is particularly pronounced in creator-friendly regulatory environments such as the European Union.
Lawyers occupy a pivotal position in this unfolding landscape. They draft sophisticated licensing agreements, litigate precedent-setting cases, conduct risk assessments of training datasets, and shape the emerging legislative frameworks that will govern the field for decades.
The sheer volume and high financial stakes of AI-related disputes guarantee sustained demand for their services. Some defend technology companies on the grounds of transformative use. Others advocate on behalf of creators seeking fair compensation.
The present clashes reveal an industry that has not yet found stable equilibrium. They occur between American and Chinese laboratories, between technology firms and content creators, and between the drive for rapid advancement and the assertion of ownership rights. The very dynamics that render AI so powerfully transformative are also what render it persistently contentious.
Achieving lasting balance will demand sustained negotiation, the development of new industry norms, and likely a certain amount of pragmatic inconsistency along the way. The conversation is far from over. Oh yeah… and for that snarky expression, “learn to code”… that’s toast, too. AI is coming for everyone.