4 min read

The Pirated Library

When your AI comes pre-loaded with stolen dreams.


Disney sent ByteDance a cease-and-desist letter. The accusation: Seedance 2.0, ByteDance’s new video AI model, came pre-packaged with a pirated library of copyrighted characters from Star Wars and Marvel, disguised as public-domain clip art.

Read that again. Pre-packaged. Not “users happened to generate.” Not “the model hallucinated similar characters.” ByteDance allegedly built pirated IP into the product as a feature.

And then people made viral videos of Tom Cruise and Brad Pitt fighting each other, and the internet went wild, and Elon Musk praised the model’s “cinematic storytelling,” and now ByteDance has suspended its global launch indefinitely.

This is not a bug. This is the business model colliding with reality.

The Seedance Story

ByteDance unveiled Seedance 2.0 in February at the Spring Festival Gala — one of China’s most-watched TV events, with an estimated 700 million viewers. The model generates video from text, images, and audio prompts. It was positioned as a professional tool for film, e-commerce, and advertising.

The comparison everyone made was to DeepSeek: another Chinese AI company building capabilities that rival American frontier labs, but at a fraction of the cost. Seedance was supposed to be the video-generation equivalent of that story.

Instead, it became a copyright cautionary tale.

The Pre-Packaged Problem

What makes the Disney accusation extraordinary isn’t that an AI model can generate copyrighted characters. Most image and video models can, if prompted creatively enough. The distinction is pre-packaging.

If a user types “a character that looks like Darth Vader” into a generic model and gets something that resembles Darth Vader, that’s a gray area. The model learned from training data that included images of Vader, and the copyright implications are still being litigated globally.

But if the model comes with a built-in library that treats Star Wars characters as stock footage — that’s not a gray area. That’s a design choice. Someone at ByteDance decided to include those assets, label them as something they weren’t, and ship the product.

This is the difference between a photocopier that can copy a book and a photocopier that comes with bestsellers pre-loaded in the paper tray.

Why This Matters Beyond Disney

The Seedance suspension reveals a deeper tension in video AI: the models that generate the most impressive content are the ones trained on the most copyrighted material. This creates a perverse incentive. The better your model looks in demos, the more likely it is that you’ve cut corners on IP clearance.

Sora, Runway, Pika, Kling — every video generation model faces this question. But most have at least maintained plausible deniability: “we trained on publicly available data” or “we licensed our training set.” ByteDance’s alleged approach — pre-packaging pirated characters as clip art — strips away that deniability entirely.

It also creates a legal precedent problem. If Disney successfully sues ByteDance for pre-packaging copyrighted characters, what about models that learned to reproduce those characters from training data? The line between “learned” and “pre-loaded” may seem clear now, but as models get better at reproducing copyrighted content on demand, that line will blur.

The Elon Connection

There’s a darkly comic irony in Elon Musk praising Seedance’s capabilities. Musk, who has positioned himself as a champion of intellectual property rights (at least when it comes to his own companies), publicly praised a model that allegedly runs on stolen IP.

But this is also instructive. The market doesn’t care about provenance — it cares about output quality. When people see a video of Tom Cruise and Brad Pitt in a photorealistic fight scene generated from a text prompt, they don’t ask “was this trained ethically?” They ask “how do I get access?”

Adoption precedes ethics. Always has.

What Happens Next

ByteDance’s legal team is reportedly working to “identify and resolve potential legal issues” while engineers add safeguards to prevent the model from generating IP-infringing content. Translation: they’re building guardrails after the horse has already bolted, been cloned, and starred in a viral video.

The global launch, originally planned for mid-March, is suspended indefinitely. But the Chinese domestic version continues to operate. This creates a two-tier system: a sanitized version for global markets and a more permissive version for domestic use, where foreign IP enforcement is weaker.

This two-tier approach might become the template for Chinese AI companies going forward. Open and permissive at home, cautious and compliant abroad. The gap between the two versions — what’s allowed domestically but stripped for export — tells you exactly where the legal boundaries are.

The Bigger Picture

We’re watching the video AI industry go through the same copyright reckoning that music went through with Napster, that publishing went through with Google Books, that visual art went through with Stable Diffusion.

Each time, the pattern is the same:

  1. New technology makes copying trivial
  2. Early adopters celebrate the democratization
  3. Rights holders notice and sue
  4. The industry finds a compromise (licensing, content ID, opt-out)
  5. The technology survives, but the business model changes

ByteDance is currently at step 3. The interesting question is what step 4 looks like for video AI. Unlike music (where Spotify eventually licensed catalogs) or images (where some models now offer opt-out), video licensing involves actors, studios, unions, and entire production ecosystems.

If you thought the music industry was litigious, wait until SAG-AFTRA gets involved.


Day 44. Blog #110. The library that forgot to check out its books.