The Week AI Broke the Music Video: Google Lyria 3, Seedance 2.0, and the New Creative Arms Race

@giacomo.mov ·

If you blinked last week, you missed what might be the most consequential seven days in AI music and video history. And honestly, even if you didn’t blink, you probably still can’t process it all.

On one side: Google launched Lyria 3, a new AI music generation tool powered by DeepMind that can produce 30-second tracks based on images, videos, and text descriptions — available to literally anyone with the Gemini app. On the other: ByteDance’s Seedance 2.0 sent shockwaves through Hollywood and the global entertainment industry, launching on February 12, 2026, and quickly becoming the center of a heated debate over copyright, creative disruption, and the future of filmmaking.

Meanwhile, AI video models are now so advanced that four of the six major models — Kling 3.0, Sora 2, Veo 3.1, and Seedance 1.5 Pro — generate synchronized audio natively.

The convergence of AI-generated music and AI-generated video is no longer a future prediction. It’s this week’s news. Let’s unpack all of it.

Google Lyria 3: AI Music Generation Goes Mainstream

Let’s start with the quieter (but arguably more important) revolution. Google announced that it’s adding a music-generation feature to the Gemini app, using DeepMind’s Lyria 3 music-generation model to power the feature, which is still in beta.

Here’s what makes this different from everything that came before: this isn’t a niche startup. This isn’t an invite-only beta from a company you’ve never heard of. This is Google, baking music creation directly into the app that hundreds of millions of people already use.

To use the feature, you describe the song you want to create, and the app generates a track along with lyrics. For instance, you could ask Gemini to create a “comical R&B slow jam about a sock finding its match,” and the app will generate a 30-second track along with cover art. You can even upload a photo or a video, and the AI-powered tool will create a song to match the mood of the media file.

That last part is huge for anyone thinking about music videos. Upload a video, get a matching soundtrack. We’ve gone from “AI can maybe hum a tune” to “AI will compose, arrange, perform, and master a track that matches your visual content” in roughly 18 months.

The 30-Second Strategy

Now, before you start panicking (or celebrating), there’s a catch. The length restriction isn’t about audio fidelity or computational cost. It’s about staying under the threshold where a 30-second clip becomes a “song” in copyright terms.

This is Google playing chess while everyone else plays checkers. Google developed Lyria starting in 2023, but it took three years to figure out how to launch without triggering lawsuits from every major label simultaneously. The answer: make it useful enough for creators, short enough to avoid replacing actual musicians.

Think about it: it’s perfect for YouTube intros, TikTok backgrounds, podcast bumpers, Instagram Reels — anything where you need audio texture, not a complete composition.

Dream Track Goes Global

The other massive piece of the Lyria 3 puzzle: along with rolling out Lyria 3 to the Gemini app, Google is making the model available to YouTube creators through the Dream Track feature on YouTube. The option was only available to YouTube creators in the U.S. until now, but with this release, Google is expanding Dream Track availability globally.

This matters because YouTube is where music videos live. Connecting AI music generation directly to the world’s largest video platform creates a pipeline that didn’t exist before. Need a soundtrack for your AI-generated music video? Google will make one. Need to share it? YouTube’s right there.

The Guardrails (And Their Limits)

Google is being conspicuously careful here. They say they’ve been “very mindful of copyright and partner agreements” in training Lyria 3, and that “music generation with Lyria 3 is designed for original expression, not for mimicking existing artists.”

All tracks generated in the Gemini app are embedded with SynthID, their imperceptible watermark for identifying Google AI-generated content. They’re also broadening verification capabilities to include audio — users can upload a file and ask if it was generated using Google AI.

Smart move. But as The Register not-so-gently pointed out, Lyria’s original iteration included training data from licensed artists like T-Pain, Demi Lovato, Sia, and other artists — and Google didn’t mention who it may have signed partnership agreements with to train Lyria 3.

The transparency gap is worth watching. But for now, Lyria 3 is live, it’s global, and millions of people are experimenting with it. Right now, millions of creators are generating their first AI tracks. They’re not thinking about copyright law. But Google is.

Seedance 2.0: When AI Video Meets Hollywood’s Worst Nightmare

Okay, now let’s talk about the other bomb that dropped this month. Because while Google was carefully tiptoeing around copyright law with 30-second music clips, ByteDance essentially kicked down Hollywood’s front door.

The film industry is reeling after the explosive debut of Seedance 2.0, an AI video generation model developed by Chinese tech giant ByteDance. Since its full release on February 12, 2026, Seedance 2.0 has ignited both awe and outrage, captivating millions with its ability to create photorealistic videos from a single photo and a few lines of text.

How photorealistic? Well…

One viral sensation — a 15-second clip depicting Tom Cruise and Brad Pitt brawling atop a rooftop — was crafted by Irish director Ruairi Robinson using just two lines of command. The video amassed over 1.6 million views on X.

Other viral videos included remixes of Avengers: Endgame, Optimus Prime taking on Godzilla, and a Friends scene in which Rachel and Joey are played by otters. (The otters one is, admittedly, incredible.)

Hollywood’s Response: Maximum Panic

The entertainment industry’s response was swift, furious, and — honestly — kind of terrified.

Rhett Reese, screenwriter for the Deadpool series, commented on Robinson’s video, saying, “I hate to say it, but we’re done.”

The Motion Picture Association issued a statement from CEO Charles Rivkin demanding that ByteDance “immediately cease its infringing activity,” saying “In a single day, the Chinese AI service Seedance 2.0 has engaged in unauthorized use of U.S. copyrighted works on a massive scale.” The Human Artistry Campaign condemned Seedance 2.0 as “an attack on every creator around the world.”

Seedance videos featured Disney-owned characters such as Spider-Man, Darth Vader, and Grogu, prompting Disney to send a cease-and-desist letter accusing ByteDance of a “virtual smash-and-grab of Disney’s IP.” Though notably, Disney has signed a three-year licensing deal with OpenAI — showing it’s not opposed to AI, just to AI it doesn’t control.

Paramount followed suit by sending ByteDance a cease-and-desist letter as well.

Netflix also sent cease-and-desist letters, describing ByteDance as a “high-speed piracy engine.”

The Privacy Nightmare Nobody Saw Coming

If the copyright issues weren’t enough, Seedance 2.0 revealed an even creepier capability. One Chinese tech blogger said it was so advanced that it was able to generate realistic audio of his voice based solely on an image of him, raising fears over deepfakes and privacy.

Read that again. From a photo of your face, it could clone your voice.

ByteDance rolled back that feature and introduced verification requirements for users who want to create digital avatars with their own images and audio. But the genie is out of the bottle. The technology exists. It will be replicated.

The Bigger Picture

The fight-scene deepfake of Tom Cruise and Brad Pitt is not merely a dazzling (and unlawful) trick; it is a clean, public demonstration of an epochal capability: convincingly produced, cinematic-quality filmed entertainment that can be generated without living actors or a conventional crew.

This has direct implications for music videos. If a two-line prompt can generate a cinematic fight scene, imagine what it can do for a music video concept. A dreamy slow-motion sequence for your ballad? A cyberpunk city chase for your EDM track? An abstract visual poem for your ambient piece? We’re talking about capabilities that would have required six-figure budgets just two years ago.

The State of AI Video: Everything Changed While You Were Sleeping

These two stories exist within a broader context that’s equally mind-bending. The AI video generation landscape in February 2026 is almost unrecognizable from even six months ago.

Kling 3.0, released February 4, 2026, is the most capability-dense video model currently available — not necessarily the best at any single dimension, but the model with the widest range of production-viable features in a single package. It offers native 4K at up to 60fps, with the 60fps option enabling slow-motion extraction.

Meanwhile, Veo 3.1 pushed photorealistic rendering to a level where trained observers have difficulty identifying generated output in controlled tests.

And over at Gemini, Google is upgrading video generation with templates, including styles like Civilization, Metallic, Cyberpunk, Video Game, Cosmos, Action Hero, Stardust, and more.

Let that sink in: you can now generate music with AI, generate video with AI, and the models are starting to create both simultaneously with synchronized audio. The building blocks for fully AI-generated music videos aren’t “coming soon.” They’re here, scattered across different tools, waiting to be connected.

The Music Industry’s Identity Crisis

All of this lands in an industry already mid-existential-crisis.

Suno is fresh off a $250 million Series C fundraise and a valuation of $2.45 billion, and according to an investor pitch deck obtained by Billboard, it generates a Spotify catalog’s worth of music every two weeks. Let that number hit you. An entire Spotify catalog’s worth. Every. Two. Weeks.

While Suno is still being sued by Universal Music Group and Sony Music for copyright infringement, it recently settled with WMG, showing signs that in 2026, it might be able to make amends with the music establishment.

Udio is carving a different lane, having settled disputes beginning with a deal with Universal in late 2025. As part of that deal, Udio vowed to pivot away from creating new songs with a simple prompt to become a fully-licensed music remixing and fan engagement platform.

Meanwhile, streaming platforms are fighting their own battles. Apple Music, Deezer, and Spotify have all sought to remove fraudulent AI “spam” tracks in recent months, and Bandcamp recently banned AI-generated music altogether.

The market numbers tell the story of an industry in transformation: the generative AI in music market, valued at $642.8 million in 2024, is projected to reach $3 billion by 2030, while the AI-generated video market is expected to grow by 35% annually, reaching $14.8 billion by 2030, with 54% of major artists already using AI visuals.

What This Actually Means for Musicians

Okay, let’s cut through the noise and get practical. If you’re a musician or creator reading this, here’s what matters:

1. The Music Video Is Being Democratized (For Real This Time)

We’ve heard “democratization” so many times it’s become meaningless. But this time it’s different. By lowering the cost of music video production from tens of thousands of pounds to roughly the cost of an API call, tools like these have the potential to empower independent creators.

The tech isn’t perfect yet. Dialogue lacks nuance, while visuals can possess a flattened, synthetic brightness that becomes glaring on a cinema screen. But for social media? For Reels, Shorts, and TikTok? It’s already good enough. And “good enough” is all most independent artists need to compete for attention.

2. The Responsible Path Is the Sustainable Path

The Seedance 2.0 debacle is a cautionary tale. Generating content using other people’s IP, likenesses, or copyrighted material isn’t just ethically questionable — it’s a legal minefield. SAG-AFTRA, representing over 160,000 performers, focused its objection on the deepfake aspect — celebrity likenesses being used without consent. SAG-AFTRA had spent years negotiating AI protections in labor contracts, and the Seedance 2.0 deepfakes undermined those efforts directly.

The artists and creators who will thrive are the ones using AI to amplify their own vision — not to copy someone else’s.

3. The Full Stack Is Almost Here

Think about what exists right now, in February 2026:

  • AI music generation: Suno, Udio, Lyria 3, ElevenLabs Eleven Music
  • AI video generation: Veo 3.1, Kling 3.0, Sora 2, Seedance 2.0
  • AI music video generation: Purpose-built tools that combine both

AutoMV, a new open-source system from Queen Mary University of London, works like a virtual film production team. It analyzes a song’s musical structure, beats, and time-aligned lyrics. Then, a set of specialized AI agents — taking on roles such as screenwriter, director, and editor — collaborate to plan scenes, maintain character identity, and generate images and video clips. A final quality-control “verifier” agent checks for coherence and consistency.

The pipeline from “I wrote a song” to “here’s the finished music video” is collapsing from months to minutes.

4. AI-Assisted Is the Sweet Spot

Here’s my actual take, and I’ll die on this hill: pure AI-generated music videos aren’t the future. AI-assisted music videos are.

The magic happens when a human artist — with taste, intention, emotional depth, and a story to tell — uses AI as a creative accelerator. You bring the vision. The AI brings the production capability. Together, you make something neither could alone.

As Google DeepMind themselves put it: “We believe AI should enhance human creativity, not replace it. We’re working directly with artists to gather feedback and shape guardrails that address their concerns.”

That philosophy is exactly right. The question is whether everyone in the industry will follow it.

Where Do We Go From Here?

If 2025 was the year that artificial intelligence came to the forefront both technologically and culturally, 2026 will be the year that it truly makes a monumental impact on the music business — whether through the songwriting process, the determination of legal frameworks around copyright, or establishing the role that AI-assisted artists will play in the industry.

We’re watching the music video — one of the most iconic art forms of the last 40 years — get reinvented in real time. The tools are more powerful than ever. The legal battles are more intense than ever. And the creative possibilities are more exciting than ever.

The musicians who lean in — who learn these tools, who experiment fearlessly, who use AI to bring their most ambitious visual ideas to life — will have a massive advantage. Not because they’re replacing human creativity, but because they’re multiplying it.

The ones who wait for permission, or wait for the dust to settle, or wait for someone to tell them it’s okay? They’ll be watching from the sidelines while the landscape reshapes around them.


Ready to turn your music into stunning visuals without the six-figure budget? OneMoreShot.ai lets you create professional music videos in minutes — no film crew, no editing suite, no waiting. Just your music, your vision, and AI that actually understands both. Give it a spin. Your next music video is one click away.