The Recording Industry Association of America and record labels allege that the leading AI music generators trained on their artists’ work without permission.
The music industry has officially declared war on Suno and Udio, two of the most prominent AI music generators. A group of music labels including Universal Music Group, Warner Music Group, and Sony Music Group has filed lawsuits in US federal court on Monday morning alleging copyright infringement on a “massive scale.”
The plaintiffs seek damages up to $150,000 per work infringed. The lawsuit against Suno is filed in Massachusetts, while the case against Udio’s parent company Uncharted Inc. was filed in New York. Suno and Udio did not immediately respond to a request to comment.
“Unlicensed services like Suno and Udio that claim it’s ‘fair’ to copy an artist’s life’s work and exploit it for their own profit without consent or pay set back the promise of genuinely innovative AI for us all,” Recording Industry Association of America chair and CEO Mitch Glazier said in a press release.
I hate to say it, but I kinda hope the music copyright cartel wins this one, only for the precedent it would set about things like proprietary use of MS Copilot output being an infringement of GPL-licensed code.
GPL code is the least concern, you can always just say the AI-generated code is GPL. What about training on leaked proprietary code? The training data already known to include medical records, CSAM, etc., wouldn't be surprised if it also contained proprietary code.
Having all AI-generated code be either "viral" copyleft or illegal to use at all would certainly be better than allowing massive laundering of GPL-licensed code for exploitation in proprietary software.
That's the argument I would be making, but it certainly isn't Microsoft's (Copilot), OpenAI's (Codex), etc's position: they think the output is sufficiently laundered from the GPL training data so as not to constitute a derivative work (which means none of the original licenses -- "open source" or otherwise -- would apply, and the recipient could do whatever they want).
Edit: actually, to be more clear, I would take either of two positions:
That the presence of GPL (or in general, copyleft) code in the training dataset requires all output to be GPL (or in general, copyleft).
That the presence of both GPL code and code under incompatible licenses in the training dataset means that the AI output cannot legally be used at all.
(Position #2 seems more likely, as the license for proprietary code would be violated, too. It's just that I don't care about that; I only care about protecting the copyleft parts.)
i am vaguely familiar with software licensing is GPL type of open source?
You could say that, LOL. It's the OG of "copyleft" licenses (the guy that made it invented the concept), although "permissive" licenses (BSD, MIT) existed before.
"Copyleft" and "permissive" are the two major categories of Free Software (a.k.a. "open source", although that term has different connotations) license. The difference between them is that "copyleft" requires future modifications by people other than the copyright holders to be released under the same terms, while "permissive" does not. In other words, "copyleft" protects the freedom of future users to control their computer by being able to modify the software on it, while "permissive" maximizes the freedom of other developers to do whatever they want with the code (including using it in proprietary apps, to exploit people).