In case you missed it, Spotify has apparently been coaching its personal music AI that ought to permit them to seize a number of the AI hype on Wall Road. Nevertheless it brings again appear unhealthy recollections.
There was a time when the music enterprise had a easy rule: “We are going to by no means let one other MTV construct a enterprise on our backs”. That philosophy arose from watching the arbitrage as worth created by artists was extracted by platforms that had nothing to do with creating it. That spectacle formed the business’s deep reluctance to license digital music within the early years of the web. “By no means” was speculated to imply by no means.
I took them at their phrase.
However in fact, “by no means” turned out to be conditional. The business made exception after exception till the rule dissolved completely. First got here the absurd statutory shortcut of the DMCA protected harbor period. Then YouTube. Then iTunes. Then Spotify. Then Twitter and Fb, social media. Then TikTok. Every time, platforms have been allowed to scale first and renegotiate later (and Twitter nonetheless hasn’t paid). Every time, the value of admission for the platform was astonishingly low in comparison with the worth extracted from music and musicians. In lots of instances, astonishingly low in comparison with their present market worth in companies which are completely depending on creatives. (You may most likely have put Amazon in that class.)
A few of these offers got here wrapped in what regarded, on the time, like significant compensation — headline-grabbing advances and what have been described as “fairness participation.” In actuality, these advances have been finite and the fairness was usually a skinny sliver, whereas the long-term impact was to commoditize artist royalties and shift sturdy worth towards the platforms. That’s one motive so many artists got here to resent and in lots of instances overtly despise Spotify and the “huge pool” mannequin. All of the whereas being instructed how transformative Spotify’s algorithm is with out explaining how the fantastic algorithm misses 80% of the music on the platform.
And now we arrive on the newest collapse of “by no means”: Spotify’s announcement that it’s creating its personal music AI and derivative-generation instruments.
If you happen to disliked Spotify earlier than, you could detest what comes subsequent.
This second is completely different — however in some ways it’s the similar basic drawback MTV created. Artists and labels supplied the core asset — their recordings — totally free or practically free, and the platform constructed a robust enterprise by packaging that worth and promoting it again to them. Distribution monetized entry to music; AI monetizes the music itself.
In response to Music Enterprise Worldwide:
Spotify’s framing seems to supply one thing of a center floor. [New CEO] Söderström shouldn’t be arguing for open distribution of AI derivatives throughout the web. As an alternative, he’s positioning Spotify because the platform the place this interplay ought to occur – the place the followers, the royalty pool, and the expertise exist already.
Proper, our followers and his pathetic “royalty pool.” And that is speculated to make us like you?
The Coaching Hole
Which brings us to the query Spotify has not answered — the query that issues greater than any characteristic announcement or product demo:
What did they practice on?
Was it Epidemic Sound? Was it licensed catalog? Public area recordings? Consumer uploads? Pirated materials?
All are equally potential.
However way more more likely to me: Did Spotify practice on the recordings licensed for streaming and Spotify’s personal platform person information derived from the followers we drove to their service — quietly amassed, normalized, and ingested into AI over years?
Spotify has not mentioned.
And that silence issues.
The Transparency Hole
Creators at present have no significant visibility into whether or not their work has already been absorbed into Spotify’s generative methods. No disclosure. No audit path. No licensing registry. No opt-in construction. No compensation framework. The unknowns should not theoretical — they’re structural:
- Had been your recordings used for coaching?
- Do your performances now exist inside mannequin weights?
- Was consent ever obtained?
- Was compensation ever contemplated?
- Can outputs reproduce protected expression derived out of your work?
If Spotify skilled on catalog licensed to them for a completely completely different goal with out express, knowledgeable permission from rights holders and performers, then AI derivatives should not merely a brand new characteristic. They’re a massively infringing second layer of worth extraction constructed on high of the primary exploitation — the unique recordings that creators already struggled to monetize pretty.
This isn’t innovation. It’s recursion.
Platform Knowledge: The Quiet Asset
Spotify possesses one of many largest behavioral and audio datasets within the historical past of recorded music that was licensed to them for a completely completely different goal — not simply recordings, however stems, utilization patterns, listener interactions, metadata, and efficiency analytics. If that corpus was used — formally or informally — as coaching enter for this Spotify AI instrument that magically appeared, then Spotify’s AI is constructed not simply on music, however on the amassed inventive labor of tens of millions of artists.
But creators have been by no means requested. No discover. No clarification. No disclosure.
It should even be mentioned that there’s a associated governance query. Daniel Ek’s funding within the defense-AI firm Helsing has been broadly reported, and Helsing’s methods like all superior AI rely on large-scale mannequin coaching, information pipelines, and machine studying infrastructure. Spotify supposedly has individually developed its personal AI capabilities.
This raises a slender however reputable transparency query: is there any technological, information, personnel, or infrastructure overlap — any “crosstalk” — between AI growth related to Helsing’s automated weapons and the fashions deployed inside Spotify? No public proof at present suggests such interplay, and the businesses function in several domains, however the absence of disclosure leaves creators and stakeholders unable to evaluate whether or not safeguards, firewalls, and governance boundaries exist. The place highly effective AI methods coexist beneath shared management affect, transparency about separation is as vital as transparency about coaching itself.
The core problem shouldn’t be merely licensing. It’s transparency. A platform can not convert custodial entry into coaching rights whereas declining to elucidate the place its coaching information got here from.
That’s why this quote from MBW belies the standard exceptionally brief sighted and moronic pablum from the Spotify government group:
Requested on the decision whether or not AI music platforms like Suno, Udio and Stability may themselves turn into DSPs and take share from Spotify, Norström pushed again: “No rightsholder is in opposition to our imaginative and prescient. We just about have the entire business behind us.”
In fact, the premise of the query is one I’ve been questioning about myself—I assume that Suno and Udio totally intend to get into the DSP recreation. However Spotify’s government blew proper previous that considerate query and answered a query he wasn’t requested which could be very related to us: “We have now just about the entire business behind us.”
Oh, effectively, you truly don’t. And it might be very informative to know precisely what makes you say that since you haven’t disclosed something about what ever the “it” is that you simply suppose the entire business is behind.
Spotify’s Shadow Library Downside
Throughout the AI sector, a now-familiar sample has emerged: Prepare first. Clarify later — if ever.
The music business has already seen this logic elsewhere: large ingestion adopted by retroactive justification. The query now’s whether or not Spotify — a licensed, mainstream platform for its music service — is replicating that very same sample inside a closed AI ecosystem for which it has no licenses which have been introduced.
So the query have to be requested clearly:
Is Spotify’s AI spinoff engine constructed completely on disclosed, licensed coaching sources? Or is that this merely a platform-contained model of shadow-library coaching?
As a result of if fashions ingested:
- Unlicensed recordings
- Consumer-uploaded infringing materials
- Catalog works with out express coaching disclosure
- Performances missing performer consciousness
then AI derivatives danger turning into a backdoor exploitation mechanism working exterior conventional consent constructions.A spinoff engine constructed on undisclosed coaching provenance shouldn’t be a creator instrument. It’s a legal responsibility hole. , sort of like Anna’s Archive.
A Direct Response to Gustav Söderström : What Coaching Would Really Be Required?
Launching a real music technology or spinoff engine would require large, structured coaching, together with:
1. Massive-Scale Audio Corpus
Hundreds of thousands of full-length recordings throughout genres, eras, and manufacturing kinds to show fashions musical construction, timbre, association, and efficiency nuance. Now the place would possibly these come from?
2. Stem-Stage and Multitrack Knowledge
Separated vocals, devices, and manufacturing layers to permit recombination, remixing, and stylistic transformation.
3. Efficiency and Voice Modeling
Intensive vocal and instrumental recordings to seize phrasing, tone, articulation, and expressive traits — the very parts tied to performer id.
4. Metadata and Behavioral Indicators
Tempo, key, style, temper, playlist placement, skip charges, and listener engagement information to information mannequin outputs towards commercially viable patterns.
5. Model and Similarity Encoding
Statistical mapping of musical traits enabling the system to generate “within the fashion of” outputs — the core mechanism behind spinoff technology.
6. Iterative Retraining at Scale
Steady ingestion and refinement utilizing newly out there recordings and platform information to enhance constancy and relevance.
7. Funding for all the above
No generative music system of consequence may be constructed with out huge coaching publicity to actual recordings and performances, and the expense.
Which returns us to the unresolved query:
The place did Spotify acquire that coaching information?
As a result of the difficulty shouldn’t be whether or not Spotify may license coaching materials. The difficulty is that Spotify has not defined — in any respect — how its coaching corpus was assembled.
Opacity is the issue.
Personhood Indicators: Coaching on Recordings Is Coaching on Individuals
Spotify can describe AI derivatives as “music instruments,” however coaching on recordings is not only coaching on songs. Recordings include personhood indicators — the distinctive human identifiers embedded in efficiency and manufacturing that allow a system be taught who somebody is (or can sound like), not merely what the composition is.
Personhood indicators embody (non-exhaustively):
- Voice id markers (timbre, formants, prosody, accent, breath, idiosyncratic phrasing)
- Instrumental efficiency fingerprints (assault, vibrato, timing micro-variance, articulation, swing really feel)
- Studio-musician signatures (the “nonfeatured” musicians who are sometimes most identifiable to different musicians)
- Songwriter kinds harmonic signatures, prosodic alignment, and lyric id markers
- Manufacturing cues tied to an artist’s model (adlibs, signature FX chains, cadence habits, recurring supply patterns)
A contemporary generative system doesn’t must “copy Monitor X” to take advantage of these indicators. It will probably summary them — compress them into representations and weights — after which reconstruct outputs that commerce on id whereas claiming no specific recording was reproduced.
That’s why “licensing” isn’t the actual threshold query right here. The brink questions are disclosure and permission:
- Did Spotify extract personhood indicators from performances on its platform?
- Had been these indicators used to coach methods that may output tokenized “feels like” content material?
- Are there credible guardrails that stop the mannequin from producing identity-proximate vocals/instrumental efficiency?
- And might creators confirm any of this with out having to sue first?
If Spotify’s coaching information provenance is opaque, then creators can not know whether or not their identity-bearing performances have been transformed into mannequin worth which is the start of commoditization of music in AI. And when the platform monetizes “derivatives” (aka competing outputs) it dangers constructing a brand new income layer (for Spotify) on high of the very human indicators that performers have been by no means requested to contribute.
The Asymmetry Downside
Spotify is aware of what it skilled on. Creators don’t. That asymmetry alone is a structural concern.
When a platform possesses full data of coaching inputs, mannequin structure, and monetization pathways — whereas creators lack even fundamental disclosure — the bargaining imbalance turns into absolute. Transparency shouldn’t be optionally available on this context. It’s the minimal situation for legitimacy.
With out it, creators can not:
- Assert rights
- Consider consent
- Measure market displacement
- Perceive whether or not their work formed mannequin habits
- And even know whether or not their id, voice, or efficiency has already been absorbed into machine methods
As each bully is aware of, opacity redistributes energy.
Derivatives or Displacement?
Spotify frames AI derivatives as inventive empowerment — followers remixing, artists increasing, new income streams rising. However the core financial query stays unanswered:
Are these instruments supplementing human creation or substituting for it?
If spinoff methods can generate stylistically constant outputs from skilled materials, then the worth captured by the mannequin originates in human recordings — recordings whose function in coaching stays undisclosed. In that situation, AI derivatives should not merely instruments. They’re artificial rivals constructed from the inventive DNA of the unique artists. Sort of like MTV.
The excellence between assistive and substitutional AI is financial, not rhetorical.
The Query That Will Not Go Away
Spotify might proceed to discuss AI derivatives within the language of alternative, scale, and inventive democratization. However none of that resolves the underlying problem:
What did they practice on?
Till Spotify supplies clear, verifiable disclosure concerning the origin of its coaching information — not merely licensing claims, however precise transparency — each spinoff output carries an unresolved provenance drawback. And within the age of generative methods, undisclosed coaching is an actual danger to the artists who feed the beast.
Framed this manner, the hurt shouldn’t be merely copy of a copyrighted recording; it’s the extraction and commercialization of identity-linked indicators from performances doubtlessly impacting featured and nonfeatured performers alike. Spotify’s failure (or refusal) to reveal coaching provenance turns into a part of the hurt, as a result of it prevents anybody from assessing consent, compensation, or displacement.
And it makes it unimaginable to grasp what worth Spotify needs to license, a lot much less whether or not we wish them to do it in any respect or practice our replacements.
As a result of possibly, simply possibly, we don’t what one other Spotify to construct a enterprise on our backs.
[This post first appeared on MusicTech.Solutions]



