General Knowledge (93)
Ableton Live 12.3 has arrived: Stem Separation, Splice integration and everything else you need to know
Written by Sounds SpaceAbleton Live 12.3 has arrived, and this one’s got a little bit of everything:
AI-powered stem separation, native Splice integration, better bounce workflows, notable Push 3 improvements, and a clutch of smaller but genuinely useful workflow refinements. The update is free to existing Live 12 users, and Ableton is running a limited-time 25% discount for newcomers and Pack purchases, so whether you’re already deep into Live or thinking about joining the ecosystem, now’s a good moment to take stock.
Below, I break down the features that matter most, what they’ll let you do differently in the studio or on stage, and a few practical tips for getting the most out of the new tools.
The big headline: stem separation built into Live Suite
The flagship addition in 12.3 is built-in stem separation for Live Suite. This isn’t a third-party plug-in or cloud-only service — Ableton has integrated local, offline stem splitting that separates an audio clip into four component stems: Vocals, Drums, Bass, and Others. Practically, that means you can take any sample, loop, or full stereo track and pull it apart inside Live, then remix, rearrange or resample the resulting parts without leaving your DAW.
Why this matters: Stem separation used to be a workflow that required external services or specialized tools that sent audio to the cloud. Having a fast, local option directly inside Live closes the loop: faster iteration, less context switching, and no file juggling. For remixers, producers who build stems for collaborators, or anyone doing sample-based creativity, the ability to isolate a vocal line or a drum bus without leaving your session is huge. Expect creative uses beyond remixing, to think creative sidechaining, subbing in new drums under a vocal stem, or extracting a texture from the “Others” stem to turn into an ambient pad.
A practical tip: stem separation quality varies with the source material. Clear, well-separated recordings (dry vocals, distinct drum hits) give the best results; heavily distorted or extremely dense mixes may produce artifacts. Use the stems as starting material: resample them, run them through effects racks, and don’t be afraid to combine stems back together after processing.
Splice integration: search, audition, and drop — inside Live
Ableton’s Splice integration is more than a link — it brings Splice’s sample library into Live’s Browser so you can search, audition in sync and key, then drag samples into your project without switching apps. The “Search with Sound” feature is particularly neat: you can capture audio from your set (or drag a clip into the Splice panel) and ask Splice to find samples that fit the rhythms and harmonic content of what you already have. That can turn a friction-filled sample hunt into a fast, creative playground.
Why this matters: searching for the right sample used to be a deep rabbit hole — dozens of browser tabs, trawling keyword searches, guessing about tempo and key. Native integration means auditioning is instantaneous and context-aware: the samples are previewed in time with your project, so you hear how they groove before committing. For fast sketching and late-night idea sessions, that’s a serious time-saver.
Bounce Groups and smarter offline workflows
Another workflow-focused upgrade is Bounce Groups: the ability to render an entire group (with processing) to a single audio file. This lets you commit CPU-heavy group chains to audio without losing the option to keep your original tracks for later edits — a clean balance between commit-and-mix efficiency and flexibility. It’s the kind of workflow improvement you don’t notice until you need it, then you wonder how you ever lived without it.
Other audio/bouncing improvements under the hood include faster, more reliable bounce operations and fixes for edge cases on different platforms. If you run large projects or play Live sets that depend on pre-rendered stems, these changes will smooth your workflow and (importantly) reduce last-minute rendering headaches.
Push 3 and hardware improvements
Live 12.3 isn’t just a software update: Push 3 gets form-and-function updates that expand what you can do in standalone and tethered modes. Notably, Push 3 in standalone mode can now work with class-compliant audio interfaces — meaning more ins/outs and a more flexible standalone setup without relying on ADAT tricks. Push’s expressive grid also gets new XY-style control modes and improved step-sequencing with touch-sensitive velocity control, plus a new Rhythm Generator view for drum programming. If you own Push 3, these firmware/software upgrades broaden its standalone studio potential.
Why this matters for live performers: being able to plug a wider range of audio interfaces into Push 3 without complex routing opens up richer live setups. For producers, the XY mode and improved sequencing make Push more tactile and creative for beat design and expressive performance.
Smaller but meaningful updates
12.3 also brings a handful of thoughtful enhancements that will please power users:
-
Auto Pan → Auto Pan-Tremolo: The Auto Pan device gets more modes and dynamic responsiveness, including dedicated tremolo behavior and level-based response shaping — great for rhythmic modulation and dynamic pumping effects.
-
A/B states for devices: You can set A and B states for instruments and effects and flip between them easily. That’s a huge boon for sound design and comparative mixing — quickly audition two radically different device settings without losing your place.
-
New Packs and Max for Live tools: Live Standard and Suite users get new creative devices (Patterns, Sting) and updates to Expressive Chords and Sequencers — useful for generative ideas and getting out of production ruts.
These are the kinds of improvements that might not make the big headlines but end up improving daily workflows — faster experimentation, easier comparisons, and more creative starting points.
Performance, platform fixes, and stability
As with most iterative updates, Ableton has bundled a number of stability and compatibility fixes across platforms. Release notes mention fixes for stem separation failures on certain macOS configurations and improvements to the Splice UI behavior, among other bug fixes. The public beta cycle surfaced issues and Ableton addressed several of them before pushing the stable release, which is reassuring for users who rely on Live in critical sessions.
If you depend on specific third-party plug-ins or unusual workflows, it’s always wise to test 12.3 on a copy of your projects first — don’t overwrite production sessions until you confirm all the plug-ins and setups behave as expected.
Pricing, availability, and the limited-time offer
Ableton is releasing 12.3 as a free update for everyone already on Live 12 (so if you’re on Live 12 Standard or Suite, it’s yours at no extra cost). For newcomers or those upgrading from much older versions, Ableton is running a limited-time promotion: 25% off Live 12 and Packs (and 20% off Push 3 and related hardware) for a short window around the release. If you were sitting on the fence about upgrading or buying in, that’s a practical savings window to consider.
Who should care — and who might want to wait
-
Remixers and sample-based producers: stem separation and Splice integration are direct wins. Faster sample hunting plus local stem extraction changes the way you can build remixes and reworks.
-
Live performers and Push 3 owners: Push 3’s standalone expansions and Bounce Groups make set preparation and standalone jamming more powerful.
-
Sound designers and experimental producers: A/B device states, Patterns, Sting, and Max for Live updates give fresh sound-design workflows.
-
Users running older or very plugin-heavy projects: test first. While Ableton has patched many issues, complex third-party setups sometimes reveal edge cases.
Quick hands-on tips to get started with 12.3
-
Try stem separation on a few different source types — a dry acapella, a full mixed track, and a drum loop — to learn the tool’s strengths and limitations. Resample processed stems to hide artifacts.
-
Open the Splice panel and experiment with Search with Sound: drag a loop or capture a section, let Splice find matches, and audition samples in sync. The faster previews will change how you hunt for sounds.
-
Bounce a group, then compare: use Bounce Groups to commit heavy group processing to audio, then A/B with the original to confirm what you’ve gained or lost.
-
Explore Push 3’s new modes if you own one — class-compliant audio support alone opens up new routing possibilities for standalone rigs.
Final thoughts
Ableton Live 12.3 is a model of how mature DAWs continue to evolve: headline features (stem separation, Splice integration) that grab attention, plus a steady stream of smaller but meaningful workflow and hardware updates. The inclusion of local stem separation is a game-changer for many workflows, while Splice integration tightens the loop between inspiration and production. For Live users, the update is essentially a must-install; for newcomers, the limited-time discount sweetens the deal.
If you use Live in any professional or semi-professional capacity, treat 12.3 as a strong, practical refinement — one that delivers immediate creative tools and sensible performance improvements rather than risky, half-baked experimentation. Fire up a non-critical project, poke around the new panels and devices, and you’ll likely find something that speeds up your process or sparks a new idea.
Suno and Warner Music Group: what the settlement means for artists, AI, and the future of music
Written by Sounds SpaceSuno and Warner Music Group: what the settlement means for artists, AI, and the future of music
This week’s surprising — and, for some, seismic — news that Warner Music Group (WMG) has settled its lawsuit with AI music generator Suno and entered into a commercial partnership marks a turning point in the music industry’s standoff with generative AI. After more than a year of litigation, public debate and fear among creators about the misuse of their work, the two companies announced a deal that aims to balance Suno’s rapid technical progress with protections and revenue for artists. Below, I unpack what happened, what’s in the agreement as reported, the likely consequences for creators and platforms, and what the settlement tells us about how music and AI will coexist going forward.
The headline: settlement + partnership, not a courtroom victory
At the heart of the story is a straightforward pivot: Warner, which had been litigating against Suno for alleged copyright infringement, is no longer pursuing that courtroom route — instead, the companies have struck a licensing and commercial partnership. As part of the agreement, Suno will phase out its current broad-use models in favour of licensed models, implement new restrictions on downloads (including limiting or blocking downloads for free users and capping paid-user downloads), and introduce mechanisms that let Warner artists and songwriters control whether and how their names, voices, likenesses, and compositions are used on Suno’s platform. Suno also announced the acquisition of Songkick — a live-music discovery brand previously under Warner’s control — as part of the broader transaction.
That combination — settling the legal claim while creating a commercial path forward — signals both sides’ strategic thinking. Warner gained contractual safeguards and revenue opportunities without protracted litigation; Suno gained industry legitimacy and access to an enormous catalog of artist assets and marketing channels. But the wrinkle is important: the agreement appears to be opt-in for artists. That means the platform can offer realistic “artist-like” outputs only for those creators who choose to participate, while others are (in principle) protected from unauthorized stylistic or likeness use.
What the agreement reportedly includes (the practical bits)
Reporting across major outlets and Suno’s own announcement gives us a reasonably clear list of concrete changes and commitments:
• Suno will launch new, licensed AI models next year to replace its current open models, designed to operate under licensing terms that compensate rights-holders.
• Downloads of AI-generated audio will be restricted: free accounts will no longer be able to download songs; paid accounts will have download caps and the option to pay more for higher allowances.
• Artists and songwriters signed to Warner will be able to opt in (or not) to have their voices, names, likenesses, and compositions used; participating artists will receive compensation as negotiated.
• Suno acquired Songkick from Warner; Songkick will remain as a fan destination under Suno’s ownership.
Warner and Suno have not publicly disclosed the financial terms of the settlement, nor the exact revenue split for artists — understandably, these are sensitive commercial details. But the structural commitments are what matter most for industry precedent: an existing AI vendor has accepted licensing obligations, and a major label has accepted a commercial route instead of purely legal enforcement.
Why this matters: three immediate implications
-
It normalizes licensing as the dominant path forward. For much of 2024–2025 the dispute between labels and AI companies felt binary: either platforms would be forced to stop using copyrighted content in training and outputs, or labels would license their works. This settlement validates the licensing route. If licensing becomes the industry norm, it means artists (and labels) can directly capture value when their style or likeness is used, rather than relying on uncertain litigation victories.
-
It sets an opt-in model for artist control. The opt-in approach that Warner and Suno are implementing is significant because it preserves artist agency. Rather than an all-or-nothing ban, artists can choose new revenue streams while controlling the use of their voice and other personal rights. That could become a blueprint for other deals — but it also raises questions about bargaining power and transparency in the offers artists receive.
-
It changes product economics for AI platforms. Requiring paid accounts and capping downloads shifts the economics of Suno’s product from an ad/scale-first free model to a hybrid subscription/licensing model. That could slow viral, mass-free proliferation of AI-created tracks, while creating predictable revenue lines that can be shared with rights-holders. It’s a step away from the “wild west” era of unfettered generation.
The broader legal landscape: one settlement, more work to do
It’s important to remember this is not the end of legal friction. Universal Music Group previously settled with AI platform Udio, and both Sony and Universal have ongoing or recent legal entanglements with AI startups. Suno’s deal with Warner doesn’t automatically resolve disputes with other labels or publishers, nor does it fully answer hard questions about training datasets, informed consent, or derivative uses. The settlement shows parties can negotiate, but it does not replace necessary public policy discussions about how copyright law should adapt to generative models.
Moreover, artist advocacy groups and creators have voiced concerns about transparency and fairness in earlier label-AI deals. Opt-in programs look good on paper, but their fairness depends on the visibility of contract terms and the bargaining leverage individual artists — especially less established ones — can realistically exercise. Without standardised transparency practices or industry-wide minimums, a fragmented patchwork of label-specific offers could leave many creators at a disadvantage.
For artists: opportunity and risk
For major, established artists, the Suno–Warner deal represents a new monetization channel and a way to control fan experiences. Imagine paid fan interactions where users can generate licensed remixes or AI-collabs that include an artist’s signature timbre or phrasing — with the artist paid for each use. That can be lucrative.
But for less-visible artists, the risks are real. Labels often negotiate on behalf of roster artists; depending on contract terms, an artist might find their likeness or composition licensed without direct negotiation with them, or receive royalties according to label agreements that artists already critique as opaque. The real question is whether these licensing mechanisms will be transparent and equitable at the artist level, not just at the label level.
For platforms and startups: a new playbook — at a price
For generative-audio startups, the message is clear: build licensing into the product roadmap early. Models trained on copyrighted works without clearance create legal risk and, as we’ve seen, can end with either an injunction or a negotiated settlement that forces product redesign. For investors and founders, the Suno–Warner deal shows a path to scale that includes paying for rights and restructuring user experiences — but it comes with higher operating costs and a need for careful artist relations.
This outcome also tilts the playing field toward platforms that can shoulder licensing costs and administrative complexity: well-funded startups, major tech companies or ventures backed by institutional capital. Smaller open-source projects and hobbyist tools may find it harder to offer competitive functionality without proper rights clearance.
Policy and the public interest: what regulators should watch
The Suno–Warner settlement may reduce the immediate urgency for aggressive legislative fixes, but regulators still have work to do. Key policy areas include: clarifying whether training on copyrighted material qualifies as fair use, mandating transparency around datasets, and ensuring a reasonable framework for attribution and compensation. There’s also a consumer-protection angle — users should understand when they are generating content that directly resembles a living artist and what rights attach to that output.
Finally, antitrust considerations could emerge — as major labels and well-funded AI firms stitch commercial relationships, regulators should watch for exclusivity deals that could lock out competitors or limit creative alternatives for artists and fans.
The cultural question: will AI-made music co-exist with human artistry?
Beyond contracts and code lies a cultural debate. AI music platforms can democratize production, allowing novices to create richer-sounding tracks and fans to experience new kinds of interactivity. But they also risk diluting artistic labor if the market floods with indistinguishable, cheaply generated tracks. The Suno–Warner settlement nudges the ecosystem toward a model where human artistry is recognized as a monetizable input to AI outputs — a compromise that, if implemented fairly, might preserve creative incentives.
That said, commercial deals alone won’t answer aesthetic questions about authenticity and taste. Those will be decided over time by listeners, curators, and creators themselves. If fans value the human story behind songs — the voice, lived experience, performance — then human artists retain a cultural edge. If, instead, novelty and volume dominate streaming economies, artists will need to adapt their business models accordingly.
Final thoughts: a test case, not the final chapter
The Suno–Warner settlement is a high-profile test case in an industry grappling with technological change. It shows litigation and negotiation can coexist: the threat of legal action pushed an AI startup to the bargaining table, and the settlement produced a commercial framework that may become a template for other deals. But the devil is in the details — the fairness of compensation, the transparency of contracts, and the long-term policy framework will determine whether this outcome is a durable solution or merely a stopgap.
For artists, the takeaway is simple but urgent: engage with these developments, seek clarity about contracts, and insist on transparency. For platforms, the lesson is equally practical: if you want to scale in music, plan to pay — financially and reputationally — to licence the human artistry that makes music meaningful.
We’re in the early chapters of the AI-music story. The Suno–Warner deal doesn’t end the debate, but it moves the conversation from purely adversarial litigation to negotiated commerce — and that shift has consequences for how music will be made, shared and valued in the years ahead.
Sources: reporting and company announcements from TechCrunch, Suno’s official blog, the Los Angeles Times, The Guardian and Pitchfork informed this analysis.
If you’d like, I can now:
• Draft a short “artist-facing” explainer summarizing what this deal means for an individual artist (rights, revenue and questions to ask).
• Or create a one-page checklist for independent musicians to protect their work and negotiate AI licensing terms. Which would help you more?
When the majors embraced the machine: what KLAY’s licensing deals with the “big three” mean for music’s AI future
Written by Sounds SpaceWhen the majors embraced the machine: what KLAY’s licensing deals with the “big three” mean for music’s AI future
The music industry has been wrestling with artificial intelligence for the better part of this decade — first in courtrooms and headlines, then in negotiation rooms and press releases. This week’s news that Los Angeles–based AI music startup KLAY Vision Inc. has struck licensing deals with the three major record groups — Universal Music Group, Sony Music (and their publishing arms), and Warner Music Group — marks a clear inflection point. For the first time, the companies that own the lion’s share of recorded and published music have formally licensed their catalogs to an AI-driven streaming product intended to let fans remake songs in new styles. That’s business-as-usual turned on its head: the record industry is no longer only defending copyright against automated imitators — it’s now actively building a commercial framework to monetize AI creativity.
Below, I unpack what the KLAY deals actually are, why they matter, the unanswered questions they raise for artists and listeners, and what they suggest about how the music business will adapt to — and profit from — generative audio technologies.
What KLAY says it does (and what the labels licensed)
KLAY bills itself as a music-technology company building a streaming product with integrated AI tools that let users remix and “remake” songs. The startup says it trained its systems on thousands of licensed recordings and compositions, and that the agreements with Universal, Sony, and Warner grant KLAY the rights to use those catalogs both to power its machine-learning models and to enable listeners to create personalized AI variations of the original songs. In press materials and label statements, the emphasis is on a paid, controlled system that compensates rights-holders and gives participating artists “control” over how their work is used.
Put simply: instead of AI engines scraping the internet for music (and getting sued for it), KLAY has secured the catalogs it needs officially — licensing the underlying recordings and publishing rights so it can legally train models and offer derivative experiences to subscribers. This is the fundamental pivot from conflict to commerce.
Why the labels decided to license, not litigate
The move reflects a broader cooling of the industry’s initial, lawsuit-heavy reaction to AI-generated music. Early skirmishes — including high-profile cases targeting companies accused of training models on copyrighted recordings without permission — exposed how difficult it is to police every synthetic output on major platforms. Licensing gives labels an avenue to recapture value from AI activity rather than leave it to unregulated creators and deepfake projects.
From a business perspective, the labels are treating AI as a new revenue stream: licensed models can be monetized through subscriptions, micro-payments for individual remixes, and potential new sync/derivative licensing arrangements for commercial use. By agreeing on terms now, labels lock in payment frameworks and control mechanisms rather than ceding the field to adversaries. Reuters and other industry outlets reporting on the deals note that this is the first time a single AI company has landed agreements across all three majors — a symbolic and practical milestone.
What this could mean for artists and songwriters
The labels stress payments and artist protection, but practical outcomes will depend on contract details that aren’t yet public. Important questions include:
-
Revenue split mechanics: Will revenues from AI-remixes be shared with performers, songwriters, producers and publishers in the same way as traditional streams, or will there be a bespoke split for AI-derived works? Early statements promise “proper recognition and reward,” but the breakdown — and whether smaller artists get a fair share — is not yet transparent.
-
Opt-in vs opt-out: Some label statements imply artist participation will be managed through label agreements, but will individual artists be able to opt out of allowing their work to be used by AI remixes? Opt-in models (where artists explicitly authorize usage) are friendlier to artist autonomy; blanket label-level licensing is administratively simpler but risks leaving artists surprised by new forms of derivative use.
-
Moral rights and voice cloning: Even when a song is licensed, voice impersonation or highly realistic vocal “clones” of living artists raise unique reputational and ethical concerns. Will artists be able to veto impersonations that sound like them, or require additional consent for likeness-based uses? The deals suggest labels will attempt to govern these uses, but public-facing safeguards remain to be seen.
In short: licensing opens the door to new income streams, but it also creates novel bargaining points and potential pitfalls. Musicians and their managers should read the fine print closely as product specifics unfold.
How KLAY’s product model could work in practice
Based on available reporting, KLAY appears to be positioning itself as a hybrid of a streaming service and an interactive creative tool. Imagine a subscription app where a fan can pull up a favorite track and choose stylistic options — tempo, instrumentation, genre twist, vocal emphasis — and the AI generates a personalized version on the fly. That output would be playable inside the app, shared within the platform, and possibly downloadable or licensed for other uses depending on rights negotiated.
Business models that have been floated by industry commentators include:
-
Subscription-first: paying for access to the AI remixing engine and licensed catalog (similar to an advanced streaming tier).
-
Per-use micropayments: small fees for individual remixes or for the right to download/share an AI-created version.
-
Creator tools and marketplace: artists and producers could sell presets, commercial licenses, or custom remixes created with the platform.
All of these require plumbing to allocate money back to the correct rights-holders and to catalogue derivative uses for accounting — a nontrivial metadata and royalty-engineering challenge.
Risks and friction points
While the deals are a big step toward normalization, several hazards remain:
-
Quality dilution and platform clutter: Generative tools can flood platforms with low-quality, AI-generated tracks masquerading as new music. That trend has already raised concerns on existing streaming services where “AI spam” has pushed down discoverability for human artists. A licensed approach may mitigate some of that, but discovery algorithms and curation will matter.
-
Deepfake vocals and likeness rights: Even with catalogs licensed, using an artist’s vocal likeness without separate consent is a distinct legal and ethical issue. Platforms and labels will need enforceable rules about vocal impersonation and identity-protecting mechanisms.
-
Creative credit and attribution: How will AI-remixes credit the original artist, songwriter, and the machine model? Proper attribution standards are crucial for maintaining transparency and preserving the lineage of creative works.
-
Inequality of bargaining power: Major-label artists and top-tier songwriters are likely to get better negotiated terms than independent or legacy artists signed on older deals. Without industry-wide standards, smaller creators could be left behind.
Why is this also an opportunity
For listeners, platforms like KLAY could unlock new ways to interact with music: personalization, reimagined versions of favorites, and collaborative experiences where fans co-create with AI. For creators, licensed AI can serve as an additional income source and a promotional tool: remixes can introduce songs to new audiences, and interactive experiences can enhance fan engagement.
More broadly, the deals hint at an industry strategy: partner with AI entrepreneurs and formalize revenue pathways rather than attempting to shut down innovation entirely. That approach promises to steer AI development toward mutually beneficial products — if the commercial terms are fair and transparent.
What to watch next
If you’re tracking this story, here are the developments that will matter most in the coming months:
-
Public launch and product UX: How KLAY implements artist controls, attribution, and payment flows in its actual app will show whether the deals were symbolic or substance-driven. A consumer launch will reveal whether listeners embrace remixable subscriptions.
-
Artist reactions and opt-ins: Watch for statements or actions from major artists about whether they authorize voice-cloning-like features for their work, and whether there’s an opt-out movement among musicians uncomfortable with the new model.
-
Regulatory scrutiny and industry standards: Governments and industry bodies may step in to set guidelines around deepfakes, voice rights, and transparent royalty reporting. The deals with the majors could influence those standards.
-
Extensions and competition: Other AI firms and major labels may ink similar deals, or the majors might partner with multiple AI players. Competition could be healthy — pushing platforms to improve artist protections and user experience.
Final take: pragmatic adaptation, with caveats
KLAY’s licensing deals with Universal, Sony and Warner represent a pragmatic shift: the music industry is no longer fighting the generative-AI tide by default. Instead, it’s trying to steer that tide into monetizable channels. For fans and creators, that can be a win — new experiences and revenue streams — but the outcome depends on how equitably and transparently the underlying systems are built.
This moment is a reminder that technology doesn’t just disrupt markets: it forces social and legal frameworks to catch up. Licensing the catalogs was step one. Step two is delivering a platform that truly balances innovation, artist rights, and fair payment. If the majors and KLAY can design that system together — with clear artist consent, transparent accounting, and sensible limits on voice impersonation — the music ecosystem may emerge more diverse and more lucrative for creators. If not, the next chapter will be courtrooms and headlines all over again.
Either way, the industry’s choice to collaborate with — rather than simply contest — AI companies signals that generative audio is here to stay. The crucial question now is who writes the rules, and whether those rules will protect the people who make the music.
“Music producers are rejecting AI”: what the new studies tell us — and what it means for music’s future
Written by Sounds Space“Music producers are rejecting AI”: what the new studies tell us and what it means for music’s future
Brother — the headlines have been loud this week: a new survey says over 80% of music producers are actively against AI-generated songs, while separate listener-focused research shows most people can’t tell AI-made tracks from human ones. That collision — creators rejecting the tech at the same time listeners can’t reliably detect it — is the perfect storm for arguments about ethics, jobs, creativity, and regulation. Below, I’ll walk you through the findings, explain why they matter, and give a blunt take on what could happen next.
The headline findings:
• A global survey of music producers, commissioned by Tracklib, reports that a very large majority of producers are sceptical or hostile to generative-AI use in music: adoption is low and rejection high — the press release frames it as “music producers are rejecting AI,” with only a small percentage actively using generative tools.
• At the same time, Deezer and Ipsos ran listening tests and found that ~97% of people couldn’t reliably tell AI-generated tracks from human-made ones in blind tests — a striking demonstration that AI music quality has reached near-parity for the ear of the average listener. That same study shows big public support for transparency (label AI-made tracks) and concerns about ethics and copyright.
• Major outlets reporting on these themes also point to real-world effects: AI tracks are already rising on streaming platforms and even some charts, which is intensifying industry debate about promotion, royalties, and verification.
What the producers’ rejection actually means
When a study says “80%+ of producers are against AI,” don’t read that as a single emotion — it’s a mix of reasons:
-
Artistic integrity and craft. Many producers see production choices as part of artistic identity. Machines that spit out passable arrangements or vocal lines feel like a shortcut that removes the human voice from the process. Tracklib’s survey captures that sentiment: a lot of rejection is because producers feel AI undermines the craft.
-
Quality concerns — and nuance. Some producers reject full-song generation but still use AI for small tasks (like reference stems, mastering presets, or chord suggestions). The “rejecting AI” headline often conceals that many are open to assistive AI, just not to AI that replaces core creative decisions.
-
Copyright and fairness. Producers worry about training data (AI trained on copyrighted music without consent), displacement, and how royalties should be split if models borrow from identifiable artist outputs. The Deezer/Ipsos work also shows public support for ethical training and labeling — so this is both a creator and consumer issue.
-
Economic risk. If streaming platforms allow AI tracks to surge without clear rules for attribution and payment, session work, ghost production, and publishing could lose income. That’s why producers are defensive — it’s their livelihood, not just ideology.
Why listeners’ inability to tell matters
The Deezer/Ipsos finding that ~97% of listeners failed to detect AI tracks is a game-changer for three reasons:
-
Market-level indistinguishability. If audiences can’t tell, then AI tracks can compete on the same platforms, playlists, and algorithms. That puts human artists in direct competition with low-cost synthetic output unless platforms treat them differently.
-
Transparency demand. The public’s unease in that study shows people want to know. If platforms don’t label AI content, they risk trust erosion and regulators stepping in. Deezer has already started tagging AI submissions — an early institutional response.
-
Policy & chart implications. Charts and awards rely on attribution; if synthetic works crowd charts, the industry must decide whether to treat them equally. We’re already seeing cases of AI songs making chart noise and being removed or disputed. That bubbling conflict is the reason labels, distributors, and streaming services are scrambling.
The contradictions: why both things can be true at once
It seems paradoxical: producers reject AI while listeners don’t care or can’t detect it. But both can be true because they’re answering different questions. Producers are speaking about process, ethics, and livelihoods; listeners are responding to end-product enjoyment. The industry is now forced to reconcile process (how music was made and who benefits) with product (what people enjoy or stream).
The pressure points: where conflict will be decided
Here are the battlegrounds we should watch:
-
Streaming platform policies. Will platforms require AI labels? Will they remove AI tracks from editorial playlists? Deezer’s tagging is a blueprint; other platforms may follow or resist.
-
Copyright litigation & licensing. Lawsuits are already happening and will accelerate. Courts will decide whether training models on copyrighted works is infringement, and that ruling will determine cost structures for AI music firms.
-
Royalties & metadata standards. If AI tracks use samples or stylistic mimicry, who gets paid? The industry needs metadata standards and traceability to assign rights and royalties correctly — otherwise, payments and charts will be gamed.
-
Consumer-facing transparency. Mandatory labelling is likely to be politically popular and technically feasible. Expect regulators to push for clarity so consumers know whether they’re listening to AI.
-
Artist and union pressure. Creators’ associations and unions will push for protections, training data consent, and revenue-sharing terms. If producers maintain unified resistance, they can shape licensing frameworks more than tech companies expect.
The likely near-term scenarios (what I actually expect)
-
Hybrid adoption increases. Most producers will continue using AI as assistants (mixing tools, stems, mastering suggestions) rather than full-song generators. The “pure AI” artist will mostly be an outlier or novelty that occasionally breaks through virally.
-
Platform responses: tags + restrictions. Platforms will adopt transparent tagging and may restrict AI-only tracks from editorial playlists and algorithmic boosts until legal/ethical frameworks are set — Deezer is already an early mover here. Expect other major services to follow to avoid reputational and legal risk.
-
Regulatory and legal clarifications. Courts and regulators will clarify what “training on copyrighted works” means. If rulings favor copyright holders, AI music firms will need licensing deals akin to the way streaming licenses catalogs, which raises costs and reduces the free-for-all.
-
An uneven marketplace. We’ll end up with two flavors of music commerce: human-made content protected, promoted, and monetized under traditional deals; and AI-generated content that lives in a parallel space with different rules, visibility, and monetization models. Chart systems may split or flag AI entries.
What this means for producers, labels, and listeners: practical takeaways
For producers:
-
Lean into what makes you uniquely human: emotional risk-taking, imperfections, personal storytelling, and brand as a creator. These are the qualities AI struggles to authentically replicate.
-
Treat AI as a tool, not a competitor. Learn to use assistive AI for workflow speedups (mastering, stem separation, idea generation) while protecting your signature creative decisions.
For labels & distributors:
-
Invest in metadata standards and provenance systems now. If you don’t capture who did what — and whether a track used AI — you’ll lose control over chart inclusion, payouts, and legal defenses.
-
Think about curated playlists as a premium human-made space; reserve some editorial real estate for verified human artistry to protect brand value.
For listeners:
-
Expect transparency features (filters, labels, opt-outs). If you want only human-made music, the tools to enforce that preference are likely coming.
Final thought, this is a crossroads, not an end
AI music is no longer a futuristic demo or toy. The technology is good enough to matter commercially and culturally, and the industry’s response will shape careers and revenue flows for a generation. Producers’ concern is legitimate — it’s about artistry and survival. Listeners’ reaction is equally important: they show curiosity but also want honesty.
The healthy path is a negotiated one: transparency, licensing, and new business models that protect creators while allowing innovation. If the industry can’t agree, regulation will force the issue — and that could either protect creators or ossify the market. Either way, producers rejecting AI today is a loud signal: the music world wants rules before mass adoption. That moment of negotiation is the real story behind the headline.
Suno raises $250M — what this means for music, makers, and the industry
Written by Sounds SpaceSuno raises $250M — what this means for music, makers, and the industry
Big money is chasing big ideas. AI music startup Suno just closed a $250 million Series C round at a reported $2.45 billion post-money valuation — a jump that underlines investor faith in generative audio, and at the same time stokes every debate happening around AI and creative work right now. This isn't just another VC headline: it has real consequences for artists, labels, startups and anyone who cares about how music gets made, credited, and monetized.
Belo,w I break down the news, why investors are writing huge checks, the tensions and lawsuits now in play, and what musicians and music businesses should actually do next.
The headlines (fast)
-
Suno raised $250M in a Series C led by Menlo Ventures, with participation from groups including NVIDIA’s venture arm and others. The round reportedly values the company at $2.45 billion.
-
Published reporting indicates Suno’s growth has been rapid (user growth and monetization metrics are being cited in industry coverage) — some outlets claim high revenue and quick adoption.
-
At the same time, Suno is entangled in copyright challenges and criticism from musicians and labels about how generative music models were trained and how outputs resemble existing works. The legal and cultural fights are unresolved.
Why investors doubled down — a simple math + product story
Investors put money where they see leverage: a product that can scale rapidly with relatively low marginal cost, a big addressable market, and defensible tech or distribution. Suno checks several boxes:
-
Mass-market potential. Text-to-music and AI-assisted songwriting promise to democratize music production: people who never touched a DAW can generate a song from a prompt. That expands the total market of music creators — and therefore potential paying customers, creators who want tool subscriptions, and businesses that need licensed music.
-
Platform economics. If Suno can combine an easy generator with a marketplace, publishing or licensing deals, it can earn recurring revenue on subscriptions, commercial licenses, and partnerships (e.g., plugin bundles, enterprise licensing for games, ads, content creators). Multiple recent write-ups highlight Suno’s strong monetization trajectory, which helps justify a big valuation.
-
Strategic investors. With participation from deep-tech and media investors, Suno gains both capital and routes into key partners — think distributed compute (NVIDIA), media relationships, and scaling channels. Those backers see AI audio as infrastructure that will be embedded across content pipelines.
Put together, the logic is: invest now to own the interface and distribution layer for AI music before incumbents (DAWs, labels, platform giants) lock it down.
But there’s a huge “but”: copyright, ethics, and cultural pushback
Money buys growth, not goodwill. The same scaling that makes Suno attractive to investors also amplifies the sticky problems everyone warned about.
-
Lawsuits and label pushback. Major labels and some artists have filed or threatened legal action alleging Suno trained models on copyrighted recordings without adequate consent or compensation. Those suits could force licensing deals, statutory damages, or changes in how models are built and monetized — each of which would hit margins and product claims.
-
Creative legitimacy. Critics argue that clicking “generate” is not the same as composing or performing — and that mass-producing music risks diluting cultural value and displacing paid work. Defenders say Suno’s tools augment creators and open new pipelines for musical discovery. That debate isn't just philosophical; it affects regulation, platform policy, and public perception.
-
Platform policy risk. Streaming platforms and distributors are already experimenting with AI content rules. If major streaming platforms restrict AI-generated content or demand specific metadata/labels and licensing, Suno’s commercial playbook may have to adapt quickly.
What will the $250M will likely fund?
Sizable rounds don’t just cover payroll. Watch for:
-
Product R&D at scale. Faster models, higher-quality vocals, longer-form music, and tools that enable customization and stems for mixing. Suno will invest in making outputs sound more polished and controllable — a route to premium tiers.
-
Legal and licensing teams. Expect Suno to beef up legal resources and negotiate licensing deals with publishers/labels — whether proactively or under pressure. That’s expensive but strategically necessary if they want to serve commercial customers.
-
Distribution and enterprise partnerships. Integrations with DAWs, game engines, ad platforms, and content platforms. With partners like NVIDIA (or NVentures’ involvement), cloud and inference costs could be optimized.
-
Creator economy features. Marketplaces, revenue shares, creator profiles, and tools for human+AI collaboration that let musicians add their voice and retain rights — those features can help reduce backlash and create a community that Suno can monetize.
How this affects musicians, producers, and indie labels
Short answer: the world changes, but you can steer your outcome.
-
New income streams, if you adapt. Musicians who learn to use AI as a co-producer, who sell curated stems, or who offer performance/production services on top of AI demos can create new revenue models. Tools that reduce routine work free time for higher-value creative tasks.
-
Competition for stock music and low-end beats. Expect price pressure when it comes to commoditized background music and simple beats. If you sell simple loops or generic beats, you’ll feel it first. Level up your offerings: unique sounds, signature production, bespoke performances.
-
Negotiation leverage with platforms. As AI content grows, musicians and rights-holders should demand better metadata, attribution, and share of revenue when models or platforms use their material. Collective bargaining, unionization, or industry-wide licenses may become necessary.
-
Credibility and branding matter more. Human-driven musical artistry becomes a differentiator. Fans will value authenticity, live performance, and artist narratives. Musicians who emphasize craft and story can maintain premium positioning.
What labels and publishers will do and why it matters:
Labels don’t want to be sidelined. Expect:
-
Licensing deals and partnerships with AI vendors — labels get paid and retain control over how recordings are used. That's a comfortable outcome for rights-holders if the economics work.
-
Litigation as leverage. Suits can force settlement talks or shape regulation. Even if litigation is slow, it creates a bargaining chip and public pressure.
-
New product offerings. Labels might launch their own AI tools or partner with platforms to offer branded, safe AI sounds — turning risk into opportunity.
Regulatory and platform questions to watch
-
Transparency mandates. Will platforms require AI-generated tracks to be labeled? How will attribution be tracked across distribution chains?
-
Compulsory licensing vs. negotiated deals. Regulators could push for clear rules on training data rights and compensation (analogous to sample-clearing regimes). That would change unit economics for every AI music company.
-
Data rights and model audits. Governments or industry bodies might demand audits of training datasets or new standards for demonstrable provenance.
Regulation is messy and slow — but high valuations like this make regulation inevitable, as the stakes are significant.
Practical takeaways for creators and music businesses
If you’re a musician, producer or indie label:
-
Experiment with AI but keep control. Use Suno and other tools to prototype quickly, but always decide which outputs become public and how you protect your brand and rights.
-
Treat AI as a collaborator, not a replacement. Use it to sketch ideas, expand arrangements, or create stems — then add your unique performance, production or vocal signature.
-
Document provenance and contracts. When licensing music, be explicit about AI usage in contracts and distribution metadata.
-
Learn to monetize beyond streaming. Offer stems, bundles, custom production, licensing for indie games/ads, or educational content. Diversify income.
-
Get savvy about IP. If you care about how models use existing works, follow industry litigation, and consider joining collectives lobbying for fair terms.
Final thought: The music ecosystem will adapt
Suno’s $250M raise is a landmark signal: investors expect AI to be central to the next wave of music creation and distribution. That will bring both opportunity and friction. Artists who adapt, insist on fair terms, and leverage the new tools creatively can thrive. The ones who don't will face a reshaped marketplace where commoditized sounds are abundant and differentiation is everything.
This moment is a crossroads. Massive capital accelerates capability; the public, artists and policymakers will decide the boundaries. For musicians: learn the tools, protect your craft, and keep making music that only a human can make — because even in a future of perfect algorithms, humans will still be the ones who turn sound into meaning.
Serato DJ: Host of new Crate management options lead its 4.0 update
Written by Sounds SpaceSerato DJ: A host of new Crate-management options lead its 4.0 update
Serato has just rolled out the biggest library overhaul in years. With the 4.0 update — released as Serato DJ Pro / Lite 4.0 and promoted through a public beta — the company has refocused its attention away from flashy effects and toward something DJs quietly crave: better ways to organize and access music. The result is a smarter, faster, and more flexible library system built around powerful crate management tools that make prepping sets and finding the right track on the fly far less frustrating.
Below, I unpack the most important additions, why they matter, and how DJs of all stripes can use the new library features to speed up workflows and boost creativity.
What’s new? The highlights
While 4.0 includes a variety of UX and performance tweaks, the marquee improvements center on crates — Serato’s core organizational unit. The most important additions are:
-
Crate Search — quickly find the crate you need by name. No more scrolling through long lists. (Pro-only feature).
-
Favorite Crates (Shortcuts / Pinning) — pin go-to crates to the top of your list for instant access during a set.
-
Crate Color—color—code crates using a right-click palette for visual separation and faster scanning.
-
Crate Sort — sort crates by alphabet, date created, or arrange custom orders.
-
Show Track in Crates — right-click any track and instantly see every crate the track lives in. Great for checking overlaps and curating transitions.
-
Streaming Tracks in Crates — add tracks from streaming services into crates alongside local files, reducing the friction of mixed-source sets.
-
Track Rating (emoji or stars) — rate tracks in the library using stars or playful emojis (watermelons, ghosts, etc.) for fast recall.
-
Crate Info — see crate metadata like total track count, total runtime, and file size in the status bar. Useful for timing sets.
-
Resizable Album Artwork & UI polishing — resize artwork in the library view, improved file relocation, safer USB ejection, and automatic analysis on import.
These features combine to make Serato’s library both more informative and more manipulable — two qualities that add up to real time saved during prep and performance.
Why crate management matters
Organizing music isn’t glamorous, but it’s foundational. A DJ’s library is the toolbox for every set; when tools are poorly organized, the performance suffers. The new crate system attacks several long-standing pain points:
-
Speed of access: Crate Search + Favorites removes the need to dig through nested folders or remember obscure crate names. In a busy club or festival, shaving seconds from a search can change the course of a set.
-
Visual scanning: Color coding and resizable album art let DJs build a visual language for their library — e.g., blue crates for house, red for peak-time weapons, green for vocal edits — speeding recognition when adrenaline’s high.
-
Context awareness: “Show Track in Crates” and Crate Info remove guesswork. Want to know if that remix is already in another playlist or how long a crate will last on the dancefloor? Now you can know instantly.
-
Bridging streaming + local files: Being able to store streaming tracks in crates reduces the friction when you mix paid downloads and on-demand tracks. It simplifies set building in hybrid workflows.
Put succinctly: better crate tools mean less cognitive load during prep and performance, and more headspace for creative decisions.
Practical workflows — how to use the new feature immediately.
1. Build a “go-to” rack: Create favorite crates for different parts of your set (warmup, peak, halftime, closer) and pin them. During a set, jump straight to the correct vibe without hunting.
2. Color code by energy & key: Use crate colors to encode energy levels or harmonic compatibility. For instance, green = compatible in key, orange = high-energy, blue = vocal tracks. Visual cues speed selection under pressure.
3. Use Track Rating as a preflight checklist: During practice, tag tracks with an emoji or star rating to indicate readiness: 5-star = performance-ready, 3-star = needs edits or cueing, 1-star = archive. Filter by rating when compiling a set.
4. Merge streaming & local crates for gig-specific builds: Add your subscription service tracks into crates with your local edits — then use Crate Info to verify runtime and track counts for set length planning.
5. Prep faster with Analyze on Import: Switch on automatic analysis so newly added tracks are ready to beatmatch and use stems/features without manual scanning. This is handy when hot-swapping files from USB drives.
Comparison: how Serato’s changes stack up
Competitors like Rekordbox and Traktor have offered various library conveniences for years. What’s notable about Serato 4.0 is the breadth of small, practical quality-of-life improvements delivered together — the kind of iterative polish that DJs notice in the field. Features like “Show Track in Crates” are reminiscent of functions some rivals implemented earlier, but Serato's approach to bringing multiple improvements at once (color, ratings, pinning, crate metadata) means you get a cohesive, modern library experience rather than a single standout tool.
Potential fixes & what to watch
No big software release is flawless. Early public beta threads point to a few things DJs should watch:
-
Crate Search in Lite: Some features (like crate search) are Pro-only. If you’re on Lite, check which tools you’ll miss before updating.
-
Library merging & external drives: while file relocation and hot-swap behavior have improved, DJs with complex external drive setups should test the new automatic merging carefully before relying on it in live shows.
-
Workflow muscle memory: any library overhaul may temporarily slow DJs who are used to older shortcuts — but the tradeoff is faster, clearer workflows after a short adjustment period.
Serato has emphasized community feedback during the public beta, so many small issues will likely be patched quickly. If you rely on Serato for gigs, don’t upgrade mid-gig season without confirming stability in your specific setup.
Pricing and availability
Serato released 4.0 as a public beta initially, with the company listing the update and downloadable builds on its site. Serato DJ Pro remains a paid product with purchase and subscription options, while Serato DJ Lite continues to be free — though some of the new features are Pro-only. Check Serato’s official downloads and release notes for the exact availability and the latest stable build.
Final thoughts — a library update that matters
Serato DJ 4.0 isn’t headline-grabbing because it adds a new effect or a flashy performance feature. It matters because it addresses the mundane, everyday work of DJs — organizing, finding, and prepping music — with practical tools that shave time off routine tasks and reduce friction in the creative process. For working DJs, producers who moonlight as selectors, mobile DJs, and club residents, those minutes reclaimed from searching and guessing add up to better sets and less stress.
If you’re a Serato user: test the public beta in a safe environment, try the crate color and favorites system, and begin using the rating system to build a library taxonomy that works for you. If you’re still on another platform, Serato 4.0 narrows the gap in library ergonomics and gives you one more good reason to reevaluate your DJ software workflow.
SoundCloud's Game-Changing Leap: Artists Claim 100% of Royalties in New Era of Creator Empowerment
Written by Sounds SpaceSoundCloud's Game-Changing Leap: Artists Claim 100% of Royalties in New Era of Creator Empowerment
By Elena Vasquez, Music Business Daily, November 14, 2025
In a seismic shift that's sending ripples through the music industry, SoundCloud has unveiled a subscription model that hands artists the full reins on their earnings. Announced on October 30, 2025, the platform's new "All-In-One Artist Subscription" allows independent creators to retain 100% of royalties from music distributed across more than 60 streaming services, including Spotify, Apple Music, TikTok, and YouTube Music. No more 20% cuts from the distributor—every penny from streams, downloads, and syncs goes straight to the artist. This isn't just a tweak; it's a declaration of war on the opaque royalty structures that have long plagued emerging talent.
For years, SoundCloud has been the digital sandbox where bedroom producers and viral hopefuls test their sound, amassing over 175 million tracks and a fiercely loyal community of "music obsessives." But monetization? That was the rub. Under the previous SoundCloud for Artists Premier plan, creators kept about 80% of distribution royalties after the platform's slice, a model that mirrored competitors like DistroKid and TuneCore but still left many feeling shortchanged in an era where a single stream nets less than a tenth of a cent. Now, with this update effective for payouts starting November 2025, SoundCloud is betting big on transparency and artist-first economics to reclaim its throne as the go-to hub for indie music.
The timing couldn't be more poignant. As major labels consolidate power amid AI-driven disruptions and TikTok's algorithm-fueled virality, independent artists are squeezed harder than ever. A 2024 MIDiA Research report pegged the average indie artist's annual streaming revenue at just $1,200—barely enough for a month's rent in Brooklyn. SoundCloud's move feels like a lifeline, especially for the platform's core demographic: Gen Z and millennial creators who upload demos, remixes, and EPs without the safety net of a label advance.
From Uploads to Uplifts: SoundCloud's Evolution
SoundCloud's story is one of scrappy innovation turned industry staple. Founded in 2007 by Swedish entrepreneurs Alex Ljung and Eric Wahlforss, it started as a simple audio-sharing site for sound designers and DJs. By 2013, it had exploded into a discovery engine, launching careers like those of Billie Eilish (whose early tracks racked up millions of plays) and Post Malone. But growth brought growing pains: a 2017 funding crunch nearly sank the company, forcing it to pivot toward monetization tools like repost chains and basic royalties.
Enter SoundCloud for Artists in 2018, which bundled distribution, analytics, and promo tools into tiered plans. The Premier tier, at $99 annually (or about $8.25 monthly), promised 80% royalties on distributed tracks—a step up from free uploaders but still trailing the "zero-cut" dreams peddled by upstarts like Amuse. Artists grumbled about hidden fees, payout delays (net 60 days), and the platform's infamous "repayable advances" that could claw back earnings if streams tanked.
Fast-forward to 2025, and SoundCloud's under new stewardship after a 2023 acquisition by SiriusXM, which injected $200 million into creator tools. CEO Michael Herd, a former Spotify exec, has steered the ship toward "sustainable independence," citing data that 70% of SoundCloud's 40 million monthly users are undiscovered artists seeking direct fan connections. The All-In-One plan is the culmination of this vision: for that same $8.25 monthly fee, artists get unlimited distribution, built-in monetization, and zero platform cuts on royalties, Fan Support tips, or merch sales.
Digging into the fine print reveals a model built for scale. Tracks uploaded via the plan beam out to 60+ DSPs (digital service providers) instantly, tapping SoundCloud's 76 million global users for immediate plays. Royalties? 100% artist-retained, with only a nominal payout processing fee (as low as $0.50 per transaction via partners like PayPal or bank transfer) once earnings hit $25. No upfront costs for vinyl-on-demand releases—artists can press limited-edition runs without bulk minimums—or for linking merch stores, where SoundCloud takes zilch from sales.
It's not without caveats. The 100% share applies only to new distributions post-November 2025; legacy earnings might still reflect the old 80/20 split. And while SoundCloud touts "no hidden fees," tax forms and two-factor authentication are mandatory for payouts, a nod to fraud prevention amid rising cyber threats in creator economies. Still, for a Nashville trap producer like DXRKNOVA, who credits the platform for doubling her monthly earnings through targeted promo, it's a no-brainer upgrade.
The All-In-One Blueprint: Features That Fuel Futures
At its heart, the All-In-One Subscription is a Swiss Army knife for the solo artist hustler. Distribution is the star: upload once, and your track hits Spotify playlists, Apple Music's algorithmic feeds, and even niche spots like Deezer or Tidal. SoundCloud's edge? Its proprietary "Instant Reach" algorithm, which funnels new releases to superfans and tastemakers within hours, potentially spiking plays before competitors even index the file.
Monetization layers on seamlessly. Fan Support lets listeners tip directly—every cent to the artist, displayed as a "top supporter" badge on profiles to foster loyalty. Merch integration is plug-and-play: link your Teespring or Bandcamp store, and SoundCloud's storefront widget handles discovery with zero fees. For physical die-hards, the vinyl-on-demand tool turns digital hits into 7-inch pressings, shipped globally without inventory risks. "It's like having a label without the middleman," says Brooklyn DJ nextdimensional, whose EP Portal saw a 40% revenue bump after bundling streams with fan tips and limited-edition merch.
Analytics round out the package: real-time insights on listener demographics, stream sources, and even AI-powered "mood mapping" to suggest collab opportunities. At $99/year, it's competitively priced against DistroKid's $22.99 annual unlimited plan (which takes 0% but lacks built-in audience tools) or TuneCore's $29.99/single (with varying cuts). SoundCloud's pitch: Why fragment your workflow when one dashboard handles discovery, distribution, and dollars?
Critics might quibble over the subscription lock-in—free users still face upload limits and no distribution—but for pros, it's a steal. Early adopters report 15-20% earnings lifts, per internal SoundCloud data, thanks to the platform's 320 million tracks-strong library acting as a perpetual promo engine.
Voices from the Upload Trenches: Artist Reactions Pour In
The announcement lit up X like a viral remix. "SoundCloud the MVP baby 100% Royalties >," tweeted electronic artist Siren Diamant, capturing the euphoria among bedroom producers. South African DJ Tezz echoed the sentiment: "Major Update for Independent Artists... keep 100% of their royalties, not just on SoundCloud, but also on Spotify, Apple Music, TikTok and more." NME amplified the buzz, headlining it as a "100 per cent of distribution royalties" win, while Mixmag hailed it for dance music creators.
Not everyone's popping champagne uncorked. Your EDM's analysis cut through the hype: "When each stream is still worth a fraction of a cent, this indicates that artist's 'keeping' their royalties may not mean much at all." Indie vocal coach Randy Morano noted the fee: "100% royalties for creators (minus an $8.25/mo fee). A huge shift toward artist-first monetization." And Charm Cracker, a self-produced multi-instrumentalist, plugged referral codes, underscoring the grassroots uptake: "They are bypassing the competition... Spotify is not working for me, really."
Interviews with early switchers paint a nuanced picture. "I've been grinding SoundCloud since 2019, scraping by on 80% scraps," shares LA-based rapper J. Mira, who dropped her debut mixtape Echo Chamber last week. "Now? That extra 20% covers my plug-ins. It's not riches, but it's respect." For global acts like Romio, an Italian Web3 musician blending NFTs with beats, the plan aligns with blockchain's decentralization ethos: "Big news for indie artists! 100% of distribution royalties for tracks released via SoundCloud."
Even skeptics concede the momentum. Rolling Stone Brasil covered it in Portuguese, emphasizing accessibility for non-English markets, while Exclaim! touted it as a North American boon. As one anonymous SoundCloud vet put it: "It's the push we needed to treat music like a business, not a hobby."
Ripples in the Royalty Pool: Broader Industry Shifts
This isn't happening in a vacuum. The music biz is in flux: Universal Music Group's 2024 TikTok standoff exposed streaming's fragility, while AI tools like Suno democratize production but dilute per-stream value. Platforms like Bandcamp thrive on direct-to-fan models (100% artist cuts, minus 10-15% fees), but lack distribution scale. SoundCloud bridges that gap, potentially siphoning users from Spotify for Artists (70% royalties post-label cuts) or Apple's $9.99 pro tier.
Economically, it's a calculated risk. SoundCloud's revenue—pegged at $300 million annually—leans on ads and subs, not distribution vig. By waiving the 20% (estimated $50 million yearly), they're banking on volume: more uploads mean more engagement, data sales, and ecosystem lock-in. Analysts at Billboard predict a 25% subscriber bump by Q2 2026, especially if viral hits like the platform's past (e.g., Lil Nas X's "Old Town Road") recur.
For marginalized creators—women, POC, LGBTQ+ artists who comprise 60% of SoundCloud's base per a 2025 USC Annenberg study—this levels the field. No more gatekept advances or predatory deals; just tools to build empires. "We're not just streaming; we're stewarding," Herd said in a company blog, though exec quotes remain sparse.
A Symphony of Change: What's Next?
As November's payouts roll out—first 100% checks hitting December—watch for copycats. Will DistroKid drop its unlimited model to zero cuts? Could Spotify counter with fan-direct royalties? SoundCloud's bet is bold: in a $28 billion streaming market dominated by three majors, empowering the underdogs might just rewrite the score.
For now, artists like J. Mira are tuning up. "This feels like the remix we deserved," she says. In an industry often accused of lip-syncing equity, SoundCloud is finally letting creators take the mic—and keep the mic drop.
The previously leaked Behringer BQ-10 analogue sequencer has officially launched — here’s everything you need to know
Written by Sounds SpaceThe previously leaked Behringer BQ-10 analogue sequencer has officially launched — here’s everything you need to know
Behringer has quietly moved another long-rumoured product from the “leak” pile to the shop window: the BQ-10 analogue sequencer. If the name and styling look familiar, that’s because the BQ-10 is a modern reimagining of the classic Korg SQ-10 design — rebuilt with contemporary connectivity and Eurorack friendliness in mind. Whether you’re a vintage-sequencing obsessive, a modular tinkerer, or a bedroom producer who loves hands-on pattern creation, the BQ-10 is designed to slot into many setups without demanding a PhD in patching.
What the BQ-10 actually is
At its heart, the BQ-10 is an analogue, voltage-controlled step sequencer giving you 24 total steps arranged as three 8-step lanes. Channels A and B can operate independently as two 8-step sequences or be combined into a single 16-step melodic lane; channel C remains dedicated as an 8-step control lane for filter, amplitude or other timbral modulations. That flexible three-lane architecture is what gives the BQ-10 a lot of creative mileage: you can have two melodic lines and a separate automation lane running at once, or stitch lanes together for longer phrases.
Physically, it’s presented as a desktop unit that can be removed and mounted into Eurorack — useful if you want both a tidy desktop workstation and rack integration. The layout is classic: knobs and switches for each step, portamento and duty-cycle/gate controls, dedicated sync and CV jacks, plus MIDI for integrating with modern gear.
Key specs and features (quick list)
-
24 analogue steps arranged as 3 × 8-step lanes (A, B, C). A+B = 16 steps when combined.
-
CV and gate outputs for full voltage control; per-lane voltage range switches for compatibility with different synth standards.
-
MIDI & USB — clock sync and MIDI note/Gate output for DAW/hardware integration.
-
Clock generator (internal), adjustable roughly from very slow to very fast — and CV-controllable, plus external clock input for sync.
-
Per-channel portamento (glide) and a global duty-cycle control to shape gate length.
-
Eight playback modes for A/B, including series, parallel, alternating and random options to vary sequence behaviour.
-
Desktop & Eurorack capable — the unit detaches from its desktop case for rack mounting and lists a modest HP size on modular databases.
Those bullet points are the essence of what delivers the classic SQ-10 vibe with modern usability. The BQ-10 keeps the tactile, real-time programming workflow that analogue sequencers are loved for, while making them far less awkward to integrate with modern setups.
How it differs from the Korg SQ-10 (and why that matters)
Behringer’s take is recognisably inspired by the Korg SQ-10, but with a few practical updates. The addition of MIDI/USB and CV-friendly voltage range switches is the biggest functional upgrade — it means the BQ-10 can talk to vintage Hz/V synths, modern 1V/oct modular gear, and computer setups without a lot of external conversion. The dedicated third lane (C) being explicitly pitched for modulation tasks is also a modern convenience: rather than patching an external sequencer to a VCF or VCA, you can build timbral changes into your patterns from the same interface.
There are aesthetic and ergonomic tweaks too: Behringer has chosen a darker control surface and wood end-cheeks for a retro but contemporary look, and the ability to run as a standalone desktop box or snap it into Eurorack opens it up to more workflows than the original hardware typically supported. For many users that “plug and play” modular compatibility is a decisive factor.
Real-world usability: what it’s good at
-
Hands-on compositional jamming. The step pots and physical transport controls invite experimentation — perfect for quickly sketching motifs or improvising live sequences.
-
Modulation-driven patches. Use channel C to twist filter cutoff, VCA level, or any CV-controllable parameter on your synth to create evolving grooves without external CV routing.
-
Integration with DAWs/hardware. MIDI/USB means the BQ-10 can either be the central timing source for hardware rigs or slotted into a software workflow; conversely, the CV outputs let it remain a pure-analogue controller when you want that sound.
-
Modular-friendly sequencing. If you build with Eurorack, the detachable form factor and full CV outputs let the BQ-10 live as a compact step engine in someone’s case.
It’s not trying to be a sequencer DAW replacement; its charm and strength are immediacy, character and hands-on performance — the traits that made step sequencers indispensable in classic electro and electronic music.
Pricing and availability
Behringer has listed the BQ-10 at an accessible price point — around US$149 / £149 in many retailers’ listings — which positions it as an affordable entry into hardware sequencing and modular-adjacent setups. Preorders have appeared across multiple dealers, and Behringer’s own product page and press materials indicate the unit is now officially launched and shipping through retailers in batches. Expect initial demand to outpace supply for a while; popular Behringer launches tend to sell quickly.
Who should consider buying one?
-
Beginners to hands-on sequencing who want a quick, low-cost way to learn analogue sequencing concepts without a full modular investment.
-
Modular users seeking an inexpensive, compact step engine to run melodic and modulation lanes in their rack.
-
Live performers and jam musicians who want a tactile pattern machine that responds well to on-the-fly changes.
-
Producers after character — if you like the slightly unpredictable, squelchy charm of analogue CV and manual knob tweaks, the BQ-10 brings that to low cost.
If your main studio workflow is tightly DAW-centric and you need deep pattern memory, song mode, step automation recall and long-form sequencing, you may find the BQ-10’s analogue, patch-centric workflow less convenient than a dedicated sequencer with full CV-to-MIDI mapping and presets. But for raw hands-on sequencing, it’s hard to beat for the price.
A few practical caveats and community reactions
-
Purists will notice differences between the original Korg SQ-10 and Behringer’s implementation — some step counts and nuanced behaviours vary. That said, many users welcome the modern connectivity, even if small purist gripes appear on synth forums.
-
Supply timing. As with many new Behringer launches, availability may vary regionally, and some retailers show preorder lead times — so if you need one immediately, double-check local stock and shipping windows.
Final thoughts — is the BQ-10 worth the buzz?
The BQ-10 is exactly what it promises to be: a relatively inexpensive, hands-on analogue sequencer that wears its heritage on its sleeve while adding pragmatic modern features. For anyone who’s ever loved the immediacy of knob-per-step sequencing and wanted to plug it directly into modular or MIDI setups without extra interface boxes, the BQ-10 is a compelling option. At its price point and with its combination of CV, gate, MIDI and Eurorack compatibility, it looks set to be a popular choice for curious beginners and seasoned hardware lovers alike.
If you want to dive deeper before buying, check the Behringer product page for full specs and manuals, watch hands-on demo videos from retailers and reviewers to hear how it behaves with your kind of synth, and read user threads for early impressions on build quality and integration tips. Happy sequencing!
Pandora Launches 2 AM: A Bold New Home for DJ Set Culture
Written by Sounds SpacePandora Launches 2 AM: A Bold New Home for DJ Set Culture
In a streaming market saturated with algorithm-driven playlists and on-demand single tracks, Pandora is staking a new claim in the world of electronic music with 2 AM, its brand-new station and DJ takeover series. The idea is simple yet ambitious: to elevate DJ sets as curation, offering exclusive mixes and artist-led programming rather than relying purely on automated playlists. The official positioning? 2 AM is being billed as “the home for DJ set culture on Pandora.” In this post, we’ll dig into what 2 AM is, why it matters, how it’s different from existing DJ/dance streaming options, and what challenges and opportunities lie ahead.
What Exactly Is 2 AM?
At its core, 2 AM is a curated music channel and a rotating “DJ takeover” format in which established DJs handpick tracks, mix them into playlists, and deliver hosted intros or commentary. The first wave of takeovers comes from big names in electronic music — such as Above & Beyond, John Summit, and Don Diablo — who will lend their own sensibilities to the station.
Pandora intends for 2 AM to cover the full spectrum of dance music: from EDM and tech house to nu-disco, chillout, and beyond. The playlists will be refreshed regularly, ensuring a steady stream of new and evolving content.
What sets 2 AM apart from Pandora’s existing dance stations is the emphasis on the DJ as curator, not an algorithm. Rather than simply mirroring what listener behavior suggests, 2 AM intends to reflect what DJs are actually spinning in clubs and festival sets — the tracks that move crowds, not just the ones data suggests you might like.
Why It Matters — For Listeners, Artists & DJs
1. Reconnecting Listeners to the Community and Energy of DJ Culture
One criticism of much digital music consumption is that it strips away context. You might listen to a track, but you don’t hear why it was chosen or how it works in a larger DJ set. With 2 AM, Pandora is attempting to re-embed that context: the intros, the track transitions, the DJ’s narrative flow.
For passionate fans of electronic music, 2 AM offers something more immersive than a static playlist. It’s closer to being at a club, but in your ears. It gives listeners an inside track on what’s trending in dancefloors globally — not just what an algorithm infers from your history.
2. Spotlighting DJs as Creative Curators
Many streaming services downplay the role of the DJ, reducing them to a single “official artist” credit or dropping them from playlists entirely. 2 AM flips that paradigm. The DJ becomes the central creative voice: you listen through them, not just to them.
For those DJs whose work centers on live sets, remixes, or hybrid genre explorations, 2 AM offers a new canvas: curated radio-style shows, unique intros, and perhaps even a way to test unreleased edits or blends.
3. An Avenue for Deeper Discovery
Another benefit of 2 AM is its potential as a discovery engine. Instead of being guided by algorithmic similarity suggestions, listeners might follow a DJ’s personal tracklist and stumble into underground or emerging artists they wouldn’t have found otherwise. This gives Pandora another differentiator in a market awash with “similar to what you already listened to” stale recommendations.
What Makes 2 AM Different from Other Platforms?
DJ-centered content isn’t new — platforms like Mixcloud, SoundCloud, or YouTube host DJ sets and live mixes. But 2 AM offers something that many of those platforms lack:
-
Native integration with Pandora’s streaming infrastructure. No need to bounce between apps; the station lives within Pandora’s ecosystem.
-
High production standards: These aren’t user-uploaded club sets; they’re artist-approved, hosted, and styled for streaming consumption.
-
Scheduled rotation + freshness: Rather than one static mix, 2 AM will continuously refresh with new DJ takeovers, shifting styles, and seasonal changes.
-
Curator identity built in: The DJ is not anonymous. Their personality, voice, and taste become a part of the listening experience.
In short, 2 AM is not just “DJ sets in Pandora” — it’s DJ programming inside a streaming radio experience.

Risks, Hurdles & Considerations
As bold as 2 AM sounds, Pandora will face challenges in making it work sustainably:
Metadata & Rights Management
DJ sets often include a blend of tracks, edits, remixes, or bootlegs. Licensing that material for streaming — especially when included in continuous, curated mixes — can be tricky. Pandora will need robust metadata, royalty accounting, and legal agreements to ensure rights are properly cleared.
Balancing Mass Appeal with Niche Integrity
The broader Pandora audience might not always gravitate toward extended DJ sets or deeper cuts. Pandora must strike a balance between accessibility (making 2 AM inviting for casual listeners) and authenticity (sustaining the respect of die-hard EDM/dance fans). If the mixes lean too “safe,” they risk alienating the core audience; lean too underground, and they might lose mainstream appeal.
Scheduling & Rotation
To keep listeners coming back, Pandora will need to manage scheduling carefully — rotating DJ takeovers, refreshing playlists, and introducing new voices regularly. Stagnation is a risk if the same DJs or subgenres dominate for too long.
Measuring ROI & User Engagement
How will Pandora measure success? Will listener metrics (hours streamed, listener retention) justify the investment in curated DJ content versus algorithmic programming? Pandora will likely monitor how 2AM engagement compares with its existing stations.
What to Watch Going Forward
Here are a few signals I’ll be keeping an eye on:
-
New DJ lineups — will Pandora bring in more underground or rising talent, not just established names?
-
Listener data — how long do users stay tuned? Is 2 AM driving new registrations or retention?
-
Community features — will Pandora add the ability to “follow DJs,” get notifications about upcoming takeovers, or explore past sets?
-
Cross-promotion with events — could Pandora link 2 AM DJ sets to live festival streams or event tie-ins?
-
Artist exclusives & premieres — the ability to debut exclusive remixes or edits through 2 AM could become a talking point and competitive edge.
Conclusion: Why 2 AM Could Be a Turning Point
Pandora’s 2 AM is more than just another streaming channel — it’s a bet on returning agency to the human curators in music. In an era where algorithmic discovery dominates, giving DJs a prominent voice is a statement: dance music is about more than data; it’s about taste, context, and narrative flow.
If Pandora can get the licensing, rotation, and listener onboarding right, 2 AM could become a meaningful bridge between club culture and everyday listening. It offers fans a new way to engage, DJs a new platform to express, and Pandora a fresh point of differentiation in a crowded streaming market.
As 2 AM rolls out and evolves, it could redefine how we experience DJ sets — not just as occasional live events, but as serialized audio journeys you can tune into nightly, weekly, or whenever the mood strikes.
Spotify’s “SongDNA”: A New Era of Music Discovery Through Credits
Written by Sounds SpaceSpotify’s “SongDNA”: A New Era of Music Discovery Through Credits
In a world where music streaming is dominated by algorithms, playlists, and endless recommendations, Spotify seems to be taking a fresh, human-centric approach. Recently, tech insiders discovered that Spotify is developing a new feature called “SongDNA.” This potential update could change how listeners discover music — not just by sound or genre, but by the people behind the songs.
Although the company hasn’t officially announced it yet, clues buried in Spotify’s app code hint at an exciting new direction for music discovery — one that highlights the unsung heroes of every track: the writers, producers, and engineers who make the magic happen.
What Is SongDNA?
SongDNA appears to be a feature in development that will give users the ability to explore music through its credits. Instead of focusing purely on the artist or the song’s title, Spotify’s new approach aims to let users dive into the DNA of a track — the creative contributors who shaped it.
This discovery came from Jane Manchun Wong, a well-known app researcher who frequently uncovers unreleased features in popular apps. Wong found references to “SongDNA” inside Spotify’s code, along with screenshots of what appears to be a new interface that lists each contributor on a track and connects their work across other songs.
For example, if you were listening to a song produced by Mark Ronson, SongDNA might let you tap his name to see all the tracks he’s ever produced — across artists, genres, and eras. Imagine discovering connections between your favorite songs and realizing they share the same songwriter, engineer, or session musician. That’s the kind of experience SongDNA could deliver.
A Shift from Algorithms to Humans
For years, Spotify’s success has been driven by its algorithmic curation — playlists like Discover Weekly or Release Radar use data and behavior patterns to serve you the next best track. But SongDNA hints at a more organic and human discovery path, one rooted in creativity, connection, and collaboration.
It represents a philosophical shift. Music discovery has long been about what you like. SongDNA introduces the idea of who you like — the creative fingerprints behind your taste. Instead of browsing by genre, you could browse by creator network.
This approach could redefine how fans think about their favorite songs. Many listeners don’t realize that some of today’s biggest pop hits were written or produced by the same handful of creative masterminds. For example, Max Martin has written dozens of chart-topping hits for Britney Spears, The Weeknd, Taylor Swift, and Katy Perry. SongDNA could make those invisible connections instantly visible — revealing the invisible threads that tie the music industry together.

Why This Matters for Artists and Producers
The potential benefits of SongDNA go far beyond fans. For songwriters, producers, and engineers, this could be a game-changer.
One of the biggest frustrations in the modern music industry is the lack of visibility for the people who work behind the scenes. While major artists get the spotlight, songwriters and producers often remain anonymous to the average listener — even though their contributions shape the sound of entire generations.
SongDNA could finally bridge that gap. By integrating credits directly into the discovery experience, Spotify would give recognition where it’s due — making it easier for industry professionals to build a public portfolio within the world’s largest streaming platform.
This transparency could also create new networking opportunities. A producer might find a new vocalist to collaborate with. A songwriter could discover others who share a similar creative style. Even music fans could use this feature to better understand the craftsmanship behind their favorite hits.
How SongDNA Could Work
While Spotify hasn’t confirmed any details, leaks suggest that SongDNA might include:
-
A detailed credit breakdown: listing every person involved in the song’s creation — from writers and producers to mixers and mastering engineers.
-
Clickable profiles: allowing users to tap a contributor’s name to see all the other songs they’ve worked on.
-
Connected discovery: recommendations based on shared contributors rather than listening habits — e.g., “Other tracks produced by Metro Boomin.”
-
Interactive visual design: perhaps showcasing “creative webs” connecting different artists and their teams.
Essentially, Spotify could evolve into an IMDB for music — a place where you can trace every creative connection in the industry.
The Competition: Tidal and Beyond
Spotify isn’t the first streaming platform to explore song credits. Tidal, for example, has long championed detailed credit listings, allowing users to view who contributed to a track. However, SongDNA seems to go beyond simple listings — it’s designed to make those credits discoverable.
Instead of credits being buried in a menu, they could become the foundation for new recommendations. This would give Spotify a fresh discovery edge — a more narrative, human layer to its data-driven ecosystem.
It also aligns with a broader industry trend: recognizing creators. From TikTok tagging original sound creators to YouTube’s music metadata updates, platforms are realizing the importance of crediting every contributor. SongDNA fits perfectly into this cultural moment.
Challenges Spotify May Face
As exciting as SongDNA sounds, it won’t be easy to implement. There are a few major hurdles Spotify must overcome:
-
Incomplete metadata – Many songs, especially older ones or indie releases, lack accurate or complete credit data. This could make the feature inconsistent at first.
-
Licensing and data integration – Spotify may need to partner with publishers, PROs, and metadata providers to gather and verify this information.
-
User interface complexity – Displaying all this data clearly and attractively will require careful UX design to avoid overwhelming users.
-
Adoption and awareness – Listeners need to care about credits. While hardcore music fans will love this, casual users may not explore it unless Spotify promotes it effectively.
If the company pulls it off, however, SongDNA could become one of Spotify’s most valuable discovery tools in years.
The Bigger Picture: Rediscovering the Human Side of Music
In a time when AI and automation dominate the headlines, SongDNA feels refreshingly human. It reminds us that music is not just content; it’s a collaboration — a product of emotion, creativity, and teamwork.
By surfacing the people behind the songs, Spotify could spark a cultural shift: encouraging fans to value who made the music as much as who performed it. For upcoming artists, that visibility could translate into recognition, respect, and new opportunities.
If Spotify launches SongDNA, it could redefine how the next generation interacts with music — not just consuming it, but exploring its creative lineage.
Final Thoughts
At the moment, SongDNA remains in development, and Spotify has yet to make an official statement. However, the discovery of this feature in Spotify’s app code — confirmed by multiple sources like TechCrunch and Digital Trends — indicates that the company is serious about exploring this direction.
Whether it launches next month or next year, one thing is clear: Spotify wants to make music discovery more personal, more connected, and more human.
If SongDNA becomes a reality, it could mark the next great evolution in streaming — turning every play into a journey through the creative universe of the artists, writers, and producers who make music come alive.
More...
How to Go Viral on YouTube as a Musician: The Complete 2025 Guide to Explosive Growth
Written by Sounds SpaceHow to Go Viral on YouTube as a Musician: The Complete 2025 Guide to Explosive Growth
In today’s digital era, YouTube is the most powerful launchpad for musicians — a platform where talent meets opportunity. It’s where unknown artists become global sensations overnight, where creativity meets algorithm, and where one video can change your life forever.
But here’s the truth: going viral on YouTube isn’t just luck. It’s a careful blend of strategy, timing, creativity, and consistency. If you’re an independent musician dreaming of viral fame, this guide will teach you how to break through the noise, attract massive attention, and build a fanbase that lasts.
🎬 1. Understanding the YouTube Algorithm (2025 Edition)
The YouTube algorithm isn’t your enemy — it’s your biggest ally once you understand how it works.
In 2025, YouTube’s algorithm focuses on three key things:
-
Watch time: How long viewers watch your video before leaving.
-
Engagement: Likes, comments, shares, and saves.
-
Viewer satisfaction: Whether people come back for more (measured through return visits and session time).
👉 What this means for you as a musician:
You can’t just drop a music video and expect it to go viral. You need to make people stay, react emotionally, and want to share it.
Action Tips:
-
Start your video strong — the first 10 seconds are critical.
-
Use storytelling: show behind-the-scenes clips before the drop or performance.
-
Add captions for accessibility and better retention.
-
Encourage engagement by saying things like:
“Comment your favorite line,” or “Tag a friend who needs to hear this!”
The longer people stay on your video and interact, the more YouTube promotes it.
🎵 2. Create a Magnetic Brand Identity
Before you go viral, people must know who you are and what you stand for. Your brand is what separates you from every other talented artist trying to make it.
Ask yourself:
-
What emotion or vibe does my music bring?
-
What colors, visuals, and energy represent me?
-
How do I want people to feel when they watch my content?
For example:
-
The Weeknd built mystery and cinematic storytelling.
-
Doja Cat mixes humor, chaos, and creativity.
-
Billie Eilish blends dark aesthetics with vulnerability.
Action Tips:
-
Use the same color palette, fonts, and logo across your videos and thumbnails.
-
Include a consistent intro/outro in every video.
-
Show your personality in short-form content between big releases.
Your music might get people’s attention — but your identity keeps them watching.
🎥 3. Optimize Your Music Videos for the Algorithm
You can’t go viral without optimization. Even the best song will struggle if YouTube can’t understand or recommend it properly.
a) Titles That Grab Attention
Your title should include:
-
Your artist name
-
The song name
-
A keyword or hook to attract non-fans
Example:
Instead of “Dreams – Official Video,”
Try “Dreams – The Dream X (A Deep House Journey You’ll Feel in Your Soul)”
b) Eye-Catching Thumbnails
Your thumbnail is your first impression. It must make people feel something.
-
Use close-up facial expressions (emotion sells)
-
Keep the text under 4 words
-
Add contrast (dark background, bright subject)
c) Tags, Descriptions, and Hashtags
Include related artists, genres, and moods in your tags and descriptions.
Example for a tech house artist:
Tags: tech house 2025, club music, underground dance, house mix, subaholic
Your description should tell a mini story about the track — not just credits.
Include a call-to-action:
“Stream this track on Spotify”
“Follow for weekly studio sessions”
“Comment where you’re listening from!”
These help YouTube categorize your video and suggest it to similar audiences.
⚡ 4. Use Shorts to Explode Your Reach
YouTube Shorts are the fastest route to virality in 2025.
They get millions of views because they’re short, addictive, and pushed heavily by YouTube’s algorithm.
What to Post:
-
Snippets of your best hooks or drops
-
Studio sessions or making-of moments
-
Reaction videos to your own track
-
Mashups, remixes, or duets
-
Before & after mixing/mastering clips
Why It Works:
-
Shorts are discoverable even by people who’ve never heard of you.
-
You can upload daily without overwhelming your audience.
-
Viral Shorts can funnel new fans directly to your full music video or Spotify page.
Pro Tip:
End your Shorts with a hook like:
“Full track out now — check the pinned comment.”
This drives curious viewers straight to your main video or streaming link.
🌍 5. Collaborate and Cross-Promote
Collaboration is the shortcut to virality.
When two musicians (or even creators from different niches) collaborate, they share audiences, creating exponential exposure.
Collaboration Ideas:
-
Feature another artist on your track
-
Invite YouTubers or dancers to use your song in their videos
-
Remix a trending song with your unique twist
-
Perform live sessions with local artists or producers
Example:
If a travel vlogger uses your chill house track in their video — and that video hits 1M views — your name and song will get thousands of organic clicks.
Never underestimate the power of shared audiences.
🔍 6. Leverage Trends Without Losing Authenticity
Jumping on trends is one of the fastest ways to ride the viral wave — but you must do it your way.
What to Do:
-
Use trending sounds or challenges, but remix them in your style.
-
Add your instrument or vocals to trending songs.
-
React to music news or other viral moments in your niche.
For instance:
If there’s a viral Afro House trend, you could post:
“What if I made this viral Afro House beat with only nature sounds?”
That’s unique, trend-driven, and creative — a perfect viral combo.
💬 7. Engage Like a Human, Not a Brand
The most viral musicians on YouTube act like real people, not companies.
Reply to comments, thank fans, and be part of your community. When fans feel a connection, they’ll share your videos out of loyalty — not just because of the music.
Engagement Tips:
-
Reply to at least 10 comments on each video.
-
Pin a fun or emotional comment at the top.
-
Ask interactive questions like:
“Where should I perform next?”
“Who wants this track on Spotify?”
The more engagement you spark, the more YouTube boosts your video’s reach.
🔊 8. Tell Stories, Don’t Just Drop Music
Every viral song has a story behind it — and that’s what people remember.
Was your song inspired by heartbreak, travel, a late-night session, or a random voice note?
Tell that story visually or verbally in your content.
Examples:
-
Upload a video titled “The Story Behind My Viral Track”
-
Share clips from your songwriting process
-
Post a mini documentary about your journey
People connect with stories, not just songs.
Your story makes your music human — and humans go viral.
🧠 9. Master Consistency and Strategy
Virality isn’t about one lucky video. It’s about consistent quality.
Many artists blow up after their 10th or 20th upload — not their first.
Suggested Schedule:
-
1 full music video per month
-
2-3 Shorts per week
-
1 behind-the-scenes or vlog-style upload every two weeks
Use YouTube Analytics to track:
-
Which videos get the most clicks (CTR)
-
Which videos get longer watch time
-
When your viewers are online
Then double down on what works.
📈 10. Promote Outside of YouTube
YouTube rewards traffic from other platforms.
If people click your YouTube link from Instagram, TikTok, or Twitter, YouTube sees that as external validation and boosts your ranking.
Strategies:
-
Post a 15-second teaser on Instagram Reels and direct followers to YouTube.
-
Use TikTok trends to tease your song.
-
Share your video link in music Reddit communities (like r/WeAreTheMusicMakers).
-
Send your video to music blogs or Discord groups for early traction.
Even 1000 views from loyal followers can kickstart the YouTube algorithm.
🔥 11. Create a Viral Moment on Purpose
You can plan virality.
Look at artists like Oliver Tree, Rosalía, or Tate McRae — their “viral” moments were calculated.
Here’s how to engineer your own:
-
Pick a unique visual hook (an outfit, dance, or camera trick).
-
Pair it with your strongest chorus or beat drop.
-
Encourage participation — ask fans to remix, dance, or duet your video.
-
Add humor or surprise — people love the unexpected.
Example:
If your track has a powerful drop, film a dramatic visual (like smashing something symbolic or changing outfits instantly). These become shareable memes and help your music spread.
💡 12. Don’t Chase Views — Build a Movement
The goal isn’t just to go viral once. It’s to build a loyal audience that stays.
When you focus only on views, you burn out fast. But when you focus on impact and connection, virality becomes a side effect.
Your content should:
-
Inspire
-
Entertain
-
Offer value (like tutorials, motivation, or authenticity)
Virality fades — community lasts.
🏁 Final Thoughts: Going Viral Is a Skill You Can Learn
Every artist who ever went viral started exactly where you are — with zero views, zero subscribers, and one dream.
If you take one thing from this guide, let it be this:
Virality is not about being the best — it’s about being consistent, creative, and authentic enough that the world finally takes notice.
So don’t wait for luck. Start experimenting. Upload that song. Post that short. Tell your story.
Because one of your next uploads could be the moment everything changes.
🎧 Quick Summary Checklist
✅ Optimize titles, thumbnails, and descriptions
✅ Use Shorts weekly
✅ Collaborate with others
✅ Engage with every comment
✅ Post consistently
✅ Promote across platforms
✅ Build your brand identity
✅ Tell your story authentically
✨ Remember:
Going viral is only the beginning.
Building a legacy is what comes after.
How to Blow Up on Spotify with Zero Investment
Written by Sounds SpaceHow to Blow Up on Spotify with Zero Investment
In today’s music industry, getting your music heard can feel like an uphill battle. Labels seem to have endless marketing budgets, playlists are flooded with major artists, and ads can cost more than what most indie musicians make in a month. But here’s the truth — you can still blow up on Spotify with zero investment if you approach it smartly, creatively, and consistently.
Money helps, sure. But strategy, consistency, and authenticity beat paid ads any day of the week. In this guide, we’ll walk you through exactly how to make your Spotify growth explode without spending a dime.
1. Start With Your Identity — Know Who You Are as an Artist
Before you post, promote, or pitch, you need clarity on who you are and what your music stands for. Think of your identity as the foundation of your Spotify strategy.
Ask yourself:
-
What emotion do I want my listeners to feel?
-
What kind of people would connect deeply with my sound?
-
What makes my music unique compared to others in my genre?
When you have a clear artist identity, every decision — from your cover art to your social media content — becomes easier. You’ll attract the right audience instead of trying to please everyone.
And that’s key: niche equals power. The smaller and more specific your target audience, the faster your Spotify presence will grow.
2. Optimize Your Spotify Profile
First impressions matter. When someone lands on your Spotify page, they should instantly understand your vibe.
Here’s how to optimize your artist profile for free:
-
High-quality profile picture: Use a clean, sharp image that reflects your brand. You can even shoot one yourself using portrait mode and natural lighting.
-
Custom banner: Create one for free using Canva. Add your logo or a simple background that matches your latest release.
-
Compelling bio: Write a short, emotional story about you and your sound. Keep it real. People connect to honesty.
-
Artist pick: Pin your latest single or a playlist that represents your sound.
-
Canvas videos: Use free editing tools like CapCut to create short looping visuals that grab attention.
A professional-looking profile doesn’t require money — it requires effort and vision.
3. Release Music Consistently
Spotify’s algorithm loves active artists. If you want to grow organically, you can’t release one song a year and expect results.
Instead, plan a release schedule — even if it’s one single every 6–8 weeks. The more you release, the more data Spotify gets about your sound and audience. That data helps Spotify recommend your music to new listeners through algorithmic playlists like:
-
Release Radar
-
Discover Weekly
-
Radio
Each release acts as a new opportunity to get discovered. Treat every song as a door that could lead to thousands of new fans.
4. Create a Network With Other Artists
You don’t need a budget to build a powerful network. You just need time and communication skills.
Here’s how to do it:
-
Follow and engage with artists in your genre on Instagram and TikTok.
-
Share their songs in your playlists or stories.
-
Reach out for collaborations. Offer to remix, feature, or co-promote each other’s tracks.
When two artists collaborate, they merge audiences. It’s one of the most effective free growth strategies on Spotify.
You can even create a Spotify playlist featuring all your collaborators, then cross-promote it. This helps every artist involved and multiplies exposure.
5. Use TikTok and Instagram Reels to Drive Streams
Short-form video content is the engine behind most Spotify growth stories today. Artists are blowing up every week from a single viral clip.
But you don’t need a professional camera or paid ads — your phone and creativity are enough.
Ideas that work:
-
Show your song’s creation process (from beat to vocals).
-
Make a “POV” video that captures the emotion behind your track.
-
Tell a story behind the lyrics.
-
Ask people to use your sound in their videos.
Even one viral video can push thousands of streams to your track overnight. Consistency is key — post every day if you can.
Pro tip: Always include your Spotify link in your bio and pin your most engaging video at the top of your profile.
6. Build Your Own Playlists
You don’t need to beg curators for placement — you can be your own playlist curator.
Create Spotify playlists around specific moods, genres, or activities — and include your music alongside popular songs that fit the same vibe.
Examples:
-
“Deep Chill House Vibes”
-
“Afro Tech Night Drive”
-
“Late Night Feelings”
Share these playlists on your social media, forums, and Reddit groups. Over time, your playlists can gain followers, and your track will naturally get consistent plays from them.
Tip: Change the order of songs weekly — Spotify often rewards active playlists with more visibility.
7. Submit to Spotify’s Editorial Team (For Free!)
A lot of artists don’t realize you can pitch your song directly to Spotify’s editorial team — completely free — through Spotify for Artists.
You need to:
-
Upload your song at least 7 days before release.
-
Fill out every detail (genre, mood, location, etc.).
-
Write a strong pitch explaining what makes your track special.
Even if you don’t land on an editorial playlist, your submission increases your chances of appearing in algorithmic playlists later on. It’s 100% worth doing for every release.
8. Engage With Your Listeners
When fans feel connected to you, they stick around — and Spotify notices that engagement.
Do this:
-
Reply to DMs and comments.
-
Post Spotify “stories” through your artist profile.
-
Use Spotify Clips (short videos that appear on your tracks).
-
Shout out fans who add your songs to their playlists.
That sense of community keeps listeners loyal, and loyal listeners trigger Spotify’s algorithm to recommend your tracks to more people.
9. Analyze Your Data and Adjust
Spotify gives you powerful analytics for free through Spotify for Artists. Use it.
Check your stats weekly:
-
Which songs are performing best?
-
Where are your listeners located?
-
What playlists are driving the most plays?
This helps you focus your promotion efforts where they matter most. For example, if you see you’re getting streams from Germany, start engaging with that audience on Instagram and TikTok.
10. Stay Patient, Consistent, and Authentic
The artists who blow up without money all share three traits: patience, consistency, and authenticity.
You might not get results in the first month, or even the first three. But if you keep showing up — releasing, engaging, and creating — the compound effect will kick in. One day, your song will land in someone’s playlist, go viral, or get picked up by Spotify’s algorithm.
Remember: every big artist today started from zero.
Your grind, story, and energy can take you there — even without a dollar spent.
Final Thoughts
Blowing up on Spotify with zero investment isn’t about luck — it’s about understanding the system, leveraging free tools, and building real connections.
Focus on your art, tell your story, and use every free platform available to amplify your music. The moment you start thinking like a brand instead of just an artist, you’ll notice the difference.
Money can buy exposure — but authenticity builds empires.
At just $99, the Akai MPK Mini IV looks to “set a new standard” for portable music creation
Written by Sounds SpaceAt just $99, the Akai MPK Mini IV looks to “set a new standard” for portable music creation
Akai’s MPK Mini series has long been the go-to compact controller for producers who want real music-making power in a backpack-sized package. The newly announced MPK Mini IV keeps that spirit alive — but this time Akai has pushed the little controller into territory that used to belong to larger, more expensive gear. For $99, the MPK Mini IV bundles a redesigned keybed, proper pitch and mod wheels, a full-color screen, a big preset library, and modern connectivity — a combination that genuinely feels like a fresh attempt to “set a new standard” for what a portable MIDI keyboard should be able to do.
What’s different this generation
On the surface, the MPK Mini IV still looks like the compact, 25-note unit we know: small footprint, octave buttons, RGB MPC-style pads, and assignable encoders. But a handful of changes make it a noticeably more expressive and studio-ready tool:
-
Real pitch and modulation wheels. The old red joystick is gone — replaced with traditional wheels that players expect for expressive pitch bends and modulation. That’s a subtle change, but it matters when you’re trying to play leads or sculpt sounds live.
-
New keybed design. Akai says the MPK Mini IV uses an updated keybed for better feel and velocity response, which helps when you want more nuanced performances from only 25 keys. For laptop-and-backpack sessions, that improved tactile response matters more than it sounds.
-
Full-color display and workflow controls. The addition of a screen and expanded transport controls brings more on-device feedback and faster hands-on control of DAW functions, reducing the need to stare at your computer while you tweak.
-
USB-C + 5-pin MIDI out. Modern connectivity arrives with USB-C for bus power and bidirectional MIDI, while a dedicated 5-pin MIDI output keeps the Mini relevant for hardware-heavy setups and live rigs. That makes the unit as comfortable onstage as it is in a bedroom studio.
Those updates don’t all sound revolutionary individually, but together they change how you can use a pocket-sized controller — from sketching ideas in cafes to integrating with outboard synths on stage.
Value: hardware + software at an entry-level price
One of the most headline-grabbing elements is the price: $99 USD. For that figure, Akai isn’t just offering a small keyboard — it ships with access to a bundled “Studio Instrument Collection” of more than 1,000 presets and sounds from AIR, Moog, and Akai, designed to get you making full productions straight away. For new producers or anyone who wants immediate, high-quality tones without hunting for separate plugins, this bundle significantly boosts the out-of-the-box value.
When you compare this to purchasing a small hardware controller plus even a single high-quality VST, the MPK Mini IV’s price becomes more persuasive: hardware + software that’s useful in real production workflows at a mass-market price point. Music tech outlets have framed it as Akai trying to reassert the MPK Mini as the affordable standard-bearer for portable creation — and the specs back that up.
Who this is for (and who might benefit most)
The MPK Mini line historically lives in two worlds: beginners who need a small, affordable entry into MIDI controllers, and experienced producers who want a compact unit to travel with. The IV seems deliberately aimed at both groups:
-
Beginners: The included sound library and one-to-one parameter mapping make it easier to start producing without juggling multiple purchases. The hardware is approachable and straightforward, and the screen helps demystify parameter changes.
-
Traveling producers / live performers: Real wheels, 5-pin MIDI, and USB-C power make the controller viable as part of a live rig or hardware-centric studio. The lighter footprint and robust mapping allow you to control both software and external synths.
-
Beatmakers and finger-drummers: The MPK’s RGB MPC pads remain central to the device’s appeal. Pressure-sensitive, RGB-backlit pads with assignable encoders are ideal for both drum programming and expressive pad performance.
If you’re someone who already owns a large controller or prefers full-sized keys and aftertouch, this isn’t a replacement — it’s a complement. But if you’ve been waiting for a truly portable controller that doesn’t force trade-offs in playability or connectivity, the IV makes a compelling case.
Practical workflow improvements
A few real-world workflow touches are worth highlighting because they change how fast you can move from idea to finished loop:
-
DAW mapping and transport controls: The expanded transport controls and DAW-friendly shortcuts are mapped to major DAWs out of the box, so basic recording and editing tasks are literally at your fingertips. That reduces awkward menu diving when you’re in the zone.
-
Screen feedback: Seeing parameter values on a small screen cuts down guesswork — useful when editing synth parameters or switching presets on stage.
-
Preset integration with the hardware: Akai’s promise of tight software/hardware integration (1-to-1 parameter mapping) means that the knobs and encoders won’t require endless MIDI learn steps for common instruments included in the bundle. That smoothness matters for fast creative sessions.
These details make the MPK Mini IV feel less like a cheap, stripped-down controller and more like a thoughtfully designed mini-instrument.
Limitations and realistic expectations
No magic: there are natural limits to what a 25-key controller can replace. If your workflow relies on full-sized, weighted keys, poly-aftertouch, or an abundance of knobs and faders, you’ll still want larger hardware. Some other limitations to keep in mind:
-
Key size and range. With only 25 keys and octave switches, complex piano parts still require either careful programming or a second, larger keyboard for serious performance.
-
Aftertouch / deeper expression. If you require polyphonic aftertouch or advanced expressive tech built into the keybed, that’s usually reserved for higher-tier controllers. The IV’s upgrades improve feel, but they don’t turn it into a high-end stage keyboard.
-
Build expectations. At $99, the MPK Mini IV is impressively specified, but it’s still a compact plastic device. For the majority of producers, that’s fine; for frequent onstage abuse, you’ll want a protective case and careful handling.
The competition and where the MPK Mini IV fits
The compact controller segment is competitive: Arturia’s MiniLab series, Novation’s Launchkey Mini line, and Akai’s own prior generations have occupied similar price brackets. What the MPK Mini IV tries to do differently is combine pro-grade expressive elements (wheels, better keybed), modern connectivity (USB-C + 5-pin MIDI), and a generous software package — all at an entry-level price. That cocktail of features positions it as an attractive first-choice for folks who want a one-box start-to-finish creative tool.
Final thoughts: Does it “set a new standard”?
Calling any single product “the new standard” is bold — and the claim relies on how you weigh portability, playability, and price. But Akai’s MPK Mini IV does something important: it narrows the compromise between compactness and functionality. By bringing expressiveness (real wheels), better playability (new keybed), modern I/O (USB-C and 5-pin MIDI), and an extensive sound library together at $99, Akai has raised expectations about what a sub-$100 compact controller can deliver. For many creators — students, traveling producers, bedroom beatmakers — that’s a meaningful shift.
If you’re shopping for a small controller right now, the MPK Mini IV deserves a serious look. It isn’t a replacement for pro-stage keyboards or full-sized controllers, but as a portable creative companion that helps you make music faster and with fewer compromises, it arguably does set a new bar for value and usefulness in this category.
Sources & further reading: Akai’s official MPK Mini IV page (product details and Studio Instrument Collection), MusicTech’s article announcing the MPK Mini IV, MusicRadar’s feature, retailer/spec pages (Sweetwater), and industry press coverage used to summarize specs and feature context.
