Fact-checked by Grok 2 weeks ago

Music and artificial intelligence


Music and artificial intelligence refers to the application of computational algorithms and models to tasks such as music composition, generation, , , , and recommendation, enabling systems to process audio data, emulate styles, and create novel outputs. The field traces its origins to mid-20th-century experiments in computer-generated sound, exemplified by Australia's , the world's first programmable computer to play music publicly in 1951 through programmed tones approximating tunes like "." Early efforts relied on rule-based and probabilistic methods, such as the 1957 Suite, the first computer-composed score using Markov chains to generate string quartets.
Subsequent advancements shifted toward architectures, including autoregressive models like for high-fidelity audio synthesis and generative adversarial networks for multi-instrument track creation, allowing AI to produce coherent pieces mimicking human composers. Notable achievements encompass AI systems generating full songs in specific styles, such as the 2016 "Daddy's Car," an AI-composed track emulating with human-provided , and modern tools producing original multi-track from text prompts. By 2024, approximately 60 million individuals had utilized generative AI for or creation, reflecting widespread adoption in , , and experimentation. These capabilities have enhanced efficiency in areas like algorithmic harmonization and style transfer, though empirical assessments indicate AI excels in pattern replication rather than novel causal innovation akin to human intuition. The integration of AI in music has sparked controversies, particularly regarding , as generative models trained on vast copyrighted datasets without permission raise infringement claims; in 2024, the sued platforms like Suno and Udio for allegedly reproducing protected works in outputs. Purely AI-generated compositions lack copyright eligibility under U.S. due to the absence of authorship, complicating for assisted works and prompting debates on , compensation for training , and potential displacement of creators. While stakeholders advocate protections to preserve incentives for original artistry, AI's deterministic emulation of statistical correlations from underscores ongoing questions about artistic and economic sustainability in music production.

Historical Development

Early Computational Approaches

The earliest documented instance of a computer generating music occurred in 1951 with the , Australia's first programmable digital computer, developed at the . Programmer Geoff Hill utilized the machine to produce simple tones by oscillating its mercury delay line memory at audio frequencies, playing familiar tunes such as "" and "Blue Danube Waltz" during public demonstrations from August 7 to 9. This approach served primarily as a diagnostic test for the computer's acoustic output rather than artistic composition, relying on manual programming of frequency oscillations without algorithmic generation of novel sequences. Preceding CSIRAC, preliminary experiments with computer-generated sounds took place in 1948–1949 at the , where programmed the to produce basic musical notes as part of exploring machine capabilities in pattern generation. These efforts involved scaling outputs to audible pitches, marking an initial foray into computational audio but limited to isolated tones without structured . A significant advancement came in 1957 with the development of MUSIC I by at Bell Laboratories, the first widely used program for synthesizing waveforms on an computer. MUSIC I enabled the computation of sound samples through , where basic waveforms like sines were combined and filtered to approximate instrument timbres, laying groundwork for software-based music production. That same year, Lejaren Hiller and Leonard Isaacson composed the Illiac Suite for using the ILLIAC I computer at the University of Illinois, employing probabilistic models to generate note sequences based on statistical analysis of existing music, such as Bach chorales. The suite's four movements progressively incorporated , from deterministic rules to fully processes, demonstrating early techniques that simulated musical decision-making via computational probability rather than human intuition. These pioneering efforts in the were constrained by computational limitations, including slow processing speeds and low memory, resulting in outputs that prioritized feasibility over complexity or expressiveness. They focused on rule-based and generation, influencing subsequent developments in computer-assisted music without yet incorporating learning from data.

Algorithmic and Rule-Based Systems

Algorithmic and rule-based systems for music generation employ explicit procedural instructions, often grounded in music theory or mathematical formalisms, to produce scores or audio without reliance on statistical learning from corpora. These methods typically involve defining constraints—such as avoidance of parallel fifths in , adherence to harmonic functions, or probabilistic distributions for note selection—and iteratively generating and evaluating musical elements against them. Unlike later data-driven approaches, they prioritize interpretability and direct emulation of compositional principles, enabling reproducible outputs but requiring manual encoding of . Pioneering computational implementations emerged in the mid-1950s. In 1956, J.M. Pinkerton developed the "Banal Tune-Maker," an early program using first-order Markov chains derived from 50 British folk tunes to probabilistically generate melodies, marking one of the initial forays into automated tune creation via transitional probabilities rather than pure randomness. The following year, Lejaren Hiller and Leonard Isaacson produced the Illiac Suite for on the ILLIAC I computer, combining Markov processes for melodic generation with hierarchical screening rules to enforce contrapuntal norms, including voice independence and resolution of dissonances, across its four movements. This work premiered on August 9, 1957, at the University of Illinois and represented a hybrid of selection and deterministic validation, yielding music that satisfied traditional fugal structures while incorporating chance elements. Iannis Xenakis advanced algorithmic techniques by integrating probabilistic mathematics into compositional practice, beginning with manual stochastic models in works like Pithoprakta (1956) and extending to computer execution. His ST/10-1+2 (1962) employed simulations on an IBM 7090 to compute glissandi densities, durations, and registers from gamma distributions, automating granular sound distributions across 10 percussionists and 2 string orchestras to realize macroscopic sonic forms from microscopic rules. Xenakis's methods emphasized causal links between mathematical parameters and perceptual outcomes, influencing subsequent formalist approaches. Later refinements focused on stricter rule enforcement for pedagogical styles. In 1984, William Schottstaedt at Stanford's Center for Computer Research in Music and Acoustics (CCRMA) implemented Automatic Species Counterpoint, a program that codified Johann Joseph Fux's 1725 guidelines to generate multi-voice accompaniments to a given , prioritizing stepwise motion, dissonance treatment, and intervallic variety through . This system extended to higher , demonstrating how rule hierarchies could yield stylistically coherent without probabilistic variance. Such systems highlighted the computational tractability of codified theory but exposed limitations in ; exhaustive rule sets struggled to capture idiomatic nuances like thematic development or cultural idioms, often producing formulaic results that prioritized conformity over innovation. Empirical evaluations, such as listener tests on Illiac Suite segments, revealed preferences for rule-screened outputs over unchecked randomness, underscoring the causal role of structured constraints in perceptual coherence.

Machine Learning Integration

The integration of into music emerged in the late and , transitioning from deterministic rule-based and algorithmic methods to probabilistic, data-driven models capable of inferring musical patterns from examples. This shift enabled systems to generalize beyond predefined rules, learning stylistic nuances such as , , and transitions through on corpora of existing music. Early applications focused on predictive generation and analysis, leveraging techniques like neural networks and Markov models to model sequential dependencies inherent in musical structures. A seminal example was Michael C. Mozer's system, introduced in a NeurIPS , which utilized a architecture to compose melodies. Trained on datasets of Bach chorales, CONCERT predicted subsequent notes based on prior context, incorporating multiple timescales to capture both local dependencies and longer-term phrase structures without explicit rule encoding. The model demonstrated emergent musical coherence, generating stylistically plausible sequences that adhered to training data distributions, though limited to monophonic lines due to the era's computational constraints. Subsequent refinements, detailed in Mozer's work, extended this predictive framework to explore connections between surface-level predictions and deeper perceptual hierarchies in music cognition. Parallel developments applied to and performance tasks. Hidden Markov models (HMMs), popularized in the 1990s for sequence modeling, were adapted for audio analysis, such as beat tracking and chord estimation, by representing music as hidden states with observable emissions like or features. Gaussian mixture models and early clustering techniques facilitated classification and similarity search, processing symbolic data or raw audio spectrograms to identify patterns in large datasets. These methods, while rudimentary compared to later , provided empirical evidence of ML's efficacy in handling variability in musical expression, outperforming hand-crafted heuristics in tasks requiring adaptation to diverse corpora. By the early , integration expanded to interactive systems, with kernel-based methods like support vector machines enabling real-time and . For instance, ML-driven models analyzed performer inputs to generate harmonious responses, as explored in connectionist frameworks for ensemble simulation. This era's advancements, constrained by data scarcity and processing power, emphasized on annotated datasets, foreshadowing scalable applications in recommendation engines that used —rooted in matrix factorization—to infer user preferences from listening histories. Despite biases toward Western classical training data in early corpora, these systems established causal links between learned representations and perceptual validity, validated through listener evaluations showing above-chance coherence ratings.

Recent Advancements and Market Emergence

In 2024 and 2025, advancements in generative for music have centered on text-to-music and text-to-song models capable of producing full compositions with vocals, , and structure from prompts. Suno, launched publicly in December 2023, gained prominence for generating complete songs including and vocals, while Udio, released in April 2024, offered enhanced controllability and audio-to-audio extensions for iterative refinement. Stability AI's Stable Audio, updated in 2024, focused on high-fidelity audio clips up to three minutes, emphasizing generation for stems and loops. These tools leveraged diffusion models and architectures trained on vast datasets, achieving levels of musical that incorporate genre-specific elements, , and , though outputs often exhibit limitations in long-form originality and emotional depth compared to human compositions. Market emergence accelerated with widespread adoption, as approximately 60 million individuals utilized generative AI for music or lyrics creation in 2024, representing 10% of surveyed consumers. The global generative in music market reached $569.7 million in 2024, projected to grow to $2.79 billion by 2030 at a exceeding 30%, driven by cloud-based platforms holding over 70% share. Key players like AIVA, operational since 2016 but expanding commercially, targeted professional composition assistance, while startups such as Suno secured over $160 million in sector funding in 2024 alone. By October 2025, Suno entered funding talks for more than $100 million at a $2 billion valuation, signaling investor confidence amid rising demand for AI-assisted production tools. Challenges in market integration include disputes, with major labels suing Suno and Udio in 2024 over usage, prompting negotiations for licensing deals by mid-2025. Advancements in generation and detection emerged as responses, enabling ethical and to mitigate infringement risks. Despite concerns—where proliferation could overwhelm independent artists—these developments have democratized access, with tools integrated into platforms like BandLab for loop suggestions and real-time . Overall, the sector's growth reflects a shift toward hybrid human- workflows, though on sustained commercial viability remains limited by unresolved legal and creative authenticity debates.

Technical Foundations

Symbolic Music Representations

Symbolic music representations encode musical structures as discrete, abstract symbols—such as pitches, durations, velocities, and rhythmic patterns—rather than continuous audio waveforms, enabling computational manipulation at a high semantic level. The , introduced in 1983 by manufacturers including and , serves as the dominant format, storing event-based sequences like note onset, offset, and controller changes in binary files typically under 1 MB for complex pieces. Complementary formats include , an XML-based standard developed by Recordare in 2000 for interchangeable notation that preserves layout and symbolic markings like dynamics and articulations, and for simpler textual encoding of melodies. These representations abstract away and acoustics, focusing on performative instructions that require via virtual instruments for auditory output. In AI-driven music tasks, symbolic supports and through tokenization into vocabularies of events or multidimensional piano-roll grids, where rows denote pitches and columns time steps, facilitating training on datasets exceeding millions of files. models, including (LSTM) networks and transformers, process these as autoregressive s to compose coherent structures, as evidenced in a 2023 survey identifying over 50 architectures for symbolic tasks like continuation and . -based representations extend this by modeling notes as nodes and relationships (e.g., chords) as edges, improving tasks like classification with convolutional networks achieving up to 85% accuracy on symbolic datasets. Advantages of symbolic approaches in AI composition include computational efficiency—requiring orders of magnitude less data and processing than audio models—and structural editability, such as transposing keys or altering tempos post-generation without waveform resampling. This enables real-time applications like interactive improvisation, where low-latency models generate MIDI streams under 100 ms delay. However, limitations arise in capturing expressive nuances like microtiming or instrumental timbre, necessitating hybrid post-processing for realism. Notable AI systems leveraging symbolic representations include Google's framework, which employs NoteSequence protobufs derived from for models like MusicVAE, trained on corpora such as the Lakh Dataset containing over 176,000 pieces. Recent advancements feature MusicLang, a 2024 transformer-based model fine-tuned on symbolic data for controllable generation via prompts like tags, and NotaGen, which outputs classical scores in from latent embeddings. XMusic, proposed in 2025, extends this to multimodal inputs (e.g., text or ) for generalized symbolic output, demonstrating improved in multi-instrument arrangements. These systems underscore symbolic methods' role in scalable, interpretable music AI, with ongoing research addressing alignment between symbolic predictions and perceptual quality.

Audio and Waveform Generation

Audio and waveform generation in AI for music involves neural networks that produce raw audio signals directly, bypassing symbolic notations like to synthesize continuous s at sample rates such as 16 kHz or higher. This approach enables the creation of realistic timbres, harmonies, and rhythms but demands substantial computational resources due to the high dimensionality of audio data—typically 16,000 samples per second for CD-quality sound. Early methods relied on autoregressive models, which predict each audio sample sequentially based on prior ones, achieving breakthroughs in naturalness over traditional synthesizers like those using sinusoidal oscillators or formant synthesis. WaveNet, introduced by DeepMind in September 2016, marked a pivotal advancement with its dilated architecture for generating raw audio waveforms in an autoregressive manner. Trained on large datasets of speech and extended to , WaveNet produced higher-fidelity outputs than parametric vocoders, with mean opinion scores indicating superior naturalness in blind tests. Building on this, Google's NSynth dataset and model, released in April 2017, applied WaveNet-inspired autoencoders to synthesis, enabling interpolation between instrument timbres—such as blending and sounds—to create novel hybrid tones from 1,000+ instruments across 289,000 notes. NSynth's representation allowed for continuous variation in pitch, timbre, and envelope, demonstrating AI's capacity to generalize beyond discrete categories in audio generation. Subsequent models scaled up complexity for full music tracks. OpenAI's , launched April 30, 2020, employed a multi-scale VQ-VAE to compress raw 44.1 kHz audio into discrete tokens, followed by autoregressive transformers conditioned on , , and to generate up to 20-second clips with rudimentary vocals. Trained on 1.2 million songs, Jukebox highlighted challenges like mode collapse in alternatives and the need for hierarchical modeling to handle long sequences, requiring hours of computation on V100 GPUs for short outputs. By 2023, Meta's MusicGen shifted to efficient token-based autoregression using EnCodec compression at 32 kHz with multiple codebooks, enabling text- or melody-conditioned generation of high-quality music up to 30 seconds from 20,000 hours of licensed training data. Diffusion models emerged as alternatives, iteratively denoising latent representations to produce audio. , released in December 2022, fine-tuned —a —on spectrogram images, converting generated mel-s back to waveforms via vocoders, thus leveraging vision for music clips conditioned on text prompts like "jazzy beats." This spectrogram-to-audio pipeline offered faster inference than pure waveform while capturing expressive musical structures, though limited to short durations due to computational constraints. These techniques underscore ongoing trade-offs: autoregressive fidelity versus controllability, with hybrid approaches like latent in models such as AudioLDM further optimizing for longer, coherent generations. Despite progress, issues persist in maintaining long-term coherence, avoiding artifacts, and scaling to professional production lengths without hallucinations or stylistic drift.

Music Information Retrieval

Music Information Retrieval (MIR) refers to the interdisciplinary field that develops computational techniques to analyze, organize, and retrieve information from music sources, including audio signals, symbolic notations, and . Core tasks include feature extraction from audio—such as Mel-frequency cepstral coefficients (MFCCs) for analysis or chroma features for harmony detection—and subsequent processing for applications like similarity search or classification. These methods bridge , , and music to enable automated handling of large music corpora. Early MIR efforts emerged in the 1990s with query-by-humming systems, exemplified by research presented at the International Computer Music Conference in 1993, which matched user-hummed melodies against databases using dynamic programming for . The field formalized with the founding of the International Society for Music Information Retrieval (ISMIR) in 2000, whose inaugural conference in , marked a milestone in fostering collaborative research. By 2024, ISMIR conferences featured over 120 peer-reviewed papers annually, reflecting sustained growth in addressing challenges like scalable indexing and cross-modal retrieval. Integration of , particularly , has transformed by replacing hand-crafted features with data-driven models. Traditional approaches relied on rule-based similarity metrics, but supervised classifiers like support vector machines achieved genre classification accuracies around 70-80% on benchmarks such as the GTZAN dataset in early evaluations. Deep learning advancements, including convolutional neural networks (CNNs) for analysis and recurrent networks for sequential data, have pushed accuracies above 85% for tasks like mood detection and instrument recognition, as demonstrated in ISMIR proceedings from 2015 onward. Frameworks like AMUSE, developed in 2006 and extended with neural components, facilitate end-to-end learning for feature extraction and retrieval. In practical deployments, MIR powers music recommendation systems by computing content-based similarities, complementing ; for instance, platforms analyze acoustic features to suggest tracks with matching or . Song identification services like employ audio fingerprinting—hashing constellations of spectral peaks—to match short clips against databases in milliseconds, even amid noise, processing billions of queries annually since its 2002 launch. These AI-enhanced systems underscore MIR's causal role in enabling efficient discovery, though challenges persist in handling symbolic-audio mismatches and cultural biases in training data.

Hybrid and Multimodal Techniques

Hybrid techniques in AI music generation integrate symbolic representations, such as MIDI sequences or rule-based structures, with neural network-based audio synthesis to address limitations in purely data-driven models, including long-term coherence and structural fidelity. Symbolic methods provide explicit control over harmony, rhythm, and form, while neural approaches excel in generating expressive waveforms; combining them enables hierarchical modeling where high-level symbolic planning guides low-level audio rendering. For instance, a 2024 framework proposed hybrid symbolic-waveform modeling to capture music's hierarchical dependencies, using symbolic tokens for global structure and diffusion models for local timbre variation, demonstrating improved coherence in generated piano pieces compared to end-to-end neural baselines. Similarly, MuseHybridNet, introduced in 2024, merges variational autoencoders with generative adversarial networks to produce thematic music, leveraging hybrid conditioning for motif consistency and stylistic diversity. These hybrid approaches mitigate issues like in pure neural generation by enforcing musical rules symbolically, as seen in systems estimating difficulty via convolutional and recurrent networks hybridized with symbolic feature extraction, achieving 85% accuracy on benchmark datasets in 2022. Recent advancements, such as thematic conditional GANs from 2025, further refine this by incorporating variational for , yielding music aligned with user-specified themes while preserving acoustic realism. Empirical evaluations indicate hybrids outperform single-modality models in metrics like melodic repetition and harmonic validity, though computational overhead remains a challenge, requiring optimized pipelines. Multimodal techniques extend this by fusing non-audio inputs—text, images, or video—with data, enabling conditioned that aligns outputs across senses for applications like synchronized . Models process text prompts for or alongside visual cues for , using cross-attention mechanisms to bridge modalities. Spotify's LLark, released in October 2023, exemplifies this as a trained on audio-text pairs, supporting tasks from captioning to continuation while handling multimodal queries for classification with 72% top-1 accuracy. MusDiff, a 2025 diffusion-based framework, integrates text and image inputs to generate with enhanced cross-modal consistency, outperforming text-only baselines in subjective quality assessments by incorporating visual semantics for rhythmic alignment. Advanced multimodal systems like MuMu-LLaMA (December 2024) employ large models to orchestrate across music, image, and text, producing full compositions from mixed prompts via pretrained encoders, with evaluations showing superior diversity in polyphonic outputs. Mozart's Touch ( 2025) uses a lightweight pipeline with multi-modal captioning and bridging for efficient synthesis, reducing parameters by 90% compared to full-scale models while maintaining fidelity in short clips. These methods reveal causal links between input modalities and musical elements, such as image-derived from motion cues, but face hurdles in alignment and propagation from corpora. Overall, hybrids and multimodals advance toward versatile AI music tools, verifiable through benchmarks like cross-modal retrieval F1-scores exceeding 0.8 in controlled studies.

Applications in Music

Composition and Songwriting Assistance

AI systems assist human composers and songwriters by generating musical elements such as melodies, progressions, harmonies, and in response to textual prompts, stylistic inputs, or partial user-provided material. These tools leverage models, including transformers and generative adversarial networks, trained on vast datasets of existing music to suggest completions, variations, or full segments that align with specified genres, moods, or structures. By automating repetitive ideation tasks, they enable users to overcome creative blocks and iterate rapidly, though outputs frequently require human refinement for coherence and emotional depth. One prominent example is AIVA, launched in 2016 and updated through 2025, which functions as a virtual co-composer specializing in instrumental tracks across more than 250 styles, including classical, film scores, and pop. Users input parameters like duration, tempo, and instrumentation, after which AIVA employs recurrent neural networks to produce editable files that can be integrated into digital audio workstations. The system has been used for soundtracks in and advertisements, with its algorithm emphasizing harmonic and structural rules derived from analyzing thousands of scores, yet it cannot originate concepts absent from its training corpus. Platforms like Suno, introduced in 2023 with version 4 released by late 2024, extend assistance to vocal-inclusive songwriting by generating complete tracks—including , melodies, and arrangements—from prompts such as "upbeat rock song about ." Features like song editing, custom personas for stylistic consistency, and audio-to-reimagined covers allow iterative refinement, making it suitable for prototyping demos or brainstorming hooks. Udio, emerging in 2024, similarly supports text-to-music creation with high-fidelity vocals and instrumentation, emphasizing emotional expressiveness through diffusion models that extend short clips into full songs, aiding users in exploring unconventional progressions without prior notation skills. Despite these capabilities, AI-assisted composition faces scrutiny for lacking true intentionality or cultural context, as models replicate patterns from licensed datasets rather than innovating from first principles, potentially homogenizing outputs across users. Empirical evaluations, such as blind tests by musicians, indicate that while AI excels at technical proficiency, human-composed works often score higher on perceived authenticity and narrative coherence. Songwriters report using these tools primarily for inspiration—e.g., generating 80% of initial ideas before manual overhaul—rather than as standalone creators, preserving artistic agency amid rapid technological iteration.

Performance and Real-Time Interaction

Artificial intelligence systems for music performance and interaction generate or modify audio in response to live human inputs, such as performer notes, variations, or stylistic cues, facilitating collaborative , , and augmented . These systems emphasize low-latency to mimic natural ensemble dynamics, often employing transformer-based models or to predict and adapt to musical progressions. Unlike offline generation tools, real-time variants prioritize sequential chunking of audio output, typically in 2-second segments, to enable ongoing human-AI during live sessions. Yamaha's Music Technology analyzes performer input via microphones, cameras, and sensors in real time, comparing it against to predict deviations, errors, and expressive nuances derived from human data. It supports synchronization for solo instruments, multi-person groups, or orchestras, adapting to individual styles and integrating with external devices like lighting. Demonstrated in the "JOYFUL " on December 21, 2023, at Hall, the system accompanied a performance of Beethoven's using the "Daredemo " interface. Google DeepMind's Lyria RealTime model produces 48 kHz stereo audio continuously, allowing users to steer generation via text prompts for genre blending, mood adjustment, or direct controls over , , density, and brightness. Integrated into tools like MusicFX DJ, it responds interactively to inputs akin to a collaborator, supporting applications from to production. Similarly, 's Magenta RealTime, released on June 20, 2025, as an open-weights model, generates high-fidelity audio with a real-time factor of 1.6—producing 2 seconds of output in 1.25 seconds on accessible hardware like TPUs. It conditions outputs on prior audio and style embeddings, enabling manipulation for multi-instrumental exploration during live performance. Research prototypes like ReaLJam employ learning-tuned transformers for human-AI , incorporating anticipation mechanisms to predict and visualize future musical actions, thus minimizing perceived . A 2025 user with experienced musicians found sessions enjoyable and musically coherent, highlighting the system's adaptive communication. Other frameworks, such as Intelligent Music Performance Systems, typologize design principles for and expressivity, while low- symbolic generators like SAGE-Music target by prioritizing attribute-conditioned outputs. These advancements underscore a shift toward paradigms, though challenges persist in achieving seamless causal responsiveness beyond statistical pattern matching.

Production, Mixing, and Mastering

has enabled automated assistance in music mixing by analyzing audio tracks to recommend adjustments in levels, equalization, dynamics, and spatial imaging. iZotope's Neutron 5, released in , incorporates a Assistant that uses to evaluate track relationships via inter-plugin communication, suggesting initial balances, curves, and compression settings tailored to genre and content. This approach processes waveforms and spectral data to identify masking issues and apply corrective processing, reducing manual iteration time for producers. In mastering, AI tools apply final loudness normalization, stereo enhancement, and limiting to prepare tracks for distribution. LANDR's mastering engine, updated through 2024, generates genre-specific masters by training on vast datasets of professional references, achieving results comparable to initial human passes in clarity and balance, though customizable via controls for and dynamics. , launched in 2023, features a Mastering Assistant that leverages neural networks to match input audio to reference styles, optimizing for platforms like streaming services with integrated metering. These systems often employ models trained on annotated audio corpora to predict perceptual improvements. Empirical comparisons indicate limitations in AI-driven mastering, with a 2025 study finding that outputs exhibit higher levels, reduced (averaging 20-30% narrower than human masters), and elevated penalties under BS.1770 standards, attributed to over-reliance on aggregated training data rather than contextual nuance. Human engineers outperform in preserving artistic intent for non-standard mixes, as prioritizes statistical averages over subjective variations. Despite this, adoption has grown, with tools like Cryo Mix providing instant enhancements for independent producers, integrating seamlessly into digital audio workstations as of 2024. Hybrid workflows, combining suggestions with manual refinement, predominate in professional settings to mitigate these shortcomings.

Recommendation and Personalization

Artificial intelligence has transformed music recommendation systems by leveraging algorithms to analyze user listening histories, preferences, and behavioral patterns, enabling highly personalized suggestions that enhance user satisfaction and platform retention. Early implementations combined , which identifies similarities between users' tastes, with content-based methods that match tracks to acoustic features like and ; these approaches underpin features such as 's Discover Weekly, launched in 2015, which generates weekly playlists of 30 new songs per user based on billions of listening sessions processed through neural networks. By 2022, algorithmic recommendations accounted for at least 30% of all songs streamed on , correlating with higher user engagement metrics, including increased daily and session lengths, as platforms report that personalized feeds drive over 40% of total streams in some cases. Advancements in , including convolutional neural networks, have refined by incorporating data such as , audio waveforms, and even user sentiment from interactions, allowing systems to predict preferences with greater accuracy than traditional matrix factorization alone. For instance, Spotify's DJ, introduced in 2023, employs large models to contextualize recommendations with commentary tailored to individual tastes, interpreting queries and orchestrating tools for dynamic playlist curation. Approximately 75% of major streaming services, including , , and , integrate such AI-driven , which empirical studies link to improved subscription retention rates by 20-30% through reduced churn from irrelevant suggestions. However, these systems face challenges from the influx of AI-generated tracks, which by 2024 comprised a notable portion of recommendations—user reports indicate up to 80% of some Discover Weekly playlists containing such content—potentially diluting quality and fostering filter bubbles that limit exposure to diverse human-created music. Recent innovations extend to agentic frameworks, where models like those at use scalable preference optimization to adapt outputs based on loops, outperforming static embeddings in taste alignment by metrics such as rate improvements of 15-25% in internal evaluations. Despite these gains, suggests algorithmic curation can homogenize listening habits, with studies showing reduced genre diversity in recommendations over time unless explicitly engineered for , as pure preference-matching prioritizes familiarity over novelty. Platforms mitigate this through hybrid techniques blending explicit diversity objectives into loss functions, though real-world deployment often trades off against short-term engagement to avoid alienating core s. Overall, 's role in recommendation underscores a causal link between data-driven and economic viability for streaming services, yet it demands ongoing scrutiny of training data biases and generative content proliferation to sustain long-term trust.

Notable Systems and Tools

Research-Oriented Projects

Google's Magenta project, initiated in 2016, represents a foundational research effort in applying machine learning to music generation and creativity, utilizing TensorFlow to develop models that assist in composition, performance, and sound synthesis. Key outputs include NSynth (2017), which synthesizes novel sounds by interpolating between learned audio embeddings from over 300,000 instrument notes, and MusicVAE (2018), a variational autoencoder for generating diverse symbolic music sequences while preserving structural coherence. The project emphasizes open-source dissemination, with tools like Magenta Studio integrating into digital audio workstations such as Ableton Live for real-time generation, and recent advancements like MagentaRT (2025), a 800-million-parameter model enabling low-latency music creation during live performance. Evaluations highlight Magenta's focus on enhancing human creativity rather than autonomous replacement, though limitations in long-form coherence and stylistic fidelity persist compared to human compositions. OpenAI's , released on April 30, 2020, advances raw audio generation by training a vector quantized variational autoencoder (VQ-VAE) on 1.2 million songs across 125 genres, producing up to 1.5-minute clips complete with rudimentary vocals and . The system conditions outputs on artist, genre, and lyrics, achieving sample quality via hierarchical tokenization of waveforms into discrete units, but requires significant computational resources—training on 64 V100 GPUs for nine days—and struggles with factual lyric accuracy and extended durations due to autoregressive dependencies. As a artifact rather than a deployable tool, 's sample explorer demonstrates capabilities in emulating styles like or , yet human evaluations rate its outputs below professional tracks in overall appeal and coherence. Other notable academic initiatives include Stanford's Center for Computer Research in Music and Acoustics (CCRMA), which integrates for tasks like modeling and interactive performance since the 1970s, evolving to incorporate for expressive in projects like the Chloroma model for choral arrangements. Peer-reviewed studies from such efforts underscore empirical challenges, such as scarcity in niche genres leading to biased outputs favoring Western classical datasets, necessitating diverse training corpora for causal validity in generation models. These projects collectively prioritize methodological innovation over commercialization, with metrics like and Frechet audio distance used to quantify progress, though real-world musical utility remains constrained by lacks in and emotional depth derivable from first-principles acoustic physics.

Commercial and Open-Source Generators

Suno, a commercial AI platform for generating full songs from text prompts, was launched in December 2023 through a partnership with Microsoft. It produces tracks including lyrics, vocals, and instrumentation, with versions enabling up to four-minute compositions even on free tiers; its V4 model, released in November 2024, improved audio fidelity and lyric coherence, while V5 in September 2025 added support for user-uploaded audio samples and enhanced dynamics. A mobile app followed in July 2024, expanding accessibility. Udio, another text-to-music service debuted in April 2024, specializes in synthesizing realistic audio tracks with vocals from descriptive prompts or provided lyrics, positioning itself as a direct competitor to Suno by emphasizing high-fidelity output and user remix capabilities. AIVA, established in 2016 by Luxembourg-based Aiva Technologies, focuses on compositional assistance across over 250 styles, particularly excelling in orchestral and classical genres; it gained formal recognition as a composer from SACEM in 2017, allowing generated works to receive performance rights licensing. Open-source alternatives have democratized access to AI music generation tools. Meta's MusicGen, introduced in June as part of the AudioCraft library, employs an autoregressive to produce music clips conditioned on text descriptions or melodic prompts, supporting up to 30-second high-quality samples at 32kHz . AudioCraft, released in , extends this with integrated models like AudioGen for sound effects and EnCodec for efficient , providing a comprehensive for training and inference on raw audio data via and repositories. These frameworks enable researchers and developers to fine-tune models locally, though they require significant computational resources for optimal performance compared to cloud-based services.

Economic Impacts

Market Growth and Revenue Dynamics

The generative AI segment within the music , encompassing tools for , generation, and production assistance, reached an estimated market size of USD 569.7 million in 2024. This figure reflects rapid adoption driven by accessible platforms that lower barriers to music creation, with compound annual growth rates (CAGR) projected between 26.5% and 30.4% through 2030 or beyond, potentially expanding the to USD 2.8 billion by 2030 or up to USD 7.4 billion by 2035. Broader AI applications in music, including recommendation and systems, contributed to a global AI music valuation of approximately USD 2.9 billion in 2024, dominated by cloud-based solutions holding 71.4% share. Revenue dynamics hinge on and subscription models, with leading generative tools like Suno generating over USD 100 million in annual recurring revenue as of October 2025, supported by 12 million and a 67% in text-to-song generation. Competitor Udio trails with 28% share and 4.8 million users, while earlier entrants like AIVA maintain niche positions through licensing for media and advertising. These models prioritize high-margin software-as-a-service approaches, yielding profit margins akin to , though scalability is constrained by ongoing lawsuits from major labels alleging unauthorized training data use. Investments underscore optimism, with Suno in talks for a USD 100 million raise at a USD 2 billion valuation in October 2025, signaling venture capital's focus on AI's potential to create novel content streams despite legal headwinds.
Key Generative AI Music PlatformsEst. Annual Revenue (2025)User Base
Suno>USD 100M12M67%
UdioNot disclosed4.8M28%
While AI-driven growth introduces efficiencies—such as cost reductions in and personalized streaming recommendations enhancing retention—its revenue impact on the USD 29.6 billion global recorded remains marginal, representing less than 2% as of 2024. Projections of explosive expansion, including CISAC's forecast of the generative AI music rising from €3 billion to €64 billion by 2028, assume resolved licensing disputes and widespread integration; however, from current suggests tempered realism, as AI outputs often supplement rather than supplant human-created content due to and preferences among consumers and platforms. Institutional investors are increasingly music rights catalogs bolstered by AI for royalty forecasting, potentially stabilizing traditional s amid AI disruption.

Effects on Royalties and Distribution

The proliferation of AI-generated music has led to an influx of low-cost content on streaming platforms, diluting the royalty pool available to human-created works. In streaming economics, royalties are typically distributed from a fixed revenue pot divided by total streams, meaning increased volume from AI-assisted or fully synthetic tracks reduces per-stream payouts for established artists. A 2025 Deutsche Bank Research report notes that the surge in independent, AI-assisted releases could fragment the royalty pool, complicating monetization for major labels and reducing earnings for creators reliant on high-value streams. Similarly, Fitch Ratings warned in October 2025 that AI-generated music erodes demand for human content, potentially impairing royalty-backed asset-backed securities and overall artist income as synthetic tracks capture disproportionate streams. Platforms have observed direct impacts on distribution and payout integrity. reported receiving approximately 20,000 AI-generated tracks daily by mid-2025, with 70% of their streams originating from automated bots designed to inflate play counts and siphon royalties from legitimate creators. This bot-driven manipulation exploits algorithmic recommendations, prioritizing volume over quality and diverting funds intended for human artists. responded in September 2025 by enhancing detection tools to block harmful AI content that degrades listener experience and routes royalties to fraudulent actors, though critics argue that broader integration still favors cost-efficient synthetic music over human works, potentially lowering overall revenue shares for artists. Legal and policy responses underscore ongoing challenges to equitable distribution. The (RIAA) filed suits against AI firms like Suno and Udio in June 2024, alleging unauthorized use of copyrighted recordings for training data, which indirectly affects royalty flows by enabling unlicensed synthetic outputs to compete in distribution channels. Despite global recorded music revenues reaching $29.6 billion in 2024 per the IFPI Global Music Report—up 5% year-over-year with streaming exceeding $20 billion—industry bodies like IFPI and RIAA have emphasized that unmitigated AI proliferation risks systemic royalty erosion without opt-in licensing or transparency mandates for training datasets. These dynamics highlight a causal tension: while AI lowers and distribution for creators, it disproportionately burdens royalty-dependent artists by commoditizing content supply.

Employment and Skill Shifts

A study commissioned by the International Confederation of Societies of Authors and Composers (CISAC), representing over 4 million creators, projects that generative AI will place 24% of music sector revenues at risk by 2028, resulting in a cumulative loss of €10 billion over five years, as AI-generated content captures 20% of streaming revenues and 60% of music library revenues. Similarly, a 2024 survey by of over 4,200 music creators found 23% of revenues at risk by 2028, equivalent to AUD$519 million in damages, with 82% of respondents expressing concern that AI could undermine their ability to earn a living from music. These projections stem from AI's capacity to automate tasks in , engineering, potentially reducing demand for session musicians, stock music providers, and entry-level producers, as tools generate tracks efficiently without human labor. In , a 2025 Centre national de la musique report highlights automation of repetitive technical roles like mixing and rights detection, exacerbating fears of remuneration dilution amid daily uploads of over 20,000 AI-generated tracks on platforms like , comprising 18% of new releases. While displacement risks concentrate in routine creative and technical functions, AI introduces skill shifts toward hybrid competencies, requiring musicians and producers to master , data literacy, and AI tool integration to oversee outputs and ensure quality. Traditional roles evolve, with producers transitioning to creative directors who guide systems rather than manually adjusting elements, as seen in workflows using platforms like Suno for . Surveys indicate 38% of creators already incorporate to assist processes, suggesting via upskilling could mitigate losses, though a generational divide persists in technical proficiency. This demands training in validating AI-generated content for originality and coherence, preserving human judgment in areas like emotional nuance where falls short. Emerging roles reflect these shifts, including music designers who leverage AI to craft full songs, virtual star creators developing AI personas with integrated lyrics and visuals, and AI voice agents managing cloned vocal licensing for projects. Hologram specialists also arise, handling AI-enhanced performances of artists, as in productions. These positions prioritize oversight and curation over pure execution, potentially offsetting some displacements if regulated to favor human-AI , though critics from creator organizations argue tech firms capture disproportionate value without equitable . Overall, while AI automates low-skill tasks, sustained hinges on workers' ability to pivot to supervisory and innovative applications. The development of AI music generation systems has raised significant concerns, primarily due to the use of copyrighted sound recordings in training datasets without explicit permission from rights holders. Major record labels, represented by the (RIAA), filed lawsuits against AI companies Suno and Udio on June 24, 2024, alleging "mass infringement" through the unauthorized copying and exploitation of copyrighted works to train models capable of generating new music tracks. The complaints assert that these models were trained on datasets encompassing a substantial portion of commercially released sound recordings, enabling outputs that mimic protected styles, genres, and specific elements of existing songs, with potential statutory damages of up to $150,000 per infringed work. Independent artists have also pursued class-action suits, such as one filed against Suno in June 2025 by Anthony Justice and 5th Wheel Records, claiming the platform's AI generated tracks that directly competed with and diluted their original works. AI developers like Suno and Udio have defended their practices by invoking the doctrine under U.S. , arguing that training processes are transformative and do not involve verbatim reproduction or distribution of copyrighted material in outputs. This position aligns with rulings in non-music AI cases, such as Bartz v. and Kadrey v. in 2025, where courts deemed ingestion of copyrighted texts for model training as due to its transformative nature in creating new generative capabilities without supplanting original markets. However, these precedents contrast with decisions like v. Ross Intelligence in February 2025, where a court rejected for AI training on proprietary legal databases, emphasizing that commercial exploitation without licensing undermined the doctrine's market harm factor. Music-specific litigation remains unresolved as of October 2025, with no appellate rulings conclusively applying to audio training data, leaving uncertainty over whether ingestion alone constitutes infringement or if outputs must demonstrably copy protected elements. In response to these disputes, licensing frameworks for training have emerged as a potential resolution, with industry players negotiating agreements to authorize use of musical works. Rightsify, a music provider, offers licensed datasets specifically for model training, enabling developers to access cleared content from catalogs while compensating rights holders. Germany's GEMA proposed a licensing model in September 2024 that extends beyond one-time fees to include ongoing royalties tied to -generated outputs derived from licensed training . Recent deals include Spotify's October 2025 partnerships with , , and to develop music tools using licensed catalogs, ensuring and . Additionally, major publishers signed licensing pacts with lyrics provider in October 2025, granting access to and content for training while addressing attribution. and neared "landmark" comprehensive licensing deals with firms by early October 2025, signaling a shift from litigation toward compensated access, though such arrangements may increase training costs significantly compared to unlicensed scraping. Despite these advances, most music models to date have relied on unauthorized datasets aggregated from public sources like streaming platforms, prompting U.S. Office scrutiny into whether such practices align with existing law.

Authorship, Originality, and Human-AI Collaboration

In the United States, protection for AI-generated musical works requires demonstrable authorship, as affirmed by the U.S. Copyright Office in its January 29, 2025, report on generative AI outputs, which states that such works are registrable only if a has contributed sufficient expressive elements beyond mere prompts or selections. This position aligns with judicial precedents like Thaler v. Perlmutter (2023), where the U.S. District Court for the District of Columbia ruled that AI cannot be an author under law, emphasizing that protection demands intellectual contribution to . For music specifically, purely AI-composed tracks—generated without modification of core elements like , , or structure—remain ineligible for , leaving them in the and vulnerable to unrestricted use. Originality in AI-assisted music hinges on the human's role in infusing novel, creative expression, as models trained on vast datasets often produce outputs derivative of existing patterns rather than inventive ones. The U.S. Copyright Office guidance from March 16, 2023, clarifies that registration applicants must disclaim -generated portions, with protection extending solely to human-authored additions, such as custom lyrics or arrangements that exhibit minimal creativity under the Feist Publications, Inc. v. Rural Telephone Service Co. (1991) standard. A March 21, 2025, U.S. Court of Appeals decision reinforced this by denying to fully -generated visual works, a principle extending to audio where outputs lack human-driven fixation of original expression, potentially complicating claims in genres like electronic or reliant on algorithmic generation. Critics argue this framework undervalues 's pattern synthesis as a form of originality, but courts prioritize causal human agency over machine computation. Human-AI in music typically involves humans providing iterative inputs—such as detailed prompts for , , or —followed by AI outputs to assert authorship, as explored in a 2023 ISMIR conference paper modeling flexible AI roles from ideation to refinement. For instance, producer employed in 2023 to generate emulating Eminem's , then refined them manually for his track "Crazy," claiming over the human-curated final product. Empirical studies, including a September 2025 preprint on text-to-music interfaces, show experienced producers using for (e.g., stem generation) while applying expertise in mixing and personalization, enhancing efficiency without supplanting creative control. An ACM from 2024 involving undergraduate composers demonstrated that hybrid workflows—where handles initial drafts and humans iterate for —yield tracks eligible for when human contributions dominate expressive choices. Such raises questions of attribution, with some advocating shared credits, though legal systems currently vest with the human collaborator.

Bias, Fairness, and Cultural Representation

AI music generation models exhibit significant representational biases stemming from the composition of datasets, which predominantly feature genres. A 2025 study by researchers at Mohamed bin Zayed University of (MBZUAI) analyzed major datasets used for generative music models and found that approximately 94% of the data derives from musical styles, such as pop, rock, and classical, while only 5.7% originates from non- genres like Indian ragas or African rhythms. This imbalance reflects the availability of digitized music corpora, which are skewed toward commercially dominant sources like streaming platforms, rather than deliberate exclusion, leading models to underperform on diverse cultural inputs. For instance, when evaluated on the GlobalDISCO —a globally balanced collection spanning 79 countries—state-of-the-art models demonstrated marked disparities in generating coherent outputs for non- regions and genres, with fidelity dropping by up to 40% compared to benchmarks. These dataset-driven biases raise fairness concerns, as models trained on such corpora may marginalize underrepresented traditions, potentially perpetuating cultural hierarchies in -assisted music production. Empirical tests in the MBZUAI study revealed that unmodified models generated outputs with lower structural coherence and stylistic accuracy for non-Western prompts, attributing this to insufficient training exposure rather than algorithmic flaws . Complementary on marginalized genres highlights that even when non-Western is included, it constitutes less than 15% in aggregate across surveyed datasets, limiting model adaptability and risking outputs that superficially mimic rather than authentically recreate cultural nuances. Initiatives like the Music RAI project, launched in 2024, seek to mitigate this by developing techniques for smaller, targeted datasets of underrepresented music, employing explainable to identify and reduce over-reliance on Western-centric patterns without artificially inflating minority representations, which could introduce contrived biases. Cultural representation challenges extend to risks of inadvertent appropriation, where models blend elements from scarce non-Western into outputs, potentially diluting original contexts without crediting sources. A 2024 analysis noted that while AI can synthesize fusions, the scarcity of high-quality training for Global South genres—averaging 14.6% coverage—results in outputs that prioritize structures over idiomatic scales or rhythms, as evidenced by listener evaluations showing reduced perceived . Addressing fairness requires causal interventions at the level, such as curated expansions, but suggests that simply diversifying inputs improves generation without compromising overall model utility, as demonstrated in augmented training experiments yielding 20-30% gains in non-Western fidelity. Nonetheless, academic sources, often institutionally aligned with mandates, occasionally overemphasize narratives while underreporting how market-driven availability causally determines these outcomes, underscoring the need for transparent auditing in model .

Controversies and Deepfakes

Artist and Industry Backlash

In April 2024, over 200 prominent musicians, including , , , and , signed an organized by the Artist Rights Alliance demanding legislative protections against the unauthorized use of artists' voices and likenesses in -generated music, citing risks to creative incentives and economic livelihoods. The letter emphasized that tools trained on copyrighted works undermine artistry and called for regulations to prevent "predatory" , reflecting widespread concerns over 's potential to flood markets with low-cost imitations. The (RIAA), representing major labels such as , Sony Music Entertainment, and , escalated industry opposition through lawsuits filed against AI music generators Suno and Udio on June 24, 2024, in U.S. federal courts in and . These suits allege that the companies trained their models on vast datasets of copyrighted sound recordings scraped without permission or licensing, enabling the generation of music that directly mimics protected works and competes with original releases. By September 2025, the RIAA amended the complaint against Suno, accusing it of further infringement via "stream ripping" techniques to pirate songs from platforms like for training data, highlighting ongoing disputes over AI developers' practices. Complementing legal actions, music industry stakeholders issued the "Principles for Music Creation with AI" on June 24, 2024, advocating for transparent, licensed AI development that respects copyrights and compensates creators, as articulated by organizations including the RIAA and labels. These efforts underscore a unified push against unlicensed AI tools, driven by evidence that such systems ingest billions of streams of protected material to produce outputs that dilute royalties and for human-made music. Critics within the industry argue that without enforceable boundaries, AI exacerbates existing revenue pressures, though proponents of the technology counter that backlash stems partly from fears of disruption rather than inherent illegality.

Specific Deepfake Cases and Responses

In April 2023, an anonymous creator released "Heart on My Sleeve," an AI-generated song using voice cloning technology to mimic and , which quickly gained over 15 million streams across , , and other platforms. The track's realistic synthesis, achieved via tools like those from , sparked widespread fan engagement but alarmed the industry over unauthorized voice replication. (UMG), representing both artists, responded by pressuring streaming services to remove the song, citing potential infringement on publicity rights and contractual obligations to protect artist likenesses; platforms including and complied within days. During the May 2024 rap feud between and , released "Taylor Made Freestyle," incorporating AI-synthesized voices of and to diss Lamar, which was uploaded to and viewed millions of times. The estate swiftly issued a cease-and-desist letter, arguing the violated the late rapper's right of publicity and risked misleading the public; deleted the track hours later to avoid litigation. publicly condemned the use of for deceased artists, emphasizing ethical boundaries in voice manipulation. These incidents prompted broader industry actions, including Tennessee's ELVIS Act, signed into on March 21, 2024, which criminalizes unauthorized commercial use of -generated deepfakes replicating a performer's voice or likeness without consent. Major labels escalated enforcement, with disclosing over 75,000 takedowns of infringing deepfakes by March 2025, often via DMCA notices to platforms hosting cloned content. In the UK, (PPL) threatened the first deepfake-specific lawsuit in March 2024 against an service generating unauthorized artist vocals, aiming to establish precedents for performers' rights. Federally, the proposed NO FAKES Act, introduced in 2023 and reintroduced in 2024, seeks to create a national right of against deepfakes, supported by testimony from artists like highlighting risks to vocal integrity.

Debates on Creativity and Replacement

The debate centers on whether can exhibit genuine in generation or merely recombine existing patterns from training data, and whether it poses an existential threat to human musicians. Proponents argue that augments human by accelerating ideation and enabling novel combinations; a 2024 study of 663 practitioners found that higher acceptance correlated with enhanced perceived (β = 0.60, p < 0.001), particularly among those with formal and software experience, suggesting tools expand creative efficiency and freedom. Similarly, experimental research indicates generative boosts individual output novelty and usefulness in creative tasks, with less skilled creators benefiting most (up to 11% increase). However, these gains come at a cost to collective diversity, as -assisted works converge toward similar outputs, potentially homogenizing musical akin to a "social dilemma" where individual advantages erode broader variety. Critics contend AI-generated music lacks authentic rooted in , , and , functioning instead as statistical without causal understanding or emotional depth. Evaluations of AI-composed pieces show that while music quality ratings remain stable regardless of manipulated AI attributes like , acceptance of AI as a "" leads to higher scores, implying in rather than inherent merit. Award-winning Joel Beckerman has emphasized that AI cannot replicate the personal in works, such as themes drawn from real-life heartbreak, nor program the unexplained essence of . This view aligns with concerns that AI's nature—trained on vast human corpora without compensation—undermines originality, as evidenced by industry warnings of market saturation diluting independent artistry by 2025. Regarding replacement, over 200 prominent musicians, including and , signed an April 2024 open letter via the decrying 's predatory use to infringe on voices, likenesses, and royalties, urging safeguards to prevent devaluing human artistry while acknowledging ethical 's potential as a tool. Experts maintain will not fully supplant humans, citing irreplaceable elements like live performances and emotional resonance that demand physical presence and biological authenticity, though it may displace routine tasks such as commercial jingles or background scoring. Empirical limits persist: struggles with unprogrammable spontaneity and contextual adaptation, ensuring human roles in high-value creation endure, even as tools like vocal synthesis reshape production pipelines.

References

  1. [1]
    A systematic review of artificial intelligence-based music generation
    Dec 15, 2022 · In this work, we analyze the scope and trends of the research on artificial intelligence-based music generation by performing a systematic review of the ...
  2. [2]
    Geoff Hill First Demonstrates Computer Music on the University of ...
    During August 7 to 9, 1951 Geoff Hill, a computer programmer with perfect pitch, programmed the University of Melbourne CSIR Mk1, the first stored-program ...
  3. [3]
    The Intersection of Music and Generative AI - Founding Minds
    Jun 13, 2024 · In 1957, the “Illiac Suite” by Lejaren Hiller and Leonard Isaacson became the first musical score generated with the help of a computer. This ...
  4. [4]
    Music and artificial intelligence - Wikipedia
    History. edit. In the 1950s and the 1960s, music made by artificial intelligence was not fully original, but generated from templates that people had already ...History · Software applications · Musical deepfakes · Technical approaches
  5. [5]
    How AI is transforming the creative economy and music industry
    Apr 3, 2024 · In 2016, AI created “Daddy's Car (opens in a new window),” a song generated in the style of The Beatles with human-written lyrics. AI analyzed a ...
  6. [6]
    60 million people used AI to create music in 2024, IMS Business ...
    Apr 24, 2025 · 10% of consumers surveyed said they had used generative AI to create music or lyrics in 2024.
  7. [7]
    Advancing deep learning for expressive music composition ... - Nature
    Jul 31, 2025 · Through the power of artificial intelligence (AI), users can obtain generated melodies, piece harmonization, and composer style emulation.<|separator|>
  8. [8]
    Case Studies: Legal Battles Over AI-Generated Music - CoverNet
    Jul 16, 2024 · In June 2024, the RIAA filed lawsuits against AI music platforms Suno and Udio, accusing them of copyright infringement. The RIAA claimed these ...
  9. [9]
    Legal Riffs: Music Industry Alleges AI Is Out of Tune
    Jul 10, 2024 · The complaints allege that gen AI technology produced by Suno and Udio directly infringes on copyrights owned by Universal Music Group (UMG) Records.
  10. [10]
    Music That Is Entirely AI-Generated Cannot Be Copyrighted, but ...
    Jun 24, 2024 · While entirely AI-generated music cannot be copyrighted, the circumstances are different for AI-assisted songs.
  11. [11]
    Defining Authorship for the Copyright of AI-Generated Music
    Jan 5, 2025 · Under the U.S. Copyright Office provision, AI-generated art cannot be protected under copyright because of the human authorship clause. But, by ...Missing: controversies | Show results with:controversies
  12. [12]
    AI and the Sound of Music - The Yale Law Journal
    Nov 22, 2024 · AI is rapidly changing the way music is created. This transformation is part of the long evolution of music creation, which has become ...
  13. [13]
    CSIR Mk1 & CSIRAC, Trevor Pearcey & Geoff Hill, Australia, 1951
    The first piece of digital computer music was created by Geoff Hill and Trevor Pearcey on the CSIR Mk1 in 1951 as a way of testing the machine rather than a ...
  14. [14]
    How Australia played the world's first music on a computer
    Jun 26, 2016 · Shortly after CSIRAC first played music, in 1951 the BBC recorded a Ferranti Mark 1 computer playing music in Manchester, England. That is the ...
  15. [15]
    Making Music with Computers - CHM Revolution
    The “Colonel Bogey March” wasn't new in 1951. But hearing it from a computer was. The performance by Australia's first computer, CSIRAC, was among the earliest, ...
  16. [16]
    (PDF) Turing and the History of Computer Music - ResearchGate
    Jun 19, 2025 · The story of Turing's pioneering work in creating the first computer-generated musical notes in Manchester in 1948–1949 is told, as well as ...
  17. [17]
    Electronic and Computer Music – ETHW
    Mathews wrote MUSIC-I, which was the first program to produce digital audio through use of an IBM 704 computer; the CSIRAC and Ferranti only produced analog ...
  18. [18]
    ILLIAC Suite - Illinois Distributed Museum
    The Illiac Suite is the first computer-composed music, created by Lejaren Hiller and Leonard Isaacson using the ILLIAC I, with the last movement using non- ...
  19. [19]
    The First Significant Computer Music Composition
    In 1957 Lejaren Hiller ... collaborated on the first significant computer music composition, the Illiac Suite, composed on the University of Illinois ...
  20. [20]
  21. [21]
    The History of Algorithmic Composition - CCRMA - Stanford University
    The idea of utilizing formal instructions and processes to create music dates back in musical history as far back as the ancient Greeks.
  22. [22]
    [PDF] AI Methods in Algorithmic Composition: A Comprehensive Survey
    The earliest use of computers to compose music dates back to the mid-1950s, roughly at the same time as the concept of Artificial Intelligence was coined at the ...<|control11|><|separator|>
  23. [23]
    [PDF] 69 A Functional Taxonomy of Music Generation Systems
    The first attempts at generating melodies with computers date back to 1956, when. Pinkerton built a first order Markov model, the “Banal Tune-Maker”, based on a ...<|separator|>
  24. [24]
    [PDF] Revisiting the Illiac Suite – a rule based approach to stochastic ...
    Feb 1, 2010 · Hiller and Isaacson describe the Illiac Suite as a chronological record of experiments. The general idea is to use screening rules to accept or ...
  25. [25]
    38. The Algorithmic Music of Iannis Xenakis–What's Next?
    Oct 9, 2024 · Iannis Xenakis was a pioneer in algorithmic composition of music and art. He combined architecture, mathematics, music, and performance art ...
  26. [26]
    [PDF] Automatic Species Counterpoint - Stanford CCRMA
    AUTOMATIC SPECIES COUNTERPOINT by. Bill Schottstaedt. May 1984. Species counterpoint as presented by J. J. Fux in Gradus Ad Parnassum appears to be a ready ...Missing: date | Show results with:date
  27. [27]
    [PDF] From Artificial Neural Networks to Deep Learning for Music Generation
    Apr 7, 2020 · The Illiac Suite is the first score composed by a computer [20] and was an early example of algorithmic music composition, making use of ...<|separator|>
  28. [28]
    [PDF] Connectionist Music Composition Based on Melodic and Stylistic ...
    We describe a recurrent connectionist network, called CONCERT, that uses a set of melodies written in a given style to compose new melodies in that.
  29. [29]
    [PDF] Artificial Intelligence, Machine Learning, and Music Understanding
    AI and Machine Learning are used in music for computer accompaniment, interactive composition, and music synthesis, enabling new ways of thinking about music.
  30. [30]
    What is the best AI Music Generator right now? - CometAPI
    Aug 14, 2025 · What it is: Suno is a consumer-facing, text-to-song generator that made rapid headway by producing full songs (instrumental and vocal content) ...
  31. [31]
    Best AI Music Generators Compared: Suno Vs Udio Vs Stable Audio
    Rating 4.4 · Review by Marga BagusMay 20, 2025 · To provide a clear comparison of AI music creation tools, I evaluated all three platforms across key performance metrics: Feature, Suno, Udio ...
  32. [32]
    10 Best Stable Audio Alternatives for Streamlined AI Music Production
    Udio - Best Stable Audio Alternative For Potent Text-To-Music Generator. Udio is a renowned AI music generator celebrated for its ability to streamline music ...
  33. [33]
    AI Music Generators: The Future of Composition in 2025 - MIDIGEN
    Sep 7, 2025 · In 2025, AI music generation has reached unprecedented levels of sophistication, with systems capable of understanding complex musical theory, ...Missing: advancements | Show results with:advancements
  34. [34]
    Global Generative Ai In Music Market Size & Outlook
    The global generative ai in music market size was estimated at USD 569.7 million in 2024 and is projected to reach USD 2,794.7 million by 2030, growing at a ...
  35. [35]
    AI in Music Industry Statistics 2025: Market Growth & Trends
    Nov 5, 2024 · Global AI in the Music market was valued at $2.9 billion in 2024. Cloud-based solutions dominate with 71.4% of AI music market share.
  36. [36]
    20 Best Suno AI Alternatives & Competitors in 2025 - Techjockey
    More versatile AI music tools similar to Suno AI are getting popular, such as AIVA, Boomy, Soundraw, and Loudly. These AI music generators offer finer control ...
  37. [37]
    AI Music Generator Suno in Funding Talks at $2 Billion Valuation
    Oct 17, 2025 · Suno Inc., a startup that generates music using artificial intelligence, is in talks to raise over $100 million at a valuation of more than ...
  38. [38]
    AI music generation is now really, really good(*) – Joe Pater
    Jan 6, 2025 · I just found out how good AI music generation has become, and have spent a little time finding out what I could about the technology.
  39. [39]
    Three music AI trends to watch in 2025: Recording & takeaways
    Nov 4, 2024 · Three music AI trends to watch in 2025: Recording & takeaways · 1. Synthetic data · 2. Model controllability · 3. AI content & copyright detection.Follow Along · 1. Synthetic Data · 3. Ai Content & Copyright...<|separator|>
  40. [40]
    AI's Impact On Music In 2025: Licensing, Creativity And Industry ...
    Dec 30, 2024 · By mid-2025, AI companies could operate with near-impunity, churning out music based on unlicensed datasets. The stakes have never been higher ...
  41. [41]
    How AI is Changing the Music Industry in 2025 - Eventbrite
    May 22, 2025 · Musicians are turning to AI programs to create new music, like BandLab, which suggests musical loops based on prompts. Listeners are discovering ...Protecting Creative Work · Embracing Music Innovation · Losing Income To Ai
  42. [42]
    10 Ways AI Is Likely to Shape the Music Industry in 2025 -
    Dec 29, 2024 · By 2025, AI will power recording, fan interaction, and influence creation, distribution, and consumption, becoming a firmly embedded element of ...
  43. [43]
    Music formats - symbolic and audio - MidiTok's docs
    There are three major file formats for symbolic music: MIDI, abc and musicXML. ... The common file extensions are .mxl and .musicxml . Next. The MIDI ...
  44. [44]
    [PDF] MUSPY: A TOOLKIT FOR SYMBOLIC MUSIC GENERATION
    In contrast, MusicXML, as a sheet music exchanging format, has the concepts of notes, measures and symbolic musi- cal markings and contains visual layout ...
  45. [45]
    Symbolic Music Generation Using Deep Neural Networks | Towards AI
    May 22, 2024 · MIDI, an ultimate way to store symbolic music generation, is the format used most often in the modern music industry. It includes almost ...
  46. [46]
    [PDF] SYMBOLIC MUSIC REPRESENTATIONS FOR CLASSIFICATION ...
    In this paper, we investigate matrix (piano roll), se- quence, and graph representations and their corresponding neural architectures, in combination with ...
  47. [47]
    A Survey on Deep Learning for Symbolic Music Generation
    This paper attempts to provide a task-oriented survey of symbolic music generation based on deep learning techniques, covering most of the currently popular ...<|separator|>
  48. [48]
    SAGE-Music: Low-Latency Symbolic Music Generation via Attribute ...
    Oct 1, 2025 · Low-latency symbolic music generation is essential for real-time improvisation and human-AI co-creation. Existing transformer-based models, ...
  49. [49]
    How does AI create music - j-labs
    Jun 2, 2021 · There are multiple approaches that can be utilized to perform the task of music generation, such as Markov chains, grammar-based and rule-based models, and ...
  50. [50]
    salu133445/muspy: A toolkit for symbolic music generation - GitHub
    Dataset management system for commonly used datasets with interfaces to PyTorch and TensorFlow. Data I/O for common symbolic music formats (e.g., MIDI, MusicXML ...
  51. [51]
    MusicLang - a controllable model for symbolic music generation ...
    Mar 14, 2024 · I'm working with u/symphonia on MusicLang, a foundation model for symbolic music generation. After 18 months of R&D, we've released our work on our HuggingFace.
  52. [52]
    XMusic: Towards a Generalized and Controllable Symbolic Music ...
    Jan 15, 2025 · This paper presents a generalized symbolic music generation framework, XMusic, which supports flexible prompts (ie, images, videos, texts, tags, and humming)
  53. [53]
  54. [54]
    [1609.03499] WaveNet: A Generative Model for Raw Audio - arXiv
    Sep 12, 2016 · This paper introduces WaveNet, a deep neural network for generating raw audio waveforms. The model is fully probabilistic and autoregressive.
  55. [55]
    Jukebox | OpenAI
    Apr 30, 2020 · We're introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles.
  56. [56]
    NSynth: Neural Audio Synthesis - Google Magenta
    Apr 6, 2017 · NSynth is a novel music synthesis approach using deep neural networks to generate sounds at the level of individual samples, unlike traditional ...
  57. [57]
    WaveNet - Google DeepMind
    Introduced in 2016, WaveNet was one of the first AI models to generate natural-sounding speech. Since then, it has inspired research, products, and applications ...
  58. [58]
    [2005.00341] Jukebox: A Generative Model for Music - arXiv
    Apr 30, 2020 · We introduce Jukebox, a model that generates music with singing in the raw audio domain. We tackle the long context of raw audio using a multi-scale VQ-VAE.
  59. [59]
    [2306.05284] Simple and Controllable Music Generation - arXiv
    Jun 8, 2023 · We introduce MusicGen, a single Language Model (LM) that operates over several streams of compressed discrete music representation, ie, tokens.
  60. [60]
    Meta's new MusicGen AI was trained on 20k hours of (licensed) music
    Jun 12, 2023 · It's called MusicGen, and is described as a simple and controllable music generation LM [language model] with textual and melodic conditioning.
  61. [61]
    riffusion/riffusion-hobby: Stable diffusion for real-time music generation
    Dec 27, 2022 · Riffusion is a library for real-time music and audio generation with stable diffusion. Read about it at https://www.riffusion.com/about and try ...
  62. [62]
    riffusion/riffusion-model-v1 - Hugging Face
    Riffusion is a latent text-to-image diffusion model capable of generating spectrogram images given any text input. These spectrograms can be converted into ...
  63. [63]
    Recent developments in Generative AI for Audio - AssemblyAI
    Jul 27, 2023 · Over the past decade, we've witnessed significant advancements in AI-powered audio generation techniques, including music and speech synthesis.
  64. [64]
    Applications and Advances of Artificial Intelligence in Music ... - arXiv
    Sep 3, 2024 · This paper provides a systematic review of the latest research advancements in AI music generation, covering key technologies, models, datasets, evaluation ...
  65. [65]
    Music Information Retrieval - an overview | ScienceDirect Topics
    1. Introduction to Music Information Retrieval · 2. Music Data Representations and Feature Extraction · 3. Core Computational Techniques in Music Information ...Introduction to Music... · Core Computational... · Applications of Music...
  66. [66]
    [PDF] Music Information Retrieval: Recent Developments and Applications
    We present here the main techniques for music content description, focusing on the analysis of music audio signals. This description is cru- cial for MIR ...
  67. [67]
    [PDF] ISMIR 2024 Tutorial: Exploring 25 Years of Music Information Retrieval
    Nov 10, 2024 · Query-by-humming research began in the 1990s. • T. Kageyama, K. Mochizuki, and Y. Takashima: Melody retrieval with humming, ICMC 1993.Missing: milestones | Show results with:milestones
  68. [68]
    ISMIR | International Society for Music Information Retrieval
    Since its inception in 2000, the ISMIR community mailing list has grown into a forum of over 1,800 members from across the world, and routinely receives ...Ismir 2025 · Transactions of the... · About the Society · Past ConferencesMissing: history milestones
  69. [69]
    Bridging technology and musical creativity: ISMIR 2024
    Mar 17, 2025 · ISMIR 2024 published a total of 123 papers, showcasing cutting-edge research in the field of music information retrieval, including some of the ...
  70. [70]
    Advancements in the Music Information Retrieval Framework ...
    AMUSE (Advanced MUSic Explorer) was created 2006 as an open-source Java framework for various music information retrieval tasks like feature extraction, ...
  71. [71]
    Music Information Retrieval and Contemporary Classical Music
    Sep 1, 2020 · Humphrey et al. (2015) presented an important overview of deep learning techniques for music informatics. I believe that an important ...
  72. [72]
    A Tutorial on Deep Learning for Music Information Retrieval - ar5iv
    The goal of this paper is to boost the interest of beginners by providing a comprehensive tutorial and reducing the barriers to entry into deep learning for MIR ...
  73. [73]
    AI in the Music Industry – Part 4: AI in Music Recognition and ...
    Feb 26, 2024 · In this part of our series on AI in the music industry, we will look at how music recognition and music recommendations work in music streaming services.
  74. [74]
    [PDF] Music Retrieval System Using Query-by-Humming
    A famous example of this type of technique that is being used in real-world applications is 'Shazam'. They use a method known as Audio Fingerprinting [3] to ...
  75. [75]
    [PDF] Hybrid Symbolic-Waveform Modeling of Music - CEUR-WS.org
    Oct 19, 2024 · Abstract. Generative modeling of music is a challenging task due to the hierarchical structure and long-term dependencies.
  76. [76]
    A Novel Framework for Automatic Music Generation Using ‎Hybrid AI ...
    This research introduces a novel and strong approach is MuseHybridNet- a hybrid model designed to carry ‎forward the boundaries of AI-generated music. ‎ What ...
  77. [77]
    A hybrid deep learning approach for musical difficulty estimation of ...
    Musical Instrument Digital Interface (MIDI)1 is the most commonly used format for symbolic music. It was initially designed to allow communication between a ...
  78. [78]
    A theme music generation model based on hybrid variational ...
    In this paper, we propose a thematic music generation model based on a generative adversarial network, named Thematic Music Conditional Generative Adversarial ...
  79. [79]
    LLark: A Multimodal Foundation Model for Music - Spotify Research
    Oct 13, 2023 · LLark is a research exploration into the question: how can we build a flexible multimodal language model for music understanding?
  80. [80]
    MusDiff: A multimodal-guided framework for music generation
    We propose MusDiff, a multimodal music generation framework that combines text and image inputs to enhance music quality and cross-modal consistency.
  81. [81]
    MuMu-LLaMA: Multi-modal Music Understanding and Generation ...
    Dec 9, 2024 · This framework integrates LLMs to comprehend input music and generate music across various modalities, utilizing pretrained models for music, images, and ...
  82. [82]
    Mozart's Touch: a lightweight multimodal music generation ...
    Jul 7, 2025 · The framework consists of three key components: Multi-modal Captioning Module, Large Language Model (LLM) Understanding & Bridging Module, and ...
  83. [83]
    Multi-modal music techniques for synthesizing high-quality audio ...
    Sep 26, 2025 · Such advancements enable the synthesizer to exert precise control over musical composition and instrument processing at the note level, ...
  84. [84]
    A Survey on Multimodal Music Emotion Recognition - arXiv
    Apr 26, 2025 · The paper introduces a four-stage MMER framework, including multimodal data selection, feature extraction, feature processing, and final emotion prediction.
  85. [85]
    12 AI Music Generators That Create Original Songs in 2025
    Sep 3, 2025 · Explore our list of AI music generators that creative professionals can use for videos, podcasts, and social media content.
  86. [86]
    AI Music Composition Tools: Top 15 Picks of 2025 - Staccato AI
    Jan 8, 2025 · 6. Lemonaide: AI HipHop Generator · 7. Unison MIDI Wizard: Chord & Melody Generator · 8. Hookpad Aria: Songwriting Assistant Within Hookpad.
  87. [87]
    Top AI Tools That Actually Help Songwriters - Indie International
    Sep 29, 2025 · Top AI Tools That Actually Help Songwriters · Lyric Generation And Brainstorming Tools · AI-Assisted Melody Creation · Chord Progression Generators.Lyric Generation And... · Ai-Assisted Melody Creation · Ai-Based Beat And Groove...
  88. [88]
    AIVA, the AI Music Generation Assistant
    AIVA is an AI music generation assistant that allows you to generate new songs in more than 250 different styles, in a matter of seconds.Missing: capabilities | Show results with:capabilities
  89. [89]
    AIVA AI In-Depth Review (2025): Your AI Music Co-Composer
    Oct 7, 2025 · AIVA is a sophisticated and powerful AI composer that shines for instrumental music creation. Its combination of high-quality generation, deep ...
  90. [90]
    AIVA at the AI Art Gallery - NVIDIA
    AIVA is an AI music composer trained on thousands of scores of music, helping people create their own personalized soundtracks.
  91. [91]
    Suno | AI Music
    Create stunning original music for free in seconds using AI. Make your own masterpieces, share with friends, and discover music from artists worldwide.About · Pricing · Knowledge Base · Making Music
  92. [92]
    Suno's Latest AI Music Generator: 5 Reasons It's a Game-Changer
    Dec 1, 2024 · New Features and Upgrades in Suno v4 · Covers: Upload any audio track and reimagine it through custom prompts. · Personas: Create and save ...
  93. [93]
    Udio | AI Music Generator - Official Website
    Discover, create, and share music with the world. Use the latest technology to create AI music in seconds.
  94. [94]
    Meet Udio — the most realistic AI music creation tool I've ever tried
    Apr 10, 2024 · Udio is the latest artificial intelligence music tool to hit the market, coming out of stealth with a bang as it unveils an uncanny ability to capture emotion ...
  95. [95]
    Could The Future Of Music Composition Be AI? Analysing ... - Medium
    Jul 1, 2022 · AIVA can reduce that learning process down to a couple of hours.” AI can't create innately, it relies on datasets and past examples of music to ...
  96. [96]
  97. [97]
    Using Udio AI for Music Production — A Fun and Useful Tool, or The ...
    Dec 3, 2024 · In this article, I'm going to break down how I used Udio AI during the songwriting phase of a production, which I then recorded, mixed, ...
  98. [98]
    [2508.04651] Live Music Models - arXiv
    Aug 6, 2025 · These models demonstrate a new paradigm for AI-assisted music creation that emphasizes human-in-the-loop interaction for live music performance.
  99. [99]
    [2502.21267] ReaLJam: Real-Time Human-AI Music Jamming with ...
    Feb 28, 2025 · We introduce ReaLJam, an interface and protocol for live musical jamming sessions between a human and a Transformer-based AI agent trained with reinforcement ...
  100. [100]
    Magenta RealTime: An Open-Weights Live Music Model
    Jun 20, 2025 · Magenta RealTime is an open-weights live music model that allows interactive creation, control, and performance of music in real-time. It is ...
  101. [101]
    Research and Development - AI Music Ensemble Technology
    We are developing an AI system that plays together with you as an ensemble partner. The AI compares in real time the notes you are playing to the sheet music.
  102. [102]
    Lyria RealTime - Google DeepMind
    Lyria RealTime unlocks the world of real-time music generation, allowing anyone to interactively create, control, and perform music in the moment.
  103. [103]
    Intelligent Music Performance Systems: Towards a Design Framework
    Oct 31, 2023 · This article attempts to define and typologize the main principles for the design of Intelligent Music Performance Systems (IMPS).
  104. [104]
  105. [105]
    Can AI mix better than a real mix engineer? We put iZotope's AI ...
    Dec 12, 2024 · Neutron 5 is a suite of 10 processors ranging from the conventional to the exotic. Specifically, these are Gate, Equalizer, Compressor, ...
  106. [106]
  107. [107]
  108. [108]
    LANDR Mastering Plugin review - MusicRadar
    Rating 4.0 · Review by Stuart AdamsFeb 13, 2024 · Pros · Easy to get good results. · Impressive frequency balancing. · AI-generated baseline settings can be subsequently modified.What is it? · Performance and verdict
  109. [109]
  110. [110]
    Listening without ears: Artificial intelligence in audio mastering
    Nov 1, 2018 · The research empirically studies the actual practices surrounding algorithmic technologies through interviews with those affected by and ...
  111. [111]
    [PDF] Machine learning in audio mastering: a comparative study - DergiPark
    Mar 30, 2025 · Empirical research reveals AI mastering causes greater distortion, narrower dynamic range, and higher loudness penalties. In contrast ...
  112. [112]
    Creative places for artificial intelligence in audio mastering
    In this paper, we consider the impact of artificial intelligence (AI) in the creative economy of music production.Missing: studies | Show results with:studies
  113. [113]
    Cryo Mix | AI Mixing & Mastering
    Elevate your tracks with Cryo Mix's AI Mixing and Mastering tool, offering instant, precise enhancements. Ideal for music professionals and enthusiasts ...
  114. [114]
    Mixing and Mastering: A Blend of Technology and Human Touch
    Jul 16, 2024 · AI-driven tools for mixing and mastering have emerged, promising to expedite these complex processes. These tools analyze your tracks and provide automated ...
  115. [115]
    Exploring the Role of AI and Personalization in Music Streaming
    Feb 14, 2024 · Leading music streaming platforms rely heavily on AI algorithms to understand user preferences and deliver personalized playlists and recommendations.
  116. [116]
    How Spotify Uses AI (And What You Can Learn from It)
    Jan 26, 2024 · Spotify's Discover Weekly is a series of custom-made AI playlists, uniquely crafted each Monday for every user by the platform's AI algorithms. ...
  117. [117]
    How to break free of Spotify's algorithm | MIT Technology Review
    Aug 16, 2024 · According to a 2022 report published by Distribution Strategy Group, at least 30% of songs streamed on Spotify are recommended by AI. The ...
  118. [118]
    Music content personalized recommendation system based on a ...
    Dec 12, 2023 · This paper proposed a personalized music recommendation system powered by deep learning, explicitly utilizing convolutional neural networks (CNNs)
  119. [119]
    Contextualized Recommendations Through Personalized Narratives ...
    Dec 18, 2024 · Launched in 2023, DJ is a personalized AI guide that deeply understands listeners' music tastes, providing tailored song selections and ...
  120. [120]
    AI Music Statistics 2025: Trends, Insights, and Future Predictions
    The global AI-generated music market is expected to reach approximately $2 billion, growing at a compound annual growth rate (CAGR) of nearly 30% from previous ...
  121. [121]
  122. [122]
    Personalizing Agentic AI to Users' Musical Tastes with Scalable ...
    In this post, we present our approach using LLM-based agentic systems that interpret queries, orchestrate tools, and adapt through preference learning, going ...
  123. [123]
    Diversity by Design in Music Recommender Systems
    In this overview, we address the concept of diversity in music recommendation, and taking a value-driven approach we review diversity-related methodologies.
  124. [124]
    Assessing the Impact of Music Recommendation Diversity on Listeners
    We show that exposure to specific levels of music recommendation diversity may be responsible for long-term impacts on listeners' attitudes.<|control11|><|separator|>
  125. [125]
    The impact of algorithmically driven recommendation systems on ...
    Feb 9, 2023 · The impact of streaming platforms on musical production, consumption and culture. Anxieties about “algorithms” have been a regular feature of these debates.
  126. [126]
    Google Magenta
    It is a collection of music creativity tools built on Magenta's open source models, using cutting-edge machine learning techniques for music generation. Read ...Demos · Magenta Studio - Ableton Live... · Getting Started · Blog posts
  127. [127]
    Magenta: Music and Art Generation with Machine Intelligence - GitHub
    Magenta is a research project exploring the role of machine learning in the process of creating art and music.
  128. [128]
    Behind Magenta, the tech that rocked I/O
    May 20, 2019 · Magenta, which started nearly three years ago, is an open-source research project powered by TensorFlow that explores the role of machine learning as a tool in ...
  129. [129]
    OpenAI Jukebox Sample Explorer
    Jukebox Sample Explorer ; 5b_lyrics, Unseen lyrics, Heavy Metal ; 5b_lyrics, Novel artists and styles, Country Pop ; 5b, No lyrics conditioning, Rock ; 5b_lyrics ...
  130. [130]
    How to create music using Suno AI? - Analytics Vidhya
    Feb 6, 2025 · Launched in late 2023, it is a partnership between Microsoft and Suno AI, offering a unique approach to AI-driven audio technology. Unlike other ...<|separator|>
  131. [131]
    Suno explained: How to use the viral AI song generator for free
    Feb 14, 2025 · In Suno's latest V4 model (released in November 2024), you can create full four-minute songs, even with a free account. The songs and lyrics are ...What is it? · How does it work? · How do you use it?
  132. [132]
    Suno releases 'most powerful' model yet, days after record labels ...
    Sep 24, 2025 · Suno has introduced the latest version of its AI music generator, v5, which it describes as its “most advanced music model yet.”.
  133. [133]
    Suno (platform) - Wikipedia
    On July 1, 2024, a mobile app for Suno was released. On November 19, 2024, Suno upgraded its AI song model program to v4.
  134. [134]
    New AI music generator Udio synthesizes realistic music on demand
    Apr 10, 2024 · A new AI music-synthesis service that can create novel high-fidelity musical audio from written prompts, including user-provided lyrics.
  135. [135]
    AI composers create music for video games | AIVA Project - CORDIS
    Jul 3, 2020 · AIVA is the first 'virtual artist' to be recognised by SACEM(opens in new window) – the French professional association representing the ...Missing: history | Show results with:history
  136. [136]
    Meta Releases Open-Source Generative AI Text-to-Music Composer ...
    Jun 14, 2023 · Meta has released a generative AI model for producing music named MusicGen, capable of turning a written prompt and music sample into an original composition.<|separator|>
  137. [137]
    facebook/musicgen-large - Hugging Face
    MusicGen is a text-to-music model capable of genreating high-quality music samples conditioned on text descriptions or audio prompts.<|separator|>
  138. [138]
    Introducing AudioCraft: A Generative AI Tool For Audio and Music
    Aug 2, 2023 · Our latest AI tool that generates high-quality, realistic audio and music from text. AudioCraft consists of three models: MusicGen, AudioGen and EnCodec.
  139. [139]
    facebookresearch/audiocraft - GitHub
    AudioCraft contains inference and training code for two state-of-the-art AI generative models producing high-quality audio: AudioGen and MusicGen. Installation.
  140. [140]
    AudioCraft - AI at Meta
    AudioGen and MusicGen represent two variants of audio generative models. While AudioGen is focused on text-to-sound generation and learned to produce audio from ...
  141. [141]
    Generative AI In Music Market Size | Industry Report, 2030
    AI's ability to generate high-quality music compositions, harmonies, and even lyrics has attracted significant attention from artists, producers, and businesses ...
  142. [142]
    Top 20 Companies in Generative AI in Music Market 2025
    Global Generative AI in Music Market Size is projected to Grow from USD 558.4 Million in 2024 to USD 7410.4 Million by 2035, at a CAGR of 26.5% during the ...
  143. [143]
  144. [144]
    Suno vs Udio: The Battle for AI Music Supremacy
    Rating 4.8 (1,250) · Free · MultimediaJan 15, 2025 · Suno currently leads with 67% market share and 12+ million active users, while Udio holds a strong 28% with 4.8 million users.Missing: AIVA | Show results with:AIVA
  145. [145]
  146. [146]
    AI Music Generator Suno in Funding Talks at $2 Billion Valuation
    Oct 17, 2025 · The startup is generating more than $100 million in annual recurring revenue and has recently been in discussions with major record labels to ...Missing: AIVA | Show results with:AIVA
  147. [147]
    [PDF] GLOBAL MUSIC REPORT 2025 - IFPI
    Mar 19, 2025 · Global recorded music revenues saw a tenth consecutive year of growth in 2024, up 4.8% to US$29.6 billion. The growth tells a story of an ...
  148. [148]
    Global economic study shows human creators' future at risk ... - CISAC
    Dec 2, 2024 · The market for Gen AI music and audiovisual content will rise from €3 billion now to €64 billion in 2028; Revenues of Gen AI services in music ...Missing: size | Show results with:size
  149. [149]
    How is AI Used in the Music Industry? Trends & Innovations - Miquido
    Apr 9, 2025 · Thanks to predictable streaming revenues and AI-driven royalty forecasts, institutional investors are pouring billions into music rights.<|separator|>
  150. [150]
  151. [151]
    AI music boom may hit artist royalties, warns Fitch Ratings
    Oct 17, 2025 · Fitch Ratings warns that the rise of AI-generated music could diminish demand for artist-created content and impact royalty-backed music ...
  152. [152]
    How GenAI music is hacking the system – and getting paid
    Jul 25, 2025 · The DSP Deezer now reports receiving 20,000 AI-generated tracks per day, and on the consumption side, 70% of streams of that AI-generated music ...
  153. [153]
    Spotify Strengthens AI Protections for Artists, Songwriters, and ...
    Sep 25, 2025 · That kind of harmful AI content degrades the user experience for listeners and often attempts to divert royalties to bad actors. The future of ...
  154. [154]
    Spotify working on AI music tools with major record labels - BBC
    Oct 16, 2025 · However, critics say adding more AI to the platform would result in less streaming revenue for human artists. Also part of the deal are music ...
  155. [155]
    Record Companies Bring Landmark Cases for Responsible AI ...
    Jun 24, 2024 · RIAA today announced the filing of two copyright infringement cases based on the mass infringement of copyrighted sound recordings copied and exploited without ...
  156. [156]
    IFPI Global Report 2025: Music Revenues Climb 5% Amid AI ...
    Mar 19, 2025 · IFPI Global Report 2025: Music Revenues Up 5% to $29.6 Billion Amid AI Copyright Concerns. Streaming revenues surpass $20 billion for the first ...
  157. [157]
    Largest report on AI in music reveals devastating impact
    Aug 19, 2024 · The report's economic findings are startling, showing that by 2028, 23% of music creators' revenues will be at risk due to generative AI, an ...
  158. [158]
    The Impact of Artificial Intelligence on Music Production - NHSJS
    Mar 20, 2025 · This paper examines the role of AI in music production, focusing on its influence on creativity, ethical considerations, and economic implications.
  159. [159]
    [PDF] AI in the Music Industry - Centre national de la musique
    The rapid advance of AI is giving rise to legitimate fears about the future of jobs in the music industry, fuelling caution among employers and internal ...
  160. [160]
    How AI is Changing the Music Industry in 2025 - Yapsody
    Rating 9.5/10 (14) Jul 31, 2025 · Platforms like Splice and BandLab are using AI to help creators find the perfect sound instantly. Even more revolutionary, some tools can ...Ai In The Music Industry... · Ai As A Songwriting Partner... · Browse Categories<|separator|>
  161. [161]
    Are these the AI music industry jobs of the future? - Global News
    Aug 24, 2025 · AI is opening up new job categories in the music industry. Some of them are pretty cool. Others are concerning.
  162. [162]
    These entertainment jobs are most vulnerable to AI, study says
    Jan 30, 2024 · In the music and sound-recording industry, the tasks most likely to be affected by AI are voice generation and cloning, music generation and ...<|control11|><|separator|>
  163. [163]
    Major record companies sue Suno, Udio for 'mass infringement' of ...
    Jun 24, 2024 · The RIAA calls “mass infringement of copyrighted sound recordings copied and exploited without permission by two multi-million-dollar music generation services ...
  164. [164]
    [PDF] Suno-complaint-file-stamped20.pdf - RIAA
    Jun 24, 2024 · Plaintiffs are record companies or recorded music businesses that, together, own or exclusively control copyrights in a great majority of the ...
  165. [165]
    [PDF] Udio-Complaint-6.24.241.pdf - RIAA
    Jun 24, 2024 · Plaintiffs currently commercially exploit, and at all relevant times have commercially exploited, all the sound recordings listed in Exhibit A. ...
  166. [166]
    Suno argues none of the millions of tracks made on its platform ...
    Aug 26, 2025 · The indie artist class action lawsuit – filed by Anthony Justice and 5th Wheel Records in June 2025 – specifically targets the impact of AI ...
  167. [167]
    Suno and Udio slam label lawsuits... but the RIAA hits back
    Aug 2, 2024 · AI music services Suno and Udio are being sued by labels in the US, and now we know more about how they plan to defend themselves.<|separator|>
  168. [168]
    A New Look at Fair Use: Anthropic, Meta, and Copyright in AI Training
    Jul 3, 2025 · Bartz v. Anthropic and Kadrey v. Meta offer the first significant judicial guidance on how courts will look to apply the fair use doctrine to AI training ...
  169. [169]
    Fair Use and AI Training: Two Recent Decisions Highlight the ...
    Jul 8, 2025 · In each case, the court found that, on the facts before it, the use of copyrighted works to train an AI model was highly transformative and fair ...
  170. [170]
    Using copyrighted content to train AI without permission is not 'fair ...
    Feb 12, 2025 · A US court has ruled on whether using copyrighted material without permission to train AI amounts to “fair use” under copyright law.
  171. [171]
    A Tale of Three Cases: How Fair Use Is Playing Out in AI Copyright ...
    Jul 7, 2025 · In this article, we review the court's reasoning in each of these three cases to elucidate how the battle for a fair use finding is playing out for AI ...
  172. [172]
    Rightsify | AI music training data
    Rightsify provides licensed music data for AI model builders, synthetic data & intelligent licensing for developers, AI labs & brands worldwide.Gramosynth · BGM · GCX · Contact Us
  173. [173]
    GEMA proposes licensing model for AI-generated music
    Sep 30, 2024 · Through its new licensing model, GEMA proposes that authors should receive compensation beyond a one-time payment for training data. It ...
  174. [174]
  175. [175]
    Major music publishers ink AI licensing deals with lyrics and music ...
    Oct 15, 2025 · Lyrics and music data company Musixmatch says it has signed AI licensing deals with the publishing arms of all three major music companies.
  176. [176]
    Record labels could strike 'landmark' AI music licensing deals 'within ...
    Oct 2, 2025 · Two of the three major music companies, Universal Music Group and Warner Music Group, are reportedly nearing licensing deals with AI music ...
  177. [177]
    Copyright Office Weighs In on AI Training and Fair Use - Skadden Arps
    May 15, 2025 · While a number of copyright holders have licensed their data for such usage, most of the content is used by AI models without authorization.
  178. [178]
    Copyright and Artificial Intelligence | U.S. Copyright Office
    Copyright and Artificial Intelligence analyzes copyright law and policy issues raised by artificial intelligence.Spring 2023 AI Listening · Studies · Registration Guidance for...Missing: generators | Show results with:generators
  179. [179]
    Copyright Office Releases Part 2 of Artificial Intelligence Report
    Jan 29, 2025 · It concludes that the outputs of generative AI can be protected by copyright only where a human author has determined sufficient expressive ...
  180. [180]
    Copyright Denied for Generative AI: Remarks on Thaler v Perlmutter
    The Court held that a work must be a result of human authorship to receive copyright protection and therefore upheld the decision of the United States Copyright ...
  181. [181]
    [PDF] Works Containing Material Generated by Artificial Intelligence
    Mar 16, 2023 · Based on these developments, the Office concludes that public guidance is needed on the registration of works containing AI-generated content.
  182. [182]
    US Court Bans Copyright for AI Art – What Musicians & Creators ...
    Mar 26, 2025 · On March 21, 2025, the US Court of Appeals issued a landmark ruling that denies copyright protection for AI-generated works.
  183. [183]
    US Copyright Office on AI: Human creativity still matters, legally - WIPO
    Apr 24, 2025 · You can't copyright fully AI-generated work but you may be able to protect your sufficiently original selection and arrangement or modification ...
  184. [184]
    [PDF] Human-AI music creation: Understanding the perceptions ... - ISMIR
    The Human-AI Collaboration Model demonstrates our belief that the role that AI plays on creative thinking is highly flexible within music creation and that ...
  185. [185]
    AI-Generated Music Artists and Hybrid Collaborations - John Rector
    Sep 26, 2025 · AI-Driven Music Creation: Advances in AI now enable generation of song elements – from lyrics and melodies to vocals and fully produced ...
  186. [186]
    AI-Assisted Music Production: A User Study on Text-to-Music Models
    Sep 27, 2025 · This study explores the interaction of experienced music producers with an interface combining a Text-to-Music model and an AI-based music ...<|separator|>
  187. [187]
    Exploring the Collaborative Co-Creation Process with AI: A Case ...
    To bridge this gap, we conducted a case study in a college-level course where nine undergraduate students were tasked with creating three original music tracks ...
  188. [188]
    (PDF) Collaborative AI in Music Composition: Human-AI Symbiosis ...
    May 7, 2025 · This research evaluates how partnership between humans and AI components transforms musical education along with the process of composition for ...
  189. [189]
    Identifying bias in generative music models | NAACL - MBZUAI
    Apr 28, 2025 · A team from MBZUAI tested the performance of models in non-Western music, with the aim of making them more inclusive of diverse genres.
  190. [190]
    Bias beyond Borders: Global Inequalities in AI-Generated Music
    Oct 2, 2025 · Analyzing GlobalDISCO, we find that state-of-the-art music generation models are highly biased across both world regions and genres, and that ...
  191. [191]
    MBZUAI study reveals bias in AI music models
    May 2, 2025 · The results suggest modifying models is not enough, addressing dataset bias is essential to building inclusive music generation systems. The ...
  192. [192]
    [2412.04100] Missing Melodies: AI Music Generation and its "Nearly ...
    Dec 5, 2024 · However, around 40% of these datasets include some form of non-Western music, genres from the Global South account for only 14.6% of the data.
  193. [193]
    Music RAI: Responsible AI International Community to Reduce Bias ...
    Music RAI aims to address bias in AI music generation and analysis, which often favors Western music, by developing tools for marginalized genres.
  194. [194]
    Reducing Barriers to the Use of Marginalised Music Genres in AI
    Jul 18, 2024 · This paper to explore the eXplainable AI (XAI) challenges and opportunities associated with reducing barriers to using marginalised genres of music with AI ...
  195. [195]
    [PDF] Music for All: Representational Bias and Cross-Cultural Adaptability ...
    Apr 29, 2025 · Music models show bias towards Western music, with only 5.7% of datasets from non-Western genres, leading to under-representation of non- ...
  196. [196]
    [PDF] Leveraging small datasets for ethical and responsible AI music making
    May 14, 2025 · 6% of available music datasets used for AI model training repre- sent non-Western music [31]. Whilst there have been initiatives to build ...<|control11|><|separator|>
  197. [197]
    Billie Eilish, Nicki Minaj, Stevie Wonder and more musicians ...
    Apr 2, 2024 · A group of more than 200 high-profile musicians have signed an open letter calling for protections against the predatory use of artificial intelligence.
  198. [198]
    Record labels claim AI generator Suno ripped their ... - The Verge
    Sep 22, 2025 · Major record labels have escalated their lawsuit against Suno, alleging that the AI startup knowingly pirated songs from YouTube to train ...
  199. [199]
    MUSIC INDUSTRY UNITES TO PROTECT THE RIGHTS OF ...
    Jun 24, 2024 · The music industry is uniting to protect musicians by introducing "Principles for Music Creation with AI" to ensure responsible AI use and ...
  200. [200]
    AI-Generated Fake 'Drake'/'Weeknd' Collaboration Alarms Industry
    Apr 17, 2023 · UPDATED: Are you ready for deep-fake-Drake? Many fans are, at least as an amusing lark, as an apparently artificial-intelligence-generated “ ...
  201. [201]
    The Ballad of 'Deepfake Drake' - The New York Times
    Apr 28, 2023 · The Ballad of 'Deepfake Drake'. A song generated by A.I. could have far-reaching consequences for the future of the music industry.
  202. [202]
    Drake and The Weeknd AI song pulled from Spotify and Apple - BBC
    Apr 18, 2023 · Heart On My Sleeve is no longer available on Apple Music, Spotify ... 'Disgraceful' deep-fake AI video condemned by presidential candidate.
  203. [203]
    Biggest AI Music Stories 2024: Drake, TikTok, Lawsuits & More
    Dec 10, 2024 · The biggest AI music stories of 2024 involve a Drake diss track, the TikTok-UMG fight, several lawsuits, Tenneessee's ELVIS Act on deepfakes ...
  204. [204]
    UK music body threatens first-of-its-kind AI deepfake lawsuit
    Mar 19, 2024 · UK music body threatens first-of-its-kind AI deepfake lawsuit.<|control11|><|separator|>
  205. [205]
    AI Could Endanger the Music Industry. Here's How Congress Can ...
    May 27, 2025 · “The NO FAKES Act would give each of us the ability to say when and how AI deepfakes of our voices and likenesses can be used,” McBride told ...<|separator|>
  206. [206]
    Exploring the Impact of Artificial Intelligence on the Creativity ... - NIH
    Apr 15, 2025 · This study investigates the relationship between artificial intelligence (AI) tools and the creative abilities of music practitioners.
  207. [207]
    Generative AI enhances individual creativity but reduces ... - Science
    Jul 12, 2024 · We find that access to generative AI ideas causes stories to be evaluated as more creative, better written, and more enjoyable, especially among less creative ...<|control11|><|separator|>
  208. [208]
    Human, I wrote a song for you: An experiment testing the influence ...
    This study examines the evaluation of musical performances of artificial intelligence (AI) and the acceptance of AI music generators as musicians.
  209. [209]
    Why AI can't completely replace human musicians - CNBC
    Dec 29, 2023 · Although AI probably won't completely displace human musicians, that doesn't mean the technology won't have an impact on the music industry.
  210. [210]
    Hundreds of musicians warn against replacing human artists with AI
    Apr 2, 2024 · Top musicians among hundreds warning against replacing human artists with AI ... (AI) to infringe upon and devalue the rights of human artists.
  211. [211]
    AI music isn't going away. Here are 4 big questions about what's next
    Apr 25, 2024 · A group of visual artists sued companies behind AI image generators. And in the music realm, three publishing companies, Universal, Concord and ...