Music Generation
131 papers with code • 0 benchmarks • 24 datasets
Music Generation is the task of generating music or music-like sounds from a model or algorithm. The goal is to produce a sequence of notes or sound events that are similar to existing music in some way, such as having the same style, genre, or mood.
Benchmarks
These leaderboards are used to track progress in Music Generation
Libraries
Use these libraries to find Music Generation models and implementationsDatasets
Latest papers
Exploring XAI for the Arts: Explaining Latent Space in Generative Music
We increase the explainability of the model by: i) using latent space regularisation to force some specific dimensions of the latent space to map to meaningful musical attributes, ii) providing a user interface feedback loop to allow people to adjust dimensions of the latent space and observe the results of these changes in real-time, iii) providing a visualisation of the musical attributes in the latent space to help people understand and predict the effect of changes to latent space dimensions.
JEN-1: Text-Guided Universal Music Generation with Omnidirectional Diffusion Models
Despite the task's significance, prevailing generative models exhibit limitations in music quality, computational efficiency, and generalization.
MusicLDM: Enhancing Novelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies
Diffusion models have shown promising results in cross-modal generation tasks, including text-to-image and text-to-audio generation.
Graph-based Polyphonic Multitrack Music Generation
Graphs can be leveraged to model polyphonic multitrack symbolic music, where notes, chords and entire sections may be linked at different levels of the musical hierarchy by tonal and rhythmic relationships.
Polyffusion: A Diffusion Model for Polyphonic Score Generation with Internal and External Controls
We propose Polyffusion, a diffusion model that generates polyphonic music scores by regarding music as image-like piano roll representations.
VampNet: Music Generation via Masked Acoustic Token Modeling
We introduce VampNet, a masked acoustic token modeling approach to music synthesis, compression, inpainting, and variation.
EmoGen: Eliminating Subjective Bias in Emotional Music Generation
In this paper, we propose EmoGen, an emotional music generation system that leverages a set of emotion-related music attributes as the bridge between emotion and music, and divides the generation into two stages: emotion-to-attribute mapping with supervised clustering, and attribute-to-music generation with self-supervised learning.
Simple and Controllable Music Generation
We tackle the task of conditional music generation.
MuseCoco: Generating Symbolic Music from Text
In contrast, symbolic music offers ease of editing, making it more accessible for users to manipulate specific musical elements.
GETMusic: Generating Any Music Tracks with a Unified Representation and Diffusion Framework
Our proposed representation, coupled with the non-autoregressive generative model, empowers GETMusic to generate music with any arbitrary source-target track combinations.