Music Generation
Music generation research aims to create systems capable of producing high-quality, diverse, and controllable music from various inputs like text, images, or other audio. Current efforts focus on refining diffusion models and transformers, often incorporating techniques like distillation and multi-modal conditioning to improve efficiency, controllability, and the realism of generated music, including vocals and accompaniment. This field is significant for its potential to revolutionize music creation, composition assistance, and music-related applications in entertainment, education, and therapy, while also raising important questions about copyright and ethical considerations.
Papers
PAGURI: a user experience study of creative interaction with text-to-music models
Francesca Ronchini, Luca Comanducci, Gabriele Perego, Fabio Antonacci
MuseBarControl: Enhancing Fine-Grained Control in Symbolic Music Generation through Pre-Training and Counterfactual Loss
Yangyang Shu, Haiming Xu, Ziqin Zhou, Anton van den Hengel, Lingqiao Liu