no code implementations • 29 Jan 2025 • Marco Pasini, Stefan Lattner, George Fazekas
Efficiently compressing high-dimensional audio signals into a compact and informative latent space is crucial for various tasks, including generative modeling and music information retrieval (MIR).
no code implementations • 27 Nov 2024 • Marco Pasini, Javier Nistal, Stefan Lattner, George Fazekas
Autoregressive models are typically applied to sequences of discrete tokens, but recent research indicates that generating sequences of continuous embeddings in an autoregressive manner is also feasible.
no code implementations • 25 Nov 2024 • Elona Shatri, Daniel Raymond, George Fazekas
In this paper, we explore the intersection of technology and cultural preservation by developing a self-supervised learning framework for the classification of musical symbols in historical manuscripts.
no code implementations • 25 Nov 2024 • Elona Shatri, Kalikidhar Palavala, George Fazekas
The generation of handwritten music sheets is a crucial step toward enhancing Optical Music Recognition (OMR) systems, which rely on large and diverse datasets for optimal performance.
no code implementations • 10 Nov 2024 • Soumya Sai Vanka, Lennart Hannink, Jean-Baptiste Rolland, George Fazekas
Diff-MST, a deep learning model for mixing style transfer, forecasts mixing console parameters for tracks using a reference song.
no code implementations • 27 Aug 2024 • Elona Shatri, George Fazekas
This study emphasises the role of pixel-wise segmentation in advancing accurate music symbol recognition, contributing to knowledge discovery in OMR.
1 code implementation • 12 Aug 2024 • Marco Pasini, Stefan Lattner, George Fazekas
We demonstrate that Music2Latent outperforms existing continuous audio autoencoders in sound quality and reconstruction accuracy while achieving competitive performance on downstream MIR tasks using its latent representations.
2 code implementations • 2 Aug 2024 • Benno Weck, Ilaria Manco, Emmanouil Benetos, Elio Quinton, George Fazekas, Dmitry Bogdanov
Motivated by this, we introduce MuChoMusic, a benchmark for evaluating music understanding in multimodal language models focused on audio.
no code implementations • 11 Jul 2024 • Soumya Sai Vanka, Christian Steinmetz, Jean-Baptiste Rolland, Joshua Reiss, George Fazekas
Mixing style transfer automates the generation of a multitrack mix for a given set of tracks by inferring production attributes from a reference song.
1 code implementation • 18 Jul 2023 • Eleanor Row, Jingjing Tang, George Fazekas
In this paper, we outline the curation process for obtaining and sorting the repertoire, the pipeline for creating the Original and Variation pairs, and our analysis of the dataset.
no code implementations • 6 Apr 2023 • Soumya Sai Vanka, Maryam Safi, Jean-Baptiste Rolland, George Fazekas
The integration of artificial intelligence (AI) technology in the music industry is driving a significant change in the way music is being composed, produced and mixed.
no code implementations • 31 Jan 2023 • Yuqiang Li, Shengchen Li, George Fazekas
Results display a general phenomenon of over-fitting from two aspects, the pitch embedding space and the test loss of the single-token grid encoding.
3 code implementations • 18 Apr 2022 • John Muradeli, Cyrus Vahidi, Changhong Wang, Han Han, Vincent Lostanlen, Mathieu Lagrange, George Fazekas
Joint time-frequency scattering (JTFS) is a convolutional operator in the time-frequency domain which extracts spectrotemporal modulations at various rates and scales.
no code implementations • 7 Jun 2017 • Keunwoo Choi, George Fazekas, Kyunghyun Cho, Mark Sandler
The results highlight several important aspects of music tagging and neural networks.
13 code implementations • 14 Sep 2016 • Keunwoo Choi, George Fazekas, Mark Sandler, Kyunghyun Cho
We introduce a convolutional recurrent neural network (CRNN) for music tagging.
no code implementations • 17 Aug 2016 • Keunwoo Choi, George Fazekas, Brian McFee, Kyunghyun Cho, Mark Sandler
Descriptions are often provided along with recommendations to help users' discovery.
1 code implementation • 8 Jul 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.
no code implementations • 7 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).
11 code implementations • 1 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).
4 code implementations • 18 Apr 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.