Search Results for author: George Fazekas

Found 20 papers, 8 papers with code

Music2Latent2: Audio Compression with Summary Embeddings and Autoregressive Decoding

no code implementations29 Jan 2025 Marco Pasini, Stefan Lattner, George Fazekas

Efficiently compressing high-dimensional audio signals into a compact and informative latent space is crucial for various tasks, including generative modeling and music information retrieval (MIR).

Audio Compression Denoising +3

Continuous Autoregressive Models with Noise Augmentation Avoid Error Accumulation

no code implementations27 Nov 2024 Marco Pasini, Javier Nistal, Stefan Lattner, George Fazekas

Autoregressive models are typically applied to sequences of discrete tokens, but recent research indicates that generating sequences of continuous embeddings in an autoregressive manner is also feasible.

Audio Generation

Low-Data Classification of Historical Music Manuscripts: A Few-Shot Learning Approach

no code implementations25 Nov 2024 Elona Shatri, Daniel Raymond, George Fazekas

In this paper, we explore the intersection of technology and cultural preservation by developing a self-supervised learning framework for the classification of musical symbols in historical manuscripts.

Classification Few-Shot Learning +1

Synthesising Handwritten Music with GANs: A Comprehensive Evaluation of CycleWGAN, ProGAN, and DCGAN

no code implementations25 Nov 2024 Elona Shatri, Kalikidhar Palavala, George Fazekas

The generation of handwritten music sheets is a crucial step toward enhancing Optical Music Recognition (OMR) systems, which rely on large and diverse datasets for optimal performance.

Style Transfer

Diff-MSTC: A Mixing Style Transfer Prototype for Cubase

no code implementations10 Nov 2024 Soumya Sai Vanka, Lennart Hannink, Jean-Baptiste Rolland, George Fazekas

Diff-MST, a deep learning model for mixing style transfer, forecasts mixing console parameters for tracks using a reference song.

Deep Learning Style Transfer

Knowledge Discovery in Optical Music Recognition: Enhancing Information Retrieval with Instance Segmentation

no code implementations27 Aug 2024 Elona Shatri, George Fazekas

This study emphasises the role of pixel-wise segmentation in advancing accurate music symbol recognition, contributing to knowledge discovery in OMR.

Information Retrieval Instance Segmentation +7

Music2Latent: Consistency Autoencoders for Latent Audio Compression

1 code implementation12 Aug 2024 Marco Pasini, Stefan Lattner, George Fazekas

We demonstrate that Music2Latent outperforms existing continuous audio autoencoders in sound quality and reconstruction accuracy while achieving competitive performance on downstream MIR tasks using its latent representations.

Audio Compression Information Retrieval +1

MuChoMusic: Evaluating Music Understanding in Multimodal Audio-Language Models

2 code implementations2 Aug 2024 Benno Weck, Ilaria Manco, Emmanouil Benetos, Elio Quinton, George Fazekas, Dmitry Bogdanov

Motivated by this, we introduce MuChoMusic, a benchmark for evaluating music understanding in multimodal language models focused on audio.

Multimodal Reasoning Multiple-choice +1

Diff-MST: Differentiable Mixing Style Transfer

no code implementations11 Jul 2024 Soumya Sai Vanka, Christian Steinmetz, Jean-Baptiste Rolland, Joshua Reiss, George Fazekas

Mixing style transfer automates the generation of a multitrack mix for a given set of tracks by inferring production attributes from a reference song.

Style Transfer

JAZZVAR: A Dataset of Variations found within Solo Piano Performances of Jazz Standards for Music Overpainting

1 code implementation18 Jul 2023 Eleanor Row, Jingjing Tang, George Fazekas

In this paper, we outline the curation process for obtaining and sorting the repertoire, the pipeline for creating the Original and Variation pairs, and our analysis of the dataset.

Information Retrieval Music Information Retrieval +1

Adoption of AI Technology in the Music Mixing Workflow: An Investigation

no code implementations6 Apr 2023 Soumya Sai Vanka, Maryam Safi, Jean-Baptiste Rolland, George Fazekas

The integration of artificial intelligence (AI) technology in the music industry is driving a significant change in the way music is being composed, produced and mixed.

An Comparative Analysis of Different Pitch and Metrical Grid Encoding Methods in the Task of Sequential Music Generation

no code implementations31 Jan 2023 Yuqiang Li, Shengchen Li, George Fazekas

Results display a general phenomenon of over-fitting from two aspects, the pitch embedding space and the test loss of the single-token grid encoding.

Feature Engineering Music Generation +1

Differentiable Time-Frequency Scattering on GPU

3 code implementations18 Apr 2022 John Muradeli, Cyrus Vahidi, Changhong Wang, Han Han, Vincent Lostanlen, Mathieu Lagrange, George Fazekas

Joint time-frequency scattering (JTFS) is a convolutional operator in the time-frequency domain which extracts spectrotemporal modulations at various rates and scales.

Audio Generation Resynthesis

Explaining Deep Convolutional Neural Networks on Music Classification

1 code implementation8 Jul 2016 Keunwoo Choi, George Fazekas, Mark Sandler

Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.

Chord Recognition Classification +6

Towards Playlist Generation Algorithms Using RNNs Trained on Within-Track Transitions

no code implementations7 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).

Automatic tagging using deep convolutional neural networks

11 code implementations1 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).

Music Tagging

Text-based LSTM networks for Automatic Music Composition

4 code implementations18 Apr 2016 Keunwoo Choi, George Fazekas, Mark Sandler

In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.

Diversity

Cannot find the paper you are looking for? You can Submit a new open access paper.