Search Results for author: Philippe Esling

Found 26 papers, 14 papers with code

Continuous descriptor-based control for deep audio synthesis

1 code implementation27 Feb 2023 Ninon Devis, Nils Demerlé, Sarah Nabi, David Genova, Philippe Esling

Despite significant advances in deep models for music generation, the use of these techniques remains restricted to expert users.

Audio Synthesis Continuous Control +1

SingSong: Generating musical accompaniments from singing

no code implementations30 Jan 2023 Chris Donahue, Antoine Caillon, Adam Roberts, Ethan Manilow, Philippe Esling, Andrea Agostinelli, Mauro Verzetti, Ian Simon, Olivier Pietquin, Neil Zeghidour, Jesse Engel

We present SingSong, a system that generates instrumental music to accompany input vocals, potentially offering musicians and non-musicians alike an intuitive new way to create music featuring their own voice.

Audio Generation Retrieval

Challenges in creative generative models for music: a divergence maximization perspective

no code implementations16 Nov 2022 Axel Chemla--Romeu-Santos, Philippe Esling

The development of generative Machine Learning (ML) models in creative practices, enabled by the recent improvements in usability and availability of pre-trained models, is raising more and more interest among artists, practitioners and performers.

Creative divergent synthesis with generative models

1 code implementation16 Nov 2022 Axel Chemla--Romeu-Santos, Philippe Esling

Machine learning approaches now achieve impressive generation capabilities in numerous domains such as image, audio or video.

Streamable Neural Audio Synthesis With Non-Causal Convolutions

no code implementations14 Apr 2022 Antoine Caillon, Philippe Esling

As our method is based on a post-training reconfiguration of the model, we show that it is able to transform models trained without causal constraints into a streaming model.

Audio Generation Audio Synthesis

RAVE: A variational autoencoder for fast and high-quality neural audio synthesis

1 code implementation9 Nov 2021 Antoine Caillon, Philippe Esling

By leveraging a multi-band decomposition of the raw waveform, we show that our model is the first able to generate 48kHz audio signals, while simultaneously running 20 times faster than real-time on a standard laptop CPU.

Audio Synthesis Representation Learning

Signal-domain representation of symbolic music for learning embedding spaces

1 code implementation8 Sep 2021 Mathieu Prang, Philippe Esling

We evaluate the ability to learn meaningful features from this representation from a musical point of view.

Energy Consumption of Deep Generative Audio Models

no code implementations6 Jul 2021 Constance Douwes, Philippe Esling, Jean-Pierre Briot

In most scientific domains, the deep learning community has largely focused on the quality of deep generative models, resulting in highly accurate and successful solutions.

Spectrogram Inpainting for Interactive Generation of Instrument Sounds

1 code implementation15 Apr 2021 Théis Bazin, Gaëtan Hadjeres, Philippe Esling, Mikhail Malt

Modern approaches to sound synthesis using deep neural networks are hard to control, especially when fine-grained conditioning information is not available, hindering their adoption by musicians.

Image Generation

Creativity in the era of artificial intelligence

no code implementations13 Aug 2020 Philippe Esling, Ninon Devis

As creativity is a highly context-prone concept, we underline the limits and deficiencies of current AI, requiring to move towards artificial creativity.

Timbre latent space: exploration and creative aspects

no code implementations4 Aug 2020 Antoine Caillon, Adrien Bitton, Brice Gatinet, Philippe Esling

Recent studies show the ability of unsupervised models to learn invertible audio representations using Auto-Encoders.

Neural Granular Sound Synthesis

no code implementations4 Aug 2020 Adrien Bitton, Philippe Esling, Tatsuya Harada

In this setting the learned grain space is invertible, meaning that we can continuously synthesize sound when traversing its dimensions.

Audio Generation

Ultra-light deep MIR by trimming lottery tickets

1 code implementation31 Jul 2020 Philippe Esling, Theis Bazin, Adrien Bitton, Tristan Carsault, Ninon Devis

We show that our proposal can remove up to 90% of the model parameters without loss of accuracy, leading to ultra-light deep MIR models.

Audio Classification Drum Transcription +3

Diet deep generative audio models with structured lottery

1 code implementation31 Jul 2020 Philippe Esling, Ninon Devis, Adrien Bitton, Antoine Caillon, Axel Chemla--Romeu-Santos, Constance Douwes

This hypothesis states that extremely efficient small sub-networks exist in deep models and would provide higher accuracy than larger models if trained in isolation.

Vector-Quantized Timbre Representation

1 code implementation13 Jul 2020 Adrien Bitton, Philippe Esling, Tatsuya Harada

Although its definition is usually elusive, it can be seen from a signal processing viewpoint as all the spectral features that are perceived independently from pitch and loudness.

Cross-modal variational inference for bijective signal-symbol translation

no code implementations10 Feb 2020 Axel Chemla--Romeu-Santos, Stavros Ntalampiras, Philippe Esling, Goffredo Haus, Gérard Assayag

Extraction of symbolic information from signals is an active field of research enabling numerous applications especially in the Musical Information Retrieval domain.

Audio Generation Density Estimation +6

Multi-Step Chord Sequence Prediction Based on Aggregated Multi-Scale Encoder-Decoder Network

1 code implementation12 Nov 2019 Tristan Carsault, Andrew McLeod, Philippe Esling, Jérôme Nika, Eita Nakamura, Kazuyoshi Yoshii

In this paper, we postulate that this comes from the multi-scale structure of musical information and propose new architectures based on an iterative temporal aggregation of input labels.

Using musical relationships between chord labels in automatic chord extraction tasks

1 code implementation12 Nov 2019 Tristan Carsault, Jérôme Nika, Philippe Esling

Recent researches on Automatic Chord Extraction (ACE) have focused on the improvement of models based on machine learning.

Universal audio synthesizer control with normalizing flows

1 code implementation Digital Audio Effects (DaFX) 2019 2019 Philippe Esling, Naotake Masuda, Adrien Bardet, Romeo Despres, Axel Chemla--Romeu-Santos

By using this formulation, we show that we can address simultaneously automatic parameter inference, macro-control learning and audio-based preset exploration within a single model.

A database linking piano and orchestral MIDI scores with application to automatic projective orchestration

no code implementations19 Oct 2018 Léopold Crestel, Philippe Esling, Lena Heng, Stephen McAdams

This article introduces the Projective Orchestral Database (POD), a collection of MIDI scores composed of pairs linking piano scores to their corresponding orchestrations.

Modulated Variational auto-Encoders for many-to-many musical timbre transfer

no code implementations ICLR 2019 Adrien Bitton, Philippe Esling, Axel Chemla--Romeu-Santos

We define timbre transfer as applying parts of the auditory properties of a musical instrument onto another.

Sound Audio and Speech Processing

Generative timbre spaces: regularizing variational auto-encoders with perceptual metrics

1 code implementation Conference 2018 Philippe Esling, Axel Chemla--Romeu-Santos, Adrien Bitton

Based on this, we introduce a method for descriptor-based synthesis and show that we can control the descriptors of an instrument while keeping its timbre structure.

Sound Audio and Speech Processing

Live Orchestral Piano, a system for real-time orchestral music generation

no code implementations5 Sep 2016 Léopold Crestel, Philippe Esling

This paper introduces the first system for performing automatic orchestration based on a real-time piano input.

Music Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.