no code implementations • 24 Mar 2024 • Hanzhi Yin, Gang Cheng, Christian J. Steinmetz, Ruibin Yuan, Richard M. Stern, Roger B. Dannenberg
We describe a novel approach for developing realistic digital models of dynamic range compressors for digital audio production by analyzing their analog prototypes.
no code implementations • 2 Nov 2023 • Shubhr Singh, Christian J. Steinmetz, Emmanouil Benetos, Huy Phan, Dan Stowell
Deep learning models such as CNNs and Transformers have achieved impressive performance for end-to-end audio tagging.
1 code implementation • 22 May 2023 • Christopher Mitcheltree, Christian J. Steinmetz, Marco Comunità, Joshua D. Reiss
Low frequency oscillator (LFO) driven audio effects such as phaser, flanger, and chorus, modify an input signal using time-varying filters and delays, resulting in characteristic sweeping or widening effects.
no code implementations • 10 Apr 2023 • Scott H. Hawley, Christian J. Steinmetz
We investigate applying audio manipulations using pretrained neural network-based autoencoders as an alternative to traditional signal processing methods, since the former may provide greater semantic or perceptual organization.
no code implementations • 1 Nov 2022 • Marco Comunità, Christian J. Steinmetz, Huy Phan, Joshua D. Reiss
Deep learning approaches for black-box modelling of audio effects have shown promise, however, the majority of existing work focuses on nonlinear effects with behaviour on relatively short time-scales, such as guitar amplifiers and distortion.
3 code implementations • 6 Mar 2022 • Joseph Turian, Jordie Shier, Humair Raj Khan, Bhiksha Raj, Björn W. Schuller, Christian J. Steinmetz, Colin Malloy, George Tzanetakis, Gissel Velarde, Kirk McNally, Max Henry, Nicolas Pinto, Camille Noufi, Christian Clough, Dorien Herremans, Eduardo Fonseca, Jesse Engel, Justin Salamon, Philippe Esling, Pranay Manocha, Shinji Watanabe, Zeyu Jin, Yonatan Bisk
The aim of the HEAR benchmark is to develop a general-purpose audio representation that provides a strong basis for learning in a wide variety of tasks and scenarios.
1 code implementation • 6 Dec 2021 • Christian J. Steinmetz, Joshua D. Reiss
Applications of deep learning for audio effects often focus on modeling analog effects or learning to control effects to emulate a trained audio engineer.
1 code implementation • 7 Oct 2021 • Joseph T. Colonel, Christian J. Steinmetz, Marcus Michelen, Joshua D. Reiss
In this work, we address some of these limitations by learning a direct mapping from the target magnitude response to the filter coefficient space with a neural network trained on millions of random filters.
1 code implementation • 4 Oct 2021 • Christian J. Steinmetz, Joshua D. Reiss
In this work, we propose WaveBeat, an end-to-end approach for joint beat and downbeat tracking operating directly on waveforms.
1 code implementation • 15 Jul 2021 • Christian J. Steinmetz, Vamsi Krishna Ithapu, Paul Calamia
Deep learning approaches have emerged that aim to transform an audio signal so that it sounds as if it was recorded in the same room as a reference recording, with applications both in audio post-production and augmented reality.
1 code implementation • 11 Feb 2021 • Christian J. Steinmetz, Joshua D. Reiss
Deep learning approaches have demonstrated success in modeling analog audio effects.
1 code implementation • 20 Oct 2020 • Christian J. Steinmetz, Jordi Pons, Santiago Pascual, Joan Serrà
Applications of deep learning to automatic multitrack mixing are largely unexplored.
Audio and Speech Processing Sound
2 code implementations • 8 Oct 2020 • Christian J. Steinmetz, Joshua D. Reiss
By processing audio signals in the time-domain with randomly weighted temporal convolutional networks (TCNs), we uncover a wide range of novel, yet controllable overdrive effects.