no code implementations • 18 Jun 2022 • Zhepei Wang, Ritwik Giri, Shrikant Venkataramani, Umut Isik, Jean-Marc Valin, Paris Smaragdis, Mike Goodwin, Arvindh Krishnaswamy
In this work, we propose Exformer, a time-domain architecture for target speaker extraction.
no code implementations • 16 Jun 2022 • Jean-Marc Valin, Ritwik Giri, Shrikant Venkataramani, Umut Isik, Arvindh Krishnaswamy
In real life, room effect, also known as room reverberation, and the present background noise degrade the quality of speech.
no code implementations • 8 Jun 2021 • Ritwik Giri, Shrikant Venkataramani, Jean-Marc Valin, Umut Isik, Arvindh Krishnaswamy
The presence of multiple talkers in the surrounding environment poses a difficult challenge for real-time speech communication systems considering the constraints on network size and complexity.
1 code implementation • 18 Jun 2020 • Yu-Che Wang, Shrikant Venkataramani, Paris Smaragdis
Supervised learning for single-channel speech enhancement requires carefully labeled training examples where the noisy mixture is input into the network and the network is trained to produce an output close to the ideal target.
Audio and Speech Processing Sound
no code implementations • 20 Feb 2020 • Jonah Casebeer, Umut Isik, Shrikant Venkataramani, Arvindh Krishnaswamy
Many neural speech enhancement and source separation systems operate in the time-frequency domain.
2 code implementations • 22 Oct 2019 • Efthymios Tzinis, Shrikant Venkataramani, Zhepei Wang, Cem Subakan, Paris Smaragdis
In the first step we learn a transform (and it's inverse) to a latent space where masking-based separation performance using oracles is optimal.
Ranked #32 on
Speech Separation
on WSJ0-2mix
no code implementations • 7 Nov 2018 • Prem Seetharaman, Gordon Wichern, Shrikant Venkataramani, Jonathan Le Roux
Isolating individual instruments in a musical mixture has a myriad of potential applications, and seems imminently achievable given the levels of performance reached by recent deep learning methods.
1 code implementation • 5 Nov 2018 • Efthymios Tzinis, Shrikant Venkataramani, Paris Smaragdis
We present a monophonic source separation system that is trained by only observing mixtures with no ground truth separation information.
no code implementations • 5 Oct 2018 • Shrikant Venkataramani, Paris Smaragdis
The performance of single channel source separation algorithms has improved greatly in recent times with the development and deployment of neural networks.
1 code implementation • 6 May 2017 • Shrikant Venkataramani, Jonah Casebeer, Paris Smaragdis
We present an auto-encoder neural network that can act as an equivalent to short-time front-end transforms.
Sound