Search Results for author: Zhong-Qiu Wang

Found 9 papers, 4 papers with code

ESPnet-SE++: Speech Enhancement for Robust Speech Recognition, Translation, and Understanding

1 code implementation19 Jul 2022 Yen-Ju Lu, Xuankai Chang, Chenda Li, Wangyou Zhang, Samuele Cornell, Zhaoheng Ni, Yoshiki Masuyama, Brian Yan, Robin Scheibler, Zhong-Qiu Wang, Yu Tsao, Yanmin Qian, Shinji Watanabe

To showcase such integration, we performed experiments on carefully designed synthetic datasets for noisy-reverberant multi-channel ST and SLU tasks, which can be used as benchmark corpora for future research.

Automatic Speech Recognition Robust Speech Recognition +5

Locate This, Not That: Class-Conditioned Sound Event DOA Estimation

no code implementations8 Mar 2022 Olga Slizovskaia, Gordon Wichern, Zhong-Qiu Wang, Jonathan Le Roux

Existing systems for sound event localization and detection (SELD) typically operate by estimating a source location for all classes at every time instant.

Sound Event Localization and Detection

Conditional Diffusion Probabilistic Model for Speech Enhancement

1 code implementation10 Feb 2022 Yen-Ju Lu, Zhong-Qiu Wang, Shinji Watanabe, Alexander Richard, Cheng Yu, Yu Tsao

Speech enhancement is a critical component of many user-oriented audio applications, yet current systems still suffer from distorted and unnatural outputs.

Speech Enhancement Speech Synthesis

The Cocktail Fork Problem: Three-Stem Audio Separation for Real-World Soundtracks

2 code implementations19 Oct 2021 Darius Petermann, Gordon Wichern, Zhong-Qiu Wang, Jonathan Le Roux

The cocktail party problem aims at isolating any source of interest within a complex acoustic scene, and has long inspired audio source separation research.

Audio Source Separation

Multi-microphone Complex Spectral Mapping for Utterance-wise and Continuous Speech Separation

2 code implementations4 Oct 2020 Zhong-Qiu Wang, Peidong Wang, DeLiang Wang

Although our system is trained on simulated room impulse responses (RIR) based on a fixed number of microphones arranged in a given geometry, it generalizes well to a real array with the same geometry.

Speaker Separation Speech Separation

Sequential Multi-Frame Neural Beamforming for Speech Separation and Enhancement

no code implementations18 Nov 2019 Zhong-Qiu Wang, Hakan Erdogan, Scott Wisdom, Kevin Wilson, Desh Raj, Shinji Watanabe, Zhuo Chen, John R. Hershey

This work introduces sequential neural beamforming, which alternates between neural network based spectral separation and beamforming based spatial separation.

Speaker Separation Speech Enhancement +3

Deep Learning Based Phase Reconstruction for Speaker Separation: A Trigonometric Perspective

no code implementations22 Nov 2018 Zhong-Qiu Wang, Ke Tan, DeLiang Wang

This study investigates phase reconstruction for deep learning based monaural talker-independent speaker separation in the short-time Fourier transform (STFT) domain.

Speaker Separation

End-to-End Speech Separation with Unfolded Iterative Phase Reconstruction

no code implementations26 Apr 2018 Zhong-Qiu Wang, Jonathan Le Roux, DeLiang Wang, John R. Hershey

In addition, we train through unfolded iterations of a phase reconstruction algorithm, represented as a series of STFT and inverse STFT layers.

Speech Separation

Cannot find the paper you are looking for? You can Submit a new open access paper.