Search Results for author: Zhongweiyang Xu

Found 5 papers, 1 papers with code

Multi-Source Music Generation with Latent Diffusion

1 code implementation10 Sep 2024 Zhongweiyang Xu, Debottam Dutta, Yu-Lin Wei, Romit Roy Choudhury

Its goal is to use one single diffusion model to generate mutually-coherent music sources, that are then mixed to form the music.

FAD Music Generation

uSee: Unified Speech Enhancement and Editing with Conditional Diffusion Models

no code implementations2 Oct 2023 Muqiao Yang, Chunlei Zhang, Yong Xu, Zhongweiyang Xu, Heming Wang, Bhiksha Raj, Dong Yu

Speech enhancement aims to improve the quality of speech signals in terms of quality and intelligibility, and speech editing refers to the process of editing the speech according to specific user needs.

Denoising Self-Supervised Learning +2

Dual-Path Cross-Modal Attention for better Audio-Visual Speech Extraction

no code implementations9 Jul 2022 Zhongweiyang Xu, Xulin Fan, Mark Hasegawa-Johnson

Most current research upsamples the visual features along the time dimension so that audio and video features are able to align in time.

Speech Extraction

Learning to Separate Voices by Spatial Regions

no code implementations9 Jul 2022 Zhongweiyang Xu, Romit Roy Choudhury

We consider the problem of audio voice separation for binaural applications, such as earphones and hearing aids.

Cannot find the paper you are looking for? You can Submit a new open access paper.