Search Results for author: Chih-Wei Wu

Found 7 papers, 6 papers with code

ODAQ: Open Dataset of Audio Quality

2 code implementations30 Dec 2023 Matteo Torcoli, Chih-Wei Wu, Sascha Dick, Phillip A. Williams, Mhd Modar Halimeh, William Wolcott, Emanuel A. P. Habets

Research into the prediction and analysis of perceived audio quality is hampered by the scarcity of openly available datasets of audio signals accompanied by corresponding subjective quality scores.

A Generalized Bandsplit Neural Network for Cinematic Audio Source Separation

1 code implementation5 Sep 2023 Karn N. Watcharasupat, Chih-Wei Wu, Yiwei Ding, Iroro Orife, Aaron J. Hipple, Phillip A. Williams, Scott Kramer, Alexander Lerch, William Wolcott

Cinematic audio source separation is a relatively new subtask of audio source separation, with the aim of extracting the dialogue, music, and effects stems from their mixture.

Audio Source Separation

Looking Similar, Sounding Different: Leveraging Counterfactual Cross-Modal Pairs for Audiovisual Representation Learning

no code implementations12 Apr 2023 Nikhil Singh, Chih-Wei Wu, Iroro Orife, Mahdi Kalayeh

We additionally compare this approach to a strong baseline where we remove speech before pretraining, and find that dub-augmented training is more effective, including for paralinguistic and audiovisual tasks where speech removal leads to worse performance.

Contrastive Learning counterfactual +1

Learning to Fuse Music Genres with Generative Adversarial Dual Learning

1 code implementation5 Dec 2017 Zhiqian Chen, Chih-Wei Wu, Yen-Cheng Lu, Alexander Lerch, Chang-Tien Lu

FusionGAN is a novel genre fusion framework for music generation that integrates the strengths of generative adversarial networks and dual learning.

Music Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.