Search Results for author: Xuchen Song

Found 8 papers, 2 papers with code

A Long-Tail Friendly Representation Framework for Artist and Music Similarity

no code implementations8 Sep 2023 Haoran Xiang, Junyu Dai, Xuchen Song, Furao Shen

The investigation of the similarity between artists and music is crucial in music retrieval and recommendation, and addressing the challenge of the long-tail phenomenon is increasingly important.

Metric Learning Music Recommendation +1

ALCAP: Alignment-Augmented Music Captioner

1 code implementation21 Dec 2022 Zihao He, Weituo Hao, Wei-Tsung Lu, Changyou Chen, Kristina Lerman, Xuchen Song

Music captioning has gained significant attention in the wake of the rising prominence of streaming media platforms.

Contrastive Learning Music Captioning +1

Supervised Metric Learning for Music Structure Features

no code implementations18 Oct 2021 Ju-Chiang Wang, Jordan B. L. Smith, Wei-Tsung Lu, Xuchen Song

Music structure analysis (MSA) methods traditionally search for musically meaningful patterns in audio: homogeneity, repetition, novelty, and segment-length regularity.

Metric Learning

Modeling the Compatibility of Stem Tracks to Generate Music Mashups

no code implementations26 Mar 2021 Jiawen Huang, Ju-Chiang Wang, Jordan B. L. Smith, Xuchen Song, Yuxuan Wang

A music mashup combines audio elements from two or more songs to create a new work.

High-resolution Piano Transcription with Pedals by Regressing Onset and Offset Times

3 code implementations5 Oct 2020 Qiuqiang Kong, Bochen Li, Xuchen Song, Yuan Wan, Yuxuan Wang

In addition, previous AMT systems are sensitive to the misaligned onset and offset labels of audio recordings.

Sound Audio and Speech Processing

Noise Robust TTS for Low Resource Speakers using Pre-trained Model and Speech Enhancement

no code implementations26 May 2020 Dongyang Dai, Li Chen, Yu-Ping Wang, Mu Wang, Rui Xia, Xuchen Song, Zhiyong Wu, Yuxuan Wang

Firstly, the speech synthesis model is pre-trained with both multi-speaker clean data and noisy augmented data; then the pre-trained model is adapted on noisy low-resource new speaker data; finally, by setting the clean speech condition, the model can synthesize the new speaker's clean voice.

Speech Enhancement Speech Synthesis

Cannot find the paper you are looking for? You can Submit a new open access paper.