no code implementations • 7 Jul 2024 • Qiqi He, Xuchen Song, Weituo Hao, Ju-Chiang Wang, Wei-Tsung Lu, Wei Li
For the case where the artist information is available, we extend the audio-based model to take multimodal inputs and develop a framework, called MultiModal Contrastive (MMC) learning, to enhance the training.
no code implementations • 14 Dec 2023 • Julian D. Parker, Janne Spijkervet, Katerina Kosta, Furkan Yesiler, Boris Kuznetsov, Ju-Chiang Wang, Matt Avent, Jitong Chen, Duc Le
End-to-end generation of musical audio using deep learning techniques has seen an explosion of activity recently.
no code implementations • 2 Oct 2023 • Yun-Ning Hung, Ju-Chiang Wang, Minz Won, Duc Le
To our knowledge, this is the first attempt to study the effects of scaling up both model and training data for a variety of MIR tasks.
no code implementations • 19 Jun 2023 • Wei-Tsung Lu, Ju-Chiang Wang, Yun-Ning Hung
Multitrack music transcription aims to transcribe a music audio input into the musical notes of multiple instruments simultaneously.
no code implementations • 4 Jun 2023 • Mojtaba Heydari, Ju-Chiang Wang, Zhiyao Duan
Singing voice beat and downbeat tracking posses several applications in automatic music production, analysis and manipulation.
no code implementations • 1 Feb 2023 • Kin Wai Cheuk, Keunwoo Choi, Qiuqiang Kong, Bochen Li, Minz Won, Ju-Chiang Wang, Yun-Ning Hung, Dorien Herremans
Jointist consists of an instrument recognition module that conditions the other two modules: a transcription module that outputs instrument-specific piano rolls, and a source separation module that utilizes instrument information and transcription results.
no code implementations • 3 Jan 2023 • Daiyu Zhang, Ju-Chiang Wang, Katerina Kosta, Jordan B. L. Smith, Shicen Zhou
Experiments for Chinese lyric-to-melody generation show that the proposed framework is able to model key characteristics of rhythm and pitch distributions in the dataset, and in a subjective evaluation, the melodies generated by our system were rated as similar to or better than those of a state-of-the-art alternative.
no code implementations • 4 Nov 2022 • Yin Zhu, Qiuqiang Kong, Junjie Shi, Shilei Liu, Xuzhou Ye, Ju-Chiang Wang, Junping Zhang
Binaural rendering of ambisonic signals is of broad interest to virtual reality and immersive media.
no code implementations • 22 Jun 2022 • Kin Wai Cheuk, Keunwoo Choi, Qiuqiang Kong, Bochen Li, Minz Won, Amy Hung, Ju-Chiang Wang, Dorien Herremans
However, its novelty necessitates a new perspective on how to evaluate such a model.
Ranked #4 on Music Transcription on Slakh2100
no code implementations • 29 May 2022 • Ju-Chiang Wang, Yun-Ning Hung, Jordan B. L. Smith
Conventional music structure analysis algorithms aim to divide a song into segments and to group them with abstract labels (e. g., 'A', 'B', and 'C').
no code implementations • 18 Oct 2021 • Ju-Chiang Wang, Jordan B. L. Smith, Wei-Tsung Lu, Xuchen Song
Music structure analysis (MSA) methods traditionally search for musically meaningful patterns in audio: homogeneity, repetition, novelty, and segment-length regularity.
no code implementations • 26 Mar 2021 • Ju-Chiang Wang, Jordan B. L. Smith, Jitong Chen, Xuchen Song, Yuxuan Wang
This paper presents a novel supervised approach to detecting the chorus segments in popular music.
no code implementations • 26 Mar 2021 • Jiawen Huang, Ju-Chiang Wang, Jordan B. L. Smith, Xuchen Song, Yuxuan Wang
A music mashup combines audio elements from two or more songs to create a new work.