1 code implementation • 21 Feb 2025 • Yoonjin Chung, Pilsun Eu, Junwon Lee, Keunwoo Choi, Juhan Nam, Ben Sangbae Chon
Although being widely adopted for evaluating generated audio signals, the Fr\'echet Audio Distance (FAD) suffers from significant limitations, including reliance on Gaussian assumptions, sensitivity to sample size, and high computational complexity.
no code implementations • 19 Feb 2025 • Seungheon Doh, Keunwoo Choi, Juhan Nam
We present TalkPlay, a multimodal music recommendation system that reformulates the recommendation task as large language model token generation.
no code implementations • 15 Jan 2025 • Mathieu Lagrange, Junwon Lee, Modan Tailleur, Laurie M. Heller, Keunwoo Choi, Brian McFee, Keisuke Imoto, Yuki Okamoto
This paper presents Task 7 at the DCASE 2024 Challenge: sound scene synthesis.
1 code implementation • 11 Nov 2024 • Seungheon Doh, Keunwoo Choi, Daeyong Kwon, Taesu Kim, Juhan Nam
To achieve this, a conversational music retrieval system should seamlessly engage in multi-turn conversation by 1) understanding user queries and 2) responding with natural language and retrieved music.
1 code implementation • 23 Oct 2024 • Junwon Lee, Modan Tailleur, Laurie M. Heller, Keunwoo Choi, Mathieu Lagrange, Brian McFee, Keisuke Imoto, Yuki Okamoto
Despite significant advancements in neural text-to-audio generation, challenges persist in controllability and evaluation.
1 code implementation • 31 Jul 2023 • Seungheon Doh, Keunwoo Choi, Jongpil Lee, Juhan Nam
In addition, we trained a transformer-based music captioning model with the dataset and evaluated it under zero-shot and transfer-learning settings.
no code implementations • 10 Jul 2023 • Sangshin Oh, Minsung Kang, Hyeongi Moon, Keunwoo Choi, Ben Sangbae Chon
To achieve successful deployment of AI research, it is crucial to understand the demands of the industry.
no code implementations • 19 Mar 2023 • Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam
We introduce a framework that recommends music based on the emotions of speech.
no code implementations • 1 Feb 2023 • Kin Wai Cheuk, Keunwoo Choi, Qiuqiang Kong, Bochen Li, Minz Won, Ju-Chiang Wang, Yun-Ning Hung, Dorien Herremans
Jointist consists of an instrument recognition module that conditions the other two modules: a transcription module that outputs instrument-specific piano rolls, and a source separation module that utilizes instrument information and transcription results.
3 code implementations • 26 Nov 2022 • Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam
This paper introduces effective design choices for text-to-music retrieval systems.
1 code implementation • 14 Nov 2022 • Chang-Bin Jeon, Hyeongi Moon, Keunwoo Choi, Ben Sangbae Chon, Kyogu Lee
Second, to overcome the absence of existing multi-singing datasets for a training purpose, we present a strategy for construction of multiple singing mixtures using various single-singing datasets.
no code implementations • 21 Jul 2022 • Keunwoo Choi, Sangshin Oh, Minsung Kang, Brian McFee
"Foley" refers to sound effects that are added to multimedia during post-production to enhance its perceived acoustic properties, e. g., by simulating the sounds of footsteps, ambient environmental sounds, or visible objects on the screen.
no code implementations • 22 Jun 2022 • Kin Wai Cheuk, Keunwoo Choi, Qiuqiang Kong, Bochen Li, Minz Won, Amy Hung, Ju-Chiang Wang, Dorien Herremans
However, its novelty necessitates a new perspective on how to evaluate such a model.
Ranked #4 on
Music Transcription
on Slakh2100
1 code implementation • 23 Nov 2021 • Minz Won, Janne Spijkervet, Keunwoo Choi
The target audience for this web book is researchers and practitioners who are interested in state-of-the-art music classification research and building real-world applications.
no code implementations • 2 Mar 2021 • Keunwoo Choi, Yuxuan Wang
Optionally, LRID-Net is facilitated with modality dropouts to handle a missing modality.
no code implementations • 28 Oct 2020 • Qiuqiang Kong, Keunwoo Choi, Yuxuan Wang
Music classification is a task to classify a music piece into labels such as genres or composers.
1 code implementation • 2 Oct 2020 • Sunghyeon Kim, Hyeyoon Lee, Sunjong Park, Jinho Lee, Keunwoo Choi
In this study, we train deep neural networks to classify composer on a symbolic domain.
no code implementations • 24 Jul 2020 • Sanna Wager, Keunwoo Choi, Simon Durand
The purpose of speech dereverberation is to remove quality-degrading effects of a time-invariant impulse response filter from the signal.
2 code implementations • 9 Jun 2019 • Keunwoo Choi, Kyunghyun Cho
We introduce DrummerNet, a drum transcription system that is trained in an unsupervised manner.
Sound Audio and Speech Processing
4 code implementations • 4 Jun 2018 • Kyungyun Lee, Keunwoo Choi, Juhan Nam
Since the vocal component plays a crucial role in popular music, singing voice detection has been an active research topic in music information retrieval.
2 code implementations • 13 Sep 2017 • Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler
Following their success in Computer Vision and other areas, deep learning techniques have recently become widely adopted in Music Information Retrieval (MIR) research.
1 code implementation • 6 Sep 2017 • Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler
In this paper, we empirically investigate the effect of audio preprocessing on music tagging with deep neural networks.
7 code implementations • 19 Jun 2017 • Keunwoo Choi, Deokjin Joo, Ju-ho Kim
We introduce Kapre, Keras layers for audio and music signal preprocessing.
no code implementations • 7 Jun 2017 • Keunwoo Choi, George Fazekas, Kyunghyun Cho, Mark Sandler
The results highlight several important aspects of music tagging and neural networks.
3 code implementations • 27 Mar 2017 • Keunwoo Choi, György Fazekas, Mark Sandler, Kyunghyun Cho
In this paper, we present a transfer learning approach for music classification and regression tasks.
13 code implementations • 14 Sep 2016 • Keunwoo Choi, George Fazekas, Mark Sandler, Kyunghyun Cho
We introduce a convolutional recurrent neural network (CRNN) for music tagging.
no code implementations • 17 Aug 2016 • Keunwoo Choi, George Fazekas, Brian McFee, Kyunghyun Cho, Mark Sandler
Descriptions are often provided along with recommendations to help users' discovery.
1 code implementation • 8 Jul 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.
no code implementations • 7 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).
11 code implementations • 1 Jun 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).
4 code implementations • 18 Apr 2016 • Keunwoo Choi, George Fazekas, Mark Sandler
In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.