Search Results for author: Keunwoo Choi

Found 26 papers, 15 papers with code

LP-MusicCaps: LLM-Based Pseudo Music Captioning

1 code implementation31 Jul 2023 Seungheon Doh, Keunwoo Choi, Jongpil Lee, Juhan Nam

In addition, we trained a transformer-based music captioning model with the dataset and evaluated it under zero-shot and transfer-learning settings.

Language Modelling Large Language Model +3

A Demand-Driven Perspective on Generative Audio AI

no code implementations10 Jul 2023 Sangshin Oh, Minsung Kang, Hyeongi Moon, Keunwoo Choi, Ben Sangbae Chon

To achieve successful deployment of AI research, it is crucial to understand the demands of the industry.

Audio Generation

Textless Speech-to-Music Retrieval Using Emotion Similarity

no code implementations19 Mar 2023 Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam

We introduce a framework that recommends music based on the emotions of speech.

Retrieval

Jointist: Simultaneous Improvement of Multi-instrument Transcription and Music Source Separation via Joint Training

no code implementations1 Feb 2023 Kin Wai Cheuk, Keunwoo Choi, Qiuqiang Kong, Bochen Li, Minz Won, Ju-Chiang Wang, Yun-Ning Hung, Dorien Herremans

Jointist consists of an instrument recognition module that conditions the other two modules: a transcription module that outputs instrument-specific piano rolls, and a source separation module that utilizes instrument information and transcription results.

Chord Recognition Instrument Recognition +1

Toward Universal Text-to-Music Retrieval

3 code implementations26 Nov 2022 Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam

This paper introduces effective design choices for text-to-music retrieval systems.

Music Classification Retrieval +2

MedleyVox: An Evaluation Dataset for Multiple Singing Voices Separation

1 code implementation14 Nov 2022 Chang-Bin Jeon, Hyeongi Moon, Keunwoo Choi, Ben Sangbae Chon, Kyogu Lee

Second, to overcome the absence of existing multi-singing datasets for a training purpose, we present a strategy for construction of multiple singing mixtures using various single-singing datasets.

Music Source Separation Super-Resolution

A Proposal for Foley Sound Synthesis Challenge

no code implementations21 Jul 2022 Keunwoo Choi, Sangshin Oh, Minsung Kang, Brian McFee

"Foley" refers to sound effects that are added to multimedia during post-production to enhance its perceived acoustic properties, e. g., by simulating the sounds of footsteps, ambient environmental sounds, or visible objects on the screen.

Music Classification: Beyond Supervised Learning, Towards Real-world Applications

1 code implementation23 Nov 2021 Minz Won, Janne Spijkervet, Keunwoo Choi

The target audience for this web book is researchers and practitioners who are interested in state-of-the-art music classification research and building real-world applications.

Classification Information Retrieval +4

Listen, Read, and Identify: Multimodal Singing Language Identification of Music

no code implementations2 Mar 2021 Keunwoo Choi, Yuxuan Wang

Optionally, LRID-Net is facilitated with modality dropouts to handle a missing modality.

Language Identification

Large-Scale MIDI-based Composer Classification

no code implementations28 Oct 2020 Qiuqiang Kong, Keunwoo Choi, Yuxuan Wang

Music classification is a task to classify a music piece into labels such as genres or composers.

Classification General Classification +1

Dereverberation using joint estimation of dry speech signal and acoustic system

no code implementations24 Jul 2020 Sanna Wager, Keunwoo Choi, Simon Durand

The purpose of speech dereverberation is to remove quality-degrading effects of a time-invariant impulse response filter from the signal.

Room Impulse Response (RIR) Speech Dereverberation

Deep Unsupervised Drum Transcription

2 code implementations9 Jun 2019 Keunwoo Choi, Kyunghyun Cho

We introduce DrummerNet, a drum transcription system that is trained in an unsupervised manner.

Sound Audio and Speech Processing

Revisiting Singing Voice Detection: a Quantitative Review and the Future Outlook

4 code implementations4 Jun 2018 Kyungyun Lee, Keunwoo Choi, Juhan Nam

Since the vocal component plays a crucial role in popular music, singing voice detection has been an active research topic in music information retrieval.

Information Retrieval Music Information Retrieval +1

A Tutorial on Deep Learning for Music Information Retrieval

2 code implementations13 Sep 2017 Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler

Following their success in Computer Vision and other areas, deep learning techniques have recently become widely adopted in Music Information Retrieval (MIR) research.

Information Retrieval Music Information Retrieval +2

A Comparison of Audio Signal Preprocessing Methods for Deep Neural Networks on Music Tagging

1 code implementation6 Sep 2017 Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler

In this paper, we empirically investigate the effect of audio preprocessing on music tagging with deep neural networks.

Music Tagging

Transfer learning for music classification and regression tasks

3 code implementations27 Mar 2017 Keunwoo Choi, György Fazekas, Mark Sandler, Kyunghyun Cho

In this paper, we present a transfer learning approach for music classification and regression tasks.

Classification General Classification +4

Explaining Deep Convolutional Neural Networks on Music Classification

1 code implementation8 Jul 2016 Keunwoo Choi, George Fazekas, Mark Sandler

Deep convolutional neural networks (CNNs) have been actively adopted in the field of music information retrieval, e. g. genre classification, mood detection, and chord recognition.

Chord Recognition Classification +6

Towards Playlist Generation Algorithms Using RNNs Trained on Within-Track Transitions

no code implementations7 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We introduce a novel playlist generation algorithm that focuses on the quality of transitions using a recurrent neural network (RNN).

Automatic tagging using deep convolutional neural networks

11 code implementations1 Jun 2016 Keunwoo Choi, George Fazekas, Mark Sandler

We present a content-based automatic music tagging algorithm using fully convolutional neural networks (FCNs).

Music Tagging

Text-based LSTM networks for Automatic Music Composition

4 code implementations18 Apr 2016 Keunwoo Choi, George Fazekas, Mark Sandler

In this paper, we introduce new methods and discuss results of text-based LSTM (Long Short-Term Memory) networks for automatic music composition.

Cannot find the paper you are looking for? You can Submit a new open access paper.