1 code implementation • 31 Jul 2023 • Seungheon Doh, Keunwoo Choi, Jongpil Lee, Juhan Nam
In addition, we trained a transformer-based music captioning model with the dataset and evaluated it under zero-shot and transfer-learning settings.
1 code implementation • 25 Mar 2022 • Sangeun Kum, Jongpil Lee, Keunhyoung Luke Kim, Taehyoung Kim, Juhan Nam
We address the issue by using pseudo labels from vocal pitch estimation models given unlabeled data.
no code implementations • 9 Aug 2020 • Jongpil Lee, Nicholas J. Bryan, Justin Salamon, Zeyu Jin, Juhan Nam
For this task, it is typically necessary to define a similarity metric to compare one recording to another.
no code implementations • 9 Aug 2020 • Jongpil Lee, Nicholas J. Bryan, Justin Salamon, Zeyu Jin, Juhan Nam
For this, we (1) outline past work on the relationship between metric learning and classification, (2) extend this relationship to multi-label data by exploring three different learning approaches and their disentangled versions, and (3) evaluate all models on four tasks (training time, similarity retrieval, auto-tagging, and triplet prediction).
no code implementations • 23 Jul 2020 • Seungheon Doh, Jongpil Lee, Tae Hong Park, Juhan Nam
Word embedding pioneered by Mikolov et al. is a staple technique for word representations in natural language processing (NLP) research which has also found popularity in music information retrieval tasks.
1 code implementation • 5 Jul 2019 • Jeong Choi, Jongpil Lee, Jiyoung Park, Juhan Nam
Audio-based music classification and tagging is typically based on categorical supervised learning with a fixed set of labels.
no code implementations • 27 Jun 2019 • Jongpil Lee, Jiyoung Park, Juhan Nam
Supervised music representation learning has been performed mainly using semantic labels such as music genres.
no code implementations • 20 Jun 2019 • Jeong Choi, Jongpil Lee, Jiyoung Park, Juhan Nam
Music classification and tagging is conducted through categorical supervised learning with a fixed set of labels.
2 code implementations • 18 Jul 2018 • Jongpil Lee, Kyungyun Lee, Jiyoung Park, Jang-Yeon Park, Juhan Nam
Recently deep learning based recommendation systems have been actively explored to solve the cold-start problem using a hybrid approach.
no code implementations • 4 Dec 2017 • Jongpil Lee, Taejun Kim, Jiyoung Park, Juhan Nam
Music, speech, and acoustic scene sound are often handled separately in the audio domain because of their different signal characteristics.
2 code implementations • 28 Oct 2017 • Taejun Kim, Jongpil Lee, Juhan Nam
Recent work has shown that the end-to-end approach using convolutional neural network (CNN) is effective in various types of machine learning tasks.
2 code implementations • 18 Oct 2017 • Jiyoung Park, Jongpil Lee, Jangyeon Park, Jung-Woo Ha, Juhan Nam
In this paper, we present a supervised feature learning approach using artist labels annotated in every single track as objective meta data.
Sound Audio and Speech Processing
1 code implementation • 21 Jun 2017 • Jongpil Lee, Juhan Nam
Music tag words that describe music audio by text have different levels of abstraction.
3 code implementations • 6 Mar 2017 • Jongpil Lee, Jiyoung Park, Keunhyoung Luke Kim, Juhan Nam
Recently, the end-to-end approach that learns hierarchical representations from raw data using deep convolutional neural networks has been successfully explored in the image, text and speech domains.
1 code implementation • 6 Mar 2017 • Jongpil Lee, Juhan Nam
Second, we extract audio features from each layer of the pre-trained convolutional networks separately and aggregate them altogether given a long audio clip.