no code implementations • 11 Feb 2025 • Sungnyun Kim, Kangwook Jang, Sangmin Bae, Sungwoo Cho, Se-Young Yun
Audio-visual speech recognition (AVSR) has become critical for enhancing speech recognition in noisy environments by integrating both auditory and visual modalities.
no code implementations • 28 Oct 2024 • Sangmin Bae, Adam Fisch, Hrayr Harutyunyan, Ziwei Ji, Seungyeon Kim, Tal Schuster
In this work, we revisit "layer tying" as form of parameter sharing in Transformers, and introduce novel methods for converting existing LLMs into smaller "Recursive Transformers" that share parameters across layers, with minimal loss of performance.
no code implementations • 14 Oct 2024 • Yongjin Yang, Sihyeon Kim, Hojung Jung, Sangmin Bae, Sangmook Kim, Se-Young Yun, Kimin Lee
Fine-tuning text-to-image diffusion models with human feedback is an effective method for aligning model behavior with human intentions.
no code implementations • 26 Jul 2024 • Sihyeon Kim, Boryeong Cho, Sangmin Bae, Sumyeong Ahn, Se-Young Yun
To address this issue, previous studies have focused on mitigating hallucinations by employing contrastive decoding (CD) with augmented images, which amplifies the contrast with the original image.
1 code implementation • 20 Jul 2024 • Yunseon Choi, Sangmin Bae, Seonghyun Ban, Minchan Jeong, Chuheng Zhang, Lei Song, Li Zhao, Jiang Bian, Kee-Eung Kim
With the advent of foundation models, prompt tuning has positioned itself as an important technique for directing model behaviors and eliciting desired responses.
1 code implementation • 4 Jul 2024 • Sungnyun Kim, Kangwook Jang, Sangmin Bae, Hoirin Kim, Se-Young Yun
Cross-modal attention modules are introduced to enrich video features with audio information so that speech variability can be taken into account when training on the video temporal dynamics.
1 code implementation • 4 Jun 2024 • Namgyu Ho, Sangmin Bae, Taehyeon Kim, Hyunjik Jo, Yireun Kim, Tal Schuster, Adam Fisch, James Thorne, Se-Young Yun
We introduce the Block Transformer which adopts hierarchical global-to-local modeling to autoregressive transformers to mitigate the inference bottlenecks associated with self-attention.
1 code implementation • 22 May 2024 • Felix den Breejen, Sangmin Bae, Stephen Cha, Se-Young Yun
Based on this observation, we propose to pretrain ICL-transformers on a new forest dataset generator which creates datasets that are unrealistic, but have complex decision boundaries.
no code implementations • 5 May 2024 • June-Woo Kim, Miika Toikkanen, Sangmin Bae, Minseok Kim, Ho-Young Jung
To address this, we propose RepAugment, an input-agnostic representation-level augmentation technique that outperforms SpecAugment, but is also suitable for respiratory sound classification with waveform pretrained models.
1 code implementation • 15 Dec 2023 • June-Woo Kim, Sangmin Bae, Won-Yang Cho, Byungjo Lee, Ho-Young Jung
Despite the remarkable advances in deep learning technology, achieving satisfactory performance in lung sound classification remains a challenge due to the scarcity of available data.
Ranked #8 on
Audio Classification
on ICBHI Respiratory Sound Database
(using extra training data)
1 code implementation • 14 Nov 2023 • Yujin Kim, Jaehong Yoon, Seonghyeon Ye, Sangmin Bae, Namgyu Ho, Sung Ju Hwang, Se-Young Yun
The dynamic nature of knowledge in an ever-changing world presents challenges for language models trained on static data; the model in the real world often requires not only acquiring new knowledge but also overwriting outdated information into updated ones.
no code implementations • 13 Nov 2023 • Felix den Breejen, Sangmin Bae, Stephen Cha, Tae-Young Kim, Seoung Hyun Koh, Se-Young Yun
While interests in tabular deep learning has significantly grown, conventional tree-based models still outperform deep learning methods.
1 code implementation • 11 Nov 2023 • June-Woo Kim, Chihyeon Yoon, Miika Toikkanen, Sangmin Bae, Ho-Young Jung
In this work, we propose a straightforward approach to augment imbalanced respiratory sound data using an audio diffusion model as a conditional neural vocoder.
Ranked #7 on
Audio Classification
on ICBHI Respiratory Sound Database
(using extra training data)
2 code implementations • 9 Oct 2023 • Sangmin Bae, Jongwoo Ko, Hwanjun Song, Se-Young Yun
To tackle the high inference latency exhibited by autoregressive language models, previous studies have proposed an early-exiting framework that allocates adaptive computation paths for each token based on the complexity of generating the subsequent token.
1 code implementation • 23 May 2023 • Sangmin Bae, June-Woo Kim, Won-Yang Cho, Hyerim Baek, Soyoun Son, Byungjo Lee, Changwan Ha, Kyongpil Tae, Sungnyun Kim, Se-Young Yun
Respiratory sound contains crucial information for the early diagnosis of fatal lung diseases.
Ranked #6 on
Audio Classification
on ICBHI Respiratory Sound Database
(using extra training data)
1 code implementation • CVPR 2023 • Sangmook Kim, Sangmin Bae, Hwanjun Song, Se-Young Yun
In this work, we first demonstrate that the superiority of two selector models depends on the global and local inter-class diversity.
1 code implementation • CVPR 2023 • Sungnyun Kim, Sangmin Bae, Se-Young Yun
Fortunately, the recent self-supervised learning (SSL) is a promising approach to pretrain a model without annotations, serving as an effective initialization for any downstream tasks.
no code implementations • 29 Sep 2021 • Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun
This paper proposes a novel contrastive learning framework, called Self-Contrastive (SelfCon) Learning, that self-contrasts within multiple outputs from the different levels of a multi-exit network.
2 code implementations • 29 Jun 2021 • Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun
To this end, we propose Self-Contrastive (SelfCon) learning, which self-contrasts within multiple outputs from the different levels of a single network.
3 code implementations • 6 Jun 2021 • Gihun Lee, Minchan Jeong, Yongjin Shin, Sangmin Bae, Se-Young Yun
In federated learning, a strong global model is collaboratively learned by aggregating clients' locally trained models.
no code implementations • 6 Dec 2020 • Taehyeon Kim, Sangmin Bae, Jin-woo Lee, Seyoung Yun
Federated learning has emerged as an innovative paradigm of collaborative machine learning.
1 code implementation • 13 Oct 2020 • Sungnyun Kim, Gihun Lee, Sangmin Bae, Se-Young Yun
Contrastive learning has shown remarkable results in recent self-supervised approaches for visual representation.
1 code implementation • 24 Apr 2020 • Gihun Lee, Sangmin Bae, Jaehoon Oh, Se-Young Yun
With the success of deep learning in various fields and the advent of numerous Internet of Things (IoT) devices, it is essential to lighten models suitable for low-power devices.