Search Results for author: Sangmin Bae

Found 23 papers, 15 papers with code

MoHAVE: Mixture of Hierarchical Audio-Visual Experts for Robust Speech Recognition

no code implementations11 Feb 2025 Sungnyun Kim, Kangwook Jang, Sangmin Bae, Sungwoo Cho, Se-Young Yun

Audio-visual speech recognition (AVSR) has become critical for enhancing speech recognition in noisy environments by integrating both auditory and visual modalities.

Audio-Visual Speech Recognition Computational Efficiency +3

Relaxed Recursive Transformers: Effective Parameter Sharing with Layer-wise LoRA

no code implementations28 Oct 2024 Sangmin Bae, Adam Fisch, Hrayr Harutyunyan, Ziwei Ji, Seungyeon Kim, Tal Schuster

In this work, we revisit "layer tying" as form of parameter sharing in Transformers, and introduce novel methods for converting existing LLMs into smaller "Recursive Transformers" that share parameters across layers, with minimal loss of performance.

Knowledge Distillation

Automated Filtering of Human Feedback Data for Aligning Text-to-Image Diffusion Models

no code implementations14 Oct 2024 Yongjin Yang, Sihyeon Kim, Hojung Jung, Sangmin Bae, Sangmook Kim, Se-Young Yun, Kimin Lee

Fine-tuning text-to-image diffusion models with human feedback is an effective method for aligning model behavior with human intentions.

Diversity

VACoDe: Visual Augmented Contrastive Decoding

no code implementations26 Jul 2024 Sihyeon Kim, Boryeong Cho, Sangmin Bae, Sumyeong Ahn, Se-Young Yun

To address this issue, previous studies have focused on mitigating hallucinations by employing contrastive decoding (CD) with augmented images, which amplifies the contrast with the original image.

Hard Prompts Made Interpretable: Sparse Entropy Regularization for Prompt Tuning with RL

1 code implementation20 Jul 2024 Yunseon Choi, Sangmin Bae, Seonghyun Ban, Minchan Jeong, Chuheng Zhang, Lei Song, Li Zhao, Jiang Bian, Kee-Eung Kim

With the advent of foundation models, prompt tuning has positioned itself as an important technique for directing model behaviors and eliciting desired responses.

Few-Shot Text Classification Q-Learning +5

Learning Video Temporal Dynamics with Cross-Modal Attention for Robust Audio-Visual Speech Recognition

1 code implementation4 Jul 2024 Sungnyun Kim, Kangwook Jang, Sangmin Bae, Hoirin Kim, Se-Young Yun

Cross-modal attention modules are introduced to enrich video features with audio information so that speech variability can be taken into account when training on the video temporal dynamics.

Audio-Visual Speech Recognition speech-recognition +1

Block Transformer: Global-to-Local Language Modeling for Fast Inference

1 code implementation4 Jun 2024 Namgyu Ho, Sangmin Bae, Taehyeon Kim, Hyunjik Jo, Yireun Kim, Tal Schuster, Adam Fisch, James Thorne, Se-Young Yun

We introduce the Block Transformer which adopts hierarchical global-to-local modeling to autoregressive transformers to mitigate the inference bottlenecks associated with self-attention.

Language Modeling Language Modelling

Fine-tuned In-Context Learning Transformers are Excellent Tabular Data Classifiers

1 code implementation22 May 2024 Felix den Breejen, Sangmin Bae, Stephen Cha, Se-Young Yun

Based on this observation, we propose to pretrain ICL-transformers on a new forest dataset generator which creates datasets that are unrealistic, but have complex decision boundaries.

In-Context Learning

RepAugment: Input-Agnostic Representation-Level Augmentation for Respiratory Sound Classification

no code implementations5 May 2024 June-Woo Kim, Miika Toikkanen, Sangmin Bae, Minseok Kim, Ho-Young Jung

To address this, we propose RepAugment, an input-agnostic representation-level augmentation technique that outperforms SpecAugment, but is also suitable for respiratory sound classification with waveform pretrained models.

Data Augmentation Sound Classification

Stethoscope-guided Supervised Contrastive Learning for Cross-domain Adaptation on Respiratory Sound Classification

1 code implementation15 Dec 2023 June-Woo Kim, Sangmin Bae, Won-Yang Cho, Byungjo Lee, Ho-Young Jung

Despite the remarkable advances in deep learning technology, achieving satisfactory performance in lung sound classification remains a challenge due to the scarcity of available data.

Ranked #8 on Audio Classification on ICBHI Respiratory Sound Database (using extra training data)

Audio Classification Contrastive Learning +2

Carpe Diem: On the Evaluation of World Knowledge in Lifelong Language Models

1 code implementation14 Nov 2023 Yujin Kim, Jaehong Yoon, Seonghyeon Ye, Sangmin Bae, Namgyu Ho, Sung Ju Hwang, Se-Young Yun

The dynamic nature of knowledge in an ever-changing world presents challenges for language models trained on static data; the model in the real world often requires not only acquiring new knowledge but also overwriting outdated information into updated ones.

Continual Learning Question Answering +1

Fine-Tuning the Retrieval Mechanism for Tabular Deep Learning

no code implementations13 Nov 2023 Felix den Breejen, Sangmin Bae, Stephen Cha, Tae-Young Kim, Seoung Hyun Koh, Se-Young Yun

While interests in tabular deep learning has significantly grown, conventional tree-based models still outperform deep learning methods.

Deep Learning Retrieval +1

Adversarial Fine-tuning using Generated Respiratory Sound to Address Class Imbalance

1 code implementation11 Nov 2023 June-Woo Kim, Chihyeon Yoon, Miika Toikkanen, Sangmin Bae, Ho-Young Jung

In this work, we propose a straightforward approach to augment imbalanced respiratory sound data using an audio diffusion model as a conditional neural vocoder.

Ranked #7 on Audio Classification on ICBHI Respiratory Sound Database (using extra training data)

Audio Classification Sound Classification

Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding

2 code implementations9 Oct 2023 Sangmin Bae, Jongwoo Ko, Hwanjun Song, Se-Young Yun

To tackle the high inference latency exhibited by autoregressive language models, previous studies have proposed an early-exiting framework that allocates adaptive computation paths for each token based on the complexity of generating the subsequent token.

Re-thinking Federated Active Learning based on Inter-class Diversity

1 code implementation CVPR 2023 Sangmook Kim, Sangmin Bae, Hwanjun Song, Se-Young Yun

In this work, we first demonstrate that the superiority of two selector models depends on the global and local inter-class diversity.

Active Learning Diversity +1

Coreset Sampling from Open-Set for Fine-Grained Self-Supervised Learning

1 code implementation CVPR 2023 Sungnyun Kim, Sangmin Bae, Se-Young Yun

Fortunately, the recent self-supervised learning (SSL) is a promising approach to pretrain a model without annotations, serving as an effective initialization for any downstream tasks.

Representation Learning Self-Supervised Learning

Self-Contrastive Learning

no code implementations29 Sep 2021 Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun

This paper proposes a novel contrastive learning framework, called Self-Contrastive (SelfCon) Learning, that self-contrasts within multiple outputs from the different levels of a multi-exit network.

Contrastive Learning

Self-Contrastive Learning: Single-viewed Supervised Contrastive Framework using Sub-network

2 code implementations29 Jun 2021 Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun

To this end, we propose Self-Contrastive (SelfCon) learning, which self-contrasts within multiple outputs from the different levels of a single network.

Contrastive Learning

Preservation of the Global Knowledge by Not-True Distillation in Federated Learning

3 code implementations6 Jun 2021 Gihun Lee, Minchan Jeong, Yongjin Shin, Sangmin Bae, Se-Young Yun

In federated learning, a strong global model is collaboratively learned by aggregating clients' locally trained models.

Continual Learning Federated Learning +1

MixCo: Mix-up Contrastive Learning for Visual Representation

1 code implementation13 Oct 2020 Sungnyun Kim, Gihun Lee, Sangmin Bae, Se-Young Yun

Contrastive learning has shown remarkable results in recent self-supervised approaches for visual representation.

Contrastive Learning Linear evaluation +1

SIPA: A Simple Framework for Efficient Networks

1 code implementation24 Apr 2020 Gihun Lee, Sangmin Bae, Jaehoon Oh, Se-Young Yun

With the success of deep learning in various fields and the advent of numerous Internet of Things (IoT) devices, it is essential to lighten models suitable for low-power devices.

Math

Cannot find the paper you are looking for? You can Submit a new open access paper.