Search Results for author: Kang Min Yoo

Found 26 papers, 13 papers with code

Attribute Injection for Pretrained Language Models: A New Benchmark and an Efficient Method

1 code implementation COLING 2022 Reinald Kim Amplayo, Kang Min Yoo, Sang-Woo Lee

Metadata attributes (e. g., user and product IDs from reviews) can be incorporated as additional inputs to neural-based NLP models, by expanding the architecture of the models to improve performance.

Aligning Large Language Models through Synthetic Feedback

no code implementations23 May 2023 Sungdong Kim, Sanghwan Bae, Jamin Shin, Soyoung Kang, Donghyun Kwak, Kang Min Yoo, Minjoon Seo

In this work, we propose a novel framework for alignment learning with almost no human labor and no dependency on pre-aligned LLMs.

Language Modelling

Memory-Efficient Fine-Tuning of Compressed Large Language Models via sub-4-bit Integer Quantization

no code implementations23 May 2023 Jeonghoon Kim, Jung Hyun Lee, Sungdong Kim, Joonsuk Park, Kang Min Yoo, Se Jung Kwon, Dongsoo Lee

Such a strategy compresses the size of the model considerably, leading to a lower inference latency upon deployment and a reduction in the overall memory required.

Model Compression Natural Language Understanding +1

Probing Out-of-Distribution Robustness of Language Models with Parameter-Efficient Transfer Learning

no code implementations27 Jan 2023 Hyunsoo Cho, Choonghyun Park, Junyeop Kim, Hyuhng Joon Kim, Kang Min Yoo, Sang-goo Lee

As the size of the pre-trained language model (PLM) continues to increase, numerous parameter-efficient transfer learning methods have been proposed recently to compensate for the tremendous cost of fine-tuning.

Language Modelling Transfer Learning

Critic-Guided Decoding for Controlled Text Generation

no code implementations21 Dec 2022 Minbeom Kim, Hwanhee Lee, Kang Min Yoo, Joonsuk Park, Hwaran Lee, Kyomin Jung

In this work, we propose a novel critic decoding method for controlled language generation (CriticControl) that combines the strengths of reinforcement learning and weighted decoding.

Language Modelling reinforcement-learning +2

Prompt-Augmented Linear Probing: Scaling Beyond The Limit of Few-shot In-Context Learners

no code implementations21 Dec 2022 Hyunsoo Cho, Hyuhng Joon Kim, Junyeob Kim, Sang-Woo Lee, Sang-goo Lee, Kang Min Yoo, Taeuk Kim

Through in-context learning (ICL), large-scale language models are effective few-shot learners without additional model fine-tuning.

Language Modelling

Enhancing Out-of-Distribution Detection in Natural Language Understanding via Implicit Layer Ensemble

1 code implementation20 Oct 2022 Hyunsoo Cho, Choonghyun Park, Jaewook Kang, Kang Min Yoo, Taeuk Kim, Sang-goo Lee

Out-of-distribution (OOD) detection aims to discern outliers from the intended data distribution, which is crucial to maintaining high reliability and a good user experience.

Contrastive Learning intent-classification +5

AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models

no code implementations8 Oct 2022 Se Jung Kwon, Jeonghoon Kim, Jeongin Bae, Kang Min Yoo, Jin-Hwa Kim, Baeseong Park, Byeongwook Kim, Jung-Woo Ha, Nako Sung, Dongsoo Lee

To combine parameter-efficient adaptation and model compression, we propose AlphaTuning consisting of post-training quantization of the pre-trained language model and fine-tuning only some parts of quantized parameters for a target task.

Language Modelling Model Compression +1

Continuous Decomposition of Granularity for Neural Paraphrase Generation

1 code implementation COLING 2022 Xiaodong Gu, Zhaowei Zhang, Sang-Woo Lee, Kang Min Yoo, Jung-Woo Ha

While Transformers have had significant success in paragraph generation, they treat sentences as linear sequences of tokens and often neglect their hierarchical information.

Paraphrase Generation

Self-Generated In-Context Learning: Leveraging Auto-regressive Language Models as a Demonstration Generator

no code implementations16 Jun 2022 Hyuhng Joon Kim, Hyunsoo Cho, Junyeob Kim, Taeuk Kim, Kang Min Yoo, Sang-goo Lee

Large-scale pre-trained language models (PLMs) are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed demonstrations on a prompt without being explicitly tuned for the desired downstream task.

text-classification Text Classification +1

Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations

no code implementations25 May 2022 Kang Min Yoo, Junyeob Kim, Hyuhng Joon Kim, Hyunsoo Cho, Hwiyeol Jo, Sang-Woo Lee, Sang-goo Lee, Taeuk Kim

Despite recent explosion of interests in in-context learning, the underlying mechanism and the precise impact of the quality of demonstrations remain elusive.

Language Modelling

Mutual Information Divergence: A Unified Metric for Multimodal Generative Models

1 code implementation25 May 2022 Jin-Hwa Kim, Yunji Kim, Jiyoung Lee, Kang Min Yoo, Sang-Woo Lee

Based on a recent trend that multimodal generative evaluations exploit a vison-and-language pre-trained model, we propose the negative Gaussian cross-mutual information using the CLIP features as a unified metric, coined by Mutual Information Divergence (MID).

Hallucination Pair-wise Detection (1-ref) Hallucination Pair-wise Detection (4-ref) +5

Generating Information-Seeking Conversations from Unlabeled Documents

no code implementations25 May 2022 Gangwoo Kim, Sungdong Kim, Kang Min Yoo, Jaewoo Kang

In this paper, we introduce a novel framework, SIMSEEK, (Simulating information-Seeking conversation from unlabeled documents), and compare its two variants.

Conversational Search

Efficient Attribute Injection for Pretrained Language Models

no code implementations16 Sep 2021 Reinald Kim Amplayo, Kang Min Yoo, Sang-Woo Lee

Metadata attributes (e. g., user and product IDs from reviews) can be incorporated as additional inputs to neural-based NLP models, by modifying the architecture of the models, in order to improve their performance.

Self-Guided Contrastive Learning for BERT Sentence Representations

1 code implementation ACL 2021 Taeuk Kim, Kang Min Yoo, Sang-goo Lee

In this work, we propose a contrastive learning method that utilizes self-guidance for improving the quality of BERT sentence representations.

Contrastive Learning Data Augmentation +1

Reward Optimization for Neural Machine Translation with Learned Metrics

1 code implementation15 Apr 2021 Raphael Shu, Kang Min Yoo, Jung-Woo Ha

Results show that the reward optimization with BLEURT is able to increase the metric scores by a large margin, in contrast to limited gain when training with smoothed BLEU.

Machine Translation NMT +1

Variational Hierarchical Dialog Autoencoder for Dialog State Tracking Data Augmentation

1 code implementation EMNLP 2020 Kang Min Yoo, Hanbit Lee, Franck Dernoncourt, Trung Bui, Walter Chang, Sang-goo Lee

Recent works have shown that generative data augmentation, where synthetic samples generated from deep generative models complement the training dataset, benefit NLP tasks.

Data Augmentation dialog state tracking +4

Don't Just Scratch the Surface: Enhancing Word Representations for Korean with Hanja

3 code implementations IJCNLP 2019 Kang Min Yoo, Taeuk Kim, Sang-goo Lee

We propose a simple yet effective approach for improving Korean word representations using additional linguistic annotation (i. e. Hanja).

Cross-Lingual Transfer Headline generation +1

Data Augmentation for Spoken Language Understanding via Joint Variational Generation

no code implementations7 Sep 2018 Kang Min Yoo, Youhyun Shin, Sang-goo Lee

Data scarcity is one of the main obstacles of domain adaptation in spoken language understanding (SLU) due to the high cost of creating manually tagged SLU datasets.

Data Augmentation Domain Adaptation +1

Improving Visually Grounded Sentence Representations with Self-Attention

no code implementations2 Dec 2017 Kang Min Yoo, Youhyun Shin, Sang-goo Lee

Sentence representation models trained only on language could potentially suffer from the grounding problem.

Visual Grounding

Learning to Compose Task-Specific Tree Structures

1 code implementation10 Jul 2017 Jihun Choi, Kang Min Yoo, Sang-goo Lee

For years, recursive neural networks (RvNNs) have been shown to be suitable for representing text into fixed-length vectors and achieved good performance on several natural language processing tasks.

Natural Language Inference Sentiment Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.