Search Results for author: Sang-goo Lee

Found 36 papers, 17 papers with code

Aligning Language Models to Explicitly Handle Ambiguity

no code implementations18 Apr 2024 Hyuhng Joon Kim, Youna Kim, Cheonbok Park, Junyeob Kim, Choonghyun Park, Kang Min Yoo, Sang-goo Lee, Taeuk Kim

However, conversational agents built upon even the most recent large language models (LLMs) face challenges in processing ambiguous inputs, primarily due to the following two hurdles: (1) LLMs are not directly trained to handle inputs that are too ambiguous to be properly managed; (2) the degree of ambiguity in an input can vary according to the intrinsic knowledge of the LLMs, which is difficult to investigate.

Question Answering

Proxy-based Item Representation for Attribute and Context-aware Recommendation

1 code implementation11 Dec 2023 Jinseok Seol, Minseok Gang, Sang-goo Lee, Jaehui Park

Additionally, that the proxy embeddings are shared across all items allows the infrequent items to borrow training signals of frequent items in a unified model structure and end-to-end manner.

 Ranked #1 on Recommendation Systems on Amazon Beauty (using extra training data)

Attribute Recommendation Systems

Universal Domain Adaptation for Robust Handling of Distributional Shifts in NLP

1 code implementation23 Oct 2023 Hyuhng Joon Kim, Hyunsoo Cho, Sang-Woo Lee, Junyeob Kim, Choonghyun Park, Sang-goo Lee, Kang Min Yoo, Taeuk Kim

When deploying machine learning systems to the wild, it is highly desirable for them to effectively leverage prior knowledge to the unfamiliar domain while also firing alarms to anomalous inputs.

Universal Domain Adaptation

CELDA: Leveraging Black-box Language Model as Enhanced Classifier without Labels

no code implementations5 Jun 2023 Hyunsoo Cho, Youna Kim, Sang-goo Lee

Additionally, our proposed methodology can be applied universally to any LM and has the potential to scale to larger models, making it a more viable option for utilizing large LMs.

Language Modelling text-classification +1

Probing Out-of-Distribution Robustness of Language Models with Parameter-Efficient Transfer Learning

no code implementations27 Jan 2023 Hyunsoo Cho, Choonghyun Park, Junyeop Kim, Hyuhng Joon Kim, Kang Min Yoo, Sang-goo Lee

As the size of the pre-trained language model (PLM) continues to increase, numerous parameter-efficient transfer learning methods have been proposed recently to compensate for the tremendous cost of fine-tuning.

Language Modelling Transfer Learning

Enhancing Out-of-Distribution Detection in Natural Language Understanding via Implicit Layer Ensemble

1 code implementation20 Oct 2022 Hyunsoo Cho, Choonghyun Park, Jaewook Kang, Kang Min Yoo, Taeuk Kim, Sang-goo Lee

Out-of-distribution (OOD) detection aims to discern outliers from the intended data distribution, which is crucial to maintaining high reliability and a good user experience.

Contrastive Learning intent-classification +5

Self-Generated In-Context Learning: Leveraging Auto-regressive Language Models as a Demonstration Generator

no code implementations16 Jun 2022 Hyuhng Joon Kim, Hyunsoo Cho, Junyeob Kim, Taeuk Kim, Kang Min Yoo, Sang-goo Lee

Large-scale pre-trained language models (PLMs) are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed demonstrations on a prompt without being explicitly tuned for the desired downstream task.

In-Context Learning text-classification +2

Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations

no code implementations25 May 2022 Kang Min Yoo, Junyeob Kim, Hyuhng Joon Kim, Hyunsoo Cho, Hwiyeol Jo, Sang-Woo Lee, Sang-goo Lee, Taeuk Kim

Despite recent explosion of interests in in-context learning, the underlying mechanism and the precise impact of the quality of demonstrations remain elusive.

In-Context Learning Language Modelling

Exploiting Session Information in BERT-based Session-aware Sequential Recommendation

1 code implementation22 Apr 2022 Jinseok Seol, Youngrok Ko, Sang-goo Lee

In recommendation systems, utilizing the user interaction history as sequential information has resulted in great performance improvement.

Sequential Recommendation

False Negative Distillation and Contrastive Learning for Personalized Outfit Recommendation

no code implementations13 Oct 2021 Seongjae Kim, Jinseok Seol, Holim Lim, Sang-goo Lee

The first challenge is that outfit recommendation often requires a complex and large model that utilizes visual information, incurring huge memory and time costs.

Contrastive Learning Data Augmentation +3

Self-Guided Contrastive Learning for BERT Sentence Representations

1 code implementation ACL 2021 Taeuk Kim, Kang Min Yoo, Sang-goo Lee

In this work, we propose a contrastive learning method that utilizes self-guidance for improving the quality of BERT sentence representations.

Contrastive Learning Data Augmentation +2

Masked Contrastive Learning for Anomaly Detection

1 code implementation18 May 2021 Hyunsoo Cho, Jinseok Seol, Sang-goo Lee

However, the primary objective of contrastive learning is to learn task-agnostic features without any labels, which is not entirely suited to discern anomalies.

Anomaly Detection Contrastive Learning +1

Contrastive Learning for Unsupervised Image-to-Image Translation

no code implementations7 May 2021 Hanbit Lee, Jinseok Seol, Sang-goo Lee

Image-to-image translation aims to learn a mapping between different groups of visually distinguishable images.

Contrastive Learning Translation +1

Semantics-Preserving Adversarial Training

no code implementations23 Sep 2020 Wonseok Lee, Hanbit Lee, Sang-goo Lee

Adversarial training is a defense technique that improves adversarial robustness of a deep neural network (DNN) by including adversarial examples in the training data.

Adversarial Robustness

IDS at SemEval-2020 Task 10: Does Pre-trained Language Model Know What to Emphasize?

no code implementations SEMEVAL 2020 Jaeyoul Shin, Taeuk Kim, Sang-goo Lee

We propose a novel method that enables us to determine words that deserve to be emphasized from written text in visual media, relying only on the information from the self-attention distributions of pre-trained language models (PLMs).

Language Modelling

Multilingual Chart-based Constituency Parse Extraction from Pre-trained Language Models

1 code implementation Findings (EMNLP) 2021 Taeuk Kim, Bowen Li, Sang-goo Lee

As it has been unveiled that pre-trained language models (PLMs) are to some extent capable of recognizing syntactic concepts in natural language, much effort has been made to develop a method for extracting complete (binary) parses from PLMs without training separate parsers.

Constituency Parsing Cross-Lingual Transfer

Are Pre-trained Language Models Aware of Phrases? Simple but Strong Baselines for Grammar Induction

1 code implementation ICLR 2020 Taeuk Kim, Jihun Choi, Daniel Edmiston, Sang-goo Lee

With the recent success and popularity of pre-trained language models (LMs) in natural language processing, there has been a rise in efforts to understand their inner workings.

Variational Hierarchical Dialog Autoencoder for Dialog State Tracking Data Augmentation

1 code implementation EMNLP 2020 Kang Min Yoo, Hanbit Lee, Franck Dernoncourt, Trung Bui, Walter Chang, Sang-goo Lee

Recent works have shown that generative data augmentation, where synthetic samples generated from deep generative models complement the training dataset, benefit NLP tasks.

Data Augmentation dialog state tracking +4

Summary Level Training of Sentence Rewriting for Abstractive Summarization

no code implementations WS 2019 Sanghwan Bae, Taeuk Kim, Jihoon Kim, Sang-goo Lee

As an attempt to combine extractive and abstractive summarization, Sentence Rewriting models adopt the strategy of extracting salient sentences from a document first and then paraphrasing the selected ones to generate a summary.

Abstractive Text Summarization Extractive Text Summarization +3

Don't Just Scratch the Surface: Enhancing Word Representations for Korean with Hanja

3 code implementations IJCNLP 2019 Kang Min Yoo, Taeuk Kim, Sang-goo Lee

We propose a simple yet effective approach for improving Korean word representations using additional linguistic annotation (i. e. Hanja).

Cross-Lingual Transfer Headline Generation +1

SNU IDS at SemEval-2019 Task 3: Addressing Training-Test Class Distribution Mismatch in Conversational Classification

1 code implementation SEMEVAL 2019 Sanghwan Bae, Jihun Choi, Sang-goo Lee

We present several techniques to tackle the mismatch in class distributions between training and test data in the Contextual Emotion Detection task of SemEval 2019, by extending the existing methods for class imbalance problem.

Emotion Recognition in Conversation General Classification

SNU_IDS at SemEval-2019 Task 3: Addressing Training-Test Class Distribution Mismatch in Conversational Classification

1 code implementation6 Mar 2019 Sanghwan Bae, Jihun Choi, Sang-goo Lee

We present several techniques to tackle the mismatch in class distributions between training and test data in the Contextual Emotion Detection task of SemEval 2019, by extending the existing methods for class imbalance problem.

General Classification

Data Augmentation for Spoken Language Understanding via Joint Variational Generation

no code implementations7 Sep 2018 Kang Min Yoo, Youhyun Shin, Sang-goo Lee

Data scarcity is one of the main obstacles of domain adaptation in spoken language understanding (SLU) due to the high cost of creating manually tagged SLU datasets.

Data Augmentation Domain Adaptation +1

Dynamic Compositionality in Recursive Neural Networks with Structure-aware Tag Representations

2 code implementations7 Sep 2018 Taeuk Kim, Jihun Choi, Daniel Edmiston, Sanghwan Bae, Sang-goo Lee

Most existing recursive neural network (RvNN) architectures utilize only the structure of parse trees, ignoring syntactic tags which are provided as by-products of parsing.

Natural Language Inference Sentence +2

Element-wise Bilinear Interaction for Sentence Matching

no code implementations SEMEVAL 2018 Jihun Choi, Taeuk Kim, Sang-goo Lee

When we build a neural network model predicting the relationship between two sentences, the most general and intuitive approach is to use a Siamese architecture, where the sentence vectors obtained from a shared encoder is given as input to a classifier.

Natural Language Inference Paraphrase Identification +1

Improving Visually Grounded Sentence Representations with Self-Attention

no code implementations2 Dec 2017 Kang Min Yoo, Youhyun Shin, Sang-goo Lee

Sentence representation models trained only on language could potentially suffer from the grounding problem.

Sentence Visual Grounding

Style2Vec: Representation Learning for Fashion Items from Style Sets

1 code implementation14 Aug 2017 Hanbit Lee, Jinseok Seol, Sang-goo Lee

With the rapid growth of online fashion market, demand for effective fashion recommendation systems has never been greater.

Image Classification Recommendation Systems +3

A Syllable-based Technique for Word Embeddings of Korean Words

no code implementations WS 2017 Sanghyuk Choi, Taeuk Kim, Jinseok Seol, Sang-goo Lee

Word embedding has become a fundamental component to many NLP tasks such as named entity recognition and machine translation.

Machine Translation named-entity-recognition +4

Learning to Compose Task-Specific Tree Structures

1 code implementation10 Jul 2017 Jihun Choi, Kang Min Yoo, Sang-goo Lee

For years, recursive neural networks (RvNNs) have been shown to be suitable for representing text into fixed-length vectors and achieved good performance on several natural language processing tasks.

Natural Language Inference Sentiment Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.