no code implementations • 2 Aug 2024 • Youna Kim, Hyuhng Joon Kim, Cheonbok Park, Choonghyun Park, Hyunsoo Cho, Junyeob Kim, Kang Min Yoo, Sang-goo Lee, Taeuk Kim
When using large language models (LLMs) in knowledge-intensive tasks, such as open-domain question answering, external context can bridge a gap between external knowledge and LLM's parametric knowledge.
no code implementations • 24 Jun 2024 • Choonghyun Park, Hyuhng Joon Kim, Junyeob Kim, Youna Kim, Taeuk Kim, Hyunsoo Cho, Hwiyeol Jo, Sang-goo Lee, Kang Min Yoo
Based on the findings, we further train the classifier with the dataset augmented by FAILOpt prompt.
no code implementations • 18 Apr 2024 • Hyuhng Joon Kim, Youna Kim, Cheonbok Park, Junyeob Kim, Choonghyun Park, Kang Min Yoo, Sang-goo Lee, Taeuk Kim
In interactions between users and language model agents, user utterances frequently exhibit ellipsis (omission of words or phrases) or imprecision (lack of exactness) to prioritize efficiency.
1 code implementation • 11 Dec 2023 • Jinseok Seol, Minseok Gang, Sang-goo Lee, Jaehui Park
Additionally, that the proxy embeddings are shared across all items allows the infrequent items to borrow training signals of frequent items in a unified model structure and end-to-end manner.
Ranked #1 on Recommendation Systems on Amazon Fashion (using extra training data)
1 code implementation • 23 Oct 2023 • Hyuhng Joon Kim, Hyunsoo Cho, Sang-Woo Lee, Junyeob Kim, Choonghyun Park, Sang-goo Lee, Kang Min Yoo, Taeuk Kim
When deploying machine learning systems to the wild, it is highly desirable for them to effectively leverage prior knowledge to the unfamiliar domain while also firing alarms to anomalous inputs.
no code implementations • 5 Jun 2023 • Hyunsoo Cho, Youna Kim, Sang-goo Lee
Additionally, our proposed methodology can be applied universally to any LM and has the potential to scale to larger models, making it a more viable option for utilizing large LMs.
no code implementations • 27 Jan 2023 • Hyunsoo Cho, Choonghyun Park, Junyeop Kim, Hyuhng Joon Kim, Kang Min Yoo, Sang-goo Lee
As the size of the pre-trained language model (PLM) continues to increase, numerous parameter-efficient transfer learning methods have been proposed recently to compensate for the tremendous cost of fine-tuning.
no code implementations • 21 Dec 2022 • Hyunsoo Cho, Hyuhng Joon Kim, Junyeob Kim, Sang-Woo Lee, Sang-goo Lee, Kang Min Yoo, Taeuk Kim
Through in-context learning (ICL), large-scale language models are effective few-shot learners without additional model fine-tuning.
1 code implementation • 20 Oct 2022 • Hyunsoo Cho, Choonghyun Park, Jaewook Kang, Kang Min Yoo, Taeuk Kim, Sang-goo Lee
Out-of-distribution (OOD) detection aims to discern outliers from the intended data distribution, which is crucial to maintaining high reliability and a good user experience.
no code implementations • 16 Jun 2022 • Hyuhng Joon Kim, Hyunsoo Cho, Junyeob Kim, Taeuk Kim, Kang Min Yoo, Sang-goo Lee
Large-scale pre-trained language models (PLMs) are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed demonstrations on a prompt without being explicitly tuned for the desired downstream task.
no code implementations • 25 May 2022 • Kang Min Yoo, Junyeob Kim, Hyuhng Joon Kim, Hyunsoo Cho, Hwiyeol Jo, Sang-Woo Lee, Sang-goo Lee, Taeuk Kim
Despite recent explosion of interests in in-context learning, the underlying mechanism and the precise impact of the quality of demonstrations remain elusive.
1 code implementation • 22 Apr 2022 • Jinseok Seol, Youngrok Ko, Sang-goo Lee
In recommendation systems, utilizing the user interaction history as sequential information has resulted in great performance improvement.
no code implementations • 11 Mar 2022 • Jinseok Seol, Seongjae Kim, Sungchan Park, Holim Lim, Hyunsoo Na, EunYoung Park, Dohee Jung, Soyoung Park, Kangwoo Lee, Sang-goo Lee
The rapid growth of the online fashion market brought demands for innovative fashion services and commerce platforms.
no code implementations • 13 Oct 2021 • Seongjae Kim, Jinseok Seol, Holim Lim, Sang-goo Lee
The first challenge is that outfit recommendation often requires a complex and large model that utilizes visual information, incurring huge memory and time costs.
1 code implementation • ACL 2021 • Taeuk Kim, Kang Min Yoo, Sang-goo Lee
In this work, we propose a contrastive learning method that utilizes self-guidance for improving the quality of BERT sentence representations.
1 code implementation • 18 May 2021 • Hyunsoo Cho, Jinseok Seol, Sang-goo Lee
However, the primary objective of contrastive learning is to learn task-agnostic features without any labels, which is not entirely suited to discern anomalies.
no code implementations • 7 May 2021 • Hanbit Lee, Jinseok Seol, Sang-goo Lee
Image-to-image translation aims to learn a mapping between different groups of visually distinguishable images.
no code implementations • 23 Sep 2020 • Wonseok Lee, Hanbit Lee, Sang-goo Lee
Adversarial training is a defense technique that improves adversarial robustness of a deep neural network (DNN) by including adversarial examples in the training data.
no code implementations • SEMEVAL 2020 • Jaeyoul Shin, Taeuk Kim, Sang-goo Lee
We propose a novel method that enables us to determine words that deserve to be emphasized from written text in visual media, relying only on the information from the self-attention distributions of pre-trained language models (PLMs).
1 code implementation • Findings (EMNLP) 2021 • Taeuk Kim, Bowen Li, Sang-goo Lee
As it has been unveiled that pre-trained language models (PLMs) are to some extent capable of recognizing syntactic concepts in natural language, much effort has been made to develop a method for extracting complete (binary) parses from PLMs without training separate parsers.
1 code implementation • ICLR 2020 • Taeuk Kim, Jihun Choi, Daniel Edmiston, Sang-goo Lee
With the recent success and popularity of pre-trained language models (LMs) in natural language processing, there has been a rise in efforts to understand their inner workings.
1 code implementation • EMNLP 2020 • Kang Min Yoo, Hanbit Lee, Franck Dernoncourt, Trung Bui, Walter Chang, Sang-goo Lee
Recent works have shown that generative data augmentation, where synthetic samples generated from deep generative models complement the training dataset, benefit NLP tasks.
no code implementations • WS 2019 • Sanghwan Bae, Taeuk Kim, Jihoon Kim, Sang-goo Lee
As an attempt to combine extractive and abstractive summarization, Sentence Rewriting models adopt the strategy of extracting salient sentences from a document first and then paraphrasing the selected ones to generate a summary.
Ranked #5 on Extractive Text Summarization on CNN / Daily Mail
Abstractive Text Summarization Extractive Text Summarization +3
3 code implementations • IJCNLP 2019 • Kang Min Yoo, Taeuk Kim, Sang-goo Lee
We propose a simple yet effective approach for improving Korean word representations using additional linguistic annotation (i. e. Hanja).
no code implementations • ACL 2019 • Jihun Choi, Taeuk Kim, Sang-goo Lee
We present a latent variable model for predicting the relationship between a pair of text sequences.
1 code implementation • SEMEVAL 2019 • Sanghwan Bae, Jihun Choi, Sang-goo Lee
We present several techniques to tackle the mismatch in class distributions between training and test data in the Contextual Emotion Detection task of SemEval 2019, by extending the existing methods for class imbalance problem.
Ranked #6 on Emotion Recognition in Conversation on EC
1 code implementation • 6 Mar 2019 • Sanghwan Bae, Jihun Choi, Sang-goo Lee
We present several techniques to tackle the mismatch in class distributions between training and test data in the Contextual Emotion Detection task of SemEval 2019, by extending the existing methods for class imbalance problem.
no code implementations • 7 Sep 2018 • Jihun Choi, Taeuk Kim, Sang-goo Lee
We propose a method of stacking multiple long short-term memory (LSTM) layers for modeling sentences.
Ranked #10 on Sentiment Analysis on SST-5 Fine-grained classification
no code implementations • 7 Sep 2018 • Kang Min Yoo, Youhyun Shin, Sang-goo Lee
Data scarcity is one of the main obstacles of domain adaptation in spoken language understanding (SLU) due to the high cost of creating manually tagged SLU datasets.
2 code implementations • 7 Sep 2018 • Taeuk Kim, Jihun Choi, Daniel Edmiston, Sanghwan Bae, Sang-goo Lee
Most existing recursive neural network (RvNN) architectures utilize only the structure of parse trees, ignoring syntactic tags which are provided as by-products of parsing.
no code implementations • SEMEVAL 2018 • Jihun Choi, Taeuk Kim, Sang-goo Lee
When we build a neural network model predicting the relationship between two sentences, the most general and intuitive approach is to use a Siamese architecture, where the sentence vectors obtained from a shared encoder is given as input to a classifier.
1 code implementation • SEMEVAL 2018 • Taeuk Kim, Jihun Choi, Sang-goo Lee
We present a novel neural architecture for the Argument Reasoning Comprehension task of SemEval 2018.
1 code implementation • SEMEVAL 2018 • Taeuk Kim, Jihun Choi, Sang-goo Lee
We present a novel neural architecture for the Argument Reasoning Comprehension task of SemEval 2018.
no code implementations • 2 Dec 2017 • Kang Min Yoo, Youhyun Shin, Sang-goo Lee
Sentence representation models trained only on language could potentially suffer from the grounding problem.
no code implementations • WS 2017 • Youhyun Shin, Sang-goo Lee
Recently, there has been increased interest in utilizing characters or subwords for natural language processing (NLP) tasks.
1 code implementation • 14 Aug 2017 • Hanbit Lee, Jinseok Seol, Sang-goo Lee
With the rapid growth of online fashion market, demand for effective fashion recommendation systems has never been greater.
no code implementations • WS 2017 • Sanghyuk Choi, Taeuk Kim, Jinseok Seol, Sang-goo Lee
Word embedding has become a fundamental component to many NLP tasks such as named entity recognition and machine translation.
1 code implementation • 10 Jul 2017 • Jihun Choi, Kang Min Yoo, Sang-goo Lee
For years, recursive neural networks (RvNNs) have been shown to be suitable for representing text into fixed-length vectors and achieved good performance on several natural language processing tasks.
Ranked #58 on Natural Language Inference on SNLI