no code implementations • 31 Jul 2024 • Taehyun Cho, Seungyub Han, Kyungjae Lee, Seokhun Ju, Dohyeong Kim, Jungwoo Lee
Distributional reinforcement learning improves performance by effectively capturing environmental stochasticity, but a comprehensive theoretical understanding of its effectiveness remains elusive.
Distributional Reinforcement Learning reinforcement-learning +1
1 code implementation • 1 Jul 2024 • Takyoung Kim, Kyungjae Lee, Young Rok Jang, Ji Yong Cho, Gangwoo Kim, Minseok Cho, Moontae Lee
Interactions with billion-scale large language models typically yield long-form responses due to their extensive parametric capacities, along with retrieval-augmented features.
1 code implementation • 9 Jun 2024 • Seungone Kim, Juyoung Suk, Ji Yong Cho, Shayne Longpre, Chaeeun Kim, Dongkeun Yoon, Guijin Son, Yejin Cho, Sheikh Shafayat, Jinheon Baek, Sue Hyun Park, Hyeonbin Hwang, Jinkyung Jo, Hyowon Cho, Haebin Shin, Seongyun Lee, Hanseok Oh, Noah Lee, Namgyu Ho, Se June Joo, Miyoung Ko, Yoonjoo Lee, Hyungjoo Chae, Jamin Shin, Joel Jang, Seonghyeon Ye, Bill Yuchen Lin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, Minjoon Seo
To overcome these limitations, we introduce the BiGGen Bench, a principled generation benchmark designed to thoroughly evaluate nine distinct capabilities of LMs across 77 diverse tasks.
no code implementations • 29 May 2024 • Dohyeong Kim, Taehyun Cho, Seungyub Han, Hojun Chung, Kyungjae Lee, Songhwai Oh
Furthermore, the proposed method has been evaluated on continuous control tasks and showed the best performance among other RCRL algorithms satisfying the constraints.
no code implementations • 22 May 2024 • Hajung Kim, Chanhwi Kim, Hoonick Lee, Kyochul Jang, Jiwoo Lee, Kyungjae Lee, Gangwoo Kim, Jaewoo Kang
Transforming natural language questions into SQL queries is crucial for precise data retrieval from electronic health record (EHR) databases.
1 code implementation • 2 May 2024 • Seungone Kim, Juyoung Suk, Shayne Longpre, Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, Minjoon Seo
Proprietary LMs such as GPT-4 are often employed to assess the quality of responses from various LMs.
no code implementations • 21 Mar 2024 • Kyungjae Lee, Dasol Hwang, Sunghyun Park, Youngsoo Jang, Moontae Lee
Despite the promise of RLHF in aligning LLMs with human preferences, it often leads to superficial alignment, prioritizing stylistic changes over improving downstream performance of LLMs.
no code implementations • NeurIPS 2023 • Taehyun Cho, Seungyub Han, Heesoo Lee, Kyungjae Lee, Jungwoo Lee
Distributional reinforcement learning algorithms have attempted to utilize estimated uncertainty for exploration, such as optimism in the face of uncertainty.
Distributional Reinforcement Learning reinforcement-learning +1
no code implementations • 24 Oct 2023 • Wookje Han, Jinsol Park, Kyungjae Lee
Information-seeking questions in long-form question answering (LFQA) often prove misleading due to ambiguity or false presupposition in the question.
no code implementations • 25 Sep 2023 • Joonhyung Lee, Sangbeom Park, Jeongeun Park, Kyungjae Lee, Sungjoon Choi
Particularly, we focus on two aspects of the place task: stability robustness and contextual reasonableness of object placements.
1 code implementation • 8 Aug 2023 • Sang-eun Han, Yeonseok Jeong, Seung-won Hwang, Kyungjae Lee
Our experiments show that our framework not only ensures monotonicity, but also outperforms state-of-the-art multi-source QA methods on Natural Questions.
no code implementations • 7 Jun 2023 • Kyungjae Lee, Sang-eun Han, Seung-won Hwang, Moontae Lee
This paper studies the problem of open-domain question answering, with the aim of answering a diverse range of questions leveraging knowledge resources.
no code implementations • 6 Apr 2023 • Yongho Song, Dahyun Lee, Myungha Jang, Seung-won Hwang, Kyungjae Lee, Dongha Lee, Jinyeong Yeo
The long-standing goal of dense retrievers in abtractive open-domain question answering (ODQA) tasks is to learn to capture evidence passages among relevant passages for any given query, such that the reader produce factually correct outputs from evidence passages.
2 code implementations • 7 Feb 2023 • Joel Jang, Seungone Kim, Seonghyeon Ye, Doyoung Kim, Lajanugen Logeswaran, Moontae Lee, Kyungjae Lee, Minjoon Seo
Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known as multitask-prompted fine-tuning (MT), have shown the capability to generalize to unseen tasks.
Ranked #9 on Question Answering on StoryCloze
1 code implementation • NeurIPS 2023 • Dohyeong Kim, Kyungjae Lee, Songhwai Oh
In safety-critical robotic tasks, potential failures must be reduced, and multiple constraints must be met, such as avoiding collisions, limiting energy consumption, and maintaining balance.
Distributional Reinforcement Learning reinforcement-learning +3
no code implementations • CVPR 2023 • MyeongAh Cho, Minjung Kim, Sangwon Hwang, Chaewon Park, Kyungjae Lee, Sangyoun Lee
Furthermore, as the relationship between context and motion is important in order to identify the anomalies in complex and diverse scenes, we propose a Context--Motion Interrelation Module (CoMo), which models the relationship between the appearance of the surroundings and motion, rather than utilizing only temporal dependencies or motion information.
1 code implementation • Findings (ACL) 2022 • Kyungjae Lee, Wookje Han, Seung-won Hwang, Hwaran Lee, Joonsuk Park, Sang-Woo Lee
To this end, we first propose a novel task--Continuously-updated QA (CuQA)--in which multiple large-scale updates are made to LMs, and the performance is measured with respect to the success in adding and updating knowledge while retaining existing knowledge.
1 code implementation • 21 Mar 2022 • Junbum Cha, Kyungjae Lee, Sungrae Park, Sanghyuk Chun
Domain generalization (DG) aims to learn a generalized model to an unseen target domain using only limited source domains.
Ranked #3 on Domain Generalization on TerraIncognita
no code implementations • ICLR 2022 • Sung Woo Park, Kyungjae Lee, Junseok Kwon
We propose a novel probabilistic framework for modeling stochastic dynamics with the rigorous use of stochastic optimal control theory.
no code implementations • 29 Sep 2021 • Tae Hyun Cho, Sungyeob Han, Heesoo Lee, Kyungjae Lee, Jungwoo Lee
Distributional reinforcement learning aims to learn distribution of return under stochastic environments.
no code implementations • 27 Sep 2021 • Sangbeom Park, Yoonbyung Chai, Sunghyun Park, Jeongeun Park, Kyungjae Lee, Sungjoon Choi
In this paper, we present a semi-autonomous teleoperation framework for a pick-and-place task using an RGB-D sensor.
no code implementations • ACL 2021 • Kyungjae Lee, Seung-won Hwang, Sang-eun Han, Dohyeon Lee
This paper studies the bias problem of multi-hop question answering models, of answering correctly without correct reasoning.
no code implementations • EACL 2021 • Kyungho Kim, Kyungjae Lee, Seung-won Hwang, Young-In Song, SeungWook Lee
This paper studies the problem of generatinglikely queries for multimodal documents withimages.
4 code implementations • NeurIPS 2021 • Junbum Cha, Sanghyuk Chun, Kyungjae Lee, Han-Cheol Cho, Seunghyun Park, Yunsung Lee, Sungrae Park
Domain generalization (DG) methods aim to achieve generalizability to an unseen target domain by using only training data from the source domains.
Ranked #25 on Domain Generalization on TerraIncognita
no code implementations • NeurIPS 2020 • Kyungjae Lee, Hongjun Yang, Sungbin Lim, Songhwai Oh
In simulation, the proposed estimator shows favorable performance compared to existing robust estimators for various $p$ values and, for MAB problems, the proposed perturbation strategy outperforms existing exploration methods.
no code implementations • 2 Mar 2020 • MyeongAh Cho, Taeoh Kim, Ig-Jae Kim, Kyungjae Lee, Sangyoun Lee
Due to the lack of databases, HFR methods usually exploit the pre-trained features on a large-scale visual database that contain general facial information.
no code implementations • 7 Jan 2020 • Joosung Lee, Sangwon Hwang, Kyungjae Lee, Woo Jin Kim, Junhyeop Lee, Tae-young Chung, Sangyoun Lee
Visual odometry is an essential key for a localization module in SLAM systems.
no code implementations • IJCNLP 2019 • Kyungjae Lee, Sunghyun Park, Hojae Han, Jinyoung Yeo, Seung-won Hwang, Juho Lee
This paper studies the problem of supporting question answering in a new language with limited training resources.
2 code implementations • TACL 2019 • Jihyeok Kim, Reinald Kim Amplayo, Kyungjae Lee, Sua Sung, Minji Seo, Seung-won Hwang
The performance of text classification has improved tremendously using intelligently engineered neural-based models, especially those injecting categorical metadata as additional information, e. g., using user/product information for sentiment classification.
Ranked #4 on Sentiment Analysis on User and product information (Yelp 2013 (Acc) metric)
no code implementations • 31 Jan 2019 • Kyungjae Lee, Sungyub Kim, Sungbin Lim, Sungjoon Choi, Songhwai Oh
By controlling the entropic index, we can generate various types of entropy, including the SG entropy, and a different entropy results in a different class of the optimal policy in Tsallis MDPs.
no code implementations • 27 Sep 2018 • Sungjoon Choi, Sanghoon Hong, Kyungjae Lee, Sungbin Lim
To this end, we present a novel framework referred to here as ChoiceNet that can robustly infer the target distribution in the presence of inconsistent data.
1 code implementation • 14 Jun 2018 • Reinald Kim Amplayo, Kyungjae Lee, Jinyeong Yeo, Seung-won Hwang
We are the first to use translations as domain-free contexts for sentence classification.
Ranked #7 on Text Classification on TREC-6
no code implementations • NeurIPS 2018 • Kyungjae Lee, Sungjoon Choi, Songhwai Oh
Third, we propose a maximum causal Tsallis entropy imitation learning (MCTEIL) algorithm with a sparse mixture density network (sparse MDN) by modeling mixture weights using a sparsemax distribution.
1 code implementation • CVPR 2020 • Sungjoon Choi, Sanghoon Hong, Kyungjae Lee, Sungbin Lim
In this paper, we focus on weakly supervised learning with noisy training data for both classification and regression problems. We assume that the training outputs are collected from a mixture of a target and correlated noise distributions. Our proposed method simultaneously estimates the target distribution and the quality of each data which is defined as the correlation between the target and data generating distributions. The cornerstone of the proposed method is a Cholesky Block that enables modeling dependencies among mixture distributions in a differentiable manner where we maintain the distribution over the network weights. We first provide illustrative examples in both regression and classification tasks to show the effectiveness of the proposed method. Then, the proposed method is extensively evaluated in a number of experiments where we show that it constantly shows comparable or superior performances compared to existing baseline methods in the handling of noisy data.
no code implementations • 19 Sep 2017 • Kyungjae Lee, Sungjoon Choi, Songhwai Oh
The proposed sparse MDP is compared to soft MDPs which utilize causal entropy regularization.
1 code implementation • 3 Sep 2017 • Sungjoon Choi, Kyungjae Lee, Sungbin Lim, Songhwai Oh
The proposed uncertainty-aware learning from demonstration method outperforms other compared methods in terms of safety using a complex real-world driving dataset.
no code implementations • 12 Aug 2016 • Sungjoon Choi, Kyungjae Lee, Andy Park, Songhwai Oh
The performance of KDMRL is extensively evaluated in two sets of experiments: grid world and track driving experiments.