Search Results for author: KyungMin Kim

Found 13 papers, 5 papers with code

Make the Pertinent Salient: Task-Relevant Reconstruction for Visual Control with Distractions

no code implementations13 Oct 2024 KyungMin Kim, JB Lanier, Pierre Baldi, Charless Fowlkes, Roy Fox

Training in the presence of visual distractions is particularly difficult due to the high variation they introduce to representation learning.

Model-based Reinforcement Learning Representation Learning

Realizable Continuous-Space Shields for Safe Reinforcement Learning

no code implementations2 Oct 2024 KyungMin Kim, Davide Corsi, Andoni Rodriguez, JB Lanier, Benjami Parellada, Pierre Baldi, Cesar Sanchez, Roy Fox

For real-world robotic domains, it is essential to define safety specifications over continuous state and action spaces to accurately account for system dynamics and compute new actions that minimally deviate from the agent's original decision.

Deep Reinforcement Learning reinforcement-learning +1

Rethinking Data Augmentation for Robust LiDAR Semantic Segmentation in Adverse Weather

2 code implementations2 Jul 2024 Junsung Park, KyungMin Kim, Hyunjung Shim

Motivated by this issue, we identified key factors of adverse weather and conducted a toy experiment to pinpoint the main causes of performance degradation: (1) Geometric perturbation due to refraction caused by fog or droplets in the air and (2) Point drop due to energy absorption and occlusions.

Data Augmentation LIDAR Semantic Segmentation +2

Reinforcement Learning from Delayed Observations via World Models

1 code implementation18 Mar 2024 Armin Karamzade, KyungMin Kim, Montek Kalsi, Roy Fox

In standard reinforcement learning settings, agents typically assume immediate feedback about the effects of their actions after taking them.

continuous-control Continuous Control +2

Persuasion in Veto Bargaining

no code implementations19 Oct 2023 Jenny S Kim, KyungMin Kim, Richard Van Weelden

We consider the classic veto bargaining model but allow the agenda setter to engage in persuasion to convince the veto player to approve her proposal.

Selective Perception: Optimizing State Descriptions with Reinforcement Learning for Language Model Actors

no code implementations21 Jul 2023 Kolby Nottingham, Yasaman Razeghi, KyungMin Kim, JB Lanier, Pierre Baldi, Roy Fox, Sameer Singh

Large language models (LLMs) are being applied as actors for sequential decision making tasks in domains such as robotics and games, utilizing their general world knowledge and planning abilities.

Decision Making Language Modelling +3

Selective Generation for Controllable Language Models

1 code implementation18 Jul 2023 Minjae Lee, KyungMin Kim, Taesoo Kim, Sangdon Park

$\texttt{SGen}^{\texttt{Sup}}$, a direct modification of the selective prediction, is a supervised learning algorithm which exploits entailment-labeled data, annotated by humans.

Conformal Prediction Hallucination +3

Relation-Aware Language-Graph Transformer for Question Answering

1 code implementation2 Dec 2022 Jinyoung Park, Hyeong Kyu Choi, Juyeon Ko, Hyeonjin Park, Ji-Hoon Kim, Jisu Jeong, KyungMin Kim, Hyunwoo J. Kim

To address these issues, we propose Question Answering Transformer (QAT), which is designed to jointly reason over language and graphs with respect to entity relations in a unified manner.

MedQA Question Answering +1

Meta-node: A Concise Approach to Effectively Learn Complex Relationships in Heterogeneous Graphs

no code implementations26 Oct 2022 Jiwoong Park, Jisu Jeong, KyungMin Kim, Jin Young Choi

To tackle this challenge, we propose a novel concept of meta-node for message passing that can learn enriched relational knowledge from complex heterogeneous graphs without any meta-paths and meta-graphs by explicitly modeling the relations among the same type of nodes.

Contrastive Learning Graph Learning +1

An Investigation on Hardware-Aware Vision Transformer Scaling

no code implementations29 Sep 2021 Chaojian Li, KyungMin Kim, Bichen Wu, Peizhao Zhang, Hang Zhang, Xiaoliang Dai, Peter Vajda, Yingyan Lin

In particular, when transferred to PiT, our scaling strategies lead to a boosted ImageNet top-1 accuracy of from $74. 6\%$ to $76. 7\%$ ($\uparrow2. 1\%$) under the same 0. 7G FLOPs; and when transferred to the COCO object detection task, the average precision is boosted by $\uparrow0. 7\%$ under a similar throughput on a V100 GPU.

Image Classification object-detection +2

Winning the CVPR'2021 Kinetics-GEBD Challenge: Contrastive Learning Approach

1 code implementation22 Jun 2021 Hyolim Kang, Jinwoo Kim, KyungMin Kim, Taehyun Kim, Seon Joo Kim

Generic Event Boundary Detection (GEBD) is a newly introduced task that aims to detect "general" event boundaries that correspond to natural human perception.

Boundary Detection Contrastive Learning +1

Keeping the Listener Engaged: a Dynamic Model of Bayesian Persuasion

no code implementations16 Mar 2020 Yeon-Koo Che, KyungMin Kim, Konrad Mierendorff

We consider a dynamic model of Bayesian persuasion in which information takes time and is costly for the sender to generate and for the receiver to process, and neither player can commit to their future actions.

Cannot find the paper you are looking for? You can Submit a new open access paper.