Search Results for author: Byung-Jun Lee

Found 7 papers, 2 papers with code

Batch Reinforcement Learning with Hyperparameter Gradients

no code implementations ICML 2020 Byung-Jun Lee, Jongmin Lee, Peter Vrancx, Dongho Kim, Kee-Eung Kim

We consider the batch reinforcement learning problem where the agent needs to learn only from a fixed batch of data, without further interaction with the environment.

Continuous Control reinforcement-learning +1

Offline Imitation Learning by Controlling the Effective Planning Horizon

no code implementations18 Jan 2024 Hee-Jun Ahn, Seong-Woong Shim, Byung-Jun Lee

In offline imitation learning (IL), we generally assume only a handful of expert trajectories and a supplementary offline dataset from suboptimal behaviors to learn the expert policy.

Imitation Learning

Local Metric Learning for Off-Policy Evaluation in Contextual Bandits with Continuous Actions

1 code implementation24 Oct 2022 Haanvid Lee, Jongmin Lee, Yunseon Choi, Wonseok Jeon, Byung-Jun Lee, Yung-Kyun Noh, Kee-Eung Kim

We consider local kernel metric learning for off-policy evaluation (OPE) of deterministic policies in contextual bandits with continuous action spaces.

Metric Learning Multi-Armed Bandits +1

OptiDICE: Offline Policy Optimization via Stationary Distribution Correction Estimation

1 code implementation21 Jun 2021 Jongmin Lee, Wonseok Jeon, Byung-Jun Lee, Joelle Pineau, Kee-Eung Kim

We consider the offline reinforcement learning (RL) setting where the agent aims to optimize the policy solely from the data without further environment interactions.

Offline RL Reinforcement Learning (RL)

Representation Balancing Offline Model-based Reinforcement Learning

no code implementations ICLR 2021 Byung-Jun Lee, Jongmin Lee, Kee-Eung Kim

We present a new objective for model learning motivated by recent advances in the estimation of stationary distribution corrections.

Model-based Reinforcement Learning Offline RL +2

Cannot find the paper you are looking for? You can Submit a new open access paper.