no code implementations • 1 Jun 2022 • Wonjoon Goo, Scott Niekum
In this work, we argue that it is not only viable but beneficial to explicitly model the behavior policy for offline RL because the constraint can be realized in a stable way with the trained model.
no code implementations • 7 Feb 2022 • Harshit Sikchi, Akanksha Saran, Wonjoon Goo, Scott Niekum
We propose a new framework for imitation learning -- treating imitation as a two-player ranking-based game between a policy and a reward.
no code implementations • 5 Oct 2021 • Wonjoon Goo, Scott Niekum
The goal of offline reinforcement learning (RL) is to find an optimal policy given prerecorded trajectories.
1 code implementation • 8 Mar 2021 • Farzan Memarian, Wonjoon Goo, Rudolf Lioutikov, Scott Niekum, Ufuk Topcu
We introduce Self-supervised Online Reward Shaping (SORS), which aims to improve the sample efficiency of any RL algorithm in sparse-reward environments by automatically densifying rewards.
1 code implementation • 9 Feb 2020 • Wonjoon Goo, Scott Niekum
A central goal of meta-learning is to find a learning rule that enables fast adaptation across a set of tasks, by learning the appropriate inductive bias for that set.
2 code implementations • 9 Jul 2019 • Daniel S. Brown, Wonjoon Goo, Scott Niekum
The performance of imitation learning is typically upper-bounded by the performance of the demonstrator.
3 code implementations • 12 Apr 2019 • Daniel S. Brown, Wonjoon Goo, Prabhat Nagarajan, Scott Niekum
A critical flaw of existing inverse reinforcement learning (IRL) methods is their inability to significantly outperform the demonstrator.
1 code implementation • 29 Jun 2018 • Wonjoon Goo, Scott Niekum
Due to burdensome data requirements, learning from demonstration often falls short of its promise to allow users to quickly and naturally program robots.