no code implementations • 14 Nov 2023 • Nicholas E. Corrado, Josiah P. Hanna
We empirically evaluate PROPS on both continuous-action MuJoCo benchmark tasks as well as discrete-action tasks and demonstrate that (1) PROPS decreases sampling error throughout training and (2) improves the data efficiency of on-policy policy gradient algorithms.
no code implementations • 27 Oct 2023 • Nicholas E. Corrado, Yuxiao Qu, John U. Balis, Adam Labiosa, Josiah P. Hanna
In offline reinforcement learning (RL), an RL agent learns to solve a task using only a fixed dataset of previously collected data.
1 code implementation • 26 Oct 2023 • Nicholas E. Corrado, Josiah P. Hanna
Recently, data augmentation (DA) has emerged as a method for leveraging domain knowledge to inexpensively generate additional data in reinforcement learning (RL) tasks, often yielding substantial improvements in data efficiency.