no code implementations • 14 Oct 2024 • James R. Han, Hugues Thomas, Jian Zhang, Nicholas Rhinehart, Timothy D. Barfoot
In simulation, we show that DR-MPC substantially outperforms prior work, including traditional DRL and residual DRL models.
no code implementations • 31 Jan 2024 • Jiezhi Yang, Khushi Desai, Charles Packer, Harshil Bhatia, Nicholas Rhinehart, Rowan Mcallister, Joseph Gonzalez
We propose CARFF, a method for predicting future 3D scenes given past observations.
no code implementations • NeurIPS 2021 • Nicholas Rhinehart, Jenny Wang, Glen Berseth, John D. Co-Reyes, Danijar Hafner, Chelsea Finn, Sergey Levine
We study this question in dynamic partially-observed environments, and argue that a compact and general learning objective is to minimize the entropy of the agent's state visitation estimated using a latent state-space model.
1 code implementation • ICML Workshop URL 2021 • Arnaud Fickinger, Natasha Jaques, Samyak Parajuli, Michael Chang, Nicholas Rhinehart, Glen Berseth, Stuart Russell, Sergey Levine
Unsupervised reinforcement learning (RL) studies how to leverage environment statistics to learn useful behaviors without the cost of reward engineering.
no code implementations • ICML Workshop URL 2021 • Nicholas Rhinehart, Jenny Wang, Glen Berseth, John D Co-Reyes, Danijar Hafner, Chelsea Finn, Sergey Levine
We study this question in dynamic partially-observed environments, and argue that a compact and general learning objective is to minimize the entropy of the agent's state visitation estimated using a latent state-space model.
1 code implementation • 21 Apr 2021 • Nicholas Rhinehart, Jeff He, Charles Packer, Matthew A. Wright, Rowan Mcallister, Joseph E. Gonzalez, Sergey Levine
Humans have a remarkable ability to make decisions by accurately reasoning about future events, including the future behaviors and states of mind of other agents.
no code implementations • 12 Apr 2021 • Dhruv Shah, Benjamin Eysenbach, Gregory Kahn, Nicholas Rhinehart, Sergey Levine
We describe a robotic learning system for autonomous exploration and navigation in diverse, open-world environments.
no code implementations • 17 Dec 2020 • Dhruv Shah, Benjamin Eysenbach, Gregory Kahn, Nicholas Rhinehart, Sergey Levine
We propose a learning-based navigation system for reaching visually indicated goals and demonstrate this system on a real mobile robot platform.
no code implementations • ICLR 2021 • Avi Singh, Huihan Liu, Gaoyue Zhou, Albert Yu, Nicholas Rhinehart, Sergey Levine
Reinforcement learning provides a general framework for flexible decision making and control, but requires extensive data collection for each new task that an agent needs to learn.
no code implementations • ICLR 2021 • Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg
Safe exploration presents a major challenge in reinforcement learning (RL): when active data collection requires deploying partially trained policies, we must ensure that these policies avoid catastrophically unsafe regions, while still enabling trial and error learning.
2 code implementations • ICML 2020 • Angelos Filos, Panagiotis Tigas, Rowan Mcallister, Nicholas Rhinehart, Sergey Levine, Yarin Gal
Out-of-training-distribution (OOD) scenarios are a common challenge of learning agents at deployment, typically leading to arbitrary deductions and poorly-informed decisions.
no code implementations • 18 Mar 2020 • Xinshuo Weng, Jianren Wang, Sergey Levine, Kris Kitani, Nicholas Rhinehart
Through experiments on a robotic manipulation dataset and two driving datasets, we show that SPFNet is effective for the SPF task, our forecast-then-detect pipeline outperforms the detect-then-forecast approaches to which we compared, and that pose forecasting performance improves with the addition of unlabeled data.
1 code implementation • ICLR 2021 • Glen Berseth, Daniel Geng, Coline Devin, Nicholas Rhinehart, Chelsea Finn, Dinesh Jayaraman, Sergey Levine
Every living organism struggles against disruptive environmental forces to carve out and maintain an orderly niche.
2 code implementations • ICCV 2019 • Nicholas Rhinehart, Rowan Mcallister, Kris Kitani, Sergey Levine
For autonomous vehicles (AVs) to behave appropriately on roads populated by human-driven vehicles, they must be able to reason about the uncertain intentions and decisions of other drivers from rich perceptual information.
no code implementations • CVPR 2020 • Jiaqi Guan, Ye Yuan, Kris M. Kitani, Nicholas Rhinehart
Automatically reasoning about future human behaviors is a difficult problem but has significant practical applications to assistive systems.
1 code implementation • ICLR 2020 • Nicholas Rhinehart, Rowan Mcallister, Sergey Levine
Yet, reward functions that evoke desirable behavior are often difficult to specify.
no code implementations • ICLR 2019 • Arjun Sharma, Mohit Sharma, Nicholas Rhinehart, Kris M. Kitani
The use of imitation learning to learn a single policy for a complex task that has multiple modes or hierarchical structure can be challenging.
no code implementations • 27 Sep 2018 • Nicholas Rhinehart, Anqi Liu, Kihyuk Sohn, Paul Vernaza
We propose a novel approach to regularizing generative adversarial networks (GANs) leveraging learned {\em structured Gibbs distributions}.
no code implementations • ECCV 2018 • Nicholas Rhinehart, Kris M. Kitani, Paul Vernaza
We propose a method to forecast a vehicle's ego-motion as a distribution over spatiotemporal paths, conditioned on features (e. g., from LIDAR and images) embedded in an overhead map.
no code implementations • 22 Jun 2018 • Xinlei Pan, Eshed Ohn-Bar, Nicholas Rhinehart, Yan Xu, Yilin Shen, Kris M. Kitani
The learning process is interactive, with a human expert first providing input in the form of full demonstrations along with some subgoal states.
no code implementations • 20 Jun 2018 • Tanmay Shankar, Nicholas Rhinehart, Katharina Muelling, Kris M. Kitani
We introduce a novel deterministic policy gradient update, DRAG (i. e., DeteRministically AGgrevate) in the form of a deterministic actor-critic variant of AggreVaTeD, to train our neural parser.
no code implementations • NeurIPS 2017 • Arun Venkatraman, Nicholas Rhinehart, Wen Sun, Lerrel Pinto, Martial Hebert, Byron Boots, Kris M. Kitani, J. Andrew Bagnell
We seek to combine the advantages of RNNs and PSRs by augmenting existing state-of-the-art recurrent neural networks with Predictive-State Decoders (PSDs), which add supervision to the network's internal state representation to target predicting future observations.
no code implementations • ICLR 2018 • Anubhav Ashok, Nicholas Rhinehart, Fares Beainy, Kris M. Kitani
Our approach takes a larger `teacher' network as input and outputs a compressed `student' network derived from the `teacher' network.
no code implementations • ICCV 2017 • Nicholas Rhinehart, Kris M. Kitani
We address the problem of incrementally modeling and forecasting long-term goals of a first-person camera wearer: what the user will do, where they will go, and what goal they seek.
no code implementations • CVPR 2016 • Nicholas Rhinehart, Kris M. Kitani
When people observe and interact with physical spaces, they are able to associate functionality to regions in the environment.
no code implementations • 27 Oct 2014 • Nicholas Rhinehart, Jiaji Zhou, Martial Hebert, J. Andrew Bagnell
We present an efficient algorithm with provable performance for building a high-quality list of detections from any candidate set of region-based proposals.