no code implementations • 23 Dec 2024 • Haohong Lin, Xin Huang, Tung Phan-Minh, David S. Hayden, huan zhang, Ding Zhao, Siddhartha Srinivasa, Eric M. Wolff, Hongge Chen
Then, CCDiff maximizes controllability while adhering to realism by automatically identifying and injecting causal structures directly into the diffusion process, providing structured guidance to enhance both realism and controllability.
no code implementations • 19 Dec 2024 • Xin Huang, Eric M. Wolff, Paul Vernaza, Tung Phan-Minh, Hongge Chen, David S. Hayden, Mark Edmonds, Brian Pierce, Xinxin Chen, Pratik Elias Jacob, Xiaobai Chen, Chingiz Tairbekov, Pratik Agarwal, Tianshi Gao, Yuning Chai, Siddhartha Srinivasa
We present DriveGPT, a scalable behavior model for autonomous driving.
no code implementations • 19 Dec 2024 • Yi Xu, Yuxin Hu, Zaiwei Zhang, Gregory P. Meyer, Siva Karthik Mustikovela, Siddhartha Srinivasa, Eric M. Wolff, Xin Huang
Human drivers rely on commonsense reasoning to navigate diverse and dynamic real-world scenarios.
no code implementations • 10 Oct 2023 • Yi Ru Wang, Jiafei Duan, Dieter Fox, Siddhartha Srinivasa
To address this gap, we introduce NEWTON, a repository and benchmark for evaluating the physics reasoning skills of LLMs.
no code implementations • 12 Oct 2022 • Gaoyue Zhou, Liyiming Ke, Siddhartha Srinivasa, Abhinav Gupta, Aravind Rajeswaran, Vikash Kumar
Offline reinforcement learning (ORL) holds great promise for robot learning due to its ability to learn from arbitrary pre-generated experience.
3 code implementations • 11 Sep 2022 • Samuel K. Ainsworth, Jonathan Hayase, Siddhartha Srinivasa
The success of deep learning is due in large part to our ability to solve certain massive non-convex optimization problems with relative ease.
1 code implementation • 13 Jul 2022 • Vincent Roulet, Siddhartha Srinivasa, Maryam Fazel, Zaid Harchaoui
We present the implementation of nonlinear control algorithms based on linear and quadratic approximations of the objective from a functional viewpoint.
no code implementations • 10 Oct 2021 • Mohak Bhardwaj, Sanjiban Choudhury, Byron Boots, Siddhartha Srinivasa
If new search problems are sufficiently similar to problems solved during training, the learned policy will choose a good edge evaluation ordering and solve the motion planning problem quickly.
3 code implementations • 12 Dec 2020 • Samuel Ainsworth, Kendall Lowrey, John Thickstun, Zaid Harchaoui, Siddhartha Srinivasa
We study the estimation of policy gradients for continuous-time systems with known dynamics.
no code implementations • 13 Nov 2020 • Liyiming Ke, Jingqiang Wang, Tapomayukh Bhattacharjee, Byron Boots, Siddhartha Srinivasa
Billions of people use chopsticks, a simple yet versatile tool, for fine manipulation of everyday objects.
1 code implementation • 28 Sep 2020 • William Agnew, Christopher Xie, Aaron Walsman, Octavian Murad, Caelen Wang, Pedro Domingos, Siddhartha Srinivasa
By using these priors over the physical properties of objects, our system improves reconstruction quality not just by standard visual metrics, but also performance of model-based control on a variety of robotics manipulation tasks in challenging, cluttered environments.
no code implementations • L4DC 2020 • Colin Summers, Kendall Lowrey, Aravind Rajeswaran, Siddhartha Srinivasa, Emanuel Todorov
We introduce Lyceum, a high-performance computational ecosystem for robot learning.
1 code implementation • NeurIPS 2019 • Samuel Ainsworth, Matt Barnes, Siddhartha Srinivasa
In many environments, only a relatively small subset of the complete state space is necessary in order to accomplish a given task.
no code implementations • 16 Jul 2019 • Mohak Bhardwaj, Sanjiban Choudhury, Byron Boots, Siddhartha Srinivasa
If new search problems are sufficiently similar to problems solved during training, the learned policy will choose a good edge evaluation ordering and solve the motion planning problem quickly.
no code implementations • 30 May 2019 • Liyiming Ke, Sanjiban Choudhury, Matt Barnes, Wen Sun, Gilwoo Lee, Siddhartha Srinivasa
We show that the state-of-the-art methods such as GAIL and behavior cloning, due to their choice of loss function, often incorrectly interpolate between such modes.
1 code implementation • 2 Apr 2019 • Rosario Scalise, Jesse Thomason, Yonatan Bisk, Siddhartha Srinivasa
We collect over 13 hours of egocentric manipulation data for training a model to reason about whether a robot successfully placed unseen objects in or on one another.
1 code implementation • CVPR 2019 • Liyiming Ke, Xiujun Li, Yonatan Bisk, Ari Holtzman, Zhe Gan, Jingjing Liu, Jianfeng Gao, Yejin Choi, Siddhartha Srinivasa
We present the Frontier Aware Search with backTracking (FAST) Navigator, a general framework for action decoding, that achieves state-of-the-art results on the Room-to-Room (R2R) Vision-and-Language navigation challenge of Anderson et.
Ranked #4 on Vision-Language Navigation on Room2Room
1 code implementation • 24 Jan 2019 • Lawrence Chan, Dylan Hadfield-Menell, Siddhartha Srinivasa, Anca Dragan
Learning preferences implicit in the choices humans make is a well studied problem in both economics and computer science.
no code implementations • 22 Jan 2019 • Sumit Kumar, Shushman Choudhary, Siddhartha Srinivasa
Our aim is to reduce the expected number of collision checks by creating a belief model of the configuration space using results from collision tests.
no code implementations • 24 Oct 2018 • Lerrel Pinto, Aditya Mandalika, Brian Hou, Siddhartha Srinivasa
This paper proposes a sample-efficient yet simple approach to learning closed-loop policies for nonprehensile manipulation.
no code implementations • 20 May 2018 • Rosario Scalise, Yonatan Bisk, Maxwell Forbes, Daqing Yi, Yejin Choi, Siddhartha Srinivasa
Robotic agents that share autonomy with a human should leverage human domain knowledge and account for their preferences when completing a task.
2 code implementations • ICML 2018 • Ahmed Hefny, Zita Marinho, Wen Sun, Siddhartha Srinivasa, Geoffrey Gordon
Predictive state policy networks consist of a recursive filter, which keeps track of a belief about the state of the environment, and a reactive policy that directly maps beliefs to actions, to maximize the cumulative reward.
no code implementations • 12 Jan 2018 • Min Chen, Stefanos Nikolaidis, Harold Soh, David Hsu, Siddhartha Srinivasa
The trust-POMDP model provides a principled approach for the robot to (i) infer the trust of a human teammate through interaction, (ii) reason about the effect of its own actions on human trust, and (iii) choose actions that maximize team performance over the long term.
no code implementations • 20 Nov 2017 • Sanjiban Choudhury, Siddhartha Srinivasa, Sebastian Scherer
We are interested in planning algorithms that actively infer the underlying structure of the valid configuration space during planning in order to find solutions with minimal effort.
1 code implementation • NeurIPS 2017 • Sanjiban Choudhury, Shervin Javdani, Siddhartha Srinivasa, Sebastian Scherer
By leveraging this property, we are able to significantly reduce computational complexity from exponential to linear in the number of edges.
Robotics
no code implementations • 24 Feb 2014 • Shervin Javdani, Yuxin Chen, Amin Karbasi, Andreas Krause, J. Andrew Bagnell, Siddhartha Srinivasa
Instead of minimizing uncertainty per se, we consider a set of overlapping decision regions of these hypotheses.