no code implementations • 11 Mar 2025 • Yuki Shirai, Arvind Raghunathan, Devesh K. Jha
In this paper, we present a novel framework for simultaneously designing trajectories of robots, objects, and contacts efficiently for contact-rich manipulation.
no code implementations • 14 Jan 2025 • Yuxin Chen, Devesh K. Jha, Masayoshi Tomizuka, Diego Romeres
This reward is then used to fine-tune the pre-trained policy with reinforcement learning (RL), resulting in alignment of pre-trained policy with new human preferences while still solving the original task.
no code implementations • 17 Oct 2024 • Shivam Vats, Devesh K. Jha, Maxim Likhachev, Oliver Kroemer, Diego Romeres
Our approach, called RecoveryChaining, uses a hybrid action space, where the model-based controllers are provided as additional \emph{nominal} options which allows the recovery policy to decide how to recover, when to switch to a nominal controller and which controller to switch to even with \emph{sparse rewards}.
no code implementations • 17 Dec 2023 • Xinghao Zhu, Devesh K. Jha, Diego Romeres, Lingfeng Sun, Masayoshi Tomizuka, Anoop Cherian
Automating the assembly of objects from their parts is a complex problem with innumerable applications in manufacturing, maintenance, and recycling.
no code implementations • 11 Dec 2023 • Lingfeng Sun, Devesh K. Jha, Chiori Hori, Siddarth Jain, Radu Corcodel, Xinghao Zhu, Masayoshi Tomizuka, Diego Romeres
Designing robotic agents to perform open vocabulary tasks has been the long-standing goal in robotics and AI.
no code implementations • 25 Sep 2023 • Kei Ota, Devesh K. Jha, Krishna Murthy Jatavallabhula, Asako Kanezaki, Joshua B. Tenenbaum
In particular, we estimate the contact patch between a grasped object and its environment using force and tactile observations to estimate the stability of the object during a contact formation.
no code implementations • 18 May 2023 • Eric Rosen, Devesh K. Jha
We address the problem of teleoperating an industrial robot manipulator via a commercially available Virtual Reality (VR) interface.
no code implementations • 15 Mar 2023 • Yuki Shirai, Devesh K. Jha, Arvind U. Raghunathan
This requirement makes manipulation extremely challenging as a robot has to reason about complex frictional interactions with uncertainty in physical properties of the object and the environment.
no code implementations • 10 Mar 2023 • Kei Ota, Devesh K. Jha, Hsiao-Yu Tung, Joshua B. Tenenbaum
We evaluate our method on several part-mating tasks with novel objects using a robot equipped with a vision-based tactile sensor.
no code implementations • 2 Dec 2022 • Devesh K. Jha, Siddarth Jain, Diego Romeres, William Yerazunis, Daniel Nikovski
In this paper, we present a system for human-robot collaborative assembly using learning from demonstration and pose estimation, so that the robot can adapt to the uncertainty caused by the operation of humans.
no code implementations • 23 Oct 2022 • Tim Schneider, Boris Belousov, Georgia Chalvatzaki, Diego Romeres, Devesh K. Jha, Jan Peters
Robotic manipulation stands as a largely unsolved problem despite significant advances in robotics and machine learning in recent years.
no code implementations • 28 Sep 2022 • Seiji Shaw, Devesh K. Jha, Arvind Raghunathan, Radu Corcodel, Diego Romeres, George Konidaris, Daniel Nikovski
In this paper, we present constrained dynamic movement primitives (CDMP) which can allow for constraint satisfaction in the robot workspace.
no code implementations • 22 Mar 2022 • Yuki Shirai, Devesh K. Jha, Arvind Raghunathan, Diego Romeres
Generalizable manipulation requires that robots be able to interact with novel objects and environment.
no code implementations • 5 Mar 2022 • Yuki Shirai, Devesh K. Jha, Arvind Raghunathan, Diego Romeres
In our formulation, we explicitly consider joint chance constraints for complementarity as well as states to capture the stochastic evolution of dynamics.
no code implementations • 20 Nov 2021 • Devesh K. Jha, Diego Romeres, William Yerazunis, Daniel Nikovski
This can be used to learn a suitable representation of the skill that can be generalized to novel positions of one of the parts involved in the assembly, for example the hole in a peg-in-hole (PiH) insertion task.
no code implementations • 6 Jun 2021 • Arvind U. Raghunathan, Devesh K. Jha, Diego Romeres
PYROBOCOP is a lightweight Python-based package for control and optimization of robotic systems described by nonlinear Differential Algebraic Equations (DAEs).
no code implementations • 20 Mar 2021 • Devesh K. Jha
On the other hand, memory estimation of the symbolic sequence helps to extract the predictive patterns in the discretized data.
no code implementations • 16 Feb 2021 • Kei Ota, Devesh K. Jha, Asako Kanezaki
Previous work has shown that this is mostly due to instability during training of deep RL agents when using larger networks.
no code implementations • 14 Nov 2020 • Kei Ota, Devesh K. Jha, Diego Romeres, Jeroen van Baar, Kevin A. Smith, Takayuki Semitsu, Tomoaki Oiki, Alan Sullivan, Daniel Nikovski, Joshua B. Tenenbaum
The physics engine augmented with the residual model is then used to control the marble in the maze environment using a model-predictive feedback over a receding horizon.
no code implementations • 31 Oct 2020 • Kei Ota, Devesh K. Jha, Tadashi Onishi, Asako Kanezaki, Yusuke Yoshiyasu, Yoko SASAKI, Toshisada Mariyama, Daniel Nikovski
The main novelty of the proposed approach is that it allows a robot to learn an end-to-end policy which can adapt to changes in the environment during execution.
no code implementations • 22 Jul 2020 • Yifang Liu, Diego Romeres, Devesh K. Jha, Daniel Nikovski
One of the main challenges in peg-in-a-hole (PiH) insertion tasks is in handling the uncertainty in the location of the target hole.
no code implementations • 26 Mar 2020 • Wenyu Zhang, Skyler Seto, Devesh K. Jha
The purpose of these agents is to quickly adapt and/or generalize their notion of physics of interaction in the real world based on certain features about the interacting objects that provide different contexts to the predictive models.
no code implementations • 3 Mar 2020 • Kei Ota, Yoko SASAKI, Devesh K. Jha, Yusuke Yoshiyasu, Asako Kanezaki
Specifically, we train a deep convolutional network that can predict collision-free paths based on a map of the environment-- this is then used by a reinforcement learning algorithm to learn to closely follow the path.
1 code implementation • ICML 2020 • Kei Ota, Tomoaki Oiki, Devesh K. Jha, Toshisada Mariyama, Daniel Nikovski
We believe that stronger feature propagation together with larger networks (and thus larger search space) allows RL agents to learn more complex functions of states and thus improves the sample efficiency.
no code implementations • 25 Feb 2020 • Alberto Dalla Libera, Diego Romeres, Devesh K. Jha, Bill Yerazunis, Daniel Nikovski
In this paper, we propose a derivative-free model learning framework for Reinforcement Learning (RL) algorithms based on Gaussian Process Regression (GPR).
no code implementations • 27 Jan 2020 • Wenyu Zhang, Devesh K. Jha, Emil Laftchiev, Daniel Nikovski
In the most general setting of these types of problems, one or more samples of data across multiple time series can be assigned several concurrent fault labels from a finite, known set and the task is to predict the possibility of fault occurrence over a desired time horizon.
no code implementations • 22 Jan 2020 • Patrik Kolaric, Devesh K. Jha, Arvind U. Raghunathan, Frank L. Lewis, Mouhacine Benosman, Diego Romeres, Daniel Nikovski
Motivated by these problems, we try to formulate the problem of trajectory optimization and local policy synthesis as a single optimization problem.
Model-based Reinforcement Learning
reinforcement-learning
+2
no code implementations • 3 Jul 2019 • Ankush Chakrabarty, Devesh K. Jha, Gregery T. Buzzard, Yebin Wang, Kyriakos Vamvoudakis
We develop a method for obtaining safe initial policies for reinforcement learning via approximate dynamic programming (ADP) techniques for uncertain systems evolving with discrete-time dynamics.
no code implementations • 15 May 2019 • Arvind U. Raghunathan, Anoop Cherian, Devesh K. Jha
To this end, we introduce the Gradient-based Nikaido-Isoda (GNI) function which serves: (i) as a merit function, vanishing only at the first-order stationary points of each player's optimization problem, and (ii) provides error bounds to a stationary Nash point.
no code implementations • 13 Mar 2019 • Kei Ota, Devesh K. Jha, Tomoaki Oiki, Mamoru Miura, Takashi Nammoto, Daniel Nikovski, Toshisada Mariyama
Our experiments show that our RL agent trained with a reference path outperformed a model-free PID controller of the type commonly used on many robotic platforms for trajectory tracking.
no code implementations • 26 Sep 2017 • Devesh K. Jha, Nurali Virani, Jan Reimann, Abhishek Srivastav, Asok Ray
In the second example, the data set is taken from NASA's data repository for prognostics of bearings on rotating shafts.