Search Results for author: Krishan Rana

Found 9 papers, 4 papers with code

SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Robot Task Planning

no code implementations12 Jul 2023 Krishan Rana, Jesse Haviland, Sourav Garg, Jad Abou-Chakra, Ian Reid, Niko Suenderhauf

To ensure the scalability of our approach, we: (1) exploit the hierarchical nature of 3DSGs to allow LLMs to conduct a 'semantic search' for task-relevant subgraphs from a smaller, collapsed representation of the full graph; (2) reduce the planning horizon for the LLM by integrating a classical path planner and (3) introduce an 'iterative replanning' pipeline that refines the initial plan using feedback from a scene graph simulator, correcting infeasible actions and avoiding planning failures.

Robot Task Planning

Contrastive Language, Action, and State Pre-training for Robot Learning

no code implementations21 Apr 2023 Krishan Rana, Andrew Melnik, Niko Sünderhauf

In this paper, we introduce a method for unifying language, action, and state information in a shared embedding space to facilitate a range of downstream tasks in robot learning.

Retrieval

Residual Skill Policies: Learning an Adaptable Skill-based Action Space for Reinforcement Learning for Robotics

1 code implementation4 Nov 2022 Krishan Rana, Ming Xu, Brendan Tidd, Michael Milford, Niko Sünderhauf

Furthermore, the downstream RL agent is limited to learning structurally similar tasks to those used to construct the skill space.

Reinforcement Learning (RL)

Zero-Shot Uncertainty-Aware Deployment of Simulation Trained Policies on Real-World Robots

no code implementations10 Dec 2021 Krishan Rana, Vibhavari Dasagi, Jesse Haviland, Ben Talbot, Michael Milford, Niko Sünderhauf

While deep reinforcement learning (RL) agents have demonstrated incredible potential in attaining dexterous behaviours for robotics, they tend to make errors when deployed in the real world due to mismatches between the training and execution environments.

Continuous Control Reinforcement Learning (RL)

Bayesian Controller Fusion: Leveraging Control Priors in Deep Reinforcement Learning for Robotics

no code implementations21 Jul 2021 Krishan Rana, Vibhavari Dasagi, Jesse Haviland, Ben Talbot, Michael Milford, Niko Sünderhauf

More importantly, given the risk-aversity of the control prior, BCF ensures safe exploration and deployment, where the control prior naturally dominates the action distribution in states unknown to the policy.

reinforcement-learning Reinforcement Learning (RL) +1

Critic Guided Segmentation of Rewarding Objects in First-Person Views

1 code implementation20 Jul 2021 Andrew Melnik, Augustin Harter, Christian Limberg, Krishan Rana, Niko Suenderhauf, Helge Ritter

This work discusses a learning approach to mask rewarding objects in images using sparse reward signals from an imitation learning dataset.

Imitation Learning

Multiplicative Controller Fusion: Leveraging Algorithmic Priors for Sample-efficient Reinforcement Learning and Safe Sim-To-Real Transfer

1 code implementation11 Mar 2020 Krishan Rana, Vibhavari Dasagi, Ben Talbot, Michael Milford, Niko Sünderhauf

We present a novel approach to model-free reinforcement learning that can leverage existing sub-optimal solutions as an algorithmic prior during training and deployment.

Robot Navigation

Residual Reactive Navigation: Combining Classical and Learned Navigation Strategies For Deployment in Unknown Environments

no code implementations24 Sep 2019 Krishan Rana, Ben Talbot, Vibhavari Dasagi, Michael Milford, Niko Sünderhauf

In this work we focus on improving the efficiency and generalisation of learned navigation strategies when transferred from its training environment to previously unseen ones.

Cannot find the paper you are looking for? You can Submit a new open access paper.