Search Results for author: Rudolf Lioutikov

Found 12 papers, 8 papers with code

Open the Black Box: Step-based Policy Updates for Temporally-Correlated Episodic Reinforcement Learning

1 code implementation21 Jan 2024 Ge Li, Hongyi Zhou, Dominik Roth, Serge Thilges, Fabian Otto, Rudolf Lioutikov, Gerhard Neumann

Current advancements in reinforcement learning (RL) have predominantly focused on learning step-based policies that generate actions for each perceived state.

Reinforcement Learning (RL)

Movement Primitive Diffusion: Learning Gentle Robotic Manipulation of Deformable Objects

no code implementations15 Dec 2023 Paul Maria Scheikl, Nicolas Schreiber, Christoph Haas, Niklas Freymuth, Gerhard Neumann, Rudolf Lioutikov, Franziska Mathis-Ullrich

Policy learning in robot-assisted surgery (RAS) lacks data efficient and versatile methods that exhibit the desired motion quality for delicate surgical interventions.

Imitation Learning

MP3: Movement Primitive-Based (Re-)Planning Policy

no code implementations22 Jun 2023 Fabian Otto, Hongyi Zhou, Onur Celik, Ge Li, Rudolf Lioutikov, Gerhard Neumann

We introduce a novel deep reinforcement learning (RL) approach called Movement Primitive-based Planning Policy (MP3).

Reinforcement Learning (RL)

Goal-Conditioned Imitation Learning using Score-based Diffusion Policies

1 code implementation5 Apr 2023 Moritz Reuss, Maximilian Li, Xiaogang Jia, Rudolf Lioutikov

To the best of our knowledge this is the first work that a) represents a behavior policy based on such a decoupled SDM b) learns an SDM based policy in the domain of GCIL and c) provides a way to simultaneously learn a goal-dependent and a goal-independent policy from play-data.

Denoising Imitation Learning

ProDMPs: A Unified Perspective on Dynamic and Probabilistic Movement Primitives

no code implementations4 Oct 2022 Ge Li, Zeqi Jin, Michael Volpp, Fabian Otto, Rudolf Lioutikov, Gerhard Neumann

MPs can be broadly categorized into two types: (a) dynamics-based approaches that generate smooth trajectories from any initial state, e. g., Dynamic Movement Primitives (DMPs), and (b) probabilistic approaches that capture higher-order statistics of the motion, e. g., Probabilistic Movement Primitives (ProMPs).

Numerical Integration

Distributional Depth-Based Estimation of Object Articulation Models

1 code implementation12 Aug 2021 Ajinkya Jain, Stephen Giguere, Rudolf Lioutikov, Scott Niekum

Our core contributions include a novel representation for distributions over rigid body transformations and articulation model parameters based on screw theory, von Mises-Fisher distributions, and Stiefel manifolds.

Benchmarking Object

Self-Supervised Online Reward Shaping in Sparse-Reward Environments

1 code implementation8 Mar 2021 Farzan Memarian, Wonjoon Goo, Rudolf Lioutikov, Scott Niekum, Ufuk Topcu

We introduce Self-supervised Online Reward Shaping (SORS), which aims to improve the sample efficiency of any RL algorithm in sparse-reward environments by automatically densifying rewards.

ScrewNet: Category-Independent Articulation Model Estimation From Depth Images Using Screw Theory

1 code implementation24 Aug 2020 Ajinkya Jain, Rudolf Lioutikov, Caleb Chuck, Scott Niekum

Robots in human environments will need to interact with a wide variety of articulated objects such as cabinets, drawers, and dishwashers while assisting humans in performing day-to-day tasks.

Benchmarking

Cannot find the paper you are looking for? You can Submit a new open access paper.