Search Results for author: Philipp Becker

Found 18 papers, 12 papers with code

KalMamba: Towards Efficient Probabilistic State Space Models for RL under Uncertainty

no code implementations21 Jun 2024 Philipp Becker, Niklas Freymuth, Gerhard Neumann

We propose KalMamba, an efficient architecture to learn representations for RL that combines the strengths of probabilistic SSMs with the scalability of deterministic SSMs.

Computational Efficiency Mamba +2

Vlearn: Off-Policy Learning with Efficient State-Value Function Estimation

no code implementations7 Mar 2024 Fabian Otto, Philipp Becker, Ngo Anh Vien, Gerhard Neumann

This transfer to deep methods is not straightforward and requires novel design choices such as robust policy updates, twin value function networks to avoid an optimization bias, and importance weight clipping.

Efficient Exploration reinforcement-learning

Information-Theoretic Trust Regions for Stochastic Gradient-Based Optimization

1 code implementation31 Oct 2023 Philipp Dahlinger, Philipp Becker, Maximilian Hüttenrauch, Gerhard Neumann

Before each update, it solves the trust region problem for an optimal step size, resulting in a more stable and faster optimization process.

Combining Reconstruction and Contrastive Methods for Multimodal Representations in RL

1 code implementation10 Feb 2023 Philipp Becker, Sebastian Mossburger, Fabian Otto, Gerhard Neumann

Here, different self-supervised loss functions have distinct advantages and limitations depending on the information density of the underlying sensor modality.

reinforcement-learning Reinforcement Learning (RL) +1

On Uncertainty in Deep State Space Models for Model-Based Reinforcement Learning

1 code implementation17 Oct 2022 Philipp Becker, Gerhard Neumann

We show that RSSMs use a suboptimal inference scheme and that models trained using this inference overestimate the aleatoric uncertainty of the ground truth system.

Model-based Reinforcement Learning reinforcement-learning +3

Inferring Versatile Behavior from Demonstrations by Matching Geometric Descriptors

1 code implementation17 Oct 2022 Niklas Freymuth, Nicolas Schreiber, Philipp Becker, Aleksandar Taranovic, Gerhard Neumann

We find that the geometric descriptors greatly help in generalizing to new task configurations and that combining them with our distribution-matching objective is crucial for representing and reproducing versatile behavior.

Imitation Learning

End-to-End Learning of Hybrid Inverse Dynamics Models for Precise and Compliant Impedance Control

no code implementations27 May 2022 Moritz Reuss, Niels van Duijkeren, Robert Krug, Philipp Becker, Vaisakh Shaj, Gerhard Neumann

These models need to precisely capture the robot dynamics, which consist of well-understood components, e. g., rigid body dynamics, and effects that remain challenging to capture, e. g., stick-slip friction and mechanical flexibilities.

Friction

Specializing Versatile Skill Libraries using Local Mixture of Experts

1 code implementation8 Dec 2021 Onur Celik, Dongzhuoran Zhou, Ge Li, Philipp Becker, Gerhard Neumann

This local and incremental learning results in a modular MoE model of high accuracy and versatility, where both properties can be scaled by adding more components on the fly.

Incremental Learning Reinforcement Learning (RL)

Switching Recurrent Kalman Networks

no code implementations16 Nov 2021 Giao Nguyen-Quynh, Philipp Becker, Chen Qiu, Maja Rudolph, Gerhard Neumann

In addition, driving data can often be multimodal in distribution, meaning that there are distinct predictions that are likely, but averaging can hurt model performance.

Autonomous Driving Time Series +1

Versatile Inverse Reinforcement Learning via Cumulative Rewards

no code implementations15 Nov 2021 Niklas Freymuth, Philipp Becker, Gerhard Neumann

Inverse Reinforcement Learning infers a reward function from expert demonstrations, aiming to encode the behavior and intentions of the expert.

reinforcement-learning Reinforcement Learning +1

Action-Conditional Recurrent Kalman Networks For Forward and Inverse Dynamics Learning

2 code implementations20 Oct 2020 Vaisakh Shaj, Philipp Becker, Dieter Buchler, Harit Pandya, Niels van Duijkeren, C. James Taylor, Marc Hanheide, Gerhard Neumann

We adopt a recent probabilistic recurrent neural network architecture, called Re-current Kalman Networks (RKNs), to model learning by conditioning its transition dynamics on the control actions.

Friction

Expected Information Maximization: Using the I-Projection for Mixture Density Estimation

1 code implementation ICLR 2020 Philipp Becker, Oleg Arenz, Gerhard Neumann

Such behavior is appealing whenever we deal with highly multi-modal data where modelling single modes correctly is more important than covering all the modes.

Density Estimation Traffic Prediction

Recurrent Kalman Networks: Factorized Inference in High-Dimensional Deep Feature Spaces

3 code implementations17 May 2019 Philipp Becker, Harit Pandya, Gregor Gebhardt, Cheng Zhao, James Taylor, Gerhard Neumann

In order to integrate uncertainty estimates into deep time-series modelling, Kalman Filters (KFs) (Kalman et al., 1960) have been integrated with deep learning models, however, such approaches typically rely on approximate inference techniques such as variational inference which makes learning more complex and often less scalable due to approximation errors.

Image Imputation Imputation +4

Cannot find the paper you are looking for? You can Submit a new open access paper.