Search Results for author: Oleg Arenz

Found 10 papers, 6 papers with code

LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning

1 code implementation1 Mar 2023 Firas Al-Hafez, Davide Tateo, Oleg Arenz, Guoping Zhao, Jan Peters

Recent methods for imitation learning directly learn a $Q$-function using an implicit reward formulation rather than an explicit reward function.

Continuous Control Imitation Learning +4

A Unified Perspective on Natural Gradient Variational Inference with Gaussian Mixture Models

1 code implementation23 Sep 2022 Oleg Arenz, Philipp Dahlinger, Zihan Ye, Michael Volpp, Gerhard Neumann

The two currently most effective methods for GMM-based variational inference, VIPS and iBayes-GMM, both employ independent natural gradient updates for the individual components and their weights.

Variational Inference

Self-supervised Sequential Information Bottleneck for Robust Exploration in Deep Reinforcement Learning

no code implementations12 Sep 2022 Bang You, Jingming Xie, Youping Chen, Jan Peters, Oleg Arenz

Recent works based on state-visitation counts, curiosity and entropy-maximization generate intrinsic reward signals to motivate the agent to visit novel states for exploration.

Efficient Exploration reinforcement-learning +3

Integrating Contrastive Learning with Dynamic Models for Reinforcement Learning from Images

1 code implementation2 Mar 2022 Bang You, Oleg Arenz, Youping Chen, Jan Peters

Recent methods for reinforcement learning from images use auxiliary tasks to learn image features that are used by the agent's policy or Q-function.

Contrastive Learning Data Augmentation +3

A First-Order Method for Estimating Natural Gradients for Variational Inference with Gaussians and Gaussian Mixture Models

no code implementations29 Sep 2021 Oleg Arenz, Zihan Ye, Philipp Dahlinger, Gerhard Neumann

Effective approaches for Gaussian variational inference are MORE, VOGN, and VON, which are zero-order, first-order, and second-order, respectively.

Variational Inference

Non-Adversarial Imitation Learning and its Connections to Adversarial Methods

1 code implementation8 Aug 2020 Oleg Arenz, Gerhard Neumann

We also show that our non-adversarial formulation can be used to derive novel algorithms by presenting a method for offline imitation learning that is inspired by the recent ValueDice algorithm, but does not rely on small policy updates for convergence.

Imitation Learning

Deep Adversarial Reinforcement Learning for Object Disentangling

no code implementations8 Mar 2020 Melvin Laux, Oleg Arenz, Jan Peters, Joni Pajarinen

The ARL framework utilizes an adversary, which is trained to steer the original agent, the protagonist, to challenging states.

Object reinforcement-learning +1

Expected Information Maximization: Using the I-Projection for Mixture Density Estimation

1 code implementation ICLR 2020 Philipp Becker, Oleg Arenz, Gerhard Neumann

Such behavior is appealing whenever we deal with highly multi-modal data where modelling single modes correctly is more important than covering all the modes.

Density Estimation Traffic Prediction

Trust-Region Variational Inference with Gaussian Mixture Models

no code implementations10 Jul 2019 Oleg Arenz, Mingjun Zhong, Gerhard Neumann

For efficient improvement of the GMM approximation, we derive a lower bound on the corresponding optimization objective enabling us to update the components independently.

Variational Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.