1 code implementation • 1 Mar 2023 • Firas Al-Hafez, Davide Tateo, Oleg Arenz, Guoping Zhao, Jan Peters
Recent methods for imitation learning directly learn a $Q$-function using an implicit reward formulation rather than an explicit reward function.
1 code implementation • 23 Sep 2022 • Oleg Arenz, Philipp Dahlinger, Zihan Ye, Michael Volpp, Gerhard Neumann
The two currently most effective methods for GMM-based variational inference, VIPS and iBayes-GMM, both employ independent natural gradient updates for the individual components and their weights.
no code implementations • 12 Sep 2022 • Bang You, Jingming Xie, Youping Chen, Jan Peters, Oleg Arenz
Recent works based on state-visitation counts, curiosity and entropy-maximization generate intrinsic reward signals to motivate the agent to visit novel states for exploration.
1 code implementation • 2 Mar 2022 • Bang You, Oleg Arenz, Youping Chen, Jan Peters
Recent methods for reinforcement learning from images use auxiliary tasks to learn image features that are used by the agent's policy or Q-function.
no code implementations • 29 Sep 2021 • Oleg Arenz, Zihan Ye, Philipp Dahlinger, Gerhard Neumann
Effective approaches for Gaussian variational inference are MORE, VOGN, and VON, which are zero-order, first-order, and second-order, respectively.
1 code implementation • 8 Aug 2020 • Oleg Arenz, Gerhard Neumann
We also show that our non-adversarial formulation can be used to derive novel algorithms by presenting a method for offline imitation learning that is inspired by the recent ValueDice algorithm, but does not rely on small policy updates for convergence.
no code implementations • 8 Mar 2020 • Melvin Laux, Oleg Arenz, Jan Peters, Joni Pajarinen
The ARL framework utilizes an adversary, which is trained to steer the original agent, the protagonist, to challenging states.
1 code implementation • ICLR 2020 • Philipp Becker, Oleg Arenz, Gerhard Neumann
Such behavior is appealing whenever we deal with highly multi-modal data where modelling single modes correctly is more important than covering all the modes.
no code implementations • 10 Jul 2019 • Oleg Arenz, Mingjun Zhong, Gerhard Neumann
For efficient improvement of the GMM approximation, we derive a lower bound on the corresponding optimization objective enabling us to update the components independently.
1 code implementation • ICML 2018 • Oleg Arenz, Gerhard Neumann, Mingjun Zhong
Inference from complex distributions is a common problem in machine learning needed for many Bayesian methods.