Search Results for author: Peter Nickl

Found 4 papers, 3 papers with code

Variational Learning is Effective for Large Deep Networks

1 code implementation27 Feb 2024 Yuesong Shen, Nico Daheim, Bai Cong, Peter Nickl, Gian Maria Marconi, Clement Bazan, Rio Yokota, Iryna Gurevych, Daniel Cremers, Mohammad Emtiyaz Khan, Thomas Möllenhoff

We give extensive empirical evidence against the common belief that variational learning is ineffective for large neural networks.

The Memory Perturbation Equation: Understanding Model's Sensitivity to Data

1 code implementation30 Oct 2023 Peter Nickl, Lu Xu, Dharmesh Tailor, Thomas Möllenhoff, Mohammad Emtiyaz Khan

Understanding model's sensitivity to its training data is crucial but can also be challenging and costly, especially during training.

Variational Hierarchical Mixtures for Probabilistic Learning of Inverse Dynamics

no code implementations2 Nov 2022 Hany Abdulsamad, Peter Nickl, Pascal Klink, Jan Peters

We derive two efficient variational inference techniques to learn these representations and highlight the advantages of hierarchical infinite local regression models, such as dealing with non-smooth functions, mitigating catastrophic forgetting, and enabling parameter sharing and fast predictions.

regression Variational Inference

A Variational Infinite Mixture for Probabilistic Inverse Dynamics Learning

1 code implementation10 Nov 2020 Hany Abdulsamad, Peter Nickl, Pascal Klink, Jan Peters

Probabilistic regression techniques in control and robotics applications have to fulfill different criteria of data-driven adaptability, computational efficiency, scalability to high dimensions, and the capacity to deal with different modalities in the data.

Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.