1 code implementation • NeurIPS 2023 • Kim A. Nicoli, Christopher J. Anders, Lena Funcke, Tobias Hartung, Karl Jansen, Stefan Kühn, Klaus-Robert Müller, Paolo Stornati, Pan Kessel, Shinichi Nakajima
In this paper, we propose a novel and powerful method to harness Bayesian optimization for Variational Quantum Eigensolvers (VQEs) -- a hybrid quantum-classical protocol used to approximate the ground state of a quantum Hamiltonian.
no code implementations • 10 Jun 2024 • Philipp Misof, Pan Kessel, Jan E. Gerken
Equivariant neural networks have in recent years become an important technique for guiding architecture selection for neural networks with many applications in domains ranging from medical image analysis to quantum chemistry.
no code implementations • 23 Mar 2024 • Lorenz Vaitl, Ludwig Winkler, Lorenz Richter, Pan Kessel
Recent work shows that path gradient estimators for normalizing flows have lower variance compared to standard estimators for variational inference, resulting in improved training.
no code implementations • 5 Mar 2024 • Jan E. Gerken, Pan Kessel
We show that deep ensembles become equivariant for all inputs and at all training times by simply using data augmentation.
no code implementations • 18 Jul 2023 • Andreas Loukas, Pan Kessel, Vladimir Gligorijevic, Richard Bonneau
In silico screening uses predictive models to select a batch of compounds with favorable properties from a library for experimental validation.
no code implementations • 27 Feb 2023 • Kim A. Nicoli, Christopher J. Anders, Tobias Hartung, Karl Jansen, Pan Kessel, Shinichi Nakajima
In this work, we first point out that the tunneling problem is also present for normalizing flows but is shifted from the sampling to the training phase of the algorithm.
no code implementations • 17 Jul 2022 • Lorenz Vaitl, Kim A. Nicoli, Shinichi Nakajima, Pan Kessel
We propose an algorithm to estimate the path-gradient of both the reverse and forward Kullback-Leibler divergence for an arbitrary manifestly invertible normalizing flow.
1 code implementation • 17 Jun 2022 • Lorenz Vaitl, Kim A. Nicoli, Shinichi Nakajima, Pan Kessel
Recent work has established a path-gradient estimator for simple variational Gaussian distributions and has argued that the path-gradient is particularly beneficial in the regime in which the variational distribution approaches the exact target distribution.
1 code implementation • 10 Jun 2022 • Ann-Kathrin Dombrowski, Jan E. Gerken, Klaus-Robert Müller, Pan Kessel
Counterfactuals can explain classification decisions of neural networks in a human interpretable way.
no code implementations • 22 Nov 2021 • Kim A. Nicoli, Christopher Anders, Lena Funcke, Tobias Hartung, Karl Jansen, Pan Kessel, Shinichi Nakajima, Paolo Stornati
Crucially, these models allow for the direct estimation of the free energy at a given point in parameter space.
no code implementations • 23 Aug 2021 • Siddhant Agarwal, Nicola Tosi, Pan Kessel, Doris Breuer, Grégoire Montavon
Using a dataset of 10, 525 two-dimensional simulations of the thermal evolution of the mantle of a Mars-like planet, we show that deep learning techniques can produce reliable parameterized surrogates (i. e. surrogates that predict state variables such as temperature based only on parameters) of the underlying partial differential equations.
no code implementations • ICML Workshop INNF 2021 • Ann-Kathrin Dombrowski, Jan E Gerken, Pan Kessel
Normalizing flows are diffeomorphisms which are parameterized by neural networks.
no code implementations • 18 Dec 2020 • Ann-Kathrin Dombrowski, Christopher J. Anders, Klaus-Robert Müller, Pan Kessel
Explanation methods shed light on the decision process of black-box classifiers such as deep neural networks.
1 code implementation • ICML 2020 • Christopher J. Anders, Plamen Pasliev, Ann-Kathrin Dombrowski, Klaus-Robert Müller, Pan Kessel
Explanation methods promise to make black-box classifiers more transparent.
no code implementations • 14 Jul 2020 • Kim A. Nicoli, Christopher J. Anders, Lena Funcke, Tobias Hartung, Karl Jansen, Pan Kessel, Shinichi Nakajima, Paolo Stornati
In this work, we demonstrate that applying deep generative machine learning models for lattice field theory is a promising route for solving problems where Markov Chain Monte Carlo (MCMC) methods are problematic.
no code implementations • 29 Oct 2019 • Kim A. Nicoli, Shinichi Nakajima, Nils Strodthoff, Wojciech Samek, Klaus-Robert Müller, Pan Kessel
We propose a general framework for the estimation of observables with generative neural samplers focusing on modern deep generative neural networks that provide an exact sampling probability.
2 code implementations • NeurIPS 2019 • Ann-Kathrin Dombrowski, Maximilian Alber, Christopher J. Anders, Marcel Ackermann, Klaus-Robert Müller, Pan Kessel
Explanation methods aim to make neural networks more trustworthy and interpretable.
no code implementations • 26 Mar 2019 • Kim Nicoli, Pan Kessel, Nils Strodthoff, Wojciech Samek, Klaus-Robert Müller, Shinichi Nakajima
In this comment on "Solving Statistical Mechanics Using Variational Autoregressive Networks" by Wu et al., we propose a subtle yet powerful modification of their approach.
no code implementations • 23 Oct 2018 • Kim A. Nicoli, Pan Kessel, Michael Gastegger, Kristof T. Schütt
In this work, we extend the SchNet architecture by using weighted skip connections to assemble the final representation.