1 code implementation • ICML 2020 • Voot Tangkaratt, Bo Han, Mohammad Emtiyaz Khan, Masashi Sugiyama
Learning from demonstrations can be challenging when the quality of demonstrations is diverse, and even more so when the quality is unknown and there is no additional information to estimate the quality.
no code implementations • 8 Mar 2023 • Eren Mehmet Kıral, Thomas Möllenhoff, Mohammad Emtiyaz Khan
This simplifies all three difficulties for many cases, providing flexible parametrizations through group's action, simple gradient computation through reparameterization, and updates that always stay on the manifold.
no code implementations • 20 Feb 2023 • Wu Lin, Valentin Duruisseaux, Melvin Leok, Frank Nielsen, Mohammad Emtiyaz Khan, Mark Schmidt
Riemannian submanifold optimization with momentum is computationally challenging because ensuring iterates remain on the submanifold often requires solving difficult differential equations.
no code implementations • 21 Oct 2022 • Alexandre Piche, Valentin Thomas, Joseph Marino, Rafael Pardinas, Gian Maria Marconi, Christopher Pal, Mohammad Emtiyaz Khan
However, learning the value function via bootstrapping often leads to unstable training due to fast-changing target values.
no code implementations • 12 Oct 2022 • Ganesh Tata, Gautham Krishna Gudur, Gopinath Chennupati, Mohammad Emtiyaz Khan
Calibration can reduce overconfident predictions of deep neural networks, but can calibration also accelerate training?
no code implementations • 4 Oct 2022 • Thomas Möllenhoff, Mohammad Emtiyaz Khan
Sharpness-aware minimization (SAM) and related adversarial deep-learning methods can drastically improve generalization, but their underlying mechanisms are not yet fully understood.
1 code implementation • NeurIPS 2021 • Vincent Adam, Paul E. Chang, Mohammad Emtiyaz Khan, Arno Solin
Sparse variational Gaussian process (SVGP) methods are a common choice for non-conjugate Gaussian process inference because of their computational benefits.
no code implementations • 22 Jul 2021 • Wu Lin, Frank Nielsen, Mohammad Emtiyaz Khan, Mark Schmidt
In this paper, we propose new structured second-order methods and structured adaptive-gradient methods obtained by performing natural-gradient descent on structured parameter spaces.
1 code implementation • 17 Jul 2021 • Ayush Jain, P. K. Srijith, Mohammad Emtiyaz Khan
Deep Gaussian Processes (DGPs) are multi-layer, flexible extensions of Gaussian processes but their training remains challenging.
no code implementations • 9 Jul 2021 • Mohammad Emtiyaz Khan, Håvard Rue
The rule, derived from Bayesian principles, yields a wide-range of algorithms from fields such as optimization, deep learning, and graphical models.
1 code implementation • NeurIPS 2021 • Mohammad Emtiyaz Khan, Siddharth Swaroop
Humans and animals have a natural ability to quickly adapt to their surroundings, but machine-learning models, when subjected to changes, often require a complete retraining from scratch.
no code implementations • 4 Jun 2021 • Alexandre Piché, Valentin Thomas, Joseph Marino, Gian Maria Marconi, Christopher Pal, Mohammad Emtiyaz Khan
However, training is often unstable due to fast-changing target Q-values, and target networks are employed to regularize the Q-value estimation and stabilize training by using an additional set of lagging parameters.
1 code implementation • 11 Apr 2021 • Alexander Immer, Matthias Bauer, Vincent Fortuin, Gunnar Rätsch, Mohammad Emtiyaz Khan
Marginal-likelihood based model-selection, even though promising, is rarely used in deep learning due to estimation difficulties.
no code implementations • 15 Feb 2021 • Wu Lin, Frank Nielsen, Mohammad Emtiyaz Khan, Mark Schmidt
Natural-gradient descent (NGD) on structured parameter spaces (e. g., low-rank covariances) is computationally challenging due to difficult Fisher-matrix computations.
1 code implementation • 9 Jul 2020 • Paul E. Chang, William J. Wilkinson, Mohammad Emtiyaz Khan, Arno Solin
Gaussian process (GP) regression with 1D inputs can often be performed in linear time via a stochastic differential equation formulation.
1 code implementation • NeurIPS 2020 • Pingbo Pan, Siddharth Swaroop, Alexander Immer, Runa Eschenhagen, Richard E. Turner, Mohammad Emtiyaz Khan
Continually learning new skills is important for intelligent systems, yet standard deep learning methods suffer from catastrophic forgetting of the past.
4 code implementations • ICML 2020 • Xiangming Meng, Roman Bachmann, Mohammad Emtiyaz Khan
Our work provides a principled approach for training binary neural networks which justifies and extends existing approaches.
1 code implementation • ICML 2020 • Wu Lin, Mark Schmidt, Mohammad Emtiyaz Khan
The Bayesian learning rule is a natural-gradient variational inference method, which not only contains many existing learning algorithms as special cases but also enables the design of new algorithms.
1 code implementation • 29 Oct 2019 • Wu Lin, Mohammad Emtiyaz Khan, Mark Schmidt
Our generalization enables us to establish a connection between Stein's lemma and the reparamterization trick to derive gradients of expectations of a large class of functions under weak assumptions.
no code implementations • 15 Sep 2019 • Voot Tangkaratt, Bo Han, Mohammad Emtiyaz Khan, Masashi Sugiyama
However, the quality of demonstrations in reality can be diverse, since it is easier and cheaper to collect demonstrations from a mix of experts and amateurs.
1 code implementation • 7 Jun 2019 • Wu Lin, Mohammad Emtiyaz Khan, Mark Schmidt
Natural-gradient methods enable fast and simple algorithms for variational inference, but due to computational difficulties, their use is mostly limited to \emph{minimal} exponential-family (EF) approximations.
1 code implementation • NeurIPS 2019 • Kazuki Osawa, Siddharth Swaroop, Anirudh Jain, Runa Eschenhagen, Richard E. Turner, Rio Yokota, Mohammad Emtiyaz Khan
Importantly, the benefits of Bayesian principles are preserved: predictive probabilities are well-calibrated, uncertainties on out-of-distribution data are improved, and continual-learning performance is boosted.
1 code implementation • NeurIPS 2019 • Mohammad Emtiyaz Khan, Alexander Immer, Ehsan Abedi, Maciej Korzepa
Deep neural networks (DNN) and Gaussian processes (GP) are two powerful models with several theoretical connections relating them, but the relationship between their training methods is not well understood.
2 code implementations • 27 May 2019 • Jiaxin Shi, Mohammad Emtiyaz Khan, Jun Zhu
Inference in Gaussian process (GP) models is computationally challenging for large data, and often difficult to approximate with a small number of inducing points.
no code implementations • 8 Apr 2019 • Badr-Eddine Chérief-Abdellatif, Pierre Alquier, Mohammad Emtiyaz Khan
Our work in this paper presents theoretical justifications in favor of online algorithms relying on approximate Bayesian methods.
1 code implementation • 19 Dec 2018 • Simone Parisi, Voot Tangkaratt, Jan Peters, Mohammad Emtiyaz Khan
Actor-critic methods can achieve incredible performance on difficult reinforcement learning problems, but they are also prone to instability.
2 code implementations • NeurIPS 2018 • Aaron Mishkin, Frederik Kunstner, Didrik Nielsen, Mark Schmidt, Mohammad Emtiyaz Khan
Uncertainty estimation in large deep-learning models is a computationally challenging task, where it is difficult to form even a Gaussian approximation to the posterior distribution.
1 code implementation • 12 Jul 2018 • Mohammad Emtiyaz Khan, Didrik Nielsen
Bayesian inference plays an important role in advancing machine learning, but faces computational challenges when applied to complex models such as deep neural networks.
3 code implementations • ICML 2018 • Mohammad Emtiyaz Khan, Didrik Nielsen, Voot Tangkaratt, Wu Lin, Yarin Gal, Akash Srivastava
Uncertainty computation in deep learning is essential to design robust and reliable systems.
no code implementations • 22 May 2018 • Chao Li, Mohammad Emtiyaz Khan, Zhun Sun, Gang Niu, Bo Han, Shengli Xie, Qibin Zhao
Exact recovery of tensor decomposition (TD) methods is a desirable property in both unsupervised learning and scientific data analysis.
1 code implementation • ICLR 2018 • Wu Lin, Nicolas Hubacher, Mohammad Emtiyaz Khan
Recent efforts on combining deep models with probabilistic graphical models are promising in providing flexible models that are also easy to interpret.
no code implementations • 4 Dec 2017 • Mohammad Emtiyaz Khan, Zuozhu Liu, Voot Tangkaratt, Yarin Gal
Overall, this paper presents Vprop as a principled, computationally-efficient, and easy-to-implement method for Bayesian deep learning.
no code implementations • 15 Nov 2017 • Mohammad Emtiyaz Khan, Wu Lin, Voot Tangkaratt, Zuozhu Liu, Didrik Nielsen
We present the Variational Adaptive Newton (VAN) method which is a black-box optimization method especially suitable for explorative-learning tasks such as active learning and reinforcement learning.
1 code implementation • 19 May 2017 • Hongyi Ding, Mohammad Emtiyaz Khan, Issei Sato, Masashi Sugiyama
We model the intensity of each sequence as an infinite mixture of latent functions, each of which is obtained using a function drawn from a Gaussian process.
2 code implementations • 13 Mar 2017 • Mohammad Emtiyaz Khan, Wu Lin
In this paper, we propose a new algorithm called Conjugate-computation Variational Inference (CVI) which brings the best of the two worlds together -- it uses conjugate computations for the conjugate terms and employs stochastic gradients for the rest.
no code implementations • 31 Oct 2015 • Mohammad Emtiyaz Khan, Reza Babanezhad, Wu Lin, Mark Schmidt, Masashi Sugiyama
We also give a convergence-rate analysis of our method and many other previous methods which exploit the geometry of the space.
no code implementations • 13 Oct 2015 • Mattia Carpin, Stefano Rosati, Mohammad Emtiyaz Khan, Bixio Rimoldi
We address the problem of localizing non-collaborative WiFi devices in a large region.
no code implementations • 5 Jun 2013 • Mohammad Emtiyaz Khan, Aleksandr Y. Aravkin, Michael P. Friedlander, Matthias Seeger
Latent Gaussian models (LGMs) are widely used in statistics and machine learning.