no code implementations • 14 Oct 2020 • Onur Teymur, Jackson Gorham, Marina Riabiz, Chris. J. Oates
Several researchers have proposed minimisation of maximum mean discrepancy (MMD) as a method to quantise probability measures, i. e., to approximate a target distribution by a representative point set.
1 code implementation • NeurIPS 2020 • Jackson Gorham, Anant Raj, Lester Mackey
Stein discrepancies (SDs) monitor convergence and non-convergence in approximate inference when exact integration and sampling are intractable.
1 code implementation • 9 May 2019 • Wilson Ye Chen, Alessandro Barp, François-Xavier Briol, Jackson Gorham, Mark Girolami, Lester Mackey, Chris. J. Oates
Stein Points are a class of algorithms for this task, which proceed by sequentially minimising a Stein discrepancy between the empirical measure and the target and, hence, require the solution of a non-convex optimisation problem to obtain each new point.
1 code implementation • ICML 2018 • Wilson Ye Chen, Lester Mackey, Jackson Gorham, François-Xavier Briol, Chris. J. Oates
An important task in computational statistics and machine learning is to approximate a posterior distribution $p(x)$ with an empirical measure supported on a set of representative points $\{x_i\}_{i=1}^n$.
no code implementations • ICML 2017 • Jackson Gorham, Lester Mackey
We develop a theory of weak convergence for KSDs based on Stein's method, demonstrate that commonly used KSDs fail to detect non-convergence even for Gaussian targets, and show that kernels with slowly decaying tails provably determine convergence for a large class of target distributions.
no code implementations • 21 Nov 2016 • Jackson Gorham, Andrew B. Duncan, Sebastian J. Vollmer, Lester Mackey
Stein's method for measuring convergence to a continuous target distribution relies on an operator characterizing the target and Stein factor bounds on the solutions of an associated differential equation.
no code implementations • NeurIPS 2015 • Jackson Gorham, Lester Mackey
To improve the efficiency of Monte Carlo estimation, practitioners are turning to biased Markov chain Monte Carlo procedures that trade off asymptotic exactness for computational speed.