no code implementations • ICML 2020 • Veronika Rockova
There has been a growing realization of the potential of Bayesian machine learning as a platform that can provide both flexible modeling, accurate predictions as well as coherent uncertainty statements.
no code implementations • 16 Apr 2024 • Sean O'Hagan, Jungeum Kim, Veronika Rockova
Finally, we successfully apply our approach to the problem of masked image classification using deep generative models.
no code implementations • 8 Dec 2023 • Jungeum Kim, Veronika Rockova
After training, our deep learning approach enables rapid evaluations of the Bayes factor estimator at any fictional data arriving from either hypothesized model, not just the observed data $Y_0$.
1 code implementation • 26 Oct 2023 • Jiguang Li, Robert Gibbons, Veronika Rockova
In our simulation study, we show that our method reliably recovers both the factor dimensionality as well as the latent structure on high-dimensional synthetic data even for small samples.
no code implementations • 31 May 2023 • Jungeum Kim, Veronika Rockova
We show polynomial mixing of Twiggy Bayesian CART without assuming that the signal is connected on a tree.
no code implementations • 1 Jul 2020 • Yi Liu, Veronika Rockova
Thompson sampling is a heuristic algorithm for the multi-armed bandit problem which has a long tradition in machine learning.
BIG-bench Machine Learning Interpretable Machine Learning +3
3 code implementations • 14 Sep 2019 • Wei Jiang, Malgorzata Bogdan, Julie Josse, Blazej Miasojedow, Veronika Rockova, Traumabase group
We consider the problem of variable selection in high-dimensional settings with missing observations among the covariates.
Methodology Applications Computation
no code implementations • 1 Oct 2018 • Veronika Rockova, Enakshi Saha
Laying the foundations for the theoretical analysis of Bayesian forests, Rockova and van der Pas (2017) showed optimal posterior concentration under conditionally uniform tree priors.
no code implementations • NeurIPS 2018 • Nicholas Polson, Veronika Rockova
As an aside, we show that SS-DL does not overfit in the sense that the posterior concentrates on smaller networks with fewer (up to the optimal number of) nodes and links.
1 code implementation • 9 Jan 2018 • Gemma E. Moran, Veronika Rockova, Edward I. George
In a similar way, we show that conjugate priors for linear regression, which induce prior dependence, can lead to such underestimation in the Bayesian high-dimensional regression setting.
Methodology
1 code implementation • 29 Aug 2017 • Sameer K. Deshpande, Veronika Rockova, Edward I. George
We propose a Bayesian procedure for simultaneous variable and covariance selection using continuous spike-and-slab priors in multivariate linear regression models where q possibly correlated responses are regressed onto p predictors.
Methodology