no code implementations • 2 Sep 2022 • Dang Trung Kien, Neo Han Wei, Sanjay Chaudhuri
In this article, we describe a {\tt R} package for sampling from an empirical likelihood-based posterior using a Hamiltonian Monte Carlo method.
no code implementations • 2 Sep 2022 • Sanjay Chaudhuri, Teng Yin
Furthermore, we discuss Bayesian model selection using empirical likelihood and extend our two-step Metropolis Hastings algorithm to a reversible jump Markov chain Monte Carlo procedure to sample from the resulting posterior.
no code implementations • 12 Nov 2020 • Sanjay Chaudhuri, Subhroshekhar Ghosh, David J. Nott, Kim Cuc Pham
The expected log-likelihood is then estimated by an empirical likelihood where the only inputs required are a choice of summary statistic, it's observed value, and the ability to simulate the chosen summary statistics for any parameter value under the model.
no code implementations • 3 Oct 2019 • Subhro Ghosh, Sanjay Chaudhuri
In the Bayesian setting, we rigorously establish the posterior consistency of procedures based on these ideas, where instead of a parametric likelihood, an empirical likelihood is used to define the posterior distribution.
no code implementations • 3 Oct 2018 • Sanjay Chaudhuri, Subhro Ghosh, David J. Nott, Kim Cuc Pham
Many scientifically well-motivated statistical models in natural, engineering and environmental sciences are specified through a generative process, but in some cases it may not be possible to write down a likelihood for these models analytically.
no code implementations • 12 Mar 2015 • Sanjay Chaudhuri
Rules for comparing degree of association among the vertices of such Gaussian graphical models are also developed.