1 code implementation • 15 May 2024 • Takuo Matsubara

In probabilistic prediction, a parametric probability distribution is often specified on the space of output variables, and a point estimate of the output-distribution parameter is produced for each input by a model.

1 code implementation • 25 Jun 2023 • Takuo Matsubara, Niek Tax, Richard Mudd, Ido Guy

This paper proposes a new metric to measure the calibration error of probabilistic binary classifiers, called test-based calibration error (TCE).

1 code implementation • 16 Jun 2022 • Takuo Matsubara, Jeremias Knoblauch, François-Xavier Briol, Chris. J. Oates

Discrete state spaces represent a major computational challenge to statistical inference, since the computation of normalisation constants requires summation over large or possibly infinite sets, which can be impractical.

1 code implementation • 15 Apr 2021 • Takuo Matsubara, Jeremias Knoblauch, François-Xavier Briol, Chris. J. Oates

Generalised Bayesian inference updates prior beliefs using a loss function, rather than a likelihood, and can therefore be used to confer robustness against possible mis-specification of the likelihood.

1 code implementation • 16 Oct 2020 • Takuo Matsubara, Chris J. Oates, François-Xavier Briol

Our approach constructs a prior distribution for the parameters of the network, called a ridgelet prior, that approximates the posited Gaussian process in the output space of the network.

no code implementations • 19 May 2018 • Sho Sonoda, Isao Ishikawa, Masahiro Ikeda, Kei Hagihara, Yoshihiro Sawano, Takuo Matsubara, Noboru Murata

We prove that the global minimum of the backpropagation (BP) training problem of neural networks with an arbitrary nonlinear activation is given by the ridgelet transform.

Cannot find the paper you are looking for? You can
Submit a new open access paper.

Contact us on:
hello@paperswithcode.com
.
Papers With Code is a free resource with all data licensed under CC-BY-SA.