no code implementations • 1 Sep 2022 • Robert C. Williamson, Zac Cranko
In this paper we systematically develop the theory of loss functions for such problems from a novel perspective whose basic ingredients are convex sets with a particular structure.
no code implementations • 25 Jul 2022 • Robert C. Williamson, Zac Cranko
We introduce two new classes of measures of information for statistical experiments which generalise and subsume $\phi$-divergences, integral probability metrics, $\mathfrak{N}$-distances (MMD), and $(f,\Gamma)$ divergences between two or more distributions.
no code implementations • 29 Sep 2021 • Dylan M. Paiton, David Schultheiss, Matthias Kuemmerer, Zac Cranko, Matthias Bethge
We undertake analysis to characterize the geometry of the boundary, which is more curved within the adversarial subspace than within a random subspace of equal dimensionality.
no code implementations • 11 Feb 2020 • Zac Cranko, Zhan Shi, Xinhua Zhang, Richard Nock, Simon Kornblith
The problem of adversarial examples has highlighted the need for a theory of regularisation that is general enough to apply to exotic function classes, such as universal approximators.
no code implementations • 25 Sep 2019 • Zac Cranko, Zhan Shi, Xinhua Zhang, Simon Kornblith, Richard Nock
Distributional robust risk (DRR) minimisation has arisen as a flexible and effective framework for machine learning.
no code implementations • 19 Feb 2019 • Zac Cranko, Robert C. Williamson, Richard Nock
The study of a machine learning problem is in many ways is difficult to separate from the study of the loss function being used.
no code implementations • 4 Sep 2018 • Zac Cranko, Simon Kornblith, Zhan Shi, Richard Nock
Robust risk minimisation has several advantages: it has been studied with regards to improving the generalisation properties of models and robustness to adversarial perturbation.
1 code implementation • 13 Jun 2018 • Hisham Husain, Zac Cranko, Richard Nock
Privacy enforces an information theoretic barrier on approximation, and we show how to reach this barrier with guarantees on the approximation of the target non private density.
no code implementations • 8 Jun 2018 • Zac Cranko, Aditya Krishna Menon, Richard Nock, Cheng Soon Ong, Zhan Shi, Christian Walder
A key feature of our result is that it holds for all proper losses, and for a popular subset of these, the optimisation of this central measure appears to be independent of the loss.
no code implementations • 22 Mar 2018 • Zac Cranko, Richard Nock
There has recently been a steady increase in the number iterative approaches to density estimation.
1 code implementation • NeurIPS 2017 • Richard Nock, Zac Cranko, Aditya Krishna Menon, Lizhen Qu, Robert C. Williamson
In this paper, we unveil a broad class of distributions for which such convergence happens --- namely, deformed exponential families, a wide superset of exponential families --- and show tight connections with the three other key GAN parameters: loss, game and architecture.