Search Results for author: Robert C. Williamson

Found 34 papers, 4 papers with code

Geometry and Stability of Supervised Learning Problems

no code implementations4 Mar 2024 Facundo Mémoli, Brantley Vose, Robert C. Williamson

We introduce a notion of distance between supervised learning problems, which we call the Risk distance.

Four Facets of Forecast Felicity: Calibration, Predictiveness, Randomness and Regret

no code implementations25 Jan 2024 Rabanus Derr, Robert C. Williamson

Machine learning has traditionally focused on types of losses and their corresponding regret.

A General Framework for Learning under Corruption: Label Noise, Attribute Noise, and Beyond

no code implementations17 Jul 2023 Laura Iacovissi, Nan Lu, Robert C. Williamson

Corruption is frequently observed in collected data and has been extensively studied in machine learning under different corruption models.

Attribute

Insights From Insurance for Fair Machine Learning

no code implementations26 Jun 2023 Christian Fröhlich, Robert C. Williamson

Tracing the interaction of uncertainty, fairness and responsibility in insurance provides a fresh perspective on fairness in machine learning.

Fairness

The Geometry of Mixability

no code implementations23 Feb 2023 Armando J. Cabrera Pacheco, Robert C. Williamson

Mixable loss functions are of fundamental importance in the context of prediction with expert advice in the online setting since they characterize fast learning rates.

The Geometry and Calculus of Losses

no code implementations1 Sep 2022 Robert C. Williamson, Zac Cranko

In this paper we systematically develop the theory of loss functions for such problems from a novel perspective whose basic ingredients are convex sets with a particular structure.

Tailoring to the Tails: Risk Measures for Fine-Grained Tail Sensitivity

no code implementations5 Aug 2022 Christian Fröhlich, Robert C. Williamson

As a concrete example, we focus on divergence risk measures based on f-divergence ambiguity sets, which are a widespread tool used to foster distributional robustness of machine learning systems.

BIG-bench Machine Learning

Fairness and Randomness in Machine Learning: Statistical Independence and Relativization

no code implementations27 Jul 2022 Rabanus Derr, Robert C. Williamson

In this paper, we dissect the role of statistical independence in fairness and randomness notions regularly used in machine learning.

BIG-bench Machine Learning Fairness

Information Processing Equalities and the Information-Risk Bridge

no code implementations25 Jul 2022 Robert C. Williamson, Zac Cranko

We introduce two new classes of measures of information for statistical experiments which generalise and subsume $\phi$-divergences, integral probability metrics, $\mathfrak{N}$-distances (MMD), and $(f,\Gamma)$ divergences between two or more distributions.

Risk Measures and Upper Probabilities: Coherence and Stratification

no code implementations7 Jun 2022 Christian Fröhlich, Robert C. Williamson

Machine learning typically presupposes classical probability theory which implies that aggregation is built upon expectation.

BIG-bench Machine Learning

What killed the Convex Booster ?

no code implementations19 May 2022 Yishay Mansour, Richard Nock, Robert C. Williamson

A landmark negative result of Long and Servedio established a worst-case spectacular failure of a supervised learning trio (loss, algorithm, model) otherwise praised for its high precision machinery.

PAC-Bayesian Bound for the Conditional Value at Risk

no code implementations NeurIPS 2020 Zakaria Mhammedi, Benjamin Guedj, Robert C. Williamson

Conditional Value at Risk (CVaR) is a family of "coherent risk measures" which generalize the traditional mathematical expectation.

Fairness

A Primal-Dual link between GANs and Autoencoders

no code implementations NeurIPS 2019 Hisham Husain, Richard Nock, Robert C. Williamson

First, we find that the $f$-GAN and WAE objectives partake in a primal-dual relationship and are equivalent under some assumptions, which then allows us to explicate the success of WAE.

Generalization Bounds

Proper-Composite Loss Functions in Arbitrary Dimensions

no code implementations19 Feb 2019 Zac Cranko, Robert C. Williamson, Richard Nock

The study of a machine learning problem is in many ways is difficult to separate from the study of the loss function being used.

Density Estimation

Adversarial Networks and Autoencoders: The Primal-Dual Relationship and Generalization Bounds

no code implementations3 Feb 2019 Hisham Husain, Richard Nock, Robert C. Williamson

First, we find that the $f$-GAN and WAE objectives partake in a primal-dual relationship and are equivalent under some assumptions, which then allows us to explicate the success of WAE.

Generalization Bounds

Fairness risk measures

1 code implementation24 Jan 2019 Robert C. Williamson, Aditya Krishna Menon

In this paper, we propose a new definition of fairness that generalises some existing proposals, while allowing for generic sensitive features and resulting in a convex objective.

Fairness

Minimax Lower Bounds for Cost Sensitive Classification

no code implementations20 May 2018 Parameswaran Kamalaruban, Robert C. Williamson

The cost-sensitive classification problem plays a crucial role in mission-critical machine learning applications, and differs with traditional classification by taking the misclassification costs into consideration.

BIG-bench Machine Learning Binary Classification +2

Exp-Concavity of Proper Composite Losses

no code implementations20 May 2018 Parameswaran Kamalaruban, Robert C. Williamson, Xinhua Zhang

In special cases like the Aggregating Algorithm (\cite{vovk1995game}) with mixable losses and the Weighted Average Algorithm (\cite{kivinen1999averaging}) with exp-concave losses, it is possible to achieve $O(1)$ regret bounds.

Computational Efficiency

Constant Regret, Generalized Mixability, and Mirror Descent

no code implementations NeurIPS 2018 Zakaria Mhammedi, Robert C. Williamson

For a given entropy $\Phi$, losses for which a constant regret is possible using the \textsc{GAA} are called $\Phi$-mixable.

Open-Ended Question Answering

Provably Fair Representations

1 code implementation12 Oct 2017 Daniel McNamara, Cheng Soon Ong, Robert C. Williamson

These provable properties can be used in a governance model involving a data producer, a data user and a data regulator, where there is a separation of concerns between fairness and target task utility to ensure transparency and prevent perverse incentives.

BIG-bench Machine Learning Fairness

f-GANs in an Information Geometric Nutshell

1 code implementation NeurIPS 2017 Richard Nock, Zac Cranko, Aditya Krishna Menon, Lizhen Qu, Robert C. Williamson

In this paper, we unveil a broad class of distributions for which such convergence happens --- namely, deformed exponential families, a wide superset of exponential families --- and show tight connections with the three other key GAN parameters: loss, game and architecture.

The cost of fairness in classification

no code implementations25 May 2017 Aditya Krishna Menon, Robert C. Williamson

We study the problem of learning classifiers with a fairness constraint, with three main contributions towards the goal of quantifying the problem's inherent tradeoffs.

Classification Fairness +1

A Modular Theory of Feature Learning

no code implementations9 Nov 2016 Daniel McNamara, Cheng Soon Ong, Robert C. Williamson

We propose the idea of a risk gap induced by representation learning for a given prediction context, which measures the difference in the risk of some learner using the learned features as compared to the original inputs.

Representation Learning

Fast rates in statistical and online learning

no code implementations9 Jul 2015 Tim van Erven, Peter D. Grünwald, Nishant A. Mehta, Mark D. Reid, Robert C. Williamson

For bounded losses, we show how the central condition enables a direct proof of fast rates and we prove its equivalence to the Bernstein condition, itself a generalization of the Tsybakov margin condition, both of which have played a central role in obtaining fast rates in statistical learning.

Density Estimation Learning Theory

An Average Classification Algorithm

no code implementations4 Jun 2015 Brendan van Rooyen, Aditya Krishna Menon, Robert C. Williamson

When working with a high or infinite dimensional kernel, it is imperative for speed of evaluation and storage issues that as few training samples as possible are used in the kernel expansion.

Classification General Classification

Learning with Symmetric Label Noise: The Importance of Being Unhinged

1 code implementation NeurIPS 2015 Brendan van Rooyen, Aditya Krishna Menon, Robert C. Williamson

However, Long and Servedio [2010] proved that under symmetric label noise (SLN), minimisation of any convex potential over a linear function class can result in classification performance equivalent to random guessing.

Binary Classification Classification +1

A Theory of Feature Learning

no code implementations1 Apr 2015 Brendan van Rooyen, Robert C. Williamson

Feature Learning aims to extract relevant information contained in data sets in an automated fashion.

Learning in the Presence of Corruption

no code implementations1 Apr 2015 Brendan van Rooyen, Robert C. Williamson

In this paper we develop a general framework for tackling such problems as well as introducing upper and lower bounds on the risk for learning in the presence of corruption.

Generalized Mixability via Entropic Duality

no code implementations24 Jun 2014 Mark D. Reid, Rafael M. Frongillo, Robert C. Williamson, Nishant Mehta

Mixability is a property of a loss which characterizes when fast convergence is possible in the game of prediction with expert advice.

From Stochastic Mixability to Fast Rates

no code implementations NeurIPS 2014 Nishant A. Mehta, Robert C. Williamson

In the non-statistical prediction with expert advice setting, there is an analogous slow and fast rate phenomenon, and it is entirely characterized in terms of the mixability of the loss $\ell$ (there being no role there for $\mathcal{F}$ or $\mathsf{P}$).

Generalised Mixability, Constant Regret, and Bayesian Updating

no code implementations10 Mar 2014 Mark D. Reid, Rafael M. Frongillo, Robert C. Williamson

Mixability of a loss is known to characterise when constant regret bounds are achievable in games of prediction with expert advice through the use of Vovk's aggregating algorithm.

Le Cam meets LeCun: Deficiency and Generic Feature Learning

no code implementations20 Feb 2014 Brendan van Rooyen, Robert C. Williamson

"Deep Learning" methods attempt to learn generic features in an unsupervised fashion from a large unlabelled data set.

Mixability in Statistical Learning

no code implementations NeurIPS 2012 Tim V. Erven, Peter Grünwald, Mark D. Reid, Robert C. Williamson

We show that, in the special case of log-loss, stochastic mixability reduces to a well-known (but usually unnamed) martingale condition, which is used in existing convergence theorems for minimum description length and Bayesian inference.

Bayesian Inference

Composite Multiclass Losses

no code implementations NeurIPS 2011 Elodie Vernet, Mark D. Reid, Robert C. Williamson

We also show that the integral representation for binary proper losses can not be extended to multiclass losses.

General Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.