no code implementations • 24 Nov 2022 • Antonia Marcu
Over the past years, the crucial role of data has largely been shadowed by the field's focus on architectures and training procedures.
no code implementations • 15 Feb 2022 • Dominic Belcher, Antonia Marcu, Adam Prügel-Bennett
In this paper we show that the expected generalisation performance of a learning machine is determined by the distribution of risks or equivalently its logarithm -- a quantity we term the risk entropy -- and the fluctuations in a quantity we call the training ratio.
no code implementations • 22 Nov 2021 • Antonia Marcu, Adam Prügel-Bennett
The community lacks theory-informed guidelines for building good data sets.
1 code implementation • NeurIPS 2021 • Antonia Marcu, Adam Prügel-Bennett
Data distortion is commonly applied in vision models during both training (e. g methods like MixUp and CutMix) and evaluation (e. g. shape-texture bias and robustness).
5 code implementations • 27 Feb 2020 • Ethan Harris, Antonia Marcu, Matthew Painter, Mahesan Niranjan, Adam Prügel-Bennett, Jonathon Hare
Finally, we show that a consequence of the difference between interpolating MSDA such as MixUp and masking MSDA such as FMix is that the two can be combined to improve performance even further.
Ranked #3 on Image Classification on Fashion-MNIST
no code implementations • 11 Nov 2019 • Antonia Marcu, Adam Prügel-Bennett
In this paper, a new approach to computing the generalisation performance is presented that assumes the distribution of risks, $\rho(r)$, for a learning scenario is known.