Search Results for author: Luca Arnaboldi

Found 12 papers, 6 papers with code

Online Learning and Information Exponents: On The Importance of Batch size, and Time/Complexity Tradeoffs

1 code implementation4 Jun 2024 Luca Arnaboldi, Yatin Dandi, Florent Krzakala, Bruno Loureiro, Luca Pesce, Ludovic Stephan

We study the impact of the batch size $n_b$ on the iteration time $T$ of training two-layer neural networks with one-pass stochastic gradient descent (SGD) on multi-index target functions of isotropic covariates.

Repetita Iuvant: Data Repetition Allows SGD to Learn High-Dimensional Multi-Index Functions

1 code implementation24 May 2024 Luca Arnaboldi, Yatin Dandi, Florent Krzakala, Luca Pesce, Ludovic Stephan

Here, we investigate the training dynamics of two-layer shallow neural networks trained with gradient-based algorithms, and discuss how they learn pertinent features in multi-index models, that is target functions with low-dimensional relevant directions.

Computational Efficiency

NLP Verification: Towards a General Methodology for Certifying Robustness

no code implementations15 Mar 2024 Marco Casadio, Tanvi Dinkar, Ekaterina Komendantskaya, Luca Arnaboldi, Matthew L. Daggitt, Omri Isac, Guy Katz, Verena Rieser, Oliver Lemon

We propose a number of practical NLP methods that can help to quantify the effects of the embedding gap; and in particular we propose the metric of falsifiability of semantic subspaces as another fundamental metric to be reported as part of the NLP verification pipeline.

The Benefits of Reusing Batches for Gradient Descent in Two-Layer Networks: Breaking the Curse of Information and Leap Exponents

1 code implementation5 Feb 2024 Yatin Dandi, Emanuele Troiani, Luca Arnaboldi, Luca Pesce, Lenka Zdeborová, Florent Krzakala

In particular, multi-pass GD with finite stepsize is found to overcome the limitations of gradient flow and single-pass GD given by the information exponent (Ben Arous et al., 2021) and leap exponent (Abbe et al., 2023) of the target function.

Vehicle: Bridging the Embedding Gap in the Verification of Neuro-Symbolic Programs

1 code implementation12 Jan 2024 Matthew L. Daggitt, Wen Kokke, Robert Atkey, Natalia Slusarz, Luca Arnaboldi, Ekaterina Komendantskaya

Neuro-symbolic programs -- programs containing both machine learning components and traditional symbolic code -- are becoming increasingly widespread.

Escaping mediocrity: how two-layer networks learn hard generalized linear models with SGD

2 code implementations29 May 2023 Luca Arnaboldi, Florent Krzakala, Bruno Loureiro, Ludovic Stephan

These insights are grounded in the reduction of SGD dynamics to a stochastic process in lower dimensions, where escaping mediocrity equates to calculating an exit time.

ANTONIO: Towards a Systematic Method of Generating NLP Benchmarks for Verification

no code implementations6 May 2023 Marco Casadio, Luca Arnaboldi, Matthew L. Daggitt, Omri Isac, Tanvi Dinkar, Daniel Kienitz, Verena Rieser, Ekaterina Komendantskaya

In particular, many known neural network verification methods that work for computer vision and other numeric datasets do not work for NLP.

From high-dimensional & mean-field dynamics to dimensionless ODEs: A unifying approach to SGD in two-layers networks

1 code implementation12 Feb 2023 Luca Arnaboldi, Ludovic Stephan, Florent Krzakala, Bruno Loureiro

This manuscript investigates the one-pass stochastic gradient descent (SGD) dynamics of a two-layer neural network trained on Gaussian data and labels generated by a similar, though not necessarily identical, target function.

Why Robust Natural Language Understanding is a Challenge

no code implementations21 Jun 2022 Marco Casadio, Ekaterina Komendantskaya, Verena Rieser, Matthew L. Daggitt, Daniel Kienitz, Luca Arnaboldi, Wen Kokke

With the proliferation of Deep Machine Learning into real-life applications, a particular property of this technology has been brought to attention: robustness Neural Networks notoriously present low robustness and can be highly sensitive to small input perturbations.

Natural Language Understanding

Vehicle: Interfacing Neural Network Verifiers with Interactive Theorem Provers

no code implementations10 Feb 2022 Matthew L. Daggitt, Wen Kokke, Robert Atkey, Luca Arnaboldi, Ekaterina Komendantskya

However, although work has managed to incorporate the results of these verifiers to prove larger properties of individual systems, there is currently no general methodology for bridging the gap between verifiers and interactive theorem provers (ITPs).

Automated Theorem Proving

A Generic Methodology for the Statistically Uniform & Comparable Evaluation of Automated Trading Platform Components

no code implementations21 Sep 2020 Artur Sokolovsky, Luca Arnaboldi

This work is one of the first in applying this rigorous statistically-backed approach to the field of financial markets and we hope this may be a springboard for more research.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.