Search Results for author: François Charton

Found 12 papers, 6 papers with code

Salsa Fresca: Angular Embeddings and Pre-Training for ML Attacks on Learning With Errors

no code implementations2 Feb 2024 Samuel Stevens, Emily Wenger, Cathy Li, Niklas Nolte, Eshika Saxena, François Charton, Kristin Lauter

Our architecture improvements enable scaling to larger-dimension LWE problems: this work is the first instance of ML attacks recovering sparse binary secrets in dimension $n=1024$, the smallest dimension used in practice for homomorphic encryption applications of LWE where sparse binary secrets are proposed.

Math

Learning the greatest common divisor: explaining transformer predictions

1 code implementation29 Aug 2023 François Charton

Models trained from uniform operands only learn a handful of GCD (up to $38$ GCD $\leq100$).

Length Generalization in Arithmetic Transformers

no code implementations27 Jun 2023 Samy Jelassi, Stéphane d'Ascoli, Carles Domingo-Enrich, Yuhuai Wu, Yuanzhi Li, François Charton

We find that relative position embeddings enable length generalization for simple tasks, such as addition: models trained on $5$-digit numbers can perform $15$-digit sums.

Position

Contrastive Distillation Is a Sample-Efficient Self-Supervised Loss Policy for Transfer Learning

no code implementations21 Dec 2022 Chris Lengerich, Gabriel Synnaeve, Amy Zhang, Hugh Leather, Kurt Shuster, François Charton, Charysse Redwood

Traditional approaches to RL have focused on learning decision policies directly from episodic decisions, while slowly and implicitly learning the semantics of compositional representations needed for generalization.

Few-Shot Learning Language Modelling +2

What is my math transformer doing? -- Three results on interpretability and generalization

1 code implementation31 Oct 2022 François Charton

This paper investigates the failure cases and out-of-distribution behavior of transformers trained on matrix inversion and eigenvalue decomposition.

Math

SALSA: Attacking Lattice Cryptography with Transformers

no code implementations11 Jul 2022 Emily Wenger, Mingjie Chen, François Charton, Kristin Lauter

Currently deployed public-key cryptosystems will be vulnerable to attacks by full-scale quantum computers.

Cryptanalysis

End-to-end symbolic regression with transformers

3 code implementations22 Apr 2022 Pierre-Alexandre Kamienny, Stéphane d'Ascoli, Guillaume Lample, François Charton

Symbolic regression, the task of predicting the mathematical expression of a function from the observation of its values, is a difficult task which usually involves a two-step procedure: predicting the "skeleton" of the expression up to the choice of numerical constants, then fitting the constants by optimizing a non-convex loss function.

regression Symbolic Regression

Deep Symbolic Regression for Recurrent Sequences

no code implementations12 Jan 2022 Stéphane d'Ascoli, Pierre-Alexandre Kamienny, Guillaume Lample, François Charton

Symbolic regression, i. e. predicting a function from the observation of its values, is well-known to be a challenging task.

regression Symbolic Regression

A deep language model to predict metabolic network equilibria

no code implementations7 Dec 2021 François Charton, Amaury Hayat, Sean T. McQuade, Nathaniel J. Merrill, Benedetto Piccoli

We show that deep learning models, and especially architectures like the Transformer, originally intended for natural language, can be trained on randomly generated datasets to predict to very high accuracy both the qualitative and quantitative features of metabolic networks.

Language Modelling Machine Translation +1

Linear algebra with transformers

1 code implementation3 Dec 2021 François Charton

Transformers can learn to perform numerical computations from examples only.

Automated Theorem Proving Few-Shot Learning

Learning advanced mathematical computations from examples

1 code implementation ICLR 2021 François Charton, Amaury Hayat, Guillaume Lample

Using transformers over large generated datasets, we train models to learn mathematical properties of differential systems, such as local stability, behavior at infinity and controllability.

Deep Learning for Symbolic Mathematics

7 code implementations ICLR 2020 Guillaume Lample, François Charton

Neural networks have a reputation for being better at solving statistical or approximate problems than at performing calculations or working with symbolic data.

Cannot find the paper you are looking for? You can Submit a new open access paper.