no code implementations • ICML 2020 • Ugo Tanielian, Thibaut Issenhuth, Elvis Dohmatob, Jeremie Mary
Typical architectures of Generative Adversarial Networks make use of a unimodal latent/input distribution transformed by a continuous generator.
1 code implementation • 10 Dec 2024 • Reza Bayat, Mohammad Pezeshki, Elvis Dohmatob, David Lopez-Paz, Pascal Vincent
Neural networks often learn simple explanations that fit the majority of the data while memorizing exceptions that deviate from these explanations. This behavior leads to poor generalization when the learned explanations rely on spurious correlations.
no code implementations • 7 Oct 2024 • Elvis Dohmatob, Yunzhen Feng, Arjun Subramonian, Julia Kempe
Within the scaling laws paradigm, which underpins the training of large neural networks like ChatGPT and Llama, we consider a supervised regression setting and establish the existance of a strong form of the model collapse phenomenon, a critical performance degradation due to synthetic data in the training corpus.
no code implementations • 7 Oct 2024 • Arjun Subramonian, Samuel J. Bell, Levent Sagun, Elvis Dohmatob
Machine learning models may capture and amplify biases present in data, leading to disparate test performance across social groups.
no code implementations • 11 Jun 2024 • Yunzhen Feng, Elvis Dohmatob, Pu Yang, Francois Charton, Julia Kempe
Large Language Models (LLM) are increasingly trained on data generated by other LLM, either because generated text and images become part of the pre-training corpus, or because synthetized data is used as a replacement for expensive human-annotation.
no code implementations • 12 Feb 2024 • Elvis Dohmatob, Yunzhen Feng, Julia Kempe
In the era of proliferation of large language and image generation models, the phenomenon of "model collapse" refers to the situation whereby as a model is trained recursively on data generated from previous generations of itself over time, its performance degrades until the model eventually becomes completely useless, i. e the model collapses.
no code implementations • 10 Feb 2024 • Elvis Dohmatob, Yunzhen Feng, Pu Yang, Francois Charton, Julia Kempe
We discover a wide range of decay phenomena, analyzing loss of scaling, shifted scaling with number of generations, the ''un-learning" of skills, and grokking when mixing human and synthesized data.
no code implementations • 4 Oct 2023 • Vivien Cabannes, Elvis Dohmatob, Alberto Bietti
Learning arguably involves the discovery and memorization of abstract rules.
no code implementations • 1 Aug 2023 • Elvis Dohmatob, Meyer Scetbon
In this paper, we investigate the impact of test-time adversarial attacks on linear regression models and determine the optimal level of robustness that any model can reach while maintaining a given level of standard predictive performance (accuracy).
no code implementations • 31 Jan 2023 • Meyer Scetbon, Elvis Dohmatob
However, we show that this strategy can be arbitrarily sub-optimal in the case of general Mahalanobis attacks.
1 code implementation • 4 Nov 2022 • Morgane Goibert, Thomas Ricatte, Elvis Dohmatob
In this paper, we investigate the impact of neural networks (NNs) topology on adversarial robustness.
no code implementations • 18 Oct 2022 • Virginie Do, Elvis Dohmatob, Matteo Pirotta, Alessandro Lazaric, Nicolas Usunier
We consider Contextual Bandits with Concave Rewards (CBCR), a multi-objective bandit problem where the desired trade-off between the rewards is defined by a known concave objective function, and the reward vector depends on an observed stochastic context.
no code implementations • 13 Sep 2022 • Nicolas Usunier, Virginie Do, Elvis Dohmatob
In this paper, we propose the first efficient online algorithm to optimize concave objective functions in the space of rankings which applies to every concave and smooth objective function, such as the ones found for fairness of exposure.
1 code implementation • 1 Jul 2022 • Insu Han, Mike Gartrell, Elvis Dohmatob, Amin Karbasi
In this work, we develop a scalable MCMC sampling algorithm for $k$-NDPPs with low-rank kernels, thus enabling runtime that is sublinear in $n$.
no code implementations • 25 Mar 2022 • Elvis Dohmatob, Chuan Guo, Morgane Goibert
Finally, we show that if a decision-region is compact, then it admits a universal adversarial perturbation with $L_2$ norm which is $\sqrt{d}$ times smaller than the typical $L_2$ norm of a data point.
no code implementations • 22 Mar 2022 • Elvis Dohmatob, Alberto Bietti
To better understand these factors, we provide a precise study of the adversarial robustness in different scenarios, from initialization to the end of training in different regimes, as well as intermediate scenarios, where initialization still plays a role due to "lazy" training.
2 code implementations • ICLR 2022 • Insu Han, Mike Gartrell, Jennifer Gillenwater, Elvis Dohmatob, Amin Karbasi
However, existing work leaves open the question of scalable NDPP sampling.
1 code implementation • 4 Jun 2021 • Elvis Dohmatob
More precisely, if $n$ is the number of training examples, $d$ is the input dimension, and $k$ is the number of hidden neurons in a two-layer neural network, we prove for a large class of activation functions that, if the model memorizes even a fraction of the training, then its Sobolev-seminorm is lower-bounded by (i) $\sqrt{n}$ in case of infinite-width random features (RF) or neural tangent kernel (NTK) with $d \gtrsim n$; (ii) $\sqrt{n}$ in case of finite-width RF with proportionate scaling of $d$ and $k$; and (iii) $\sqrt{n/k}$ in case of finite-width NTK with proportionate scaling of $d$ and $k$.
no code implementations • NeurIPS 2020 • Elena Smirnova, Elvis Dohmatob
Entropy regularization, smoothing of Q-values and neural network function approximator are key components of the state-of-the-art reinforcement learning (RL) algorithms, such as Soft Actor-Critic~\cite{haarnoja2018soft}.
no code implementations • 12 Nov 2020 • Elvis Dohmatob
We measure the quality of such a hyperplane by its margin $\gamma(w)$, defined as minimum distance between any of the points $x_i$ and the hyperplane.
2 code implementations • ICLR 2021 • Mike Gartrell, Insu Han, Elvis Dohmatob, Jennifer Gillenwater, Victor-Emmanuel Brunel
Determinantal point processes (DPPs) have attracted significant attention in machine learning for their ability to model subsets drawn from a large item collection.
no code implementations • 17 Jun 2020 • Elvis Dohmatob
(1) We use optimal transport theory to derive variational formulae for the Bayes-optimal error a classifier can make on a given classification problem, subject to adversarial attacks.
no code implementations • 8 Jun 2020 • Ugo Tanielian, Thibaut Issenhuth, Elvis Dohmatob, Jeremie Mary
Typical architectures of Generative AdversarialNetworks make use of a unimodal latent distribution transformed by a continuous generator.
no code implementations • 20 Sep 2019 • Elena Smirnova, Elvis Dohmatob
Entropy regularized algorithms such as Soft Q-learning and Soft Actor-Critic, recently showed state-of-the-art performance on a number of challenging reinforcement learning (RL) tasks.
no code implementations • 27 Jun 2019 • Morgane Goibert, Elvis Dohmatob
We study Label-Smoothing as a means for improving adversarial robustness of supervised deep-learning models.
no code implementations • 14 Jun 2019 • Louis Faury, Ugo Tanielian, Flavian vasile, Elena Smirnova, Elvis Dohmatob
This manuscript introduces the idea of using Distributionally Robust Optimization (DRO) for the Counterfactual Risk Minimization (CRM) problem.
1 code implementation • NeurIPS 2019 • Mike Gartrell, Victor-Emmanuel Brunel, Elvis Dohmatob, Syrine Krichene
Our method imposes a particular decomposition of the nonsymmetric kernel that enables such tractable learning algorithms, which we analyze both theoretically and experimentally.
no code implementations • 23 Feb 2019 • Elena Smirnova, Elvis Dohmatob, Jérémie Mary
Our formulation results in a efficient algorithm that accounts for a simple re-weighting of policy actions in the standard policy iteration scheme.
no code implementations • 17 Nov 2018 • Mike Gartrell, Elvis Dohmatob, Jon Alberdi
While DPPs have substantial expressive power, they are fundamentally limited by the parameterization of the kernel matrix and their inability to capture nonlinear interactions between items within sets.
no code implementations • 8 Oct 2018 • Elvis Dohmatob
This manuscript presents some new impossibility results on adversarial robustness in machine learning, a very important yet largely open problem.
no code implementations • NeurIPS 2016 • Elvis Dohmatob, Arthur Mensch, Gael Varoquaux, Bertrand Thirion
We propose a multivariate online dictionary-learning method for obtaining decompositions of brain images with structured and sparse components (aka atoms).
no code implementations • 31 May 2016 • Fouad Hadj-Selem, Tommy Lofstedt, Elvis Dohmatob, Vincent Frouin, Mathieu Dubois, Vincent Guillemot, Edouard Duchesnay
Nesterov's smoothing technique can be used to minimize a large number of non-smooth convex structured penalties but reasonable precision requires a small smoothing parameter, which slows down the convergence speed.
no code implementations • 22 Dec 2015 • Gaël Varoquaux, Michael Eickenberg, Elvis Dohmatob, Bertand Thirion
The total variation (TV) penalty, as many other analysis-sparsity problems, does not lead to separable factors or a proximal operatorwith a closed-form expression, such as soft thresholding for the $\ell\_1$ penalty.
no code implementations • 12 Dec 2014 • Alexandre Abraham, Elvis Dohmatob, Bertrand Thirion, Dimitris Samaras, Gael Varoquaux
Functional Magnetic Resonance Images acquired during resting-state provide information about the functional organization of the brain through measuring correlations between brain areas.