no code implementations • 3 Jul 2024 • Etai Littwin, Omid Saremi, Madhu Advani, Vimal Thilak, Preetum Nakkiran, Chen Huang, Joshua Susskind
A recent successful approach that falls under the JEPA framework is self-distillation, where an online encoder is trained to predict the output of the target encoder, sometimes using a lightweight predictor network.
no code implementations • 13 Jun 2024 • Preetum Nakkiran, Arwen Bradley, Hattie Zhou, Madhu Advani
We present an accessible first course on diffusion models and flow matching for machine learning, aimed at a technical audience with no diffusion experience.
no code implementations • 19 Aug 2020 • Julia Steinberg, Madhu Advani, Haim Sompolinsky
We find that sparse expansion of the input of a student perceptron network both increases its capacity and improves the generalization performance of the network when learning a noisy rule from a teacher perceptron when these expansions are pruned after learning.
no code implementations • 2 Jun 2019 • Stefano Recanatesi, Matthew Farrell, Madhu Advani, Timothy Moore, Guillaume Lajoie, Eric Shea-Brown
Datasets such as images, text, or movies are embedded in high-dimensional spaces.
no code implementations • 3 Jun 2018 • Yamini Bansal, Madhu Advani, David D. Cox, Andrew M. Saxe
To solve this constrained optimization problem, our method employs Lagrange multipliers that act as integrators of error over training and identify `support vector'-like examples.
1 code implementation • ICLR 2018 • Andrew Michael Saxe, Yamini Bansal, Joel Dapello, Madhu Advani, Artemy Kolchinsky, Brendan Daniel Tracey, David Daniel Cox
The practical successes of deep neural networks have not been matched by theoretical progress that satisfyingly explains their behavior.
no code implementations • NeurIPS 2016 • Madhu Advani, Surya Ganguli
In this work we demonstrate, when the signal distribution and the likelihood function associated with the noise are both log-concave, that optimal MMSE performance is asymptotically achievable via another M-estimation procedure.
no code implementations • 18 Jan 2016 • Madhu Advani, Surya Ganguli
Our analysis uncovers fundamental limits on the accuracy of inference in high dimensions, and reveals that widely cherished inference algorithms like maximum likelihood (ML) and maximum-a posteriori (MAP) inference cannot achieve these limits.