Search Results for author: Madhu Advani

Found 8 papers, 1 papers with code

How JEPA Avoids Noisy Features: The Implicit Bias of Deep Linear Self Distillation Networks

no code implementations3 Jul 2024 Etai Littwin, Omid Saremi, Madhu Advani, Vimal Thilak, Preetum Nakkiran, Chen Huang, Joshua Susskind

A recent successful approach that falls under the JEPA framework is self-distillation, where an online encoder is trained to predict the output of the target encoder, sometimes using a lightweight predictor network.

Decoder Self-Supervised Learning

Step-by-Step Diffusion: An Elementary Tutorial

no code implementations13 Jun 2024 Preetum Nakkiran, Arwen Bradley, Hattie Zhou, Madhu Advani

We present an accessible first course on diffusion models and flow matching for machine learning, aimed at a technical audience with no diffusion experience.

A new role for circuit expansion for learning in neural networks

no code implementations19 Aug 2020 Julia Steinberg, Madhu Advani, Haim Sompolinsky

We find that sparse expansion of the input of a student perceptron network both increases its capacity and improves the generalization performance of the network when learning a noisy rule from a teacher perceptron when these expansions are pruned after learning.

Minnorm training: an algorithm for training over-parameterized deep neural networks

no code implementations3 Jun 2018 Yamini Bansal, Madhu Advani, David D. Cox, Andrew M. Saxe

To solve this constrained optimization problem, our method employs Lagrange multipliers that act as integrators of error over training and identify `support vector'-like examples.

Generalization Bounds

An equivalence between high dimensional Bayes optimal inference and M-estimation

no code implementations NeurIPS 2016 Madhu Advani, Surya Ganguli

In this work we demonstrate, when the signal distribution and the likelihood function associated with the noise are both log-concave, that optimal MMSE performance is asymptotically achievable via another M-estimation procedure.

Vocal Bursts Intensity Prediction

Statistical Mechanics of High-Dimensional Inference

no code implementations18 Jan 2016 Madhu Advani, Surya Ganguli

Our analysis uncovers fundamental limits on the accuracy of inference in high dimensions, and reveals that widely cherished inference algorithms like maximum likelihood (ML) and maximum-a posteriori (MAP) inference cannot achieve these limits.

Bayesian Inference Vocal Bursts Intensity Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.