no code implementations • 19 Aug 2020 • Julia Steinberg, Madhu Advani, Haim Sompolinsky
We find that sparse expansion of the input of a student perceptron network both increases its capacity and improves the generalization performance of the network when learning a noisy rule from a teacher perceptron when these expansions are pruned after learning.
no code implementations • 2 Jun 2019 • Stefano Recanatesi, Matthew Farrell, Madhu Advani, Timothy Moore, Guillaume Lajoie, Eric Shea-Brown
Datasets such as images, text, or movies are embedded in high-dimensional spaces.
no code implementations • 3 Jun 2018 • Yamini Bansal, Madhu Advani, David D. Cox, Andrew M. Saxe
To solve this constrained optimization problem, our method employs Lagrange multipliers that act as integrators of error over training and identify `support vector'-like examples.
1 code implementation • ICLR 2018 • Andrew Michael Saxe, Yamini Bansal, Joel Dapello, Madhu Advani, Artemy Kolchinsky, Brendan Daniel Tracey, David Daniel Cox
The practical successes of deep neural networks have not been matched by theoretical progress that satisfyingly explains their behavior.
no code implementations • NeurIPS 2016 • Madhu Advani, Surya Ganguli
In this work we demonstrate, when the signal distribution and the likelihood function associated with the noise are both log-concave, that optimal MMSE performance is asymptotically achievable via another M-estimation procedure.
no code implementations • 18 Jan 2016 • Madhu Advani, Surya Ganguli
Our analysis uncovers fundamental limits on the accuracy of inference in high dimensions, and reveals that widely cherished inference algorithms like maximum likelihood (ML) and maximum-a posteriori (MAP) inference cannot achieve these limits.