no code implementations • ICLR 2022 • Ruixiang Zhang, Shuangfei Zhai, Etai Littwin, Josh Susskind
We show that the low-rank approximation of NFKs derived from unsupervised generative models and supervised learning models gives rise to high-quality compact representations of data, achieving competitive results on a variety of machine learning tasks.
no code implementations • 2 Jul 2021 • Shih-Yu Sun, Vimal Thilak, Etai Littwin, Omid Saremi, Joshua M. Susskind
Deep linear networks trained with gradient descent yield low rank solutions, as is typically studied in matrix factorization.
no code implementations • 1 Jul 2021 • Etai Littwin, Omid Saremi, Shuangfei Zhai, Vimal Thilak, Hanlin Goh, Joshua M. Susskind, Greg Yang
We analyze the learning dynamics of infinitely wide neural networks with a finite sized bottle-neck.
no code implementations • 8 May 2021 • Greg Yang, Etai Littwin
To facilitate this proof, we develop a graphical notation for Tensor Programs.
no code implementations • NeurIPS 2020 • Etai Littwin, Ben Myara, Sima Sabah, Joshua Susskind, Shuangfei Zhai, Oren Golan
Modern neural network performance typically improves as model size increases.
1 code implementation • NeurIPS 2020 • Etai Littwin, Tomer Galanti, Lior Wolf, Greg Yang
{\em Hypernetworks} are architectures that produce the weights of a task-specific {\em primary network}.
no code implementations • 28 Jan 2020 • Etai Littwin, Tomer Galanti, Lior Wolf
We derive finite width and depth corrections for the Neural Tangent Kernel (NTK) of ResNets and DenseNets.
no code implementations • ICML Workshop Deep_Phenomen 2019 • Etai Littwin, Lior Wolf
The Hessian of neural networks can be decomposed into a sum of two matrices: (i) the positive semidefinite generalized Gauss-Newton matrix G, and (ii) the matrix H containing negative eigenvalues.
no code implementations • 25 Sep 2019 • Etai Littwin, Lior Wolf
A critical part of the training process of neural networks takes place in the very first gradient steps post initialization.
no code implementations • NeurIPS 2018 • Etai Littwin, Lior Wolf
Normalization techniques play an important role in supporting efficient and often more effective training of deep neural networks.
no code implementations • 8 Nov 2016 • Etai Littwin, Lior Wolf
Deep Residual Networks present a premium in performance in comparison to conventional networks of the same depth and are trainable at extreme depths.
no code implementations • CVPR 2016 • Etai Littwin, Lior Wolf
Deep learning techniques are renowned for supporting effective transfer learning.
no code implementations • CVPR 2015 • Etai Littwin, Hadar Averbuch-Elor, Daniel Cohen-Or
In this paper, we introduce a spherical embedding technique to position a given set of silhouettes of an object as observed from a set of cameras arbitrarily positioned around the object.