1 code implementation • NeurIPS 2019 • Cory Stephenson, Jenelle Feather, Suchismita Padhy, Oguz Elibol, Hanlin Tang, Josh Mcdermott, SueYeon Chung
Higher level concepts such as parts-of-speech and context dependence also emerge in the later layers of the network.
no code implementations • ICLR Workshop LLD 2019 • Tyler Lee, Ting Gong, Suchismita Padhy, Andrew Rouditchenko, Anthony Ndirango
We demonstrate that, in scenarios with limited labeled training data, one can significantly improve the performance of three different supervised classification tasks individually by up to 6% through simultaneous training with these additional self-supervised tasks.
no code implementations • ICLR 2021 • Cory Stephenson, Suchismita Padhy, Abhinav Ganesh, Yue Hui, Hanlin Tang, SueYeon Chung
Understanding how large neural networks avoid memorizing training data is key to explaining their high generalization performance.
no code implementations • 28 May 2019 • Suchismita Padhy, Jenelle Feather, Cory Stephenson, Oguz Elibol, Hanlin Tang, Josh Mcdermott, SueYeon Chung
The success of deep neural networks in visual tasks have motivated recent theoretical and empirical work to understand how these networks operate.