no code implementations • 26 Nov 2023 • Cédric Gerbelot, Avetik Karagulyan, Stefani Karp, Kavya Ravichandran, Menachem Stern, Nathan Srebro
Although statistical learning theory provides a robust framework to understand supervised learning, many theoretical aspects of deep learning remain unclear, in particular how different architectures may lead to inductive bias when trained using gradient based methods.
no code implementations • 1 Nov 2023 • Sam Dillavou, Benjamin D Beyer, Menachem Stern, Andrea J Liu, Marc Z Miskin, Douglas J Durian
Standard deep learning algorithms require differentiating large nonlinear networks, a process that is slow and power-hungry.
no code implementations • 10 Jan 2022 • Jacob F Wycoff, Sam Dillavou, Menachem Stern, Andrea J Liu, Douglas J Durian
In a neuron network, synapses update individually using local information, allowing for entirely decentralized learning.