no code implementations • 5 Mar 2025 • Devon Jarvis, Verena Klar, Richard Klein, Benjamin Rosman, Andrew Saxe
While relearning of lost knowledge has been shown in acute brain injuries such as stroke, it has not been widely supported in chronic cognitive diseases such as SD.
no code implementations • 27 Jan 2025 • Yedi Zhang, Aaditya K. Singh, Peter E. Latham, Andrew Saxe
For the merged parametrization, we show the training dynamics has two fixed points and the loss trajectory exhibits a single, abrupt drop.
no code implementations • 25 Jun 2024 • Jirko Rubruck, Jan P. Bauer, Andrew Saxe, Christopher Summerfield
We identify hallmarks of this early OCS phase and illustrate how these signatures are observed in deep linear networks and larger, more complex (and nonlinear) convolutional neural networks solving a hierarchical learning task based on MNIST and CIFAR10.
no code implementations • 18 Jun 2024 • Yedi Zhang, Andrew Saxe, Peter E. Latham
We then show that, under symmetry conditions on the data, these networks have the same learning dynamics as linear networks.
1 code implementation • 10 Jun 2024 • Daniel Kunin, Allan Raventós, Clémentine Dominé, Feng Chen, David Klindt, Andrew Saxe, Surya Ganguli
While the impressive performance of modern neural networks is often attributed to their capacity to efficiently extract task-relevant features from data, the mechanisms underlying this rich feature learning regime remain elusive, with much of our theoretical understanding stemming from the opposing lazy regime.
1 code implementation • 3 Jun 2024 • Stefano Sarao Mannelli, Yaraslau Ivashynka, Andrew Saxe, Luca Saglietti
A wide range of empirical and theoretical works have shown that overparameterisation can amplify the performance of neural networks.
no code implementations • 28 Feb 2024 • Jin Hwa Lee, Stefano Sarao Mannelli, Andrew Saxe
Diverse studies in systems neuroscience begin with extended periods of curriculum training known as `shaping' procedures.
1 code implementation • 1 Dec 2023 • Yedi Zhang, Peter E. Latham, Andrew Saxe
This is the first work to calculate the duration of the unimodal phase in learning as a function of the depth at which modalities are fused within the network, dataset statistics, and initialization.
no code implementations • 29 Jun 2023 • Timo Flesch, Valerio Mante, William Newsome, Andrew Saxe, Christopher Summerfield, David Sussillo
A recent paper (Flesch et al, 2022) describes behavioural and neural data suggesting that task representations are gated in the prefrontal cortex in both humans and macaques.
no code implementations • 17 Jun 2023 • Nishil Patel, Sebastian Lee, Stefano Sarao Mannelli, Sebastian Goldt, Andrew Saxe
Reinforcement learning (RL) algorithms have proven transformative in a range of domains.
no code implementations • 10 Oct 2022 • Timo Flesch, Andrew Saxe, Christopher Summerfield
How do humans and other animals learn new tasks?
no code implementations • 16 Jun 2022 • Aaditya K. Singh, David Ding, Andrew Saxe, Felix Hill, Andrew K. Lampinen
Through controlled experiments, we show that training a speaker with two listeners that perceive differently, using our method, allows the speaker to adapt to the idiosyncracies of the listeners.
1 code implementation • 18 May 2022 • Sebastian Lee, Stefano Sarao Mannelli, Claudia Clopath, Sebastian Goldt, Andrew Saxe
Continual learning - learning new tasks in sequence while maintaining performance on old tasks - remains particularly challenging for artificial neural networks.
1 code implementation • 22 Mar 2022 • Timo Flesch, David G. Nagy, Andrew Saxe, Christopher Summerfield
Here, we propose novel computational constraints for artificial neural networks, inspired by earlier work on gating in the primate prefrontal cortex, that capture the cost of interleaved training and allow the network to learn two tasks in sequence without forgetting.
1 code implementation • 9 Jul 2021 • Sebastian Lee, Sebastian Goldt, Andrew Saxe
Using each teacher to represent a different task, we investigate how the relationship between teachers affects the amount of forgetting and transfer exhibited by the student when the task switches.
no code implementations • 15 Jun 2021 • Luca Saglietti, Stefano Sarao Mannelli, Andrew Saxe
To study the former, we provide an exact description of the online learning setting, confirming the long-standing experimental observation that curricula can modestly speed up learning.
no code implementations • 9 Jun 2021 • Federica Gerace, Luca Saglietti, Stefano Sarao Mannelli, Andrew Saxe, Lenka Zdeborová
Transfer learning can significantly improve the sample efficiency of neural networks, by exploiting the relatedness between a data-scarce target task and a data-abundant source task.
no code implementations • NeurIPS 2020 • Yinan Cao, Christopher Summerfield, Andrew Saxe
Studies suggesting that representations in deep networks resemble those in biological brains have mostly relied on one specific learning rule: gradient descent, the workhorse behind modern deep learning.
no code implementations • 16 Apr 2020 • Andrew Saxe, Stephanie Nelli, Christopher Summerfield
In this Perspective, our goal is to offer a roadmap for systems neuroscience research in the age of deep learning.
Neurons and Cognition
no code implementations • 17 Jul 2018 • Maxwell Nye, Andrew Saxe
Specifically, we train deep neural networks to learn two simple functions with known efficient solutions: the parity function and the fast Fourier transform.
1 code implementation • NeurIPS 2016 • Chuan-Yung Tsai, Andrew Saxe, David Cox
We present a novel neural network algorithm, the Tensor Switching (TS) network, which generalizes the Rectified Linear Unit (ReLU) nonlinearity to tensor-valued hidden units.
no code implementations • NeurIPS 2011 • Maneesh Bhand, Ritvik Mudur, Bipin Suresh, Andrew Saxe, Andrew Y. Ng
In this work we focus on that component of adaptation which occurs during an organism's lifetime, and show that a number of unsupervised feature learning algorithms can account for features of normal receptive field properties across multiple primary sensory cortices.
no code implementations • NeurIPS 2009 • Ian Goodfellow, Honglak Lee, Quoc V. Le, Andrew Saxe, Andrew Y. Ng
Our evaluation metrics can also be used to evaluate future work in unsupervised deep learning, and thus help the development of future algorithms.