Search Results for author: Blake Richards

Found 15 papers, 5 papers with code

Mixture-of-Depths: Dynamically allocating compute in transformer-based language models

1 code implementation2 Apr 2024 David Raposo, Sam Ritter, Blake Richards, Timothy Lillicrap, Peter Conway Humphreys, Adam Santoro

Our method enforces a total compute budget by capping the number of tokens ($k$) that can participate in the self-attention and MLP computations at a given layer.

Addressing Sample Inefficiency in Multi-View Representation Learning

no code implementations17 Dec 2023 Kumar Krishna Agrawal, Arna Ghosh, Adam Oberman, Blake Richards

In this work, we provide theoretical insights on the implicit bias of the BarlowTwins and VICReg loss that can explain these heuristics and guide the development of more principled recommendations.

Representation Learning Self-Supervised Learning

Learning to combine top-down context and feed-forward representations under ambiguity with apical and basal dendrites

no code implementations9 Dec 2023 Nizar Islah, Guillaume Etter, Mashbayar Tugsbayar, Tugce Gurbuz, Blake Richards, Eilif Muller

When input stimuli are ambiguous and relevant contextual information is available, the apical feedback modulates the basal signals to recover unambiguous sensory representations.

Anatomy Temporal Sequences

Synaptic Weight Distributions Depend on the Geometry of Plasticity

1 code implementation30 May 2023 Roman Pogodin, Jonathan Cornford, Arna Ghosh, Gauthier Gidel, Guillaume Lajoie, Blake Richards

Overall, our work shows that the current paradigm in theoretical work on synaptic plasticity that assumes Euclidean synaptic geometry may be misguided and that it should be possible to experimentally determine the true geometry of synaptic plasticity in the brain.

Transfer Entropy Bottleneck: Learning Sequence to Sequence Information Transfer

no code implementations29 Nov 2022 Damjan Kalajdzievski, Ximeng Mao, Pascal Fortier-Poisson, Guillaume Lajoie, Blake Richards

When presented with a data stream of two statistically dependent variables, predicting the future of one of the variables (the target stream) can benefit from information about both its history and the history of the other variable (the source stream).

Contrastive Retrospection: honing in on critical steps for rapid learning and generalization in RL

1 code implementation NeurIPS 2023 Chen Sun, Wannan Yang, Thomas Jiralerspong, Dane Malenfant, Benjamin Alsbury-Nealy, Yoshua Bengio, Blake Richards

Distinct from other contemporary RL approaches to credit assignment, ConSpec takes advantage of the fact that it is easier to retrospectively identify the small set of steps that success is contingent upon (and ignoring other states) than it is to prospectively predict reward at every taken step.

Contrastive Learning Out-of-Distribution Generalization +1

The neuroconnectionist research programme

no code implementations8 Sep 2022 Adrien Doerig, Rowan Sommers, Katja Seeliger, Blake Richards, Jenann Ismael, Grace Lindsay, Konrad Kording, Talia Konkle, Marcel A. J. van Gerven, Nikolaus Kriegeskorte, Tim C. Kietzmann

Artificial Neural Networks (ANNs) inspired by biology are beginning to be widely used to model behavioral and neural data, an approach we call neuroconnectionism.

Philosophy

On Neural Architecture Inductive Biases for Relational Tasks

1 code implementation9 Jun 2022 Giancarlo Kerg, Sarthak Mittal, David Rolnick, Yoshua Bengio, Blake Richards, Guillaume Lajoie

Recent work has explored how forcing relational representations to remain distinct from sensory representations, as it seems to be the case in the brain, can help artificial systems.

Inductive Bias Out-of-Distribution Generalization

Investigating Power laws in Deep Representation Learning

no code implementations11 Feb 2022 Arna Ghosh, Arnab Kumar Mondal, Kumar Krishna Agrawal, Blake Richards

Access to task relevant labels at scale is often scarce or expensive, motivating the need to learn from unlabelled datasets with self-supervised learning (SSL).

Representation Learning Scene Recognition +1

Towards Scaling Difference Target Propagation by Learning Backprop Targets

1 code implementation31 Jan 2022 Maxence Ernoult, Fabrice Normandin, Abhinav Moudgil, Sean Spinney, Eugene Belilovsky, Irina Rish, Blake Richards, Yoshua Bengio

As such, it is important to explore learning algorithms that come with strong theoretical guarantees and can match the performance of backpropagation (BP) on complex tasks.

The functional specialization of visual cortex emerges from training parallel pathways with self-supervised predictive learning

no code implementations NeurIPS 2021 Shahab Bakhtiari, Patrick Mineault, Timothy Lillicrap, Christopher Pack, Blake Richards

We show that when we train a deep neural network architecture with two parallel pathways using a self-supervised predictive loss function, we can outperform other models in fitting mouse visual cortex.

Cannot find the paper you are looking for? You can Submit a new open access paper.