no code implementations • 10 Jan 2025 • Satchel Grant, Noah D. Goodman, James L. McClelland
To what degree do NNs induce abstract, mutable, slot-like numeric variables, and in what situations do these representations emerge?
1 code implementation • CVPR 2024 • Drew A. Hudson, Daniel Zoran, Mateusz Malinowski, Andrew K. Lampinen, Andrew Jaegle, James L. McClelland, Loic Matthey, Felix Hill, Alexander Lerchner
We introduce SODA, a self-supervised diffusion model, designed for representation learning.
1 code implementation • 6 Jun 2023 • Takateru Yamakoshi, James L. McClelland, Adele E. Goldberg, Robert D. Hawkins
Accounts of human language processing have long appealed to implicit ``situation models'' that enrich comprehension with relevant but unstated world knowledge.
no code implementations • 7 Oct 2022 • Andrew J. Nam, Mustafa Abdool, Trevor Maxfield, James L. McClelland
As a step toward understanding how transformer-based systems generalize, we explore the question of OODG in small scale transformers trained with examples from a known distribution.
Out-of-Distribution Generalization
Systematic Generalization
no code implementations • 6 Oct 2022 • Andrew J. Nam, Mengye Ren, Chelsea Finn, James L. McClelland
Large language models have recently shown promising progress in mathematical reasoning when fine-tuned with human-generated sequences walking through a sequence of solution steps.
no code implementations • 2 Oct 2022 • YuXuan Li, James L. McClelland
Transformer networks have seen great success in natural language processing and machine vision, where task objectives such as next word prediction and image classification benefit from nuanced context sensitivity across high-dimensional inputs.
1 code implementation • 14 Jul 2022 • Ishita Dasgupta, Andrew K. Lampinen, Stephanie C. Y. Chan, Hannah R. Sheahan, Antonia Creswell, Dharshan Kumaran, James L. McClelland, Felix Hill
We evaluate state of the art large language models, as well as humans, and find that the language models reflect many of the same patterns observed in humans across these tasks $\unicode{x2014}$ like humans, models answer more accurately when the semantic content of a task supports the logical inferences.
no code implementations • 5 Apr 2022 • Andrew K. Lampinen, Ishita Dasgupta, Stephanie C. Y. Chan, Kory Matthewson, Michael Henry Tessler, Antonia Creswell, James L. McClelland, Jane X. Wang, Felix Hill
In summary, explanations can support the in-context learning of large LMs on challenging tasks.
1 code implementation • 7 Dec 2021 • Andrew K. Lampinen, Nicholas A. Roy, Ishita Dasgupta, Stephanie C. Y. Chan, Allison C. Tam, James L. McClelland, Chen Yan, Adam Santoro, Neil C. Rabinowitz, Jane X. Wang, Felix Hill
Inferring the abstract relational and causal structure of the world is a major challenge for reinforcement-learning (RL) agents.
no code implementations • 10 Jul 2021 • Andrew J. Nam, James L. McClelland
We also find that most of those who master the task can describe a valid solution strategy, and such participants perform better on transfer puzzles than those whose strategy descriptions are vague or incomplete.
3 code implementations • 8 May 2020 • Andrew K. Lampinen, James L. McClelland
We demonstrate the effectiveness of this framework across a wide variety of tasks and computational paradigms, ranging from regression to image classification and reinforcement learning.
no code implementations • 12 Dec 2019 • James L. McClelland, Felix Hill, Maja Rudolph, Jason Baldridge, Hinrich Schütze
We take language to be a part of a system for understanding and communicating about situations.
no code implementations • ICLR 2020 • Felix Hill, Andrew Lampinen, Rosalia Schneider, Stephen Clark, Matthew Botvinick, James L. McClelland, Adam Santoro
The question of whether deep neural networks are good at generalising beyond their immediate training experience is of critical importance for learning-based approaches to AI.
2 code implementations • 23 May 2019 • Andrew K. Lampinen, James L. McClelland
How can deep learning systems flexibly reuse their knowledge?
1 code implementation • 23 Oct 2018 • Andrew M. Saxe, James L. McClelland, Surya Ganguli
An extensive body of empirical research has revealed remarkable regularities in the acquisition, organization, deployment, and neural representation of human semantic knowledge, thereby raising a fundamental conceptual question: what are the theoretical principles governing the ability of neural networks to acquire, organize, and deploy abstract knowledge by integrating across many individual experiences?
no code implementations • 27 Oct 2017 • Andrew K. Lampinen, James L. McClelland
Standard deep learning systems require thousands or millions of examples to learn a concept, and cannot integrate new concepts easily.
2 code implementations • 20 Dec 2013 • Andrew M. Saxe, James L. McClelland, Surya Ganguli
We further exhibit a new class of random orthogonal initial conditions on weights that, like unsupervised pre-training, enjoys depth independent learning times.