Search Results for author: James L. McClelland

Found 16 papers, 8 papers with code

Causal interventions expose implicit situation models for commonsense language understanding

1 code implementation6 Jun 2023 Takateru Yamakoshi, James L. McClelland, Adele E. Goldberg, Robert D. Hawkins

Accounts of human language processing have long appealed to implicit ``situation models'' that enrich comprehension with relevant but unstated world knowledge.

World Knowledge

Achieving and Understanding Out-of-Distribution Generalization in Systematic Reasoning in Small-Scale Transformers

no code implementations7 Oct 2022 Andrew J. Nam, Mustafa Abdool, Trevor Maxfield, James L. McClelland

As a step toward understanding how transformer-based systems generalize, we explore the question of OODG in small scale transformers trained with examples from a known distribution.

Out-of-Distribution Generalization Systematic Generalization

Learning to Reason With Relational Abstractions

no code implementations6 Oct 2022 Andrew J. Nam, Mengye Ren, Chelsea Finn, James L. McClelland

Large language models have recently shown promising progress in mathematical reasoning when fine-tuned with human-generated sequences walking through a sequence of solution steps.

Mathematical Reasoning

Systematic Generalization and Emergent Structures in Transformers Trained on Structured Tasks

no code implementations2 Oct 2022 YuXuan Li, James L. McClelland

Transformer networks have seen great success in natural language processing and machine vision, where task objectives such as next word prediction and image classification benefit from nuanced context sensitivity across high-dimensional inputs.

Image Classification Systematic Generalization

Language models show human-like content effects on reasoning tasks

1 code implementation14 Jul 2022 Ishita Dasgupta, Andrew K. Lampinen, Stephanie C. Y. Chan, Hannah R. Sheahan, Antonia Creswell, Dharshan Kumaran, James L. McClelland, Felix Hill

We evaluate state of the art large language models, as well as humans, and find that the language models reflect many of the same patterns observed in humans across these tasks $\unicode{x2014}$ like humans, models answer more accurately when the semantic content of a task supports the logical inferences.

Language Modelling Logical Reasoning +2

Systematic human learning and generalization from a brief tutorial with explanatory feedback

no code implementations10 Jul 2021 Andrew J. Nam, James L. McClelland

We also find that most of those who master the task can describe a valid solution strategy, and such participants perform better on transfer puzzles than those whose strategy descriptions are vague or incomplete.

High School Mathematics Systematic Generalization +1

Transforming task representations to perform novel tasks

3 code implementations8 May 2020 Andrew K. Lampinen, James L. McClelland

We demonstrate the effectiveness of this framework across a wide variety of tasks and computational paradigms, ranging from regression to image classification and reinforcement learning.

Image Classification Zero-Shot Learning

Extending Machine Language Models toward Human-Level Language Understanding

no code implementations12 Dec 2019 James L. McClelland, Felix Hill, Maja Rudolph, Jason Baldridge, Hinrich Schütze

We take language to be a part of a system for understanding and communicating about situations.

Environmental drivers of systematicity and generalization in a situated agent

no code implementations ICLR 2020 Felix Hill, Andrew Lampinen, Rosalia Schneider, Stephen Clark, Matthew Botvinick, James L. McClelland, Adam Santoro

The question of whether deep neural networks are good at generalising beyond their immediate training experience is of critical importance for learning-based approaches to AI.

Unity

A mathematical theory of semantic development in deep neural networks

1 code implementation23 Oct 2018 Andrew M. Saxe, James L. McClelland, Surya Ganguli

An extensive body of empirical research has revealed remarkable regularities in the acquisition, organization, deployment, and neural representation of human semantic knowledge, thereby raising a fundamental conceptual question: what are the theoretical principles governing the ability of neural networks to acquire, organize, and deploy abstract knowledge by integrating across many individual experiences?

Semantic Similarity Semantic Textual Similarity

One-shot and few-shot learning of word embeddings

no code implementations27 Oct 2017 Andrew K. Lampinen, James L. McClelland

Standard deep learning systems require thousands or millions of examples to learn a concept, and cannot integrate new concepts easily.

Few-Shot Learning Sentence +1

Exact solutions to the nonlinear dynamics of learning in deep linear neural networks

3 code implementations20 Dec 2013 Andrew M. Saxe, James L. McClelland, Surya Ganguli

We further exhibit a new class of random orthogonal initial conditions on weights that, like unsupervised pre-training, enjoys depth independent learning times.

Unsupervised Pre-training

Cannot find the paper you are looking for? You can Submit a new open access paper.