Search Results for author: Michael Littman

Found 24 papers, 3 papers with code

Selecting Context Clozes for Lightweight Reading Compliance

no code implementations NAACL (BEA) 2022 Greg Keim, Michael Littman

We explore a novel approach to reading compliance, leveraging large language models to select inline challenges that discourage skipping during reading.

Computably Continuous Reinforcement-Learning Objectives are PAC-learnable

no code implementations9 Mar 2023 Cambridge Yang, Michael Littman, Michael Carbin

In particular, for the analysis that considers only sample complexity, we prove that if an objective given as an oracle is uniformly continuous, then it is PAC-learnable.

General Reinforcement Learning reinforcement-learning +2

Model-based Lifelong Reinforcement Learning with Bayesian Exploration

2 code implementations20 Oct 2022 Haotian Fu, Shangqun Yu, Michael Littman, George Konidaris

We propose a model-based lifelong reinforcement-learning approach that estimates a hierarchical Bayesian posterior distilling the common structure shared across different tasks.

reinforcement-learning Reinforcement Learning +1

Meta-Learning Parameterized Skills

1 code implementation7 Jun 2022 Haotian Fu, Shangqun Yu, Saket Tiwari, Michael Littman, George Konidaris

We propose a novel parameterized skill-learning algorithm that aims to learn transferable parameterized skills and synthesize them into a new action space that supports efficient learning in long-horizon tasks.

Meta-Learning Robot Manipulation

Does DQN really learn? Exploring adversarial training schemes in Pong

no code implementations20 Mar 2022 Bowen He, Sreehari Rammohan, Jessica Forde, Michael Littman

In this work, we study two self-play training schemes, Chainer and Pool, and show they lead to improved agent performance in Atari Pong compared to a standard DQN agent -- trained against the built-in Atari opponent.

Learning Generalizable Behavior via Visual Rewrite Rules

no code implementations9 Dec 2021 Yiheng Xie, Mingxuan Li, Shangqun Yu, Michael Littman

Though deep reinforcement learning agents have achieved unprecedented success in recent years, their learned policies can be brittle, failing to generalize to even slight modifications of their environments or unfamiliar situations.

Descriptive

On the (In)Tractability of Reinforcement Learning for LTL Objectives

no code implementations24 Nov 2021 Cambridge Yang, Michael Littman, Michael Carbin

In recent years, researchers have made significant progress in devising reinforcement-learning algorithms for optimizing linear temporal logic (LTL) objectives and LTL-like objectives.

reinforcement-learning Reinforcement Learning +1

Reinforcement Learning with General LTL Objectives is Intractable

no code implementations AAAI Workshop CLeaR 2022 Cambridge Yang, Michael Littman, Michael Carbin

In recent years, researchers have made significant progress in devising reinforcement-learning algorithms for optimizing linear temporal logic (LTL) objectives and LTL-like objectives.

reinforcement-learning Reinforcement Learning +1

Learning Finite Linear Temporal Logic Specifications with a Specialized Neural Operator

no code implementations7 Nov 2021 Homer Walke, Daniel Ritter, Carl Trimbach, Michael Littman

Finite linear temporal logic ($\mathsf{LTL}_f$) is a powerful formal representation for modeling temporal sequences.

Temporal Sequences

Coarse-Grained Smoothness for RL in Metric Spaces

no code implementations23 Oct 2021 Omer Gottesman, Kavosh Asadi, Cameron Allen, Sam Lobel, George Konidaris, Michael Littman

We propose a new coarse-grained smoothness definition that generalizes the notion of Lipschitz continuity, is more widely applicable, and allows us to compute significantly tighter bounds on Q-functions, leading to improved learning.

Decision Making

Bayesian Exploration for Lifelong Reinforcement Learning

no code implementations29 Sep 2021 Haotian Fu, Shangqun Yu, Michael Littman, George Konidaris

A central question in reinforcement learning (RL) is how to leverage prior knowledge to accelerate learning in new tasks.

reinforcement-learning Reinforcement Learning +1

Task Scoping: Generating Task-Specific Abstractions for Planning in Open-Scope Models

no code implementations17 Oct 2020 Michael Fishman, Nishanth Kumar, Cameron Allen, Natasha Danas, Michael Littman, Stefanie Tellex, George Konidaris

Unfortunately, planning to solve any specific task using an open-scope model is computationally intractable - even for state-of-the-art methods - due to the many states and actions that are necessarily present in the model but irrelevant to that problem.

Teaching with IMPACT

no code implementations14 Mar 2019 Carl Trimbach, Michael Littman

Like many problems in AI in their general form, supervised learning is computationally intractable.

ReNeg and Backseat Driver: Learning from Demonstration with Continuous Human Feedback

no code implementations ICLR 2019 Jacob Beck, Zoe Papakipos, Michael Littman

Our framework learns continuous control from sub-optimal demonstration and evaluative feedback collected before training.

Continuous Control

Measuring and Characterizing Generalization in Deep Reinforcement Learning

no code implementations7 Dec 2018 Sam Witty, Jun Ki Lee, Emma Tosch, Akanksha Atrey, Michael Littman, David Jensen

We re-examine what is meant by generalization in RL, and propose several definitions based on an agent's performance in on-policy, off-policy, and unreachable states.

reinforcement-learning Reinforcement Learning +2

Finding Options that Minimize Planning Time

no code implementations16 Oct 2018 Yuu Jinnai, David Abel, D. Ellis Hershkowitz, Michael Littman, George Konidaris

We formalize the problem of selecting the optimal set of options for planning as that of computing the smallest set of options so that planning converges in less than a given maximum of value-iteration passes.

Personalized Education at Scale

no code implementations24 Sep 2018 Sam Saarinen, Evan Cater, Michael Littman

Tailoring the presentation of information to the needs of individual students leads to massive gains in student outcomes~\cite{bloom19842}.

Reinforcement Learning Reinforcement Learning (RL)

State Abstractions for Lifelong Reinforcement Learning

no code implementations ICML 2018 David Abel, Dilip Arumugam, Lucas Lehnert, Michael Littman

We introduce two new classes of abstractions: (1) transitive state abstractions, whose optimal form can be computed efficiently, and (2) PAC state abstractions, which are guaranteed to hold with respect to a distribution of tasks.

reinforcement-learning Reinforcement Learning +1

Policy and Value Transfer in Lifelong Reinforcement Learning

no code implementations ICML 2018 David Abel, Yuu Jinnai, Sophie Yue Guo, George Konidaris, Michael Littman

We consider the problem of how best to use prior experience to bootstrap lifelong learning, where an agent faces a series of task instances drawn from some task distribution.

reinforcement-learning Reinforcement Learning +1

Mean Actor Critic

2 code implementations1 Sep 2017 Cameron Allen, Kavosh Asadi, Melrose Roderick, Abdel-rahman Mohamed, George Konidaris, Michael Littman

We propose a new algorithm, Mean Actor-Critic (MAC), for discrete-action continuous-state reinforcement learning.

Atari Games reinforcement-learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.