Search Results for author: Yannick Schroecker

Found 12 papers, 4 papers with code

Structured State Space Models for In-Context Reinforcement Learning

2 code implementations NeurIPS 2023 Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Foerster, Satinder Singh, Feryal Behbahani

We propose a modification to a variant of S4 that enables us to initialise and reset the hidden state in parallel, allowing us to tackle reinforcement learning tasks.

Continuous Control Meta-Learning +1

Meta-Gradients in Non-Stationary Environments

no code implementations13 Sep 2022 Jelena Luketina, Sebastian Flennerhag, Yannick Schroecker, David Abel, Tom Zahavy, Satinder Singh

We support these results with a qualitative analysis of resulting meta-parameter schedules and learned functions of context features.

Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality

no code implementations26 May 2022 Tom Zahavy, Yannick Schroecker, Feryal Behbahani, Kate Baumli, Sebastian Flennerhag, Shaobo Hou, Satinder Singh

Finding different solutions to the same problem is a key aspect of intelligence associated with creativity and adaptation to novel situations.

Bootstrapped Meta-Learning

1 code implementation ICLR 2022 Sebastian Flennerhag, Yannick Schroecker, Tom Zahavy, Hado van Hasselt, David Silver, Satinder Singh

We achieve a new state-of-the art for model-free agents on the Atari ALE benchmark and demonstrate that it yields both performance and efficiency gains in multi-task meta-learning.

Efficient Exploration Few-Shot Learning +1

Active Learning within Constrained Environments through Imitation of an Expert Questioner

no code implementations1 Jul 2019 Kalesha Bullard, Yannick Schroecker, Sonia Chernova

Active learning agents typically employ a query selection algorithm which solely considers the agent's learning objectives.

Active Learning Imitation Learning

Generative predecessor models for sample-efficient imitation learning

no code implementations ICLR 2019 Yannick Schroecker, Mel Vecerik, Jonathan Scholz

We propose Generative Predecessor Models for Imitation Learning (GPRIL), a novel imitation learning algorithm that matches the state-action distribution to the distribution observed in expert demonstrations, using generative models to reason probabilistically about alternative histories of demonstrated states.

Imitation Learning Robot Manipulation

Imitating Latent Policies from Observation

2 code implementations21 May 2018 Ashley D. Edwards, Himanshu Sahni, Yannick Schroecker, Charles L. Isbell

In this paper, we describe a novel approach to imitation learning that infers latent policies directly from state observations.

Imitation Learning

State Aware Imitation Learning

no code implementations NeurIPS 2017 Yannick Schroecker, Charles L. Isbell

Imitation learning is the study of learning how to act given a set of demonstrations provided by a human expert.

Imitation Learning

State Space Decomposition and Subgoal Creation for Transfer in Deep Reinforcement Learning

no code implementations24 May 2017 Himanshu Sahni, Saurabh Kumar, Farhan Tejani, Yannick Schroecker, Charles Isbell

To address this issue, we develop a framework through which a deep RL agent learns to generalize policies from smaller, simpler domains to more complex ones using a recurrent attention mechanism.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.