Search Results for author: Vernon Lawhern

Found 4 papers, 2 papers with code

A Narration-based Reward Shaping Approach using Grounded Natural Language Commands

no code implementations31 Oct 2019 Nicholas Waytowich, Sean L. Barton, Vernon Lawhern, Garrett Warnell

While this problem can be addressed through reward shaping, such approaches typically require a human expert with specialized knowledge.

Starcraft Starcraft II

Grounding Natural Language Commands to StarCraft II Game States for Narration-Guided Reinforcement Learning

no code implementations24 Apr 2019 Nicholas Waytowich, Sean L. Barton, Vernon Lawhern, Ethan Stump, Garrett Warnell

While deep reinforcement learning techniques have led to agents that are successfully able to learn to perform a number of tasks that had been previously unlearnable, these techniques are still susceptible to the longstanding problem of {\em reward sparsity}.

Reinforcement Learning (RL) Starcraft +1

Compact Convolutional Neural Networks for Classification of Asynchronous Steady-state Visual Evoked Potentials

1 code implementation12 Mar 2018 Nicholas R. Waytowich, Vernon Lawhern, Javier O. Garcia, Jennifer Cummings, Josef Faller, Paul Sajda, Jean M. Vettel

Steady-State Visual Evoked Potentials (SSVEPs) are neural oscillations from the parietal and occipital regions of the brain that are evoked from flickering visual stimuli.

EEG General Classification +1

Deep TAMER: Interactive Agent Shaping in High-Dimensional State Spaces

2 code implementations28 Sep 2017 Garrett Warnell, Nicholas Waytowich, Vernon Lawhern, Peter Stone

While recent advances in deep reinforcement learning have allowed autonomous learning agents to succeed at a variety of complex tasks, existing algorithms generally require a lot of training data.

reinforcement-learning Reinforcement Learning (RL) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.