no code implementations • 24 Jun 2019 • Mark Woodward, Chelsea Finn, Karol Hausman
In this paper, we investigate if, and how, a "helper" agent can be trained to interactively adapt their behavior to maximize the reward of another agent, whom we call the "prime" agent, without observing their reward or receiving explicit demonstrations.
no code implementations • 24 Jun 2019 • Mark Woodward, Chelsea Finn, Karol Hausman
Most importantly, we find that our approach produces an agent that is capable of learning interactively from a human user, without a set of explicit demonstrations or a reward function, and achieving significantly better performance cooperatively with a human than a human performing the task alone.
2 code implementations • 21 Feb 2017 • Mark Woodward, Chelsea Finn
Recent advances in one-shot learning have produced models that can learn from a handful of labeled examples, for passive classification and regression tasks.