no code implementations • 12 May 2023 • Sourena Khanzadeh, Samad Alias Nyein Chan, Richard Valenzano, Manar Alalfi
This paper presents an approach that evaluates best-first search methods to code refactoring.
no code implementations • 17 Dec 2021 • Rodrigo Toro Icarte, Ethan Waldie, Toryn Q. Klassen, Richard Valenzano, Margarita P. Castro, Sheila A. McIlraith
Here we show that RMs can be learned from experience, instead of being specified by the user, and that the resulting problem decomposition can be used to effectively solve partially observable RL problems.
Partially Observable Reinforcement Learning Problem Decomposition +2
3 code implementations • 6 Oct 2020 • Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano, Sheila A. McIlraith
First, we propose reward machines, a type of finite state machine that supports the specification of reward functions while exposing reward function structure.
no code implementations • 5 Oct 2020 • Rodrigo Toro Icarte, Richard Valenzano, Toryn Q. Klassen, Phillip Christoffersen, Amir-Massoud Farahmand, Sheila A. McIlraith
Learning memoryless policies is efficient and optimal in fully observable environments.
Partially Observable Reinforcement Learning reinforcement-learning +1
1 code implementation • ICML 2018 • Rodrigo Toro Icarte, Toryn Klassen, Richard Valenzano, Sheila Mcilraith
In this paper we propose Reward Machines {—} a type of finite state machine that supports the specification of reward functions while exposing reward function structure to the learner and supporting decomposition.