Search Results for author: Sheila A. McIlraith

Found 18 papers, 5 papers with code

PRP Rebooted: Advancing the State of the Art in FOND Planning

no code implementations18 Dec 2023 Christian Muise, Sheila A. McIlraith, J. Christopher Beck

Fully Observable Non-Deterministic (FOND) planning is a variant of classical symbolic planning in which actions are nondeterministic, with an action's outcome known only upon execution.

Remembering to Be Fair: Non-Markovian Fairness in Sequential Decision Making

no code implementations8 Dec 2023 Parand A. Alamdari, Toryn Q. Klassen, Elliot Creager, Sheila A. McIlraith

In this paper we investigate the notion of fairness in the context of sequential decision making where multiple stakeholders can be affected by the outcomes of decisions.

Decision Making Fairness

Learning Symbolic Representations for Reinforcement Learning of Non-Markovian Behavior

no code implementations8 Jan 2023 Phillip J. K. Christoffersen, Andrew C. Li, Rodrigo Toro Icarte, Sheila A. McIlraith

Recent work has leveraged Knowledge Representation (KR) to provide a symbolic abstraction of aspects of the state that summarize reward-relevant properties of the state-action history and support learning a Markovian decomposition of the problem in terms of an automaton over the KR.

reinforcement-learning Reinforcement Learning (RL)

Noisy Symbolic Abstractions for Deep RL: A case study with Reward Machines

no code implementations20 Nov 2022 Andrew C. Li, Zizhao Chen, Pashootan Vaezipoor, Toryn Q. Klassen, Rodrigo Toro Icarte, Sheila A. McIlraith

Natural and formal languages provide an effective mechanism for humans to specify instructions and reward functions.

Learning to Follow Instructions in Text-Based Games

1 code implementation8 Nov 2022 Mathieu Tuli, Andrew C. Li, Pashootan Vaezipoor, Toryn Q. Klassen, Scott Sanner, Sheila A. McIlraith

Text-based games present a unique class of sequential decision making problem in which agents interact with a partially observable, simulated environment via actions and observations conveyed through natural language.

Decision Making Instruction Following +2

Challenges to Solving Combinatorially Hard Long-Horizon Deep RL Tasks

1 code implementation3 Jun 2022 Andrew C. Li, Pashootan Vaezipoor, Rodrigo Toro Icarte, Sheila A. McIlraith

Deep reinforcement learning has shown promise in discrete domains requiring complex reasoning, including games such as Chess, Go, and Hanabi.

Learning Reward Machines: A Study in Partially Observable Reinforcement Learning

no code implementations17 Dec 2021 Rodrigo Toro Icarte, Ethan Waldie, Toryn Q. Klassen, Richard Valenzano, Margarita P. Castro, Sheila A. McIlraith

Here we show that RMs can be learned from experience, instead of being specified by the user, and that the resulting problem decomposition can be used to effectively solve partially observable RL problems.

Partially Observable Reinforcement Learning Problem Decomposition +2

Be Considerate: Objectives, Side Effects, and Deciding How to Act

no code implementations4 Jun 2021 Parand Alizadeh Alamdari, Toryn Q. Klassen, Rodrigo Toro Icarte, Sheila A. McIlraith

We endow RL agents with the ability to contemplate such impact by augmenting their reward based on expectation of future return by others in the environment, providing different criteria for characterizing impact.

Decision Making Reinforcement Learning (RL)

Planning from Pixels using Inverse Dynamics Models

no code implementations ICLR 2021 Keiran Paster, Sheila A. McIlraith, Jimmy Ba

Learning task-agnostic dynamics models in high-dimensional observation spaces can be challenging for model-based RL agents.

Reward Machines: Exploiting Reward Function Structure in Reinforcement Learning

3 code implementations6 Oct 2020 Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano, Sheila A. McIlraith

First, we propose reward machines, a type of finite state machine that supports the specification of reward functions while exposing reward function structure.

counterfactual Counterfactual Reasoning +3

Interpretable Sequence Classification via Discrete Optimization

1 code implementation6 Oct 2020 Maayan Shvo, Andrew C. Li, Rodrigo Toro Icarte, Sheila A. McIlraith

Our automata-based classifiers are interpretable---supporting explanation, counterfactual reasoning, and human-in-the-loop modification---and have strong empirical performance.

Classification counterfactual +4

Towards the Role of Theory of Mind in Explanation

no code implementations6 May 2020 Maayan Shvo, Toryn Q. Klassen, Sheila A. McIlraith

Theory of Mind is commonly defined as the ability to attribute mental states (e. g., beliefs, goals) to oneself, and to others.

Attribute

Towards Neural-Guided Program Synthesis for Linear Temporal Logic Specifications

no code implementations31 Dec 2019 Alberto Camacho, Sheila A. McIlraith

Synthesizing a program that realizes a logical specification is a classical problem in computer science.

Program Synthesis

Towards Empathetic Planning

no code implementations14 Jun 2019 Maayan Shvo, Sheila A. McIlraith

Critical to successful human interaction is a capacity for empathy - the ability to understand and share the thoughts and feelings of another.

Finite LTL Synthesis with Environment Assumptions and Quality Measures

no code implementations31 Aug 2018 Alberto Camacho, Meghyn Bienvenu, Sheila A. McIlraith

In this paper, we investigate the problem of synthesizing strategies for linear temporal logic (LTL) specifications that are interpreted over finite traces -- a problem that is central to the automated construction of controllers, robot programs, and business processes.

Finite LTL Synthesis is EXPTIME-complete

no code implementations14 Sep 2016 Jorge A. Baier, Alberto Camacho, Christian Muise, Sheila A. McIlraith

LTL synthesis -- the construction of a function to satisfy a logical specification formulated in Linear Temporal Logic -- is a 2EXPTIME-complete problem with relevant applications in controller synthesis and a myriad of artificial intelligence applications.

Cannot find the paper you are looking for? You can Submit a new open access paper.