Search Results for author: Chris Nota

Found 6 papers, 1 papers with code

On the Convergence of Discounted Policy Gradient Methods

no code implementations28 Dec 2022 Chris Nota

Many popular policy gradient methods for reinforcement learning follow a biased approximation of the policy gradient known as the discounted approximation.

Policy Gradient Methods reinforcement-learning +1

Is the Policy Gradient a Gradient?

no code implementations17 Jun 2019 Chris Nota, Philip S. Thomas

The policy gradient theorem describes the gradient of the expected discounted return with respect to an agent's policy parameters.

Open-Ended Question Answering Policy Gradient Methods

Classical Policy Gradient: Preserving Bellman's Principle of Optimality

no code implementations6 Jun 2019 Philip S. Thomas, Scott M. Jordan, Yash Chandak, Chris Nota, James Kostas

We propose a new objective function for finite-horizon episodic Markov decision processes that better captures Bellman's principle of optimality, and provide an expression for the gradient of the objective.

Lifelong Learning with a Changing Action Set

1 code implementation5 Jun 2019 Yash Chandak, Georgios Theocharous, Chris Nota, Philip S. Thomas

have been well-studied in the lifelong learning literature, the setting where the action set changes remains unaddressed.

Decision Making

Asynchronous Coagent Networks

no code implementations ICML 2020 James E. Kostas, Chris Nota, Philip S. Thomas

Coagent policy gradient algorithms (CPGAs) are reinforcement learning algorithms for training a class of stochastic neural networks called coagent networks.

Hierarchical Reinforcement Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.