Search Results for author: Nathaniel Grammel

Found 6 papers, 2 papers with code

Agent Environment Cycle Games

no code implementations28 Sep 2020 Justin K. Terry, Nathaniel Grammel, Benjamin Black, Ananth Hari, Caroline Horsch, Luis Santos

Partially Observable Stochastic Games (POSGs) are the most general and common model of games used in Multi-Agent Reinforcement Learning (MARL).

Multi-agent Reinforcement Learning reinforcement-learning +1

Stochastic Optimization and Learning for Two-Stage Supplier Problems

no code implementations7 Aug 2020 Brian Brubach, Nathaniel Grammel, David G. Harris, Aravind Srinivasan, Leonidas Tsepenekas, Anil Vullikanti

The main focus of this paper is radius-based (supplier) clustering in the two-stage stochastic setting with recourse, where the inherent stochasticity of the model comes in the form of a budget constraint.

Data Structures and Algorithms

Multi-Agent Informational Learning Processes

no code implementations11 Jun 2020 Justin K. Terry, Nathaniel Grammel

We introduce a new mathematical model of multi-agent reinforcement learning, the Multi-Agent Informational Learning Processor "MAILP" model.

Multi-agent Reinforcement Learning reinforcement-learning +1

Revisiting Parameter Sharing in Multi-Agent Deep Reinforcement Learning

2 code implementations27 May 2020 J. K. Terry, Nathaniel Grammel, Sanghyun Son, Benjamin Black, Aakriti Agrawal

Next, we formally introduce methods to extend parameter sharing to learning in heterogeneous observation and action spaces, and prove that these methods allow for convergence to optimal policies.

Multi-agent Reinforcement Learning reinforcement-learning +1

Scenario Submodular Cover

no code implementations10 Mar 2016 Nathaniel Grammel, Lisa Hellerstein, Devorah Kletenik, Patrick Lin

In contrast, in Stochastic Submodular Cover, the variables of the input distribution are assumed to be independent, and the distribution of each variable is given as input.

Cannot find the paper you are looking for? You can Submit a new open access paper.