ColosseumRL: A Framework for Multiagent Reinforcement Learning in $N$-Player Games

10 Dec 2019  ·  Alexander Shmakov, John Lanier, Stephen McAleer, Rohan Achar, Cristina Lopes, Pierre Baldi ·

Much of recent success in multiagent reinforcement learning has been in two-player zero-sum games. In these games, algorithms such as fictitious self-play and minimax tree search can converge to an approximate Nash equilibrium. While playing a Nash equilibrium strategy in a two-player zero-sum game is optimal, in an $n$-player general sum game, it becomes a much less informative solution concept. Despite the lack of a satisfying solution concept, $n$-player games form the vast majority of real-world multiagent situations. In this paper we present a new framework for research in reinforcement learning in $n$-player games. We hope that by analyzing behavior learned by agents in these environments the community can better understand this important research area and move toward meaningful solution concepts and research directions. The implementation and additional information about this framework can be found at https://colosseumrl.igb.uci.edu/.

PDF Abstract
No code implementations yet. Submit your code now

Categories


Multiagent Systems

Datasets


Introduced in the Paper:

ColosseumRL