Search Results for author: Anuj Mahajan

Found 18 papers, 7 papers with code

Generalization Across Observation Shifts in Reinforcement Learning

no code implementations7 Jun 2023 Anuj Mahajan, Amy Zhang

We focus on bisimulation metrics, which provide a powerful means for abstracting task relevant components of the observation and learning a succinct representation space for training the agent using reinforcement learning.

reinforcement-learning

marl-jax: Multi-Agent Reinforcement Leaning Framework

1 code implementation24 Mar 2023 Kinal Mehta, Anuj Mahajan, Pawan Kumar

We present marl-jax, a multi-agent reinforcement learning software package for training and evaluating social generalization of the agents.

Multi-agent Reinforcement Learning reinforcement-learning +1

Trust-Region-Free Policy Optimization for Stochastic Policies

no code implementations15 Feb 2023 Mingfei Sun, Benjamin Ellis, Anuj Mahajan, Sam Devlin, Katja Hofmann, Shimon Whiteson

In this paper, we show that the trust region constraint over policies can be safely substituted by a trust-region-free constraint without compromising the underlying monotonic improvement guarantee.

Generalization in Cooperative Multi-Agent Systems

no code implementations31 Jan 2022 Anuj Mahajan, Mikayel Samvelyan, Tarun Gupta, Benjamin Ellis, Mingfei Sun, Tim Rocktäschel, Shimon Whiteson

Specifically, we study generalization bounds under a linear dependence of the underlying dynamics on the agent capabilities, which can be seen as a generalization of Successor Features to MAS.

Generalization Bounds Multi-agent Reinforcement Learning

Reinforcement Learning in Factored Action Spaces using Tensor Decompositions

no code implementations27 Oct 2021 Anuj Mahajan, Mikayel Samvelyan, Lei Mao, Viktor Makoviychuk, Animesh Garg, Jean Kossaifi, Shimon Whiteson, Yuke Zhu, Animashree Anandkumar

We present an extended abstract for the previously published work TESSERACT [Mahajan et al., 2021], which proposes a novel solution for Reinforcement Learning (RL) in large, factored action spaces using tensor decompositions.

Multi-agent Reinforcement Learning reinforcement-learning +1

Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning

no code implementations31 May 2021 Anuj Mahajan, Mikayel Samvelyan, Lei Mao, Viktor Makoviychuk, Animesh Garg, Jean Kossaifi, Shimon Whiteson, Yuke Zhu, Animashree Anandkumar

Algorithms derived from Tesseract decompose the Q-tensor across agents and utilise low-rank tensor approximations to model agent interactions relevant to the task.

Learning Theory Multi-agent Reinforcement Learning +3

UneVEn: Universal Value Exploration for Multi-Agent Reinforcement Learning

no code implementations6 Oct 2020 Tarun Gupta, Anuj Mahajan, Bei Peng, Wendelin Böhmer, Shimon Whiteson

VDN and QMIX are two popular value-based algorithms for cooperative MARL that learn a centralized action value function as a monotonic mixing of per-agent utilities.

Multi-agent Reinforcement Learning reinforcement-learning +3

RODE: Learning Roles to Decompose Multi-Agent Tasks

2 code implementations ICLR 2021 Tonghan Wang, Tarun Gupta, Anuj Mahajan, Bei Peng, Shimon Whiteson, Chongjie Zhang

Learning a role selector based on action effects makes role discovery much easier because it forms a bi-level learning hierarchy -- the role selector searches in a smaller role space and at a lower temporal resolution, while role policies learn in significantly reduced primitive action-observation spaces.

Clustering Starcraft +1

VIREL: A Variational Inference Framework for Reinforcement Learning

1 code implementation NeurIPS 2019 Matthew Fellows, Anuj Mahajan, Tim G. J. Rudner, Shimon Whiteson

This gives VIREL a mode-seeking form of KL divergence, the ability to learn deterministic optimal polices naturally from inference and the ability to optimise value functions and policies in separate, iterative steps.

reinforcement-learning Reinforcement Learning (RL) +1

Symmetry Learning for Function Approximation in Reinforcement Learning

no code implementations9 Jun 2017 Anuj Mahajan, Theja Tulabandhula

In this paper we explore methods to exploit symmetries for ensuring sample efficiency in reinforcement learning (RL), this problem deserves ever increasing attention with the recent advances in the use of deep networks for complex RL tasks which require large amount of training data.

reinforcement-learning Reinforcement Learning (RL)

Lifted Inference Rules With Constraints

no code implementations NeurIPS 2015 Happy Mittal, Anuj Mahajan, Vibhav G. Gogate, Parag Singla

Lifted inference rules exploit symmetries for fast reasoning in statistical rela-tional models.

Cannot find the paper you are looking for? You can Submit a new open access paper.