A collaboration of multi-agent model using an interactive interface

Multi-agent reinforcement learning algorithms scarcely attend to noisy environments, in which agents are inhibited from achieving optimal policy training and making correct decisions. This work investigates the effect of noises in multi-agent environments and proposes a multi-agent actor-critic with collaboration (MACC) model. The model uses lightweight communication to overcome the interference from noises. There are two policies for each agent in MACC: collaboration policy and behavior policy. The behavior of an agent not only depends on its own state but also be influenced by each other agent through a scalar, collaboration value. The collaboration value is generated by the collaboration policy for each individual agent, and it ensures a succinct consensus about the environment. This paper elaborates on the training of the collaboration policy and specifies how it coordinates the behavior policy in a manner of temporal abstraction mechanism, while the observation sequence is considered for more accurate perception. Several experiments on multi-agent collaboration simulation platforms demonstrate that the MACC performs better than baselines in noisy environments, especially in partially observable environments.

PDF

Datasets


  Add Datasets introduced or used in this paper

Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods


No methods listed for this paper. Add relevant methods here