Search Results for author: Jiechuan Jiang

Found 17 papers, 5 papers with code

A Survey on Game Playing Agents and Large Models: Methods, Applications, and Challenges

1 code implementation15 Mar 2024 Xinrun Xu, Yuxin Wang, Chaoyi Xu, Ziluo Ding, Jiechuan Jiang, Zhiming Ding, Börje F. Karlsson

The swift evolution of Large-scale Models (LMs), either language-focused or multi-modal, has garnered extensive attention in both academy and industry.

Towards General Computer Control: A Multimodal Agent for Red Dead Redemption II as a Case Study

2 code implementations5 Mar 2024 Weihao Tan, Ziluo Ding, Wentao Zhang, Boyu Li, Bohan Zhou, Junpeng Yue, Haochong Xia, Jiechuan Jiang, Longtao Zheng, Xinrun Xu, Yifei Bi, Pengjie Gu, Xinrun Wang, Börje F. Karlsson, Bo An, Zongqing Lu

Despite the success in specific tasks and scenarios, existing foundation agents, empowered by large models (LMs) and advanced tools, still cannot generalize to different scenarios, mainly due to dramatic differences in the observations and actions across scenarios.

Efficient Exploration

Fully Decentralized Cooperative Multi-Agent Reinforcement Learning: A Survey

no code implementations10 Jan 2024 Jiechuan Jiang, Kefan Su, Zongqing Lu

Cooperative multi-agent reinforcement learning is a powerful tool to solve many real-world cooperative tasks, but restrictions of real-world applications may require training the agents in a fully decentralized manner.

Multi-agent Reinforcement Learning reinforcement-learning

Learning from Visual Observation via Offline Pretrained State-to-Go Transformer

no code implementations NeurIPS 2023 Bohan Zhou, Ke Li, Jiechuan Jiang, Zongqing Lu

Learning from visual observation (LfVO), aiming at recovering policies from only visual observation data, is promising yet a challenging problem.

reinforcement-learning

Model-Based Decentralized Policy Optimization

no code implementations16 Feb 2023 Hao Luo, Jiechuan Jiang, Zongqing Lu

To help the policy improvement be stable and monotonic, we propose model-based decentralized policy optimization (MDPO), which incorporates a latent variable function to help construct the transition and reward function from an individual perspective.

Best Possible Q-Learning

no code implementations2 Feb 2023 Jiechuan Jiang, Zongqing Lu

To tackle this challenge, we propose best possible operator, a novel decentralized operator, and prove that the policies of agents will converge to the optimal joint policy if each agent independently updates its individual state-action value by the operator.

Multi-agent Reinforcement Learning Q-Learning

Offline Decentralized Multi-Agent Reinforcement Learning

no code implementations4 Aug 2021 Jiechuan Jiang, Zongqing Lu

In this paper, we propose a framework for offline decentralized multi-agent reinforcement learning, which exploits value deviation and transition normalization to deliberately modify the transition probabilities.

Multi-agent Reinforcement Learning Q-Learning +2

Model-Based Opponent Modeling

no code implementations4 Aug 2021 Xiaopeng Yu, Jiechuan Jiang, Wanpeng Zhang, Haobin Jiang, Zongqing Lu

When one agent interacts with a multi-agent environment, it is challenging to deal with various opponents unseen before.

Adaptive Learning Rates for Multi-Agent Reinforcement Learning

no code implementations1 Jan 2021 Jiechuan Jiang, Zongqing Lu

In multi-agent reinforcement learning (MARL), the learning rates of actors and critic are mostly hand-tuned and fixed.

Multi-agent Reinforcement Learning reinforcement-learning +1

The Emergence of Individuality in Multi-Agent Reinforcement Learning

no code implementations28 Sep 2020 Jiechuan Jiang, Zongqing Lu

EOI learns a probabilistic classifier that predicts a probability distribution over agents given their observation and gives each agent an intrinsic reward of being correctly predicted by the classifier.

Multi-agent Reinforcement Learning reinforcement-learning +1

The Emergence of Individuality

2 code implementations10 Jun 2020 Jiechuan Jiang, Zongqing Lu

EOI learns a probabilistic classifier that predicts a probability distribution over agents given their observation and gives each agent an intrinsic reward of being correctly predicted by the classifier.

Multi-agent Reinforcement Learning

Generative Exploration and Exploitation

no code implementations21 Apr 2019 Jiechuan Jiang, Zongqing Lu

Sparse reward is one of the biggest challenges in reinforcement learning (RL).

Reinforcement Learning (RL)

Learning Attentional Communication for Multi-Agent Cooperation

no code implementations NeurIPS 2018 Jiechuan Jiang, Zongqing Lu

Our model leads to efficient and effective communication for large-scale multi-agent cooperation.

Decision Making

Cannot find the paper you are looking for? You can Submit a new open access paper.