Search Results for author: Yihe Zhou

Found 5 papers, 3 papers with code

Advantage-Aware Policy Optimization for Offline Reinforcement Learning

no code implementations12 Mar 2024 Yunpeng Qing, Shunyu Liu, Jingyuan Cong, KaiXuan Chen, Yihe Zhou, Mingli Song

Offline Reinforcement Learning (RL) endeavors to leverage offline datasets to craft effective agent policy without online interaction, which imposes proper conservative constraints with the support of behavior policies to tackle the Out-Of-Distribution (OOD) problem.

D4RL reinforcement-learning +1

Powerformer: A Section-adaptive Transformer for Power Flow Adjustment

no code implementations5 Jan 2024 KaiXuan Chen, Wei Luo, Shunyu Liu, Yaoquan Wei, Yihe Zhou, Yunpeng Qing, Quan Zhang, Jie Song, Mingli Song

In this paper, we present a novel transformer architecture tailored for learning robust power system state representations, which strives to optimize power dispatch for the power flow adjustment across different transmission sections.

Is Centralized Training with Decentralized Execution Framework Centralized Enough for MARL?

1 code implementation27 May 2023 Yihe Zhou, Shunyu Liu, Yunpeng Qing, KaiXuan Chen, Tongya Zheng, Yanhao Huang, Jie Song, Mingli Song

Despite the encouraging results achieved, CTDE makes an independence assumption on agent policies, which limits agents to adopt global cooperative information from each other during centralized training.

Multi-agent Reinforcement Learning reinforcement-learning +2

Contrastive Identity-Aware Learning for Multi-Agent Value Decomposition

1 code implementation23 Nov 2022 Shunyu Liu, Yihe Zhou, Jie Song, Tongya Zheng, KaiXuan Chen, Tongtian Zhu, Zunlei Feng, Mingli Song

Value Decomposition (VD) aims to deduce the contributions of agents for decentralized policies in the presence of only global rewards, and has recently emerged as a powerful credit assignment paradigm for tackling cooperative Multi-Agent Reinforcement Learning (MARL) problems.

Contrastive Learning SMAC+

Interaction Pattern Disentangling for Multi-Agent Reinforcement Learning

1 code implementation8 Jul 2022 Shunyu Liu, Jie Song, Yihe Zhou, Na Yu, KaiXuan Chen, Zunlei Feng, Mingli Song

In this work, we introduce a novel interactiOn Pattern disenTangling (OPT) method, to disentangle not only the joint value function into agent-wise value functions for decentralized execution, but also the entity interactions into interaction prototypes, each of which represents an underlying interaction pattern within a subgroup of the entities.

Multi-agent Reinforcement Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.