中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
GCS: Graph-Based Coordination Strategy for Multi-Agent Reinforcement Learning

文献类型:会议论文

作者Jingqing Ruan; Yali Du; Xuantang Xiong; Dengpeng Xing; Dengpeng Xing; Linghui Meng; Haifeng Zhang; Jun Wang; Bo Xu
出版日期2022-05
会议日期2022-5
会议地点Online
英文摘要

Many real-world scenarios involve a team of agents that have to
coordinate their policies to achieve a shared goal. Previous studies
mainly focus on decentralized control to maximize a common reward and barely consider the coordination among control policies,
which is critical in dynamic and complicated environments. In this
work, we propose factorizing the joint team policy into a graph
generator and graph-based coordinated policy to enable coordinated behaviours among agents. The graph generator adopts an
encoder-decoder framework that outputs directed acyclic graphs
(DAGs) to capture the underlying dynamic decision structure. We
also apply the DAGness-constrained and DAG depth-constrained
optimization in the graph generator to balance efficiency and performance. The graph-based coordinated policy exploits the generated decision structure. The graph generator and coordinated
policy are trained simultaneously to maximize the discounted return. Empirical evaluations on Collaborative Gaussian Squeeze, Cooperative Navigation, and Google Research Football demonstrate
the superiority of the proposed method. The code is available at
https://github.com/Amanda-1997/GCS_aamas337.

源URL[http://ir.ia.ac.cn/handle/173211/59413]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
通讯作者Yali Du; Bo Xu
作者单位中科院自动化所
推荐引用方式
GB/T 7714
Jingqing Ruan,Yali Du,Xuantang Xiong,et al. GCS: Graph-Based Coordination Strategy for Multi-Agent Reinforcement Learning[C]. 见:. Online. 2022-5.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。