中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Efficient Hierarchical Reinforcement Learning via Mutual Information Constrained Subgoal Discovery

文献类型:会议论文

作者Kaishen Wang1,2; Jingqing Ruan1; Qingyang Zhang1; Dengpeng Xing1,2
出版日期2023-11
会议日期2023-11
会议地点长沙
英文摘要

Goal-conditioned hierarchical reinforcement learning has demonstrated impressive capabilities in addressing complex and long-horizon tasks. However, the extensive subgoal space often results in low sample efficiency and challenging exploration. To address this issue, we extract informative subgoals by constraining their generation range in mutual information distance space. Specifically, we impose two constraints on the high-level policy during off-policy training: the generated subgoals should be reached with less effort by the low-level policy, and the realization of these subgoals can facilitate achieving the desired goals. These two constraints enable subgoals to act as critical links between the current states and the desired goals, providing more effective guidance to the low-level policy. The empirical results on continuous control tasks
demonstrate that our proposed method significantly enhances the training efficiency, regardless of the dimensions of the state and action spaces, while ensuring comparable performance to state-of-the-art methods.

语种英语
源URL[http://ir.ia.ac.cn/handle/173211/56577]  
专题智能机器人系统研究
通讯作者Dengpeng Xing
作者单位1.Institute of Automation, Chinese Academy of Sciences
2.University of Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Kaishen Wang,Jingqing Ruan,Qingyang Zhang,et al. Efficient Hierarchical Reinforcement Learning via Mutual Information Constrained Subgoal Discovery[C]. 见:. 长沙. 2023-11.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。