Efficient Hierarchical Reinforcement Learning via Mutual Information Constrained Subgoal Discovery
文献类型:会议论文
作者 | Kaishen Wang1,2![]() ![]() ![]() |
出版日期 | 2023-11 |
会议日期 | 2023-11 |
会议地点 | 长沙 |
英文摘要 | Goal-conditioned hierarchical reinforcement learning has demonstrated impressive capabilities in addressing complex and long-horizon tasks. However, the extensive subgoal space often results in low sample efficiency and challenging exploration. To address this issue, we extract informative subgoals by constraining their generation range in mutual information distance space. Specifically, we impose two constraints on the high-level policy during off-policy training: the generated subgoals should be reached with less effort by the low-level policy, and the realization of these subgoals can facilitate achieving the desired goals. These two constraints enable subgoals to act as critical links between the current states and the desired goals, providing more effective guidance to the low-level policy. The empirical results on continuous control tasks |
语种 | 英语 |
源URL | [http://ir.ia.ac.cn/handle/173211/56577] ![]() |
专题 | 智能机器人系统研究 |
通讯作者 | Dengpeng Xing |
作者单位 | 1.Institute of Automation, Chinese Academy of Sciences 2.University of Chinese Academy of Sciences |
推荐引用方式 GB/T 7714 | Kaishen Wang,Jingqing Ruan,Qingyang Zhang,et al. Efficient Hierarchical Reinforcement Learning via Mutual Information Constrained Subgoal Discovery[C]. 见:. 长沙. 2023-11. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。