中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Continuous Exploration via Multiple Perspectives in Sparse Reward Environment

文献类型:会议论文

作者Chen ZP(陈忠鹏)1,2; Guan Q(关强)2
出版日期2023-12
会议日期2023-10-13
会议地点厦门国际会议中心
关键词Reinforcement Learning · Exploration Strategy · Sparse Reward · Intrinsic Motivation
英文摘要

Exploration is a major challenge in deep reinforcement learning, especially in cases where reward is sparse. Simple random exploration strategies, such as epsilon-greedy, struggle to solve the hard exploration problem in the sparse reward environment. A more effective approach to solve the hard exploration problem in the sparse reward environment is to use an exploration strategy based on intrinsic motivation, where the key point is to design reasonable and effective intrinsic reward to drive the agent to explore. This paper proposes a method called CEMP, which drives the agent to explore more effectively and continuously in the sparse reward environment. CEMP contributes a new framework for designing intrinsic reward from multiple perspectives, and can be easily integrated into various existing reinforcement learning algorithms. In addition, experimental results in a series of complex and sparse reward environments in MiniGrid demonstrate that our proposed CEMP method achieves better final performance and faster learning efficiency than ICM, RIDE, and TRPO-AE-Hash, which only calculate intrinsic reward from a single perspective.

语种英语
源URL[http://ir.ia.ac.cn/handle/173211/57193]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
通讯作者Guan Q(关强)
作者单位1.中国科学院自动化研究所
2.中国科学院大学
推荐引用方式
GB/T 7714
Chen ZP,Guan Q. Continuous Exploration via Multiple Perspectives in Sparse Reward Environment[C]. 见:. 厦门国际会议中心. 2023-10-13.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。