Continuous Exploration via Multiple Perspectives in Sparse Reward Environment
文献类型:会议论文
作者 | Chen ZP(陈忠鹏)1,2![]() ![]() |
出版日期 | 2023-12 |
会议日期 | 2023-10-13 |
会议地点 | 厦门国际会议中心 |
关键词 | Reinforcement Learning · Exploration Strategy · Sparse Reward · Intrinsic Motivation |
英文摘要 | Exploration is a major challenge in deep reinforcement learning, especially in cases where reward is sparse. Simple random exploration strategies, such as epsilon-greedy, struggle to solve the hard exploration problem in the sparse reward environment. A more effective approach to solve the hard exploration problem in the sparse reward environment is to use an exploration strategy based on intrinsic motivation, where the key point is to design reasonable and effective intrinsic reward to drive the agent to explore. This paper proposes a method called CEMP, which drives the agent to explore more effectively and continuously in the sparse reward environment. CEMP contributes a new framework for designing intrinsic reward from multiple perspectives, and can be easily integrated into various existing reinforcement learning algorithms. In addition, experimental results in a series of complex and sparse reward environments in MiniGrid demonstrate that our proposed CEMP method achieves better final performance and faster learning efficiency than ICM, RIDE, and TRPO-AE-Hash, which only calculate intrinsic reward from a single perspective. |
语种 | 英语 |
源URL | [http://ir.ia.ac.cn/handle/173211/57193] ![]() |
专题 | 数字内容技术与服务研究中心_听觉模型与认知计算 |
通讯作者 | Guan Q(关强) |
作者单位 | 1.中国科学院自动化研究所 2.中国科学院大学 |
推荐引用方式 GB/T 7714 | Chen ZP,Guan Q. Continuous Exploration via Multiple Perspectives in Sparse Reward Environment[C]. 见:. 厦门国际会议中心. 2023-10-13. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。