中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Adaptive Q-Learning for Data-Based Optimal Output Regulation With Experience Replay

文献类型:期刊论文

作者Luo, Biao1; Yang, Yin2; Liu, Derong3
刊名IEEE TRANSACTIONS ON CYBERNETICS
出版日期2018-12-01
卷号48期号:12页码:3337-3348
ISSN号2168-2267
关键词Data-based experience replay neural networks (NNs) off-policy optimal control Q-learning (QL)
DOI10.1109/TCYB.2018.2821369
通讯作者Luo, Biao(biao.luo@hotmail.com)
英文摘要In this paper, the data-based optimal output regulation problem of discrete-time systems is investigated. An off-policy adaptive Q-learning (QL) method is developed by using real system data without requiring the knowledge of system dynamics and the mathematical model of utility function. By introducing the Q-function, an off-policy adaptive QI, algorithm is developed to learn the optimal Q-function. An adaptive parameter alpha(i) in the policy evaluation is used to achieve tradeoff between the current and future Q-functions. The convergence of adaptive QI, algorithm is proved and the influence of the adaptive parameter is analyzed. To realize the adaptive QL algorithm with real system data, the actor-critic neural network (NN) structure is developed. The least-squares scheme and the batch gradient descent method are developed to update the critic and actor NN weights, respectively. The experience replay technique is employed in the learning process, which leads to simple and convenient implementation of the adaptive QL method. Finally, the effectiveness of the developed adaptive QL method is verified through numerical simulations.
WOS关键词DISCRETE-TIME-SYSTEMS ; H-INFINITY CONTROL ; SPATIALLY DISTRIBUTED PROCESSES ; UNCERTAIN NONLINEAR-SYSTEMS ; BARRIER LYAPUNOV FUNCTIONS ; POLICY ITERATION ; CONTROL DESIGN ; CONTROLLER-DESIGN ; UNKNOWN DYNAMICS ; TRACKING CONTROL
资助项目National Natural Science Foundation of China[61503377] ; National Natural Science Foundation of China[61533017] ; National Natural Science Foundation of China[U1501251] ; Qatar National Research Fund under National Priority Research Project[NPRP9-466-1-103]
WOS研究方向Automation & Control Systems ; Computer Science
语种英语
出版者IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
WOS记录号WOS:000450613100007
资助机构National Natural Science Foundation of China ; Qatar National Research Fund under National Priority Research Project
源URL[http://ir.ia.ac.cn/handle/173211/22602]  
专题自动化研究所_复杂系统管理与控制国家重点实验室_智能化团队
通讯作者Luo, Biao
作者单位1.Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
2.Hamad Bin Khalifa Univ, Coll Sci & Engn, Doha, Qatar
3.Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Guangdong, Peoples R China
推荐引用方式
GB/T 7714
Luo, Biao,Yang, Yin,Liu, Derong. Adaptive Q-Learning for Data-Based Optimal Output Regulation With Experience Replay[J]. IEEE TRANSACTIONS ON CYBERNETICS,2018,48(12):3337-3348.
APA Luo, Biao,Yang, Yin,&Liu, Derong.(2018).Adaptive Q-Learning for Data-Based Optimal Output Regulation With Experience Replay.IEEE TRANSACTIONS ON CYBERNETICS,48(12),3337-3348.
MLA Luo, Biao,et al."Adaptive Q-Learning for Data-Based Optimal Output Regulation With Experience Replay".IEEE TRANSACTIONS ON CYBERNETICS 48.12(2018):3337-3348.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。