4.5 Article

Off-policy integral reinforcement learning optimal tracking control for continuous-time chaotic systems

期刊

CHINESE PHYSICS B
卷 24, 期 9, 页码 -

出版社

IOP PUBLISHING LTD
DOI: 10.1088/1674-1056/24/9/090504

关键词

adaptive dynamic programming; approximate dynamic programming; chaotic system; optimal tracking control

资金

  1. National Natural Science Foundation of China [61304079, 61374105]
  2. Beijing Natural Science Foundation, China [4132078, 4143065]
  3. China Postdoctoral Science Foundation [2013M530527]
  4. Fundamental Research Funds for the Central Universities, China [FRF-TP-14-119A2]
  5. Open Research Project from State Key Laboratory of Management and Control for Complex Systems, China [20150104]

向作者/读者索取更多资源

This paper estimates an off-policy integral reinforcement learning (IRL) algorithm to obtain the optimal tracking control of unknown chaotic systems. Off-policy IRL can learn the solution of the HJB equation from the system data generated by an arbitrary control. Moreover, off-policy IRL can be regarded as a direct learning method, which avoids the identification of system dynamics. In this paper, the performance index function is first given based on the system tracking error and control error. For solving the Hamilton-Jacobi-Bellman (HJB) equation, an off-policy IRL algorithm is proposed. It is proven that the iterative control makes the tracking error system asymptotically stable, and the iterative performance index function is convergent. Simulation study demonstrates the effectiveness of the developed tracking control method.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据