Discrete-Time Local Value Iteration Adaptive Dynamic Programming: Admissibility and Termination Analysis

被引:33
|
作者
Wei, Qinglai [1 ]
Liu, Derong [2 ]
Lin, Qiao [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[2] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Adaptive critic designs; adaptive dynamic programming (ADP); approximate dynamic programming; local iteration; neural networks; neurodynamic programming; nonlinear systems; optimal control; OPTIMAL TRACKING CONTROL; ZERO-SUM GAME; NONLINEAR-SYSTEMS; FEEDBACK-CONTROL; CONTROL SCHEME; LEARNING CONTROL; NETWORKS; DESIGN;
D O I
10.1109/TNNLS.2016.2593743
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a novel local value iteration adaptive dynamic programming (ADP) algorithm is developed to solve infinite horizon optimal control problems for discrete-time nonlinear systems. The focuses of this paper are to study admissibility properties and the termination criteria of discrete-time local value iteration ADP algorithms. In the discrete-time local value iteration ADP algorithm, the iterative value functions and the iterative control laws are both updated in a given subset of the state space in each iteration, instead of the whole state space. For the first time, admissibility properties of iterative control laws are analyzed for the local value iteration ADP algorithm. New termination criteria are established, which terminate the iterative local ADP algorithm with an admissible approximate optimal control law. Finally, simulation results are given to illustrate the performance of the developed algorithm.
引用
收藏
页码:2490 / 2502
页数:13
相关论文
共 50 条
  • [41] Optimal consensus of a class of discrete-time linear multi-agent systems via value iteration with guaranteed admissibility
    Li, Pingchuan
    Zou, Wencheng
    Guo, Jian
    Xiang, Zhengrong
    NEUROCOMPUTING, 2023, 516 : 1 - 10
  • [43] Optimal tracking control for linear discrete-time stochastic system based on adaptive dynamic programming
    Wang, Fang
    Chen, Xin
    Wang, Wei
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 1398 - 1403
  • [44] Stable value iteration for two-player zero-sum game of discrete-time nonlinear systems based on adaptive dynamic programming
    Song, Ruizhuo
    Zhu, Liao
    NEUROCOMPUTING, 2019, 340 : 180 - 195
  • [45] Discrete-time Optimal Zero-sum Games for Nonlinear Systems via Adaptive Dynamic Programming
    Wei, Qinglai
    Song, Ruizhuo
    Xu, Yancai
    Liu, Derong
    Lin, Qiao
    2017 6TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS (DDCLS), 2017, : 357 - 364
  • [46] Value iteration and adaptive dynamic programming for data-driven adaptive optimal control design
    Bian, Tao
    Jiang, Zhong-Ping
    AUTOMATICA, 2016, 71 : 348 - 360
  • [47] Optimal Control for Unknown Discrete-Time Nonlinear Markov Jump Systems Using Adaptive Dynamic Programming
    Zhong, Xiangnan
    He, Haibo
    Zhang, Huaguang
    Wang, Zhanshan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (12) : 2141 - 2155
  • [48] MRAC for unknown discrete-time nonlinear systems based on supervised neural dynamic programming
    Fu, Hao
    Chen, Xin
    Wang, Wei
    Wu, Min
    NEUROCOMPUTING, 2020, 384 : 130 - 141
  • [49] A Generalized Policy Iteration Adaptive Dynamic Programming Algorithm for Optimal Control of Discrete-Time Nonlinear Systems with Actuator Saturation
    Lin, Qiao
    Wei, Qinglai
    Zhao, Bo
    ADVANCES IN NEURAL NETWORKS, PT II, 2017, 10262 : 60 - 65
  • [50] A Nearer Optimal and Faster Trained Value Iteration ADP for Discrete-Time Nonlinear Systems
    Hu, Junping
    Yang, Gen
    Hou, Zhicheng
    Zhang, Gong
    Yang, Wenlin
    Wang, Weijun
    IEEE ACCESS, 2021, 9 : 14933 - 14944