Fast Evaluation of Quadratic Control-Lyapunov Policy

被引:16
作者
Wang, Yang [1 ]
Boyd, Stephen [1 ]
机构
[1] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Approximate dynamic programming; model predictive control (MPC); optimization-based control; real-time convex optimization; stochastic control; MODEL-PREDICTIVE CONTROL; ALGORITHMS; SYSTEMS;
D O I
10.1109/TCST.2010.2056371
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The evaluation of a control-Lyapunov policy, with quadratic Lyapunov function, requires the solution of a quadratic program (QP) at each time step. For small problems this QP can be solved explicitly; for larger problems an online optimization method can be used. For this reason the control-Lyapunov control policy is considered a computationally intensive control law, as opposed to an "analytical" control law, such as conventional linear state feedback, linear quadratic Gaussian control, or H-infinity, too complex or slow to be used in high speed control applications. In this note we show that by precomputing certain quantities, the control-Lyapunov policy can be evaluated extremely efficiently. We will show that when the number of inputs is on the order of the square-root of the state dimension, the cost of evaluating a control-Lyapunov policy is on the same order as the cost of evaluating a simple linear state feedback policy, and less (in order) than the cost of updating a Kalman filter state estimate. To give an idea of the speeds involved, for a problem with 100 states and 10 inputs, the control-Lyapunov policy can be evaluated in around 67 mu s, on a 2 GHz AMD processor; the same processor requires 40 mu s to carry out a Kalman filter update.
引用
收藏
页码:939 / 946
页数:8
相关论文
共 58 条
[1]   Efficient solution of second order cone program for model predictive control [J].
Åkerblad, M ;
Hansson, A .
INTERNATIONAL JOURNAL OF CONTROL, 2004, 77 (01) :55-77
[2]  
Anderson E, 1990, P 1990 ACM IEEE C SU, P2
[3]  
Anderson E., 1999, LAPACK Users Guide, V3rd edn., DOI [10.1137/1.9780898719604, DOI 10.1137/1.9780898719604]
[4]  
[Anonymous], 1996, Neuro-dynamic programming
[5]  
[Anonymous], 2007, Control Techniques for Complex Networks
[6]   LEARNING TO ACT USING REAL-TIME DYNAMIC-PROGRAMMING [J].
BARTO, AG ;
BRADTKE, SJ ;
SINGH, SP .
ARTIFICIAL INTELLIGENCE, 1995, 72 (1-2) :81-138
[7]   The explicit linear quadratic regulator for constrained systems [J].
Bemporad, A ;
Morari, M ;
Dua, V ;
Pistikopoulos, EN .
AUTOMATICA, 2002, 38 (01) :3-20
[8]   Suboptimal explicit receding horizon control via approximate multiparametric quadratic programming [J].
Bemporad, A ;
Filippi, C .
JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2003, 117 (01) :9-38
[9]   Model predictive control design: New trends and tools [J].
Bemporad, Alberto .
PROCEEDINGS OF THE 45TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-14, 2006, :6678-6683
[10]  
Bertsekas D., 1996, Stochastic optimal control: the discrete-time case, V5