Inverse Optimal Control With Constraint Relaxation

被引:0
作者
Rickenbach, Rahel [1 ]
Lahr, Amon [1 ]
Zeilinger, Melanie N. [1 ]
机构
[1] Swiss Fed Inst Technol, Inst Dynam Syst & Control, CH-8092 Zurich, Switzerland
来源
IEEE CONTROL SYSTEMS LETTERS | 2025年 / 9卷
关键词
Noise measurement; Vectors; Optimal control; Noise; Linear programming; Trajectory; Standards; Limiting; Estimation error; Data mining; Constrained control; optimal control; uncertain systems; OPTIMIZATION;
D O I
10.1109/LCSYS.2025.3590879
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inverse optimal control (IOC) is a promising paradigm for learning and mimicking optimal control strategies from capable demonstrators, or gaining a deeper understanding of their intentions, by estimating an unknown objective function from one or more corresponding optimal control sequences. When computing estimates from demonstrations in environments with safety-preserving inequality constraints, acknowledging their presence in the chosen IOC method is crucial given their strong influence on the final control strategy. However, solution strategies capable of considering inequality constraints, such as the inverse Karush-Kuhn-Tucker approach, rely on their correct activation and fulfillment; a restrictive assumption when dealing with noisy demonstrations. To overcome this problem, we leverage the concept of exact penalty functions for IOC and show preservation of estimation accuracy. Considering noisy demonstrations, we then illustrate how the usage of penalty functions reduces the number of unknown variables and how their approximations enhance the estimation method's capacity to account for wrong constraint activations within a polytopic-constrained environment. The proposed method is evaluated for three systems in simulation, outperforming traditional relaxation approaches for noisy demonstrations.
引用
收藏
页码:2055 / 2060
页数:6
相关论文
共 27 条
[1]   From inverse optimal control to inverse reinforcement learning: A historical review [J].
Ab Azar, Nematollah ;
Shahmansoorian, Aref ;
Davoudi, Mohsen .
ANNUAL REVIEWS IN CONTROL, 2020, 50 :119-138
[2]   Learning for Control: An Inverse Optimization Approach [J].
Akhtar, Syed Adnan ;
Kolarijani, Arman Sharifi ;
Esfahani, Peyman Mohajerin .
IEEE CONTROL SYSTEMS LETTERS, 2022, 6 :187-192
[3]   Linear scaling geometry optimisation and transition state search in hybrid delocalised internal coordinates [J].
Billeter, SR ;
Turner, AJ ;
Thiel, W .
PHYSICAL CHEMISTRY CHEMICAL PHYSICS, 2000, 2 (10) :2177-2186
[4]   Inverse KKT: Learning cost functions of manipulation tasks from demonstrations [J].
Englert, Peter ;
Ngo Anh Vien ;
Toussaint, Marc .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2017, 36 (13-14) :1474-1488
[5]   Data-driven inverse optimization with imperfect information [J].
Esfahani, Peyman Mohajerin ;
Shafieezadeh-Abadeh, Soroosh ;
Hanasusanto, Grani A. ;
Kuhn, Daniel .
MATHEMATICAL PROGRAMMING, 2018, 167 (01) :191-234
[6]   Sampling-based Inverse Reinforcement Learning Algorithms with Safety Constraints [J].
Fischer, Johannes ;
Eyberg, Christoph ;
Werling, Moritz ;
Lauer, Martin .
2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, :791-798
[7]   On convex data-driven inverse optimal control for nonlinear, non-stationary and stochastic systems [J].
Garrabe, Emiland ;
Jesawada, Hozefa ;
Del Vecchio, Carmen ;
Russo, Giovanni .
AUTOMATICA, 2025, 173
[8]   On a Probabilistic Approach for Inverse Data-Driven Optimal Control [J].
Garrabe, Emiland ;
Jesawada, Hozefa ;
Del Vecchio, Carmen ;
Russo, Giovanni .
2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, :4411-4416
[9]   A NEW REGULARIZATION METHOD FOR MATHEMATICAL PROGRAMS WITH COMPLEMENTARITY CONSTRAINTS WITH STRONG CONVERGENCE PROPERTIES [J].
Kanzow, Christian ;
Schwartz, Alexandra .
SIAM JOURNAL ON OPTIMIZATION, 2013, 23 (02) :770-798
[10]  
Kerrigan E. C., 2000, P CONTR C, P5777