From inverse optimal control to inverse reinforcement learning: A historical review

被引:70
|
作者
Ab Azar, Nematollah [1 ]
Shahmansoorian, Aref [1 ]
Davoudi, Mohsen [1 ]
机构
[1] Imam Khomeini Int Univ IKIU, Dept Elect Engn, Qazvin, Iran
关键词
Inverse optimal control; Inverse reinforcement learning; Learning from demonstration; Imitation learning; CONTROL LYAPUNOV FUNCTIONS; OPTIMAL ADAPTIVE-CONTROL; MOVEMENT PRIMITIVES; STABILITY MARGINS; DYNAMICAL-SYSTEM; LQ DESIGN; STABILIZATION; OPTIMIZATION; IMITATION; MODELS;
D O I
10.1016/j.arcontrol.2020.06.001
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inverse optimal control (IOC) is a powerful theory that addresses the inverse problems in control systems, robotics, Machine Learning (ML) and optimization taking into account the optimal manners. This paper reviews the history of the IOC and Inverse Reinforcement Learning (IRL) approaches and describes the connections and differences between them to cover the research gap in the existing literature. The general formulation of IOC/IRL is described and the related methods are categorized based on a hierarchical approach. For this purpose, IOC methods are categorized under two classes, namely classic and modern approaches. The classic IOC is typically formulated for control systems, while IRL, as a modern approach to IOC, is considered for machine learning problems. Despite the presence of a handful of IOC/IRL methods, a comprehensive categorization of these methods is lacking. In addition to the IOC/IRL problems, this paper elaborates, where necessary, on other relevant concepts such as Learning from Demonstration (LfD), Imitation Learning (IL), and Behavioral Cloning. Some of the challenges encountered in the IOC/IRL problems are further discussed in this work, including ill-posedness, non-convexity, data availability, non-linearity, the curses of complexity and dimensionality, feature selection, and generalizability. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页码:119 / 138
页数:20
相关论文
共 50 条
  • [1] Inverse Reinforcement Learning in Tracking Control Based on Inverse Optimal Control
    Xue, Wenqian
    Kolaric, Patrik
    Fan, Jialu
    Lian, Bosen
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (10) : 10570 - 10581
  • [2] Inverse Reinforcement Learning from Failure
    Shiarlis, Kyriacos
    Messias, Joao
    Whiteson, Shimon
    AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1060 - 1068
  • [3] Inverse Reinforcement Learning: A Control Lyapunov Approach
    Tesfazgi, Samuel
    Lederer, Armin
    Hirche, Sandra
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 3627 - 3632
  • [4] Inverse-Inverse Reinforcement Learning. How to Hide Strategy from an Adversarial Inverse Reinforcement Learner
    Pattanayak, Kunal
    Krishnamurthy, Vikram
    Berry, Christopher
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 3631 - 3636
  • [5] Inferring Human-Robot Performance Objectives During Locomotion Using Inverse Reinforcement Learning and Inverse Optimal Control
    Liu, Wentao
    Zhong, Junmin
    Wu, Ruofan
    Fylstra, Bretta L.
    Si, Jennie
    Huang, He
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 2549 - 2556
  • [6] Stable Inverse Reinforcement Learning: Policies From Control Lyapunov Landscapes
    Tesfazgi, Samuel
    Sprandl, Leonhard
    Lederer, Armin
    Hirche, Sandra
    IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2024, 3 : 358 - 374
  • [7] Methodologies for Imitation Learning via Inverse Reinforcement Learning: A Review
    Zhang K.
    Yu Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (02): : 254 - 261
  • [8] Inverse reinforcement learning from summary data
    Antti Kangasrääsiö
    Samuel Kaski
    Machine Learning, 2018, 107 : 1517 - 1535
  • [9] Inverse reinforcement learning from summary data
    Kangasraasio, Antti
    Kaski, Samuel
    MACHINE LEARNING, 2018, 107 (8-10) : 1517 - 1535
  • [10] A Review of Inverse Reinforcement Learning Theory and Recent Advances
    Shao Zhifei
    Joo, Er Meng
    2012 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2012,