Model-Based Reinforcement Learning With Isolated Imaginations

被引:0
作者
Pan, Minting [1 ]
Zhu, Xiangming [1 ]
Zheng, Yitao [1 ]
Wang, Yunbo [1 ]
Yang, Xiaokang [1 ]
机构
[1] Shanghai Jiao Tong Univ, AI Inst, MoE Key Lab Artificial Intelligence, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Decoupled dynamics; model-based reinforcement learning; world model;
D O I
10.1109/TPAMI.2023.3335263
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
World models learn the consequences of actions in vision-based interactive systems. However, in practical scenarios like autonomous driving, noncontrollable dynamics that are independent or sparsely dependent on action signals often exist, making it challenging to learn effective world models. To address this issue, we propose Iso-Dream++, a model-based reinforcement learning approach that has two main contributions. First, we optimize the inverse dynamics to encourage the world model to isolate controllable state transitions from the mixed spatiotemporal variations of the environment. Second, we perform policy optimization based on the decoupled latent imaginations, where we roll out noncontrollable states into the future and adaptively associate them with the current controllable state. This enables long-horizon visuomotor control tasks to benefit from isolating mixed dynamics sources in the wild, such as self-driving cars that can anticipate the movement of other vehicles, thereby avoiding potential risks. On top of our previous work (Pan et al. 2022), we further consider the sparse dependencies between controllable and noncontrollable states, address the training collapse problem of state decoupling, and validate our approach in transfer learning setups. Our empirical study demonstrates that Iso-Dream++ outperforms existing reinforcement learning models significantly on CARLA and DeepMind Control.
引用
收藏
页码:2788 / 2803
页数:16
相关论文
共 50 条
  • [41] DATA-EFFICIENT MODEL-BASED REINFORCEMENT LEARNING FOR ROBOT CONTROL
    Sun, Ming
    Gao, Yue
    Liu, Wei
    Li, Shaoyuan
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2021, 36 (04) : 211 - 218
  • [42] A Configurable Model-Based Reinforcement Learning Framework for Disaggregated Storage Systems
    Jeong, Seunghwan
    Woo, Honguk
    IEEE ACCESS, 2023, 11 : 14876 - 14891
  • [43] Model-Based Reinforcement Learning Framework of Online Network Resource Allocation
    Bakhshi, Bahador
    Mangues-Bafalluy, Josep
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 4456 - 4461
  • [44] Data-efficient model-based reinforcement learning with trajectory discrimination
    Qu, Tuo
    Duan, Fuqing
    Zhang, Junge
    Zhao, Bo
    Huang, Wenzhen
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (02) : 1927 - 1936
  • [45] Model-Based Reinforcement Learning with Hierarchical Control for Dynamic Uncertain Environments
    Oesterdiekhoff, Annika
    Heinrich, Nils Wendel
    Russwinkel, Nele
    Kopp, Stefan
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, INTELLISYS 2024, 2024, 1066 : 626 - 642
  • [46] Delay-aware model-based reinforcement learning for continuous control
    Chen, Baiming
    Xu, Mengdi
    Li, Liang
    Zhao, Ding
    NEUROCOMPUTING, 2021, 450 : 119 - 128
  • [47] Model-Based Reinforcement Learning in Multiagent Systems with Sequential Action Selection
    Akramizadeh, Ali
    Afshar, Ahmad
    Menhaj, Mohammad Bagher
    Jafari, Samira
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2011, E94D (02): : 255 - 263
  • [48] Model-based reinforcement learning approach for federated learning resource allocation and parameter optimization
    Karami, Farzan
    Khalaj, Babak Hossein
    COMPUTER COMMUNICATIONS, 2024, 228
  • [49] Successor Features Combine Elements of Model-Free and Model-based Reinforcement Learning
    Lehnert, Lucas
    Littman, Michael L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [50] Model-Based Reinforcement Learning for Eco-Driving Control of Electric Vehicles
    Lee, Heeyun
    Kim, Namwook
    Cha, Suk Won
    IEEE ACCESS, 2020, 8 : 202886 - 202896