Additional planning with multiple objectives for reinforcement learning

被引:11
|
作者
Pan, Anqi [1 ,2 ]
Xu, Wenjun [3 ,4 ]
Wang, Lei [5 ]
Ren, Hongliang [3 ]
机构
[1] Donghua Univ, Coll Informat Sci & Technol, Shanghai 201620, Peoples R China
[2] Donghua Univ, Engn Res Ctr Digitized Text & Fash Technol, Minist Educ, Shanghai 201620, Peoples R China
[3] Natl Univ Singapore, Dept Biomed Engn, Singapore 117583, Singapore
[4] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[5] Tongji Univ, Sch Elect & Informat Engn, Shanghai 201804, Peoples R China
关键词
Reinforcement learning; Multi-objective; Robotic control; ALGORITHM;
D O I
10.1016/j.knosys.2019.105392
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most control tasks have multiple objectives that need to be achieved simultaneously, while the reward definition is the weighted combination of all objects to determine one optimal policy. This configuration has a limitation in exploration flexibility and presents difficulty in reaching a satisfied terminate condition. Although some multi-objective reinforcement learning (MORL) methods have been presented recently, they concentrate on obtaining a set of compromising options rather than one best-performed strategy. On the other hand, the existing policy-improve methods have rarely emphasized on solving multiple objectives circumstances. Inspired by the enhanced policy search methods, an additional planning technique with multiple objectives for reinforcement learning is proposed in this paper, which is denoted as RLAP-MOP. This method provides opportunities to evaluate parallel requirements at the same time and suggests several optimal feasible actions to improve long-term performance further. Meanwhile, the short-term planning adopted in this paper has advantages in maintaining safe trajectories and building more accurate approximate models, which contributes to accelerating the training program. For comparison, an RLAP with single-objective optimization is also introduced in theoretical and experimental studies. The proposed techniques are investigated on a multi-objective cartpole environment and a soft robotic palpation task. The superiorities in the improved return values and learning stability prove that the multiple objectives based additional planning is a promising assistant to improve reinforcement learning. (c) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Limit Reachability for Model-Free Reinforcement Learning of ω-Regular Objectives
    Hahn, Ernst Moritz
    Perez, Mateo
    Schewe, Sven
    Somenzi, Fabio
    Trivedi, Ashutosh
    Wojtczak, Dominik
    PROCEEDINGS OF THE 5TH INTERNATIONAL WORKSHOP ON SYMBOLIC-NUMERIC METHODS FOR REASONING ABOUT CPS AND IOT (SNR 2019), 2019, : 16 - 18
  • [22] APPLICATION OF REINFORCEMENT LEARNING IN MULTISENSOR FUSION PROBLEMS WITH CONFLICTING CONTROL OBJECTIVES
    Ou, Shichao
    Fagg, Andrew H.
    Shenoy, Prashant
    Chen, Liancheng
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2009, 15 (02) : 223 - 235
  • [23] The Intelligent Path Planning System of Agricultural Robot via Reinforcement Learning
    Yang, Jiachen
    Ni, Jingfei
    Li, Yang
    Wen, Jiabao
    Chen, Desheng
    SENSORS, 2022, 22 (12)
  • [24] Reducing the Planning Horizon Through Reinforcement Learning
    Dunbar, Logan
    Rosman, Benjamin
    Cohn, Anthony G.
    Leonetti, Matteo
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT IV, 2023, 13716 : 68 - 83
  • [25] Planning for potential: efficient safe reinforcement learning
    Floris den Hengst
    Vincent François-Lavet
    Mark Hoogendoorn
    Frank van Harmelen
    Machine Learning, 2022, 111 : 2255 - 2274
  • [26] Trajectory Planning for Hypersonic Vehicles with Reinforcement Learning
    Chi, Haihong
    Thou, Mingxin
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3721 - 3726
  • [27] Planning for potential: efficient safe reinforcement learning
    den Hengst, Floris
    Francois-Lavet, Vincent
    Hoogendoorn, Mark
    van Harmelen, Frank
    MACHINE LEARNING, 2022, 111 (06) : 2255 - 2274
  • [28] Planning-Augmented Hierarchical Reinforcement Learning
    Gieselmann, Robert
    Pokorny, Florian T.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 5097 - 5104
  • [29] Approximate planning for bayesian hierarchical reinforcement learning
    Ngo Anh Vien
    Hung Ngo
    Lee, Sungyoung
    Chung, TaeChoong
    APPLIED INTELLIGENCE, 2014, 41 (03) : 808 - 819
  • [30] REINFORCEMENT LEARNING FOR FISHING ROUTE PLANNING AND OPTIMIZATION
    Zhu, Tiantian
    Naseri, Masoud
    Dhar, Sushmit
    Ashrafi, Behrooz
    PROCEEDINGS OF ASME 2024 43RD INTERNATIONAL CONFERENCE ON OCEAN, OFFSHORE AND ARCTIC ENGINEERING, OMAE2024, VOL 6, 2024,