Spacecraft Proximity Maneuvering and Rendezvous With Collision Avoidance Based on Reinforcement Learning

被引:24
|
作者
Qu, Qingyu [1 ]
Liu, Kexin [2 ]
Wang, Wei [2 ]
Lu, Jinhu [2 ]
机构
[1] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Automat Sci & Elect Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
基金
美国国家科学基金会;
关键词
Space vehicles; Heuristic algorithms; Aerodynamics; Collision avoidance; Oscillators; Orbits; Mathematical models; Aerospace control; autonomous spacecraft rendezvous (ASR); collision avoidance; deep reinforcement learning (DRL); SLIDING MODE CONTROL;
D O I
10.1109/TAES.2022.3180271
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
The rapid development of the aerospace industry puts forward the urgent need for the evolution of autonomous spacecraft rendezvous technology, which has gained significant attention recently due to increased applications in various space missions. This article studies the relative position tracking problem of the autonomous spacecraft rendezvous under the requirement of collision avoidance. An exploration-adaptive deep deterministic policy gradient (DDPG) algorithm is proposed to train a definite control strategy for this mission. Similar to the DDPG algorithm, four neural networks are used in this method, where two of them are used to generate the deterministic policy, whereas the other two are used to score the obtained policy. Differently, adaptive noise is introduced to reduce the possibility of oscillations and divergences and to cut down the unnecessary computation by weakening the exploration of stabilization problems. In addition, in order to effectively and quickly adapt to some other similar scenarios, a metalearning-based idea is introduced by fine-tuning the prior strategy. Finally, two numerical simulations show that the trained control strategy can effectively avoid the oscillation phenomenon caused by the artificial potential function. Benefiting from this, the trained control strategy based on deep reinforcement learning technology can decrease the energy consumption by 16.44% during the close proximity phase, compared with the traditional artificial potential function method. Besides, after introducing the metalearning-based idea, a strategy available for some other perturbed scenarios can be trained in a relatively short period of time, which illustrates its adaptability.
引用
收藏
页码:5823 / 5834
页数:12
相关论文
共 50 条
  • [31] Learning-Based Navigation and Collision Avoidance Through Reinforcement for UAVs
    Azzam, Rana
    Chehadeh, Mohamad
    Hay, Oussama Abdul
    Humais, Muhammad Ahmed
    Boiko, Igor
    Zweiri, Yahya
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2024, 60 (03) : 2614 - 2628
  • [32] Deep reinforcement learning-based collision avoidance for an autonomous ship
    Chun, Do-Hyun
    Roh, Myung-Il
    Lee, Hye-Won
    Ha, Jisang
    Yu, Donghun
    OCEAN ENGINEERING, 2021, 234
  • [33] Online collision avoidance trajectory planning for spacecraft proximity operations with uncertain obstacle
    Zhang, Run-de
    Cai, Wei-wei
    Yang, Le-ping
    Si, Cheng
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART G-JOURNAL OF AEROSPACE ENGINEERING, 2022, 236 (11) : 2254 - 2270
  • [34] A collision-avoidance control algorithm for spacecraft proximity operations based on improved artificial potential function
    Xu D.
    Zhang J.
    Lixue Xuebao/Chinese Journal of Theoretical and Applied Mechanics, 2020, 52 (06): : 1581 - 1589
  • [35] Optimal-fuel Control Design of Nonlinear Spacecraft Rendezvous System with the Collision Avoidance Constraint
    Shao, Lili
    Liu, Zhongbo
    Gao, Xiangyu
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON ESTIMATION, DETECTION AND INFORMATION FUSION ICEDIF 2015, 2015, : 348 - 353
  • [36] A spacecraft attitude maneuvering path planning method based on PIO-improved reinforcement learning
    Hua B.
    Sun S.
    Wu Y.
    Chen Z.
    Zhongguo Kexue Jishu Kexue/Scientia Sinica Technologica, 2023, 53 (02): : 200 - 209
  • [37] Pedestrian Collision Avoidance Using Deep Reinforcement Learning
    Alireza Rafiei
    Amirhossein Oliaei Fasakhodi
    Farshid Hajati
    International Journal of Automotive Technology, 2022, 23 : 613 - 622
  • [38] Deep Reinforcement Learning for Collision Avoidance of Robotic Manipulators
    Sangiovanni, Bianca
    Rendiniello, Angelo
    Incremona, Gian Paolo
    Ferrara, Antonella
    Piastra, Marco
    2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, : 2063 - 2068
  • [39] Multilayered reinforcement learning for complicated collision avoidance problems
    Fujii, T
    Arai, Y
    Asama, H
    Endo, I
    1998 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-4, 1998, : 2186 - 2191
  • [40] Pedestrian Collision Avoidance Using Deep Reinforcement Learning
    Rafiei, Alireza
    Fasakhodi, Amirhossein Oliaei
    Hajati, Farshid
    INTERNATIONAL JOURNAL OF AUTOMOTIVE TECHNOLOGY, 2022, 23 (03) : 613 - 622