Autonomous Decision-Making for Aerobraking via Parallel Randomized Deep Reinforcement Learning

被引:6
作者
Falcone, Giusy [1 ,3 ]
Putnam, Zachary R. R. [2 ]
机构
[1] Univ Illinois, Champaign, IL 61801 USA
[2] Univ Illinois, Dept Aerosp Engn, Champaign, IL 61801 USA
[3] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
关键词
Space vehicles; Planetary orbits; Mars; Decision making; Computer architecture; Atmospheric modeling; Reinforcement learning; Aerobraking; deep reinforcement learning (DRL); domain randomization; ACCELEROMETER DATA; MARS; MISSION; COST;
D O I
10.1109/TAES.2022.3221697
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Aerobraking is used to insert a spacecraft into a low orbit around a planet through many orbital passages into its complex atmosphere. The aerobraking atmospheric passages are challenging because of the high variability of the atmospheric environment. This paper develops a parallel domain randomized deep reinforcement learning architecture for autonomous decision-making in a stochastic environment, such as aerobraking atmospheric passages. In this context, the architecture is used for planning aerobraking maneuvers to avoid the occurrence of thermal violations during the atmospheric aerobraking passages and target a final low-altitude orbit. The parallel domain randomized deep reinforcement learning architecture is designed to account for large variability of the physical model, as well as uncertain conditions. Also, the parallel approach speeds up the training process for simulation-based applications, and domain randomization improves resultant policy generalization. This framework is applied to the 2001 Mars Odyssey aerobraking campaign; with respect to the 2001 Mars Odyssey mission flight data and a Numerical Predictor Corrector (NPC)-based state-of-the-art heuristic for autonomous aerobraking, the proposed architecture outperforms the state-of-the-art heuristic algorithm with a decrease of 97.5% in the number of thermal violations. Furthermore, it yields a reduction of 98.7% in the number of thermal violations with respect to the Mars Odyssey mission flight data and requires 13.9% fewer orbits. Results also show that the proposed architecture can also learn a generalized policy in the presence of strong uncertainties, such as aggressive atmospheric density perturbations, different atmospheric density models, and a different simulator maximum step size and error accuracy.
引用
收藏
页码:3055 / 3070
页数:16
相关论文
共 50 条
  • [21] Decision-Making for the Autonomous Navigation of Maritime Autonomous Surface Ships Based on Scene Division and Deep Reinforcement Learning
    Zhang, Xinyu
    Wang, Chengbo
    Liu, Yuanchang
    Chen, Xiang
    SENSORS, 2019, 19 (18)
  • [22] Deep imitative reinforcement learning with gradient conflict-free for decision-making in autonomous vehicles
    Shan, Zitong
    Zhao, Jian
    Huang, Wenhui
    Zhao, Yang
    Ge, Linhe
    Zhong, Shouren
    Hu, Hongyu
    Lv, Chen
    Zhu, Bing
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2025, 173
  • [23] A novel deep reinforcement learning for POMDP-based autonomous ship collision decision-making
    Zhang, Xinyu
    Zheng, Kangjie
    Wang, Chengbo
    Chen, Jihong
    Qi, Huaiyuan
    NEURAL COMPUTING & APPLICATIONS, 2023,
  • [24] Autonomous Industrial Management via Reinforcement Learning Towards Self-Learning Agents for Decision-Making
    Espinosa-Leal, Leonardo
    Chapman, Anthony
    Westerlund, Magnus
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2020, 39 (06) : 8427 - 8439
  • [25] Deploying Reinforcement Learning for Efficient Runtime Decision-Making in Autonomous Systems
    Dastranj, Melika
    Nia, Mehran Alidoost
    Kargahi, Mehdi
    2022 CPSSI 4TH INTERNATIONAL SYMPOSIUM ON REAL-TIME AND EMBEDDED SYSTEMS AND TECHNOLOGIES (RTEST 2022), 2022,
  • [26] Reinforcement Learning Based Overtaking Decision-Making for Highway Autonomous Driving
    Li, Xin
    Xu, Xin
    Zuo, Lei
    2015 SIXTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2015, : 336 - 342
  • [27] A reinforcement learning approach to autonomous decision-making in smart electricity markets
    Peters, Markus
    Ketter, Wolfgang
    Saar-Tsechansky, Maytal
    Collins, John
    MACHINE LEARNING, 2013, 92 (01) : 5 - 39
  • [28] Review of Autonomous Driving Decision-Making Research Based on Reinforcement Learning
    Jin L.
    Han G.
    Xie X.
    Guo B.
    Liu G.
    Zhu W.
    Qiche Gongcheng/Automotive Engineering, 2023, 45 (04): : 527 - 540
  • [29] A reinforcement learning approach to autonomous decision-making in smart electricity markets
    Markus Peters
    Wolfgang Ketter
    Maytal Saar-Tsechansky
    John Collins
    Machine Learning, 2013, 92 : 5 - 39
  • [30] Decision-making of autonomous vehicles in interactions with jaywalkers: A risk-aware deep reinforcement learning approach
    Zhang, Ziqian
    Li, Haojie
    Chen, Tiantian
    Sze, N. N.
    Yang, Wenzhang
    Zhang, Yihao
    Ren, Gang
    ACCIDENT ANALYSIS AND PREVENTION, 2025, 210