Autonomous Decision-Making for Aerobraking via Parallel Randomized Deep Reinforcement Learning

被引:6
作者
Falcone, Giusy [1 ,3 ]
Putnam, Zachary R. R. [2 ]
机构
[1] Univ Illinois, Champaign, IL 61801 USA
[2] Univ Illinois, Dept Aerosp Engn, Champaign, IL 61801 USA
[3] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
关键词
Space vehicles; Planetary orbits; Mars; Decision making; Computer architecture; Atmospheric modeling; Reinforcement learning; Aerobraking; deep reinforcement learning (DRL); domain randomization; ACCELEROMETER DATA; MARS; MISSION; COST;
D O I
10.1109/TAES.2022.3221697
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Aerobraking is used to insert a spacecraft into a low orbit around a planet through many orbital passages into its complex atmosphere. The aerobraking atmospheric passages are challenging because of the high variability of the atmospheric environment. This paper develops a parallel domain randomized deep reinforcement learning architecture for autonomous decision-making in a stochastic environment, such as aerobraking atmospheric passages. In this context, the architecture is used for planning aerobraking maneuvers to avoid the occurrence of thermal violations during the atmospheric aerobraking passages and target a final low-altitude orbit. The parallel domain randomized deep reinforcement learning architecture is designed to account for large variability of the physical model, as well as uncertain conditions. Also, the parallel approach speeds up the training process for simulation-based applications, and domain randomization improves resultant policy generalization. This framework is applied to the 2001 Mars Odyssey aerobraking campaign; with respect to the 2001 Mars Odyssey mission flight data and a Numerical Predictor Corrector (NPC)-based state-of-the-art heuristic for autonomous aerobraking, the proposed architecture outperforms the state-of-the-art heuristic algorithm with a decrease of 97.5% in the number of thermal violations. Furthermore, it yields a reduction of 98.7% in the number of thermal violations with respect to the Mars Odyssey mission flight data and requires 13.9% fewer orbits. Results also show that the proposed architecture can also learn a generalized policy in the presence of strong uncertainties, such as aggressive atmospheric density perturbations, different atmospheric density models, and a different simulator maximum step size and error accuracy.
引用
收藏
页码:3055 / 3070
页数:16
相关论文
共 50 条
  • [41] A Decision-Making Approach for Complex Unsignalized Intersection by Deep Reinforcement Learning
    Li, Shanke
    Peng, Kun
    Hui, Fei
    Li, Ziqi
    Wei, Cheng
    Wang, Wenbo
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (11) : 16134 - 16147
  • [42] Decision-making for the autonomous navigation of USVs based on deep reinforcement learning under IALA maritime buoyage system
    Zhao, Yiming
    Han, Fenglei
    Han, Duanfeng
    Peng, Xiao
    Zhao, Wangyuan
    OCEAN ENGINEERING, 2022, 266
  • [43] Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving
    Hoel, Carl-Johan
    Driggs-Campbell, Katherine
    Wolff, Krister
    Laine, Leo
    Kochenderfer, Mykel J.
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2020, 5 (02): : 294 - 305
  • [44] Reliable safety decision-making for autonomous vehicles: a safety assurance reinforcement learning
    Niu, Yuchen
    Wang, Yongjie
    Xiao, Mei
    Zhu, Wenying
    Wang, Tao
    TRANSPORTMETRICA B-TRANSPORT DYNAMICS, 2025, 13 (01)
  • [45] Benchmarking Lane-changing Decision-making for Deep Reinforcement Learning
    Wang, Junjie
    Zhang, Qichao
    Zhao, Dongbin
    2021 7TH INTERNATIONAL CONFERENCE ON ROBOTICS AND ARTIFICIAL INTELLIGENCE, ICRAI 2021, 2021, : 26 - 32
  • [46] Unified Local-Cloud Decision-Making via Reinforcement Learning
    Sengupta, Kathakoli
    Shangguan, Zhongkai
    Bharadwaj, Sandesh
    Arora, Sanjay
    Ohn-Bar, Eshed
    Mancuso, Renato
    COMPUTER VISION - ECCV 2024, PT XLI, 2025, 15099 : 185 - 203
  • [47] A fast decision-making method for process planning with dynamic machining resources via deep reinforcement learning
    Wu, Wenbo
    Huang, Zhengdong
    Zeng, Jiani
    Fan, Kuan
    JOURNAL OF MANUFACTURING SYSTEMS, 2021, 58 : 392 - 411
  • [48] Decision Making for Autonomous Driving via Augmented Adversarial Inverse Reinforcement Learning
    Wang, Pin
    Liu, Dapeng
    Chen, Jiayu
    Li, Hanhan
    Chan, Ching-Yao
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 1036 - 1042
  • [49] Combining reinforcement learning with rule-based controllers for transparent and general decision-making in autonomous driving
    Likmeta, Amarildo
    Metelli, Alberto Maria
    Tirinzoni, Andrea
    Giol, Riccardo
    Restelli, Marcello
    Romano, Danilo
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2020, 131 (131)
  • [50] REINFORCEMENT LEARNING FOR DECISION-MAKING IN A BUSINESS SIMULATOR
    Garcia, Javier
    Borrajo, Fernando
    Fernandez, Fernando
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGY & DECISION MAKING, 2012, 11 (05) : 935 - 960