Robust interplanetary trajectory design under multiple uncertainties via meta-reinforcement learning

被引:8
作者
Federici, Lorenzo [1 ]
Zavoli, Alessandro [2 ]
机构
[1] Univ Arizona, Dept Syst & Ind Engn, 1127 E James E Rogers Way, Tucson, AZ 85721 USA
[2] Sapienza Univ Rome, Dept Mech & Aerosp Engn, Via Eudossiana 18, I-00184 Rome, Italy
关键词
Meta-reinforcement learning; Robust trajectory design; Closed-loop guidance; Recurrent neural network; Proximal policy optimization; Stochastic optimal control; LOW-THRUST; GUIDANCE;
D O I
10.1016/j.actaastro.2023.10.018
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This paper focuses on the application of meta-reinforcement learning to the robust design of low-thrust interplanetary trajectories in the presence of multiple uncertainties. A closed-loop control policy is used to optimally steer the spacecraft to a final target state despite the considered perturbations. The control policy is approximated by a deep recurrent neural network, trained by policy-gradient reinforcement learning on a collection of environments featuring mixed sources of uncertainty, namely dynamic uncertainty and control execution errors. The recurrent network is able to build an internal representation of the distribution of environments, thus better adapting the control to the different stochastic scenarios. The results in terms of optimality, constraint handling, and robustness on a fuel-optimal low-thrust transfer between Earth and Mars are compared with those obtained via a traditional reinforcement learning approach based on a feed-forward neural network.
引用
收藏
页码:147 / 158
页数:12
相关论文
共 43 条
[31]   Tube Stochastic Optimal Control for Nonlinear Constrained Trajectory Optimization Problems [J].
Ozaki, Naoya ;
Campagnola, Stefano ;
Funase, Ryu .
JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2020, 43 (04) :645-655
[32]   Deep learning in robotics: a review of recent research [J].
Pierson, Harry A. ;
Gashler, Michael S. .
ADVANCED ROBOTICS, 2017, 31 (16) :821-835
[33]   Coupling of system resource margins through the use of electric propulsion: Implications in preparing for the Dawn mission to Ceres and Vesta [J].
Rayman, Marc D. ;
Fraschetti, Thomas C. ;
Raymond, Carol A. ;
Russell, Christopher T. .
ACTA ASTRONAUTICA, 2007, 60 (10-11) :930-938
[34]   Minimum-Fuel Closed-Loop Powered Descent Guidance with Stochastically Derived Throttle Margins [J].
Ridderhof, Jack ;
Tsiotras, Panagiotis .
JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2021, 44 (03) :537-547
[35]  
Rubinsztejn A., 2020, AAS AIAA ASTR SPEC C
[36]   Physics-Informed Neural Networks for Optimal Planar Orbit Transfers [J].
Schiassi, Enrico ;
D'Ambrosio, Andrea ;
Drozd, Kristofer ;
Curti, Fabio ;
Furfaro, Roberto .
JOURNAL OF SPACECRAFT AND ROCKETS, 2022, 59 (03) :834-849
[37]  
Schulman J, 2018, Arxiv, DOI [arXiv:1506.02438, 10.48550/arXiv.1506.02438]
[38]  
Schulman J, 2017, Arxiv, DOI arXiv:1707.06347
[39]   Meta-learning in Reinforcement Learning [J].
Schweighofer, N ;
Doya, K .
NEURAL NETWORKS, 2003, 16 (01) :5-9
[40]   Image-Based Deep Reinforcement Meta-Learning for Autonomous Lunar Landing [J].
Scorsoglio, Andrea ;
D'Ambrosio, Andrea ;
Ghilardi, Luca ;
Gaudet, Brian ;
Curti, Fabio ;
Furfaro, Roberto .
JOURNAL OF SPACECRAFT AND ROCKETS, 2022, 59 (01) :153-165