Adaptive optics control with multi-agent model-free reinforcement learning

被引:34
|
作者
Pou, B. [1 ,2 ]
Ferreira, F. [3 ]
Quinones, E. [1 ]
Gratadour, D. [3 ,4 ]
Martin, M. [2 ]
机构
[1] Barcelona Supercotnputing Ctr BSC, C Jordi Girona 29, Barcelona 08034, Spain
[2] Univ Politecn Catalunya UPC, Comp Sci Dept, C Jordi Girona 31, Barcelona 08034, Spain
[3] Univ Paris Diderot, Univ PSL, Sorbonne Paris Cite, Sorbonne Univ,CNRS,Observ Paris,LESIA, 5 Pl Jules Janssen, F-92195 Meudon, France
[4] Australian Natl Univ, Res Sch Astron & Astrophys, Canberra, ACT 2611, Australia
关键词
QUADRATIC GAUSSIAN CONTROL; WAVE-FRONT RECONSTRUCTION;
D O I
10.1364/OE.444099
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
We present a novel formulation of closed-loop adaptive optics (AO) control as a multi-agent reinforcement learning (MARL) problem in which the controller is able to learn a non-linear policy and does not need a priori information on the dynamics of the atmosphere. We identify the different challenges of applying a reinforcement learning (RL) method to AO and, to solve them, propose the combination of model-free MARL for control with an autoencoder neural network to mitigate the effect of noise. Moreover, we extend current existing methods of error budget analysis to include a RL controller. The experimental results for an 8m telescope equipped with a 40x40 Shack-Hartmann system show a significant increase in performance over the integrator baseline and comparable performance to a model-based predictive approach, a linear quadratic Gaussian controller with perfect knowledge of atmospheric conditions. Finally, the error budget analysis provides evidence that the RL controller is partially compensating for bandwidth error and is helping to mitigate the propagation of aliasing. (C) 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement
引用
收藏
页码:2991 / 3015
页数:25
相关论文
共 50 条
  • [1] On Improving Model-Free Algorithms for Decentralized Multi-Agent Reinforcement Learning
    Mao, Weichao
    Yang, Lin F.
    Zhang, Kaiqing
    Basar, Tamer
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] Model-Free Adaptive Iterative Learning Bipartite Containment Control for Multi-Agent Systems
    Sang, Shangyu
    Zhang, Ruikun
    Lin, Xue
    SENSORS, 2022, 22 (19)
  • [3] Model-Free Reinforcement Learning for Fully Cooperative Multi-Agent Graphical Games
    Zhang, Qichao
    Zhao, Dongbin
    Lewis, Frank L.
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [4] Model-free adaptive cluster consensus control for nonlinear multi-agent systems
    Li Y.-H.
    Cui L.-Z.
    Bu X.-H.
    Guo J.-L.
    Kongzhi yu Juece/Control and Decision, 2024, 39 (01): : 345 - 352
  • [5] Multi-Agent Pattern Formation: a Distributed Model-Free Deep Reinforcement Learning Approach
    Diallo, Elhadji Amadou Oury
    Sugawara, Toshiharu
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [6] Model-Free Adaptive Learning Formation Control of Nonlinear Non-Affine Multi-Agent Systems
    Yu, Xian
    Hou, Zhongsheng
    Polycarpou, Marios M.
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 4037 - 4042
  • [7] Optimal Cooperative Control of Multi-Agent Systems Through Event-Triggered Model-Free Reinforcement Learning
    Mu, Chaoxu
    Tang, Zhuo
    Wang, Ke
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [8] Model-free adaptive formation control for unknown heterogeneous nonlinear multi-agent systems
    Jin S.-T.
    Li C.
    Ren Y.
    Hou Z.-S.
    Kongzhi yu Juece/Control and Decision, 2020, 35 (06): : 1519 - 1524
  • [9] Distributed model-free adaptive predictive control for heterogeneous nonlinear multi-agent systems
    Pan, Zhenzhen
    Hou, Zhongsheng
    Chi, Ronghu
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2022, 53 (14) : 3027 - 3041
  • [10] Event-triggered model-free adaptive iterative learning bipartite consensus control for multi-agent systems
    Zhao H.-R.
    Peng L.
    Xie L.-B.
    Wu P.-L.
    Chen Y.-H.
    Kongzhi yu Juece/Control and Decision, 2022, 37 (10): : 2552 - 2558