Performance comparison of the quantum and classical deep Q-learning approaches in dynamic environments control

被引:0
作者
Zare, Aramchehr [1 ]
Boroushaki, Mehrdad [1 ]
机构
[1] Sharif Univ Technol, Dept Energy Engn, POB 14565-114, Tehran, Iran
关键词
Quantum Deep Q-learning Network; Reinforcement Learning; Quantum Ansatz; Dynamic environments;
D O I
10.1140/epjqt/s40507-025-00381-y
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
There is a lack of adequate studies on dynamic environments control for Quantum Reinforcement Learning (QRL) algorithms, representing a significant gap in this field. This study contributes to bridging this gap by demonstrating the potential of quantum RL algorithms to effectively handle dynamic environments. In this research, the performance and robustness of Quantum Deep Q-learning Networks (DQN) were examined in two dynamic environments, Cart Pole and Lunar Lander, by using three distinct quantum Ansatz layers: RealAmplitudes, EfficientSU2, and TwoLocal. The quantum DQNs were compared with classical DQN algorithms in terms of convergence speed, loss minimization, and Q-value behavior. It was observed that the RealAmplitudes Ansatz outperformed the other quantum circuits, demonstrating faster convergence and superior performance in minimizing the loss function. To assess robustness, the pole length was increased in the Cart Pole environment, and a wind function was added to the Lunar Lander environment after the 50th episode. All three quantum Ansatz layers were found to maintain robust performance under disturbed conditions, with consistent reward values, loss minimization, and stable Q-value distributions. Although the proposed QRL demonstrates competitive results overall, classical RL can surpass them in convergence speed under specific conditions.
引用
收藏
页数:24
相关论文
共 22 条
[1]   Reinforcement Quantum Annealing: A Hybrid Quantum Learning Automata [J].
Ayanzadeh, Ramin ;
Halem, Milton ;
Finin, Tim .
SCIENTIFIC REPORTS, 2020, 10 (01)
[2]   NEURONLIKE ADAPTIVE ELEMENTS THAT CAN SOLVE DIFFICULT LEARNING CONTROL-PROBLEMS [J].
BARTO, AG ;
SUTTON, RS ;
ANDERSON, CW .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1983, 13 (05) :834-846
[3]   Variational quantum algorithms [J].
Cerezo, M. ;
Arrasmith, Andrew ;
Babbush, Ryan ;
Benjamin, Simon C. ;
Endo, Suguru ;
Fujii, Keisuke ;
McClean, Jarrod R. ;
Mitarai, Kosuke ;
Yuan, Xiao ;
Cincio, Lukasz ;
Coles, Patrick J. .
NATURE REVIEWS PHYSICS, 2021, 3 (09) :625-644
[4]   Variational Quantum Circuits for Deep Reinforcement Learning [J].
Chen, Samuel Yen-Chi ;
Yang, Chao-Han Huck ;
Qi, Jun ;
Chen, Pin-Yu ;
Ma, Xiaoli ;
Goan, Hsi-Sheng .
IEEE ACCESS, 2020, 8 :141007-141024
[5]  
Combarro E., 2023, A Practical Guide to Quantum Machine Learning and Quantum Optimization: Hands-on Approach to Modern Quantum Algorithms, V1
[6]  
Dunjko V, 2017, Arxiv, DOI arXiv:1709.02779
[7]  
Gallistel CR, 1999, J COGNITIVE NEUROSCI, V11, P126
[8]   Learning quantum systems [J].
Gebhart, Valentin ;
Santagati, Raffaele ;
Gentile, Antonio Andrea ;
Gauger, Erik M. ;
Craig, David ;
Ares, Natalia ;
Banchi, Leonardo ;
Marquardt, Florian ;
Pezze, Luca ;
Bonato, Cristian .
NATURE REVIEWS PHYSICS, 2023, 5 (03) :141-156
[9]  
Jerbi S, 2021, ADV NEUR IN, V34
[10]   Quantum reinforcement learning during human decision-making [J].
Li, Ji-An ;
Dong, Daoyi ;
Wei, Zhengde ;
Liu, Ying ;
Pan, Yu ;
Nori, Franco ;
Zhang, Xiaochu .
NATURE HUMAN BEHAVIOUR, 2020, 4 (03) :294-307