Possibilities of reinforcement learning for nuclear power plants: Evidence on current applications and beyond

被引:4
作者
Gong, Aicheng [1 ,2 ]
Chen, Yangkun [2 ]
Zhang, Junjie [2 ]
Li, Xiu [2 ]
机构
[1] China Nucl Power Engn Co Ltd, State Key Lab Nucl Power Safety Monitoring Technol, Shenzhen 518172, Peoples R China
[2] Shenzhen Int Grad Sch, Shenzhen 518055, Peoples R China
关键词
Nuclear Power Plant; Complex system; Reinforcement learning; Control; Artificial intelligence; SYSTEM; ALGORITHM; OPERATION; ENERGY; COMPLEXITY; ROBOTICS; DESIGN; LEVEL; MODEL;
D O I
10.1016/j.net.2024.01.003
中图分类号
TL [原子能技术]; O571 [原子核物理学];
学科分类号
0827 ; 082701 ;
摘要
Nuclear energy plays a crucial role in energy supply in the 21st century, and more and more Nuclear Power Plants (NPPs) will be in operation to contribute to the development of human society. However, as a typical complex system engineering, the operation and development of NPPs require efficient and stable control methods to ensure the safety and efficiency of nuclear power generation. Reinforcement learning (RL) aims at learning optimal control policies via maximizing discounted long-term rewards. The reward -oriented learning paradigm has witnessed remarkable success in many complex systems, such as wind power systems, electric power systems, coal fire power plants, robotics, etc. In this work, we try to present a systematic review of the applications of RL on these complex systems, from which we believe NPPs can borrow experience and insights. We then conduct a block -by -block investigation on the application scenarios of specific tasks in NPPs and carried out algorithmic research for different situations such as power startup, collaborative control, and emergency handling. Moreover, we discuss the possibilities of further application of RL methods on NPPs and detail the challenges when applying RL methods on NPPs. We hope this work can boost the realization of intelligent NPPs, and contribute to more and more research on how to better integrate RL algorithms into NPPs.
引用
收藏
页码:1959 / 1974
页数:16
相关论文
共 175 条
  • [1] Abdolmaleki A, 2018, Arxiv, DOI [arXiv:1806.06920, 10.48550/arXiv.1806.06920]
  • [2] Ahamed TPI, 2002, ELECTR POW SYST RES, V63, P9, DOI 10.1016/S0378-7796(02)00088-3
  • [3] Akkaya I, 2019, Arxiv, DOI arXiv:1910.07113
  • [4] AliKhaldi M, 2001, PHILOS REV, V110, P469
  • [5] A review on the applications of programmable logic controllers (PLCs)
    Alphonsus, Ephrem Ryan
    Abdullah, Mohammad Omar
    [J]. RENEWABLE & SUSTAINABLE ENERGY REVIEWS, 2016, 60 : 1185 - 1205
  • [6] Self-learning Processes in Smart Factories: Deep Reinforcement Learning for Process Control of Robot Brine Injection
    Andersen, Rasmus E.
    Madsen, Steffen
    Barlo, Alexander B. K.
    Johansen, Sebastian B.
    Nor, Morten
    Andersen, Rasmus S.
    Bogh, Simon
    [J]. 29TH INTERNATIONAL CONFERENCE ON FLEXIBLE AUTOMATION AND INTELLIGENT MANUFACTURING (FAIM 2019): BEYOND INDUSTRY 4.0: INDUSTRIAL ADVANCES, ENGINEERING EDUCATION AND INTELLIGENT MANUFACTURING, 2019, 38 : 171 - 177
  • [7] Learning dexterous in-hand manipulation
    Andrychowicz, Marcin
    Baker, Bowen
    Chociej, Maciek
    Jozefowicz, Rafal
    McGrew, Bob
    Pachocki, Jakub
    Petron, Arthur
    Plappert, Matthias
    Powell, Glenn
    Ray, Alex
    Schneider, Jonas
    Sidor, Szymon
    Tobin, Josh
    Welinder, Peter
    Weng, Lilian
    Zaremba, Wojciech
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2020, 39 (01) : 3 - 20
  • [8] [Anonymous], 1992, Reinforcement learning for robots using neural networks
  • [9] Arias J.M., 2008, An advanced course in modern nuclear physics, V581
  • [10] Arndt K, 2020, IEEE INT CONF ROBOT, P2725, DOI [10.1109/ICRA40945.2020.9196540, 10.1109/icra40945.2020.9196540]