Vibration Control with Reinforcement Learning Based on Multi-Reward Lightweight Networks

被引:0
作者
Shu, Yucheng [1 ]
He, Chaogang [1 ]
Qiao, Lihong [1 ]
Xiao, Bin [1 ]
Li, Weisheng [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 09期
基金
中国国家自然科学基金;
关键词
active vibration control; reinforcement learning; lightweight; neural network; prioritized experience replaying; reward function; ALGORITHM;
D O I
10.3390/app14093853
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This paper proposes a reinforcement learning method using a deep residual shrinkage network based on multi-reward priority experience playback for high-frequency and high-dimensional continuous vibration control. Firstly, we keep the underlying equipment unchanged and construct a vibration system simulator using FIR filters to ensure the complete fidelity of the physical model. Then, by interacting with the simulator using our proposed algorithm, we identify the optimal control strategy, which is directly applied to real-world scenarios in the form of a neural network. A multi-reward mechanism is proposed to assist the lightweight network to find a near-optimal control strategy, and a priority experience playback mechanism is used to prioritize the data to accelerate the convergence speed of the neural network and improve the data utilization efficiency. At the same time, the deep residual shrinkage network is introduced to realize adaptive denoising and lightweightness of the neural network. The experimental results indicate that under narrowband white-noise excitation ranging from 0 to 100 Hz, the DDPG algorithm achieved a vibration reduction effect of 12.728 dB, while our algorithm achieved a vibration reduction effect of 20.240 dB. Meanwhile, the network parameters were reduced by more than 7.5 times.
引用
收藏
页数:28
相关论文
共 40 条
  • [1] Vibration control of semi-active suspension system using PID controller with advanced firefly algorithm and particle swarm optimization
    Ab Talib, Mat Hussin
    Mat Darus, Intan Zaurah
    Mohd Samin, Pakharuddin
    Mohd Yatim, Hanim
    Ardani, Mohd Ibthisham
    Shaharuddin, Nik Mohd Ridzuan
    Hadi, Muhamad Sukri
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (01) : 1119 - 1137
  • [2] Baselizadeh Adel, 2022, 2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC), P1584, DOI 10.1109/SMC53654.2022.9945504
  • [3] Adaptive low-level control of autonomous underwater vehicles using deep reinforcement learning
    Carlucho, Ignacio
    De Paula, Mariano
    Wang, Sen
    Petillot, Yvan
    Acosta, Gerardo G.
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2018, 107 : 71 - 86
  • [4] Magnetic control of tokamak plasmas through deep reinforcement learning
    Degrave, Jonas
    Felici, Federico
    Buchli, Jonas
    Neunert, Michael
    Tracey, Brendan
    Carpanese, Francesco
    Ewalds, Timo
    Hafner, Roland
    Abdolmaleki, Abbas
    de las Casas, Diego
    Donner, Craig
    Fritz, Leslie
    Galperti, Cristian
    Huber, Andrea
    Keeling, James
    Tsimpoukelli, Maria
    Kay, Jackie
    Merle, Antoine
    Moret, Jean-Marc
    Noury, Seb
    Pesamosca, Federico
    Pfau, David
    Sauter, Olivier
    Sommariva, Cristian
    Coda, Stefano
    Duval, Basil
    Fasoli, Ambrogio
    Kohli, Pushmeet
    Kavukcuoglu, Koray
    Hassabis, Demis
    Riedmiller, Martin
    [J]. NATURE, 2022, 602 (7897) : 414 - +
  • [5] DE-NOISING BY SOFT-THRESHOLDING
    DONOHO, DL
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1995, 41 (03) : 613 - 627
  • [6] Feiran Zhao, 2021, IEEE Transactions on Artificial Intelligence, V2, P341, DOI 10.1109/TAI.2021.3097313
  • [7] A New Vibration Controller Design Method Using Reinforcement Learning and FIR Filters: A Numerical and Experimental Study
    Feng, Xingxing
    Chen, Hong
    Wu, Gang
    Zhang, Anfu
    Zhao, Zhigao
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (19):
  • [8] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [9] Howard AG, 2017, Arxiv, DOI arXiv:1704.04861
  • [10] Haarnoja T, 2019, Arxiv, DOI arXiv:1812.05905