Attention-Shared Multi-Agent Actor-Critic-Based Deep Reinforcement Learning Approach for Mobile Charging Dynamic Scheduling in Wireless Rechargeable Sensor Networks

被引:13
作者
Jiang, Chengpeng [1 ,2 ]
Wang, Ziyang [3 ]
Chen, Shuai [1 ,2 ]
Li, Jinglin [1 ,2 ]
Wang, Haoran [1 ,2 ]
Xiang, Jinwei [1 ,2 ]
Xiao, Wendong [1 ,2 ]
机构
[1] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
[2] Beijing Engn Res Ctr Ind Spectrum Imaging, Beijing 100083, Peoples R China
[3] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
wireless rechargeable sensor network; deep reinforcement learning; multi-agent; attention-shared; mobile charging; LIFETIME; SCHEME;
D O I
10.3390/e24070965
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
The breakthrough of wireless energy transmission (WET) technology has greatly promoted the wireless rechargeable sensor networks (WRSNs). A promising method to overcome the energy constraint problem in WRSNs is mobile charging by employing a mobile charger to charge sensors via WET. Recently, more and more studies have been conducted for mobile charging scheduling under dynamic charging environments, ignoring the consideration of the joint charging sequence scheduling and charging ratio control (JSSRC) optimal design. This paper will propose a novel attention-shared multi-agent actor-critic-based deep reinforcement learning approach for JSSRC (AMADRL-JSSRC). In AMADRL-JSSRC, we employ two heterogeneous agents named charging sequence scheduler and charging ratio controller with an independent actor network and critic network. Meanwhile, we design the reward function for them, respectively, by considering the tour length and the number of dead sensors. The AMADRL-JSSRC trains decentralized policies in multi-agent environments, using a centralized computing critic network to share an attention mechanism, and it selects relevant policy information for each agent at every charging decision. Simulation results demonstrate that the proposed AMADRL-JSSRC can efficiently prolong the lifetime of the network and reduce the number of death sensors compared with the baseline algorithms.
引用
收藏
页数:23
相关论文
共 44 条
[1]   Wireless Sensor's Civil Applications, Prototypes, and Future Integration Possibilities: A Review [J].
Ayaz, Muhammad ;
Ammad-uddin, Mohammad ;
Baig, Imran ;
Aggoune, el-Hadi M. .
IEEE SENSORS JOURNAL, 2018, 18 (01) :4-30
[2]   A deep reinforcement learning-based on-demand charging algorithm for wireless rechargeable sensor networks [J].
Cao, Xianbo ;
Xu, Wenzheng ;
Liu, Xuxun ;
Peng, Jian ;
Liu, Tang .
AD HOC NETWORKS, 2021, 110
[3]  
Chawra VK, 2021, PEER PEER NETW APPL, V14, P1316, DOI 10.1007/s12083-021-01090-w
[4]  
Feiyu Chen, 2016, 2016 12th International Conference on Mobile Ad-Hoc and Sensor Networks (MSN). Proceedings, P63, DOI 10.1109/MSN.2016.018
[5]   Efficient Mobile Energy Replenishment Scheme Based on Hybrid Mode for Wireless Rechargeable Sensor Networks [J].
Feng, Yong ;
Guo, Lei ;
Fu, Xiaodong ;
Liu, Nianbo .
IEEE SENSORS JOURNAL, 2019, 19 (21) :10131-10143
[6]  
Feng Yong, 2016, P IEEE INT C COMM IC, P1
[7]   A Newborn Particle Swarm Optimization Algorithm for Charging-Scheduling Algorithm in Industrial Rechargeable Sensor Networks [J].
Feng, Yongxin ;
Zhang, Wenbo ;
Han, Guangjie ;
Kang, Yingyun ;
Wang, Jing .
IEEE SENSORS JOURNAL, 2020, 20 (18) :11014-11027
[8]  
Foerster JN, 2024, Arxiv, DOI arXiv:1705.08926
[9]  
Graves A, 2014, Arxiv, DOI [arXiv:1410.5401, 10.48550/arXiv.1410.5401]
[10]   Evaluating the On-Demand Mobile Charging in Wireless Sensor Networks [J].
He, Liang ;
Kong, Linghe ;
Gu, Yu ;
Pan, Jianping ;
Zhu, Ting .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2015, 14 (09) :1861-1875