Energy-Efficient UAV-Enabled Data Collection via Wireless Charging: A Reinforcement Learning Approach

被引:91
作者
Fu, Shu [1 ,2 ]
Tang, Yujie [3 ]
Wu, Yuan [4 ,5 ]
Zhang, Ning [6 ]
Gu, Huaxi [2 ]
Chen, Chen [1 ]
Liu, Min [1 ]
机构
[1] Chongqing Univ, Sch Microelect & Commun Engn, Chongqing 400044, Peoples R China
[2] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[3] Algoma Univ, Dept Comp Sci & Math, Sault Ste Marie, ON P6A 2G4, Canada
[4] Univ Macau, State Key Lab Internet Things Smart City, Macau, Peoples R China
[5] Univ Macau, Dept Comp Informat Sci, Macau, Peoples R China
[6] Univ Windsor, Dept Elect & Comp Engn, Windsor, ON N9B 3P4, Canada
基金
中国国家自然科学基金;
关键词
Data collection; design of reward function; energy efficiency; Q-learning; reinforcement learning; unmanned aerial vehicle (UAV); DEPLOYMENT; NETWORKS; DESIGN; COMMUNICATION; ALLOCATION; INTERNET; CHANNEL;
D O I
10.1109/JIOT.2021.3051370
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, we study the application of unmanned aerial vehicle (UAV) for data collection with wireless charging, which is crucial for providing seamless coverage and improving system performance in the next-generation wireless networks. To this end, we propose a reinforcement learning-based approach to plan the route of UAV to collect sensor data from sensor devices scattered in the physical environment. Specifically, the physical environment is divided into multiple grids, where one spot for UAV hovering as well as the wireless charging of UAV is located at the center of each grid. Each grid has a spot for the UAV to hover, and moreover, there is a wireless charger at the center of each grid, which can provide wireless charging to UAV when it is hovering in the grid. When the UAV lacks energy, it can be charged by the wireless charger at the spot. By taking into account the collected data amount as well as the energy consumption, we formulate the problem of data collection with UAV as a Markov decision problem, and exploit Q-learning to find the optimal policy. In particular, we design the reward function considering the energy efficiency of UAV flight and data collection, based on which Q-table is updated for guiding the route of UAV. Through extensive simulation results, we verify that our proposed reward function can achieve a better performance in terms of the average throughput, delay of data collection, as well as the energy efficiency of UAV, in comparison with the conventional capacity-based reward function.
引用
收藏
页码:10209 / 10219
页数:11
相关论文
共 44 条
[11]   Integrated Resource Management for Terrestrial-Satellite Systems [J].
Fu, Shu ;
Gao, Jie ;
Zhao, Lian .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (03) :3256-3266
[12]   UAVs as Mobile Infrastructure: Addressing Battery Lifetime [J].
Galkin, Boris ;
Kibilda, Jacek ;
DaSilva, Luiz A. .
IEEE COMMUNICATIONS MAGAZINE, 2019, 57 (06) :132-137
[13]   Flight Time Minimization of UAV for Data Collection Over Wireless Sensor Networks [J].
Gong, Jie ;
Chang, Tsung-Hui ;
Shen, Chao ;
Chen, Xiang .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (09) :1942-1954
[14]  
Hu B, 2019, CHINA COMMUN, V16, P1
[15]   Cooperative Jamming for Physical Layer Security Enhancement in Internet of Things [J].
Hu, Lin ;
Wen, Hong ;
Wu, Bin ;
Pan, Fei ;
Liao, Run-Fa ;
Song, Huanhuan ;
Tang, Jie ;
Wang, Xiumin .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (01) :219-228
[16]  
Jawhar I., 2015, P INT C UNM AIRCR SY, P1
[17]   Performance Analysis of UAV Relay Assisted IoT Communication Network Enhanced With Energy Harvesting [J].
Ji, Baofeng ;
Li, Yuqi ;
Zhou, Benchuan ;
Li, Chunguo ;
Song, Kang ;
Wen, Hong .
IEEE ACCESS, 2019, 7 :38738-38747
[18]   An Energy-Efficient Data Collection Scheme Using Denoising Autoencoder in Wireless Sensor Networks [J].
Li, Guorui ;
Peng, Sancheng ;
Wang, Cong ;
Niu, Jianwei ;
Yuan, Ying .
TSINGHUA SCIENCE AND TECHNOLOGY, 2019, 24 (01) :86-96
[19]   Joint Optimization on Trajectory, Altitude, Velocity, and Link Scheduling for Minimum Mission Time in UAV-Aided Data Collection [J].
Li, Jiaxun ;
Zhao, Haitao ;
Wang, Haijun ;
Gu, Fanglin ;
Wei, Jibo ;
Yin, Hao ;
Ren, Baoquan .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (02) :1464-1475
[20]   Off-Policy Q-Learning: Set-Point Design for Optimizing Dual-Rate Rougher Flotation Operational Processes [J].
Li, Jinna ;
Chai, Tianyou ;
Lewis, Frank L. ;
Fan, Jialu ;
Ding, Zhengtao ;
Ding, Jinliang .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2018, 65 (05) :4092-4102