Deep Reinforcement Learning Based Data Collection with Charging Stations

被引:1
作者
Hao, Fuxin [1 ]
Hu, Yifan [2 ]
Fu, Junjie [2 ,3 ]
机构
[1] Southeast Univ, Sch Software Engn, Suzhou, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing, Peoples R China
[3] Purple Mt Labs, Nanjing, Peoples R China
来源
2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC | 2023年
基金
中国国家自然科学基金;
关键词
data collection; deep reinforcement learning; wireless communication; wireless charging;
D O I
10.1109/CCDC58219.2023.10327135
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Incorporating efficient charging strategies into the trajectory planning process for unmanned aerial vehicles (UAVs) data collection missions remains a difficult task. In this paper, we propose a reinforcement learning (RL) approach for training trajectory planning policies which jointly considers data collection and charging. Firstly, a trajectory planning optimization problem constrained by charging and other environmental constraints is formulated. Secondly, a Markov decision process is constructed based on the proposed optimization problem. Then, the deep RL algorithm DDQN is employed to obtain the optimal trajectory planning policies, where the convolutional layers in the Q-network are utilized to extract the charging and other environmental information for decision-making. Finally, a custom data collection environment is built, and the simulation results demonstrate that the UAV successfully learns to collect more data through charging as well as satisfying the safety constraints guided by the trained policy.
引用
收藏
页码:3344 / 3349
页数:6
相关论文
共 16 条
[1]  
[Anonymous], MICROMACHINES
[2]  
Bayerlein H., 2020, GLOBECOM 2020 2020 I, P1
[3]   Resonant Beam Charging-Powered UAV-Assisted Sensing Data Collection [J].
Chen, Weichao ;
Zhao, Shengjie ;
Shi, Qingjiang ;
Zhang, Rongqing .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (01) :1086-1090
[4]   Vehicle Routing Problems for Drone Delivery [J].
Dorling, Kevin ;
Heinrichs, Jordan ;
Messier, Geoffrey G. ;
Magierowski, Sebastian .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (01) :70-85
[5]  
Elloumi M, 2018, IEEE WCNC
[6]  
Esrafilian O., 2018, IEEE IOT J, V6, P1791
[7]   Energy-Constrained Completion Time Minimization in UAV-Enabled Internet of Things [J].
Gu, Jiangchun ;
Wang, Haichao ;
Ding, Guoru ;
Xu, Yitao ;
Xue, Zhen ;
Zhou, Huaji .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (06) :5491-5503
[8]   A Novel UAV-Enabled Data Collection Scheme for Intelligent Transportation System Through UAV Speed Control [J].
Li, Xiong ;
Tan, Jiawei ;
Liu, Anfeng ;
Vijayakumar, Pandi ;
Kumar, Neeraj ;
Alazab, Mamoun .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (04) :2100-2110
[9]   Energy-Efficient UAV Control for Effective and Fair Communication Coverage: A Deep Reinforcement Learning Approach [J].
Liu, Chi Harold ;
Chen, Zheyu ;
Tang, Jian ;
Xu, Jie ;
Piao, Chengzhe .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (09) :2059-2070
[10]  
Liu J, 2018, IEEE CONF COMPUT, P553, DOI 10.1109/INFCOMW.2018.8406973