Joint Optimization Framework for Minimization of Device Energy Consumption in Transmission Rate Constrained UAV-Assisted IoT Network

被引:21
作者
Mondal, Abhishek [1 ]
Mishra, Deepak [2 ]
Prasad, Ganesh [1 ]
Hossain, Ashraf [1 ]
机构
[1] Natl Inst Technol Silchar, Dept Elect & Commun Engn, Silchar 788010, India
[2] Univ New South Wales, Sch Elect Engn & Telecommun, Sydney, NSW 2052, Australia
关键词
Trajectory; Unmanned aerial vehicles; Energy consumption; Optimization; Energy efficiency; Data collection; Internet of Things; Internet of Things (IoT); joint optimization; reinforcement learning (RL); state-action-reward-state-action (SARSA); unmanned aerial vehicle (UAV); TRAJECTORY OPTIMIZATION; FAIR COMMUNICATION; EFFICIENT; DESIGN; ALLOCATION; INTERNET; SYSTEM;
D O I
10.1109/JIOT.2021.3128883
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Due to their high maneuverability and flexible deployment, unmanned aerial vehicles (UAVs) could be an alternative option for a scenario where Internet of Things (IoT) devices consume high energy to achieve the required data rate when they are far away from the terrestrial base station (BS). Therefore, this article has proposed an energy-efficient UAV-assisted IoT network where a low-altitude quad-rotor UAV provides mobile data collection service from static IoT devices. We develop a novel optimization framework that minimizes the total energy consumption of all devices by jointly optimizing the UAV's trajectory, devices association, and respectively, transmit power allocation at every time slot while ensuring that every device should achieve a given data rate constraint. As this joint optimization problem is nonconvex and combinatorial, we adopt a reinforcement learning (RL)-based solution methodology that effectively decouples it into three individual optimization subproblems. The formulated optimization problem has transformed into a Markov decision process (MDP) where the UAV learns its trajectory according to its current state and corresponding action for maximizing the generated reward under the current policy. Finally, we conceive state-action-reward-state-action, a low complexity iterative algorithm for updating the current policy of UAV, that achieves an excellent computational complexity-optimality tradeoff. Numerical results validate the analysis and provide various insights on optimal UAV trajectory. The proposed methodology reduces the total energy consumption of all devices by 6.91%, 8.48%, and 9.94% in 80, 100, and 120 available time slots of UAV, respectively, compared to the particle swarm optimization (PSO) algorithm.
引用
收藏
页码:9591 / 9607
页数:17
相关论文
共 37 条
[1]   Optimal Wireless Charging Inclusive of Intellectual Routing Based on SARSA Learning in Renewable Wireless Sensor Networks [J].
Aslam, Nelofar ;
Xia, Kewen ;
Hadi, Muhammad Usman .
IEEE SENSORS JOURNAL, 2019, 19 (18) :8340-8351
[2]   Fog and IoT: An Overview of Research Opportunities [J].
Chiang, Mung ;
Zhang, Tao .
IEEE INTERNET OF THINGS JOURNAL, 2016, 3 (06) :854-864
[3]   Adaptive UAV-Trajectory Optimization Under Quality of Service Constraints: A Model-Free Solution [J].
Cui, Jingjing ;
Ding, Zhiguo ;
Deng, Yansha ;
Nallanathan, Arumugam ;
Hanzo, Lajos .
IEEE ACCESS, 2020, 8 :112253-112265
[4]   Internet of Things Mobile-Air Pollution Monitoring System (IoT-Mobair) [J].
Dhingra, Swati ;
Madda, Rajasekhara Babu ;
Gandomi, Amir H. ;
Patan, Rizwan ;
Daneshmand, Mahmoud .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (03) :5577-5584
[5]   3D UAV Trajectory Design and Frequency Band Allocation for Energy-Efficient and Fair Communication: A Deep Reinforcement Learning Approach [J].
Ding, Ruijin ;
Gao, Feifei ;
Shen, Xuemin .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (12) :7796-7809
[6]   An Overview of Internet of Things (IoT) and Data Analytics in Agriculture: Benefits and Challenges [J].
Elijah, Olakunle ;
Rahman, Tharek Abdul ;
Orikumhi, Igbafe ;
Leow, Chee Yen ;
Hindia, M. H. D. Nour .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (05) :3758-3773
[7]  
Filippone A, 2006, FLIGHT PERFORMANCE F
[8]   UAV Control for Wireless Service Provisioning in Critical Demand Areas: A Deep Reinforcement Learning Approach [J].
Ho, Tai Manh ;
Kim-Khoa Nguyen ;
Cheriet, Mohamed .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (07) :7138-7152
[9]   Reinforcement Learning for Decentralized Trajectory Design in Cellular UAV Networks With Sense-and-Send Protocol [J].
Hu, Jingzhi ;
Zhang, Hongliang ;
Song, Lingyang .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (04) :6177-6189
[10]   Adaptive UAV-Assisted Geographic Routing With Q-Learning in VANET [J].
Jiang, Shanshan ;
Huang, Zhitong ;
Ji, Yuefeng .
IEEE COMMUNICATIONS LETTERS, 2021, 25 (04) :1358-1362