共 51 条
Multi-UAV-Assisted Federated Learning for Energy-Aware Distributed Edge Training
被引:20
作者:
Tang, Jianhang
[1
]
Nie, Jiangtian
[2
]
Zhang, Yang
[3
]
Xiong, Zehui
[4
]
Jiang, Wenchao
[4
]
Guizani, Mohsen
[5
]
机构:
[1] Guizhou Univ, State Key Lab Publ Big Data, Guiyang 550025, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[3] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 210000, Peoples R China
[4] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore, Singapore
[5] Mohamed Bin Zayed Univ Artificial Intelligence, Machine Learning Dept, Abu Dhabi, U Arab Emirates
来源:
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT
|
2024年
/
21卷
/
01期
基金:
中国国家自然科学基金;
关键词:
UAV;
federated learning;
resource allocation;
client selection;
DRL;
RESOURCE-ALLOCATION;
ASSOCIATION;
SELECTION;
D O I:
10.1109/TNSM.2023.3298220
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Unmanned aerial vehicle (UAV)-assisted mobile edge computing (MEC) has largely extended the border and capacity of artificial intelligence of things (AIoT) by providing a key element for enabling flexible distributed data inputs, computing capacity, and high mobility. To enhance data privacy for AIoT applications, federated learning (FL) is becoming a potential solution to perform training tasks locally on distributed IoT devices. However, with the limited onboard resources and battery capacity of each UAV node, optimization is required to achieve a large-scale and high-precision FL scheme. In this work, an optimized multi-UAV-assisted FL framework is designed, where regular IoT devices are in charge of performing training tasks, and multiple UAVs are leveraged to execute local and global aggregation tasks. An online resource allocation (ORA) algorithm is proposed to minimize the training latency by jointly deciding the selection decisions of clients and a global aggregation server. By leveraging the Lyapunov optimization technique, virtual energy queues are studied to depict the energy deficit. With the help of the actor-critic learning framework, a deep reinforcement learning (DRL) scheme is designed to improve per-round training performance. A deep neural network (DNN)-based actor module is designed to derive client selection decisions, and a critic module is proposed through a conventional optimization method to evaluate the obtained selection decisions. Moreover, a greedy scheme is developed to find the optimal global aggregation server. Finally, extensive simulation results demonstrate that the proposed ORA algorithm can achieve optimal training latency and energy consumption under various system settings.
引用
收藏
页码:280 / 294
页数:15
相关论文