Decentralized Task Assignment for Mobile Crowdsensing With Multi-Agent Deep Reinforcement Learning

被引:20
作者
Xu, Chenghao [1 ]
Song, Wei [1 ]
机构
[1] Univ New Brunswick, Fac Comp Sci, Fredericton, NB E3B 5A3, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Task analysis; Sensors; Resource management; Privacy; Metaheuristics; Costs; Routing; Graph embedding; learning-communication; mobile crowdsensing (MCS); multi-agent deep reinforcement learning (DRL); QMIX; task assignment; ALLOCATION;
D O I
10.1109/JIOT.2023.3268846
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Task assignment is a fundamental research problem in mobile crowdsensing (MCS) since it directly determines an MCS system's practicality and economic value. Due to the complex dynamics of tasks and workers, task assignment problems are usually NP-hard, and approximation-based methods are preferred to impractical optimal methods. In the literature, a graph neural network-based deep reinforcement learning (GDRL) method is proposed in Xu and Song (2022) to solve routing problems in MCS and shows high performance and time efficiency. However, GDRL, as a centralized method, has to cope with the limitation in scalability and the challenge of privacy protection. In this article, we propose a multi-agent deep reinforcement learning-based method named communication-QMIX-based multi-agent DRL (CQDRL) to solve a task assignment problem in a decentralized fashion. The CQDRL method not only inherits the merits of GDRL over handcrafted heuristic and metaheuristic methods but also exploits computation potentials in mobile devices and protects workers' privacy with a decentralized decision-making scheme. Our extensive experiments show that the CQDRL method can achieve significantly better performance than other traditional methods and performs fairly close to the centralized GDRL method.
引用
收藏
页码:16564 / 16578
页数:15
相关论文
共 25 条
[1]   IntelligentCrowd: Mobile Crowdsensing via Multi-Agent Reinforcement Learning [J].
Chen, Yize ;
Wang, Hao .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2021, 5 (05) :840-845
[2]  
Dai HJ, 2017, ADV NEUR IN, V30
[3]  
Foerster JN, 2016, ADV NEUR IN, V29
[4]  
GOLDEN BL, 1987, NAV RES LOG, V34, P307, DOI 10.1002/1520-6750(198706)34:3<307::AID-NAV3220340302>3.0.CO
[5]  
2-D
[6]   Multiagent Actor-Critic Network-Based Incentive Mechanism for Mobile Crowdsensing in Industrial Systems [J].
Gu, Bo ;
Yang, Xinxin ;
Lin, Ziqi ;
Hu, Weiwei ;
Alazab, Mamoun ;
Kharel, Rupak .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (09) :6182-6191
[7]   Task Allocation in Spatial Crowdsourcing: Current State and Future Directions [J].
Guo, Bin ;
Liu, Yan ;
Wang, Leye ;
Li, Victor O. K. ;
Lam, Jacqueline C. K. ;
Yu, Zhiwen .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (03) :1749-1764
[8]  
Hamilton WL, 2018, Arxiv, DOI arXiv:1706.02216
[9]   Reinforcement Learning based Recommendation with Graph Convolutional Q-network [J].
Lei, Yu ;
Pei, Hongbin ;
Yan, Hanqi ;
Li, Wenjie .
PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, :1757-1760
[10]   Deep Reinforcement Learning for Load Balancing of Edge Servers in IoV [J].
Li, Pu ;
Xie, Wenxuan ;
Yuan, Ying ;
Chen, Chen ;
Wan, Shaohua .
MOBILE NETWORKS & APPLICATIONS, 2022, 27 (04) :1461-1474