A Neural-Based Bandit Approach to Mobile Crowdsourcing

被引:0
作者
Lin, Shouxu [1 ]
Yao, Yuhang [1 ]
Zhang, Pei [2 ]
Noh, Hae Young [3 ]
Joe-Wong, Carlee [1 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[2] Univ Michigan, Ann Arbor, MI 48109 USA
[3] Stanford Univ, Stanford, CA 94305 USA
来源
PROCEEDINGS OF THE 2022 THE 23RD ANNUAL INTERNATIONAL WORKSHOP ON MOBILE COMPUTING SYSTEMS AND APPLICATIONS (HOTMOBILE '22) | 2022年
关键词
mobile crowdsourcing; combinatorial bandit; neural networks; MULTIARMED BANDIT;
D O I
10.1145/3508396.3512886
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Mobile crowdsourcing has long promised to utilize the power of mobile crowds to reduce the time and monetary cost required to perform large-scale location-dependent tasks, e.g., environmental sensing. Assigning the right tasks to the right users, however, is a longstanding challenge: different users will be better suited for different tasks, which in turn will have different contributions to the overall crowdsourcing goal. Even worse, these relationships are generally unknown a priori and may change over time, particularly in mobile settings. The diversity of devices in the Internet of Things and diversity of new application tasks that they may run exacerbate these challenges. Thus, in this paper, we formulate the mobile crowdsourcing problem as a Contextual Combinatorial Volatile Multi-armed Bandit problem. Although prior work has attempted to learn the optimal user-task assignment based on user-specific side information, such formulations assume known structure in the relationships between contextual information, user suitability for each task, and the overall crowdsourcing goal. To relax these assumptions, we propose a Neural-MAB algorithm that can learn these relationships. We show that in a simulated mobile crowdsourcing application, our algorithm significantly outperforms existing multi-armed bandit baselines in settings with both known and unknown reward structures.
引用
收藏
页码:15 / 21
页数:7
相关论文
共 24 条
  • [1] Auer P, 2003, SIAM J COMPUT, V32, P48, DOI 10.1137/S0097539701398375
  • [2] Finite-time analysis of the multiarmed bandit problem
    Auer, P
    Cesa-Bianchi, N
    Fischer, P
    [J]. MACHINE LEARNING, 2002, 47 (2-3) : 235 - 256
  • [3] Bouneffouf D, 2019, Arxiv, DOI arXiv:1904.10040
  • [4] Chen L., 2018, NeurIPS, P3247
  • [5] Chen W., 2013, INT C MACHINE LEARNI, P151
  • [6] Cho E., 2011, P 17 ACM SIGKDD INT, P1082
  • [7] Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
  • [8] MAB-Based Reinforced Worker Selection Framework for Budgeted Spatial Crowdsensing
    Gao, Xiaofeng
    Chen, Shenwei
    Chen, Guihai
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (03) : 1303 - 1316
  • [9] Hazan E, 2012, J MACH LEARN RES, V13, P2903
  • [10] Islas L., 2019, 2019 IEEE PUN SECT I, P1, DOI [10.1109/ICEV.2019.8920457, DOI 10.1109/ICEV.2019.8920457]