Challenges and strategies for running controlled crowdsourcing experiments

被引:0
作者
Ramirez, Jorge [1 ]
Baez, Marcos [2 ]
Casati, Fabio [1 ]
Cernuzzi, Luca [3 ]
Benatallah, Boualem [4 ]
机构
[1] Univ Trento, Trento, Italy
[2] Univ Claude Bernard Lyon 1, Lyon, France
[3] Catholic Univ Nuestra Senora Asuncion, Asuncion, Paraguay
[4] Univ New South Wales, Sydney, NSW, Australia
来源
2020 XLVI LATIN AMERICAN COMPUTING CONFERENCE (CLEI 2020) | 2021年
基金
俄罗斯科学基金会;
关键词
Crowdsourcing; Task Design; Controlled experiments; Crowdsourcing Platforms; MECHANICAL TURK; PLATFORM;
D O I
10.1109/CLEI52000.2020.00036
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper reports on the challenges and lessons we learned while running controlled experiments in crowdsourcing platforms. Crowdsourcing is becoming an attractive technique to engage a diverse and large pool of subjects in experimental research, allowing researchers to achieve levels of scale and completion times that would otherwise not be feasible in lab settings. However, the scale and flexibility comes at the cost of multiple and sometimes unknown sources of bias and confounding factors that arise from technical limitations of crowdsourcing platforms and from the challenges of running controlled experiments in the "wild". In this paper, we take our experience in running systematic evaluations of task design as a motivating example to explore, describe, and quantify the potential impact of running uncontrolled crowdsourcing experiments and derive possible coping strategies. Among the challenges identified, we can mention sampling bias, controlling the assignment of subjects to experimental conditions, learning effects, and reliability of crowdsourcing results. According to our empirical studies, the impact of potential biases and confounding factors can amount to a 38% loss in the utility of the data collected in uncontrolled settings; and it can significantly change the outcome of experiments. These issues ultimately inspired us to implement CrowdHub, a system that sits on top of major crowdsourcing platforms and allows researchers and practitioners to run controlled crowdsourcing projects.
引用
收藏
页码:252 / 261
页数:10
相关论文
共 56 条
[1]  
Ahmad S., 2011, P 24 ANN ACM S US IN, P53
[2]  
Balahur A., 2010, P INT C LANGUAGE RES
[3]   Rehumanized Crowdsourcing: A Labeling Framework Addressing Bias and Ethics in Machine Learning [J].
Barbosa, Nata M. ;
Chen, Monchu .
CHI 2019: PROCEEDINGS OF THE 2019 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2019,
[4]   AUTOMAN: A Platform for Integrating Human-Based and Digital Computation [J].
Barowy, Daniel W. ;
Curtsinger, Charlie ;
Berger, Emery D. ;
McGregor, Andrew .
ACM SIGPLAN NOTICES, 2012, 47 (10) :639-654
[5]  
Bernstein M., 2010, P 23ND ANN ACM S US, V58, P313, DOI DOI 10.1145/2791285
[6]  
Blanco R, 2011, PROCEEDINGS OF THE 34TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR'11), P923
[7]   Amazon's Mechanical Turk: A New Source of Inexpensive, Yet High-Quality, Data? [J].
Buhrmester, Michael ;
Kwang, Tracy ;
Gosling, Samuel D. .
PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2011, 6 (01) :3-5
[8]  
Cheng J., 2013, Proceedings of the 24th acm conference on hypertext and social media, P214, DOI 10.1145/2481492.2481519
[9]   SciCrowd: Towards a Hybrid, Crowd-Computing System for Supporting Research Groups in Academic Settings [J].
Correia, Antonio ;
Schneider, Daniel ;
Paredes, Hugo ;
Fonseca, Benjamim .
COLLABORATION AND TECHNOLOGY (CRIWG 2018), 2018, 11001 :34-41
[10]   Evaluating Amazon's Mechanical Turk as a Tool for Experimental Behavioral Research [J].
Crump, Matthew J. C. ;
McDonnell, John V. ;
Gureckis, Todd M. .
PLOS ONE, 2013, 8 (03)