Methods to detect low quality data and its implication for psychological research

被引:167
作者
Buchanan, Erin M. [1 ]
Scofield, John E. [2 ]
机构
[1] Missouri State Univ, Dept Psychol, 901 S Natl Ave, Springfield, MO 65804 USA
[2] Univ Missouri, Dept Psychol Sci, Columbia, MO 65211 USA
关键词
Amazon Mechanical Turk; Survey automation; Participant screening; Data quality; MECHANICAL TURK; PARTICIPANTS;
D O I
10.3758/s13428-018-1035-6
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
Web-based data collection methods such as Amazon's Mechanical Turk (AMT) are an appealing option to recruit participants quickly and cheaply for psychological research. While concerns regarding data quality have emerged with AMT, several studies have exhibited that data collected via AMT are as reliable as traditional college samples and are often more diverse and representative of noncollege populations. The development of methods to screen for low quality data, however, has been less explored. Omitting participants based on simple screening methods in isolation, such as response time or attention checks may not be adequate identification methods, with an inability to delineate between high or low effort participants. Additionally, problematic survey responses may arise from survey automation techniques such as survey bots or automated form fillers. The current project developed low quality data detection methods while overcoming previous screening limitations. Multiple checks were employed, such as page response times, distribution of survey responses, the number of utilized choices from a given range of scale options, click counts, and manipulation checks. This method was tested on a survey taken with an easily available plug-in survey bot, as well as compared to data collected by human participants providing both high effort and randomized, or low effort, answers. Identified cases can then be used as part of sensitivity analyses to warrant exclusion from further analyses. This algorithm can be a promising tool to identify low quality or automated data via AMT or other online data collection platforms.
引用
收藏
页码:2586 / 2596
页数:11
相关论文
共 31 条
[1]   Measuring Resilience With the RS-14: A Tale of Two Samples [J].
Aiena, Bethany J. ;
Baczwaski, Brandy J. ;
Schulenberg, Stefan E. ;
Buchanan, Erin M. .
JOURNAL OF PERSONALITY ASSESSMENT, 2015, 97 (03) :291-300
[2]  
[Anonymous], PSYCHOL EXPT INTERNE
[3]  
[Anonymous], 3 ASS ADV ART INT HU
[4]  
[Anonymous], MOTE
[5]  
[Anonymous], P ACM SIGIR 2010 WOR
[6]  
[Anonymous], 2008, P 1 IEEE WORKSH INT
[7]   Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk [J].
Berinsky, Adam J. ;
Huber, Gregory A. ;
Lenz, Gabriel S. .
POLITICAL ANALYSIS, 2012, 20 (03) :351-368
[8]   Amazon's Mechanical Turk: A New Source of Inexpensive, Yet High-Quality, Data? [J].
Buhrmester, Michael ;
Kwang, Tracy ;
Gosling, Samuel D. .
PERSPECTIVES ON PSYCHOLOGICAL SCIENCE, 2011, 6 (01) :3-5
[9]   Separate but equal? A comparison of participants and data gathered via Amazon's MTurk, social media, and face-to-face behavioral testing [J].
Casler, Krista ;
Bickel, Lydia ;
Hackett, Elizabeth .
COMPUTERS IN HUMAN BEHAVIOR, 2013, 29 (06) :2156-2160
[10]   Lie for a Dime: When Most Prescreening Responses Are Honest but Most Study Participants Are Impostors [J].
Chandler, Jesse J. ;
Paolacci, Gabriele .
SOCIAL PSYCHOLOGICAL AND PERSONALITY SCIENCE, 2017, 8 (05) :500-508