Democratizing AI: non-expert design of prediction tasks

被引:5
作者
Bagrow, James P. [1 ,2 ]
机构
[1] Univ Vermont, Math & Stat, Burlington, VT 05405 USA
[2] Univ Vermont, Vermont Complex Syst Ctr, Burlington, VT 05405 USA
基金
美国国家科学基金会;
关键词
Citizen science; Supervised learning; Predictive models; Randomized control trial; Amazon mechanical turk; Novel data collection; Crowdsourcing; Interactive machine learning; Automatic machine learning; AutoML;
D O I
10.7717/peerj-cs.296
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-experts have long made important contributions to machine learning (ML) by contributing training data, and recent work has shown that non-experts can also help with feature engineering by suggesting novel predictive features. However, non-experts have only contributed features to prediction tasks already posed by experienced ML practitioners. Here we study how non-experts can design prediction tasks themselves, what types of tasks non-experts will design, and whether predictive models can be automatically trained on data sourced for their tasks. We use a crowdsourcing platform where non-experts design predictive tasks that are then categorized and ranked by the crowd. Crowdsourced data are collected for top-ranked tasks and predictive models are then trained and evaluated automatically using those data. We show that individuals without ML experience can collectively construct useful datasets and that predictive models can be learned on these datasets, but challenges remain. The prediction tasks designed by non-experts covered a broad range of domains, from politics and current events to health behavior, demographics, and more. Proper instructions are crucial for non-experts, so we also conducted a randomized trial to understand how different instructions may influence the types of prediction tasks being proposed. In general, understanding better how non-experts can contribute to ML can further leverage advances in Automatic machine learning and has important implications as ML continues to drive workplace automation.
引用
收藏
页数:23
相关论文
共 38 条
[1]   Quality Control in Crowdsourcing Systems Issues and Directions [J].
Allahbakhsh, Mohammad ;
Benatallah, Boualem ;
Ignjatovic, Aleksandar ;
Motahari-Nezhad, Hamid Reza ;
Bertino, Elisa ;
Dustdar, Schahram .
IEEE INTERNET COMPUTING, 2013, 17 (02) :76-81
[2]  
Alpaydin E., 2020, Introduction to machine learning, V4th
[3]  
[Anonymous], 2017, IEEE SYST J
[4]   Why Are There Still So Many Jobs? The History and Future of Workplace Automation [J].
Autor, David H. .
JOURNAL OF ECONOMIC PERSPECTIVES, 2015, 29 (03) :3-30
[5]   Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk [J].
Berinsky, Adam J. ;
Huber, Gregory A. ;
Lenz, Gabriel S. .
POLITICAL ANALYSIS, 2012, 20 (03) :351-368
[6]   Soylent: A Word Processor with a Crowd Inside [J].
Bernstein, Michael S. ;
Little, Greg ;
Miller, Robert C. ;
Hartmann, Bjoern ;
Ackerman, Mark S. ;
Karger, David R. ;
Crowell, David ;
Panovich, Katrina .
COMMUNICATIONS OF THE ACM, 2015, 58 (08) :85-94
[7]   Crowdsourcing Novel Childhood Predictors of Adult Obesity [J].
Bevelander, Kirsten E. ;
Kaipainen, Kirsikka ;
Swain, Robert ;
Dohle, Simone ;
Bongard, Josh C. ;
Hines, Paul D. H. ;
Wansink, Brian .
PLOS ONE, 2014, 9 (02)
[8]   Crowdsourcing Predictors of Behavioral Outcomes [J].
Bongard, Josh C. ;
Hines, Paul D. H. ;
Conger, Dylan ;
Hurd, Peter ;
Lu, Zhenyu .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2013, 43 (01) :176-185
[9]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[10]   Online panels in social science research: Expanding sampling methods beyond Mechanical Turk [J].
Chandler, Jesse ;
Rosenzweig, Cheskie ;
Moss, Aaron J. ;
Robinson, Jonathan ;
Litman, Leib .
BEHAVIOR RESEARCH METHODS, 2019, 51 (05) :2022-2038