Overcoming the pitfalls and perils of algorithms: A classification of machine learning biases and mitigation methods

被引:91
作者
van Giffen, Benjamin [1 ]
Herhausen, Dennis [2 ]
Fahse, Tobias [1 ]
机构
[1] Univ St Gallen, Inst Informat Management, St Gallen, Switzerland
[2] Vrije Univ Amsterdam, Sch Business & Econ, Amsterdam, Netherlands
关键词
Machine learning; Artificial intelligence; Bias; Mitigation methods; Case study; ARTIFICIAL-INTELLIGENCE; DISCRIMINATION; PREDICTION;
D O I
10.1016/j.jbusres.2022.01.076
中图分类号
F [经济];
学科分类号
02 ;
摘要
Over the last decade, the importance of machine learning increased dramatically in business and marketing. However, when machine learning is used for decision-making, bias rooted in unrepresentative datasets, inade-quate models, weak algorithm designs, or human stereotypes can lead to low performance and unfair decisions, resulting in financial, social, and reputational losses. This paper offers a systematic, interdisciplinary literature review of machine learning biases as well as methods to avoid and mitigate these biases. We identified eight distinct machine learning biases, summarized these biases in the cross-industry standard process for data mining to account for all phases of machine learning projects, and outline twenty-four mitigation methods. We further contextualize these biases in a real-world case study and illustrate adequate mitigation strategies. These insights synthesize the literature on machine learning biases in a concise manner and point to the importance of human judgment for machine learning algorithms.
引用
收藏
页码:93 / 106
页数:14
相关论文
共 65 条
[41]   AI in marketing, consumer research and psychology: A systematic literature review and research agenda [J].
Mariani, Marcello M. ;
Perez-Vega, Rodrigo ;
Wirtz, Jochen .
PSYCHOLOGY & MARKETING, 2022, 39 (04) :755-776
[42]   Designing Ethical Algorithms [J].
Martin, Kirsten .
MIS QUARTERLY EXECUTIVE, 2019, 18 (02) :129-142
[43]   CRISP-DM Twenty Years Later: From Data Mining Processes to Data Science Trajectories [J].
Martinez-Plumed, Fernando ;
Contreras-Ochando, Lidia ;
Ferri, Cesar ;
Hernandez-Orallo, Jose ;
Kull, Meelis ;
Lachiche, Nicolas ;
Ramirez-Quintana, Maria Jose ;
Flach, Peter .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2021, 33 (08) :3048-3061
[44]   A Survey on Bias and Fairness in Machine Learning [J].
Mehrabi, Ninareh ;
Morstatter, Fred ;
Saxena, Nripsuta ;
Lerman, Kristina ;
Galstyan, Aram .
ACM COMPUTING SURVEYS, 2021, 54 (06)
[45]  
Mitchell S., 2018, ARXIV STATAP
[46]   Does Machine Learning Automate Moral Hazard and Error? [J].
Mullainathan, Sendhil ;
Obermeyer, Ziad .
AMERICAN ECONOMIC REVIEW, 2017, 107 (05) :476-480
[47]   A method for taxonomy development and its application in information systems [J].
Nickerson, Robert C. ;
Varshney, Upkar ;
Muntermann, Jan .
EUROPEAN JOURNAL OF INFORMATION SYSTEMS, 2013, 22 (03) :336-359
[48]   Dissecting racial bias in an algorithm used to manage the health of populations [J].
Obermeyer, Ziad ;
Powers, Brian ;
Vogeli, Christine ;
Mullainathan, Sendhil .
SCIENCE, 2019, 366 (6464) :447-+
[49]   Social Data: Biases, Methodological Pitfalls, and Ethical Boundaries [J].
Olteanu, Alexandra ;
Castillo, Carlos ;
Diaz, Fernando ;
Kiciman, Emre .
FRONTIERS IN BIG DATA, 2019, 2
[50]   Consumers and Artificial Intelligence: An Experiential Perspective [J].
Puntoni, Stefano ;
Reczek, Rebecca Walker ;
Giesler, Markus ;
Botti, Simona .
JOURNAL OF MARKETING, 2021, 85 (01) :131-151