Explainable artificial intelligence (XAI) post-hoc explainability methods: risks and limitations in non-discrimination law

被引:0
作者
Daniel Vale
Ali El-Sharif
Muhammed Ali
机构
[1] Leiden University,eLaw Centre, Leiden School of Law
[2] Nova Southeastern University,College of Computing and Engineering
[3] University College London,UCL Knowledge Lab
来源
AI and Ethics | 2022年 / 2卷 / 4期
关键词
Artificial intelligence; Explainability; Discrimination; Law; Non-discrimination law; Machine learning;
D O I
10.1007/s43681-022-00142-y
中图分类号
学科分类号
摘要
Organizations are increasingly employing complex black-box machine learning models in high-stakes decision-making. A popular approach to addressing the problem of opacity of black-box machine learning models is the use of post-hoc explainability methods. These methods approximate the logic of underlying machine learning models with the aim of explaining their internal workings, so that human examiners can understand them. In turn, it has been alluded that the insights from post-hoc explainability methods can be used to help regulate black-box machine learning. This article examines the validity of these claims. By examining whether the insights derived from post-hoc explainability methods in post-model deployment can prima facie meet legal definitions in European (read European Union) non-discrimination law, we argue that machine learning post-hoc explanation methods cannot guarantee the insights they generate.
引用
收藏
页码:815 / 826
页数:11
相关论文
共 128 条
[1]  
Adadi A(2018)Peeking inside the Black-box: a survey on explainable artificial intelligence (XAI) IEEE Access 6 52138-52160
[2]  
Berrada M(2020)Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI Inf Fusion 58 82-115
[3]  
Barredo Arrieta A(2001)Statistical modeling: the two cultures (with comments and a rejoinder by the author) Stat Sci 70 245-317
[4]  
Díaz-Rodríguez N(2021)A survey on the explainability of supervised machine learning J Artif Intell Res 3 205395171562251-682
[5]  
Del Ser J(2016)How the machine ‘thinks’: understanding opacity in machine learning algorithms Big Data Soc. 8 832-553
[6]  
Bennetot A(2019)Machine learning Interpretability: a survey on methods and metrics Electronics 11 645-77
[7]  
Tabik S(2011)The European Union and human rights after the treaty of Lisbon Hum. Rights Law Rev 27 537-694
[8]  
Barbado A(2007)Direct discrimination, indirect discrimination and autonomy Oxf. J. Leg. Stud. 63 68-42
[9]  
Garcia S(2019)Techniques for interpretable machine learning Commun. ACM 30 681-549
[10]  
Gil-Lopez S(2012)Key concepts in EU anti-discrimination law EU Anti-Discrimination Law 13 94-67