The agency of computer vision models as optical instruments

被引:11
作者
Smits, Thomas [1 ]
Wevers, Melvin [2 ]
机构
[1] Univ Utrecht, Trans 10, NL-3512 JK Utrecht, Netherlands
[2] Univ Amsterdam, Urban Hist & Digital Methods, Amsterdam, Noord Holland, Netherlands
基金
欧洲研究理事会;
关键词
Agency; bias; computer vision; datasets; machine learning fairness; optical instrument;
D O I
10.1177/1470357221992097
中图分类号
G2 [信息与知识传播];
学科分类号
05 ; 0503 ;
摘要
Industry and governments have deployed computer vision models to make high-stake decisions in society. While they are often presented as neutral and objective, scholars have recognized that bias in these models might lead to the reproduction of racial, social, cultural and economic inequity. A growing body of work situates the provenance of bias in the collection and annotation of datasets that are needed to train computer vision models. This article moves from studying bias in computer vision models to the agency that is commonly attributed to them: the fact that they are universally seen as being able to make biased decisions. Building on the work of Bruno Latour and Jonathan Crary, the authors discuss computer vision models as agential optical instruments in the production of contemporary visuality. They analyse five interconnected research steps - task selection, category selection, data collection, data labelling and evaluation - of six widely cited benchmark datasets, published during a critical stage in the development of the field (2004-2020): Caltech 101, Caltech 256, PASCAL VOC, ImageNet, MS COCO and Google Open Images. They found that, despite all sorts of justifications, the selection of categories is not based on any general notion of visuality, but depends heavily upon perceived practical applications, the availability of downloadable images and, in conjunction with data collection, favours categories that can be unambiguously described by text. Second, the reliance on Flickr for data collection introduces a temporal bias in computer vision datasets. Third, by comparing aggregate accuracy rates and 'human' performance, the dataset papers introduce a false dichotomy between the agency of computer vision models and human observers. In general, the authors argue that the agency of datasets is produced by obscuring the power and subjective choices of its creators and the countless hours of highly disciplined labour of crowd workers.
引用
收藏
页码:329 / 349
页数:21
相关论文
共 43 条
  • [1] A Research Agenda for Hybrid Intelligence: Augmenting Human Intellect With Collaborative, Adaptive, Responsible, and Explainable Artificial Intelligence
    Akata, Zeynep
    Balliet, Dan
    de Rijke, Maarten
    Dignum, Frank
    Dignum, Virginia
    Eiben, Guszti
    Fokkens, Antske
    Grossi, Davide
    Hindriks, Koen
    Hoos, Holger
    Hung, Hayley
    Jonker, Catholijn
    Monz, Christof
    Neerincx, Mark
    Oliehoek, Frans
    Prakken, Henry
    Schlobach, Stefan
    van der Gaag, Linda
    van Harmelen, Frank
    van Hoof, Herke
    van Riemsdijk, Birna
    van Wynsberghe, Aimee
    Verbrugge, Rineke
    Verheij, Bart
    Vossen, Piek
    Welling, Max
    [J]. COMPUTER, 2020, 53 (08) : 18 - 28
  • [2] Barocas S., 2019, FAIRNESS MACHINE LEA
  • [3] Benjamin R., 2019, Race After Technology
  • [4] Bridle J., 2018, NEW DARK AGE TECHNOL
  • [5] Buolamwini J., 2018, PMLR, P77
  • [6] Business Insider Nederland, 2020, GOOGL AI WILL NO LON
  • [7] Crary Jonathan., 1992, Techniques of the Observer: On Vision and Modernity in the Nineteenth Century
  • [8] Crawford K., 2019, AI NOW 2019 REPORT
  • [9] Crawford Kate, 2019, Excavating AI: The Politics of Images in Machine Learning'
  • [10] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848