The great transformer: Examining the role of large language models in the political economy of AI

被引:76
作者
Luitse, Dieuwertje [1 ]
Denkena, Wiebke [1 ]
机构
[1] Univ Amsterdam, Dept Media Studies, Turfdraagsterpad 9, NL-1012 XT Amsterdam, Netherlands
关键词
Artificial intelligence; algorithmic techniques; Transformer; large language models; monopolization; platforms; PLATFORMIZATION;
D O I
10.1177/20539517211047734
中图分类号
C [社会科学总论];
学科分类号
03 ; 0303 ;
摘要
In recent years, AI research has become more and more computationally demanding. In natural language processing (NLP), this tendency is reflected in the emergence of large language models (LLMs) like GPT-3. These powerful neural network-based models can be used for a range of NLP tasks and their language generation capacities have become so sophisticated that it can be very difficult to distinguish their outputs from human language. LLMs have raised concerns over their demonstrable biases, heavy environmental footprints, and future social ramifications. In December 2020, critical research on LLMs led Google to fire Timnit Gebru, co-lead of the company's AI Ethics team, which sparked a major public controversy around LLMs and the growing corporate influence over AI research. This article explores the role LLMs play in the political economy of AI as infrastructural components for AI research and development. Retracing the technical developments that have led to the emergence of LLMs, we point out how they are intertwined with the business model of big tech companies and further shift power relations in their favour. This becomes visible through the Transformer, which is the underlying architecture of most LLMs today and started the race for ever bigger models when it was introduced by Google in 2017. Using the example of GPT-3, we shed light on recent corporate efforts to commodify LLMs through paid API access and exclusive licensing, raising questions around monopolization and dependency in a field that is increasingly divided by access to large-scale computing power.
引用
收藏
页数:14
相关论文
共 77 条
[1]   The Grey Hoodie Project: Big Tobacco, Big Tech, and the Threat on Academic Integrity [J].
Abdalla, Mohamed ;
Abdalla, Moustafa .
AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, :287-297
[2]  
Abid A., 2021, ARXIV210105783CS
[3]  
Ahmed N., 2020, ARXIV201015581
[4]  
Amoore Louise, 2020, Cloud Ethics: Algorithms and the Attributes of Ourselves and Others
[5]  
[Anonymous], 2013, GENERATING SEQUENCES
[6]  
[Anonymous], 2008, SOFTWARE STUDIES LEX
[7]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[8]   On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? [J].
Bender, Emily M. ;
Gebru, Timnit ;
McMillan-Major, Angelina ;
Shmitchell, Shmargaret .
PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, :610-623
[9]   LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT [J].
BENGIO, Y ;
SIMARD, P ;
FRASCONI, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02) :157-166
[10]  
Benjamin R., 2019, POLITY