Abstract
In recent years, AI research has become more and more computationally demanding. In natural language processing (NLP), this tendency is reflected in the emergence of large language models (LLMs) like GPT-3. These powerful neural network-based models can be used for a range of NLP tasks and their language generation capacities have become so sophisticated that it can be very difficult to distinguish their outputs from human language. LLMs have raised concerns over their demonstrable biases, heavy environmental footprints, and future social ramifications. In December 2020, critical research on LLMs led Google to fire Timnit Gebru, co-lead of the company’s AI Ethics team, which sparked a major public controversy around LLMs and the growing corporate influence over AI research. This article explores the role LLMs play in the political economy of AI as infrastructural components for AI research and development. Retracing the technical developments that have led to the emergence of LLMs, we point out how they are intertwined with the business model of big tech companies and further shift power relations in their favour. This becomes visible through the Transformer, which is the underlying architecture of most LLMs today and started the race for ever bigger models when it was introduced by Google in 2017. Using the example of GPT-3, we shed light on recent corporate efforts to commodify LLMs through paid API access and exclusive licensing, raising questions around monopolization and dependency in a field that is increasingly divided by access to large-scale computing power.
Highlights
Over the past decade, artificial intelligence (AI) technologies have evolved rapidly.1 Much of the public acclaim and many commercial breakthroughs have been due to deep learning, a specific methodology for machine learning in which complex neural networks are trained using large amounts of data (LeCun et al, 2015)
While AI research has widely switched from universal central processing units (CPUs) to graphics processing units (GPUs), which allow for increased parallelization since the late 2000s (Raina et al, 2009), the question becomes who can further scale up their compute capacity, for example, by acquiring purpose-built hardware such as application-specific integrated circuits (ASICS) or fieldprogrammable gate arrays (FPGAs). These specialized chips have the potential to significantly improve performance but come at a high cost, in terms of financial investment and—depending on the level of specialization—lack of flexibility (Hwang, 2018). Building on these observations about the growing economic and political importance of compute, this paper investigates the political economy of AI through the example of the emergence and commodification of (English language) large language models (LLMs)
With the introduction of the Transformer architecture in 2017, Google started the era of LLMs in natural language processing (NLP)
Summary
Artificial intelligence (AI) technologies have evolved rapidly. Much of the public acclaim and many commercial breakthroughs have been due to deep learning, a specific methodology for machine learning in which complex neural networks are trained using large amounts of data (LeCun et al, 2015). Much of the public acclaim and many commercial breakthroughs have been due to deep learning, a specific methodology for machine learning in which complex neural networks are trained using large amounts of data (LeCun et al, 2015). Deep learning has spurred activity in different AI subfields like Computer Vision or natural language processing (NLP), resulting in sophisticated commercial products for applications such as facial recognition or automated language generation. A growing body of research focuses on the social impact, ethics, and regulation of machine learning systems Regarding NLP, such scholarly work is devoting attention to the study of large language models (LLMs). A language model is a statistical representation of a language, which tells us the likelihood that a given sequence (a word, phrase, or sentence) occurs in this language.
Full Text
Topics from this Paper
Large Language Models
AI Research
Shift Power Relations
Exclusive Licensing
Bigger Models
+ Show 5 more
Create a personalized feed of these topics
Get StartedTalk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Similar Papers
INFORMS Journal on Data Science
May 19, 2023
Plastic and Reconstructive Surgery - Global Open
Apr 1, 2023
International Journal for Research in Applied Science and Engineering Technology
Jul 31, 2023
Radiation oncology journal
Sep 21, 2023
Journal of biomedical informatics
Oct 1, 2023
Health Care Science
Jul 24, 2023
International Journal of Oral Science
Jul 28, 2023
Rheumatology (Oxford, England)
Jun 12, 2023
Regional Anesthesia & Pain Medicine
Jun 19, 2023
Journal of pediatric urology
Oct 1, 2023
Scientific Journal of Astana IT University
Sep 30, 2023
Indian journal of surgical oncology
Sep 1, 2023
Big Data & Society
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023
Big Data & Society
Jul 1, 2023