IA como supercontrovérsia: provocando controvérsias sobre IA e sociedade com uma extensa comunidade de especialistas no Reino Unido

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 18

Original Research Article

Big Data & Society


April–June : 1–18
AI as super-controversy: Eliciting AI and © The Author(s) 2024
Article reuse guidelines:
society controversies with an extended sagepub.com/journals-permissions
DOI: 10.1177/20539517241255103
expert community in the UK journals.sagepub.com/home/bds

Noortje Marres1 , Michael Castelle1 , Beatrice Gobbo2 ,


Chiara Poletti1 and James Tripp1

Abstract
Following the release of large language models in the late 2010s, the backers of this new type of artificial intelligence (AI)
publicly affirmed that the technology is controversial and harmful to society. This situation sets contemporary AI apart
from 20th-century controversies about technnoscience, such as nuclear power and genetically modified (GM) foods, and
disrupts established assumptions concerning public controversies as occasions for technological democracy. In particular,
it challenges the idea that such controversies enable inclusion and collective processes of problem definition (‘problem-
atisation’) across societal domains. In this paper, we show how social research can contribute to addressing this challenge
of AI controversies by adopting a distinctive methodology of controversy analysis: controversy elicitation. This approach
actively selects, qualifies and evaluates controversies in terms of their capacity to problematise AI across the science and
non-science binary. We describe our implementation of this approach in a participatory study of recent AI controversies,
conducted through consultation with UK experts in AI and society. Combining an online questionnaire, social media ana-
lysis and a participatory workshop, our study suggests that civil society actors have developed distinctive strategies of
problematisation that counter the strategic affirmation of AI’s controversiality by its proponents and which centre on
the public mobilisation of AI-related incidents: demonstrations of bias, accidents and walkouts. Crucially, this emphasis
on ‘AI frictions’ does not result in the fragmentation of AI controversies, but rather enables the articulation of AI as a
‘super-controversy’: the explication of connections between technical propositions, situated troubles and structural pro-
blems in society (discrimination, inequalities and corporate power).

Keywords
Artificial intelligence, controversy analysis, elicitation methods, science and technology studies, Twitter, design for debate

This article is a part of special theme on Analysing Artificial Intelligence Controversies. To see a full list of all articles
in this special theme, please click here: https://journals.sagepub.com/page/bds/collections/
analysingartificialintelligencecontroversies

Introduction
In March 2023, the San Francisco-based start-up OpenAI 1
Centre for Interdisciplinary Methodologies, University of Warwick,
released their latest large language model (LLM) Coventry, UK
2
Generative Pre-Trained Transformer 4 (GPT-4) to their sub- Department of Design, Politecnico di Milano, Milano, Italy
scribers, including an accompanying document warning
Corresponding author:
about the potential negative consequences of this technol- Noortje Marres, Centre for Interdisciplinary Methodologies, University of
ogy,1 which led to a wave of media articles highlighting Warwick, Coventry CV4 7AL, UK.
the real dangers of this ‘controversial’ technology for Email: N.Marres@warwick.ac.uk

Creative Commons CC BY: This article is distributed under the terms of the Creative Commons Attribution 4.0 License (https://
creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission
provided the original work is attributed as specified on the SAGE and Open Access page (https://us.sagepub.com/en-us/nam/open-access-at-sage).
2 Big Data & Society

society.2 In previous years, OpenAI had issued similar more prominently in English-language media than similar
statements: in February 2019, the company initially opted assessments by industry outsiders in recent years
against releasing the GPT-2 model (despite ultimately (Dandurand et al., 2022), this does not necessarily mean
making it available several months later), due to ‘concerns that ‘AI controversy’ can be reduced to these industry asser-
about LLMs being used to generate deceptive, biased, or tions. To examine this, what is needed is a broad empirical
abusive language at scale’.3 Similar concerns were made investigation of the ‘problematisations’ of AI that have
public in May 2023 by the Chief Economist of OpenAI’s emerged from public debates about AI across the divide
parent company Microsoft, who went on record to state between ‘AI insiders’ affiliated with big tech companies
that ‘I am confident that AI will be used by bad actors, and their civil society critics.
and yes it will cause real damage’.4 Such affirmations of To this end, we have worked with UK-based experts in
the dangers, harms and costs to society of this new wave ‘AI and society’ to identify, qualify and evaluate the most
of so-called generative AI by their developers and backers important and possibly overlooked controversies about AI
can be called ‘disruptive’ in their own right, as they break in the last 10 years, as seen from the UK. We asked:
with unwritten conventions that have underpinned public which public controversies about AI during this period
communications about controversial science and innovation achieved its problematisation, and how was this done? To
in the last few decades: in notable controversies about tech- answer our question, we adopted a critical methodology
noscience in the 20th century, concerning nuclear power, of controversy elicitation. To prevent our AI controversy
climate change, and genetically modified (GM) food, the analysis from merely reproducing the strategic corporate
companies involved consistently denied, refuted, deflected discourse about future risk and harm, we used techniques
or severely qualified claims advanced by civil society of controversy mapping to selectively identify AI contro-
actors about the harms and risks that their products versies with the potential capacity to articulate problems
created for society. What is the significance of the seem- with AI across the divides between industry and civil
ingly irresponsible affirmations by AI developers and society, or more minimally put, between science and non-
backers of the disruptions, risks and harms that are science. We implemented this approach in a mixed
created by their innovations, and what are the implications methods research design, which combined online consult-
of this for the role and status of public controversies about ation with Twitter analysis and a participatory design work-
contemporary AI? shop with AI and Society experts from across domains. As
One of the striking features of recent corporate ‘warn- we will discuss, this approach led us to identify two dis-
ings’ about the dangers that AI poses for society is that tinctive features of recent controversies about AI: (1)
they echo similar warnings that have been made by AI Many significant AI controversies involve the mobilisation
critics during the last 5 years. When Geoff Hinton, the of what we call ‘AI frictions’: concrete incidents such as
‘Godfather’ of AI went public with his concerns about AI erroneous arrests based on incorrect facial recognition intel-
in May 2023, and cited these as a reason to end his employ- ligence, technical demonstrations of bias in operational
ment at Google, it was pointed out by many, in newspaper algorithms such as in recruitment, and walk outs by out-
editorials and on Twitter, that he had in 2018 failed to show raged tech company employees. (2) This proliferation of
support when the computer scientist Timnit Gebru and AI frictions, however, does not entail the fragmentation of
others on Google’s Ethics team, were fired or left the AI controversy, but rather provides occasions for the articu-
same company citing closely related concerns, namely lation of AI as a ‘super-controversy’: in AI controversies,
that Google’s LLMs present a source of significant harm any individual dispute is inexorably chained to a variety of
to society.5 In this regard, Hinton’s actions can be inter- other (overt and latent) problematics in and across research
preted as a discursive strategy of appropriation—a way of and innovation communities, government and public sector
disarming the criticisms of AI voiced by actors in society and wider societies. We argue that AI frictions enable a dis-
by adopting a modified version of those criticisms as tinctive mode of problematising AI across science and non-
one’s own, and mobilising one’s own authority as a credible science, one that links technological propositions with experi-
spokesperson of science and business to occupy the channel enced harms and structural problematics in the areas of social
of ‘public concern’ with AI, to the exclusion of out-group justice, knowledge politics and political economy.
critics.6 Indeed, this public communications strategy can
be understood as reflective of wider, institutionalised rela-
tions between industry and state, in which Silicon Valley How to uncover the capacity of AI
tech-company lobbyists seem assured that their ‘confi- controversies for problematisation across
dence’ that AI will cause widespread societal harm will
not be sufficient cause for the state to restrict its develop- the science/non-science binary?
ment in ways that are adverse to their interests (McGoey, Recent scholarship has drawn attention to the appropriation
2021). However, even if the affirmations of AI’s harmful- of critical discourse about technology and society by the
ness for society by its developers and backers have featured contemporary tech industry. Phan and colleagues (2022)
Marres et al. 3

describe how, when AI scientists at Google like Timnit and scientists do not fit this picture at all. Instead of
Gebru publicly contested the company’s lack of commit- opening up the debate about the socio-technical problems
ment to addressing the societal harms and risks associated posed by AI to outsiders, these statements reassert the
with the LLMs it was developing, the effect was not to authority of scientists and company insiders to define
endanger the company’s reputation. Instead, issues of risk techno-scientific risks and harms. Instead of facilitating
and harm became ‘ethicised’ and corporatised: they were the problematisation of AI in the sense of surfacing its com-
reframed by the company as concerns that could be dealt plexity, uncertainty and heterogeneous effects in society,
with via internal processes centred on values rather than these assertions present attempts to consolidate the defin-
regulation. Appropriation of the vocabulary of societal ition of AI as objective reality, in line with Samuel
risks and harms of technology can also be discerned in Johnson’s old empiricist dictum that if it hurts, it must be
contemporary scientific discourse on AI. For example, the real (Patey, 1986; on this point also see Kak and Meyers
computer scientist Percy Liang observed during his intro- West, 2023).
duction to the ‘Foundation Models’ workshop at Stanford In other words, the strategic use of AI’s controversiality
University in 2021: ‘what I am seeing today is the very threatens the role of public controversy as a force of the
beginning of a paradigm shift, [in the area of large language democratisation of research and innovation, and risks to
models] and I think this paradigm shift is going to have undermine its capacity for inclusion and problematisation.
profound implications […] there will also be heavy social The reasons for this are surely complex. For one, artificial
consequences that will result from each technological intelligence (AI) research has been marked since its incep-
decision’.7 Such affirmations that LLMs and indeed subse- tion by oversized, objectivist claims about its existential
quent generative AI models create new societal risks and impact on humanity, claims which partly served as displays
harms by AI proponents point towards a radicalisation of scientific authority and prowess (Suchman, 2023) Yet, as
of a recent argument by Geiger and colleagues which Kling argued in Computerization and Controversy (1996),
states that in contemporary capitalist societies, controversy in earlier decades controversy about automation did
about science and technology increasingly serves as a provide important occasions for the collective definition
valuation strategy: a way of opening up markets for new of social problems in compute-intensive societies, enabling
products and mobilising public concern to create a commu- the articulation of ‘key value conflicts and social choices’
nity of attention for said products (Geiger et al., 2014). (Kling, 1996: xiv) concerning the future of work, surveil-
In Geiger et al.’s accounts, techno-scientific controver- lance, privacy and problems of bias, as a way to ‘articulate
sies are increasingly deployed towards promotional ends. the different sides of key debates’ (xxi), and more generally,
In the last decade or so, controversy has acquired a ‘posi- to foster a greater appreciation for ‘computerisation’s key
tive’ association with the tech industry’s notion of social dimensions’ (xxii). Strikingly, many of the issues
disruption (Geiger, 2020), the Schumpeterian idea that flagged by Kling and other contributors to this volume
innovation done well is destructive of existing societal mirror today’s problems associated with AI. But where
arrangements and will challenge social conventions. An Kling could still claim—in 1996—that the societal issues
example of this is Google Glass, which was widely reported raised by the computerisation of society ‘have not received
in the media as controversial because of the fundamental due attention’ (xxii), today’s media seem flooded by state-
concerns with privacy and surveillance raised by these ments to this effect. News media today endlessly circulate
networked wearables equipped with Facial Recognition. ‘explainers’ about why AI is controversial, as in the
Reviewing these controversies, the tech magazine WIRED WIRED article on Google Glass above, and they are awash
concluded: ‘Google Glass wasn’t a failure… it raised with reports of the societal harm that AI will or will not
crucial concerns’.8 This affirmative understanding of con- cause in the form of job losses and worker exploitation, dis-
troversy as valuation stands in tension with earlier framings crimination by algorithm—as when UK students had their
of controversy in Science and Technology Studies (STS), General Standard of Secondary Education (GSCE) marks
which defined it as an enabling occasion for the democra- downgraded by algorithmic rebalancing—and disinforma-
tisation of research and innovation (Callon et al., 2011). tion and deepfakes that will sabotage democratic elections.
From this perspective, public disputes about However, it seems to us this mass reporting on AI’s harms
techno-scientific innovations, like nuclear power and genet- to society does not qualify as a collective process of
ically modified foods, enable (1) the expansion of the range problem articulation. Recent studies of AI media discourse
of actors and competencies involved in the definition of in France and Canada have shown that this discourse
techno-scientific risks and harms (inclusion), and (2) a instead consolidates the binary that AI will either solve soci-
process of the collective articulation of complex problems etal problems or cause epic harm (Crépel et al., 2021). And
at the intersection of science innovation and society far from widening the range of voices involved in AI
which are partly uncertain and affect different groups in debates, the hyperbolic reports on AI benefits and risks that
society in divergent ways (‘problematisation’). The stra- dominate the news media feature primarily scientists and
tegic assertions of AIs controversiality by AI companies industry representatives (Dandurand et al., 2022).
4 Big Data & Society

This situation not only has implications for our under- language of AI harms and risks to consolidate the authority
standing of AI controversies as occasions for the democra- of techno-science and the reality of AI. We ask: how to
tisation of knowledge; it also has consequences for the uncover controversy’s capacity for the problematisation
methodology of controversy analysis. Since the early 20th of science and technology, in a context in which predomin-
century (Mannheim, 2015 (1936)), knowledge controver- ant debates about the benefits, risks and harms of contem-
sies have been valued as empirical occasions for exploring porary AI seem focused on the consolidation of AI as an
wider relations between science and society: whenever objective reality? If, generally speaking, promissory and
there is sustained disagreement in and about science—is eschatological framings of AI appear to dominate media
light a wave or a particle? Is nuclear power safe?—the discourse, how can we identify AI discourses that enable
boundary between scientific insiders and outsiders is articulation and contestation of what is problematic about
troubled, and new ways of framing the phenomenon in AI across industry, science, media and society?
question become possible. Inspired by this insight, sociolo-
gists of science and technology developed the methodology Methodology: From the observation to the
of ‘controversy analysis’ from the 1970s onwards.
Actor-network theorists defined controversies as opportun-
elicitation of AI controversies
ities for the empirical observation of science-society inter- It was clear to us, then, that if we wished to attend to societal
actions: as controversies compel the actors involved to processes of the problematisation of AI, we would need to
account for their positions in the dispute, they are led to diverge from predominant approaches in controversy
make their relations explicit and, thus, describable by mapping. Contemporary AI controversies do not fit the
social researchers (Latour, 2005). This approach provided model of a generative empirical occasion: prominent
the basis for online controversy mapping, an approach media and expert debates about AI cannot be relied upon
that uses techniques of data analysis and visualisation to —in and of themselves—to bring to the surface qualifica-
create representations of issues that are publicly contested tions of the societal consequences of AI or to make explicit
in digital media (Rogers and Marres, 2000; Venturini and actors’ relations and interests. The usual research strategy of
Munk, 2021). However, in the case of AI, assertions of con- controversy mapping, which is to query online media for
troversiality do not seem to have the illuminating empirical controversy terms such as ‘Nuclear waste’ or ‘GM foods’,
effect that controversy analysis used to rely on: corporate and then to document the positions that actors take on the
and expert discourse about ‘AI risks’ has the effect of ren- issue on Web pages and in social media, does not work in
dering civil society critiques of AI less visible and push this case (see also Munk et al., this volume). Indeed,
debates about the political and economic interests of the Twitter data that we collected using the query ‘AI’
tech industry into the background (Stark et al., 2021). between December 2018 and March 2021—which came
Assertions of AI’s controversiality by its proponents to 36 million tweets—seemed to us unsuitable for contro-
exhibit features of a ‘pseudo-event’ (Parks, 2021): an orche- versy mapping: on initial exploration this data set appeared
strated publicity moment where ‘nothing happens’, an arti- to contain lots of publicity and little debate. If we were to
ficial rather than an empirical occasion. There is no lack of use this data to create a ‘controversy map’, there was a
expert contestation and societal debates about AI, but as real risk our analysis would merely end up reproducing
long as we ‘merely observe’ the unfolding of AI controver- solutionist and doomsaying claims about AI—and their
sies in popular and scientific media, as controversy associated authority and reality effects—without surfacing
mapping said we should do, we are most likely to capture articulations of AI as a collective problem. To locate the
strategic—not expansive and problematising—assertions latter, we decided, it was necessary to shift from the descrip-
of AIs controversiality. tion to the elicitation of AI controversy. Declining to adopt
How, then, to analyse public controversies about AI the empiricist posture of patiently tracing contestations
under these conditions? In this context, we propose, the about AI wherever they emerge, we instead decided to
task of controversy analysis is not to ‘describe’ the positions draw on design research methodology to devise strategies
and relations of actors involved in AI controversies, but to actively elicit a distinctive type of controversy about
rather to undertake an active search for problematisations AI: controversies with the capacity to problematise AI
that AI controversy may still be giving rise to today. That across the science-non-science divide.9 Design researcher
is, our hypothesis is that AI controversies today continue Donato Ricci offered this helpful summary of this methodo-
to provide opportunities for the democratisation of logical re-orientation of controversy mapping:
science and innovation, by expanding the range of actors
involved in AI debates beyond a narrow range of insiders If we [typically] spend a great deal of time to seek the
and for the shared articulation of complex problems. sources, detail the protocols and produce visual manipula-
However, such problematisations seem today in many tion[s] to identify “which data speak about which
cases crowded out in public discourse by promotional actors”, an open lab, or better a lab that opens itself up,
deployments of AI’s controversiality, which use the could be one that also strives to produce the conditions
Marres et al. 5

under which “data make the actors speak.” (Personal task of mapping research controversies about AI for the
communication, 20 January 2021) period 2012–2022 beginning with the UK, and (2) by the
notable presence of civil society organisations active in
This proposal to deploy data to elicit controversy articula- this issue area in the country.10 We thus began by selecting
tion among actors as an open process bears some similar- actors with demonstrable expertise in AI and society, which
ities to the ‘collaboratory’ proposed by Collier (2007; see we did by identifying speakers and registered participants at
also Ruppert et al., 2015), which he described as a designed academic and professional events on the broad topic of AI
form of collaboration—‘a new networked organizational and society in the UK between 2021 and 2022, and which
form that also includes social processes’ (Collier, 2007: 2) included academic, industry and civil society conferences
—and a key enabling condition for qualitative inquiry in and workshops.11 In so doing, we followed the STS assump-
an age of big data. Rather than seeking to create a compre- tion alluded to above, that the capacity of controversy about
hensive representation of the principal points of contention technoscience to surface emergent definitions of harm and
and actor positions on AI in different domains, we then risk is in part a consequence of inclusion, an effect of the
decided to use interventionist methods—in the form of an involvement of a wider range of actors in the debate about
expert consultation and a participatory workshop—to iden- uncertain and problematic science and innovation. In line
tify, qualify and evaluate a specific type of AI controversy, with this assumption, we adopt a broad definition of
namely those with the capacity to problematise AI across ‘experts in AI and society’, whom, following Funtowicz
domains, to bring new actors into the debate and to unsettle and Ravetz (1991), we define as members of an ‘extended
established definitions of harm and risk (see Alfrink et al. peer community’ and includes all ‘those with a stake in an
(2022) for a similar approach). issue who are committed to genuine debate’ (Funtowicz
To structure this process of controversy elicitation, we and Ravetz, 1991: 170).
adopted the following three-step research design: first, we Our respondents included civil society advocates and
actively configure a context for AI controversy elicitation, activists (Amnesty, EDRi, Article 19), civil servants from
by inviting a specific community of experts to assist us in AI-related government units (Information Commissioner’s
the identification of relevant controversies. As it is our Office, Centre for Data Ethics and Innovation, NHSX),
objective to surface problematisations of AI across academics with backgrounds in digital humanities, social
science and non-science, we decided to consult UK-based science and computer science and industry (AstraZeneca,
‘AI and society’ experts. In a second step, we qualify AI DeepMind), the arts (Serpentine Galleries; Ambient
controversies: we define the type of topics and incidents Information Systems) and journalism (BBC, TechCrunch).
that the consultation surfaced and the problematisations of That is also to say, our research design did not specifically
AI they entail. Next, we analysed selected AI controversies target institutional outsiders, but rather focused on actors
using social media methods and follow-up interviews, to who are involved in the articulation of AI harms, risks and
assess the extent to which they generated problematisations, benefits in ways that cut across the science/non-science
facilitated contestation and compelled exchange between binary. To counter-act authority effects, our consultation
heterogeneous actors. In the third and last step, that of explicitly encouraged respondent to give their own perspec-
evaluating AI controversies, we invited UK-based AI and tive on what makes AI controversial asking: (1) ‘What in
society experts into a design-led workshop, where we your view are the most important, and possibly overlooked,
worked together to create novel diagnoses of selected AI controversies about AI in the last 10 years’ (italics ours)?
controversies and their capacities for problematisation. To enable us to explore their answers further, we added
another question: (2) What are notable moments, publica-
tions and/or individuals associated with the controversies
Convening an extended expert community mentioned? To get at underreported issues associated with
in the UK: What are the most important AI, we asked: ‘Are there other controversies or problems
and possibly overlooked controversies in and with AI research that you believe should be considered
as part of our study?’ The first round of the consultation was
about AI in the last 10 years? sent out in November 2021 and completed by 53 experts.
To initiate our analysis of AI controversies, we began, as is We identified three different types of responses to our
customary in controversy mapping, with a ‘query’ (Rogers, consultation, which broadly correspond to the three ques-
2017). However, our initial query did not involve selecting tions we asked. A first set identifies (1) ‘controversial
and submitting a set of keywords to an online platform API, topics’, broad areas in which controversial AI develop-
such as Twitter’s, but rather took the form of an invitation ments take place, such as ‘Algorithmic decision-making’,
email that we sent to 250 UK-based experts in ‘AI and ‘AI warfare’ and ‘Environmental costs’ (see Figure 1). A
society’. Our decision to focus on this community was in second set of responses mentions specific sites, incidents,
part informed by (1) our role in the international Shaping events and objects of contestation, concrete instances in
AI consortium, a project in which we had taken on the which AI caused disruption, trouble and, in several cases,
6 Big Data & Society

Figure 1. Frequency of occurrence of AI controversy topics in the UK expert consultation (Autumn 2021).

demonstrable harm in society (Amazon’s biased hiring tool; While our consultation asked respondents to identify ‘con-
data transfers without consent between the Royal Free troversial developments’ and ‘controversies’ in and about
Hospital and DeepMind; the persecution of Uighurs in AI, a significant number (41 out of 53 respondents) identi-
China). We named these latter instances (2) ‘frictions’. fied concrete incidents involving AI in society: traffic acci-
Building on Meunier et al.’s ((2021); see also Shaffer dents involving automated vehicles, the use of racially
Shane, 2023) work on ‘algorithmic trouble’, we use the biased facial recognition by the South Wales and
term ‘AI friction’ to denote instances of AI-related harms Metropolitan police, the downgrading of GCSE marks by
occurring in specific environments in society (roads, hospi- algorithms during the UK ‘exams fiasco’. It suggests that
tals and schools), as distinct from more abstract and in the area of AI and society, ‘controversy’—in the sense
de-localised societal risks. In a third and last set of cases, of the staging of public disagreement about a specific
respondents offered (3) problematisations of AI, which knowledge proposition—may not be the main form of con-
we loosely defined as answers to the question ‘what is the testation of AI, with the emphasis placed, instead, on the
problem with AI?’; and more specifically, we understand demonstration of concrete instances of AI-related harm to
as attempts to articulate underlying contestations, difficul- specific groups and institutions in society (students,
ties and suffering associated with AI that arise at the limit BAME, the National Health Service (NHS)). Also striking,
of discourse (Barry, 2021) and which it may be challenging we found, is the comprehensive range of societal domains
to name (two examples from our consultation: ‘The claim identified in the responses, which ranged from law enforce-
that AI exists today’; ‘Deploying non-transparent systems ment, health, transport, the welfare state, education, media,
to make decisions that directly affect people’s lives’). democracy, the economy (including recruitment and work)
Especially striking in our results was the prominence of and sports. This broad range of application domains—
the aforementioned ‘AI frictions’ among the responses. which is nearly coterminous with society—has been
Marres et al. 7

identified as a sociological feature of digital innovation from among the consultation responses.13 We thus adopt
(Lupton, 2014), but here it emerges as a feature of public a relational, data-driven approach to analysing AI ‘contro-
controversy about a digital ‘object’, AI. This feature, too, versy couplings’ (Costas et al., 2021) and define the empir-
stands in contrast to techno-scientific controversies of ical object of our study as AI’s controversiality: we
earlier decades, which tended to be limited to specific soci- deliberately leave open the question of the form and
etal domains (GM foods connect across agriculture, health formats that AI controversies take on in the empirical set-
and economy; nuclear technology associates environment, tings in which we investigate them, making this a question
energy, human rights and the military). rather than an assumption in our research. To select AI con-
Our initial consultation results suggested to us that AI troversy couplings for further analysis from the many cou-
may qualify today as a super-controversy: AI controversies plings identified through our consultation, we relied on the
do not only take the form of public contestation of specific criterion of ‘research intensity’. As Question 2 of the con-
techno-scientific propositions, but arise from the linkage of sultation asked respondents to provide links to publications
specific instances of harm and risk arising in society with relevant to the controversy in question, we adopt a simple
technical propositions and wider, structural concerns. frequency count of associated research publications—in
However, it is important to note that the consultation also which we included investigative journalism, as well as
surfaced several knowledge controversies in the more civil society reports—as a way to identify AI friction-topic
familiar sense of the term, that is, public contestation of couplings for further research (see Figure 2). Also taking
expert claims. The most frequently mentioned controversial into consideration the spread of controversy friction-topic
development was facial recognition, a topic which gave rise couplings across different expert communities and societal
to public disagreements in the UK from 2018 onwards as domains, we selected the following five controversies for
civil society and academic experts asserted that facial rec- further analysis:
ognition systems in use by the Metropolitan Police fail to
meet accuracy standards, while the Metropolitan Police 1. Correctional Offender Management Profiling for
released an evaluation report which concluded that its Alternative Sanctions (COMPAS): a controversy
facial recognition systems are accurate and not racially about algorithmic discrimination in judicial systems
biased.12 Mentions of other notable topics in the consultation, sparked by the ProPublica (Angwin et al., 2016) report
such as tracking and targeting, data and corporate research ‘Machine Bias’.
culture, equally include references to expert disagreement as 2. NHS + DeepMind: a controversy about data sharing
in the case of the legality of data transfers without consent between UK public sector hospitals and big tech
between the NHS and DeepMind, and to debates about con- sparked by the Powles and Hodson (2017) paper on
troversial research papers like the Stochastic Parrots article by Google DeepMind research in the Royal Free.
Bender, Gebru et al. discussed below. 3. Facial recognition (Gaydar): a controversy about the use
Next, we made a distinction between controversy topics of machine learning-based image analysis to predict
in terms of the degree to which they elicited detailed proble- sexual orientation sparked by Wang and Kosinski
matisations in the consultation, as opposed to only being (2018).
indicated as a keyword, such as ‘surveillance’. What 4. LLMs (Stochastic Parrots): A controversy about bias in
stood out here for us is that, while some respondents large neural network models for encoding and generat-
focus on specific problems (‘pseudonymisation not fit for ing text, sparked by Bender et al. (2021) ‘On the
purpose’; lack of consent for the online scraping of personal Dangers of Stochastic Parrots’.
images by AI companies), many others specify AI’s contro- 5. Deep learning (DL) as a solution for AI: a controversy
versiality in relation to wider, structural problems such as about the capacity of DL—the use of trained multi-
racism, privatisation of the public sector, scientific integrity layer neural networks with large numbers of weight
and lack of trust. Facial recognition is associated with parameters—to sustain the claims of artificial intelligence
the unlawful use of technology in the public sector, as research (Marcus, 2018).
well as ‘bad science’ and persistent racial and gender biases.
Problematisations associated with ‘corporate research culture’ AI and society controversies on Twitter:
included references to toxic working environments in corpor- Levels of disagreement, actor
ate AI research leading to the suppression of independent
research. All these problematisations foreground power composition, forms of engagement
dynamics, and many refer to entrenched societal inequalities. Having identified these five controversies, our next step was
However, it proved difficult to categorise problematisations, to analyse whether and how these controversies enabled the
and we therefore left these aside for the time being. problematisation of AI across the science/non-science
To pursue the exploration of AI as a ‘super-controversy’, binary. Follow-up interviews with participants in the
we proceeded to delineate AI controversies by identifying online consultation offered some indication of the relevance
couplings of (1) frictions and (2) controversial topics of this focus. In relation to the controversial collaboration
8 Big Data & Society

Figure 2. Friction-topic couplings identified in the expert consultation ‘what is controversial about AI?’ Autumn 2021.

between the NHS Hospital the Royal Free (London) and (Interview 12SL); ‘[i]t gives you a feel for an issue’
Google subsidiary DeepMind, one respondent noted: ‘I (Interview 14EH), although we were also reminded
think that’s what happened with the NHS thing… [i]t sort that ‘there’s a huge amount of virtue signalling in terms
of broke out of the confines of people who were interested of what you argue with people about [on Twitter]’
in AI and privacy into something which had more general (Interview 13CR).
currency’ (Interview 14EH); another noted that ‘[T]he We created tailored Twitter data sets for each of our five
Google DeepMind Royal Free thing applies to my univer- controversies, by designing Twitter API queries to capture
sity. And I’ve spoken with people at my university here discussions about the controversial publications identified
and I speak to people at DeepMind, basically, they don’t through the consultation (see Figure 3).14 Next, to make
see what the fuss was about… [s]o there’s a really interest- sure our data sets were pertinent to our selected AI contro-
ing… important point about who actually does learn lessons versies, we manually classified ‘conversations’ in the data-
from this stuff’ (Interview 22JS). To ensure we captured the sets as in scope or out of scope.15 That is also to say, at this
unfolding of AI controversies across societal domains, stage of our research, we chose as our unit of analysis the
we decided to conduct the next step of our controversy ‘conversation’ as defined by the Twitter API: strings of
analysis with Twitter data. Not only is Twitter a tried and replies generated by a single tweet, which can be visualised
tested setting for controversy mapping (Burgess and as a descending, inverted ‘tree’ structure of replies from a
Matamoros-Fernández, 2016; Madsen and Munk, 2019), given ‘origin tweet’ (Nishi et al., 2016). The guiding
but at the time we conducted this research it was still a concern in our Twitter analysis remained with the genera-
notable site of intersection between academic, journalist, tive capacity of AI controversies, as outlined in Section 2:
activist and industry debates. Our interviewees confirmed to what extent do the papers in question enable the problem-
this understanding of Twitter as a prominent forum for atisation of AI on Twitter through sustained disagreement
tech controversies during the relevant period (2012– about determinate propositions among actors from diverse
2022), alongside discussion forum and instant messaging backgrounds (science and non-science)? We asked:
platform Reddit and Discord. They referred to Twitter as what problem definitions in relation to AI are surfaced
a site that ‘has enough experts, it’s open, [and] general’ through exchanges within extended expert communities
Marres et al. 9

Figure 3. Twitter data collection for five AI research controversies.

on Twitter? To what extent do these exchanges involve het- in many conversations, topicalising scientific quality. All
erogeneous actors, cutting across science and non-science? controversies, finally, gave rise to society and justice
To answer these questions, we analysed conversations in debates, which define harmful impacts of AI on society in
four dimensions: topics, levels of disagreement, forms of terms of discrimination, entrenched privilege, racism and
engagement and actor composition. societal inequality.
We begin by coding conversations for controversy We next tried to establish the degree of controversiality
topics, which first of all reveal significant shared concern of the identified topics, which we did by coding all in-scope
across all controversies with ethics, knowledge and social conversations for the level of disagreement, and then
justice.16 Other frequent themes are political economy as assigning these codes to the topics addressed in these con-
well as data and data protection (see Figure 4). Speaking versations (see Figure 5).17 Perhaps unsurprisingly, our
generally, this topic distribution is not dissimilar from our analysis shows that there is a general trend towards dis-
consultation findings. On Twitter, too, the impacts of AI agreement: topics in all controversies tend to be placed
deployments on disadvantaged communities featured as towards the more contentious rather than non-contentious
an especially prominent topic, as did the corporate owner- end of the spectrum. We note that the level of disagreement
ship of AI infrastructures and related barriers to the evalu- is correlated with the volume of tweets produced on a
ation and regulation of AI in society. Data featured certain topic: In COMPAS, the three largest topics by
prominently both in debates about rights (privacy) and reg- volume—bias, prediction and racism—tend to appear in
ulations (including GDPR) as did epistemic challenges of conversations with higher levels of contention. To
reliability and transparency of data processing within further qualify the controversiality of topics, we consid-
AI-based infrastructures. Epistemic concerns about the ered the ‘levels of engagement’ for each conversation:
quality of outputs of AI systems, and lack of adherence to the degree to which conversations elicited long and/or
standards of scientific rigour in AI research also featured wide threads of replies on Twitter (more about this
10 Big Data & Society

Figure 4. Most frequent themes across the five Twitter controversies.

Figure 5. Distribution of the most frequent topics per controversy according to their level of disagreement.
Marres et al. 11

below). We found that contentious topics do not automat- on balance exhibit a high width have a broad appearance,
ically correspond to a high level of engagement. It is true whereas those with higher depth are characterised by
that in COMPAS, Parrots and DL as a solution for AI, the extended length (Figure 6).
top topics by engagement tend to lean towards higher As to the overall findings of our Twitter analysis, we are
levels of contestation. However, we found the highest especially struck by the strong emphasis on societal pro-
level of disagreement for science-related topics that blems (racism, inequality) and political economy (market
belong to the theme of knowledge (light green in the concentration, data appropriation) in the selected AI contro-
visualisation), and secondly, in topics relating to versies, which aligns with the findings of our consultation.
society and justice (in black), political economy (in However, on Twitter, we also found a strong engagement
blue) and bias (in orange). with epistemic issues, in the form of concern with scientific
Next, we examined the extent to which our AI controver- integrity and the science/politics tension (research vs. advo-
sies on Twitter facilitated interaction among heterogenous cacy) in AI research. The focus on these topics correlates
actors, that is, exchanges across the science/non-science with the social positions of participants in the controversy:
binary. To this end, we categorised the Twitter accounts epistemic concerns are topicalised in controversies with
contributing to each controversy using basic occupational prominent participation of researchers, while controversies
categories (see Table 1).18 We found that only a very with strong activist engagement are more concerned with
small number of Twitter accounts appear in more than issues of regulation, ethics and justice (Table 2). While
one controversy, suggesting that our five controversies this alignment between discursive content and social pos-
mobilised different Twitter ‘communities’, which is note- ition is something social studies of scientific controversy
worthy in the light of the high topic similarity between AI would lead us to expect (Barnes, 1977), a different aspect
controversies on Twitter in Figure 4. Considering the size of AI and society controversies on Twitter does not align
of actor categories across controversies, we can see that with this approach. Controversy analysts in STS have
some controversies lean more towards the ‘activist/media argued that techno-scientific controversies surface com-
side’, such as COMPAS, and others more towards ‘scien- plexity and thereby disrupt received scientific and societal
tific/research’, such as DL as a solution for AI, and some problem definitions (Callon et al., 2011). But AI and
are hybrid, such as the Parrots and Gaydar controversies society controversies on Twitter rather seem to mobilise
which bring together researchers and activists. NHS/ established scientific and societal issue frames (transpar-
DeepMind is the one controversy where professions ency, pseudo-science, racism, inequality). Should we con-
(health) are prominent (see Table 2), and policymakers clude that AI and society controversies on Twitter
are relatively absent in all the Twitter controversies. As consolidate entrenched problem definitions? In the conclu-
such, these findings demonstrate a truism from the soci- sion, we will reflect on the significance of this for our under-
ology of knowledge, which states that the content of a con- standing of AI and society as an area of super-controversy.
troversy aligns with actors’ positions in society (Barnes,
1977), more about which below.
Finally, the calculation of ‘levels of engagement’ for Materialising AI controversies: shaping
conversations enabled us to explore the extent to which
controversies with design-led participatory
AI controversies on Twitter enabled a widening of engage-
ment with the issues at stake (inclusion) and the degree to methods
which they instigated processes of problem articulation We cannot forget of course that Twitter analysis provides a
(problematisation). We took the ‘width’ of conversations highly partial perspective on AI controversies. Our Twitter
—many different users replying—as an indication of the research, just as our consultation, captured particular
broadness of engagement, and ‘depth’—long reply chains instances of controversy: for example, the COMPAS con-
—as a mark of sustained engagement in the articulation of troversy on Twitter was dominated by reactions to a celeb-
issues.19 We found that the COMPAS controversy on rity politician, AOC.20 In reality the controversies in
Twitter was comparatively inclusive, while the Stochastic question unfolded across many different settings: for
Parrots controversy evinced the most sustained engage- example, the COMPAS controversy served as an exemplary
ment, suggesting more active involvement in problem case study in policy communities (Selbst et al., 2019) while
articulation as a collective process. We note that, according notable DL controversies were on display during the ‘AI
to this measure, inclusion is marked by a ‘reaction democ- Debates’ periodically held in Montreal in December of
racy’ (Gerbaudo, 2022) logic of bursty but brief engage- 2019, 2020 and 2022. The Stochastic Parrots controversy
ment, while the longer shapes—sustained articulation—are is arguably the only one of the five selected where
associated with strong, arguably more elitist, participation Twitter served as the primary setting of controversy, but
from researchers in controversy. We visualised this this one, too, has a far more complex topology: internal
finding by assigning an overall impressionistic shape to organisational processes at Google, including the firing of
each controversy: controversies of which conversations Timnit Gebru and walk-outs staged by Google staff are
12 Big Data & Society

key events in this controversy. For this reason, we treat our balance of power within the wider domain of AI and
Twitter analysis as providing an initial qualification of AI society. During the workshop, the invited experts worked
and society controversies in English-speaking contexts. In together in five small groups to evaluate the five controver-
the next and last step of our analysis, we used our Twitter sies, supported by material props.
data mappings as a starting point for the evaluation of AI After having been introduced to the five controver-
controversies during a participatory workshop with AI sies, participants worked with a diagnostic tool that
and society experts. we had designed specifically for this purpose—dubbed
A central feature of our methodology of controversy the ‘controversy shape shifter’—to determine the
elicitation is to mobilise the standpoints of a situated com- ‘shapes’ of the selected AI controversies. Asking
munity of experts to activate a collective process of problem participants to cut paper strips to different lengths that
articulation, which is a key affordance of controversy. In a corresponded to values assigned to controversy para-
design-led collaborative workshop that we organised in meters, we invited them to determine: the relevance of
March 2023 in London (Figure 7), we took this process the issues addressed, the degree of participation, the
of situated elicitation of controversy one step further by location of the controversy (situatedness), and the allo-
inviting 35 UK-based AI experts from science, government, cation of responsibility for addressing the problem
industry, activism and the arts, many of whom were respon- (power and solvability) (see Figures 8 and 9). We
dents to our initial consultation, to a design-led participa- offered the participants an explicitly normative overall
tory workshop. Drawing loosely on methods of evaluative framing for this evaluative exercise, by asking them to
inquiry (Marres and de Rijcke, 2020), we designed a diag- tell us: is the controversy in question in good or in
nostic exercise, in which we worked with participants to bad shape? To support their evaluations, we provided
assess the five selected AI controversies in terms of inclu- participants with controversy ‘dossiers’, which included
sion and problematisation: the extent to which they a timeline of events, an actor list, key documents and the
offered opportunities for participation, made visible pro- Twitter analysis for each controversy. On this basis,
blems with contemporary AI, and enabled shifts in the they determined the controversy’s ‘shape’ by assem-
bling long and short strips of cardboard guided by the
evaluative grid.
Table 1. Actor categories across controversies Participants were also encouraged to add notes to their
according to a zero-shot actor classification using a GPT- shapes—annotating the cardboard with pens—summarising
3.5-series model (see Footnote 18). their evaluations based on either the dossiers provided or
their own knowledge and experience.
For the purposes of this paper, we want to highlight one
important result of this process of shaping AI controversies.
Many participant annotations drew attention to the fact that
AI and society controversies operate on two levels at once.
On the one hand, they uncover highly specific issues raised
by AI systems (e.g., a flawed data sharing agreement
between public sector and industry organisations; a prob-
lematic internal approval procedure for academic publica-
tion in the tech industry) but, on the other hand, they
expose major structural problematics (the politicisation of
science, abuse of power and inequality). While the

Table 2. Main topics of disagreement, main actors and overall form of engagement in selected AI research controversies on Twitter.

Style of
Controversy Main topics Prominent participants engagement

COMPAS Bias; datasets; racism; transparency Media; activism Broad


NHS + DeepMind Data protection; consent and trust; data Activism; professions Broad
sharing purpose
Gaydar Harms; validity; pseudo-science; bias Research; activism Deep and broad
Deep learning as solution Artificial general intelligence; human Tech industry; research Broad and deep
for AI vs. machine learning
Stochastic parrots Corporate research culture; Women in tech; Tech industry; research Deep
environmental impact
Marres et al. 13

Figure 6. Forms of engagement: the ‘shapes’ of AI controversies on Twitter.

Figure 7. Shifting AI controversies: shaping and reshaping AI with experts in AI and society, Friends House London, 10 March 2023.

COMPAS case, for example, focused on the racial bias of policy communities, namely the appropriation of public
this scoring software and its underlying data against sector data by private companies, as well as growing Big
ethnic groups, the controversy equally exposed how the Tech’s control over the creation of public sector data
use of algorithmic systems in the public sector amplifies infrastructures.
entrenched racial inequalities. The NHS + DeepMind During the workshop, debates about DL as a solution for
case dealt with technical requirements on the legality of AI highlighted the role of everyday publics in the creation
data sharing agreements (which were not adhered to in training datasets for neural networks through their digital
this case), but simultaneously flagged a major structural participation. Also, what appears as a technical debate
challenge of AI to both national and international public about the ‘architecture’ of predictive models turns out to
14 Big Data & Society

Figure 8. The evaluative grid composed of five parameters: relevance, situatedness, power, participation and solvability.

Figure 9. Shaping of the DL as solution for AI controversy by AI and society experts.

have massive legal ramifications regarding the applicability neural networks to predict sexual orientation raised a
of contemporary copyright law to data infrastructures in major issue of societal harm by demonstrating that AI ana-
society. The Gaydar case’s controversy about the use of lytics could be used to expose people’s vulnerabilities, but
Marres et al. 15

also highlighted that scientists seek to take advantage of nonscience binary. On the one hand, the AI knowledge con-
specific ‘hype’ dynamics to attract attention to their publica- troversies that we identified on Twitter demonstrate a
tions. Finally, the Stochastic Parrots controversy was degree of inclusion, in so far they mobilised experts
perhaps the most dense in its articulation of intersecting across science, journalism, industry and activism on
societal problematics, highlighting ecological impact (the Twitter, However, representatives of affected communities
energy cost of training LLMs), the marginalisation of did not feature prominently in them. It seems that it is espe-
ethical considerations in research debates, and the roles of cially through the type of controversy topics articulated in
women in corporate AI research. This suggests to us that the controversies in question that the expert/non-expert
AI and society controversies, while not necessarily unset- boundary is crossed: these range from issues of social
tling established problem definitions in science and justice to scientific methodology, politics of knowledge,
society, nevertheless articulate unsettling connections ethics and political economy. We note that classic societal
between technical applications of AI, experienced harms —and indeed sociological—problems, such as societal
and structural societal problematics. These sometimes inequality, and the demarcation between science and polit-
unexpected linkages found through the exploration of the ics, play a central role in AI controversies, and arguably, are
selected controversies led us to characterise the topic of rendered publicly relevant through them (on this point, see
AI as an impressively generative ‘super-controversy’. also Roberge and Castelle, 2021). AI and society controver-
sies topicalise classic themes from the sociology of knowl-
edge—the social organisation of science, the role of
Conclusion underlying interests in the structure of knowledge, bias—
Our examination of AI and society controversies via an and entrenched social problems: power differentials and
online expert consultation, Twitter analysis and an evalu- privilege, discrimination and inequality. This contrasts
ative workshop has enabled the elicitation of AI controver- sharply with the understanding of techno-scientific contro-
sies in the following ways. Our online consultation with versy as a force of problem articulation that shifts attention
UK-based experts suggests that recent controversies about away from structural societal problematics such as class,
AI and society not only consist of public disagreements race and gender (Latour, 2005) towards ‘single issues’,
about knowledge propositions (is AI accurate? How do such as pollution or patient rights.
we know it is safe?) but involve demonstrations of socio- We therefore conclude that AI, despite the specificity of
technical frictions in society: occurrences, events and appli- its techno-scientific arrangements and many of the asso-
cations that instantiate the harm that AI does and could do ciated problems, has acquired the status of a ‘super-
in concrete places, with impacts on specific social groups, controversy’ through AI and society disputes. Contrary to
involving the products of specific companies. But the prom- the expectation that public controversies about techno-
inence of AI frictions does not mean that AI and society science surface contingent problems, AI and society contro-
controversies are fragmented. Many of these frictions versies instantiate and connect a range of structural con-
offer demonstrations of entrenched institutionalised dis- cerns across epistemic, political, economic and ethical
crimination against ethnic minorities and women amplified dimensions with AI. This could perhaps be observed most
by the uptake of algorithmic systems in the public sector starkly in the Stochastic Parrots controversy, which actua-
(COMPAS; the UK exams fiasco, the use of facial recogni- lised AI and society conflicts in the form of a located, per-
tion by the police and in schools). Our Twitter analysis both sonal experience of friction (‘Timnit Gebru fired’) and
consolidated and complicated this finding of the relative connected social justice arguments (white privilege) with tech-
prominence of a situational demonstrational logic in nical debates about the meaning of ‘meaning’ (i.e., whether
recent English language AI and society controversies. On generative models ‘understand’ their outputs) and political
the one hand, social justice issues of racial/gendered economy (corporate control over knowledge production).
inequality and entrenched privilege, as well as problems In such a super-controversy, problematisation does not
of political economy, such as the corporate ownership of just proceed through the creation of heterogeneous associ-
the means of knowledge production, featured prominently ation between specific actors and entities, as actor-network
among AI controversy topics identified on Twitter. But in theory suggested, but arises from the forging of connections
this setting, the focus on societal friction in the problem- between techno-scientific propositions (AI)-situated trou-
atisation of AI does not overshadow knowledge contro- bles and entrenched societal problems. The demonstration
versy: on Twitter we equally found AI to be the focus of of AI frictions seems to play an enabling role in this. In
contentious methodological and epistemological debates, this regard, their significance should not be understood
with disagreements about rigour, scientific quality and the solely in terms of the particularisation of AI-induced
science/politics distinction being especially prominent. harm by associating these with specific persons, experi-
We tentatively draw the following conclusions regarding ences, deployments and environments. Equally distinctive
the capacity of AI and society controversies to facilitate about AI frictions is their connective capacity: frictions
inclusion and problematisation across the science/ demonstrate to wider publics how ‘AI’ as a complicated
16 Big Data & Society

techno-scientific domain of application is, nevertheless, implications. Available at: https://openai.com/research/better-


closely and intimately connected with society-wide phe- language-models
nomena, such as structural inequality. Through contro- 4. Goswami R (2023) Microsoft economist warns of A.I. elec-
versy, the firing of a single researcher can topicalise tion interference from ‘bad actors’. CNBC, 3 May 23.
Available at: https://www.cnbc.com/2023/05/03/microsoft-
complicated connections between the political economy
economist-warns-of-ai-bad-actors-meddling-with-elections.
of knowledge production, social justice issues and epi-
html
stemic issues of what constitutes scientific rigour. It is 5. Simonite T (2021) What really happened when Google
also to say that, in the face of the strategic affirmation of Ousted Timnit Gebru. WIRED, 8 June 21. Available at:
the controversiality of AI by its developers and backers, https://www.wired.com/story/google-timnit-gebru-ai-what-
advocates and experts in AI and society have undertaken really-happened/
the problematisation of AI according to a different logic. 6. Merchant B (2023) Column: Afraid of AI? The startups
We are tempted to call this logic ‘sociological’ insofar as selling it want you to be. Los Angeles Times, 31 March 23.
it involves the demonstration of connections between spe- Available at: https://www.latimes.com/business/technology/
cific technical propositions, contextual frictions and struc- story/2023-03-31/column-afraid-of-ai-the-startups-selling-it-
tural problems across social domains. want-you-to-be
7. See Liang P (2021) Introduction to the Workshop on
Foundation Models. Centre for Research on Foundation
Acknowledgements Models, Stanford University. Available at: https://crfm.
We would like to thank all the respondents to our online consult- stanford.edu/workshop.html (accessed 23–24 August
ation ‘What makes AI controversial?’ as well as our intervie- 2021).
wees and the participants in the ‘Shifting AI Controversies’ 8. Eveleth R (2018) Google Glass wasn’t a failure. It raised crucial
workshop, without whose expert contributions this research concerns. WIRED, 12 December 18. Available at: https://www.
would not have been possible. We also thank our colleagues wired.com/story/google-glass-reasonable-expectation-of-
and students in the Centre for Interdisciplinary Methodologies privacy/
at the University of Warwick who took on key roles to help 9. We adopt this more minimal distinction of science versus
realise the workshop. We are grateful to all Shaping AI teams non-science to highlight the importance of differences in
for inspiring discussions. actor’s types of expertise and positioning in controversies
about science and technology, but without making assump-
tions about unequivocal belonging (to science, or society).
Declaration of conflicting interests It was proposed to us by Maria Puig de la Bellacasa, per-
The authors declared no potential conflicts of interest with respect sonal communication, May 2019 (see also Costas et al.,
to the research, authorship and/or publication of this article. 2021).
10. Our standpoint approach to AI research controversies does not
restrict AI controversies to national contexts, but we assume
Funding the national context as standpoint for our study.
The authors disclosed receipt of the following financial support for 11. Alan Turing Institute UK AI Showcase (2021), Ada Lovelace
the research, authorship and/or publication of this article: This at CogX (2021), Oxford AI Ethics Futures (2021), Ada
work was supported by the Economic and Social Research Lovelace Institute Data Stewardship (2021), Mozilla Fest
Council (Grant number ES/V013599/1). (2021); ArtAI Festival (2021); AI Summit (2021); UAL
Tech and Power (2021), AI World Congress (2021).
Respondents were included in our consultation based on
ORCID iDs their demonstrable expertise in the area of ‘AI and society’,
Noortje Marres https://orcid.org/0000-0002-8237-6946 which we identified through their bios, publications, and
Michael Castelle https://orcid.org/0000-0001-6754-6516 other descriptions of their activities.
Beatrice Gobbo https://orcid.org/0000-0001-7085-0017 12. Big Brother Watch (2018) Face Off: The lawless growth of
Chiara Poletti https://orcid.org/0009-0003-4402-1137 facial recognition in UK policing. Report. Available at:
https://bigbrotherwatch.org.uk/campaigns/stop-facial-recognition/
report/; Fussey P and D Murray (2019) Independent Report on
Notes the London Metropolitan Police Service’s Trial of Live Facial
1. OpenAI (2023) GPT-4 System Card. Available at: https://cdn. Recognition Technology. University of Essex, Report.
openai.com/papers/gpt-4-system-card.pdf Available at: https://repository.essex.ac.uk/24946/1/London-
2. See for example: Helmore E (12023) We are a little bit scared: Met-Police-Trial-of-Facial-Recognition-Tech-Report-2.pdf;
OpenAI CEO warns of risks of artificial intelligence. The National Physical Laboratory (2020) Metropolitan Police
Guardian, 17 March 23; Ordonez V, Dunn T and Noll E Service, Live Facial Recognition Trial. Report. Available at:
(2023) OpenAI CEO Sam Altman says AI will reshape https://www.met.police.uk/SysSiteAssets/media/downloads/
society, acknowledges risks: ‘A little bit scared of this’. central/services/accessing-information/facial-recognition/met-
ABC News, 16 March 23. evaluation-report.pdf
3. As stated in a blogpost published on the company website at 13. Couplings were identified in the data as follows: in coding
the time. OpenAI (2019) Better language models and their the consultation, we assigned responses to one of more
Marres et al. 17

categories (controversy topic, friction, and problematisa- examination of likely bias is clearly required (see also
tion). Frictions and problematisations were assigned to Ashwin et al., 2023).
topics, where the latter were not explicitly identified in the 19. We ascertained the ‘level of engagement’ for each contro-
responses. versy by calculating the width and depth of its in scope con-
14. We created these data sets by constructing Twitter Academic versations, where ‘width’ refers to the highest number of
API queries. This involved an iterative process of testing branching threads originating from an initial tweet, and
terms and URLs from the consultation, and quantitative ‘depth’ is the overall length of the conversation’s longest
aggregation of subsets using the Twitter API. We set the thread.
start date for data collection to one month before publication 20. One very large conversation in this data set originated from a
of the relevant papers. Tweets matching our query as well as tweet from Alexandria Ocasio-Cortez. https://twitter.com/
all the replies and quote tweets were downloaded using the RealSaavedra/status/1087627739861897216
Twitter ‘search all’ API endpoint. The controversial publica-
tions in question are: Angwin J, Larson J, Mattu S and
Kirchner L (2016) Machine bias. ProPublica, May 23 16.
Available at: https://www.propublica.org/article/machine- References
bias-risk-assessments-in-criminal-sentencing; Powles J and Alfrink K, Keller I, Kortuem G, et al. (2022) Contestable AI by
Hodson H (2017) Google DeepMind and healthcare in an design: Towards a framework. Minds & Machine 33: 613–639.
age of algorithms. Health and Technology (Berl) 7(4): 351– Angwin J, Larson J, Mattu S, et al. (2016) Machine bias.
367; Wang Y and Kosinski M (2018) Deep neural networks ProPublica, May 23 2016 Available at: https://www.
are more accurate than humans at detecting sexual orientation propublica.org/article/machine-bias-risk-assessments-in-criminal-
from facial images. Journal of Personality and Social sentencing
Psychology 114(2): 246–257; Bender EM, Gebru T, Ashwin J, Chhabra A and Rao V (2023) Using large language
McMillan-Major A and Shmitchell S (2021) On the dangers models for qualitative analysis can introduce serious bias.
of stochastic parrots: Can language models be too big?. In arXiv preprint arXiv:2309.17147.
Proceedings of the 2021 ACM conference on fairness, Barnes B (1977) Interests and the Growth of Knowledge. London:
accountability, and transparency, pp. 610–623; Marcus G Routledge.
(2018) Deep learning: A critical appraisal. arXiv preprint. Barry A (2021) What is an environmental problem? Theory,
15. We also set a threshold for the minimum required number of Culture & Society 38(2): 93–117.
tweets in the conversation to 10 replies and/or quote-tweets Bender EM, Gebru T, McMillan-Major A, et al. (2021) On the
(an engagement-related criterion). dangers of stochastic parrots: Can language models be too
16. Each in-scope conversation was assigned 1–4 topics that sum- big?. In Proceedings of the 2021 ACM conference on fair-
marise the main issues addressed in the conversation with ness, accountability, and transparency, pp.610–623.
relevance to the specific controversy. We then assigned each Burgess J and Matamoros-Fernández A (2016) Mapping
topic to broader themes. sociocultural controversies across digital media platforms:
17. We assigned scores on a scale from 0 (no disagreement) to 2 One week of #gamergate on Twitter, YouTube, and Tumblr.
(high indications of disagreement, indicated by the presence Communication Research and Practice 2(1): 79–96.
of several contestational comments like ‘you’re wrong’, Callon M, Lascoumes P and Barthe Y (2011) Acting in an
‘no…’, etc.). Uncertain World: An Essay on Technical Democracy.
18. We categorised accounts manually by coding Twitter bio Cambridge: MIT press.
descriptions for a sample of 1000 accounts randomly Collier J (2007) The collaboratory form in contemporary
selected for the Stochastic Parrots data set, and then, anthropology. Concept Work and Collaboration in the
based on the manual coding process, we designed a Anthropology of the Contemporary, pp.54–61.
prompt for a zero-shot actor classification of a random Costas R, de Rijcke S and Marres N (2021) “Heterogeneous cou-
sample of 1000 Twitter account bios for each data set, plings”: Operationalizing network perspectives to study
using category labelling by the GPT-3.5 model endpoint science-society interactions through social media metrics.
text-davinci-003. The term ‘zero-shot’ refers to the fact Journal of the Association for Information Science and
that the model is neither ‘fine-tuned’ on domain-specific Technology 72(5): 595–610.
example data, nor is it given explicit examples (‘few-shot’ Crépel M, Do S, Cointet JP, et al. (2021) Mapping AI issues in
learning), although model development did involve media through NLP methods. In: Computational humanities
‘instruction tuning’ (fine-tuning on human-generated dia- research conference, Amsterdam, NL, 17–19 November 2021,
logical examples) and reinforcement learning from human pp.77–91. https://ceur-ws.org/Vol-2989/long_paper22.pdf
feedback (Ouyang et al., 2022). More recent evidence sug- Dandurand G, Blottière M, Jorandon G, et al. (2022) Training the
gests fine-tuning on example data by social science News: Coverage of Canada’s AI Hype Cycle (2012–2021).
researchers would have improved our results (Ziems et al., Quebec: INRS—Urbanisation Culture Société.
2023). Inter-coder comparability was relatively low Funtowicz SO and Ravetz JR (1991) A new scientific methodology
between the manual and automatic coding. This can partly for global environmental issues. In: Costanza R (ed) Ecological
be explained by the fact that we assigned only one category Economics: The Science and Management of Sustainability.
to each bio, and prompted GPT to do the same, while many New York: Columbia University Press, pp.137–152.
bios listed multiple occupations. The GPT classifications Geiger S (2020) Silicon Valley, disruption, and the end of uncer-
were found to be generally plausible, but further tainty. Journal of Cultural Economy 13(2): 169–184.
18 Big Data & Society

Geiger S, Harrison D, Kjellberg H, et al. (2014) Being concerned Powles J and Hodson H (2017) Google deepmind and healthcare
about markets. In: Concerned Markets. Cheltenham, UK: in an age of algorithms. Health Technology 7(4): 351–367.
Edward Elgar Publishing, pp.1–18. Roberge J and Castelle M (2021) Toward an end-to-end sociology
Gerbaudo P (2022) Theorizing reactive democracy: The social of 21st-century machine learning. In: Roberge J and Castelle M
media public sphere, online crowds and the plebiscitary logic (eds.) The Cultural Life of Machine Learning: An Incursion
of online reactions. Democratic Theory 9(2): 120–138. into Critical AI Studies, Basingstoke: Palgrave, pp. 1–29.
Kak A and Meyers West S (2023) Confronting tech power, 2023 land- https://doi.org/10.1007/978-3-030-56286-1
scape. AINow Institute. https://ainowinstitute.org/2023-landscape Rogers R (2017) Foundations of digital methods: Query design.
Kling R (1996) Computerization and Controversy: Value In: Schäfer MT and van Es K (eds.) The Datafied Society:
Conflicts and Social Choices. 2nd ed. San Diego, CA: Studying Culture Through Data. Amsterdam, Netherlands:
Morgan and Kaufman. Amsterdam University Press, 75–94. https://doi.org/10.5117/
Latour B (2005) Reassembling the Social: An Introduction to 9789462981362
Actor-Network-Theory. Oxford: Oxford University Press. Rogers R and Marres N (2000) Landscaping climate change: A
Lupton D (2014) Digital Sociology. London and New York: mapping technique for understanding science and technology
Routledge. debates on the World Wide Web. Public Understanding of
Madsen AK and Munk AK (2019) Experiments with a data- Science 9(2): 141.
public: Moving digital methods into critical proximity with Ruppert E, Harvey P, Lury C, et al. (2015) Socialising Big Data:
political practice. Big Data & Society 6(1): 1–14. From Concept to Practice. Manchester: CRESC, The
Mannheim K (2015/1936) Ideology and Utopia: An Introduction University of Manchester and the Open University.
to the Sociology of Knowledge. New York: Harcourt, Brace Selbst AD, Boyd D, Friedler SA, et al. (2019) Fairness and
and Company. abstraction in sociotechnical systems. In: Proceedings of the
Marcus G (2018) Deep learning: A critical appraisal. arXiv preprint. conference on fairness, accountability, and transparency,
Marres N and de Rijcke S (2020) From indicators to indicating Atlanta, USA, 29–31 January 2019, pp.59–68. Association
interdisciplinarity: A participatory mapping methodology for for Computing Machinery. https://doi.org/10.1145/3287560.
research communities in-the-making. Quantitative Science 3287598
Studies 1(3): 1041–1055. Shaffer Shane T (2023) AI incidents and ‘networked trouble’:
McGoey L (2021) Philanthrocapitalism and the separation of powers. The case for a research agenda. Big Data & Society 10(2):
Annual Review of Law and Social Science 17: 391–409. 1–6.
Meunier A, Gray J and Ricci D (2021) A new AI lexicon: Stark L, Greene D and Hoffmann AL (2021) Critical perspec-
Algorithm trouble: Troublesome encounters with algorithms tives on governance mechanisms for AI/ML systems. In:
that go beyond computational processes. In: A New AI Roberge J and Castelle M (eds.) The Cultural Life of
Lexicon. New York: AI Now Institute, pp.1–9. Machine Learning: An Incursion into Critical AI Studies.
Nishi R, Takaguchi T, Oka K, et al. (2016) Reply trees in twitter: Basingstoke: Palgrave, pp.257–280. https://doi.org/10.1007/
Data analysis and branching process models. Social Network 978-3-030-56286-1_9
Analysis and Mining 6: 1–13. Suchman L (2023) The uncontroversial ‘thingness’ of AI. Big
Ouyang L, Wu J, Jiang X, et al. (2022) Training language models Data & Society 10(2): 1–5.
to follow instructions with human feedback. arXiv. https://doi. Venturini T and Munk AK (2021) Controversy Mapping: A Field
org/10.48550/arXiv.2203.02155 Guide. Cambridge: Polity.
Parks P (2021) Non-representational news: An intervention into Wang Y and Kosinski M (2018) Deep neural networks are more
pseudo-events. Journalism 22(1): 122–138. accurate than humans at detecting sexual orientation from
Patey DL (1986) Johnson’s refutation of Berkeley: Kicking the facial images. Journal of Personality and Social Psychology
stone again. Journal of the History of Ideas 47(1): 139–145. 114(2): 246–257.
Phan T, Goldenfein J, Mann M, et al. (2022) Economies of virtue: Ziems C, Held W, Shaikh O, et al. (2023) Can large language
The circulation of ‘ethics’ in Big Tech. Science as Culture models transform computational social science? Computational
31(1): 121–135. Linguistics 50(1): 237–291.

You might also like