Abstract
I argue that prediction is not primarily a technological means for knowing future outcomes, but a social model for extracting and concentrating discretionary power. Prediction is a ‘relational grammar’ that governs this allocation of discretion: the everyday ability to define one's situation. This extractive dynamic extends a long historical pattern, in which new methods for producing knowledge entail a redistribution of decision-making power. I focus on two contemporary domains: (1) crime and policing are emblematic of how predictive systems are extractive by design, with pre-existing interests governing what is measured and what persistently goes unmeasured. (2) The prediction of productivity demonstrates the long tradition of extracting discretion as a means to extract labour power. Time after time, making human behaviour more predictable for the client of prediction (the manager, the police officer) often means making life and work more unpredictable for the target of prediction (the employee, the urban citizen).
Introduction
The promise of prediction has become central to the real-world implications of AI and data-driven decision-making. When applied to human behaviours and social outcomes, prediction goes far beyond narrow technical definitions like “a model's output when provided with an input example” (Google, n.d.), and increasingly involves asserting probable outcomes as bases for definitive judgments - despite mounting evidence that much of what we call AI is “fundamentally dubious” (Narayanan, 2019) for assessing social phenomena like job performance or recidivism (also see Raji et al., 2022). I argue that prediction is primarily not a technological instrument for knowing future outcomes, but a social model for extracting and concentrating discretionary power: that is, people's ordinary capacity to define their situation
Viewed this way, prediction extends the long history of how new methods of producing knowledge generate a redistribution of epistemic power: that is, who declares what kind of truth about me, to count for what kinds of decisions? (e.g., Deringer, 2018; Poovey, 1998; Shapin and Schaeffer, 1985) Below, I describe prediction as a relational grammar which generates social expectations on how facts are made and what accuracy or objectivity looks like when it comes to human behaviour (similarly on the notion of fairness, see Hoffmann, 2021). It follows that such grammar prioritises certain kinds of factmaking over others. Focusing on crime and policing, I then examine how predictive systems are extractive by design. Choices in what kinds of predictive models are created to pursue what kinds of objectives generate systematic disparities between what tends to get measured and what does not. The result is that novel mechanisms of capture and analysis regularly reproduce pre-existing social asymmetries. The paper demonstrates this historical pattern by turning to predictions of productivity – a site where the technological extraction of discretion extends capital's long extraction of labour power. In sites like the Amazon warehouse, making human behaviour more predictable for the client of prediction frequently means making life and work
This approach to prediction extends the growing critical conversation around concepts like fairness, transparency, and the very notion of AI ethics. The question of how to pursue normative values in AI has increasingly raised the problem of how to define those values, and
Prediction as relational grammar
The history of technology amply demonstrates that new technologies do not enter as blank slates, but that their meaning and usefulness are derived from preexisting social relations and conflicts (e.g., Marvin, 1988). The social impact of technologies also tends to exceed their actual capabilities or implementation. The history of 20th century automation, for instance, was dominated by institutional and economic responses to what people
Prediction is particularly important to these mythological dimensions because it is at the vanguard of both (allegedly) concrete, real world benefits and the enveloping fantasy of long-term progress. The claim to predict a tangible ‘problem’ like recidivism or productivity promises a powerful empirical basis for thinking about new data-driven technologies as desirable and inevitable. Tech firms promise that algorithms can, for example, automatically analyse U.S. prisoner phone calls to predict violent crimes and suicide (Sherfinski and Asher-Schapiro, 2021), identify suspect individuals based on their head vibrations (Wright, 2021), and use facial features to predict prospective employees’ personality (HireVue). This remains the case even when many such predictions remain unproven or, at times, specifically disproven. COMPAS, a widely used recidivism prediction tool that leverages over 100 indicators, performs about as poorly as laypersons’ guesswork (Dressel and Farid, 2018). AI-driven tools were celebrated as potential difference-makers during the COVID-19 pandemic, but were soon found to be largely useless, plagued by poor input data and other fundamental problems (Heaven, 2021). Yet even as individual applications of prediction might be criticised and lampooned, such promises of real-world payoffs continue to generate emotional and cognitive investment in the idea of prediction (also see Finn, 2017).
These reports reflect a broader pattern in which social life is so often driven by the disjuncture between the substance of what passes as true in a given context and a broader sense of what kinds of things are likely to look, feel, and smell true. Foucault famously distinguished between what statements are treated as true, and statements that are considered
But what does it mean to
This very idea of modular and general relation is neither timeless nor universal. It is very much a modern way of seeing, and one which resonates with the long advent of quantification as a dominant method for understanding human and social phenomena (Deringer, 2018; Porter, 1995). In this sense, the language of data-driven prediction imposes its own
Crucially, any description of data-driven systems as objective and consistent represent an aspirational ideal rather than actual practice. In their history of scientific objectivity, Lorraine Daston and Peter Galison write that “all epistemology begins in fear—fear that the world is too labyrinthine to be threaded by reason; fear that the senses are too feeble and the intellect too frail”. (2007: 372) Prediction as relational grammar does not simply dictate a smooth, totalising blanket of mathematical reason. Equally significant is what gets concealed or devalued in the process of imposing that grammar. Critical media scholar Tara McPherson (2012) calls it a lenticular logic, referencing the lenses in old 3D postcards where one can flick between multiple images but never see them together. She argues that the history of computing and its attendant lessons are often sealed off from other domains. While McPherson takes up this point specifically around to race and computing, we can also consider how computing imposes its own ways of seeing on those by other domains. Drawing from James Scott's work on ‘seeing like a state’ (1998), Ali Alkhatib argues that the indiscriminate proliferation of a data-driven way of seeing adds up to an asphyxiating utopia: an expectation of calculability which drives out everything that does not fit. “The models AI researchers train and deploy construct worlds without space for the dimension of our lives that make us who we are” (Alkhatib, 2021: 3). How we attribute predictability – for instance, that a single statistical model can be ‘scaled’ to predict housing prices (Zillow) or ‘suspicious eye movements’ (Proctorio) across many different lives and locales – has direct repercussions for what kinds of actors are considered rational, and what kinds of behaviours are considered optimal. Crucially, the grammar of prediction tends to prefer particular kinds of data and relations over others: a preference driven not purely by epistemic reasons, but by economic and institutional conditions that make some datasets and problems more available than others. Often, datasets treated as general benchmarks, like ImageNet and GLUE, are built not through rigorous definition of what is truly representative of the target reality, but look “more like samples of convenience” (Raji et al., 2021: 5). Subsequently, any arbitrariness or partiality in such benchmarks tends to flow downstream into derivative applications in a ‘blood diamond effect’ (Prabhu and Birhane, 2020). These limitations are not purely technological in the sense that more diverse datasets and models are simply impossible; rather, they tend to replicate longstanding economic, political, and institutional asymmetries in what kinds of data and knowledge get produced, documented, reported, and archived, in the first place. As Scott himself would later note: “legibility doesn’t come cheap.” (2021: 513)
Extractive by design
Bad questions
Prediction, like any knowledge regime, has a self-fulfilling element: it sees what it knows to see, and it measures what it is used to measuring. These tendencies are shaped through longstanding economic and political asymmetries, whose influence is then set aside as ‘externalities’. But when prediction troubles get compartmentalised as ‘just a dataset problem’, they obfuscate how patterns of extraction shape the research questions and the choice of what to measure (and what to dismiss without measuring). When predictive successes are defined as ‘beating’ humans in tests of accuracy, they reify the model's own parameters as the world that matters, further burying these externalities beyond consideration.
The sociologist Luc Boltanski, with Laurent Thévenot (2006) and Eve Chiapello (2007), has shown how capitalism constructs
Today, prediction regularly appears as an epithet of approval, in which the technique
On the surface, predictive systems rarely
Bad measures
These patterns extend into the measurements themselves. As critical researchers of race and technology have shown with particular clarity, pre-existing inequalities shape what kinds of predictions roam the world in the first place, endowed with what kind of legitimacy (e.g., Benjamin, 2019; Browne, 2015). To suggest that the neutrality of data or algorithm cleanses predictions of their own historical provenance effectively describes a variation of money laundering, in which decisions like what kind of data is and is not gathered, or where the resulting ‘insights’ are deployed, are written off as someone else's problem. The predictive output, like ‘clean’ money, can then be treated as a neutral object of use. However, contemporary predictive systems’ need for massive dataset makes them heavily reliant on existing records and practices to source the data, with all their incumbent theories and assumptions about crime and policing. Ben Green (2019) describes how one University of Southern California project began by seeking to predict ‘adversarial groups’ like ISIS, but soon pivoted to targeting ‘criminal street gangs’. They ended up drawing data from the notoriously problematic LAPD gang data for the model, effectively amplifying and legitimating its errors and prejudices. All this happens despite the fact that “there is no agreement as to what predictive [policing] systems should accomplish […] nor as to which benchmarks should be used”, and even if there were, “like all evaluations of police technology, confounding factors make it impossible to measure directly its effectiveness at reducing crime” (Shapiro, 2017: 459).
At the same time, those who develop and sell surveillance technologies are incentivised to constantly produce useful justifications on the dangerousness of crime and the necessity of advanced policing. Technologies like PredPol and ShotSpotter are directly sold to law enforcement clients, preying on those clients' anxieties about being left behind and their perpetual interest in expanding budgets (Amoore, 2020: 2–4). Consumer-facing surveillance similarly increases fears of urban crime. In what has been called ‘digital vigilantism’ (Trottier, 2017), platforms like Amazon Ring's Neighbours incite and collate user data – the video footage and the neighbourhood alerts – into a picture of the world in which crime is always just around the corner. Citizen, the most aggressive company in this market niche, repackages dramatic cases of crime thwarted into ‘magic moments’ for consumption, and even deploys ‘street teams’ to make content when none is organically forthcoming (Ashworth, 2020). With Amazon Ring signing up over 400 local police departments for two-way data sharing, such business models generate new, ‘efficient’ pathways for connecting existing prejudices and perverse economic incentives with the aura of technological innovation (Bridges, 2021).
The choice of what to measure and how to measure caters to these existing social scripts, often entrenching their distribution of discretion across different subjects. But this is not simply a zero-sum game, where greater automation or rationalisation leads to less human discretion. Rather, the deployment of predictive systems can serve to protect and relegitimise bureaucratic discretion by integrating those judgments into the technology. The move towards data-driven policing often involves new systems of documentation and data-collection, as well as existing data taking on new uses and importance, following a longer pattern of modern policing as
The unmeasured
The common thread running through these different sites of data productionis that where the police remains the primary purchaser of these predictions, the kind of data and models we “get” as a society will invariably reaffirm their definitions of crime and safety. In the US context, conspicuously missing in this thriving industry is any major efforts to improve the disastrous undercollection and invisibility of data around police violence and misconduct. Kelly Gates writes that such asymmetry in the sheer availability of data entails a ‘symbolic annihilation’ (2015: 4), in which one's experience is not simply underappreciated, but directly overwritten by police-oriented data production. From Michael Brown and Eric Garner in 2014 to George Floyd and Breonna Taylor in 2020, high-profile police killings of Black Americans precipitate repeated public debate on just how prevalent such killings are. Yet each time, it is data on Black
Such asymmetries condemn certain kinds of suffering and lived experience as ‘merely anecdotal’, forced to an uphill struggle to count as ‘data’, while the police receive the numbers they need by default. Samuel Sinyangwe, co-founder of the Mapping Police Violence project, describes the recurrent barrier: ‘How often have you heard, “We don’t have the data. We don’t know if what you’re telling us is true.”’ (in Foss et al., 2017) Although more robust databases of police misconduct are becoming publicly available, they tend to come from small, independent alliances of researchers, journalists and activists working with limited resources. While the police are legally obligated to produce some kinds of misconduct data, they routinely engage in strategies of obfuscation and delay, turning theoretical availability into practical absence (Cox and Freivogel, 2021). As a result, whatever becomes transparent about those killed by police—their misdeeds, their personal life—tends to be far more exhaustive than the data that we can get about the police themselves (Gates, 2019). Predictive systems are often used to delegitimise the kinds of lives and experiences that are already too disadvantaged to generate rich ‘data’ in the first place.
Entrenching discretion
Defining the situation
Discretion typically describes the ability to assess how a rule might apply (or
These qualities link prediction indelibly to discretion. After all, predictive technologies are tasked with rationalising judgment, in both common senses of the word: they establish rule-bound processes for decision-making, and further provide scripts for justifying that judgment to insiders and outsiders (also see Hall, 2017). These functions impinge directly on the exercise of discretion. This is not to say that prediction always enforces formalisation and Kafkaesque inevitability at the expense of discretion. Rather, prediction serves to
In so many places in social life, discretion is something one exercises upon another, imposing one's definition of the situation over others. At the macro level, there is a dyadic struggle; the history of labour automation and prediction, as I discuss below, is a series of managerial extraction of worker discretion. But if new technologies enable new strategies for such extraction, deploying them into action often entails new groups of intermediaries, the
Discretion, then, describes the always unequal distribution of the power to
Predictable workers, unpredictable work
I now turn to efforts to predict productivity and worker behaviour. As I noted earlier, prediction's transfer of discretion from the target of prediction to its client often dovetails with other historical patterns of extraction, and in particular, that of labour power. Consider the case of Amazon. As one of the largest private employers in the world, Amazon is both test-bed and ground reality for prediction as the extraction of worker discretion. Its warehouses (‘fulfilment centres’) and delivery networks exemplify the synergy between intense, data-driven automation of labour on the one hand, and the systematic deprivation of information, predictivity, and discretion for the workers on the other. Here, the master metric is “the rate”: individual performances like boxes stowed, picked or packed per minute and hour. The rate is regularly used within the company as a reliable base measure for decisions including the hiring and firing of individual workers; it provides an always cooperative reference point for justifying Amazon's abnormally high rate of firings (Lecher, 2019). Yet notably, the exact rate that a worker must hit at a given warehouse is never disclosed to the workers themselves. Sociologist Nantina Vgontzas (2021), who has worked at Amazon warehouses and joined in workers’ organising efforts, recounts being trained against the rate of 400 picks per hour, and later being told by a human trainer that the actual expectation was around 360. Meanwhile, the demand for a machine-readable and predictable labour rate is further enacted through a sensory opprobrium. Coloured graphs warn employees to work faster; ‘project loving energy’, counsels a workstation screen (Delfanti, 2021b: 56).
These quantified expectations governing the algorithmic workplace cater to managers and employers’ desire for a certain kind of inhuman clarity, in which the many variations and ambiguities inherent in any act of labour are not actually eliminated, but simply neglected. The consequence is that for the worker, their own work and life becomes both less predictable
This disparity in control results in all too familiar harms to working conditions. Amazon's warehouse workers have long experienced significantly higher rates of severe injury compared to industry competitors (Greene and Alcantara, 2021). Amazon's delivery drivers are subject to similar pressures: top-down dictations of punishing rates, sometimes of up to 400 deliveries per 10-hour shift, provide profits at the macro level, while workers pay in the form of extraordinarily high injury rates and the now infamous practice of drivers having to pee in bottles or trash bags to meet the rate (Gurley, 2021). Amazon's response to these problems are to intensify this dynamics of profit-oriented datafication, rather than to restore discretion on the floor level. In 2021, Jeff Bezos’ final letter to shareholders as the CEO of Amazon prescribed “new automated staffing schedules that use sophisticated algorithms to rotate employees among jobs that use different muscle-tendon groups” (Ongweso Jr., 2021) - a solution that would implement new forms of unpredictable, opaque decision-making on behalf of workers, while retaining the non-negotiability of ‘the rate’ and punitive productivity standards.
Amazon may be a pioneer in the algorithmic destruction of workers' bodies, but its underlying economic and technical incentives are not at all unique. Coupang, South Korea's answer to Amazon (though heavily dependent on SoftBank's Saudi-backed funds, as is the case for many Silicon Valley companies), has also seen rocketing stock value and a reputation for superfast delivery on the back of algorithmic surveillance, high rates of workplace injury, and several cases of death from overwork (Lee, 2021; Seol, 2021). In delivery platforms like DoorDash, the on-demand worker is shephered into surge areas in one moment, then made to kick their heels curbside in another, never knowing how much they’ll make or how to pace their work (Gregory and Sadowski, 2021: 7; Shapiro, 2020: 2). Chinese food couriers, working for platforms like Ele.me and Meituan Waimai, describe sudden periods of intense pressure in which a delay of seconds might cost several orders’ worth of pay, the result of a business and technological model which profits from its ability to “stratif[y] the value of people's time” in rapid, finely-grained ways, extracting every possible sliver of ‘downtime’ (Chen and Ping, 2020: 1563). For the worker, it is their lived time chunked up into a pulsating mess of alarms and nudges, distractions and panic; for the manager and employer, it is a vast predictive matrix which externalises everything that the model would prefer not to predict and funnels the cost to the worker. What the model refuses to count cannot hurt it.
One extraction for another
This disparity between the predictor and the predicted reprises over a century of labour struggle, throughout which the extraction of discretion has served as a crucial instrument for the extraction of labour power. ‘Surveillance capitalism’, currently one of the most common critical frames around prediction, describes not a seismic revolution, but a variation upon the theme that is capitalism itself (Morozov, 2019). Delfanti (2021a) argues that automation
This dynamics runs right through the history of modern automation. Harry Braverman's landmark work,
These recurrent efforts to separate the worker from their discretion come hand-in-hand with longstanding cultural tropes about the nature of the worker and what is
Predicting to stay the same
Prediction
When understood as the extraction and redistribution of discretionary power, data-driven prediction appears more clearly as a reproduction of longstanding strategies for social control, rather than any radically novel approach towards human bias and irrationality. David Noble relays an evergreen definition of workplace automation from none other than Peter Drucker, perhaps the closest available personification of corporate managerial logic: “what is today called ‘automation’ is conceptually a logical extension of Taylor's scientific management [in which] productivity required that ‘doing’ be divorced from ‘planning’.” (in Noble, 2011: 231) From the gig economy worker left to kick their heels waiting for the algorithm to bestow the next job, to Amazon's patents for smart devices guiding warehouse workers on which boxes to dig for the delivery item (Hong, 2020:101), this separation of doing from planning deprives the ‘targets’ of the very information and experience that can become the basis for theorising and contesting their own situation.
To adapt Joseph Weizenbaum's famous dictum on computing, not everything that can be predicted should be predicted (1976: x). The very act of declaring something predictable, of turning it into a prediction ‘problem’, already establishes the conditions for extraction of discretion. The drive to predict deploys well rehearsed templates for decision-making in which the clients of prediction are empowered to consistently speak over and speak for the targets of prediction: the police departments over inhabitants of overpoliced neighbourhoods, and the managers and executives over warehouse workers. As technologies like facial recognition increasingly impose themselves over diverse social situations, they deliver not a rationalised environment in which the data speaks for itself, but a thin refurbishing of familiar asymmetries in which the same few might continue to judge the many with impunity.
Footnotes
Acknowledgments
This paper is an accidental byproduct of research conducted through the Canadian Social Sciences and Humanities Research Council, “Personal Truthmaking in the Data-driven Society.” An earlier version of this paper was presented at the ACM Fairness, Accountability and Transparency 2022, Seoul, South Korea, 21–24 June. Many thanks to Ben Gansky and Sean McDonald, the Forking Room Collective, Théo Lepage-Richer and Ranjodh Singh Dhaliwal, and Seth Lazar.
Declaration of conflicting interests
The author(s) declared no potential conflicts of interests with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed the receipt of the following financial support for the research, authorship, and/or publication of this article: This research was funded by the Social Sciences and Humanities Research Council of Canada (grant number 430-2019-00078).
