Abstract
This research explores how a new relation of production—the shift from human managers to algorithmic managers on digital platforms—manufactures workplace consent. While most research has argued that the task standardization and surveillance that accompany algorithmic management will give rise to the quintessential “bad job” (Kalleberg, Reskin, and Hudson, 2000; Kalleberg, 2011), I find that, surprisingly, many workers report liking and finding choice while working under algorithmic management. Drawing on a seven-year qualitative study of the largest sector in the gig economy, the ride-hailing industry, I describe how workers navigate being managed by an algorithm. I begin by showing how algorithms segment the work at multiple sites of human–algorithm interactions and how this configuration of the work process allows for more-frequent and narrow choice. I find that workers use two sets of tactics. In engagement tactics, individuals generally follow the algorithmic nudges and do not try to get around the system; in deviance tactics, individuals manipulate their input into the algorithmic management system. While the behaviors associated with these tactics are practical opposites, they both elicit consent, or active, enthusiastic participation by workers to align their efforts with managerial interests, and both contribute to workers seeing themselves as skillful agents. However, this choice-based consent can mask the more-structurally problematic elements of the work, contributing to the growing popularity of what I call the “good bad” job.
Keywords
In recent years, the number of people in the United States who show up to work by turning on an app on their smartphone has dramatically increased, redefining our understanding of work, labor, and employment issues. Dubbed on-demand, platform, or gig workers, these individuals log in to digital apps in which algorithms instead of humans manage workers, rewarding, disciplining, and evaluating them. Researchers have sounded alarms about the rise of algorithmic “tyranny” (Lehdonvirta, 2018: 27), “cruelty” (Gray and Suri, 2019: 67), and “despotism” (Griesbach et al., 2019: 1) that trap workers in an “encompassing” (Kellogg, Valentine, and Christin, 2020: 366) “invisible cage” (Rahman, 2021: 945). On-demand work, especially work on popular digital platforms such as Uber and Instacart, is often seen as the epitome of a “bad job,” i.e., characterized by variable and low wages, unpredictable schedules, no formal job ladders, and high risk (Kalleberg, Reskin, and Hudson, 2000; Kalleberg, 2011; Ravenelle, 2019). Furthermore, because on-demand workers are classified as independent contractors, they are not covered by employment protections, such as workers’ compensation, unemployment insurance, and health insurance, which are part and parcel of what is considered the quintessential high-quality “good job” (Kalleberg, 2011; Cappelli and Eldor, 2023). Yet, workers continue to opt for on-demand jobs despite the increasing availability of traditional jobs (Katz and Krueger, 2019; Kaplan et al., 2021; Garin et al., 2023) of high quality (Aeppli and Wilmers, 2022; Autor, Dube, and McGrew, 2023; Newman and Jacobs, 2023), with one-third of workers preferring algorithmic over human managers (Östergaard, 2017). These findings suggest that these individuals may not experience the work to be as deleterious as many scholars have argued (Shapiro, 2018; Gandini, 2019; Ravenelle, 2019). Thus, the puzzle remains regarding why people participate in and even claim to like such precarious work.
Emerging technology and algorithmic management, in particular, open up new questions about how work is organized and how it is experienced by people. In a growing stream of research about on-demand work and algorithmic management, studies often end with the punch line that workers are controlled by an all-encompassing, comprehensive management system (e.g., Rosenblat and Stark, 2016; Gandini, 2019; Griesbach et al., 2019; Rahman, 2021). This critical view of algorithmic management stems, in part, from an unnecessarily limited view of how workers can express agency within these systems. Because much of the research on algorithmic management centers on control, the limited literature that has focused on workers’ agency has categorized it as either collective action (Wood, Lehdonvirta, and Graham, 2018; Tassinari and Maccarrone, 2020; Lei, 2021) or individual resistance (Shapiro, 2018; Cameron and Rahman, 2022). Such research often treats consent as mere compliance with management’s objectives, describing workers as passive, unresisting dupes to managers’ designs (see Hodson, 1991 as an example). But consent, like resistance, can entail workers’ agency. Here, I define consent as individuals’ enthusiastic and active participation to align their efforts with managerial objectives and even to exceed managerial goals (Burawoy, 1979; Hodson, 1999; Padavic, 2005; Mollick and Rothbard, 2014). Through workers’ subjective well-being, i.e., finding meaning and fulfillment in their work, workers’ and managerial interests become aligned.
A rich literature in labor process theory examines the relationship between workers’ experience and workplace consent. This research highlights how social relations with coworkers and managers make jobs tolerable under precarious, bad conditions. In the absence of managers, workers in manufacturing units encouraged one another to work faster, eschewing organizational safety procedures (Hayes and Wheelright, 1984; Barker, 1993; Bernstein, 2012). In myriad manufacturing and service jobs, workers competed against one another, playing games to gain social status by beating the piece rate or baiting a customer for larger tips (Burawoy, 1979; Sherman, 2007; Sallaz, 2009). And when managers cloaked their authority under the guise of friendship and offered preferential treatment, individuals remained in demanding jobs, tolerated callous behavior from customers, and put up with irregular schedules (Anteby, 2008; Mears, 2015; Sallaz, 2019).
However nuanced, prior ways of understanding workplace consent are inadequate for fully understanding how consent is produced in algorithmically managed work. While early research considered the relationship between technology and consent (e.g., the role of the machine in workplace games, Burawoy, 1979), the role of technology and, particularly, algorithmic management is noticeably absent in contemporary research. Like electricity, the telephone, and the internal combustion engine, algorithms are an “infrastructural technology” (Barley, 2020: 8); they change the system of production and the labor process (Aneesh, 2009; Kellogg, Valentine, and Christin, 2020; Vallas and Schor, 2020), which has significant implications for the manufacturing of consent. Unlike work on the shop or service floor, work on digital platforms occurs through an electronically mediated exchange between the worker, the customer, and the platform’s algorithms. On-demand work is largely asocial, as workers have little or no in-person contact with managers and coworkers. And because workers are classified as independent contractors, platform companies do not have the legal obligations (e.g., duty of care, fiduciary responsibility) embedded in the employment relationship that might elicit workers’ participation; nor is it legal to try to develop these practices (Cappelli and Keller, 2013; Cappelli and Eldor, 2023). 1 Without coworkers or managers to act as a co-participant in the production of consent, this arrangement raises the question, how is consent generated in algorithmically managed work? To answer this question, we must first understand the granular relationship between workers and the technology they interact with at the point of production.
I examine this question at the “strategic” research site (Merton, 1987: 1) of the ride-hailing industry, the largest sector of the on-demand economy that relies on algorithmic management to direct workers. I draw on a seven-year study that includes participant observation (as both a driver and rider), longitudinal interviews (n = 136), social and print media, and official company materials. I first describe how algorithms structure the work process by segmenting work into small chunks of human–algorithm interactions. Algorithmic management scaffolds a choice set from which workers can make hundreds of choices within a short period. I show that at each site of these human–algorithmic interactions, workers make choices in the work process, and through this observation I identify two dominant tactics that workers use. In engagement tactics, drivers interact within the boundaries of the algorithmic management system: they decide when, where, and for which platform company to work, and they usually adhere to the algorithm’s nudges. In deviance tactics, by contrast, drivers manipulate the algorithmic management system, pushing against its boundaries: in this context, they declined certain rides and tried to inflate fares to obtain desired outcomes. Although the ride-hailing companies’ algorithmic managers penalized these actions if detected, drivers could easily counter the penalties. While the behaviors associated with these tactics of engagement and deviance are practical opposites, both tactics elicit consent and contribute to workers describing themselves as skillful agents. Last, I describe how workers can withdraw consent, either by leaving the platform or compromising the integrity of the algorithmic management system.
While “structured antagonism” (Edwards, 1990: 126) between management and workers always exists, so do shared interests—there must be some symbiosis to beget consent. By examining workers’ granular interactions with ride-hailing apps, this study shows that at hundreds of moments at the point of production, individuals could make choices within a highly confined management system—choices that elicited consent even when workers were deviating from some of the formal guidelines. This is what I label choice-based consent. In contrast to prior literature, which emphasizes how social mechanisms elicit consent through broadly sweeping, durable social relations (e.g., Burawoy, 1979; Anteby, 2013; Mears, 2015), this study finds that algorithmic management elicits consent through a series of narrow yet frequent choices that are associated with workers’ sense of mastery. In doing so, this study elaborates and extends the consent literature, identifying new mechanisms and outcomes. Ultimately, I argue that this algorithmically mediated choice-based consent contributes to the growing appeal of the “good bad” job—work that is attractive and meaningful in ways that mask structurally problematic elements.
On-Demand Work, Algorithmic Management, and Workplace Consent
The Rise of On-Demand Work and Its Implications
The number of workers in the on-demand economy is growing, with more than two million jobs added in the U.S. alone in 2020 (Zgola, 2021; Garin et al., 2023), and online on-demand work accounts for up to 12 percent of the global labor force (Datta et al., 2023). 2 Digital platforms such as Uber, Instacart, and Amazon Flex have radically transformed the way work is organized, managed, and experienced, challenging how scholars think about work. 3 For legal scholars, on-demand work blurs the legal classifications of employee and independent contractor, thereby complicating the issue of who is entitled to labor protections (Dubal, 2017, 2020). For economists and strategy scholars, digital platforms create frictionless marketplaces in which rating systems and incentives facilitate trust between strangers (Sundararajan, 2016; Tadelis, 2016; Jacobides, Cennamo, and Gawer, 2018). For researchers of communications and information sciences, on-demand work offers an exemplary setting for examining long-standing issues regarding transparency, fairness, and ethics (Christin, 2017; Möhlmann et al., 2021; Cameron et al., 2023). For labor scholars, on-demand work, especially on closed labor platforms like Lyft and DoorDash, is the quintessential precarious, low-paid job of poor quality (Ravenelle, 2019; Wood et al., 2019) whose decentralized nature makes it challenging for individuals to organize for better working conditions (Wood, Lehdonvirta, and Graham, 2018; Mayberry, Cameron, and Rahman, 2024). For organizational psychologists, the atomized nature of this work complicates their understanding of how social processes such as identification and socialization occur (Petriglieri, Ashford, and Wrzesniewski, 2019; Anicich, 2022; Cropanzano et al., 2023). And for organizational scholars, on-demand organizations represent a significant departure from traditional organizational forms because of the former’s reliance on independent contractors and the replacement of managers by algorithms (Aneesh, 2009; Vallas and Schor, 2020; Davis, 2022). Given the widespread implications of this new form of labor, the need for a granular understanding of on-demand work has never been greater.
Algorithmic Management and Control
In the on-demand workplace, algorithms are de facto managers embedded in the work process and have responsibilities that include hiring, allocating tasks, evaluating performance, and firing. First defined by Lee et al. (2015: 1603), algorithmic management means that software algorithms “assume managerial function and surrounding institutional devices that support algorithms in practice.” Drawing on elements of scientific management (Taylor, 1911), scholars have called algorithmic management a digital version of Taylorism because of the high degree of task standardization and decomposition, digital surveillance, information asymmetry, fine-grained measurement of labor, and piece-rate wages (Cherry, 2016; Faraj, Pachidi, and Sayegh, 2018; Curchod et al., 2020; Jarrahi et al., 2020; Kellogg, Valentine, and Christin, 2020; Rahman, 2021; Noponen et al., 2023). 4 As a result, much of the extant research has focused on how algorithmic management controls workers.
Matching algorithms either assign workers to tasks in closed labor market platforms (e.g., Instacart) or suggest potential workers to customers via open labor market platforms (e.g., Upwork). A worker’s refusal to accept a task can incur a penalty, such as receiving a status downgrade (Rosenblat and Stark, 2016), removal of privileged status (Cameron, Thomason, and Conzon, 2021), or the loss of access to future assignments (Rahman and Valentine, 2021). Algorithmically mediated telemetrics and customer ratings track, rank, and evaluate multiple aspects of workers’ performance (e.g., acceleration, response times, customer ratings; Wood et al., 2019; Cameron, 2022). The management system’s interactive nature enables companies to update the algorithms that manage workers by analyzing data from those same workers, thereby allowing the companies to experiment on workers at will (Möhlmann and Zalmanson, 2017; Veen, Barratt, and Goods, 2020; Rahman, Weiss, and Karunakaran, 2023). This process creates an environment rich in information asymmetry, in which “workers are given information in a piecemeal fashion, while remote company dispatchers maintain omniscient views” (Shapiro, 2018: 2959). This line of reasoning suggests that algorithmic management’s control would reach unprecedented levels, actualizing the notion of the panopticon.
The concept of consent allows scholars to think of workers as using their agency to participate in rather than refuse their work, even in exploitative conditions (Burawoy, 1979, 1985). 5 A few studies have examined how schedule flexibility, a common feature of on-demand work, affords agency (Occhiuto, 2017; Milkman et al., 2021). However, scholars argue that this flexibility is largely mythical because it severely diminishes as one’s economic dependence increases (Schor et al., 2020; Muralidhar, Bossen, and O’Neill, 2022), and most of the work on these platforms is done by people who need the work to survive (Gray and Suri, 2019). Thus, the question of how consent is generated within algorithmic management systems remains unresolved and is crucial for gaining insights into the growing appeal of this work.
Generating Consent in Manufacturing and Service Work Through Social Relations
Emerging from earlier studies on the informal organization and social relations, (Roy, 1952; Roethlisberger and Dickson, 2003; Mayo, 2004), a line of research has examined how consent, or workers’ active participation to align their efforts with managerial objectives, is manufactured through social relations in various white- and blue-collar jobs (e.g., IT professionals, machinists; Hodson, 1999; Barley and Kunda, 2004). Given that on-demand work is an amalgamation of features associated with manufacturing (e.g., piece-rate wages, technologically mediated pacing) and service work (e.g., customer interactions and ratings), the most useful literature in which to situate this study is that which examines how consent is produced in those workplaces. Labor process theorists have found that social relations between workers and other organizational members are key to securing consent. By offering leniency, praise, and friendship, managers can soften the harshness of work conditions while keeping their managerial authority intact. By looking the other way regarding rule-breaking (e.g., allowing smoking at a mine, tolerating craftsmen who use company time and materials to make souvenirs; Gouldner, 1954; Anteby, 2008), managers build relationships with workers on the managers’ own terms while also reinforcing managerial authority. Managers have urged workers to see their jobs as an opportunity to fully be themselves in order to distract their attention away from brutal conditions, such as in heavily surveilled call center work (Fleming and Sturdy, 2009). When supervisors praised aspects of workers’ identity that were stigmatized by broader society, workers were willing to tolerate physically and emotionally demanding conditions. For example, Sallaz (2019) documented how call centers, notorious for their strange hours and abusive customers, were one of the few places where gay men and trans women in the Philippines were praised for their subcultural identity markers (e.g., English fluency, urbaneness). And in VIP lounges, brokers built such strong strategic intimacies by flirting with and offering gifts to attractive young women that these women went out night after night, unpaid and teetering on stilettos, to brokers’ events (Mears, 2015).
Relationships among coworkers are also key in generating consent. Unlike gamification, which relies on rules designed by management to improve workers’ affective experience and to boost productivity (Deterding et al., 2011; Mollick and Rothbard, 2014), workplace games result from organic interactions between workers. In the social game of “making out,” machinists synchronized their efforts with one another to beat the piece rate by an agreed-upon amount (Burawoy, 1979). Similarly, in the “tipping game,” concierges and casino dealers categorized customers, agreeing on how much emotional labor to expend to optimize the tip-to-effort ratio (Sherman, 2007; Sallaz, 2009). These games are socially constructed: workers make the rules, receive feedback from each other, and compete against one another for rewards (Sallaz, 2013). Shared norms can induce workers to work harder than they would under a manager’s gaze. Describing the change from direct managerial oversight to being on a self-managing team, one technical worker reported having dozens of eyes upon him to make sure that he arrived on the line early, ready to work: “Now the whole team is around me and the whole team is observing what I’m doing” (Barker, 1993: 408). On assembly lines, workers shielded their efforts from managers’ eyes so they could ignore safety procedures and speed up production at their coworkers’ behest (Bernstein, 2012).
When consent is manufactured, workers have a sense of psychological and social achievement in that they have done their work well (Sallaz, 2013). Accomplishing the work, either through winning a game or developing group solidarity, elicits pride and mastery. Another byproduct of consent is that it mollifies the inherent conflict between workers and management (Sallaz, 2013) by either attenuating it through the development of friendly relationships with managers (e.g., strategic intimacies; Mears, 2015) or transferring the conflict laterally to coworkers (e.g., peer monitoring; Barker, 1993).
Manufacturing Consent in the On-Demand Economy
This rich literature on the manufacturing of consent is less applicable to on-demand work, which is largely atomized and asocial. Individuals work independently with little to no direct contact with managers and coworkers, and interactions with customers are fleeting and transactional. Thus, the social mechanisms foregrounded in the workplace consent literature are non-existent. And without a cohesive organizational culture and mission statement, normative mechanisms of control, such as shared goals or purposes (e.g., Kunda, 1992), are largely unavailable. Indeed, by being classified as independent contractors, workers are explicitly told by platform organizations that they are not organizational members, as these organizations instead call workers entrepreneurs, co-creators, service providers, consumers, or partners (Ravenelle, 2019).
Moreover, consent is manufactured at the point of production, and on-demand work has a different point of production compared to manufacturing and service work (i.e., worker–machine and customer–worker interactions, respectively; Burawoy, 1979; Leidner, 1993). In on-demand work, the labor process is reconfigured such that the point of production involves an app and its algorithms facilitating connections among multiple actors (e.g., the restaurant, customer, worker, and app in food delivery services; see Figure 1). Hence, irrespective of where exactly the work is done (i.e., digitally by an MTurker or on the road by a Deliveroo courier), the site of the worker–app–customer interaction constitutes a clearly defined system of production at which consent is manufactured, often in the absence of managers and coworkers. Indeed, managerial intent is obscured in this arrangement in which the app—not a supervisor—appears to organize the work process (Chai and Scully, 2019; Vallas and Schor, 2020; Schor, Tirrell, and Vallas, 2023). See Table 1 for an overview of management systems and consent production.

Algorithmic Labor Triangle and On-Demand Work*
Workplace Consent in Manufacturing, Service, and On-Demand Work
Research has not yet examined how algorithmically mediated changes to the labor process may affect consent. Like the internal combustion engine, algorithms are infrastructural technologies that lie at society’s core, as they have the power to “underwrite almost every aspect of daily life” (Barley, 2020: 8). Already, algorithms are transforming how jobs are filled (Schechner, 2017; Feffer, 2023), structured (MacKenzie, 2021; Lebovitz, Lifshitz-Assaf, and Levina, 2022), monitored (Levy and Barocas, 2018; Allison, 2023), and experienced (Bellesia, Mattarelli, and Bertolotti, 2023; Möhlmann, de Lima Salge, and Marabelli, 2023). Understanding how algorithmic management reconfigures work, especially at the human–algorithm interface, is crucial for understanding workers’ experiences and the broader socio-material work ecosystem (Orlikowski and Iacono, 2001; Barley, 2020). But this kind of granular analysis is missing from the literature. Scholars have hinted at how individuals’ ability to make choices and develop workarounds can keep them working under precarious conditions (e.g., Lee et al., 2015, Shapiro, 2018), but the literature lacks a more comprehensive understanding of the relationship between algorithmic management and manufacturing consent. One step toward answering these questions involves closer examination of workers’ moment-by-moment interactions with the technology, to help us understand the choices workers do have and make vis-a-vis the algorithmic management system.
Research Setting, Data Collection, and Analysis
The Ride-Hailing Industry
First launched in 2011, ride-hailing services such as Uber, Lyft, and Juno have disrupted the taxicab industry. The core innovations enabling these services are digital maps and the algorithms that coordinate the work process by matching independent, distributed drivers (working either in their own or rented cars) with customers within seconds and giving block-by-block directions. Fares dynamically adjust based on consumer demand, and performance is evaluated through customers’ ratings and drivers’ acceptance and cancellation rates. Drivers have little direct contact with company employees and few ways to voice concerns; even firing (which is euphemistically called deactivation) is done online. Work requirements vary, with most companies requiring clean driving records, no moving violations in the previous three years, state vehicle inspections, and increasingly, despite industry protests in some cities, criminal background checks. Once hired, which can take from three days to three weeks, workers can go online and begin working. Classified as independent contractors, drivers pay their own employment taxes and are not eligible for health insurance or other social benefits (e.g., workers’ compensation, unemployment insurance). Pay is based on miles driven and bonuses, demand-based incentives, and tips (if any), and it is delivered weekly, although drivers can daily request “instant pay” for a small fee ($1.49 as of December 2023). Some drivers are active on multiple platforms, and algorithmic features are similar across platforms. (See Online Appendix, Image 7 for an example of a worker using two platforms at the same time.) In this article, I use the term “ride-hail(ing)” when discussing the act of driving, I refer to “RideHail” when discussing one or more of the platform companies examined in this study, and I name a specific company (e.g., Uber, Lyft, Juno) only when doing so is necessary to contextualize a comment.
On-demand platforms explicitly promise to provide immediate access to labor for customers. Thus, ride-hailing platforms have two overlapping objectives: to have workers follow algorithmic nudges and to keep workers online and available to work on demand. Nudges encourage workers to make decisions in alignment with fluctuating customer demand, such as offering incentives to work in busy areas. Given that workers are classified as independent contractors, drivers are not contractually obligated to follow nudges, such that even if drivers do not follow a nudge but stay active online, this meets the ride-hailing platforms’ distal second objective of having drivers available to work and offer rides to passengers.
The Work Structure
The work arrangement
RideHail classifies workers as independent contractors, allowing workers to choose their working hours and days with little human oversight, which most drivers in this study liked. Workers I interviewed planned their daily schedules around doctor appointments, preferred sleeping schedules, and child care responsibilities and could take prolonged breaks—for overseas travel or to launch a new business, for example—with no questions asked. For this reason, one participant (Thiel, Philadelphia) left their union job as a carpenter in order to work for RideHail, stating, “I can control the shift. I don’t have a boss.” Being removed from direct human oversight also meant that workers no longer had to work with managers who were sexist, racist, or xenophobic. One driver had lost two previous jobs due to race-based managerial aggression and now described himself as “freed” from the White gaze. Indeed, the retort “I don’t have a boss” was the most frequent response to the question “What do you like about driving for [RideHail]?” In contrast to workers’ prior employment (primarily in retail), on-demand work provided more benefits, allowing individuals to escape bad managers and have greater discretion over their schedules. 6
The work task
The work process of a ride-hailing driver is based on a three-way interaction between the driver, the rider, and the app. Drivers begin a shift by choosing a location to open their app and swiping right to go online. A complete ride consists of (1) the app matching the driver and rider and the driver accepting; (2) the driver getting to the rider’s location and waiting for the rider to enter the vehicle; (3) the driver swiping “start ride” on the app; (4) the driver and rider interacting; (5) the driver dropping the rider off; and (6) the driver swiping “end ride” and rating the rider (see Figure 2). Work cycles may end prematurely, such as when a rider fails to show up or the app malfunctions. After the end of a work cycle, employees may stay online and wait to be matched again or go offline and stop working. In smaller cities, where trips tend to be shorter, drivers can complete as many as six or seven rides in an hour, while in larger locales they might complete only two or three. Fares are set locally and are based on a pickup fee, distance, time, and demand pricing (if any). Drivers may also be offered bonuses for completing a certain number of rides within a designated period, but this is not offered in all markets. Reports of gross hourly wages range from $12 to $30. 7

Ride-Hailing Work Cycle
Data Collection and Analysis
Given the emerging nature of on-demand work and my interest in theory development, I designed a multiple-source qualitative study, having spent seven years in the field (from 2016 to 2023). I used four overlapping data sources, which I triangulated to bolster validity (Eisenhardt, 1989): participant observation (160 total hours of driving-related activity), conversational interviews (n = 112), semi-structured interviews in 23 North American cities (n = 136), and social and print media. 8
Participant observation
To understand how algorithms were deployed and experienced, I participated in the ride-hailing industry as both a driver and a rider. From 2016 to 2019, I was a driver in a major U.S. city, using both my personal car and a rental car, the latter obtained through a platform-sponsored program. I earned income from one RideHail company. I varied my driving times and routes to widen my range of experiences: I drove the weekday morning commute and the evening bar shift, timed my airport runs with international flight arrivals, visited higher- and lower-income neighborhoods, and worked on major holidays, including twice on New Year’s Eve (the busiest day of the year). I also conducted mini experiments on myself. On some days I would try to maximize my income by chasing demand-based incentives and bonuses, while at other times I purposely ignored such incentives and did not check my earnings until the day’s end. Sometimes I manipulated the app, trying to confine my trips to a certain area, while at other times I let the app “drive” to see where I ended up. To gain perspective on drivers’ experiences in different areas, I enlisted a research assistant to drive for the same company in another U.S. city. Our ethnographic notes included reflections on busyness, ratings, incentives, interactions with in-app driver support systems, accidents, car care, and road conditions. I also attended classes, organized by a local group, on defensive driving and on my legal rights as a driver. During the same period, as a rider I kept notes on nearly all rides taken (n = 112) with individuals working for several ride-hailing companies. Some of these rides were personal, and some were specifically for this study. For example, I spent afternoons taking rides around a new city. Logs included information about how I hailed the ride, the car’s condition, app malfunctions, and overall impressions of the ride, including my rating.
Semi-structured interviews
In my first round of data collection starting in 2016, I conducted 63 semi-structured interviews with drivers across North America. 9 The protocol began with grand-tour questions: “Tell me about driving” and “What’s a good day driving?” In the second round of interviews (n = 44), which were conducted 18 to 24 months after the initial interviews, I asked workers to describe any changes since our last interview (e.g., schedule changes or app updates) and their current financial situations, and I followed up on themes that had not been addressed in the first interview. 10 In the third round (n = 29), I focused on how drivers’ work lives were changing, in particular regarding the COVID-19 pandemic, and how drivers navigated and solved problems (e.g., with customers or the apps). Most interview data in this article are from the first and second rounds of data collection. All interviews except one were conducted in English, and all interviews except ten were professionally transcribed. 11 In total, I conducted 136 interviews with 63 drivers, of whom 19 (30 percent) were female. Fifty (79 percent) reported driving as their primary source of income, and all except one reported driving to meet essential household expenses such as utility bills. Twenty-four drivers (38 percent) were active on at least two apps, though not all participants lived in cities in which multiple ride-hailing companies offered services. The amount of time employed in the industry (at Round 1) ranged from two weeks (ten rides) to seven years (18,000 rides), with the average driver having approximately 14 months of experience, 1,800 trips completed, and a 4.87/5.0 rating. 12 My sampling procedure yielded participants who were more committed than most drivers to the ride-hailing industry, reflected by a longer tenure than the three- to six-month industry average (Campbell, 2018; Hall and Krueger, 2018).
I used several sampling approaches to ensure maximum variation and participants’ anonymity, as news media have reported that riders and drivers have been blocked by Uber after having made grievances public (Isaac, 2019). I met roughly half my informants through ride-hailing—either as part of my everyday activities or through expeditions to an unfamiliar area. To increase participants’ anonymity, I often hailed rides from family members’ or friends’ phones. I also recruited informants by advertising (e.g., parking lots, web forums) and via convenience and snowball sampling. 13 As the majority of ride-hailing drivers were male and/or people of color who worked 30 or more hours per week (Campbell, 2018), I tried to interview drivers who were female, White, or worked part time, to gain minority perspectives. Interviews ranged from 35 minutes to 2.5 hours, with an average length of 65 minutes. Lastly, I collected data across multiple cities because ride-hailing and new features were introduced at various times and places. For example, shared rides, a service that matches drivers with multiple riders traveling in the same direction, were first introduced in 2015 and available only in larger cities. Interviews included drivers from areas where the industry was well established (e.g., Philadelphia), nascent (e.g., Missoula), banned (e.g., Austin), and facing pressure from unions (e.g., New York City).
Archival and social media
Unlike labor process theorists before me, I did not confine my field site to a fixed location, and I included blogs, discussion boards, YouTube videos, online articles, and company materials. 14 These documents served as useful support or allowed triangulation (Shah and Corley, 2006). I created several analytical aids from these materials, such as an industry time line and a look book of interviews with industry leaders. This unobtrusive form of data collection (Webb and Weick, 1979) provided important information about the social, legal, and political challenges in the industry as well as additional perspectives on drivers’ experiences.
Data Analysis
I analyzed data by using a grounded theory approach (Locke, 2001; Charmaz, 2006; Glaser and Strauss, 2017) with field observations, interviews, and web forum posts as primary data sources.
Stage 1: Open and focused coding
Data in the first wave were collected in four two-month rounds, with each round followed by two months of preliminary analysis. After my first round of data collection, I refined my research questions, interview schedule, and the structure of my notes. For example, I noticed that drivers often discussed pricing incentives and set earnings goals and that these themes also appeared in my field notes. Thus, in subsequent interviews, I probed to understand how drivers’ activities were shaped by incentives (e.g., ignoring overly challenging bonuses), and I paid attention to my own behavior as a driver in response to incentives. Toward the end of the first round of data collection, I began focused analysis. While my earlier analysis had informed my thinking, I put these observations aside so that I could see my data with fresh eyes as I began coding and, as Charmaz (2006: 45) termed it, “generating the bones of analysis.” Interviews and field notes were coded over five rounds. I began by open coding one-fifth of my data, which I selected for maximum variation of gender, hours worked, length of time driving, and location. Three major themes emerged: general (dis)like of work, pricing incentives, and ratings. In the next two rounds, I focused on these themes while also continuing open coding, through which two additional themes emerged: interactions with the apps (i.e., accepting a ride request) and strategies to create a good day. No new themes emerged in the last two coding rounds. I followed a similar process in the second and third waves of data collection. Throughout the process, I wrote memos, discussed ideas with colleagues, and workshopped early findings.
Stage 2: Axial coding
In the next stage, I began axial coding and iterating between the data and existing theory to build “a dense texture of relationships” (Charmaz, 2006: 60) around concepts. Based on my experience driving, I knew there was a rhythm to the work, so I began by constructing a temporal model of a routine day and a routine ride. Depending on when I worked, I drove the same streets and had identical conversations with similar people (e.g., driving professionals in silence on North Capitol Street on weekday mornings, having conversations with drunk partygoers on Wisconsin Avenue on weekend evenings). Tasks were repetitive in that each ride required the same interactions with the app I used as a driver (e.g., accepting a ride, rating the rider). Using stacks of index cards as a visual model, I laid out the five thematic codes on top of the routine day and routine ride models and, as I had more data for the latter, focused my analysis there (see Figure 2).
Prior literature has described algorithms as discrete units (Orlikowski and Scott, 2014; Curchod et al., 2020), but my informants and I experienced algorithms as distinct but interlocking parts of a larger management system. Declining a ride, for example, affected ratings, which could then affect the following week’s bonus. Building on this insight, I recoded my data around each mention of an algorithmic function and identified five functions: matching, work instructions, demand-based pricing, bonus pricing, and ratings. 15 Further coding clarified the following for each algorithmic function: its purpose, how it was communicated to drivers, how it was linked to other algorithms, and whether it influenced drivers through rewards or sanctions.
With a clear sense of the temporal work cycle and the function of each segment of the algorithmic management system, I asked myself, “What is the platform company’s algorithmic management system encouraging drivers to do? How are drivers responding to this encouragement?” In re-reading the interview transcripts, I found that one of my informants had already answered these questions, describing two goals, one organizational and one personal, that were sometimes in conflict and sometimes aligned: “The company is trying to make drivers take as many rides as possible. The driver is trying to make as much money as possible” (Arthur, L.A.). With this response, everything clicked. The ride-hailing platforms structured the work process to encourage drivers to complete the maximum number of rides possible. The tasks to move along the work process were segmented by the algorithms and were thus simple and discrete, designed so that workers would complete each task and quickly move along to the next. These tasks can be viewed as moments of choice that were built into the work process. And RideHail regularly emphasized choice, as schedule flexibility, as a benefit of the work. Advertisements, for example, highlighted how workers drove their own cars and set their own schedules (see Image 1 in the Online Appendix). Recognizing that moments of choice were also embedded in the human–algorithm interactions, I recoded my data accordingly, noting the ways in which workers talked about choice in relation to the algorithms. I identified moments of choice embedded within the two tactics that I named “engagement” and “deviance”—the first rarely resulting in sanctions and the latter often so. In engagement tactics, workers largely followed the algorithms’ nudges and stayed within the boundaries of the management system, while in deviance tactics, workers manipulated their input into the algorithmic management system. Conversations with other scholars encouraged me to consider more cases in which the drivers used the output of the algorithms to make informed choices about how to interact with the work arrangement itself, i.e., when, where, and for how long to work. Because these choices were within the boundaries of the management system and not sanctioned, I folded these data into the engagement tactic.
Stage 3: Theoretical coding
In the final round of analysis, theoretical coding, I developed relationships between categories elicited in earlier stages to “weave the fractured story back together” (Charmaz, 2006: 63). Starting with the observation that workers had different choices when interacting with algorithms, I returned to the literature to help me theorize from the data. I found literature in the sociology of work to be fruitful, as it extensively documents how consent, the ways in which workers enthusiastically align their efforts to meet or exceed managerial objectives, is generated in many jobs. After identifying key features associated with consent (site of production, mechanisms, outcomes), I then reanalyzed my data, paying particular attention to how the existing theory did and did not match my data. Moving between analyzing data, drawing models, and writing memos, I abstracted from these categories and relationships to identify how consent is produced through a series of constant yet confined choices. I also identified the underlying mechanisms and associated outcomes of consent, ultimately explaining that how consent is generated in algorithmic management systems differs from how it is produced in other settings, as well as showing how it affects workers’ engagement.
Findings
First, I describe the five functions that compose the algorithmic management system by assigning, evaluating, and pricing work tasks in order to parcel work into small segments. I then show how these constant yet confined choices elicit consent through seemingly oppositional tactics: engagement and deviance. I conclude by describing how workers could withdraw consent by either logging off with no intention of returning or by damaging the integrity of the algorithmic management system.
Algorithmic Management in the Ride-Hailing Industry
Algorithms scaffold the work process
To understand workers’ granular experiences of work, it is first necessary to understand how the work was structured. Unlike manufacturing and service work, in which human management helps to scaffold work, the algorithmic management system set up the work process. In the ride-hailing industry, algorithms scaffolded the work process by (1) matching workers to riders; (2) instructing workers how to proceed through the work tasks (e.g., giving directions); (3) setting and adjusting fares dynamically (demand-based pricing); (4) offering bonuses (e.g., “Do 50 rides in the next 5 days for an extra $50”); and (5) evaluating workers’ performance via customer ratings. See Table 2 for descriptions of all five components. (Table 2A in the Online Appendix shares the same information and includes accompanying images as examples of each algorithm.) Algorithmic management systems thus used rewards, sanctions, and specific directions to manage drivers. Ultimately, the algorithmic management system was the defining feature of the work; for example, in a typical shift a driver may have completed only a dozen rides but would have more than 100 unique interactions with the algorithms.
Components of the Algorithmic Management System That Scaffold the Work Process
Rewards
The pricing algorithms, which were demand-based incentives and bonuses, rewarded drivers if they coordinated their schedules in response to demand. Demand-based incentives could be predictable (e.g., rush hours) or sporadic based on local events. Texts and in-app messages alerted drivers regarding fare increases in busy areas: “Demand is higher than usual in Center City. Take advantage of higher-than-normal fares!”“1.2–1.8x boost—4.30PM–7PM in downtown DC!” and “Adele is playing at [venue] tonight! The streets will be full of people!!” (see Image 2 in the Online Appendix). In addition, heat maps that indicated high-demand areas popped up when drivers first opened an app and at the end of every ride (see Image 3 in the Online Appendix). Checking heat maps became a routine part of the workday, as drivers “turn[ed] on [the] app and [looked for] that very orange, bright color” (Leah, Chicago) and rushed to “try to go where the heat maps are surging” (Mercy, San Francisco).
Weekly bonuses offered extra money for completing an algorithmically determined number of rides. How the algorithmic management system determined bonuses was proprietary company knowledge and, at times, seemingly arbitrary; for example, a driver who met the quota one week might receive an easier or harder quota the following week. Bonuses and quotas were continuously communicated to drivers through texts and in-app alerts—while taking a mid-semester hiatus from the work, I still received biweekly texts (see Image 4 in the Online Appendix). Some bonuses encouraged commitment to a specific platform company. After driving with a competitor, George (Boston) received an enticing bonus offer from his former ride-hailing company: “They will pay me up to $500 on top of the money that I make for [the competitor] . . . and if I stay without logging out for one hour, they will pay me $40. [Laughs] That’s how they just got me back easily.” Other bonuses were predicated on workers having completed a certain number of consecutive rides without logging off, further encouraging workers to commit to a specific company (see Image 5 in the Online Appendix).
Sanctions
The matching and evaluation algorithms sanctioned workers who did not comply with regulations. When the algorithmic management system completed a match, the driver’s phone buzzed and displayed the distance to the pickup spot, rider’s details (including rating), and demand-based incentive amount (if any). Drivers had up to 15 seconds to accept, and they were sanctioned if they did not accept a certain percentage of rides. “We use acceptance rates to determine driver eligibility for certain incentives and help keep rider wait times short,” Uber’s guidelines noted. Repercussions for canceling rides included warnings, temporary suspensions, and permanent deactivation; consequently, most drivers reported accepting all rides (see Images 6 and 7 in the Online Appendix). When asked if he declined any rides, Jay (D.C.) replied, “Not really, because like I mentioned, I’m out here to work.” Laughing, Leah (Chicago) said, “You can decline—[but] how are you going to decline?” and reported that she had accepted every ride request in the past year.
High customer ratings were required for continued platform access, and drivers were punished if their evaluations fell below a certain number. In short, riders could fire drivers. 16 While the rate of driver deactivation due to ratings was low, the threat of poor ratings loomed large and clearly shaped drivers’ behaviors. 17 Drivers created signs that hung over back seats reminding riders that a rating of less than five stars was harmful (see Images 8 and 9 in the Online Appendix). At the end of one ride, a driver cheerily said to me, “You’ve been a five-star customer! I hope you rate me the same!” while making sure I was watching him rate me (Field notes, March 19, 2017). Drivers with high ratings received positive reinforcement, such as compliments and non-monetary badges (e.g., good conversationalist; see Images 10 and 11 in the Online Appendix). When customers’ ratings dropped, drivers received warnings, with suggestions about how to improve or a requirement to attend a remedial course at the driver’s expense. Others could be deactivated with limited information provided by RideHail and no means of recourse, which in cities with only one ride-hailing platform was the equivalent of an industry blackball (see Images 12 and 13 in the Online Appendix).
Directing
Lastly, the work-instruction algorithms directed workers through the navigation and pacing of the work. After a driver accepted a ride, these algorithms provided routes, created queues, and set timers (see Image 14 in the Online Appendix). Instructional materials described the process: “After you accept a request, tap ‘Navigate.’ The app guide[s] you to the rider. The rider will see your car approaching on their app and your ETA. When you’re close, we’ll send them a text.” Once the driver arrived, a timer appeared, dictating how long drivers must wait (60 seconds to 10 minutes based on ride type) before drivers could select “no show” and receive a new ride (see Image 15 in the Online Appendix). Navigation instructions were critical in shared rides, in which multiple riders with different destinations shared the same vehicle. Instructions urged drivers to “Always follow the app’s instructions. The route is built on efficiency, so the order of who is picked up and dropped off first varies from ride to ride” (see Image 16 in the Online Appendix). In addition, algorithms matched drivers to their rides and sent various messages and reminders. For example, drivers were placed in virtual queues in high-traffic areas, such as the airport (see Image 17 in the Online Appendix), and the queues dictated how long they must wait for a new ride. The algorithms sent nudges at opportune times to encourage drivers to stick with it. Texts appeared when drivers had taken a break (“You haven’t driven in three days. Go out there and make some money!”), and pop-ups nudged drivers as they logged off (“You’ve only driven 8 hours today! Only $18 to go until you meet yesterday’s pay-out”). Telemetrics monitored speed, acceleration, and deceleration, offering encouragement such as “Good job keeping your braking smooth!” and warnings when drivers did not meet the telemetrics standards (see Images 18 and 19, Online Appendix). Through rewarding, sanctioning, and directing work activities, the five components of the algorithmic management system structured and segmented the work such that drivers were able to make a series of constant yet confined choices when conducting their work, and I found these choices to be the building blocks of manufacturing consent.
Manufacturing Consent: Engagement and Deviance Tactics Within Algorithmic Management Systems
Although most research emphasizes the all-encompassing control of algorithmic management systems, these systems do not necessarily take away workers’ ability to exercise choice—indeed, because algorithms segmented the work into discrete tasks, there were frequent opportunities for workers to exercise choice, albeit over narrow segments of work. Regardless of whether workers used engagement or deviance tactics, at hundreds of points in every work day, workers made choices within a highly constraining system—choices that elicited their consent to a platform’s algorithmic management system, even when they skirted some of the platform’s formal guidelines. While the tactics were practical opposites in terms of how workers interacted with the management system, they both elicited consent and were associated with workers’ finding meaning and seeing themselves as skillful agents; however, the tactics differed in terms of their potential for worker–management conflict.
Engagement Tactics: Playing Within the Boundaries of the Algorithmic Management System
When deploying engagement tactics, drivers worked in conjunction with the algorithmic management system, using information from the apps to make decisions about how to navigate within the boundaries of the system. In simple engagement tactics, workers followed the algorithmic nudges, such as driving toward high-demand areas. In complex engagement tactics, workers did not follow the algorithmic nudges; instead, they used information provided by the nudges to inform their navigation of the work. Below, I describe workers’ engagement with the pricing and matching algorithms.
Simple engagement tactics
Chasing demand-based incentives
The algorithms that calculated demand-based incentives were some of the most visible algorithms in the management system and were directly linked to workers’ income. As soon as workers opened an app, and when they completed a ride or logged off—or even when not working—they received notifications about nearby increases in demand. Arthur (L.A.) described the importance of demand-based incentives
18
: [Demand-based incentives] play a critical role in whether or not you’re going to go out, because you’re dealing with economics—you don’t want to waste your time or your gas. I use it to my advantage. When it comes into your app, you want to be in those areas. [The company] pays flat rates and obviously the flat rate is not as attractive as surges.
Demand-based incentives could make the difference between breaking even and making a profit, especially as base mileage pay declined over the period of this study. 19 To chase a demand-based incentive, drivers checked heat maps, text messages, or in-app notifications before driving. George (Boston) monitored the heat map before signing on: “I’m just waiting. When the surge price is starting to go up, that’s when I put the system on. I know where to be, when, what time—every single day.” Drivers on ride-hailing forums offered complex suggestions to monitor demand pricing, such as installing software to take automated screenshots over multiple days at different points in time to be better able to predict trends. Another driver on the forums described using “one phone to drive for [one company], and the other phone to zoom out to your entire market to watch the surge areas.”
Accepting all rides
Some drivers believed the best way to work in the system was to choose to accept all rides. Ignoring information like a customer’s ratings, Ryan (Detroit) said, “I want to work. I’m not here to see what you’re rated.” Similarly, Curtis (D.C.) said, “The system is: do the work or you don’t. There’s no in-between. You got to be out here.” Drivers described cranking up the music, often rap or electronic music, so they could “have fun, cruise, and just go with the flow” (Tila, Boston). Rhythmic music soothes the nervous system, lowering brain activity (Thaut, 2013), and facilitates entering into a state of absolute absorption (Csikzentmihalyi, 1990). After listening to one driver’s music bounce between three distinct beat-heavy sounds—Afrobeats, Czech house, and Bamana—I asked how he put together his playlist. He replied, “I play what I want—this cat here [an Afrobeats artist] puts me in the zone” (Field notes, July 26, 2018).
In Leland’s car (Philadelphia), we started talking about Lindsey Stirling, a dubstep artist, who was playing via his phone. In a follow-up interview, I asked about his typical day: [Every day is] pretty [much] just like every other day. . . . I basically turn on the app, hoping I get a ping soon . . . I don’t really prepare in any sort of way . . . I can stay focused. I’m always pretty good when I’m driving. If there’s someone in the car, I guess something just kicks in, that I am able to stay focused and not slack off. . . . It’s weird eating because you want to stay online and keep earning, at the same time you’re hungry. I usually just wait until afterwards . . . I’m too focused. . . . Wait, what was the original question again?
Accepting every ride, Leland got so absorbed in driving that he avoided stopping to eat or even to use the bathroom—indeed, while responding to the question, he re-entered the same flow state he was in while driving, losing his train of thought. The combination of accepting all rides and beat-filled music helped drivers “stay focused” (Kristin, New Haven) so that “time goes fast” (Thiel, Philadelphia), so much so that at the end of the day they were surprised at how much they had earned (Innocent, Detroit). Subsequent software updates made it even easier for drivers to choose to accept all rides and stay in the flow state; they could just click the “Accept All Rides” button so they did not need to swipe “yes” for each ride.
Complex engagement tactics
Ignoring and avoiding demand-based incentives
In complex engagement tactics, drivers did not necessarily follow algorithms’ nudges but did stay within the boundaries of the management system, allowing them to see themselves as skillful agents. Not all drivers chased demand-based incentives, for example, and there were no rules stating that drivers must do so. Indeed, Tamara (Denver) ignored them: “I don’t pay much attention to the surge. I start up where I start off and just go wherever I get my ping. It’s just not worth it—too much thinking.” Due to their real-time nature, demand-based incentives changed frequently or even disappeared, leaving drivers with a bitter taste. Kimbo (Seattle) said, [Demand-based incentives are] a little annoying because it’s misleading or by the time you get there it’s gone. They’ll send out a text that says, “Adele is playing tonight, the streets will be filled with people.” But there’s going to be a lot of traffic and I’m going to be driving one person. They just exaggerate—that’s a better word—you’ll make crazy cash this weekend. It’s like they’re insulting your intelligence. [Laughs]
As following demand-based incentives did not necessarily lead to more money, some drivers tempered their expectations and/or carefully chose which incentives alerts might be profitable. Others went a step further, examining the demand-based-incentives algorithmically generated heat map and then driving in the opposite direction. Eric (D.C.) said, I don’t hang around where the surge is because there are a lot of cars around. You go a little farther to get more passengers than to go where the surge is and get less passengers. For example, get one passenger in 30 minutes and make $20, but go outside the surge and get four passengers and make $50.
By monitoring demand-based notifications and making estimates to determine whether a ride was worth their time, drivers exercised choice. Though drivers might respond in seemingly contradictory ways—George turned on his app only when he saw a demand-based incentive, while Eric drove in the opposite direction—each driver was monitoring the demand-based-incentive algorithm, evaluating its information, and responding in a way that they believed would maximize their earnings. While drivers may be rewarded for chasing a demand-based incentive to a busy area, drivers were not penalized for ignoring or avoiding these areas.
Optimizing pickup locations
Classified as independent contractors, drivers had the flexibility to determine when and where to work, and the algorithmic management system influenced how drivers enacted this flexibility. Leland (Philadelphia) tried to get the matching algorithm “working” in his favor by turning on the app two minutes before getting to his preferred spot because he believed that “the algorithm factors in how much time you’ve been waiting. The longer I’m online, the more chances I have to get a ping [ride].” Other drivers monitored the rider app, sometimes purchasing a second phone expressly for this purpose, to scope out other cars and position themselves advantageously. Amber (Missoula) said, You open the passenger app and it will show you the eight closest drivers and then you just go where they’re not. You could count all the other drivers on any given moment. It would be the 200 block of Ryman, but there were too many cars—it was too stressful . . . it wasn’t always a sure thing. So, I would go four blocks away where there were less people but more bars. And it would work out for me.
Considering the dynamic pricing algorithm, some drivers turned off an app when finishing a ride in a less desirable location and then turned it on again at a more desirable one (e.g., near the airport) to maximize their earnings. Two drivers described the different ways in which the pricing algorithm influenced how long they worked in a college town: I told my buddy this. Sometimes you got to turn on the app and go wherever the riders take you. If you go to Ypsilanti, leave it on. If the Ypsilanti person takes you to Detroit, leave it on. If the Detroit person takes you to Royal Oak, leave it on. Do that once and see where you go and how much you make. You’ll see that it doesn’t pay . . . staying in Ann Arbor is the best thing you can do as opposed to going outside. (David, Ann Arbor) I get out of Ann Arbor as fast as I can. They charge about 20 cents more a mile. . . . They charge more of a base rate . . . but your rides are much shorter. And if you’re going two miles in Ann Arbor, it could take you 10 to 12 minutes to get there . . . you can’t make any money. Between the hills and people walking, you can’t get anywhere fast. I’d rather take someone 20 miles on a highway, get there in 20, and make some money, than take someone three miles around Ann Arbor making an extra 20 cents a mile. (Boxton, Detroit)
Drivers’ ability to make different choices about when and where to work yet remain within the boundaries of the management system shows how the algorithmic management system allowed workers to make confined choices that could also reinforce their image of themselves as skillful workers.
Outcomes of Consent Produced by Engagement Tactics: Skill Development and the Eradication of Managerial Conflict
When deploying engagement tactics, drivers found meaning in the work by seeing themselves as skillful agents, deftly navigating the algorithmic management system and having a sense of mastery of their work. Drivers boasted that they had “special skills” (Jorge, D.C.), “knew how to make the money” (George, Boston), and could make things “work out” (Amber, Missoula) “to their advantage” (Arthur, LA). They worked smart (Zach, NYC), only in the times and places “worth” (Boxton, Detroit) their time. Emphasizing the skills and meaning she found in driving, Elizabeth (Detroit) said, “Oddly, it is fulfilling to me. I mean I’ve heard people describe it as a job that you can do if you have little skills or something like that . . . but I don’t think that’s the case for most, I really don’t.”
Another outcome associated with engagement tactics was the absence of conflict with the algorithmic manager. One reason workers did not see themselves having any conflict with the algorithmic management system is that their behaviors were aligned with the nudges, which, in turn, were often associated with various incentives. Often, drivers repeated that the best thing to do was to follow the system, which often worked in their favor. Thrilled at the money he earned by following the nudges, Marshall (Detroit) said, “I made $35 off some perks [incentives] and I did it by doing what I’m supposed to so . . . everything else will work itself out.” Drivers reported no real conflict with the algorithmic management system. Sebastian (Detroit) said, “Sometimes you do a back and forth with them about a cancellation fee or a fare adjustment, but that’s [about it].” Overall, drivers described feeling satisfied, noting that the work was “reasonable” (Betrand, D.C.), and said things such as, “I like it. I can’t complain at all. And honestly, I don’t think anyone can” (Maxwell, D.C.). Confident in their skills navigating the algorithmic management systems, workers said they could “drive forever” (Franklin, Detroit).
Each interaction with the pricing algorithms and the work-matching algorithms thus presented these drivers with an opportunity to exercise choice, such as by analyzing a heat map before deciding to drive toward (simple engagement) or away from (complex engagement) a high-demand area. Workers’ behaviors in simple engagement tactics were in full alignment with the algorithmic nudges and exemplified how nudges were used to encourage individuals to make a specific choice. Drivers exercised greater latitude of choice in complex engagement tactics because they did not explicitly follow the nudges but, instead, used the information presented by the system to inform their behaviors. While behaviors such as drivers manipulating their input to influence the algorithmic management system or using two phones to optimize their location were not prohibited, they were nonetheless not in accordance with RideHail’s provided norms and suggested practices.
Deviance Tactics: Pushing Against the Boundaries of the Algorithmic Management System
When deploying deviance tactics, drivers tried to get around the algorithmic management system’s directives by manipulating their input into the work-matching, pricing, and ratings algorithms and, if penalized, to counter any sanctions. In simple deviance tactics, drivers circumvented the blind matching algorithms by selecting and screening rides. 20 In complex deviance tactics, drivers tried to influence their position within the algorithmic management system either by inflating surge prices or by increasing their acceptance rates or customers’ ratings. If the system detected either of these tactics, it sanctioned drivers; however, because the penalties were applied in a consistent manner, drivers could counter them. In this way, deviance tactics elicited consent because workers were able to make choices, and these choices and their counters to sanctions stayed within the boundaries of the algorithmic management system, so ultimately, drivers remained online, which was one of RideHail’s objectives.
Simple deviance tactics
Pre-selecting riders
In simple deviance tactics, drivers attempted to circumvent the blind matching algorithms by being matched with a preferred rider. I describe the three tactics in this section in order of how frequently drivers might deploy each one, from most to least frequent. Trying to make the matching algorithms assign the driver to someone already in the car was a tactic that drivers initially reported frequently, though it became less common later in my data collection. When I began data collection in 2016, drivers in many cities recounted successfully using this tactic. My field notes describe an incident in D.C.: Marsha seemed unfazed as we navigated bumper-to-bumper rush hour traffic, but I was nauseated [by the stop and go] and frustrated with all the stops and complained after we dropped off the frat boys. “Wanna switch to a private ride?” she asked. “Sure, as long as it doesn’t hurt you.”“Naw, it’s cool.” She ended my ride and went offline. As soon as she went back online, she told me to request a new ride and—ta- da!—we were matched. As we continued to Ballston, she confessed she didn’t like shared rides as she’ll drive for half an hour and realize she’s only made $5. (Field notes, August 30, 2016)
Though against the rules, switching to a private ride was in Marsha’s interest as it paid more and pleased me, the rider. In late 2015, when ride-hailing first launched in the city where I lived, I became friendly with an informant, whom I first met via the app and who soon became my designated airport driver. I texted him whenever I needed a ride. Once in his car, I would request a ride and we would be matched immediately. I appreciated the convenience, and he appreciated the guaranteed fare. By early 2017, it took multiple attempts to be matched even though we were in the same car, and I often resorted to paying in cash. In mid 2017, I was back to using the RideHail app in its intended fashion for airport rides; around the same time across the country, drivers reported similar events. In late 2017, Ashton (Detroit) reported, [A friend asked] “I’ve got to go to the airport at 4:30 p.m. on Friday, can you take me?” Sure. It used to be that you would get in the car, and they’d request a ride, and it automatically goes to you. It’s changed dramatically. It takes maybe three or four times that [the rider] has to request a ride, cancel it, request a ride, cancel it. It gets to me.
In Ashton’s case, exercising the choice to give his friend a ride came with frustration, as he had to direct his friend to request the ride multiple times before being successfully matched. However, because Ashton directed the rider to initiate the requests rather than doing it himself, he avoided detection and potential sanctions by the algorithmic management system and was matched with his preselected rider.
Screening rides
Choosing whether to accept a ride after being assigned by the matching algorithm was another simple deviance tactic. Drivers screened rides by distance, surge amount, and customer rating or location. Arthur (L.A.) said, “I only accept rides five minutes [away] and below—especially if it’s not a surge. . . . It’s going to cost me more time and gas to get there for a three-minute ride. That is totally not worth it . . . [that’s how] I make my money.” Other drivers declined potentially problematic rides based on race and class cues. Orion (Detroit) said, I screen out riders by their name, rating, and location. [Some] people are more problematic than average—it’s not worth it. They just harass me. I’m not traveling from 10–15 minutes away to take them to a liquor store and wait for them with their kids.
Doing a cost–benefit analysis before traveling to pick up a rider, drivers used the output of the matching algorithms to calculate potential earnings versus time spent. Rejecting distant pickups, for example, allowed drivers to be available for shorter, more profitable rides.
A “Get Rides to Destination” feature allowed drivers to ask the matching algorithm to pair them with rides going in a specific direction. On discussion boards, drivers reported putting in distant destinations, hundreds of miles away, to get longer rides. Mercy (San Francisco) used it to pick up riders while commuting to her primary job. Like many drivers, I entered my home address toward the end of a shift to get another fare before stopping for the day.
I needed to be home by 3:15, so I put on “Get Rides to a Destination.” Ping—[shared ride], too much hassle, plus it’s 8 minutes away. Decline! Ping. Another [shared ride]—Decline! Can these people not give me a good ride? [Shared] rides take too long and there’s no way I’ll get home in time. Another decline and I’ll be blocked [30 seconds] from the app. Whatever. Ping. Oh good, it’s a [private] ride. Accept! I drop the lady off only a few blocks away from my house and I’m home at 3:08! (Field notes, July 29, 2018)
In this illustration, multiple choices were in play; I chose to use the destination feature and to accept only a private ride, which I thought was more likely to get me home on time. Given that I declined only two rides, I surmised that the possibility of sanctions would be low and minimal if applied (i.e., a 30-second block) as drivers were able to decline a small percentage of rides without facing serious penalties. Drivers’ ability to make these multiple, though narrow, choices could contribute to workers seeing themselves as skillful—I certainty saw myself as such when I arrived home at the desired time. Thus, for the most part, drivers could choose rides more desirable to them in hopes of maximizing their earnings and faced few to no penalties, as the choices were within the boundaries of the management system.
Blanket ride rejection
In blanket ride rejections, drivers rejected all rides of a specific type. Shared rides, whereby the algorithmic management system coordinated multiple riders traveling in the same direction, were introduced in late 2015. Shared rides were offered at a lower cost, which made the service more attractive to new customers; however, most drivers, including myself, disliked shared rides due to the low pay, circuitous routes, and querulous riders frustrated by longer travel times, and many drivers systematically rejected these rides. Yet, even though they were “too damn cheap,” drivers knew that “you gonna have to start accepting them or you gonna get blocked out of the system” (Tanya, D.C.). One driver described rejecting these rides: In the beginning they used to say [shared rides] were optional, but after a month they said it was mandatory and were sending messages saying they were going to cut me off if I kept [rejecting them]. I wrote two messages saying all the reasons I don’t like doing it. If I force straight decline the ride the app turns off, but I can take a break and turn it back on. I really don’t like having five people in the car. (Thiel, Philadelphia)
In late 2017, new incentives linked bonuses to ride quotas, making shared rides more attractive. Leland (Philadelphia) found the new bonuses motivating: “I love incentives, and [shared rides] helps when I’m trying to get multiple rides. If I’m trying to get 60 rides and I get a [shared ride], it takes 40 minutes, but I get three rides. It’s a great deal and I don’t have to worry about driving around looking for another ride.” With the new incentive, the deviance tactic of blanket ride rejection was transformed into the engagement tactic of accepting all rides, as workers’ behaviors then aligned with RideHail’s goal of real-time coordination of supply and demand. As the algorithmic management system changed the rules, deviant tactics were curtailed, and workers adapted in a way that met their income goals and aligned with RideHail’s more-proximate goals.
Complex deviance tactics
Inflating ratings
In complex deviance tactics, drivers manipulated their input to change their positions within the algorithmic management system, and if detected, drivers could be sanctioned. Ratings, namely customer and acceptance ratings, played an important role in earnings, as they determined drivers’ access to the platforms and their eligibility for special incentives. To inflate their acceptance rates, drivers might accept a ride they did not plan to complete by forcing customers to cancel a ride or by canceling it themselves. On a discussion board, one driver, who was looking to receive an incentive but whose acceptance rate was too low, described how they did exactly that: “Because I had accepted the request, my acceptance rate now jumped above the 80% threshold and unlocked the $80 bonus for me. BOOM! . . . I got my bonus and didn’t even have to complete an additional trip.” Smith (D.C.) described an encounter in which he accepted a ride but forced the customer to cancel to maintain his perfect rating: I had one time where I didn’t want to hurt my 100 percent [acceptance rating]—if you let that [timer] time out—so I accepted [the ride request] and I just wanted the person to cancel. It was the night of when Donald Trump was inaugurated, and it was downtown, and you couldn’t really get down there. So, people kept requesting me and I didn’t want to keep canceling. Every single time someone gave me a request, I hit it, but then I would just take a whole bunch of time and they eventually would cancel.
21
Drivers strove to keep their acceptance and cancellation ratings within certain thresholds to avoid penalties or the removal of rewards.
Drivers often manipulated the algorithms to inflate ratings, such as by canceling a ride prematurely to circumvent a customer’s rating. George (Boston) said, “If I make a mistake, I cancel the trip and give it to you for free, doesn’t matter where you are going. [That way] people don’t really have the chance to give me any bad rating.” Likewise, Smith (D.C.) described, I had just turned 4.93 and I picked up a woman and her kid. The app took me to the back [of the pickup location], but she was in the front and it was really cold. She was holding the kid in her hands and she was pissed off. I said, “Ma’am, I’m just going where the app sent me.” She ran in the back, got in the car, destroyed me on the ratings, and I went from a 4.93 to a 4.90 just like that. If you lose points it’s real hard to get them back, so I’ve learned [to] make sure that a person can’t rate you, close out the ride before you let them off. The ride will cancel—it will still pay you up until that point, but it’s impossible for them to rate you. I tell them straight up too—“I’m canceling you because I don’t want you to rate me.”
While only around a 4.6 rating was required for continued access to the app he was using, Smith was so deeply invested in the algorithmic management’s rating system that he exerted considerable efforts—with both direct and indirect financial consequences—to maintain a high rating. Canceling rides increased drivers’ overall cancellation rates, putting them in jeopardy of not receiving future incentives, of not being accepted into loyalty programs, or of being deactivated. More immediately, if drivers ended a ride early, they did not receive the full fare and were not covered by the platform’s liability insurance during the gap. Even so, some drivers saw forgoing income and insurance coverage as an acceptable tradeoff to maintain high customer ratings, showcasing how one’s consent to the work could eclipse one’s own best interests. Overall, drivers exercised choices to maintain high ratings and the attendant privileges.
Inflating surges
Drivers also influenced their input into the algorithmic management system in hopes of inflating fares. For example, when waiting in a virtual queue (e.g., airport queues), drivers would manipulate the work-instruction algorithm by toggling airplane mode on and off on their phones while monitoring surge prices in the driver and/or rider app. Toggling preserved their place in the queue while also signaling that there were fewer cars in it, which drivers hoped would drive up demand-based pricing. Other times, drivers tried to inflate demand-based pricing algorithms by driving to busy areas and then declining multiple rides. However, if the algorithmic management system detected consecutive declines, drivers were blocked from the app. John (Missoula) described, There was a bigger concert and I kept denying rides, waiting for the higher surge charge because I knew there was going to be one. And they blocked me for like a minute or two, which I was fine with because that’s what I wanted to do. [Laughs] I didn’t know if I was going to be able to turn back on, which I was irritated about, so I did a hard shut-off of the app a few times and it let me back in.
A more aggressive deployer of this tactic, Orion (Detroit) faced more-severe penalties from the system as he was blocked for up to 30 minutes, but he described how he obtained a $180 airport fare, more than quadruple the regular rate of $40: [RideHail] wants me to take as many rides as possible. What do I want to do? I want to make as much money as possible. They punish me for it; however, I profit more than I hurt. [Describes declining several rides waiting for a surge increase.] If I [don’t accept] three rides, I can’t log in for half an hour. I figured out ways around it—I request myself [after declining two rides]. [Laughs] I use a separate email for my rider and driver account, and I’ll go back online.
22
Here, Orion made multiple choices (e.g., logging off, rejecting rides) and used the platform’s technology—the rider app—against the algorithmic management system by monitoring fare changes in real time. Trying to counter these algorithms had costs both in terms of time and money as well as the risk of being blocked. However, aware of these potential consequences, drivers were often able to circumvent the algorithms’ penalties, for example by restarting the app when blocked or by requesting themselves from another phone before being blocked. Ultimately, this cycle of deviance tactic–sanction–counter reinforced drivers’ sense of having choices, especially when the algorithms were eventually beaten and drivers secured a higher fare.
Outcomes of Consent Produced by Deviance Tactics: Skill Development and the Amplification of Managerial Conflict
As in engagement tactics, when deploying deviance tactics, drivers found meaning by seeing themselves as skillful agents, deftly navigating the algorithmic management system. However, unlike engagement tactics, the skill expressed in deviance tactics was not in working with algorithms but, instead, in being able to counter the algorithms, such as by pre-selecting passengers (Marsha, D.C.) and declining shared rides (Jackson, Philadelphia). Drivers described their practices for interacting with the system as “complicated” (Orion, D.C.), “tricks” (Borris, Boston), or “methodolog[ies]” (Arthur, L.A.) and said that they “analyze, prepare, wait, and execute” (Smith, D.C.). Safeguarding their skills, drivers would tell me about them in hushed voices only after I promised not to use them myself, or they asked me not to include their tactics in my papers. 23 Confident in their abilities, drivers scoffed at penalties—John (Missoula) laughed when he was blocked, noting “that’s what I want it [the algorithm] to do,” and Orion (D.C.) ultimately secured a fare four times higher than the regular fare. Even with changes in the algorithmic management system, drivers who extensively deployed manipulation tactics envisioned themselves as smarter than other drivers. They spoke condescendingly about those who did not curate their matches (David, Detroit) or laughed when they saw a driver get matched with a customer they themselves had just rejected (Smith, D.C.). Describing drivers who simply followed the algorithms, Orion (D.C.) said, “They aren’t smart . . . they don’t understand the implications . . . they make awful, awful decisions and . . . I swear they [are] so proud of themselves.”
The second outcome associated with deviance tactics was the amplification of conflict with the algorithmic manager. When deploying manipulating tactics, drivers were aware of the conflict with the system. For example, Arthur (L.A.) was combative when matched to a low-paying ride, telling himself, “That’s not how this game is going to work for me,” before rejecting the ride. Smith (D.C.), who did not want me to reveal all his workarounds, said he tries to “drive smart [but he] know[s] this is not what [RideHail] wants [him] to do.” He went on to explain the inherent antagonism between him and the algorithmic management system: “We really can’t be friends with our bosses. If you don’t establish your own objective, you’re going to get caught, swept up in what they [RideHail] want you to do.” Comparing driving to his previous work in restaurants, Alphonse (D.C.) said, “In other places, you might get a complaint, but you can solve the problem right away, you don’t need to write. You can say, ‘sorry,’‘help me.’ But here, is nothing. It’s working with a machine.” While deploying deviance tactics set up drivers to be in opposition to the algorithmic management system, amplifying the potential for conflict, for the most part drivers were still able to obtain their desired rides.
When deploying deviance tactics, workers exercised greater latitude of choice about how to interact with the algorithmic management system and to manage these tradeoffs, compared to their use of engagement tactics. In simple deviance tactics, workers tried to circumvent the blind matching policy by declining certain rides. In complex deviance tactics, workers exercised the greatest latitude for choice by manipulating their input into the algorithmic management system to change their status within it. Pushing against the edges of what the management system would allow, drivers could successfully counter any penalties, occasionally using multiple pieces of technology to do so (e.g., requesting themselves through another account). Despite paying a cost in terms of time and money to deploy this tactic (e.g., paying cancellation fees after requesting oneself), drivers saw themselves as skillful agents because they obtained desired rides.
Yet, even though workers manipulated their input into the algorithmic management system when deploying deviance tactics, they were still consenting to the work because RideHail’s second objective was for the driver to remain online. Even when drivers deviated, they were still compliant with the larger management system (though not fully aligned with it), in part because all of their counter-measures were allowable within the system (i.e., programmed leniency). 24 Last, it must be emphasized that while these choices were confined, they were also constant, enacted hundreds of times within a shift. In a single ride, a driver could check the heat map to find high-demand areas (simple engagement tactic), use a passenger app to position oneself (complex engagement tactic), and reject incoming shared rides (complex engagement tactic) before accepting a private ride. Taken together, these findings show how in each tactic, workers could make frequent, albeit narrow, choices that all lay within the bounds of the algorithmic management system.
Changes in algorithmic management
Over the seven years of this study, the algorithmic management system frequently changed, and for the most part, workers were able to adjust, continuing to exercise choice. Often, changes in the pricing algorithms prompted changes in drivers’ behavior, as drivers increasingly believed that it was more profitable to deploy engagement rather than deviance tactics. As described above, many drivers shifted from rejecting to accepting all shared rides as new incentives were introduced. Drivers thought twice about multi-homing when a bonus based on the amount of uninterrupted time spent on an app was introduced. Loyalty programs such as UberPro (introduced in 2018) and Lyft Tier (revamped in 2019) provided drivers with more information and advantages (e.g., being able to see a rider’s destination before accepting a ride, priority airport queueing) only when drivers more closely aligned their behaviors with RideHail’s goals, e.g., such as by keeping their acceptance (cancellation) rates high (low). When comparing data between my first and last data collections, in 2016 and 2023, respectively, I identified fewer mentions in the later data of certain behaviors such as pre-selecting riders, presumably because the matching algorithm no longer used driver–rider proximity as its only input. Instead, workers often discussed the new choices they were offered, such as rider selection (e.g., Uber drivers could select a preference for women riders), ride type preferences (Trip Radar introduced by Uber), up front pricing (Uber, Lyft, Juno, Gett), and instant payment (Uber, Lyft). My informal interactions with two data scientists at one RideHail company suggested that programming of the algorithmic management system was geared toward efficiency optimization, which could include pricing incentives. 25 Taken together, these changes point to algorithmic management’s ability to program out deviance tactics while still emphasizing workers’ ability to exercise choice. Ultimately, these choices, which aligned with the organizations’ twofold goals of having workers follow algorithmic nudges and stay online ready to work, elicited consent. Table 3 provides a summary of all tactics.
Summary of Engagement and Deviance Tactics in Response to Algorithmic Management
Withdrawal: Leaving the Platform or Damaging the Integrity of the Algorithmic Management System
In contrast to the tactics that elicited consent to the algorithmic management system, workers could also withdraw consent from that system. Two ways in which workers withdrew consent were by logging off an app with no intention of returning or by compromising the integrity of the algorithmic management system, which could result in permanent deactivation if detected.
Logging off an app with no intention of returning
The most straightforward way to withdraw consent is by logging off a RideHail app with no intention of returning. In traditional employment, there are norms of reciprocity and information exchange in which workers take scheduled leaves and give formal notice when quitting. In contrast, when an on-demand worker logs off, they could return to the platform in hours, days, weeks, or never, and the platform is unaware of the length of and reason for their absence. Common reasons that drivers decided to stop working included low pay, confrontations with customers, disgruntlement with the app or the rides assigned, or the physical toll of driving. Noting the strain of driving all day, Peterson (Chicago) said, “It’s just not good, for your body . . . to be sitting all day. My back, knees, everything was hurting. And I had a fungus on my foot that never healed up” because his feet were near the floorboard heating vent. Other drivers were unhappy with the perceived poor working conditions. Marcel (D.C.), who had been unfairly deactivated (and later reinstated) because of a customer complaint, quit because of the low pay and the bodily toil: It wasn’t what I thought. You really have to grind for the money and when you think about the gas, you ain’t really making that much. That’s the sad thing, I don’t know if it’s capitalism or America in general, but everybody wants stuff for cheap when it comes to services. . . . I just [left] and graduated from the Academy and now I’m at an actual fire station.
Other drivers left because better earning opportunities presented themselves, such as traditional full-time work (Terry, George), an uptick of business in their other lines of work (Peterson, Tamara), or a change in family circumstances (Maxwell, Sonya). Regardless of why an individual stopped driving, leaving the industry was simple in that individuals simply never opened the RideHail apps again.
Damaging the integrity of the algorithmic management system
To be matched more quickly with more-lucrative rides, some drivers obfuscated, or “geo-spoofed,” their location so that the algorithmic management system thought they were elsewhere. This represents a withdrawal of consent because drivers were challenging the integrity of the algorithms themselves and violating the platforms’ terms of service. A brute-force technique, “the phone holding game,” required two participants—one always waiting at a location, typically the airport, in a RideHail queue, while a driver with a second phone completed rides and then returned to the airport. The two participants would swap phones each time the driver returned. The outcome was decreased down time between rides, as the phone at the airport signaled to RideHail’s algorithmic management system that the driver was on site and available for airport pickups at all times.
Other techniques were more technically sophisticated, requiring special software that bypassed certain technical checks and provided root access to override RideHail’s GPS tracking systems so drivers could appear to be at a certain location when they were not there. In the following quote from a discussion board, I describe how a driver overrides two ride-hailing companies’ GPS systems simultaneously: Smarter people can rig the system, it’s easy and simple and [RideHail1] can’t do shit about it. I have been getting into the [RideHail1] and [RideHail2] airport queues from my house. I get up in the morning. Drop my [geo-]pin at the airport waiting lot [using software that enables dropping a geo-pin from a remote location]. Take a shower, get ready and travel to the airport which is a 30-minute drive. Almost every day I get the [first ride] within 5 minutes. I have two phones, one for [RideHail1], one for [RideHail2]. If I get a [RideHail2] ride first, I pick the ride up and on the [RideHail1] phone keep the pin dropped at the waiting lot. After I have dropped the [RideHail2] ride and [make it] back to [the] airport, I have less than 10 drivers ahead of me. Just wait for 2 minutes for the next ride. Don’t have to wait ever in the stupid queue. (Shubra)
Other geo-spoofing tactics included using programs that made it appear that drivers were based in cities with higher bonuses even when the drivers were working elsewhere, using older phones or operating systems that made it harder for RideHail’s system to detect geo-spoofing, or inputting into RideHail algorithms longer routes than drivers actually took: I had my phone built like a fortress with multiple modules and roots, I didn’t just spoof but made my car drive the long routes PLUS more while taking the short [route], never got caught ’cause they simply couldn’t [catch me]. I remember pax [passengers] looking at their phone and asking why [the route on their app showed] the car was going through another highway, I would contain my laughter and tell them—“you know [RideHail], their app sucks!” (Ennard)
Given that geo-spoofing was a direct violation of RideHail’s terms of service, when the algorithmic management system detected these behaviors, drivers were sent warning messages and/or immediately blocked (see Images 20 and 21 in the Online Appendix). 26 On a discussion board, Yuvor explained his situation: “About 4 to 5 months ago my [RideHail premium service] account got blocked most likely because I had been falsifying my GPS location to my advantage for a while.” Asking for advice on a forum, he then asked whether there was any other action he could take besides unlawfully opening another account under a different identity. The best advice his fellow drivers offered was to wait six months and to change his email, phone number, address, and driver’s license photo “because [RideHail] uses AI to verify that’s the same person” (Gayord), but still it was a “useless Hail Mary” (Yuvor). Chances of reactivation were slim. As Ebenezer put it, “There’s no way to get around being banned by [RideHail]. I would consider a successful method to be one where the perpetrators don’t get arrested by the FBI” [emphasis in original].
As the above examples show, consent can be withdrawn. RideHail workers were able to exercise choice and to withdraw consent by no longer signing into the apps or by engaging in actions so fraudulent that drivers were at high risk of being flagged and banned from the apps. In contrast to traditional employment, workers in these on-demand settings can more easily quit without any explanation or warning to the platform companies. Indeed, the physical isolation of the work, frequent changes in the algorithm, low wages, and high performance pressure increase the likelihood of workers leaving (Rosenblat, 2018). When workers compromised the integrity of the algorithmic management system in ways that were flagrantly opposed to the platform’s community guidelines and terms of service, they were contesting the rules of RideHail—in essence rejecting algorithmic management and refusing to consent. Moreover, unlike with more-traditional forms of employment, RideHail could easily let go of workers with no explanation. Overall, these findings describe how both tactics—engagement and deviance—securely elicited workers’ consent on the platform companies’ (and not the workers’) terms, because when workers tried to interact with the system on their own terms, such as by geo-spoofing, the result could be immediate and permanent deactivation.
Discussion
This research opens a new line of inquiry on how digital platforms and algorithmic management shape individual agency, namely, through eliciting consent, or workers’ active participation, to align their efforts with managerial objectives. Instead of manufacturing consent through social relations, such as competition between coworkers or strategic intimacies forged with managers (e.g., Sherman, 2007; Mears, 2015), consent in this context was choice-based: it relied on workers having access to an algorithmically mediated set of options at the point of production, in this case on RideHail’s apps, in which workers could make constant, if confined, choices.
Contributions to our Understanding of Workplace Consent
While Burawoy (1979) emphasized the role of social relations in manufacturing consent, over 40 years ago he also hinted at the dynamic relationship between technology, choice, and consent. He compared what he found at the engine factory he studied, in which machinists played industrious games, with the same factory that his predecessor Roy (1952) had studied, in which machinists had loafed about: “The differences between the organization of work at Geer and Allied suggest even greater ‘quantitative’ choice within even narrower limits. Harry Braverman missed the equally important parallel tendency toward the expansion of choice within those ever-narrower limits” (Burawoy, 1979: 96). In the more than 20 years between these scholars’ studies of work at the same factory, technology had advanced such that the machine’s pace had sped up work, and the technical complexity had given workers more choices about how to accomplish this work. In my research context of on-demand work, which is largely asocial, I build on this idea and seek to “bring the technology back in” to examine how algorithmic management shapes the manufacturing of consent.
In my study, as workers described exercising choices, they emphasized their skill in interacting with the algorithms on their own terms and maximizing their interests. This is consistent with theory on workplace consent: workers cannot dissent from the things they do not like about the work without dissenting from the structure of the work that makes these things possible (Sallaz, 2013). Games helped manufacturing workers beat monotony, schedule flexibility supported taxi drivers’ aspirations to be family men, and call center agents were energized by a fast-paced learning environment. Correspondingly, one reason that ride-hailing drivers enjoyed work under algorithmic management is because it enabled them to make minute choices while executing the work.
Yet, as important as it is to take workers’ experiences seriously, it is also crucial to remember the larger conditions in which the work is situated. The relationship between labor and capital in platform work is obscured because of the invisibility of platform owners and managers, who hide behind an opaque management system (Chai and Scully, 2019; Vallas and Schor, 2020). Work arrangements amplify this information asymmetry. While classified as independent contractors, workers cannot organize (Dubal, 2023a), may not negotiate personally with managers (Maffie, 2022), lack access to official voice channels (Bucher, Schou, and Waldkirch, 2023), have minimal recourse against customer malfeasance (Laemmli, 2023), and have limited knowledge of the actual code that underpins the algorithmic management system (Christin, 2020). Indeed, the algorithmic management systems set the terms of interaction. In engagement tactics, the terms are obfuscated because workers’ choices align with or do not directly counter the algorithmic management system. In deviance tactics, workers’ choices may seem more in alignment with their own interests, especially when individuals counter sanctions; however, this countering works only because leniency is programmed into the system. Thus, workers’ interaction with the algorithmic management system is on its terms, not workers’. As noted, when drivers interact with the algorithmic management system on their own terms, such as by geo-spoofing, the result can be sudden and permanent deactivation. This research thus underscores how important choice can be for workers to secure continued and active participation in the work.
An algorithmic management system in which individuals have circumscribed choices while working is crucial to keeping the platform system running smoothly, because workers can invest in their work and develop a sense of mastery. Elegantly debunking economists’“agency problem,”White (2008: 301) noted that workers’ choice and the agency that comes with it is a solution, a “neat kind of social plumbing” in that organizations can use it to attain and maintain organizational control. In the on-demand economy, this agency, actualized as choice, is extreme in that there is a multiplication of choices in an employment model that is contractual and in work that is algorithmically segmented such that workers can make constant (if confined) choices. Yet, the terms of the labor conditions are obfuscated, and workers do not know the actual conditions of the exchange, keeping workers in a quasi-suspended state in which the choice allowed is encompassed within a socio-technical system of concentrated platform power (Wood and Lehdonvirta, 2021; Davis, 2022). The allure of having a sense of choice alongside structurally precarious work conditions results in what I call the “good bad” job.
The deployment of algorithmic management is increasing in workplaces, and how choice is designed into the system will play an important role in how workers’ consent can be elicited. In this setting, the algorithmic management system functioned similarly to an opt-in system in that individuals made small choices that further embedded them into the work process. Uber’s Trip Radar, for example, presented drivers with several different rides to choose from. Alternatively, drivers could choose to have the system automatically accept all rides on their behalf. These choices at each stage of the work process contributed to the work being compelling, pulling workers in. However, one would expect that an algorithmic management system functioning like an opt-out system would have different implications because it presumes, rather than elicits, consent. Even though an opt-out system might encourage participation in the work, it may not elicit the same positive feelings in workers (e.g., skill development). This suggests that consent generated in an opt-out algorithmic management system may be more tenuous in that it may increase conflict and pushback and weaken long-term commitment. Rahman and colleagues (2023) found that in an online open labor market, the way that choice was designed into the system increased workers’ frustration and disenchantment. In other design architectures, scholars have questioned the ethics and longevity of opt-out systems, noting that they erode participants’ trust (Johnson and Goldstein, 2003; Thaler, Sunstein, and Balz, 2012). Future research should pay close attention to how the design of algorithmic management systems may affect how consent is produced and workers’ experiences.
How choice will appear (e.g., type, frequency) depends on where algorithms are inserted to manage the work process. In piece-rate work, in which the entire work process is algorithmically managed (e.g., grocery delivery), individuals are more likely to have choices throughout the entire process. In retail work, algorithmic management is often embedded in scheduling or evaluation. Thus, workers are more likely to have choice when determining when to work or during the point of sale, in which workers may be able to influence customers’ purchases (Van Oort, 2023; cf. Wu, 2020, for an example in care work and ratings). These examples emphasize the importance of choice-based consent in algorithmically mediated work, suggesting that the location and latitude of workers’ choice depend on how algorithms are embedded within the work. Taken together, these examples underscore the importance of not examining a single feature of technology in isolation and not generalizing an entire technology wholesale.
Outcomes of algorithmically mediated choice-based consent
I found that choice-based consent in algorithmically mediated work is associated with two different worker outcomes: skill development and the potential for worker–management conflict. Prior literature has understood consent as workers’ active cooperation with the work, such as deftly dealing a hand of blackjack (Sallaz, 2009) or being an appropriately irate bill collector (Rafaeli and Sutton, 1991). To be sure, active cooperation is evident in ride-hailing, such as when drivers accept all rides and chase surges. Yet, this study also describes how deviant behavior, in which workers regularly circumvent the algorithms, produces consent because workers can successfully counter sanctions. In engagement tactics, workers’ sense of mastery and pride was rooted in being skilled in “getting over” the system. This finding is distinct from prior research showing that workers assessed their skills by ranking themselves against coworkers whose skills they could reliably assess (Burawoy, 1979). While engagement and deviant behaviors seem oppositional and are associated with different types of skill, they both elicit the same outcome, consent.
Prior research found that a byproduct of consent is that it mollifies the inherent worker–manager conflict by transferring it laterally to coworkers (e.g., Sherman, 2007). Examining a workplace without coworkers, this research finds that the potential managerial conflict is eradicated or amplified depending on the tactic. In engagement tactics, the potential for conflict is eradicated when workers see themselves as powerful, skillful actors within the algorithmic management system. In contrast, in deviance tactics, conflict between workers and management is amplified as workers counter the algorithms. While conflict is amplified, workers are still able to see themselves as powerful actors because they evade sanctions. Taken together, these findings suggest that the amplification of worker–management conflict does not necessarily prevent workers from consenting if the work is engineered so that individuals still exercise choice.
Since this study focused only on the relationship between algorithmic management and consent, future research could consider the relationship between consent and emerging technology more broadly. Different technologies, depending on how and where they are integrated in the work process, could lead to similar or different ways in which consent is produced. In Burawoy (1979), the way people related to their machines was important: workers needed a good machine, not a bad one, to make out. Surveillance devices that can be easily circumvented or removed from the labor process (e.g., body trackers, Levy, 2023) would shape the production of consent differently than would technology more deeply embedded in the work process (e.g., big data; Brayne, 2020). Moreover, the nature of the work may shape the relationship between consent and technology. Ranganathan and Benson (2020), for example, found that RFID trackers were associated with the manufacturing of consent only in simple sewing tasks and were ignored in more-complex sewing tasks. Future research should continue to explore these topics.
Broadening the locus of consent
This research suggests two exciting new avenues for research on consent: expanding the temporal horizon of consent production and creating a more interdisciplinary conceptualization of consent. Constant and confined choices, such as personalized incentives that require workers to complete a consecutive number of rides in a short period, keep individuals engaged while executing the work on a specific platform. But a defining feature of the on-demand economy and the contemporary economy, more generally, is the rise of multi-job holding (Campion, Caza, and Moss, 2020) and deconstructed work (i.e., having multiple roles within one company; Rogiers and Collings, 2024). To paraphrase the esteemed management scholar Gerald Davis (2016), our grandparents had careers, our parents had jobs, and today we have tasks. To that end, a promising avenue for future research is to understand how consent is produced beyond the execution of the labor process, such as to a job role (Schlund and Bohns, 2022) or to a mode of working (e.g., on-call work; Wood, 2020). Scheduling apps such as Qwick, Pared, Shiftgig, and Snagajob allow retail and restaurant workers to choose from among companies to offer their services. Over the seven years of this research, many drivers in my study switched between types of on-demand work, purposely eschewing the traditional labor market (cf. Ravenelle, 2020, 2023). Similarly, Adler (2021) found that cultural creatives committed to non-standard jobs (e.g., bartending) to reinforce their dedication to their artistic aspirations. As individuals’ tenure at any one specific job continues to decline, an important area for scholars to understand is how consent is generated beyond the point of production at any one specific firm.
At its heart, this article addresses the classic sociological question of why workers consent to low-quality working conditions. Yet, the concept of consent is central to many of society’s most pressing issues, such as data privacy, medical decision making, protocols for conducting research, and determining what constitutes a sexual assault. The varied fields in which consent is studied, ranging from philosophy to legal studies to psychology (Kleinig, 2010; Bohns and Schlund, 2020; Hirsch and Khan, 2020), suggest potential multidisciplinary and interdisciplinary synergies. In the book Logic of Consent, legal scholar Peter Westen (2017) identified three components of consent: freedom (one must agree free of coercion), knowledge (one must be informed of its terms), and intellectual competence (one must be able to understand). Using this framework, we can ask, how might individuals’ economic dependence on on-demand work affect whether they are consenting to their work, particularly given frequent and opaque algorithmically determined wage changes? 27 Can on-demand workers ever be truly informed of the circumstances surrounding their work if they are presented with new terms and conditions every time they sign in to an app, especially if they are in a flow state (induced by the micro-choices presented in the app)? And even if workers agree to these terms, does this give organizations the right to sell workers’ information and, in essence, financially benefit from their labor twice? The core concern in these questions is that even if individuals have legally consented, this does not necessarily mean that they feel like they have consented—which is key to the sociological definition of consent. Untangling these complex and intertwined issues is necessary given the fast-changing nature of work and technology.
Algorithmic Management and Workarounds
This research adds nuance to the issue of how to categorize workers’ responses, beyond compliance and resistance, to managerial regimes, adding to the literature on workarounds (e.g., Gasser, 1986; Beane, 2023). Prior research has, for the most part, recognized the agency of on-demand workers only when they resist (e.g., Shapiro, 2018) or engage in off-app activities (e.g., disintermediation, Maffie, 2023). Here, I found many examples of workers expressing agency and developing workarounds. Both engagement and deviance tactics could be simple or complex depending on how closely they aligned with the algorithmic management system. In complex engagement tactics, for instance, workers’ behaviors were off script, benign violations of norms, such as using the customer app to scope out demand and pre-position themselves. This distinction is important because it shows that consent is not merely compliance, and workers are not dupes. Individuals whose behaviors are largely within the algorithmic management system are not perfectly compliant, and authoring these tactics is part of what gives these individuals a sense of mastery. Moreover, in complex tactics, drivers displayed a command of their environment by using multiple pieces of technology as raw inputs for creative workarounds (e.g., requesting themselves on rider accounts). These varied tactics exemplify the “culture worlds” (Seaver, 2017: 7) and “algorithmic imaginaries” (Bucher, 2016: 30) that workers create to navigate their work. Future research should explore other ways that individuals express agency within such systems, such as via voice or creativity.
In addition, this research provides insight into the difference between behaviors that are deviant yet within the boundaries of the system and behaviors that constitute a withdrawal of consent. Most research equates the withdrawal of consent to quitting (e.g., Sallaz, 2013). This study describes another way in which workers may withdraw consent: compromising the integrity of the management system. Compared to deviance tactics, the withdrawal of consent can be even more deleterious for on-demand organizations since they rely on an always available workforce. As organizations increasingly use technology to manage and track workers, research should continue to explore the relationship between consent and technology, including when and how consent is withdrawn and its organizational implications.
Implications of Algorithmically Mediated Choice-Based Consent: The Rise of the “Good Bad” Job?
While robots are not taking over most jobs, algorithms are increasingly taking over managerial functions previously held by humans, such as hiring, firing, directing, and evaluating workers. A large class of jobs—ranging from truck driving to warehouse work to retail work to home health—are being managed by algorithms and done by those classified as independent contractors. Surprisingly, many individuals appear to like or even prefer non-standard or algorithmically managed on-demand work compared to standard employment, even when they are aware of its exploitative aspects (Fielden, Tench, and Fawkes, 2003; Ertel et al., 2005; Farrell and Morris, 2017; Östergaard, 2017; Adler, 2021). This study argues that workers’ ability to make frequent choices in algorithmic management systems is one reason they enjoy this line of work. This observation then raises the question of the broader implications of algorithmically meditated choice-based consent.
One consequence of algorithmically managed jobs is that work that is structurally considered “bad”—with variable (and often low) wages, minimal labor protections, and limited advancement opportunities—may be psychologically experienced as not so bad. Two seminal books—Kalleberg’s (2011)Good Jobs, Bad Jobs and Standing’s (2011)The Precariat—document the growing trend of precarious work, which many scholars argue only becomes more precarious when algorithmic managers are introduced (e.g., Griesbach et al., 2019). At the same time, algorithmic management allows drivers to have varied choices at multiple levels (within rides, between rides, on work itself) and in a greater type and amount than in jobs typically available for the similarly skilled part of the labor force. On-demand consolidator apps such as Gridwise and JackApp afford more choice, allowing workers to coordinate their activities to secure the highest-paying gig at any moment. These constant choices entice individuals to work longer, invest more effort, and, in essence, commit more (cf. Staw, 1981 for how small choices engender long-term commitment). And workers’ enjoyment is not faked. The risk of bad bosses is eliminated. Hours are flexible. Workers take pride in their skills. As noted, some drivers leave what Kalleberg would call a good job for on-demand work (e.g., the unionized carpenter) or remain committed to non-standard work despite other options (Adler, 2021). Similar to Ravenelle (2020), I found that people in my study made a career of on-demand work, switching between offering different app-based services—including rides (Uber), package delivery (Amazon Flex), meal delivery (DoorDash), and grocery shopping (Instacart)—instead of finding traditional off-platform work. Cultural narratives exalt these moves, lionizing hustle culture (Hill, 2020), heroizing workers during the height of the COVID-19 pandemic (Cameron, Chan, and Anteby, 2022), and exalting the “romance of entrepreneurism” (Ravenelle, 2019: 281). 28 These constrained choices, scaffolded by algorithmic management, and cultural “deep stories” (Hochschild, 2016: 128) attract adherents, providing insights into why individuals behave in ways that do not always seem aligned with their own interests.
The Janus-like tension between the structural elements and the psychological experience of doing on-demand work might explain its growing magnetism and appeal. Tensions can be generative (Smith and Lewis, 2011), and technology can amplify the generativity (Murray, Rhymer, and Sirmon, 2021). Mobile email devices reinforce the autonomy paradox, in which individuals work longer hours while claiming to have more schedule flexibility (Mazmanian, Orlikowski, and Yates, 2013), and sophisticated user interfaces keep individuals on machines longer (e.g., Schüll, 2012). Yet, technologies reflect and embody broader sociocultural values and can never be seen as socially or politically neutral. Noting the thrill of a Facebook “like” or being matched on Tinder, McMillan Cottom (2020: 446) wrote, “Knowing the extractive terms of their labor does not diminish their [workers’] enjoyment of the job. Platform capitalism owes much of its dominance to how good it feels to be captured by the platform.” Ultimately, this generativity can keep individuals participating in work that provides a psychological sense of fulfillment yet requires them to take on more economic and physical risk compared to similarly skilled, more-traditional work. Despite reassurances that Uber will not take over the world (Fleming, Rhodes, and Yu, 2019) and strong arguments from academics about why more firms should provide more good jobs (Ton, 2014, 2023), these insights suggest that these “good bad” jobs may be here to stay: as a long-term career, as interim work as people move between jobs and/or labor markets, or as a career stage for those who then enter standard employment. And these findings suggest that workers may be satisfied with these jobs.
This article also has implications for the current legal and policy debates regarding worker classification, by moving beyond the binary classification of on-demand workers as either employees or contractors (Dubal, 2017, 2020). As a new organizational form (Vallas and Schor, 2020), platforms manufacture consent and garner control in ways that are not easily defined, leaving checks like the ABC test, which is used in many states to determine whether a person is an employee or an independent contractor, largely ineffective. Moreover, given platforms’ ever-changing, multi-sided nature, there are broader challenges about how to hold them accountable through regulations (Khan, 2016; Pignot, 2023; Rahman, Karunakaran, and Cameron, 2024). Therefore, labor relations on platforms need to be understood not only through a contractual perspective. This research shows how power and control are deeply involved in the production of consent such that managerial control is achieved in a way that makes precarious work appear desirable through constant choices. Legal scholars and policy officials must pay attention to these new configurations to hold these platforms accountable, for the benefit of workers and customers alike.
Boundary Conditions, Limitations, and Future Directions
There are limitations to these claims and opportunities for future research. First, this study cannot answer how individuals came to learn a tactic. A reasonable hypothesis is that workers learned the tactics from one another; however, less than 15 percent of the drivers I interviewed were active on online forums, and when directly asked, most drivers said they learned from trial and error. My analysis suggests that workers did experiment, since sanctions for deviance tactics occurred only after multiple ride refusals, and as solitary workers, individuals had more time on their hands to experiment. While it is well documented that on-demand organizations experiment with workers as part of their operational strategy (Luca and Bazerman, 2021; Rahman, Weiss, and Karunakaran, 2023), this research suggests that workers themselves might engage in similar experimentation (cf. Beane, 2019). Other research designs and sampling strategies (e.g., longitudinal panels, diaries) could provide insights into how workers learn these tactics.
Second, this study’s claims about how consent is reconfigured are relevant to other settings that use an algorithmic management system. Although this study was undertaken among workers classified as being in an independent contractor work arrangement, this model would likely generalize to workers in standard work arrangements with a similar work process and an algorithmic management system, i.e., if RideHail drivers were suddenly to become employees (as they already have in many countries in Western Europe). Verdin and colleagues (2023) found that in Spain and Germany, independent contractors and employees at a food delivery platform were under near-identical algorithmic management systems, suggesting a similar process in manufacturing consent. Similarly, when a U.S. package delivery company converted its drivers to employees, the labor process and, hence, the process of manufacturing consent remained basically unchanged (Johnston et al., 2023). This suggests that these findings are generalizable to workplaces with workers in different classification systems.
This case study presents a set of claims that could be expanded in other settings. This research is based primarily in the U.S., and the status and experience of platform workers vary across countries according to legal and social norms. Indeed, emerging research based on global data suggests that platform workers’ experiences vary based on configuration of the algorithmic management system (Cameron and Thomason, 2023), social status (Thomason and Cameron, 2023), and the organizing and labor history of a given locale (Tassinari and Maccarrone, 2020; Woodcock, 2021). Comparative research should examine how algorithmic management changes due to variations in organizational form (e.g., open vs. closed labor markets), geographic regions (e.g., Global North vs. Global South vs. Global East), and governance structures (e.g., corporations vs. cooperatives) and how this change affects the production of consent. Driving is largely independent, simple work (Hubbard, 2000; Viscelli, 2016), and as algorithmic management is integrated into more interdependent, complex work, how consent is manufactured may also change (e.g., warehouse work supervised by human and algorithmic managers, content creation by temporary teams; Valentine et al., 2017; Vallas, Johnston and Mommadova, 2022; Mears, 2023). 29
Conclusion
As emerging technologies and shifting organizational paradigms continue to change the workplace, it is imperative to re-examine and update mainstream organizational theories (Barley, Bechky, and Milliken, 2017; Bailey et al., 2019; Barley and Orlikowski, 2023). This study takes an important step in this direction, examining how algorithmic management reconfigures long-standing theories on workplace consent. Choice-based consent illuminates the importance of constant, even if confined, choice as a mechanism that keep workers engaged, especially in jobs considered to be of poor quality. As the relationship between technology and work continues to evolve, the relationship between control and consent will continue to develop in ways that require innovative data collection and theory building.
Supplemental Material
sj-pdf-1-asq-10.1177_00018392241236163 – Supplemental material for The Making of the “Good Bad” Job: How Algorithmic Management Manufactures Consent Through Constant and Confined Choices
Supplemental material, sj-pdf-1-asq-10.1177_00018392241236163 for The Making of the “Good Bad” Job: How Algorithmic Management Manufactures Consent Through Constant and Confined Choices by Lindsey D. Cameron in Administrative Science Quarterly
Footnotes
Acknowledgements
I am extremely appreciative of the associate editor, Forrest Briscoe, and the three anonymous reviewers for their very constructive feedback and deep engagement with my research. I thank the workers at RideHail for sharing their stories. Solo-authored papers are never an individual endeavor, and I have deep gratitude for everyone who assisted in the development of this manuscript. Thanks to Yan Chen, Sean Dew, Eli Gonzalez, Vidisha Hermani, Kalie Mayberry, Brandon Nguyen, Ladasa Tiraviriyapol, and Ruiling Wen for their research assistance. I am grateful for the editorial suggestions of Kristin McGuire, Rose Ernst, and Sylvia Kudaverdian, as well as the ASQ editorial team, Joan Friedman and Ashleigh Imus. Previous versions of this manuscript benefited from comments from Jerry Davis, Jane Dutton, Beth Bechky, Seth Carnahan, Tawanna Dillahunt, Michel Anteby, Callen Anthony, Sue Ashford, Danny Armanois, Steve Barley, Matt Beane, Ethan Bernstein, Matthew Bidwell, Peter Cappelli, Stephanie Creary, Curtis Chan, Angelique Davis, Julia DiBenigno, Veena Dubal, Mauro Guillén, Kevin Lee, Tiffany Johnson, Kate Kellogg, Arvind Karunakaran, Katherine Klein, Natalia Levina, Mike Maffie, Melissa Mazmanian, Dave Mayer, Ashley Mears, Wanda Orlikowski, Nick Oucchiuto, Mike Pratt, Hatim Rahman, Aruna Ranganathan, Alex Rosenblat, Juliet Schor, Ben Shestakofsky, Gretchen Spreitzer, Julia Ticona, Jim Walsh, and Nathan Wilmers, among others. I am also thankful for the feedback and support from the May Meaning Meeting, Stanford’s future of work community, NYU Stern’s Qualitative Writing Group, NYU Stern’s Data Analysis Group, the University of Michigan’s Center for Positive Organizations, the Data and Society Research Institute, the Aspen Institute, the Women of OB writing group, and the Remember to Exhale Writing group. I am grateful to the seminar participants at the following universities and colleges: University of Tennessee, San Francisco State University, University of Western Ontario, Vanderbilt University, Cornell University, University of Oregon, University of British Columbia, University of Southern California, University of Pennsylvania, Harvard Business School, Harvard Law School, University of Delaware, University of Michigan, Stanford University, Northeastern University, New York University, Imperial College, University of Pittsburgh, Carnegie Mellon University, University of Texas at Austin, Santa Clara University, University of Chicago, Case Western University, Boston College, Temple University, MIT, London Business School, Boston University, Northeastern University, Peking University, University of Hong Kong, Harvard Business School Blackbox Lab, the University of California at Santa Barbara, and Princeton University. I am also grateful for the comments of attendees at many conferences, seminars, and workshops, too numerous to mention. Deep gratitude to the workers at TeaHaus in Ann Arbor, Michigan where much of the first draft of this manuscript was written. I am grateful for the financial support of the Rackham School, University of Michigan and Wharton’s Center for Human Resource, the MACK Center, and Wharton’s Behavioral Lab.
Correction (May 2024):
Article updated online; for details please see the
Author’s Note
The Figure 1 caption has been updated to include attribution to Duggan et al. (2020). The author’s second affiliation has also been corrected to “Institute for Advanced Study.”
1
There are continual legal challenges to the classification of those working in the on-demand economy because the work blurs the independent contractor–employee boundary (e.g., California’s AB 5/Prop 22, Massachusetts’s Suffolk Superior Court vs. Uber Technologies, Inc. and Lyft Inc.; Dubal, 2017, 2020,
). These challenges are especially relevant to closed labor market platforms, such as Lyft and Instacart, in which algorithmic management more heavily directs workers’ behaviors, as opposed to open labor market platforms such as TaskRabbit and Upwork.
2
Estimating the total number of workers in the on-demand economy is difficult (see Collins et al., 2019 and Abraham et al., 2023 for rationale). Estimates suggest that the gig economy’s workforce is growing much faster (150 percent increase 2019–2021) than the U.S. workforce as a whole (one percent increase 2019–2021) (Garin et al., 2023). Estimates range between 154 and 435 million gig workers across the globe, representing between 4.4 and 12.4 percent of the worldwide workforce (Datta et al., 2023).
3
While these labels are often used interchangeably in the literature, I use the phrase “on-demand work(ers)” as it most accurately reflects this research context. On-demand workers are a subset of gig workers (those who complete short-term work assignments such as musicians and day laborers), who are conceptually distinct from on-call workers (e.g., utility repair persons, disaster personnel). In part, this distinction exists because on-demand organizations explicitly promise to provide labor that is available on demand, and these organizations’ management systems are algorithmically engineered to have a labor buffer. This promise is not present in organizations whose workforce is primarily on-call workers; in these organizations the customers provide the labor buffer (e.g., customers wait for the next available repair person). In addition, on-demand workers and gig workers often are in independent contractor work arrangements as opposed to traditional employment work arrangements.
4
Call center work has been described using similar analogies, i.e., as an “assembly line in the head”’ (Taylor and Bain, 1999: 101), “electronic panopticon” (Bain and Taylor, 2000: 2), or “electronic sweatshop” (Fleming and Sturdy, 2011: 177). In contrast to this research, which focuses on consent and algorithmic management, much of the research on call center work has focused on workers’ resistance to digital surveillance technologies; for a review of the call center debate, see Callaghan and Thompson (2001) and
.
5
Some research has named a new term, effort, which is distinct from consent, in part because workers in the settings of these studies had shorter organizational tenures than did those in the settings in which consent was originally developed (e.g., workers in call centers tend to have shorter organizational tenures than those in manufacturing factories; Sallaz, 2015; Occhiuto, 2017). I have chosen to define consent in alignment with the broader literature for the sake of conceptual parsimony, especially given that in the U.S., organizational tenure has been declining for decades (Bidwell et al., 2013; Spreitzer, Cameron, and Garrett, 2017; Hyman, 2018; Katz and Krueger, 2019; Zgola, 2021). Moreover, due to my sampling procedure, drivers in my sample had a longer tenure (approximately 14 months) than the industry average (three to six months) (Campbell, 2018; Hall and Krueger, 2018). See the Discussion section for more on the conceptualization of consent in relation to the changes of the contemporary economy.
6
Whether workers have actual discretion over their work schedules is a source of much ongoing debate. Many of RideHail’s advertisements highlight drivers’ schedule flexibility, which RideHail argues is possible only because workers are classified as independent contractors and not as employees (see Online Appendix, Image 1). However, Dubal (2022) noted that having such schedule flexibility is not limited to those in independent contractor work arrangements. Many scholars have also argued that because many individuals in the on-demand economy are financially dependent on the work and their schedules are determined by when there is peak demand, the times that individuals work are not truly discretionary (e.g., Rosenblat, 2018; Ravenelle, 2019; Schor et al., 2020). James Parrott, the director of economic and fiscal policy at the Center for New York City Affairs at the New School, said, “It’s a fiction that the workers really have flexibility. Sure, they can go on at 2 a.m. in the morning, but they’re not going to get paid so why would they do that?” (Hu, 2022).
7
Calculating drivers’ net pay is difficult because earnings varied based on mileage rates (which varied by city), incentives offered (which varied by person as determined by the app), hours worked, the cost of operating the car (gas, insurance, depreciation, maintenance), tax rates, and RideHail’s variable service fee. Further, some cities, such as New York City, have implemented hourly minimum wages.
found that the U.S. national average after-tax wage was $11.77 per hour, which is substantially lower than the $32.06 average hourly compensation of private-sector workers and less than the $14.99 average hourly compensation of service work, the lowest-paid major occupation. These low wages, alongside opaque algorithmic management systems and the lack of employment benefits, protections, and opportunities for voice, have led many scholars to deem on-demand work, especially on closed labor platforms such as Uber and Instacart, of poor job quality.
8
Driving-related activity includes applying to, completing training for, and driving for RideHail. As a participant, I spent 100 hours doing these activities. Because ride-hailing platforms restrict driving to the state in which the car is registered, a research assistant completed the remaining hours in a different state.
9
I interviewed drivers in Ann Arbor, Michigan; Atlanta, Georgia; Austin, Texas; Baltimore, Maryland; Boston, Massachusetts; Charlottesville, Virginia; Chicago, Illinois; Denver, Colorado; Detroit, Michigan; Houston, Texas; Lewiston, Maine; Los Angeles, California; Missoula, Montana; Montreal, Quebec; New Haven, Connecticut; New York City, New York; Palo Alto, California; Port Huron, Michigan; Philadelphia, Pennsylvania; Sacramento, California; San Francisco, California; Seattle, Washington; and Washington, D.C.
10
All participants who had driven more than ten hours per month over the previous year were asked to complete a follow-up interview. Some drivers declined to participate because they were no longer driving, had moved, or were no longer interested in the study. When possible, I conducted a short exit interview about their experiences driving. I used a similar procedure in the third round of data collection.
11
I conducted one interview in French. One participant declined to be audio-recorded; in the other cases, the audio files were corrupted. Participants’ data were recorded and analyzed based on the contact summary sheet (Miles and Huberman, 1994) that was created immediately after each interview.
12
All data about income dependence, multi-homing (when workers use more than one platform for work), driver’s location, and tenure are based on the time of the first interview.
13
By far, snowball sampling was the lowest-yield sampling technique in that most drivers did not know other drivers. Three drivers had multiple members of their household who were driving. I chose not to interview more than two people from the same household.
14
Data presented in this article from postings on drivers’ discussion boards do not include the driver’s location.
15
My informants rarely used the word “algorithm” but, rather, described what the functions of the algorithms did, without explicitly naming the technology. I relied on my own experiences to determine whether an informant’s response referenced a function of the algorithmic management system. In an interesting parallel to this research, an emerging line of computational social research has begun to use the human–algorithmic interaction as a unit of analysis because this interaction provides increased statistical power (Xiong et al., 2022).
16
Platform companies did not post the exact minimum rating requirement, instead stating that the minimum average was city-specific. Drivers with ratings below a 4.6 were, however, more likely to be sanctioned. While drivers could also rate riders, these ratings did not have the same consequence. If a driver rated a rider at a three or below, they were not matched with them again; but given the large driver pool, this was not a concern mentioned in customer focus groups or message boards.
17
Leaked internal reports suggested that only 2 to 3 percent of drivers were at risk of deactivation due to poor ratings (Cook, 2015).
18
While Uber and Lyft used two different terms to refer to demand-based incentives, surges and primetime, respectively, drivers often used the term “surge” to refer to demand-based incentives on both platforms.
19
Numerous factors have been cited as contributing to the declining base pay, including drivers’ pay being decoupled from time and distance, RideHail setting individualized prices, the ever-increasing supply of drivers, and the pressure by RideHail investors to increase profit margins (Rosenblat, 2018; Isaac, 2019; Dubal, 2023b).
20
The blind matching policy is core to RideHail’s business model in that these on-demand companies promise to provide on-demand rides to customers, in any location, especially in areas traditionally underserved by other transportation services (e.g., inner-city neighborhoods; see Brown, 2018 and Cecco, 2019 for how the ride-hailing industry has increased transportation options in these traditionally underserved areas). While drivers do not need to maintain a perfect acceptance rating to maintain platform access, their overall acceptance ratings need to be quite high, generally above 95 percent (Campbell, 2018; Rosenblat, 2018).
21
Drivers typically have a 5- to 15-second window to accept an incoming ride request.
22
Given Orion’s 30-minute lock-out, this suggests that Orion was a serial deviator, since other drivers reported shorter (30-second) blocks for rejecting multiple consecutive rides. However, given his workaround strategy, Orion was still able to meet his desired goal, a higher fare.
23
Honoring confidentiality with my informants, I have not reported on any tactics that drivers did not want shared.
24
Of the four tactics, only the simple engagement tactic is in full alignment with the organization’s twofold goals, following algorithmic nudges and remaining available on the app.
25
The two informal conversations, both in 2018, were as follows. In the first informal conversation at a job talk, a RideHail employee literally followed me to my meeting with the dean, asking, “So why do the workers deviate? They should just follow the algorithm! The algorithms will make the best decisions for them.” The second informal conversation, a few months later at a conference, was nearly identical, except this time, the individual followed me into the bathroom (luckily, not into the stall). Research that includes interviews with data scientists at RideHail companies suggests an emphasis on the efficiency and optimization of the algorithmic management system (Möhlmann et al., 2021; Li, 2023). See Daub (2020) and
for a discussion of the growing prevalence of the economics-based efficiency logic in the tech industry and its shortcomings.
26
While deviance tactics, if detected, were sanctioned by RideHail’s algorithmic management system, penalties were much milder and revocable. Behaviors that constituted a withdrawal of consent resulted in immediate and permanent deactivation.
27
Related, a recent ruling by the U.S. Department of Labor under the Biden administration argues that one’s economic dependence on a company should be considered in how workers are classified (Wiessner, 2024).
28
29
Author’s Biography
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
