Abstract
Academic and public debates are increasingly concerned with the question whether and how algorithmic decision-making (ADM) may reinforce social inequality. Most previous research on this topic originates from computer science. The social sciences, however, have huge potentials to contribute to research on social consequences of ADM. Based on a process model of ADM systems, we demonstrate how social sciences may advance the literature on the impacts of ADM on social inequality by uncovering and mitigating biases in training data, by understanding data processing and analysis, as well as by studying social contexts of algorithms in practice. Furthermore, we show that fairness notions need to be evaluated with respect to specific outcomes of ADM systems and with respect to concrete social contexts. Social sciences may evaluate how individuals handle algorithmic decisions in practice and how single decisions aggregate to macro social outcomes. In this overview, we highlight how social sciences can apply their knowledge on social stratification and on substantive domains of ADM applications to advance the understanding of social impacts of ADM.
Keywords
Introduction
As the increasing use of algorithmic decision-making (ADM) has raised concerns about its social impacts and particularly about new or reinforced social inequalities, research quantifying consequences of ADM for social inequality remains in demand. Understanding the sources and effects of social inequality is one of the core competencies—and responsibilities—of the social sciences. To facilitate a cross-disciplinary discussion and additional research in this area, we use a process model of automated decision-making to highlight when and where social inequality may arise from ADM systems. Focusing on the data generation, data analysis, and implementation of ADM systems, we suggest a roadmap and research avenues for social scientists interested in answering the increased calls for the study of social impacts of ADM.
ADM is used as an umbrella term for a variety of systems that are used to assist or replace human deciders (see AlgorithmWatch, 2019). For instance, judges may use recidivism risk scores predicted by algorithms trained on decades of criminal records to determine bail decisions (Stevenson, 2018), mortgage lenders can base interest rates on default risks predicted by algorithms (Bartlett et al., 2019), and public social services may draw on algorithmic support to make decisions on financial aids (Lind and Wallentin, 2020).
ADM systems are based on predictions from models that process historical data, which contain both inputs (“predictors,” “features,” “independent variables,” “x”) and one or more outputs (“label,” “outcome,” “dependent variable,” “y”). The goal of data processing is to “learn” associations between inputs and output from the past to make predictions where the output is still unknown. Predictions are then used to decide whether some action should be taken or not. While our focus is on ADM systems that draw on some automated learning, these systems can generally vary in the complexity of how inputs determine outputs—including simple threshold rules for single input variables—, as well as in the extent to which humans are involved in the final decisions (see related definitions and surrounding discussions in AlgorithmWatch (2019) and European Parliament, Directorate General for Parliamentary Research Services et al. (2019)).
ADM seems promising as an alternative to (pure) human decision-making, as human decisions may be just as or even more biased than ADM, with ADM potentially having higher efficacy (Miller, 2018), transparency, and accountability (Mayson, 2019). However, concerns have been raised about algorithms exacerbating social inequality and discriminating against certain societal groups, for example, due to learning biases from historical training data (e.g. Zou and Schiebinger, 2018).
A recent example of an ADM system that raised such concerns is a system that has been tested by the Public Employment Service Austria (AMS). This system classifies job seekers into three groups, depending on their predicted chances to find a new employment (Lopez, 2019). The system builds groups of feature combinations based on, for example, gender, age, nationality, education, and previous contact with AMS, and predicts short- as well as long-term chances of integration into the labor market (Gamper et al., 2020). The assignment to a group can influence which kind of assistance is given to an individual: for instance, Kopf (2019) argues that while all job seekers are supported by the employment agency, individuals with low chances for re-employment would usually profit more from intensified assistance than from, for example, qualification measures. However, concerns arise if, for example, women, with all other characteristics held equal, had lower scores than men. Such concerns sparked discourse regarding the discriminatory potentials of this system (see Kopf, 2019; Lopez, 2019).
While similar decisions have been made without algorithmic assistance in the past, novel ADM systems have specific features that create new and amplify old challenges. First, these systems make use of new technical devices and facilities, unprecedented amounts of data, and novel techniques of data analysis that allow deciders to employ new decision-making strategies to approach old problems. Second, ADM systems constitute socio-technical systems that entail machines and humans (Selbst et al., 2019): they are pervaded by human decisions and cultural notions that we need to scrutinize (Seaver, 2019) to understand potential detrimental effects for society.
Scholars from various disciplines have called for examining algorithmic outcomes to avoid or mitigate undesired consequences of ADM (Kusner and Loftus, 2020; Zou and Schiebinger, 2018). Previous research from computer science (Mehrabi et al., 2019), legal studies (Wachter, 2020), and philosophical (Mittelstadt et al., 2016) perspectives discussed algorithmic, structural, and ethical problems with ADM. Joyce et al. (2021) and Liu (2021) provide a general overview of sociological perspectives on related artificial intelligence.
Drawing on previous literature, our own work on ADM systems, and a previously developed big data processing model (Weyer et al., 2018), we here highlight areas in which social scientists can (and should) use their expertise to contribute to the debate of equitable ADM. We show how a social science perspective on data generating processes, analytical challenges, and implementations can help anticipate (undesired) social impacts of ADM.
A process model of ADM
To understand how social inequality, here defined as “the unequal distribution of valued resources, opportunities, and positions among the members of a population in a given space and time” (Otte et al., 2021: 362), can arise or be amplified through ADM systems, attention needs to be paid to the distribution of opportunities and restrictions leading into and out of the ADM system. While inequality not always necessarily constitutes injustice, it is oftentimes considered an undesired property of ADM systems (see Kuppler et al. (2021) for a detailed discussion on distributive justice in ADM).
A major path via which algorithms—just like human-made decisions—may affect such distributions is discriminatory behavior. By discrimination we mean “an action or practice that excludes, disadvantages, or merely differentiates between individuals or groups of individuals on the basis of some ascribed or perceived trait” (Kohler-Hausmann, 2011), such as gender and race. Computer science research on Fair Machine Learning (Fair ML) aims to tackle discrimination by investigating how algorithms can be designed to make predictions fair, that is, without “prejudice or favoritism toward an individual or a group based on their inherent or acquired characteristics” (Mehrabi et al., 2019: 1).
Social implications of ADM systems do not only arise through biased predictions, but from implementations of decisions within a social environment. Even if fairness on the prediction level is present, disparate impact can occur (Feldman et al., 2015), for example, because impacts of ADM decisions are hard to factor into the preceding data analysis (Kusner et al., 2019), or because human deciders rely disproportionally on the ADM-based recommendations. Recent research extended the notion of fairness to include actual inequality effects resulting from algorithmic discrimination in the social context in which it is placed (see section “Data preparation and analysis—from fairness in algorithmic output to fairness in social impact”) and to frame such effects in terms of causal impact (Kasy and Abebe, 2021). To investigate these social impacts of ADM, a social science perspective becomes particularly valuable.
To discuss how ADM systems may impact social inequality, we adapt a “big data process model” (Weyer et al., 2018: 74), by breaking down the ADM process into three steps (based on Weyer et al., 2018). We discuss how social inequalities may be shaped in each step:
The different forms of biases may propagate through the ADM process and can be reinforced or mitigated along the way.
We briefly illustrate this three-step model with an example. The data basis of the Austrian AMS model that classifies individuals according to their labor market integration chances likely reflects historical unequal labor market participation rates, for example for women (data generation). The data analysis itself potentially manifests this bias, if, for example, the model resulting from training assigned women—all else being equal—lower employability scores than men (data analysis). Then, we also need to ask about the actual consequences of the system (implementation). For example, Kopf (2019) argues that women were ultimately under-represented in the lowest employability group. However, based on data reported by Gamper et al. (2020) on group assignment at the beginning of unemployment, Allhutter et al. (2020) note that the share of women was roughly double the share of men in the lowest employability group. Allhutter et al. (2020) suspect that varying conclusions may result from considering different models, time frames, or (sub-)populations. We therefore need to carefully scrutinize whether or to which extent women would be effectively disadvantaged by the system in practice. Furthermore, we need to know under which conditions human deciders adopt or disagree with the predicted scores and how job seekers subsequently change their job search behavior (also see Allhutter et al., 2020). In a feedback loop, such behavior may flow into the data basis for future model building. Finally, we need to understand how potentially discriminatory decisions on the individual level will manifest on the macro-level in the long term.
Sources of bias and social impacts along the ADM process
In the following subsections, we explore each of the potential sources of inequalities that we described in the previous section. We follow the ADM process model step by step and show how the social sciences have already contributed to researching problems related to potentially discriminatory ADM. We also identify promising research questions related to social inequality impacts of ADM that the social sciences could investigate.
Data generation—historical bias and selective participation
Algorithms can be trained on a variety of data, ranging from governmental records, such as individual labor market histories, and survey data to digital data created through individuals’ online activities and interactions with digital devices. If bias is present in the data sets used, unfair or discriminatory outputs may result. While using data to make predictions is not exclusive to algorithms, specific aspects in data generation require heightened attention in ADM. In this section, we focus on exemplary problems in biased data sets (Rodolfa et al., 2021; Sen et al., 2019) and refer to previous literature for more general introductions (e.g. Groves, 2004).
Two major sources for bias can arise in the data generation step (for a detailed overview, e.g. Mehrabi et al., 2019). The first source covers all those cases where data used to develop an ADM system contains historical discrimination. That is, an outcome is unequally distributed between individuals with different characteristics such as gender and race, after controlling for other characteristics of the individuals that cause variation in the outcome. The mechanisms creating such discrimination are manifold and depend on the concrete context, for which social sciences can provide domain-specific knowledge. In the labor market example above, historical labor market records may show that, after controlling for other individual characteristics, women had worse re-employment chances than men after losing their job in the past. Similarly, historical criminal records may insinuate that, all other characteristics of an individual held constant, black offenders had higher risks of recidivism once released from jail than white offenders.
The second source comprises biases due to selective participation and representation of social groups in data generation and collection (see Mehrabi et al., 2019). Selective participation can introduce a mismatch between the data that is used for training a prediction model and the ultimate target population that is affected in its application. If important subgroups are misrepresented in the training step, high error rates (and ultimately incorrect decisions) may result once the model is confronted with the target data in the deployment phase (Daumé III, 2017). Unequal participation in the generation and collection of digital data constitutes a particular challenge for those ADM systems that rely on them. Previous research has shown that the use of information and communication technology is often selective, for example, with respect to digital skills, age, and socio-economic status (Hargittai and Hsieh, 2013; Lutz, 2019). Models trained on such data may thus find relationships that hold only for the group of individuals using such technology. That is, individuals who are already disadvantaged because they do not use specific digital technologies could also be disadvantaged by an ADM system if the system cannot consider their behaviors and preferences (Lerman, 2013).
Social scientists are needed to identify coverage issues due to differences in social characteristics, digital skills, trust, and privacy concerns in the data collection process. Social scientists, and particularly survey researchers, can contribute to tackling representation issues of training data by applying weighting methods or improving data collections. Designing, conducting, and evaluating various forms of data collection processes such that the acquired sample resembles the target population of interest is a core task of survey research. Recent work in survey research investigates coverage and representation issues in the context of digital data and data collected via smartphones and sensors and introduces methods for adjusting non-random samples (Baker et al., 2013; Japec et al., 2015; Keusch et al., 2020). This includes, for example, pseudo-weighting approaches that allow to correct for biases due to selective participation by leveraging information from an auxiliary reference sample (Elliott and Valliant, 2017). Note that such techniques are closely related to adaptation approaches that have been proposed in computer science to account for covariate shift between training and test data (Daumé III, 2017). Weighting techniques from survey research could similarly be utilized to adjust (survey- and non-survey-based) training data if a suitable reference sample that resembles the target population can be found and both datasets include structural information about the entities of interest (e.g. socio-demographic attributes of individuals or make and type of digital devices). While applying pre-processing techniques such as re-weighting may not be feasible in all ADM contexts, recent work on post-processing predictions exemplifies how ideas from survey research (mass imputation; Yang and Kim, 2020) and computer science (multi-calibration; Hebert-Johnson et al., 2018) can be combined to tackle misrepresentation in training data (Kim et al., 2022).
In addition to historical bias and representation bias, ADM can be adversely affected by using mismeasured variables. Using proxy variables such as healthcare costs as a proxy for health needs can obscure differences in the true outcome of interest when, for example, black individuals generated lower healthcare costs than white individuals once the true health status is held constant (Obermeyer et al., 2019). Such measurement bias can be directly connected to social science work on measurement errors and thus represents one example of how social science already contributes to researching social impacts of ADM (Boeschoten et al., 2020; Jacobs and Wallach, 2021). Moreover, the contextual nature of some individual characteristics and behaviors may not be amenable to quantification and therefore, ADM system cannot cover these characteristics appropriately, such as context-sensitive combinations of protected attributes relating to intersectional discrimination (Mann and Matzner, 2019). These may be only subtly present in social interactions, lead to discrimination, and be insufficiently captured in automated analysis (see section “Data preparation and analysis—from fairness in algorithmic output to fairness in social impact”).
To conclude, biases in datasets gain renewed momentum in the context of ADM for three reasons. First, it is likely that the increased quantity of predictions and decisions that a model can make compared to a human decider will intensify inequalities that are already present in the data. Second, relying on ADM systems increases the importance of patterns in the data in comparison to the importance of heuristics of human decision-makers (but see section “Implementation—micro-interaction with ADM and macro-social outcomes”). Third, the amount of data produced and used in ADM systems has considerably increased with the advance of digital technologies. Social sciences can apply methodological and domain knowledge (a) to better understand how situation-specific biases may be present already in the data collection stage of ADM processes and (b) to explore how advances in survey methods can be used to correct such biases.
Research avenues:
How can we utilize methodological advances in survey research to correct biases in data due to selective participation and improve the data input for ADM? How can we extend research on digital divides and technical competencies to study inequality in being covered by ADM systems (Lutz, 2019)?
Data preparation and analysis—from fairness in algorithmic output to fairness in social impact
Data preparation and analysis is the step in which developers work with data and construct and refine algorithms. This process entails manifold interpretations and decisions, including, ideally, considerations on how to produce fair outputs. In this section, we outline how algorithms may produce biased predictions due to biased data or decisions during the modeling process. We give a brief survey on the mainly computer scientific research field of Fair ML. Then, we show how a social science perspective can contribute to the identification of meaningful fairness criteria in social contexts, particularly when considering social impacts of ADM systems on macro-level social outcomes and public perceptions of fairness.
Approaches in fair machine learning
Fair ML is a research field that investigates the fairness of machine learning algorithms. This research branch produces important contributions by proposing fairness metrics and improving algorithm design such that individuals are less likely to be assessed by characteristics that should not matter for taking a decision (“protected attributes”). Such steps are necessary as otherwise, algorithms might reproduce existing biases or exacerbate inequalities even when the data sources are unbiased (Aghaei et al., 2019). For example, this is the case when prediction error rates differ between groups (Rodolfa et al., 2021). Various steps in the construction of variables (“feature engineering”), such as how race is coded, may also introduce biases (Rodolfa et al., 2021).
Fairness definitions oftentimes are formal measures based on rates of correct and incorrect predictions for individuals of different social groups for which non-discrimination should be ensured (Corbett-Davies and Goel, 2018). For instance, an algorithm might be tasked with assigning job seekers into two classes: those with high or low chances of finding a new job. The algorithm could be trained with data that show past job market outcomes of job seekers. The algorithm tries to combine the characteristics of these individuals to build a model that predicts chances of labor market integration as accurately as possible. The prediction outputs can be evaluated by comparing the predicted with the observed outcomes in the data.
Several fairness definitions specify how error rates should be balanced across different groups of individuals. As an example, an algorithm may be considered fair if it results in equal false negative rates (equal opportunity; Hardt et al., 2016) or equal false positive rates (predictive equality; Rodolfa et al., 2021) between members of different groups (e.g. men and women). A related definition is equalized odds, which means that members of different groups experience both false negatives and false positives at the same rate (Hardt et al., 2016). This principle can be applied to various error metrics and their combinations (e.g. false discovery rates, false omission rates, accuracy), resulting in a variety of group-based fairness notions. Furthermore, subgroup fairness (Hebert-Johnson et al., 2018) and individual fairness (Dwork et al., 2012) notions have been proposed that expand beyond comparisons of error rates on the group level (e.g. by considering intersections of gender and race).
Research in Fair ML resulted in various methods and tools that may mitigate biases at different stages of the modeling pipeline (Berk et al., 2018; Mehrabi et al., 2019). Pre-processing techniques can be used to eliminate sources of unfairness in the data prior to model training, for example, by removing dependencies between legitimate factors and protected attributes (Johndrow and Lum, 2017). In-processing techniques aim at modifying the model building process itself, for example, by introducing fairness constraints in the objective function (Berk et al., 2017). Post-processing methods may be used to alter the output of a prediction algorithm after model training, for example, by “nudging” predictions towards the true outcome for subgroups where high errors are observed (Kim et al., 2019). These procedures have been shown to mitigate different notions of unfairness at the prediction stage of the ADM process in several applications (Friedler et al., 2018).
Competing fairness definitions and the importance of social context
Many fairness definitions and correction methods have been proposed, and it may prove difficult to choose the definition and technique that is the most appropriate for the given prediction task (see Makhlouf et al., 2020). Moreover, some fairness definitions were found incompatible with each other and in conflict with overall accuracy (Berk et al., 2018), while Selbst et al. (2019) discuss as “formalism trap” whether an appropriate mathematical definition of the complex concept of fairness was even possible.
One major concern in handling fairness boils down to the question: is it better to ignore specific individual characteristics such as gender or race altogether, or should we try to balance, for example, error rates between groups based on these features (Corbett-Davies and Goel, 2018)? Neglecting group membership may, for instance, lead to aggregation bias, meaning that one model is used for all groups although the model works worse for some of the groups (Suresh and Guttag, 2020). In the case of race, Benthall and Haynes (2019) discuss that ignoring race would still possibly lead to racial discrimination as effects of correlates relevant to race and inequality persisted. However, explicitly including race reified this category. Instead, they propose a third alternative of algorithmically finding latent categories that mirror racial segregation. From a social science perspective, this discussion extends to the question which features are considered protected attributes in the first place. While gender and race represent attributes that are commonly considered sensitive and are protected by legislation, sociological research on the intergenerational transmission of resources and education (e.g. van Doorn et al., 2011) raises questions on which concepts purely measure individual merit and which attributes may constitute “hybrid” characteristics that are (at least partly) socially inherited. Relatedly, using traditional concepts of gender and race for defining protected groups will fail to account for individuals who do not find themselves represented by those categories. Particular attention needs to be paid to intersectional discrimination that may disadvantage individuals based on multiple protected attributes at the same time, for example, gender and race: automated analysis of large data bases may contain a plethora of potential protected attributes, suggest new associations between these attributes, and thereby statistically form new groups of people that may then be discriminated against (Mann and Matzner, 2019). Social scientists can scrutinize data, analytical decisions, and outputs with respect to intersectionality in different contexts of ADM applications, and suggest groupings of protected attributes that are contextually relevant.
Eventually, fairness is context-specific. Among others, the choice of a fairness metric may depend on the outcome and which resources will be distributed (Kuppler et al., 2021). The idea of social context is not new in the realm of computer science (Selbst et al., 2019) and is part of pursuing “algorithmic realism” (Green and Viljoen, 2020). For instance, this entails the question whether a system aims at helping or punishing individuals, which implies an emphasis on disparate distributions towards either false negatives (incorrectly excluded from a positive intervention) or false positives (incorrectly included in receiving a negative intervention) (Saleiro et al., 2019). This discussion extends to the broader question on the just or desired allocation principle in a specific ADM application context. Sociological discourse on distributive justice can enlarge computer science's decision space when it comes to designing allocation systems and selecting bias correction techniques by highlighting which design choices may serve which principle (Kuppler et al., 2021).
Empirical findings on fairness perceptions
Fairness perceptions matter to ADM development for two reasons. First, they are relevant to design socially acceptable ADM systems. Second, the individual evaluation of an algorithm may contribute to how that individual interacts with and acts upon the decision of the ADM system, thereby potentially shaping inequality outcomes.
A comprehensive literature review on fairness perceptions on ADM concludes that perceptions strongly depend on context characteristics, such as the features used by the algorithm and the purpose of the algorithm (Starke et al., 2021). Participants in one study applied some justice principles relevant for human decision-making also to algorithmic decisions, but the concrete style of explaining the algorithm impacted justice perceptions only when the respondent was exposed to multiple styles (Binns et al., 2018). In addition, general trust in ML systems and the features used and not used are relevant for fairness judgments (Dodge et al., 2019). Empirically validated frameworks that define process features relevant to fairness perceptions (Grgić-Hlača et al., 2018) can build the basis for practically applicable guidelines for designing contextually fair algorithms, which is why such work is particularly attractive for future work.
Whether and how individual characteristics such as socio-demographic attributes like age and gender interact with, for example, explanation styles and the impact of the decision situation needs further research and possibly depends on individual affectedness (Pierson, 2018). Experimental evidence suggests that fairness ratings depend on whether respondents’ characteristics are involved in the algorithmic decision, and conservatives were found to be more accepting of using individual characteristics in computer-assisted bail decisions than liberals (Grgić-Hlača et al., 2020).
In conclusion, building fair algorithms is a prerequisite for arriving at fair predictions and, subsequently, decisions. Software toolkits that assist in assessing the fairness of algorithms are available (Bellamy et al., 2019; Saleiro et al., 2019). To advance Fair ML, we can intensify research on fairness perceptions in concrete ADM processes and strengthen the link between distributive justice principles and (fairness in) automated allocation systems (Kuppler et al., 2021). Moreover, Starke et al. (2021) suggest systematizing situation-specific factors such as whether a decision is high-stake or low-stake and the area of application (e.g. decisions in the criminal justice system or hiring) that may shape fairness perceptions.
Research avenues:
How do contextual information (the purpose of an algorithm) and explanations of algorithm function shape fairness perceptions of ADM processes? How do individual characteristics influence fairness perceptions? How can fairness assessment and mitigation techniques be implemented and extended beyond equalizing error rates towards serving context-specific allocation principles? How can social science provide domain-specific knowledge to define appropriate, non-discriminatory outcomes for an ADM system, including the consideration of externalities?
Implementation—micro-interaction with ADM and macro-social outcomes
Researchers from different disciplines have demonstrated that the used data and the data analysis performed do not suffice for explaining the social impacts of algorithms (Cowgill and Tucker, 2020; Kleinberg et al., 2018). The question whether the use of an algorithm will produce fair outcomes is not only a question of the fairness of predictions and decisions, but also of their actual impacts in a social environment (Kusner et al., 2019). In fact, “[…] even fair decisions at the machine learning level may not lead to equitable results in society and the decision-making process may need to compensate for these other inequities” (Rodolfa et al., 2021:304).
The notion of disparate impact helps to understand the difference between the output of an analysis and subsequent societal consequences. Disparate impact refers to effects of practices that result in unintended disadvantages for groups of individuals with certain characteristics (Barocas and Selbst, 2016). For example, even if no discrimination is intended, individuals may be affected differently due to their characteristics. Implementing notions of disparate impact in algorithms is one step to practically achieve fairer results (e.g. Feldman et al., 2015), and computer scientific research developed and applied such extended notions of fairness. Among those are suggestions to ascertain fairness by optimizing how an outcome of interest is expected to be affected in the long term (Liu et al., 2019), choosing fairness metrics that satisfy specific policy goals (Rodolfa et al., 2020), and engaging with the needs of affected population groups to adjust analyses in feedback loops (Noriega-Campero et al., 2018).
These outcome-oriented approaches seem most promising for the development of an encompassing understanding of fairness that contributes to contextually appropriate assessments. A social science perspective can help to analyze the implementation process of ADM in social contexts and to understand interaction processes at the micro-level between algorithms, affected individuals and, in some cases, human deciders, and their macro-social outcomes.
Human versus algorithmic predictions: Empirical evidence from real-life cases
Studying impacts of ADM systems in real-life cases faces the same challenge as other observational social research: it remains unclear what the outcome would have been had a decision been taken without an algorithm (see Holland, 1986). Although methods for tackling such problems of causal inference are well known to social scientists, there is so far only little research applying them to the study of social impacts of ADM (such as Cowgill and Tucker, 2017). One notable exception are recidivism prediction algorithms in the USA, where studies find mixed effects regarding the reduction of crime and racial disparity through algorithms (Berk, 2017; Kleinberg et al., 2018; Stevenson, 2018). Stevenson (2018) suggests that even if algorithms made better predictions, they might not necessarily improve relevant outcomes, and judges’ own biases could lead to a sub-optimal use of algorithmic predictions, emphasizing the need to study how human decision-makers rely on algorithms.
Previous research reports mixed findings regarding differences in the accuracy of predictions between algorithms and human deciders. Some find that humans perform worse than algorithms (Green and Chen, 2019), while others find comparable accuracy and fairness in predictions (Bansak, 2019; Dressel and Farid, 2018; Tan et al., 2018). In addition, in the context of recidivism prediction tasks, it is likely that the characteristics of the defendants will matter: given a risk assessment, human deciders deviated more strongly to unfavorable predictions for black defendants than for white defendants (Green and Chen, 2019).
Overall, the question whether an algorithm can outperform a human decider will have to be evaluated on a case-by-case basis. In those cases where the final decision remains in the hands of a human decider, we also must consider whether and how a human decider is involved and influenced by an algorithmic decision or recommendation.
From “automation bias” to “algorithmic aversion”—How human deciders (do not) adopt algorithmic recommendations
Many ADM systems, particularly those in which the stakes are high, involve a human decider who may consider algorithmic predictions in her decisions. While a machine-assisted decision may deviate from a purely human decision, human deciders will not always follow the algorithmic recommendation. Therefore, potential biases inherent in the algorithmic prediction may be alleviated or corrected by human deciders, but humans may also introduce or reinforce discrimination in the process. The interaction of a human decider with an ADM system is likely complex and requires detailed investigation. Research on “human factors” and human-computer interaction provides valuable work that can be applied to the study of ADM systems (Zerilli et al., 2019). The communication between algorithmic recommendations, human deciders, and affected individuals is likely shaped by the complexity of the underlying model. If we want the involved individuals to understand how ADM systems arrive at decisions and to uphold accountability, we need algorithms that can be explained—either by making use of inherently interpretable methods or by employing post-hoc interpretation techniques (Molnar, 2019). Differential social impacts may arise, for example, if explanations are differently effective for social groups and shape the reliance on or compliance with algorithmic recommendations.
Here, we focus on the specific problem of circumstances under which a human decider will be more likely to adopt (or override) an algorithmic recommendation. Two central phenomena characterize human reliance on algorithmic predictions: automation bias and algorithmic aversion. Automation bias refers to errors stemming from human reliance on automated systems such as ADM: while errors of omission refer to cases where someone relies on a flawed algorithmic prediction (false negatives), errors of commission refer to falsely assuming an error (false positives) (Wickens et al., 2015).
Empirical evidence for automation bias has been found, for example, in clinical decision support systems (Goddard et al., 2012). Research shows that factors such as trust and own experience shape reliance on automated systems (Burton et al., 2019; Cepera et al., 2018; Lee and See, 2004; Logg et al., 2019; Weyer et al., 2018). Moreover, Parasuraman and Manzey (2010) note that errors of commission are lower when a system serves information integration and analysis as compared to providing concrete recommendations for actions.
There also is evidence for algorithmic aversion, that is, individuals becoming less likely to rely on algorithmic predictions after experiencing false predictions (Burton et al., 2019). Experimental studies show that confidence in algorithms is lowered when algorithms make a mistake (Dietvorst et al., 2015). Moreover, humans tend to adjust their predictions more often based on human advice than based on statistical forecasting (Önkal et al., 2009). However, Grgić-Hlača et al. (2019) report that machine advice does affect participants’ predictions in the case of criminal recidivism and Araujo et al. (2020) even find evidence for algorithmic appreciation, that is, a preference for automated decisions compared to human decisions.
Empirical studies of actual adoptions of algorithmic recommendations and consequences for inequality are scarce and mostly investigate the judicial context. Results show that higher recidivism scores lead to longer sentences but judges also seem to rely less on risk scores over time (Stevenson and Doleac, 2019). If risk scores are transformed into a categorical scale (low, medium, and high risk), individuals who are placed just above a threshold value receive on average one to four additional weeks of detention before trial compared to those placed just below the threshold (Cowgill, 2018). Again, individual characteristics seem to play an important role as this effect was more pronounced for black defendants than for white defendants.
Social sciences add domain-specific knowledge and tools for understanding macro-level outcomes of human-ADM interactions
Social sciences contribute to developing fair ADM systems by bringing in their domain-specific expertise on individual behavior and social practices across social environments. A thorough analysis of ADM impacts requires such domain-specific knowledge, for example, on labor market behavior. For instance, social sciences can help to answer questions such as: how will an individual adjust her behavior when an employment agency employee decides for a specific (or no) training program based on an ADM recommendation? Could this decrease motivation as an individual feels more constrained by algorithmic decisions than by human decisions? First research documents how individuals evaluate algorithmic decisions compared to human decisions, finding both similarities and differences (Araujo et al., 2020; Binns et al., 2018; Plane et al., 2017; see section “Data preparation and analysis—from fairness in algorithmic output to fairness in social impact”). Due to potential context-dependency, more research is needed to gain a better understanding of human interpretations of algorithmic decisions.
Social sciences can help to predict and to assess outcomes of ADM processes by providing domain-specific knowledge in the fields of the ADM application (Bertelsmann Stiftung, 2020). This includes knowledge on which goals human decision-makers may follow, which factors they consider, and how these differ from the purely ADM process (see Kleinberg et al., 2018). This also entails how characteristics of the individual and its environment shape the severity of the impact of a decision derived by an ADM (see Abebe et al., 2020). Moreover, as institutional and organizational contexts may react to the implementation of ADM systems in (unintended) ways (Selbst et al., 2019), social sciences also provide methods and previous research to understand established practices in specific contexts and anticipate potential reactions. These methods can also be used to investigate established practices that shape ADM implementation. In the case of comparing algorithmic and human decisions, understanding the goals programmed into an algorithm and analyzing the goals human deciders consider when taking a decision is crucial (Kleinberg et al. 2018; Stevenson 2018).
Additional to in-depth case studies that investigate ADM in concrete contexts (e.g. Elish and Watkins, 2020), experimental research and observational studies along the lines of the research presented in this section improve our understanding of interactions within ADM systems. To show whether and how algorithmic literacy and subsequent behavior impact social inequality, we need to study how these competencies, awareness (Gran et al., 2021), and knowledge related to algorithms are distributed across social groups—for example, by age and education (Fischer and Petersen, 2018)—, and then how this knowledge translates into behavior (e.g. adjusting to the algorithm's “preferences,” see Freeman Engstrom et al. 2020).
Furthermore, social scientists can contribute to investigating how individual decisions made by ADM systems influence inequality and discrimination on the societal macro-level, that is, how single decisions accumulate to overall patterns of inequality in a population akin to the micro-macro model of sociological explanation (Coleman, 1994). Agent-based modeling (ABM) is a promising method to study how interaction on the micro-level produces macro-outcomes as it allows researchers to simulate, for example, interactions of technical and social elements of an ADM process (Gilbert, 2008). ABM could be used to model an interactional setting with three types of agents: affected individuals, algorithms, and deciders. Each affected individual has, for example, certain demographic characteristics and attitudes towards technology. Results of algorithmic predictions based on different fairness strategies can be presented to the decider. The human decider—if applicable—may consider the algorithmic decision and the affected individual's characteristics to arrive at a decision and weigh both according to her own experience, for example. The affected individual may then adapt her behavior according to her characteristics and the decision.
ABM presents many advantages as it allows researchers to represent the interplay of human and machine actors (see Calero Valdez and Ziefle, 2018) in ADM systems and dynamics over time. For example, fairness implications may only show when considering long-term effects on macro-outcomes in the population (Heidari et al., 2019; Liu et al., 2019), and simulations can be run for hundreds or thousands of rounds. Furthermore, ABM responds to calls for a stronger integration of the social environment of ADM systems to grasp their impact appropriately. ABM has already been used to study the governance of socio-technical systems (Adelt et al., 2018), and Cruz Cortés and Ghosh (2019: 3), for example, apply ABM in the context of criminal recidivism risk for a “[s]ystematic analysis […] [which] implies analyzing the data generating process, the decision-making stage, and its consequences all under the same framework.” In conclusion, simulations are promising tools to assess macro-level outcomes of ADM applications from a social science perspective.
Research avenues:
How do individuals adapt behavior preemptively or as a reaction towards an algorithmic decision? Which individual resources affect interactional behavior? Which situational and individual characteristics determine reliance on ADM systems across social contexts? How do these individual decisions and interactions aggregate to macro-social inequality outcomes, and how can researchers study such impacts using simulation techniques?
Conclusion
Synthesizing several theoretical and empirical advances in the research on the consequences of ADM systems for social inequality, this paper provides an overview geared towards social science research, with a focus on data generation, analysis, and implementation challenges. For each part of the ADM pipeline, we highlighted possible inequality issues and how social sciences can contribute to their study. Put briefly, (1) the data used may be biased, (2) the algorithm itself might rely on contextually problematic conceptualizations and formalizations of fairness—or may not consider fairness at all—, and (3) the inequality outcomes depend on concrete interactional settings that can result in cumulative disadvantages, particularly for those who have been historically disadvantaged. We summarize potential sources of inequality, related social science topics, example papers, and research avenues in Table S1 (in Supplementary Materials).
Social sciences can draw on established research to contribute to these efforts by bringing in expertise on methods, concrete social contexts, and human (inter)action to investigate how ADM systems affect (macro-)social inequality outcomes. To study algorithmic bias, social scientists can contribute to developing context-aware fairness notions, and to evaluate the scale of actual impacts that ADM systems produce in practice.
Social science research on inequality and ADM systems as well as interactions between algorithms and humans goes far beyond what we were able to cover here (e.g. Joyce et al., 2021; Liu, 2021). Other challenges range from, for example, accounting for the agency of algorithms (Lange et al., 2019), social and political challenges with respect to regulation (Mittelstadt, 2019), privacy (Anthony et al., 2017), and governance (Danaher et al., 2017), to artificial intelligence shifting power relationships (Kalluri, 2020), or other social impacts beyond inequality outcomes. Moreover, we need to understand the contexts in which ADM are applied, including established practices and interactions between human and technical elements. To this end, researchers can draw on a variety of qualitative approaches, such as ethnography (see, e.g. Lange et al., 2019) within the respective social contexts or expert interviews with individuals involved in ADM implementation.
Finally, “impacts” of ADM on social inequality do not necessarily equal to increases in disparities. Human decisions are oftentimes also biased and flawed, and algorithmic decisions could potentially display less bias than humans (Mayson, 2019) and reduce social inequality overall. However, social implications need to be thought of when designing and implementing ADM applications. We hope that this paper will assist in the development of a research framework and that it will help to enhance concrete guidelines for creating socially responsible ADM systems. Such guidelines are currently discussed and urgently needed as the supervision, assessment, and even necessity of approval of ADM is an ongoing policy debate (e.g. AlgorithmWatch, 2019).
Supplemental Material
sj-docx-1-bds-10.1177_20539517221089305 - Supplemental material for Social impacts of algorithmic decision-making: A research agenda for the social sciences
Supplemental material, sj-docx-1-bds-10.1177_20539517221089305 for Social impacts of algorithmic decision-making: A research agenda for the social sciences by Frederic Gerdon, Ruben L Bach, Christoph Kern and Frauke Kreuter in Big Data & Society
Footnotes
Acknowledgments
We thank the participants of the doctoral colloquium of the Center of Doctoral Studies in the Social and Behavioral Sciences (Sociology) at the University of Mannheim, the members of the Kreuter-Keusch research lab, as well as the anonymous reviewers for their helpful feedback and comments on the paper. We acknowledge funding from the VolkswagenStiftung for the project “Consequences of Artificial Intelligence for Urban Societies” (CAIUS) and the Baden-Württemberg Stiftung for the project “Fairness in Automated Decision making” (FairADM). The publication of this article was funded by the Mannheim Centre for European Social Research (MZES). This work was supported by the University of Mannheim's Graduate School of Economic and Social Sciences. FG led the development of the paper, RB and CK contributed to research and writing, FK conceptualized the underlying research projects and contributed to writing.
Declaration of conflicting interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by the Baden-Württemberg Stiftung, Volkswagen Foundation; The publication of the article was funded by Mannheim Centre for European Social Research (MZES).
Supplemental material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
