Abstract
Public debate about content moderation has overwhelmingly focused on removal: social media platforms deleting content and suspending users, or opting not to do so. However, removal is not the only available remedy. Reducing the visibility of problematic content is becoming a commonplace element of platform governance. Platforms use machine learning classifiers to identify content they judge misleading enough, risky enough, or offensive enough that, while it does not warrant removal according to the site guidelines, warrants demoting them in algorithmic rankings and recommendations. In this essay, I document this shift and explain how reduction works. I then raise questions about what it means to use recommendation as a means of content moderation.
It is easy to assume that, in the current debates about how social media platforms moderate problematic content, content moderation is content
While removal may be the most visible response, it is by no means the only remedy available. Besides being able to (1) remove content and users, platforms can also (2) implement age barriers, geo-blocking, or temporary holds, to keep problematic content away from some users some of the time; (3) append fact-checking labels and interstitial warnings, to alert users to problematic content before or as they encounter it; (4) impose demonetization and punitive strike systems, as disincentives for producing problematic content; and (5) provide counter-speech and model preferred norms, to raise the overall level of discourse (Goldman, 2021). Some of these strategies are not as visible or controversial as removal; others may go unnoticed because they are imposed by different product teams, intervene at different points in the platform, are more difficult for users to identify in practice, or get justified in different ways.
This essay focuses on yet another type of remedy: when platforms (6) reduce the visibility or reach of problematic content. Many social media platforms have quietly begun to identify content that they deem not quite bad enough to remove. The offending content remains on the site, still available if a user can find it directly. However, the platform limits the conditions under which it circulates: whether or how it is offered up as a recommendation or search result, in an algorithmically generated feed, or “up next” in users’ queues. This is not a new policy or technique, exactly, though it is being deployed to new ends, for an expanding list of reasons. Reducing the visibility of risky, misleading, or salacious content is now used by many platforms to curb any content they deem as nearly violating their rules, across nearly all the traditional Trust & Safety concerns.
YouTube and Facebook, which have most clearly acknowledged these techniques, call them “borderline content” policies. I am reluctant to adopt that term: using “borderline” as a pejorative has quiet resonances with assumptions too often made about both geographic borders and “borderline” mental health conditions, in ways I do not want to reify. Critics have called this “shadowbanning”
1
or “suppression”
2
—I find these terms more compelling, though I worry, along with Cotter (2021), that they may inadvertently help platforms dodge the very criticism being leveled at them. I will use
That there is not yet a settled industry term is telling. Platforms are, understandably, wary of being scrutinized for these policies—either for being interventionist and biased, or opaque and unaccountable. Some platforms have not acknowledged them publicly at all. Those that have are circumspect about it. It is not that reduction techniques are hidden entirely, but platforms benefit from letting them linger quietly in the shadow of removal policies. So, despite their widespread use, reduction policies remain largely absent from news coverage, debate, policymaking, and even much of the scholarly conversations about content moderation and platform governance.
My aim is not simply to ask whether these techniques are good or bad, or suggest how they could be improved or regulated. I want to argue that reduction techniques should be included within a broadened definition of “content moderation,” not adjacent to it. Content moderation and algorithmic recommendation are mostly treated as different platform functions, handled by different product teams, and scrutinized by different critics according to different measures and concerns (Caplan, 2019). However, if we remain analytically agnostic about these distinctions, we might instead see both content moderation and algorithmic recommendation as governance—by different means, deployed in different ways, with different justifications.
First, I will define reduction techniques and identify how they are implemented by several of the major social media platforms. Then, I will examine how the platforms explain and justify these interventions. I will explain how reduction works and offer a typology of reduction techniques. I will then make a case for why reduction must be understood as part of a broader content moderation project, and how doing so changes how we think about content moderation and platform governance.
A Methodological Note
For this research, I examined company policies, corporate communications, and some of the technical literature coming from both academia and industry. However, I also lean on conversations I have had with platform employees in Trust & Safety and adjacent areas—a dozen or more representatives from five major social media platforms. Let me make a quick note on my access to and use of these conversations.
Many Silicon Valley technology companies are reluctant to grant researchers access, especially for qualitative inquiry into the company’s decision-making. As Bonini and Gandini (2020) note, more than the technology, it may be the industry itself that is “black boxed.” This reluctance protects powerful institutions from scrutiny (Monahan & Fisher, 2015, p. 710). Nondisclosure agreements (NDA), once a way to protect trade secrets, are now a ubiquitous and routine part of employee contracts and encounters with interested outsiders, and help manage scrutiny and suppress criticism (Starr, 2020). These agreements are shaping how technologists think about those who come asking interested questions, as a matter of secrecy (Seaver, 2016).
However, this guarded industry also regularly trades on informal intra- and inter-company consultations. Managers seek out informal advice from their industry peers; consult with nonprofit, advocacy, and academic groups; attend conferences on pressing issues; and orchestrate one-off conversations with people elsewhere in their own company, or at adjacent ones. When these interactions cross the outer membrane of the corporation, sometimes they are governed by an NDA as a matter of course. However, often they are not, depending instead on informal relations and mutual trust.
I find myself with a bit of this “insider access,” where different questions can be asked, and are more likely to be answered. However, the instant I would identify a conversation as a research interview, no matter how informally I did so, that performative gesture often triggered a nervousness that any “interview” would need to be under NDA. Formality seemed to invoke a matching formality: mine an Institutional Review Board (IRB) consent process, theirs an NDA. I felt an NDA would be a substantive barrier to talking frankly and publicly about these policies later on. So instead, I promised informants that our conversations would be treated as background. I fully acknowledge that there is a privilege to my position—not so much from being employed by a technology company myself, ironically, which in some ways complicates my access to other companies—more from the working relationships developed over the course of previous research, aided by presumptions my informants may make about me as “working in industry” like them. These are not conversations that another researcher could necessarily have, and I recognize the methodological problem that raises.
Any details I learned, I tried to verify using that company’s published policies and statements. In the article, I use published evidence to present details I may have first learned in confidence. For anything else, I have not disclosed specific details, though they inform my argument. Even so, I remain concerned that these routinized logics of secrecy and informality, especially in Silicon Valley, are a problematic barrier to qualitative inquiry, and that our ethical commitments may at times hamper us from investigating and challenging the strategies of the powerful (Souleles, 2021). I am keenly aware that my efforts to know these companies are entangled in their use of me. In these conversations, am I finding out things from them, or are my questions a form of free consulting? Both, I suspect. And like other researchers, I function within the “contact zones that media industries stage carefully around us as we study them” (Caldwell, 2013, p. 165, cited in Vonderau, 2014). Still, I am unwilling to let that prevent me entirely from revealing and explaining policies these platforms continue to publicly obscure.
“Borderline Content”
YouTube announced its “borderline content” policy in a January 2019 post, though the practice had already been in place for a few months or more. This followed on the heels of nearly 2 years of blistering criticism, not only for hosting problematic content but also for amplifying it with its recommendation system. Ex-YouTube designer Guillaume Chaillot specifically blamed the recommendation algorithm for the glut of misinformation, warning that “fiction is outperforming reality”;
3
Zeynep Tufekci, commenting on a
It is worth lingering for a moment on how YouTube framed this new intervention. The title of the post, “Continuing our work to improve recommendations on YouTube,”
5
gives no indication that this is in any way a Trust & Safety concern. The post begins with a lineage of adjustments YouTube had already made, including limiting “clickbaity videos with misleading titles and descriptions,” and “getting too many similar recommendations,” part of the “hundreds of changes to improve the quality of recommendations for users” YouTube had made just in the past year. Then, the new policy is introduced as if it isn’t new at all: We’ll continue that work this year, including taking a closer look at how we can reduce the spread of content that comes close to—but doesn’t quite cross the line of—violating our Community Guidelines. To that end, we’ll begin reducing recommendations of borderline content and content that could misinform users in harmful ways—such as videos promoting a phony miracle cure for a serious illness, claiming the earth is flat, or making blatantly false claims about historic events like 9/11.
Harms are presented as falling along a spectrum: “content that comes close to—but doesn’t quite cross the line of—violating our Community Guidelines.” In this spatial understanding of harm lies a “borderline,” just left of the existing prohibitions (Maddox & Malson, 2020).
Notice that the problems identified are traditionally moderation issues—misinformation and conspiracy—but the tactic is drawn from recommendation quality: clickbait, repetitive suggestions, and so on. While discussions of online harm tend to highlight victims and perpetrators, quality interventions tend to treat users as consumers, and are more concerned about customer satisfaction than the public interest.
YouTube then quickly assures the reader that: To be clear, this will only affect recommendations of what videos to watch, not whether a video is available on YouTube. As always, people can still access all videos that comply with our Community Guidelines and, when relevant, these videos may appear in recommendations for channel subscribers and in search results. We think this change strikes a balance between maintaining a platform for free speech and living up to our responsibility to users.
Reduction techniques depend on this kind of demarcation, between what is hosted and what is recommended, between archive and algorithm. Reduction only pertains to where content is offered up, suggested: not where it lives in the archive (Gillespie, 2016). 6
YouTube later incorporated this strategy into a broader approach to governance that the company called “The Four Rs”: remove, raise, reward, and reduce. By late 2019, YouTube PR could assert that the effort to “reduce” borderline content and harmful misinformation was working: “The result is a 70% average drop in watch time of this content coming from non-subscribed recommendations in the U.S.” 7 —though as Lewis noted, such claims are nearly impossible to confirm, especially given that the category itself is defined by YouTube’s efforts to identify it. 8
Facebook and Instagram had already announced their reduction policy in May 2018,
9
after also enduring 2 years of criticism for amplifying misinformation.
10
The terminology and justifications are similar to YouTube’s: There are other types of problematic content that, although they don’t violate our policies, are still misleading or harmful and that our community has told us they don’t want to see on Facebook—things like clickbait or sensationalism. When we find examples of this kind of content, we reduce its spread in News Feed using ranking . . .
11
Facebook later published a detailed list of what pages, groups, or events they will not recommend 12 and an exhaustive “Content Distribution Guidelines” indicating what they will “demote” in the News Feed. 13 Facebook’s list includes three categories of concern, two of which echo YouTube’s: borderline content and harmful misinformation, but also low-quality junk. Facebook will not recommend “content that users broadly tell us they dislike,” meaning clickbait, “engagement bait,” contest giveaways, and links to deceptive or malicious sites; or “content that is associated with low-quality publishing,” including unoriginal or repurposed content, news that’s unclear about its provenance, or, content enjoying a surge of engagement on Facebook that’s unmatched on the wider web. Including this in their reduction strategy, just like when YouTube frames theirs in the legacy of clickbait, further confirms that this is the aim of content moderation using the tools developed in the pursuit of recommendation quality.
In July 2021, Instagram introduced “sensitive content control,” which allowed the user to adjust the degree to which sensitive content should be filtered out of the “explore” recommendations the platform offers. 14 While the announcement emphasized the agency users were being given, the fact that users could now “allow,” “limit (default),” or “limit even more” sensitive content revealed that such content already was, by default, being reduced. The company did not immediately specify what counted as sensitive; Instagram head Adam Mosseri later indicated that the intervention focuses on “sexually suggestive, firearm, and drug-related content” 15 and was separate from parallel efforts to reduce misinformation and self-harm.
Twitter, LinkedIn, and TikTok have similar reduction strategies already in place, though they have been less vocal about them. In a July 2020 promise to address COVID-19 vaccine misinformation, Twitter asserted, “Tweets that are labeled under this expanded guidance will have reduced visibility across the service . . . however, anyone following the account will still be able to see the Tweet and Retweet.”
16
They used similar language to describe their approach to election misinformation in 2020 and 2021.
17
LinkedIn has acknowledged that, in response to community guidelines violations, they may “limit the visibility of certain content, or remove it entirely.”
18
That language first appeared in LinkedIn’s “Professional Community Policies” when they were substantially rewritten in December 2019.
19
TikTok has been even coyer, perhaps because their widely praised recommendation engine is so central to their service. However, in the community guidelines, the company acknowledges that for some content—such as spam, videos under review, or videos that could be considered upsetting or depict things that may be shocking to a general audience—we may reduce discoverability, including by redirecting search results or limiting distribution in the For You feed.
20
Depending on how we broaden the definition, we can find other platforms engaged in reduction strategies that share at least a family resemblance. Tumblr, which once took a more permissive approach to sexual content (Tiidenberg et al., 2021), used to limit the circulation of explicit content by refusing to serve up search results to queries it judged explicit. Users could post pornographic images, could even tag them as “#porn”—but if a user searched for “#porn” no results would be returned. Hashtag blocking has its problems (Gerrard, 2018; Pilipets & Paasonen, 2022; Sybert, 2021). However, like reduction, it similarly demarcates between hosting content and offering it up in search results or recommendations.
Reddit’s “quarantine” policy, though structured differently (Chandrasekharan et al., 2022; Copland, 2020; DeCook, 2022), belongs as well. Reduction is just one part of the Reddit quarantine, but the effect is similar. For users who don’t already subscribe to a quarantined subreddit, no posts from within that subreddit will appear on the Reddit front page or be returned in search results or other site-wide recommendations. In other words, it’s there if you go looking for it, but it won’t be circulated broadly. Reddit’s reasoning will sound familiar: “to prevent its content from being accidentally viewed by those who do not knowingly wish to do so, or viewed without appropriate context.” 21 It is worth noting that Reddit has been more explicit and transparent about their quarantines than YouTube and Facebook have about reduction. Still, Reddit users outside the quarantined subreddit may not know why some things aren’t bubbling up on their front page anymore.
Why Not Just Remove It
If a platform deems content problematic enough to justify reducing its circulation, why not simply remove it? In other words, if the problem is at the borderline, why not just move the border? After all, Trust & Safety teams write these rules themselves, change them frequently, and reserve the right to interpret them as they see fit; why not adjust the rule and remove the content? The platforms that have acknowledged implementing these approaches have offered several kinds of reasons; I will suggest a few more that, unsurprisingly, tend to go unsaid publicly.
Facebook has offered a jumble of justifications for their reduction policy. First, they were responding to their users’ wishes—what “our community has told us they don’t want to see”—though without indicating how those wishes were articulated. Six months later, Mark Zuckerberg and his PR team offered a second, somewhat different justification: Our research suggests that no matter where we draw the lines for what is allowed, as a piece of content gets close to that line, people will engage with it more on average—even when they tell us afterwards they don’t like the content. This is a basic incentive problem that we can address by penalizing borderline content so it gets less distribution and engagement.
22
According to Zuckerberg, users are (invariably) drawn to “sensationalist and provocative content.” Wherever a line is drawn, interest balloons alongside it. This incentivizes creators to be daring, saucy, outrageous, and dangerous. By this logic, Facebook is not implicated, it is simply dealing with the inevitable outcome of policing itself: a “natural engagement pattern” it is not directly responsible for (Hallinan, 2021). This is akin to concerns about “gaming the algorithm” (Petre et al., 2019), that clever content creators will do precisely as much as they can get away with—except that, in this characterization, demand comes first. Zuckerberg included Figures 1 and 2 (that look mathematical, but are not) to explain both the problem and his solution (Intriguingly, after the reduction policy was in place, Zuckerberg not only hoped to reduce demand and supply down to an equivalent level, but to mute it further).

Mark Zuckerberg, Facebook, 15 November 2018. “Blueprint for content governance and enforcement.”
In Facebook’s view, moving the line would only shift the problem: to the left of every line, no matter where it is drawn, is a bubble of demand for the sensational and illicit that someone will fulfill. It is worth noting that this is a very specific theory of culture and taste, and I am not convinced that it is correct. Still, naturalizing this as human and inevitable allows Facebook to act responsibly without admitting responsibility.
A year later, Facebook offered a third justification: Since recommended content doesn’t come from accounts you choose to follow, it’s important that we have certain standards for what we recommend. This helps ensure we don’t recommend potentially sensitive content to those who don’t explicitly indicate that they wish to see it.
23
Consent, which Facebook infers when you like a page or friend another user, justifies a lesser standard of responsibility than a recommendation coming “from” Facebook. 24
YouTube management has generally framed their reduction policies as an acknowledgment of a growing public responsibility. In a long-form piece on the new policy,
In my informal conversations, I was also offered other, less noble reasons for not simply removing this problematic content. First, reduction is less politically risky than removal. Given the recent climate, platforms fear reprisals from conservative critics who air their outrage whenever their posts are removed. 26 Demoting reprehensible content lets platforms avoid “censoring” it, or facing charges of bias that are difficult to refute. Flip this around, and it is not difficult to imagine that reducing problematic content allows platforms to continue to benefit financially from the users who do seek it out, in the form of advertising revenue and/or data collection, while still answering public concerns by reducing its reach (Matamoros-Fernández, 2017; Siapera & Viejo-Otero, 2021).
Reduction strategies may also be preferable when the types of problematic content are difficult to identify, in flux, or difficult to police. Reducing without removing allows platforms the flexibility to intervene around quickly emerging phenomena, to go after content designed to elude prohibitions, and to curtail content they “know” is bad but have a hard time articulating why. Seen in the best light, this flexibility makes it easier to respond quickly to changing problems—from the unpredictable outbursts of White nationalism, to the evasive tactics of pro-ana groups, to the constantly evolving QAnon conspiracy. In a less flattering light, reduction also avoids public accountability, as the interventions themselves are hard to spot, and are not—yet—reported as part of the platform’s transparency obligations.
How Reduction Works
Reduction is not the exclusive purview of the Trust & Safety teams who traditionally handle content moderation. Like the conceptual distinction being made between different zones of responsibility on the platform, reduction lives at a similar institutional distinction between the teams that handle algorithm design and those that handle policing. This is not an insignificant detail: where work done within a company can reveal a great deal about how a problem is perceived, what approaches are likely to be pursued, who has to approve and justify the resources required, and how the solution will be understood and valued by the company’s senior leadership (Caplan & boyd, 2018).
Recommendation is a central component of social media platforms, but it is driven by a different set of concerns and priorities: while Trust & Safety teams
Generally these signals measure some aspect of the content understood to be positive—that is, reasons why the content should be shown: Is this video recent? Is this link recommended by this user’s friends or network? Is this post often liked by users who share a similar matrix of interests? Content with a higher score on these measures is more likely to be recommended. Sophisticated recommender systems can calculate hundreds of these signals, measure each signal against its own specific threshold, weigh them differently, and adjust dynamically depending on the user, region, situation, genre, or moment. Reduction techniques take advantage of this, but flip it: the calculation of what to recommend now includes a negative signal, indicating that a particular piece of content should not be recommended to this particular user.
There are several points along the recommendation process where reduction can take place. In its own technical literature, for example, YouTube engineers distinguish two key steps in their recommendation algorithm: inventory and ranking. 27 YouTube does not rank millions of videos every time it needs to recommend content to a user. Inventory, or what in Figure 3 is called “candidate generation,” represents a first pass at gathering videos that might be relevant to this user. YouTube’s software quickly identifies a small set of videos to even consider recommending, numbering in the hundreds, based generally on their similarity to videos the user watched before, their recent popularity, and so on. “Ranking” then orders these videos to determine which to recommend next, and in what order.

Deep neural networks for YouTube recommendations.
This means reduction can happen in several ways, with different implications for how widely “reduced” content would circulate, and to whom:
In whatever form, this is an example of what Ananny (2020) calls a “probabilistic” approach to content moderation: these platforms want “less” of the content they have deemed problematic, and use mathematical tweaks to achieve it.
The other half of the challenge, common to all content moderation efforts, is how to identify the problematic content in the first place—and do so quickly, accurately, based on limited information, and at scale. It should surprise no one that, to accomplish this, most platforms have turned to a now well-worn Silicon Valley technique: develop a machine learning classifier that can estimate what content is “problematic,” by training that classifier on a heap of data that’s already been evaluated by human raters (Ekbia & Nardi, 2014; Gorwa et al., 2020; Gray & Suri, 2019). Ideally, the judgments made by the human raters will be approximated by the machine learning classifier, which can then make the same judgment over and over on millions of pieces of content.
Facebook seems to use the same machine learning classifiers already used to identify content they may remove. 29 Posts that score highly are removed, and posts that score close to that may now be algorithmically reduced. YouTube opted to develop a separate classifier, using the same machine learning techniques, specifically to identify the content they consider borderline or harmful. 30 This meant they needed data rated by humans. The labor of rating videos as “borderline” is farmed out as piecework to thousands of independent workers, much like other forms of commercial content moderation labor (Roberts, 2019; Ruckenstein & Turunen, 2019; Steiger et al., 2021). But YouTube asserts that, here, the work is structured differently. Evaluators are given more time per evaluation and are encouraged to do research—that is, Google searches—to help assess which videos are problematic: How fishy is this video and how false are its claims? 31 (Tellingly, the training manual YouTube gives its raters to spot conspiracy videos, misinformation, and other borderline content, is the same document Google uses to train evaluators of search result quality. 32 The tools of quality, turned to the concerns of harm). Using the scores from these human evaluators, the machine learning classifier is trained to spot borderline content and harmful misinformation.
Making Visible What’s Been Made Less Visible
Given how they have conducted content moderation up to this point, reduction policies may be the most mature step platforms have yet taken (DiResta, 2018). While it may not sit well with idealized notions of the marketplace of ideas, it may nevertheless be true that healthy public spheres actually require their information landscapes be curated. Whether platform managers are genuinely motivated by a sense of responsibility, or it is just a good PR move, this may finally recognize the complex ways in which recommendation algorithms do seem to amplify some kinds of contributions over others, do seem to exacerbate the more chaotic elements of civic discourse. The technical field of recommender systems is beginning to think about what “healthy” recommendations might look like (Beutel et al., 2020; A. Singh et al., 2020), as part of a broader commitment to fairness, accountability, and transparency.
However, reduction avoids none of the legal or societal problems that already haunt content removal: the presumptive power of the arbiter, the possible biases, the inequitable impact on different user communities, and the implications for free speech (Keller, 2021). In fact, reduction techniques concentrate even more of that curatorial power in the hands of these elite, private, profit-oriented intermediaries, by embedding and obscuring their judgment and worldviews within these nearly imperceptible adjustments.
One common rejoinder to such unchecked power is to demand that platforms be transparent about and accountable to their own policies. For years, critics have pushed social media companies to be more transparent about the moderation decisions they make, 33 and the major platforms have slowly acquiesced; many now publish regular transparency reports, articulate their policies in greater detail, and offer more robust mechanisms for appealing moderation decisions (Ananny & Crawford, 2018; Suzor et al., 2019). However, reduction techniques, at least so far, remain completely invisible to these apparatuses of observability or accountability (Rieder & Hofmann, 2020). None of these interventions are reported or documented, and data about what is reduced and to what degree are not made publicly available. Except for Facebook, even the policies are not spelled out.
This is especially problematic because reduction is so hard to detect. When content is removed or a user suspended, there are at least traces left behind: the video is missing, and there is a mark where the deleted post used to be. However, there is no trace left when a post, tweet, or video simply has not circulated as far as it might have otherwise. The content remains, it can be found, commented on, and forwarded, yet it seems to not have the audience or traction that it might have. This uncertainty leaves users grasping for explanations, and is part of why so many users are suspicious that murky machinations are at work under the hood of these platforms (Caplan & Gillespie, 2020; de Keulenaar et al., 2021; Nicholas, 2022; West, 2018). Although critics may not have all the details right, or an easy way to confirm their suspicions, many have long suspected that platforms are gaming their own systems without making it known to their users. YouTube creators raise alarm bells when their work gets demonetized without explanation, forced to read the tea leaves on why revenue is down on one video and not another (Caplan & Gillespie, 2020; Cotter, 2021). Sex workers accuse platforms of shadowbanning, but must gather the evidence themselves (Are, 2021; Blunt et al., 2020, 2021; Smith et al., 2021). Conservative pundits accuse the ranking mechanisms of social media algorithms of being tipped against them, in ways that seem nefarious but cannot be pinpointed. While not all of these accusations will prove to be true, and some are politically self-serving for those making them, they tap into an underlying truth: platforms
At the same time, it is hard to imagine how platform companies would be transparent about reduction policies. How does one measure or document reduction? What should the reduced visibility of a piece of content be compared to? Because the circulation of a piece of content tomorrow depends in part on who happens to see it today, both amplification and reduction have a cumulative effect. For the same reason, we cannot know how that content would have traveled had it not been reduced. There is no “normal” reach of content to measure against. How something might have performed—versus how it did—depends on its quality, who saw and liked it early, whether it got traction and how much, what it was up against on the platform, what news was breaking at the same time, how the machine learning algorithms rated it, what criteria and thresholds they were tuned to at that moment, and on and on (Magalhães & Yu, 2022).
Algorithms in practice are frustratingly and impossibly opaque (Burrell, 2016). And that impossibility of being precise about what impact a reduction policy may have had leaves users uncertain about when and to what degree they have been subject to an intervention, allows platforms to both overstate and oversimplify the effects of their algorithms, and may hinder what the law could do to oversee platforms and prevent abuses of this power (Cobbe & Singh, 2019; Helberger, 2020; Heldt, 2020; Keller, 2021).
Conclusion: Moderation by Other Means
Major platforms, including Facebook, YouTube, Instagram, Twitter, Tumblr, TikTok, LinkedIn, and Reddit have added reduction to their content moderation techniques. They use machine learning classifiers to identify content that is misleading enough, risky enough, and problematic enough to warrant reducing its visibility, by demoting or excluding it from the algorithmic rankings and recommendations—while not going so far as to remove it. They distinguish different aspects of their platforms to ascribe different logics of responsibility and justify different interventions: passive hosting versus active recommending, what users asked for versus what they were presented with. Reduction policies are explained as either responsible oversight of algorithmic systems, or unavoidable responses to the human impulse toward the sensational: Or, they are not explained publicly at all.
The way YouTube, Facebook, and others have developed and legitimated their reduction policies reveals a great deal about how content moderation now works. This is content moderation by other means, conducted sometimes by other parts of the companies, deployed in different parts of the platform. To address growing concerns about
Whether it is the separation between signal and noise in this context is complicated, as what constitutes a disturbance is decided by multiple actors, and is not restricted to those who create the medium. What needs to be filtered constantly changes because what is considered to be an interference to the business model is also constantly in flux. (p. 186)
If reduction is a form of content moderation, then it must be included in the ongoing debates about platform responsibility. Does it benefit the public, or undermine it, when platforms regularly and quietly reduce what they deem to be misinformation, conspiracy, and “borderline content” violations? We do not know the impact of reduction techniques. We do not know whether that impact differs when what’s being reduced is White nationalism, junk news links, explicit sex work, or users struggling with the impulse to harm themselves (Gerrard, 2020). We do not know whether we can trust platforms to engage in these reduction practices thoughtfully, in ways that produce a robust but fairer public sphere. We have little access to who is making these policies and distinctions, and according to what criteria.
In fact, I suspect that reduction policies will have the effect of further normalizing the underlying logics by which social media select and circulate information. Reduction is a corrective mechanism that presumes and reifies the logic of the system it is designed to correct. It prefigures a system that takes
If platforms are unavoidably curators, always selecting for and selecting out; if asking them not to be is nonsensical; and if that power to curate might be a more mature, progressive approach than anything they’ve done until now, then what we must grapple with is a very old problem: private intermediaries with power over the public speech that circulates through them (Gillespie, 2018). What we know, not just from two decades of studying social media but from studying media for many more, is that the power to select is important and unavoidable, it can be exerted in both progressive and destructive ways—but that, again and again, the impact of these selections are felt inequitably. Whatever combination of algorithmic reductions, moderation removals, and design adjustments platforms employ, the questions driving analyses should be: Who is excluded and who is included? Who enjoys the largesse and who bears the constraints? Which groups are further marginalized and which are given center stage? These inquiries must go beyond mere calculations of profit, for sure, but also beyond the First Amendment, beyond unspecified notions of public interest, beyond reassuring notions of openness and meritocracy. We know, and should now take as axiomatic, that universalized principles of fairness cloak and reaffirm systems of inequity (Díaz & Hecht-Felella, 2021; Gerrard & Thornham, 2020; Gray & Stein, 2021; Haimson et al., 2021; Marshall, 2021; Noble, 2018; Southerton et al., 2021; Zolides, 2021). The study of platforms and their interventions, then, should put questions of social inequity front and center, and explain why they always seem to come second to other priorities.
There is a world in which reduction techniques could, at least in theory, be a salve to structural inequities of visibility: if they were used to counter how marginalized communities regularly lose ground amid the proclaimed evenhandedness of intermediaries and in the face of the bad faith tactics of their antagonists. Platforms willing to reduce content that is harmful, abusive, or in bad faith might also reduce the swollen voices of the already amplified, to make more room for the kind of pluralist civic sphere that social media has sometimes promised but never achieved. However, that is probably not what will happen. Unless reduction techniques are premised on a different theory of participation and equity, driven by an ethic of care, and fitted with accountable oversight and public debate, they could further scuttle marginalized communities—as early evidence suggests they are already doing. Marginalized communities have long been “reduced” by the centrism and conservatism of traditional media, their content dismissed as “low quality” because it doesn’t look like it is “supposed to.” Reduction on social media platforms, for different reasons, could easily do the same.
Footnotes
Acknowledgements
The author thanks the anonymous reviewers, his excellent colleagues in the Social Media Collective at MSR, and to those who helped him with these ideas or read drafts of this paper, especially Mike Ananny, Robyn Caplan, Zoe Darmé, Fernando Diaz, Stephan Dreyer, Ysabel Gerrard, Amélie Heldt, Kate Klonick, Oliver Haimson, Ashley Mears, and Dylan Mulvin. The author thanks Wendy Chun and the Digital Democracies Institute at Simon Fraser University, Oscar Westlund and the Digital Journalism Research Group at Oslo Metropolitan University, Sally Wyatt and the Dutch Digital Society 2021 conference, Ana Viseu and the Universidade Nova de Lisboa, and Joan Feigenbaum and DIMACS for the opportunities to present this as a work in progress. Finally, the author thanks Rachel Bergmann and Elizabeth Fetterolf for their superb assistance with this research.
Correction (May 2023):
Article updated online to include Figure 2 on page 5.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
