Abstract
During the onset of the COVID-19 pandemic, various officials flagged the critical threat of false information. In this study, we explore how three major social media platforms (Facebook, Twitter, and YouTube) responded to this “infodemic” during early stages of the pandemic via emergent fact-checking policies and practices, and consider what this means for ensuring a well-informed public. We accomplish this through a thematic analysis of documents published by the three platforms that address fact-checking, particularly those that focus on COVID-19. In addition to examining what the platforms said they did, we also examined what the platforms actually did in practice via a retrospective case study drawing on secondary data about the viral conspiracy video, Plandemic. We demonstrate that the platforms focused their energies primarily on the visibility of COVID-19 mis/disinformation on their sites via (often vaguely described) policies and practices rife with subjectivity. Moreover, the platforms communicated the expectation that users should ultimately be the ones to hash out what they believe is true. We argue that this approach does not necessarily serve the goal of ensuring a well-informed public, as has been the goal of fact-checking historically, and does little to address the underlying conditions and structures that permit the circulation and amplification of false information online.
In February 2020, the Director-General of the World Health Organization (WHO) cautioned that the spread of “fake news” had the potential to exacerbate the spread of COVID-19, declaring “we’re not just fighting an epidemic; we’re fighting an infodemic” (Ghebreyesus, 2020). One month later, the Associate Director of the International Fact-checking Network (IFCN) described COVID-19 as “the biggest challenge fact-checkers have ever faced” (Suárez, 2020). By this point, Facebook and YouTube had begun to develop partnerships with fact-checking organizations, and fact-checking gradually became a go-to response for managing information pollution (misinformation, disinformation, malinformation) (Wardle & Derakhshan, 2017). Because platforms are key vectors for misinformation, particularly pandemic-related misinformation (Bridgman et al., 2020), ensuring that the global public are well-informed requires assurance that platform fact-checking functions effectively and in the public’s best interest. In a very real sense, the global population’s safety and wellness hinges on the strength and quality of platforms’ regimes of fact-checking.
More broadly, platforms’ role in and response to infodemics symptomize the rise of the “platform society”—how platforms have “penetrated the heart of societies—affecting institutions, economic transactions, and social and cultural practices” (van Dijck et al., 2018, p. 2). In this state of affairs, platforms’ private interests and values do not always perfectly match those of the diversity of publics that global platforms like Facebook, Twitter, and YouTube serve. Moreover, there is an ongoing debate about what rights and responsibilities platforms have in relation to user generated content and activity (e.g., Gillespie, 2018b; Gorwa, 2019; van Dijck et al., 2018). Early reports suggest that platform fact-checking programs invoke many longstanding questions and concerns around “governance by platforms,” or the policies, design choices, and business models that structure participation on these sites (Gillespie, 2018b), particularly those related to content moderation.
The purpose of this study is to explore how platforms responded to the infodemic during early stages of the pandemic via emergent fact-checking policies and practices, particularly in the United States. We focus on how three major platforms (Facebook, Twitter, and YouTube) turned to and integrated their governance structures with the existing infrastructure of fact-checking, which has traditionally been the purview of journalists. In this way, we connect literatures on mis/disinformation and fact-checking with the growing body of work on platform governance to consider how platforms are reconfiguring fact-checking and with what effect. To accomplish this, we conducted a thematic analysis of official documents published by Facebook, Twitter, and YouTube that address fact-checking and focus on COVID-19. In addition to examining what the platforms said they did, we also examined what the platforms actually did in practice via a retrospective case study drawing on secondary data about the viral video Plandemic, which propagated various false claims and conspiracy theories about COVID-19.
Acknowledging that platform policies change rapidly, we note that this research focuses on platform policies and (in)actions that occurred early on during the COVID-19 pandemic (from March 2020 until November 2020, prior to vaccine availability). As such, this research documents an important historical moment in platforms’ nascent practices and policies around fact-checking and COVID-19-related mis/disinformation.
Overall, our findings reveal a familiar emphasis on vague and subjective policies and practices for limiting the visibility of mis/disinformation. As in past approaches to content moderation, the platforms communicate the expectation that users should ultimately be the ones to hash out what they believe is true. We argue that this deviates from the goal of fact-checking, traditionally, which is to ensure a well-informed public.
Infodemics and Information Pollution
The struggle to combat the spread of false information has become an era-defining issue. The loss of trust in institutions (media, government, education, and in our context, health organizations) in recent years has contributed to the pervasiveness and high levels of susceptibility to false information (Humprecht et al., 2020; Swire-Thompson & Lazer, 2020). Moreover, polarization, populism, fragmentation, shifts toward the online engagement-based advertising ecosystem, and the decline of local journalism have collectively established conditions conducive to “information pollution” (Humprecht et al., 2020; Wardle & Derakhshan, 2017).
Information pollution serves as an umbrella term for different kinds of problematic information. This article focuses on two main forms of information pollution: misinformation and disinformation. Misinformation refers to when people share unsubstantiated claims, rumors, and conspiracy theories that they do not recognize as untrue or inaccurate. Disinformation refers to when people intentionally share false information to propagate a certain point of view by deceiving audiences.
In the health context, information pollution is especially dangerous. Studies have consistently demonstrated how mis/disinformation have contributed to the spread of diseases (Gyenes & Mina, 2018), even ones previously thought eradicated (Swire-Thompson & Lazer, 2020 ). Indeed, early evidence suggests that the preexisting anti-vaccine propaganda has contributed to global reluctance to receive a COVID-19 vaccine (Smith et al., 2020). In the domain of health and science, mis/disinformation is defined in relation to contemporaneous expert consensus (Tan et al., 2015) and “best available evidence” (Garrett et al., 2016 p. 333). That is, a claim can be considered false if most experts agree and the preponderance of scientific evidence suggests that it is false (Vraga & Bode, 2020).
In online settings, false information often spreads farther and faster than facts (Ball & Maxmen, 2020; Lewandowsky et al., 2017; Swire-Thompson & Lazer, 2020). As such, information pollution has increased in scale and severity in recent years, as false information can quickly spread from platform to platform, evading moderation (Ferrara et al., 2020; Gyenes & Mina, 2018) and making it difficult to control. The mechanisms by which false information spreads and who accepts it are still under constant discovery. Existing work suggests that platforms’ technical infrastructures—particularly, algorithms that sort, filter, and recommend content—play a significant role in connecting people with digital spaces and conversations generative of information pollution (Gillespie, 2020; Kaiser et al., 2021; Villasenor, 2020). This state of affairs has placed new demands on platforms to engage in fact-checking.
The Evolution of Fact-Checking
Fact checking has evolved over time, and currently plays a key role in the so-called “post-truth” media landscape. Historically, fact-checking has been associated with journalism, reflecting its professionalization in the 20th century as a “fact-centered discipline” (Graves & Amazeen, 2019, p. 3). Yet, over time, fact-checking has developed into a broader infrastructure composed of an eclectic mix of people and organizations, practices and routines, principles, and tools interconnected in support of the goal of “helping people become better informed and promoting fact-based public discourse” (Graves & Amazeen, 2019, p. 1).
In the early 21st century, fact-checking significantly expanded and began to revolve around ensuring institutional accountability (Graves & Amazeen, 2019). This shift likely grew in response to a series of claims in the political sphere that proved false, with the Bush Administration’s claims about weapons of mass destruction as justification for the Iraq War as a watershed moment (Marietta et al., 2015). Simultaneously, the Internet democratized the ability to produce news-like content. This allowed independent, and sometimes amateur, fact-checking sites (e.g., Snopes.com) to establish themselves to help dispel conspiracy theories and rumors while also striving to serve as watchdogs for politicians, journalists, and other public figures (Graves & Amazeen, 2019). Soon, some major media organizations developed fact-checking arms, in response to demand for greater oversight of political campaigns and government (Graves & Amazeen, 2019).
While media organizations dominate fact-checking, comprising more than half of all fact-checking organizations, their role has weakened over time with increasingly more nonprofits, think tanks, nongovernmental organizations, and academic institutions joining the ranks (Bell, 2019; Stencel & Luther, 2021). The extraordinary levels of misinformation and disinformation in the 2016 U.S. election, further engendered demand for fact-checking with a resultant 200% increase in fact-checking organizations (Fischer, 2020). According to the Duke Reporters’ Lab census, as of June 2021, 341 fact-checking organizations are active worldwide (Stencel & Luther, 2021). Demonstrating fact-checking’s evolution into its own subindustry, in 2015, the IFCN (n.d.) was formed, and a year later the organization introduced a Code of Principles, which prioritized transparency, nonpartisanship, and fairness.
Fact-checking is now considered an essential tool for combating false information, particularly online. While past work found that fact-checking corrections can create a backfire effect, and reinforce people’s original, inaccurate beliefs (Nyhan & Reifler, 2010; Nyhan et al., 2013), more recent research has shown fact-checking to be successful in improving the accuracy of beliefs, supporting the ability to correctly evaluate claims, and reducing intentions to share untrue headlines on social media (Amazeen et al., 2015; Nyhan et al., 2020; Porter & Wood, 2020; Yaqub et al., 2020). Together, these findings suggest that fact-checking can effectively fulfill its core aim of ensuring a well-informed public, which makes it an important tool for online platforms as they have become key venues for keeping abreast of news and current events.
Fact-Checking in the Platform Era
After the 2016 U.S. presidential election, concerns about rampant mis/disinformation online escalated, and major U.S.-based platforms began to implement their own fact-checking programs in response (Ananny, 2018). As will be discussed, Facebook and YouTube have built relationships with third-party fact-checking organizations and Twitter has handled fact-checking internally. These fact-checking programs represent an extension of platforms’ “apparatus for content moderation” (Gillespie, 2020, p. 329), in which they have invested considerable resources over the past decade. Content moderation is a core mechanism by which platforms govern activity on their sites, shaping “what is seen, what is valued, what is said” (Grimmelmann, 2015, p. 42).
By looking to existing work on content moderation, we can see how the platform fact-checking program may differ from pre-existing fact-checking efforts. First, content moderation follows a particular politico-economic logic that may conflict with arbitrating truths. Content is platforms’ core commodity (Roberts, 2018), and public outcry and legal pressure have made it good business for them to address various kinds of “problematic” content on their sites. Content moderation policies and practices should be understood as “compromises—between users with different values and expectations, as well as between the demands of users and the demands of profit” (Gillespie, 2018a, p. 12). Yet, in such compromises, the values, expectations, and demands of certain users and stakeholders matter more than others. Platforms commonly rely on “tiered governance,” exempting high-profile accounts from normal enforcement of policies or subjecting them to a more lenient set of rules (Caplan & Gillespie, 2020; Horwitz, 2021). For example, under pressure from conservative politicians and advocacy groups, Facebook has urged fact-checking partners to alter “False” determinations, particularly for prolific advertisers (Pasternack, 2020).
This politico-economic logic of content moderation positions it at odds with fact-checking’s goal of achieving institutional accountability. Historically, platforms like Facebook, Twitter, and YouTube have been loath to “choose sides,” which in some respects institutional accountability presupposes. Instead, they aim to give the appearance of noninterventionist mediators (Gillespie, 2018a), “empowering all by choosing none” (Gillespie, 2010, p. 357). Platforms lean on a cyberlibertarian approach that places the onus on individual users, deemed rational actors, to work out the truth or the “best” ideas through competition in the “marketplace of ideas” (Maddox & Malson, 2020).
Second, content moderation on platforms depends upon both people and artificial intelligence (AI), but, given the global scale of content, the latter is used to detect the majority of cases and to enforce rules (Gillespie, 2018a). Algorithmic moderation is often perceived and positioned as more objective than human judgments (Gillespie, 2018a; Roberts, 2018). Yet, algorithmic moderation is rife with errors, particularly in situations in which drawing the line between acceptable and unacceptable depends upon nuanced understanding and interpretation of sociocultural context (Gillespie, 2018a). For example, fact-checking and news organizations typically fact-check a much wider range of content than platforms, including content that can be difficult for AI to evaluate, like satire, opinion pieces, and political advertisements (Stewart, 2021). Misinformation and disinformation pose a significant problem for automated detection, as they often require making complex judgments about facts that carry partial truths, involve moral concepts, or lack consensus (Stewart, 2021). Moreover, mis/disinformation differs from other kinds of problematic content for which content moderation systems were built. Unlike porn, gore, or the illegal sale of regulated goods, mis/disinformation is “designed to emulate exactly what the platform wants to distribute most” (Gillespie, 2020, p. 334). Although AI is gradually being integrated into the fact-checking infrastructure, some suggest human oversight remains key (Adair & Stencel, 2020; Graves, 2018).
Research Questions
As platforms endeavor to develop fact-checking programs, they must adapt the preexisting fact-checking infrastructure to their goals, values, and systems. In the remainder of the article, we explore this, as it has unfolded in the context of health mis/disinformation. We ask, in the early stages of the pandemic, how have platforms responded to COVID-19 and the larger infodemic crisis via policies and practices around fact-checking, and how do they position themselves within the broader infrastructure of fact-checking? By exploring these questions, we aim to contribute to discussion of how platforms are (re)shaping our information infrastructure, which has significant implications for global resilience to information pollution (Humprecht et al., 2020), particularly amid the high stakes context of the COVID-19 pandemic. How platforms enter into and alter the nature of fact-checking matters for how we arbitrate truth in the global public sphere—who is responsible for this work and how we should operationalize it.
Method
To answer the research question above, we conducted a thematic analysis of official English-language documents issued by Facebook, Twitter, and YouTube that address fact-checking, focusing on documents that mention COVID-19. To identify documents, we searched each platform’s relevant webpages (e.g., about.fb.com, blog.twitter.com, support.google.com) for the keywords “fact checking” or “fact check” and related keywords like “misinformation” and “fake news.” Our search yielded blog posts, press releases, policy documents, and help pages, all representing a subset of a larger collection of documents that concern misinformation and fact-checking on the platforms in general. We cut off data collection in November 2020 (providing insights into the platforms’ actions during the early days of the pandemic), resulting in a total of 312 documents. We then reviewed all the documents from January 2020 on, searching for keywords related to COVID-19 (COVID-19, coronavirus, pandemic). Our final dataset describing policies and practices related to the early days of the COVID-19 pandemic totaled 60 documents (see supplementary materials for a list of documents). Of note, all three authors are based in the United States, which meant that the platforms defaulted to showing policies in English and specifically for U.S.-based users.
We qualitatively coded these 60 documents via an iterative process, focusing on what the platforms said about how they were responding to the COVID-19 infodemic and how they rhetorically positioned themselves in this. First, each of our three-member research team independently coded 5 different documents each (i.e., 15 documents). We discussed codes, occasionally merging and synthesizing higher-level codes. After, we each coded six additional documents and discussed and adjusted codes again (18 documents). Finally, we divided the remaining 27 documents among us (9 per researcher) and coded them independently. At this point, we met to further develop more abstract themes. The final codebook included codes highlighting fact-checking actions taken by the platforms (including codes about third-party fact-checking partnerships, removal of content based on fact-checking, and encouraging content creators to do their own fact-checking) and their enforcement strategies used to identify and prevent the circulation of problematic information (such as limiting visibility through warning labels, demoting, or removal of content).
We also conducted a retrospective case study of the viral video Plandemic, which contained a series of false claims about COVID-19 and public health officials. Drawing on secondary sources, mainly media reports, we sought to understand (1) what actions platforms took in relation to the Plandemic videos, (2) when they took action, and (3) why they said they took action. We used this case study as a tool for reflecting on how platforms actually applied their policies and practices in practice.
Findings
In the three platforms’ descriptions of their policies and practices, we can see how they have integrated themselves within the infrastructure of fact-checking, treating it as special area of content moderation. At the highest level, the platforms described a twofold response to COVID-19 mis/disinformation: (1) verification of information and (2) limiting the visibility of false or inaccurate information. Yet, we saw that, unlike fact-checking journalists and organizations who focus on adjudicating the veracity of claims, the platforms exhibited a principal concern over the visibility of potentially false and inaccurate information. Such a concern is reflective of platforms’ commitment to remaining neutral and the expectation that users will resolve the truth for themselves.
Verification of Information
Fact-Checking Programs (and Who’s Responsible?)
Facebook’s and YouTube’s fact-checking programs center partnerships built with existing third-party organizations. In multiple blog posts, Facebook highlighted its work “with over 60 fact-checking organizations around the world that review content in more than 50 languages” [FB56]. These 60 (and counting) partners represent those that have been certified by IFCN. YouTube described a similar relationship with third-party fact-checkers, though specifying that any U.S. publisher could generate fact-checks, so long as they were “either a verified signatory of the International Fact Checking Network’s (IFCN) Code of Principles or are an authoritative publisher” (YT21).
In contrast to Facebook and YouTube, Twitter rarely referred to “fact checking” in its communications. Indeed, only a handful of Twitter documents mention “fact check[ing],” and these refer to work by external organizations independent of (and not working directly with) the platform. However, the company described efforts that amounted to fact-checking in multiple blog posts. For example, in a blog post in May 2020, Twitter announced new measures for dealing with “misleading information,” which included a section describing how the company defined and identified such tweets [TW2]. Yet, Twitter shared few other details about the internal process of fact-checking, including who performs this work. In January 2021, Twitter changed course, announcing a community-based model of fact-checking tweets called “Birdwatch.” 1 While still in piloting, this tool would enroll average users in the process of fact-checking by allowing anyone to “write notes that provide informative context” (Coleman, 2021). Then, users would rate the quality and helpfulness of these notes, which would inform “future reputation models that recognize those whose contributions are consistently found helpful by a diverse set of people” (Twitter, n.d.). This mirrors the platform’s position (and, indeed, other platforms’ positions) that the “truth” should arise from discussion and debate among the user community, as the company put it “consensus from a broad and diverse set of contributors” (Twitter, n.d.).
Making Verification Sociotechnical
The three platforms bring together people and technologies—manual and automated processes—for selecting content to be fact-checked and, under some circumstances, actually performing a function akin to verification. Automated systems identify content that may include false or inaccurate claims and some of these are sent to fact-checkers and moderation staff for human review. This approach resembles the platforms’ approach to moderating content in general, which revolves around similar algorithmic processes refined over several years (Gillespie, 2018a). All three platforms noted an increased reliance on automated processes as a result of reduced and remote workforces during the pandemic, although the platforms were subtle in communicating the extent of this reliance and shared very little in general about what their automated processes do and how they function.
For example, Twitter explained in a blog post “Our teams are using and improving on internal systems to proactively monitor content related to COVID-19. These systems help ensure we are not amplifying Tweets with these warnings or labels and detecting the high-visibility content quickly” (TW3). Similarly, YouTube noted in a blog post: “Each quarter, millions of videos that are first flagged by our automated systems are later evaluated by our human review team and determined not to violate our policies” (YT22). Facebook was a little more forthcoming, explaining that its “machine learning model identifies potential misinformation using a variety of signals. These include comments on the post that express disbelief, and whether a post is being shared by a Page that has spread misinformation in the past.” 2
As with content moderation more generally, the platforms’ communications explained that these automated processes intersect with human review. For example, Facebook noted that posts flagged as potential misinformation by its machine learning model would first be sent to paid contractors who would research claims and offer a preliminary judgment on their merit, which would then be sent to fact-checking partners to aid them in selecting stories to fact-check (FB44). Moreover, the platforms’ algorithms detect potential instances of false or misleading information based in large part on user flags. YouTube has innovated in this area by introducing a “trusted flagger” program, which grants certain users special tools and privileges, including giving their flags more weight in detecting mis/disinformation. 3
Limiting the Visibility of Mis/Disinformation
Following verification by fact-checkers, the three platforms described limiting the visibility of content determined to be false. In their communications about COVID-19 misinformation, the platforms described a multifaceted approach of removing “harmful” misinformation, demoting non-harmful misinformation, and elevating authoritative voices.
Demoting and Removing
For all three platforms, content determined to be false or inaccurate as a result of verification, is de-prioritized in feeds or removed entirely. Yet, in their communications, it was not always clear how the platforms decided which content to demote versus which to remove. Facebook stated in a blog post: “Once a post is rated false by a fact-checker, we reduce its distribution so fewer people see it” (FB51). YouTube similarly indicated, following an existing practice, they would “reduc[e] recommendations of borderline content or videos that could misinform users in harmful ways” (YT27). It was not clear from Twitter’s documents whether it demoted any false or inaccurate tweets in feeds, but the company did note in a blog post from May 2020: “Our teams are using and improving on internal systems to proactively monitor content related to COVID-19. These systems help ensure we’re not amplifying Tweets” that were judged to contain misleading or disputed claims (TW2). Possibly, this could mean that the platform demoted tweets in feeds.
Parallel to demoting misinformation, the platforms also emphasized their efforts to make authoritative posts more visible. For example, Twitter stated, “With a critical mass of expert organizations, official government accounts, health professionals, and epidemiologists on our service, our goal is to elevate and amplify authoritative health information as far as possible” (TW12). All three platforms also discussed dedicated spaces on their site for providing general information about COVID-19 updated in real time from health experts like the Centers for Disease Control and Prevention (CDC) and the WHO (see Figure 1), which were presented as a means of providing people with easy access to reliable sources. Notably, the platforms specified that this provision of information would not be universally available, at least initially, in all countries.

YouTube COVID-19 information panel linking to the CDC website.
In addition to attempting to achieve the appropriate balance in the visibility of “good” and “bad” information, platforms removed some content entirely. The decision to remove content depended on the degree to which a given false claim would be considered harmful. For example, Twitter provided a table as a visual aid to explain their practices in this regard (see Figure 2). In this table, we see that the company placed “propensity for harm” as their key determining factor for applying labels, giving warnings, or removal. Namely, a tweet must contain “misleading information” and have a “severe” propensity for harm for the platform to remove it.

Twitter’s table for determining how to deal with mis/disinformation (TW2).
Similarly, Facebook stated in a blog post “We remove COVID-19 related misinformation that could contribute to imminent physical harm,” and “claims that don’t directly result in physical harm, like conspiracy theories about the origin of the virus” are sent to fact-checking partners to be debunked (FB48). While in one statement, YouTube referenced “harm” as a factor in reducing the distribution of “borderline” content, the company’s “COVID-19 Medical Misinformation Policy” dictated that they would remove any content that “contradicts WHO or local health authorities’ guidance on Treatment, Prevention, Diagnostic, Transmission” (YT9).
Warning Labels
Both Facebook and Twitter additionally added warning labels to posts deemed false by fact-checkers (see Figures 3 and 4). For instance, in an April 2020 blog post, Facebook explained: “Once a piece of content is rated false by fact-checkers, we reduce its distribution and show warning labels with more context” [FB51]. Recall that Twitter similarly stated that it would add warning labels to tweets that include disputed claims and have a “severe” “propensity for harm” (see Figure 2). Twitter elaborated in a blog post:
We may label or place a warning on tweets to provide additional context in situations where the risks of harm associated with a Tweet are less severe but where people may still be confused or misled. This will make it easier to find facts and make informed decisions about what people see on Twitter. (TW3)
Although these warning labels do not make posts less visible in feeds—as in demotion and removal—in many cases, they obscure posts with an overlaid label that notifies users that a post has been “disputed.” “Learn More” and “See Why” buttons on the warning labels compel users to click to learn what it was about the posts’ content that earned this label (see Figures 3 and 4).

Twitter warning label for “disputed” COVID-19 information (TW2).

Facebook’s warning label for false information (Facebook, 2019).
YouTube did not indicate in any of its communications that it used warning labels in the same way as the other two platforms. Rather than actively label individual videos as false or disputed or misleading, YouTube stated that it would display fact-checks in information panels in some search results (see Figure 5):
There are a few factors that determine whether a fact check information panel will appear for any given search. Most important, there must be a relevant fact check article available from an eligible publisher. And in order to match a viewer’s needs with the information we provide, fact checks will only show when people search for a specific claim. For example, if someone searches for “did a tornado hit Los Angeles,” they might see a relevant fact check article, but if they search for a more general query like “tornado,” they may not. (YT21)
This statement established that users may see fact-checks, though did not clarify when users could expect to see fact-checks.

YouTube fact-check information panel.
In the next section, we explain how the three platforms applied the foregoing policies and practices to the Plandemic video, which went viral in early May 2020.
Plandemic as a Case Study of Platform Fact-Checking in Practice
Featuring a discredited scientist who supported claims that “a shadowy cabal of elites is using a global crisis as a cover to profiteer and entrench their power” (Newton, 2020a), Plandemic espoused various common conspiracy theories that emerged early on in the COVID-19 crisis. The most prominent false claims the video propagated included that COVID-19 was engineered, not naturally occurring; masks can make people sick; and death counts were being inflated. The video was first published to Facebook, YouTube, and Vimeo on 4 May 2020 (Frenkel, Decker, & Alba, 2020), soon finding audiences on Twitter and Instagram as well (Hatmaker, 2020). It particularly gained steam as people shared it to QAnon, conspiracy theory, and anti-vaccine Facebook groups (Frenkel et al., 2020; Newton, 2020a; Nilsen, 2020), and as microcelebrities, including a celebrity doctor, shared it (Frenkel et al., 2020). In the week that followed the video’s debut, it accrued over 8 million views across YouTube, Facebook, Twitter, and Instagram (Frenkel et al., 2020). The “star” of the video gained 130,000 new followers on Twitter following the video’s release (Newton, 2020a).
By the time Plandemic premiered, the three platforms had already instituted some new policies and practices to address coronavirus misinformation. In response to Plandemic, YouTube and Facebook removed the video from their sites on 7 May, 3 days after it was published (Frenkel et al., 2020). Facebook also indicated that they had demoted the video prior to removing it (Newton, 2020a). The decision to remove Plandemic hinged on a determination that the specific false claim that wearing a mask can make people sick, could lead to “imminent harm” (Andrews, 2020). According to The Verge, YouTube said “it did not recommend ‘Plandemic’ or surface it ‘prominently’ in search results” (Newton, 2020a). Twitter began blocking associated hashtags (e.g., #Plandemic, #PlagueOfCorruption) around the same time that the other two platforms removed the video (Wong & Solsman, 2020). Twitter opted not to remove links to the video’s website, instead marking it as “unsafe,” which limited its reach, and directed users who clicked on it to a warning message indicating that the content was “potentially spammy or unsafe” (Robertson, 2020).
On 18 August a second, feature-length Plandemic film, Plandemic: Indoctornation, was released. By this time, Facebook had blocked access to the video’s domain, banning users from sharing it (Brewster, 2020). A company spokesman also noted that “This latest video contains COVID-19 claims that our fact-checking partners have repeatedly rated false so we have reduced its distribution and added a warning label showing their findings to anyone who sees it” (Newton, 2020b, p. n.p.). Users who attempted to share the link also received a pop-up message stating the video violated the Community Standards on spam (Zadrozny, 2020). YouTube removed uploads of the full video, but noted that they had not seen many attempts to upload it (Newton, 2020b). YouTube specified that they would deal with clips of the video on a case-by-case basis (Newton, 2020b). Twitter again did not block the video’s domain, but instead displayed a warning message to users who clicked, indicating the link was “potentially spammy or unsafe” (Robertson, 2020).
With the benefit of time, the three platforms have managed to vastly minimize access to the Plandemic videos. While a handful of stray copies of the videos or clips can still be found on Facebook, Twitter, and YouTube (as evident from keyword searches), they are not easy to stumble upon or even actively seek out. On the one hand, this is a commendable feat: the three platforms effectively scrubbed the videos from their sites. On the other hand, Plandemic went viral in the first place because of these platforms, and, to some extent, once the first video went viral, significant damage had already been done. It did not matter that the platforms verified the claims of the video through their fact-checking programs, because millions of people around the world learned of Plandemic on Facebook, Twitter, and YouTube in the 3 days it took for the platforms to respond. Consistent with what we saw in the platforms’ communications, their handling of Plandemic suggests a preoccupation with visibility of (certain kinds of) misinformation first and an informed public second. The primary concern was mitigating risk through assessing potential for harm. While fact-checking informed the platforms’ decision-making, harm-assessments, rather than false determinations, prompted removal.
Discussion
Our findings suggest that, in the early days of the pandemic, Facebook, Twitter, and YouTube focused their energies primarily on managing the visibility of COVID-19 mis/disinformation on their sites via (often vaguely described) policies and practices rife with subjectivity. This represents a rearticulation of the platforms’ insistence that they are not arbiters of truth (Gillespie, 2010), as well as an extension of sociotechnical systems they have built to deal with other kinds of problematic content. The three platforms all employed practices for verifying information through fact-checking, relying on a combination of human and algorithmic actors. All three platforms then either demoted, removed, or labeled content deemed false or inaccurate. Although these policies and practices can help mitigate the spread of mis/disinformation, as we saw with Plandemic, they do not always effectively or sufficiently limit exposure. To “flatten the curve” of misinformation, as Donovan (2020) wrote, platforms need to become more accountable for the content that they make available to billions of users. As the platforms further integrate fact-checking into the heart of their services, their decisions can shape the broader infrastructure of fact-checking by encouraging solutions that repurpose past practices, but which are vaguely articulated and do not address the full range of factors that contribute to information pollution. Simply put, platforms’ focus on limiting the visibility of mis/disinformation may divert attention away from the ideal outcome of fact-checking: a well-informed global public, an outcome that is especially critical during a pandemic. Below we further discuss the ways the platforms policies and practices did not quite accomplish this outcome, and then offer some recommendations.
In our analyses, we saw that the platforms wanted to regulate the visibility of facts and falsities according to subjective (and not openly shared) risk assessments made internally. Recall that both Facebook and Twitter had policies that dictate false information would only be removed if there was a potential for harm, otherwise it would be demoted (Facebook) and labeled (Facebook and Twitter). This creates an essential hierarchy of false COVID-19 information based on judgments about which false or inaccurate claims are likely, as Facebook put it, to “contribute to imminent physical harm” [FB48]. As we saw with Plandemic, such judgments are highly subjective: while Facebook identified one claim in the video that they believed could lead to “imminent harm,” Twitter seemingly did not judge the video generative of real-world harm, based on its decision not to remove the video. Moreover, this harm-based approach also indicates that the platforms did not believe that COVID-19 mis/disinformation, on its own, could be corrosive. Even mis/disinformation not directly resulting in physical harm can erode trust in institutions, which poses a more pernicious threat to a healthy democracy and citizenry (Kavanagh & Rich, 2018). As we have seen in the spread of mis/disinformation around COVID-19, the circulation of false claims that are not causally related to physical harm have indirectly led to an increase in anti-Asian hate crimes (Reny & Barreto, 2020), allowed the pandemic to surge due to vehement refusal to wear masks and social distance (Hornik et al., 2021), and resulted in attempts to sabotage vaccine distribution (Razek et al., 2021). If platforms are serious about combating mis/disinformation, fact-checking labels and nebulous policies about what constitutes a narrow definition of “harm” likely do little to “flatten the curve” of the infodemic. Instead, platform fact-checking, at least in the early days of the COVID-19 pandemic, seems to be designed to protect the companies against liability for and bad press around real-world harms connected to activity on their sites.
Due to their risk-based approach, Facebook and Twitter also pursued a different aim than fact-checkers generally do. Fact-checking, traditionally, aims to pass a judgment on a claim to determine whether it is true or false. By contrast, by leaving false and inaccurate information up on their sites, though demoting and/or labeling the content, the platforms reaffirmed their underlying “marketplace of ideas” approach to governance (Maddox & Malson, 2020). In other words, the platforms “prioritize[d] free speech and more speech to correct the record” (Maddox & Malson, 2020, p. 8). Although their warning labels may indicate, for example, as Twitter’s labels state, that content “conflicts with guidance from public health experts regarding COVID-19,” the decision to allow mis/disinformation to remain visible on the sites (though perhaps less visible) implicitly gives some credence to the false claims. This response amounts to an impartial shrug, effectively communicating to users: “Others have determined this is false, but you decide for yourself.” When it comes to health mis/disinformation, a “you decide” approach does not make sense. For example, Twitter permitted the Plandemic video to remain on the site, explaining that people often “[dispute] the claims in real time with the link included as context” (Brewster, 2020). However, relative to other kinds of mis/disinformation (e.g., in the political realm), medical facts and public health guidelines are fairly clear cut: we judge the veracity of claims based on how well they match the facts and guidance offered by doctors, researchers, and public health experts. When it comes to COVID-19, in most cases, claims that contradict expert medical advice are not matters of “differing opinion.” Average individuals are not qualified in general to judge the credibility of medical claims or advice on their own. Moreover, it is questionable whether a user needs to link to false information to dispel it.
As the platforms lean on automated systems to support fact-checking on their sites, they urge a deeper investment in technologies of visibility within the broader fact-checking infrastructure. Certainly, there are many opportunities for automated systems to support the identification of claims to fact-check, as well as the verification and correction of claims (see Graves, 2018). However, developing systems to support fact-checking also requires significant resources (Graves, 2018), which means that the three platforms may have an outsized influence on the design, testing, and application of them. Indeed, Facebook and YouTube’s parent company, Google, have been major benefactors of the fact-checking industry in recent years, with Facebook reportedly paying fact-checking organizations hundreds of thousands of dollars each annually (Bell, 2019) and with both platforms creating grant programs to support fact-checking (FB52; YT21). Moreover, the design of such systems shapes the nature and work of fact-checking (Graves, 2018). Our analysis of the platforms’ communications indicated that their automated systems were oriented toward regulating the visibility of mis/disinformation via detection and downranking. However, it is not clear whether or how the platforms are developing systems to support the verification of information. Facebook’s and YouTube’s partnerships with third-party fact-checking organizations have given them room to relieve themselves of responsibility for fact-checking determinations. They may be reluctant to develop automated systems that would support verification, focusing instead on systems that allow them to modulate the relative visibility of “good” and “bad” information. The focus on visibility of content not only creates a loophole for unverified claims to circulate, but also directs the focus toward what is visible to users rather than rethinking the systems that give rise to viral mis/disinformation in the first place.
In short, this article demonstrates how regimes of platform fact-checking embody a cyberlibertarian ideology that drives platform decisions, interpellating their users and fact-checking partners to become subjects to their rules. Platforms, then, urge their users to absorb these ideologies and push fact-checking as a meaningful form of enforcement via their global prominence and their ample revenue streams that allow them to invest in fact-checking around the world. While our data and analysis focused on the U.S. context, the platforms’ fact-checking measures extend beyond the U.S. borders. Indeed, the platforms have historically tested measures in other countries prior to implementing them in the United States, although they tend to prioritize the United States in their communications (Karanicolas, 2020). The global impact of platform fact-checking is particularly apparent as different countries have grappled with their own local informational challenges related to the pandemic and elections. Even as new fact-checking organizations emerge in various countries, platforms’ investments in fact-checking grant them considerable power in asserting their own values and logics in collective processes of truth-making. As such, we contribute to theoretical discussions across platform and infrastructure studies, journalism studies, science and technology studies, among other fields, about how platforms (re)shape how we make sense of global events like the COVID-19 pandemic, though with a particular Silicon Valley ethos.
Our findings also give insight into some actions platforms could take to shift efforts from the core emphasis on regulating the visibility of mis/disinformation to increasing informedness and healthier, fact-based discourse within their sites. To do this, platforms could devote resources to not only telling people whether something is true or false, but to increasing users’ media and digital literacies via interventions within the platform interfaces. Moreover, many of the “informational resources” platforms present to users require a significant level of health and science literacy, which are exceedingly low in the United States (Salisbury, 2020). As Kearney and colleagues (2020) suggested, investing in “primary preventions” like “pre-bunking” rather than debunking once misinformation has spread are necessary to help users build these literacy skills. Platforms should develop methods of catching mis/disinformation before it goes viral. Engaging in modes of anticipatory moderation—for example, “virality circuit breakers” for false information (Simpson & Conner, 2020)—will be a critical element in helping to stop the spread of viral misinformation, and platforms must work together and collaborate on these anticipatory actions since viral content moves from platform to platform. As part of this, platforms could devote resources to identifying communities and influencers driving adjacent or related conversations about conspiracy theories. For example, identifying accounts historically responsible for significant proportions of anti-vaccine misinformation can allow for targeted fact-checking during the pandemic that preempts “superspreader” infodemic events as was seen with the Plandemic videos. This is especially critical as the United States often acts as a major exporter of online misinformation (Bridgman et al., 2021), and the three platforms investigated in this study have a responsibility, as U.S.-based companies, to prevent this.
Supplemental Material
sj-docx-1-sms-10.1177_20563051211069048 – Supplemental material for Fact-Checking the Crisis: COVID-19, Infodemics, and the Platformization of Truth
Supplemental material, sj-docx-1-sms-10.1177_20563051211069048 for Fact-Checking the Crisis: COVID-19, Infodemics, and the Platformization of Truth by Kelley Cotter, Julia R. DeCook and Shaheen Kanthawala in Social Media + Society
Footnotes
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Supplemental Material
Supplemental material for this article is available online.
Notes
Author Biographies
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
