Abstract
Alarmist narratives about online misinformation continue to gain traction despite evidence that its prevalence and impact are overstated. Drawing on research examining the use of big data in social science and reception studies, we identify six misconceptions about misinformation and highlight the conceptual and methodological challenges they raise. The first set of misconceptions concerns the prevalence and circulation of misinformation. First, scientists focus on social media because it is methodologically convenient, but misinformation is not just a social media problem. Second, the internet is not rife with misinformation or news, but with memes and entertaining content. Third, falsehoods do not spread faster than the truth; how we define (mis)information influences our results and their practical implications. The second set of misconceptions concerns the impact and the reception of misinformation. Fourth, people do not believe everything they see on the internet: the sheer volume of engagement should not be conflated with belief. Fifth, people are more likely to be uninformed than misinformed; surveys overestimate misperceptions and say little about the causal influence of misinformation. Sixth, the influence of misinformation on people’s behavior is overblown as misinformation often “preaches to the choir.” To appropriately understand and fight misinformation, future research needs to address these challenges.
Introduction
Concerns about misinformation are rising the world over. Today, Americans are more concerned about misinformation than sexism, racism, terrorism, or climate change (Mitchell et al., 2019). Most of them (60%) think that “made-up news” had a major impact on the 2020 election (Auxier, 2020). Internet users are more afraid of fake news than online fraud and online bullying (World Risk Poll, 2020). Numerous scientific and journalistic articles claim that online misinformation is the source of many contemporary sociopolitical issues while neglecting deeper factors such as the decline of trust in institutions (Benkler et al., 2018; Bennett & Livingston, 2020) or in the media (Newman et al., 2020).
Yet, the scientific literature is clear. Unreliable news, including false, deceptive, low-quality, or hyper partisan news, represents a minute portion of people’s information diet (Allen et al., 2020; Altay et al., 2022; Cordonier & Brest, 2021); most people do not share unreliable news (Grinberg et al., 2019; Guess et al., 2019); on average people deem fake news less plausible than true news (Acerbi et al., 2022); social media are not the only culprit (Benkler et al., 2018; Bennett & Livingston, 2020; Tsfati et al., 2020); and the influence of fake news on large sociopolitical events is overblown (Guess et al., 2020; Mercier, 2020; Nyhan, 2020).
Because of this disconnect between public discourse and empirical findings, a growing body of research argues that alarmist narratives about misinformation should be understood as a “moral panic” (Anderson, 2021; Carlson, 2020; Jungherr & Schroeder, 2021; Mitchelstein et al., 2020). In particular, these narratives can be characterized as one of the many “technopanics” (Marwick, 2008) that have emerged with the rise of digital media. These panics repeat themselves cyclically (Orben, 2020), and are fueled by a wide range of actors such as the mass media, policymakers, and experts (Cohen, 1972, p. 8). A well-known example is the broadcast of Orson Welles’ radio drama
The “Myth of the
Frequently Encountered Misconceptions About Misinformation in Column Two, and Misinterpretation of Research Results Associated With These Misconceptions in Column Three.
Misconceptions About the Prevalence and Circulation of Misinformation
Misinformation Is Just a Social Media Problem
The internet drastically reduces the cost of accessing, producing, and sharing information. Social media reduces friction even more: the cost of connecting with physically distant minds is lower than ever. The media landscape is no longer controlled by traditional gatekeepers, and misinformation, just like any other content, is easier to publish. Today’s wide access to rich digital traces makes contemporary large-scale issues like misinformation easier to study, which can give the impression that misinformation was less prevalent in the past (compared to reliable information). Alarmist headlines about the effect of social media and new technologies are widespread, such as “How technology disrupted the truth” (Viner, 2016) or “You thought fake news was bad? Deep fakes are where truth goes to die” (Schwartz, 2018). Yet, we need to be careful when comparing the big data sets that we have today with the much poorer and smaller data sets of the past. We should also resist the temptation of idealizing the past. There was no golden age in which people only believed and communicated true information (Nyhan, 2020). Conspiracy theories spread long before social media (Allcott & Gentzkow, 2017). Misinformation, such as false rumors, is a universal feature of human societies, not a modern exception (Acerbi, 2020). Before the advent of the internet, in the 70s, rumors about “sex thieves” arose from interactions between strangers in markets and spread throughout Africa (Bonhomme, 2016). In 1969, rumors about women being abducted and sold as sex slaves proliferated in the French city of Orleans (Morin, 1969). One cannot assume that misinformation is more common today simply because it is more available and measured.
Social scientists focus on social media—Twitter in particular—because it is methodologically convenient (Tufekci, 2014). But active social media users are not representative of the general population (Mellon & Prosser, 2017), and most U.S. social media users (70%) say they rarely, or never, post about political or social issues (and only 9% say they often do so; Mcclain, 2021). Traditional media may play an important role in spreading misinformation, but because of the current focus on social media, little is known about misinformation in widely used sources of information such as television or radio. According to Google Scholar, in May 2022, only seven papers had both “fake news” and “television” in their title, compared to 578 with “fake news” and “social media” (the results are similar with misinformation, 6 vs 389). Television is a gateway to misinformation from elites, most notably from politicians (Benkler et al., 2018; Tsfati et al., 2020). For instance, Trump’s tweets in his final year of office were on cable news channels for a total of 32 h (Wardle, 2021). More broadly, the detrimental consequences of mainstream partisan media, where politicians routinely spread misinformation, are well documented (e.g., Broockman & Kalla, 2022). If political elites or partisan media did not wait for social media to spread misinformation, now they can leverage social media features (e.g., hashtags) to frame topics of discussion in online public debate and to impose their partisan agenda based (sometimes) on outright lies.
These limitations invite more nuanced views on the role of social media in the misinformation problem and suggest that more work is needed on misinformation in legacy media and offline networks.
The Internet is Rife With Misinformation
During the 2016 US presidential campaign, the top 20 fake news stories on Facebook accumulated nearly nine million shares, reactions, and comments, between August 1 and November 8 (BuzzFeed, 2016). But what do nine million interactions mean on Facebook? If the 1.5 billion Facebook users in 2016 had commented, reacted, or shared just one piece of content per week, the nine million engagements with the top fake news stories would represent only 0.042% of all their actions during the period studied (Watts & Rothschild, 2017).
According to one estimate, in the US, between 2019 and 2020, traffic to untrustworthy websites increased by 70%, whereas traffic to trustworthy websites increased by 47% (Majid, 2021), which led to the following headline: “Covid-19 and the rise of misinformation and misunderstanding” (Majid, 2021). Yet, traffic to trustworthy websites is one order of magnitude higher than the traffic to untrustworthy websites. In March 2020, untrustworthy websites received 30 million additional views compared to March 2019, whereas traffic to trustworthy websites increased by two billion (Majid, 2021).
These two examples highlight the need to zoom out of the big numbers surrounding misinformation and to put them in perspective with people’s information practices. The average US internet user spends less than 10 min a day consuming news online (Allen et al., 2020) and the average French internet user spends less than 5 min (Cordonier & Brest, 2021). News consumption represents only 14% of Americans’ media diet and 3% of the time French people spend on the internet (Cordonier & Brest, 2021). Where is misinformation in all that? It represents 0.15% of the American media diet and 0.16% of the French (Cordonier & Brest, 2021). In addition, misinformation consumption is heavily skewed: 61% of the French participants did not consult any unreliable sources during the 30 days of the study (Cordonier & Brest, 2021), and a small minority of people account for most of the misinformation consumed and shared online (Grinberg et al., 2019).
Just as the internet is not rife with news, the internet is not rife with misinformation. When people want to inform themselves, they primarily access established news websites (Fletcher et al., 2018; Guess et al., 2020), or, more commonly, turn on the TV (Allen et al., 2020). Misinformation receives little online attention compared to reliable news, and, in turn, reliable news receives little online attention compared to everything else that people do (Cordonier & Brest, 2021). People do not use the internet primarily to be informed, but rather to connect with friends, shop, work, and more generally for entertainment (Cordonier & Brest, 2021). Documenting the large volume of interactions generated by misinformation is important, but it should not be taken as proof of its predominance. Big numbers are the rule on the internet, and to be properly understood, they must be contextualized within the entire information ecosystem.
That being said, it is not always possible for researchers to consider the big picture since access to social media data is restricted and often partial. As a result, many research questions remain unexplored and conclusions are sometimes uncertain, even when private companies provide access to huge amounts of data. For instance, despite its unprecedented size, the URLs dataset released by Facebook, via Social Science One’s program, only includes URLs that have been publicly shared at least 100 times. This threshold of 100 public shares overestimates by a factor of four the prevalence of misinformation on Facebook (Allen et al., 2021). Despite our limited access to social media data, methodological improvements are possible. For example, even with partial data access, it is possible to conduct cross-platform studies, which are needed to obtain accurate estimates of misinformation prevalence and to quantify the diversity of people’s news diets (Cointet et al., 2021).
Falsehoods Spread Faster than the Truth
The idea that falsehoods spread faster than the truth was amplified and legitimized by an influential article published in
Misinterpretations of that article raise the following questions: What categories should we use? Do dichotomous measures of veracity make sense? How should we define misinformation? These past few years, important efforts have been made to refine the blurred concept of “fake news” (for a review, see Tandoc et al., 2018). In practice, though, researchers use the ready-made categories constructed by fact-checkers, putting under the same heading content that differs in terms of harmfulness or facticity.
How (mis)information is defined influences the perceived scale of the problem and the solutions to fight it (Rogers, 2020). For instance, in 2016, a widely circulated BuzzFeed article showed that fake news outperformed true news on Facebook in terms of engagement before the U.S. presidential election (Silverman, 2016). However, when excluding hyperpartisan news (such as Fox News) from the fake news category, reliable news largely outperformed fake news (Rogers, 2020). Politically biased information that is not false could have harmful effects (Longhi, 2021), but does it belong in the misinformation category? As Rogers (2021) explains, “stricter definitions of misinformation (imposter sites, pseudo-science, conspiracy, extremism only) lessen the scale of the problem, while roomier ones (adding “hyperpartisan” and “junk” sites serving clickbait) increase it, albeit rarely to the point where it outperforms non-problematic (or more colloquially, mainstream) media” (p. 2). Of course, it would be impossible to find a unanimous typology that would put every piece of news in the “right” category, but it is important to be mindful of the limitations of the categories we use.
In the first part of this article, we have highlighted three misconceptions about the prevalence and circulation of misinformation. We encourage future research to broaden the scope of misinformation studies by (1) putting misinformation consumption into perspective with broader people’s information practices, as well as (2) all their online activities, and (3) moving away from blatantly false information to study subtler forms of information disorders. (See Figure 1. From the innermost circle outward: fact-checked false news; broader forms of misinformation, including, e.g., partisan interpretations; all online information; all information, including offline.)

Contextualizing misinformation prevalence and circulation.
Misconceptions About the Impact and Reception of Misinformation
In the second part of this article, we examine misconceptions about the impact and reception of misinformation. It is not because the numbers generated by misinformation are small compared to reliable information or non-news content that we can dismiss them. Small numbers can make a difference at a societal level, but to accurately assess their impact and how they translate into attitudes and behaviors, we need to understand what they mean. Until now, studies on misinformation have been dominated by experimental approaches, self-report surveys, and big data methods reducing the reception of misinformation to mere stimulus and response mechanisms, while failing to capture audience agency and contexts. Drawing on an active audience approach, the following sections explain why certain findings from quantitative studies can lead to misconceptions about the impact and reception of misinformation (see Figure 2).

Contextualizing misinformation effect and reception.
People Believe Everything They See on the Internet
Media effects are increasingly measured via big data methods (Athique, 2018, p. 59) neglecting the agency of people and paying little attention to communicative context (Livingstone, 2019). As danah Boyd and Kate Crawford (2012, p. 666) put it, “why people do things, write things, or make things can be lost in the sheer volume of numbers.” For instance, it is tempting to conflate prevalence with impact, but the diffusion of inaccurate information should be distinguished from its reception. People are more likely to share true and false news they perceive as more accurate (e.g., Altay et al., 2021), but there is also a well-documented disconnect between sharing intentions and accuracy judgments (Altay et al., 2021). Sharing and liking are not believing. People interact with misinformation for a variety of reasons: to socialize, to express skepticism, outrage or anger, to signal group membership, or simply to have a good laugh (Acerbi, 2019; Berriche & Altay, 2020; Metzger et al., 2021; Osmundsen et al., 2021). Engagement metrics are an imperfect proxy of information reception and are not sufficient to estimate the impact of misinformation. As Wagner and Boczkowski (2019, p. 881) note, “studies that measure exposure to fake news seem sometimes to easily infer acceptance of content due to mere consumption of it, therefore perhaps exaggerating potential negative effects [. . .] future studies should methodologically account for critical or ironic news sharing, as well as other forms of negotiation and resignification of false content.”
With the datafication of society, scientists’ attention shifted from empirical audiences to the digital traces that they leave. However, we should not forget the people who left these traces in the digital record, nor neglect the rich theoretical framework produced by decades of reception studies based on in situ observations of people consuming media content (Lull, 1988; Morley, 1980). People are not passive receptacles of information. They are active, interpretative, and domesticate technologies in complex and unexpected ways (Livingstone, 2019). People are more skeptical than gullible when browsing online (Fletcher & Nielsen, 2019). Trust in the media is low, but trust in information encountered on social media is even lower. And people deploy a variety of strategies to detect and counter misinformation, such as checking different sources or turning to fact-checks (Wagner & Boczkowski, 2019). People often disqualified as “irrational” or “gullible” in public discourse, such as “conspiracy theorists” or “anti-vaxxers,” deploy sophisticated verification strategies to “fact-check” the news in their own way (Tripodi, 2018) and produce “objectivist counter-expertise” (Ylä-Anttila, 2018) by doing their “own research” (Marwick & Partin, 2022). For instance, a recent in-depth analysis of around 15,000 comments, based on mix-methods, showed how “anti-vaxxers” cite scientific studies on Facebook groups to support their positions and challenge the objectivity of mainstream media (Berriche, 2021).
Given people’s skepticism toward information encountered online and the low prevalence of misinformation in their media diet, interventions aimed at reducing the acceptance of misinformation are bound to have smaller effects than interventions increasing trust in reliable sources of information (Acerbi et al., 2022). More broadly, enhancing trust in reliable sources should be a priority over fostering distrust in unreliable sources (Altay, 2022).
A Large Number of People are Misinformed
Headlines about the ubiquity of misbeliefs are rampant in the media and are most often based on surveys. But how well do surveys measure misbeliefs? Luskin and colleagues (2018) analyzed the design of 180 media surveys with closed-ended questions measuring belief in misinformation. They found that more than 90% of these surveys lacked an explicit “Don’t know” or “Not sure” option and used formulations encouraging guessing such as “As far as you know . . .,” or “Would you say that . . .” Often, participants answer these questions by guessing the correct answer and report holding beliefs that they did not hold before the survey (Graham, 2021). Not providing, or not encouraging “Don’t know” answers is known to increase guessing even more (Luskin & Bullock, 2011). Guessing would not be a major issue if it only added noise to the data. To find out, Luskin and colleagues (2018) tested the impact of not providing “Don’t know” answers and encouraging guessing on the prevalence of misbeliefs. They found that it overestimates the proportion of incorrect answers by nine percentage points (25 to 16), and, when considering only people who report being confident in holding a misperception, it overestimates incorrect answers by 20 percentage points (25 to 5). In short, survey items measuring misinformation overestimate the extent to which people are misinformed, eclipsing the share of those who are simply uninformed.
In the same vein, conspiratorial beliefs are notoriously difficult to measure and surveys tend to exaggerate their prevalence (Clifford et al., 2019). For instance, participants in survey experiments display a preference for positive response options (yes vs no, or agree vs disagree) which inflates agreement with statements, including conspiracy theories, by up to 50% (Hill & Roberts, 2021; Krosnick, 2018). Moreover, the absence of “Don’t know” options, together with the impossibility to express one’s preference for conventional explanations in comparison to conspiratorial explanations, greatly overestimate the prevalence of conspiratorial beliefs (Clifford et al., 2019). These methodological problems contributed to unsupported alarmist narratives about the prevalence of conspiracy theories, such as Qanon going mainstream (Uscinski et al., 2022a).
Moreover, the misperceptions that surveys measure are skewed toward politically controversial and polarizing misperceptions, which are not representative of the misperceptions that people actually hold (Nyhan, 2020). This could contribute to fueling affective polarization by emphasizing differences between groups instead of similarities and inflate the prevalence of misbeliefs. When misperceptions become group markers, participants use them to signal group membership—whether they truly believe the misperceptions or not (Bullock et al., 2013). Responses to factual questions in survey experiments are known to be vulnerable to “partisan cheerleading” (Bullock et al., 2013; Prior et al., 2015), in which, instead of stating their true beliefs, participants give politically congenial responses. Quite famously, a large share of Americans believed that Donald Trump’s inauguration in 2017 was more crowded than Barack Obama’s in 2009, despite being presented with visual evidence to the contrary. Partisanship does not directly influence people’s perceptions: misperceptions about the size of the crowds were largely driven by expressive responding and guessing. Respondents who supported President Trump “intentionally provide misinformation” to reaffirm their partisan identity (Schaffner & Luks, 2018, p. 136). The extent to which expressive responding contributes to the overestimation of other political misbeliefs is debated (Nyhan, 2020), but it is probably significant.
Solutions have been proposed to overcome these flaws and measure misbeliefs more accurately, such as including confidence-in-knowledge measures (Graham, 2020) and considering only participants who firmly and confidently say they believe misinformation items as misinformed (Luskin et al., 2018). Yet, even when people report confidently holding misbeliefs, these misbeliefs are highly unstable across time, much more so than beliefs (Graham, 2021). For instance, the responses of people saying they are 100% certain that climate change is not occurring have the same measurement properties as responses of people saying they are 100% certain the continents are not moving or that the sun goes around the Earth (Graham, 2021). A participant’s response at time
So far, we have seen that even well-designed surveys overestimate the prevalence of misbeliefs. A further issue is that surveys unreliably measure exposure to misinformation and the occurrence of rare events such as fake news exposure. People report being exposed to a substantial amount of misinformation and recall having been exposed to particular fake headlines (Allcott & Gentzkow, 2017). To estimate the reliability of these measures, Allcott and Gentzkow (2017) showed participants the 14 most popular fake news during the American election campaign, together with 14 made-up “placebo fake news.” 15% of participants declared having been exposed to one of the 14 “real fake news,” but 14% also declared having been exposed to one of the 14 “fake news placebos.”
During the pandemic, many people supposedly engaged in extremely dangerous hygiene practices to fight COVID-19 because of misinformation encountered on social media, such as drinking diluted bleach (Islam et al., 2020). This led to headlines such as “COVID-19 disinformation killed thousands of people, according to a study” (Paris Match Belgique, 2020). Yet, the study is silent regarding causality, and cannot be taken as evidence that misinformation had a causal impact on people’s behavior (France info, 2020). For instance, 39% of Americans reported having engaged in at least one cleaning practice not recommended by the CDC, 4% of Americans reported drinking or gargling a household disinfectant, while another 4% reported drinking or gargling diluted bleach (Gharpure et al., 2020). These percentages should not be taken at face value. A replication of the survey found that these worrying responses are entirely attributable to problematic respondents who also reported “recently having had a fatal heart attack” or “eating concrete for its iron content” at a rate similar to that of ingesting household cleaners (Litman et al., 2020; reminiscent of the “lizardman’s constant” by Alexander, 2013). The authors conclude that “Once inattentive, mischievous, and careless respondents are taken out of the analytic sample we find no evidence that people ingest cleansers to prevent Covid-19 infection” (Alexander, 2013, p. 1). This is not to say that COVID-19 misinformation had no harmful effects (such as creating confusion or eroding trust in reliable information), but rather that surveys using self-reported measures of rare and dangerous behaviors should be interpreted with caution.
Misinformation Has a Strong Influence on People’s Behavior
Sometimes, people believe what they see on the internet and engagement metrics do translate into belief. Yet, even when misinformation is believed, it does not necessarily mean that it changed anyone’s mind or behavior. First, people largely consume politically congenial misinformation (Guess et al., 2019, 2021). That is, they consume misinformation they already agree with, or are predisposed to accept. Congenial misinformation “preaches to the choir” and is unlikely to have drastic effects beyond reinforcing previously held beliefs. Second, even when misinformation changes people’s minds and leads to the formation of new (mis)beliefs, it is not clear if these (mis)beliefs ever translate into behaviors. Attitudes are only weak predictors of behaviors. This problem is well known in public policies as the value-action gap (Kollmuss & Agyeman, 2002). Most notoriously, people report being increasingly concerned about the environment without adjusting their behaviors accordingly (Landry et al., 2018).
Common misbeliefs, such as conspiracy theories, are likely to be cognitively held in such a way that limits their influence on behaviors (Mercier, 2020; Mercier & Altay, 2022). For instance, the behavioral consequences that follow from common misbeliefs are often at odds with what we would expect from people actually believing them. As Jonathan Kay (2011, p. 185) noted, “one of the great ironies of the Truth movement is that its activists typically hold their meetings in large, unsecured locations such as college auditoriums—even as they insist that government agents will stop at nothing to protect their conspiracy for world domination from discovery.” Often, these misbeliefs are likely to be post hoc rationalizations of pre-existing attitudes, such as distrust of institutions.
In the real world, it is difficult to measure how much attitude change misinformation causes, and it is a daunting task to assess its impact on people’s behavior. Surveys relying on correlational data tell us little about causation. For example, belief in conspiracy theories is associated with many costly behaviors, such as COVID-19 vaccine refusal (Uscinski et al., 2022b). Does this mean that vaccine hesitancy is caused by conspiracy theories? No, it could be that both vaccine hesitancy and belief in conspiracy theories are caused by other factors, such as low trust in institutions (Mercier & Altay, 2022; Uscinski et al., 2022b). A few ingenious studies allowed some causal inferences to be drawn. For instance, Kim and Kim (2019) used a longitudinal survey to capture people’s beliefs and behaviors both before and after the diffusion of the “Obama is a Muslim” rumor. They found that after the rumor spread, more people were likely to believe that Obama was a Muslim. Yet, this effect was “driven almost entirely by those predisposed to dislike Obama” (p. 307), and the diffusion of the rumor had no measurable effect on people’s intention to vote for Obama. This should not come as a surprise, considering that even political campaigns and political advertising only have weak and indirect effects on voters (Kalla & Broockman, 2018). As David Karpf (2019) writes “Generating social media interactions is easy; mobilizing activists and persuading voters is hard.”
The idea that exposure to misinformation (or information) has a strong and direct influence on people’s attitudes and behaviors comes from a misleading analogy of social influence according to which ideas infect human minds like viruses infect human bodies. Americans did not vote for Trump in 2016 because they were brainwashed. There is no such thing as “brainwashing” (Mercier, 2020). Information is not passed from brain to brain like a virus is passed from body to body. When humans communicate, they constantly reinterpret the messages they receive, and modify the ones they send (Claidière et al., 2014). The same tweet will create very different mental representations in each brain that reads it, and the public representations people leave behind them, in the form of digital traces, are only an imperfect proxy of their private mental representations. The virus metaphor, all too popular during the COVID-19 pandemic—think of the “infodemic” epithet—is misleading (Simon & Camargo, 2021). It is reminiscent of outdated models of communication (e.g., “hypodermic needle model”) assuming that audiences were passive and easily swayed by pretty much everything they heard or read (Lasswell, 1927). As Anderson (2021) notes “we might see the role of Facebook and other social media platforms as returning us to a pre-Katz and Lazarsfeld era, with fears that Facebook is “radicalizing the world” and that Russian bots are injecting disinformation directly in the bloodstream of the polity.” These premises are at odds with what we know about human psychology and clash with decades of data from communication studies.
Conclusion
We identified six misconceptions about the prevalence and impact of misinformation and examined the conceptual and methodological challenges they raise. First, social media makes the perfect villain; but before blaming it for the misinformation problem, more work needs to be done on legacy media and offline networks. Second, the misinformation problem should be evaluated at the scale of the information ecosystem, for example, by including news consumption and news avoidance in the equation. Third, we should be mindful of the categories that we use, like “fake” versus “true” news since they influence our results and their practical implications. Fourth, more qualitative and quantitative reception studies are needed to understand people’s informational practices. Digital traces do not always mean what we expect them to, and often, to fully understand them, fine-grained analyses are needed. Fifth, the quantity of misinformed people is likely to be overestimated. Surveys measuring misbeliefs should include “Don’t know” or “Not sure” options and avoid wording that encourages guessing. Sixth, we should resist monocausal explanations and blaming misinformation for complex socioeconomic problems. People mostly consume information they are predisposed to accept; this acceptance should not be conflated with attitude or behavioral change. More broadly, conclusions drawn from engagement metrics, online experiments, or surveys need to be taken with a grain of salt, as they tend to overestimate the prevalence of misbeliefs, and tell us little about the causal influence of misinformation and its reception. In the lines below, we detail practical avenues for future research to overcome some of these challenges.
We need to move away from fake news and blatantly false information as their prevalence and effects are likely minimal. Instead, we should investigate subtler forms of misinformation that produce biased perceptions of reality. For instance, partisan media use true information in misleading ways by selectively reporting on some issues but not others and framing them in a politically congruent manner (Broockman & Kalla, 2022).
We need to better interpret data collected via computational methods or self-report surveys. So far, misinformation studies have been dominated by big data methods or experimental approaches. This research could benefit from being completed by active audience research (Livingstone, 2019). For instance, instead of focusing on media effects where participants are often implicitly depicted as passive, it would be more fruitful to study the different ways people use the information they consume online. Similarly, digital ethnographies are needed to understand the reception of misinformation in our complex digital ecosystem.
We need to rely on more complex models of influence. Metaphors like “contagion” and “infodemic” provide an overly simplified—and incorrect—view of how influence works (Simon & Camargo, 2021). People do not change their minds when exposed to “fake news” on social media. Social media does not turn people into conspiracy theorists. Individual predispositions, such as conspiracy mentality, mediate the relationship between social media use and belief in conspiracy theories (Uscinski et al., 2022b). These mediating variables, which explain for instance why some people but not others visit untrustworthy websites, should be investigated further.
We need to ask better research questions. The misconceptions outlined above (mis)guided some research agenda by focusing excessively on why people “fall for fake news” or on how to make people more skeptical. Yet, most people do not fall for fake news and are (overly) skeptical of news on social media. This excessive focus on biases, common in psychology, left some pressing questions unanswered. For example, we know very little about the cognitive and social tools that people use to verify information online. Future studies should move from a perspective centered on the identification of isolated failures of individual cognition to an approach attentive to audience critical skills. Our digital environment empowered us with resources such as fact-checking websites, search engines and Wikipedia, making it easier than ever before to verify content encountered online and offline. Yet, research on the uses (and misuses) of these new information search and verification tools is drastically lacking.
The study of misconceptions does not entail a focus on misinformation. People commonly hold false beliefs and engage in detrimental behaviors because they do not trust institutions and choose not to inform themselves. Instead of focusing on the small number of people who consume news from unreliable sources, it would be more fruitful to focus on the large share of people who are overly skeptical of reliable sources and rarely consume any news (Allen et al., 2020; Cordonier & Brest, 2021). Despite legitimate concerns about misinformation, people are more likely to be uninformed than misinformed (Li & Wagner, 2020), even during the COVID-19 pandemic (Cushion et al., 2020).
To appropriately understand and fight misinformation, it is crucial to have these conceptual and methodological blind spots in mind. Just like misinformation, misinformation on misinformation could have deleterious effects (Altay et al., 2020; Jungherr & Schroeder, 2021; Miró-Llinares & Aguerri, 2021; Van Duyn & Collier, 2019), such as diverting society’s attention and resources from deeper socioeconomic issues or fueling people’s mistrust of the media even more (Altay & Acerbi, 2022). Misinformation is mostly a symptom of deeper sociopolitical problems rather than a cause of these problems (Ramaciotti Morales et al., 2022). Fighting the symptoms can help, but it should not divert us from the real causes, nor overshadow the need to fight for access to accurate, transparent, and quality information.
Footnotes
Acknowledgements
We would like to thank Dominique Cardon, Héloïse Théro, and Shaden Shabayek for proofreading our manuscript and for their precious feedbacks and corrections.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Project De Facto Observatory of Information x Connecting Europe Facility-2020-FR-IA-0291 Reboot Foundation x.
