Abstract
Effective science communication is challenging when scientific messages are informed by a continually updating evidence base and must often compete against misinformation. We argue that we need a new program of science communication as collective intelligence—a collaborative approach, supported by technology. This would have four key advantages over the typical model where scientists communicate as individuals: scientific messages would be informed by (a) a wider base of aggregated knowledge, (b) contributions from a diverse scientific community, (c) participatory input from stakeholders, and (d) better responsiveness to ongoing changes in the state of knowledge.
Keywords
Many of the pressing challenges that societies face today, from climate change to global pandemics, require decisions informed by the best available scientific evidence. Ideally, citizens should have access to good quality scientific knowledge that they can trust. However, citizens may have difficulties accessing scientific information and grasping the technical terms used. Some of the difficulty can be mitigated by a better style of science communication, for example, using clearer and jargon-free language (Hanel & Mehler, 2019; Martínez & Mammola, 2021), more intuitive presentation formats (Sirota & Juanchich, 2019), effective graphics (Harold et al., 2016), and narratives that resonate with people (Freling et al., 2020). Similarly, there is a case for supporting people’s competencies to critically engage with information (Brodsky et al., 2021; Hertwig & Grüne-Yanoff, 2017). While these aspects are important, it is also essential to consider the content of these messages: what is the best evidence and who is involved in generating it. Scientific knowledge is continually updating, and new evidence now emerges rapidly, with gaps, uncertainties, and ambiguities in the data and its interpretation. A new program of science communication is needed that can address these complexities and derive clear messages that (a) reflect the best available evidence and (b) are delivered in a way that maintains public trust.
Currently, individual scientists are incentivized to rapidly disseminate their findings, often at the expense of quality control (Higginson & Munafò, 2016). This can harm the reliability of scientific messages as well as public trust in them. Furthermore, scientific messages compete in a contested and complex online landscape that favors partisanship over reasoned debate (Lorenz-Spreen et al., 2020). Especially where evidence conflicts with political or commercial interests, organized efforts to misinform, sow public confusion, or advance conspiracy theories have distorted public discourse (Koehler, 2016), threatened evidence-based policy making (Vériter et al., 2020), and personally targeted individual prominent scientists (Mann, 2015). In this commentary, we argue that to combat the challenges of today’s information landscape, science communication must go beyond “one-person reporting” and harness the collective knowledge and expertise of many scientists worldwide to provide high quality information and engage with stakeholders. In short, we propose to approach science communication as a collective intelligence process.
In its broadest form, “collective intelligence” can be seen as a collaborative approach to problem-solving, typically supported by technological tools, which allows for real-time co-ordination and mobilization of knowledge that is distributed among many individuals (Suran et al., 2021). To some extent, the scientific process already embeds collective intelligence, as scientific knowledge is informed by reasoned argument between scientists, generating better outputs through peer evaluation and debate (Mercier, 2016). Here, we focus on harnessing the most advantageous characteristics of existing collective intelligence systems that would benefit science communication (see, e.g., online Supplementary Table). We explain why and how these characteristics could be an effective way to address specific obstacles present in the traditional, “one-person reporting” model of science communication.
Aggregating Distributed Knowledge
Collective intelligence can help science communication by aggregating knowledge that is distributed among individual scientists. First, aggregating data and evidence can build a more complete picture of the current state of scientific inquiry, leading to more confidence in the reliability of a scientific proposition. For example, distributed networks of laboratories can aggregate samples for an experimental protocol, spreading the time and labor costs of data collection and evidence syntheses (Coles et al., 2022). Monitoring and aggregating evidence can also increasingly be done in real time with new Artificial Intelligence (AI) tools, for example, using machine learning to screen databases for relevant evidence.
Second, aggregating independent expert judgments can mitigate bias in evidence interpretation and enhance accurate assessment. Furthermore, communicating judgments that fairly represent those of a collective avoids the false balance that may be presented if an audience only hears from a few, unrepresentative experts (Koehler, 2016). Showing the distribution of judgments can highlight when there is a consensus or, when judgments differ, it can illustrate the uncertainties involved in interpreting the available evidence and experts’ level of confidence in the state of knowledge. Critically, technologically supported aggregation methods allow experts to add their judgments independently, reducing the risk of biases that can be introduced through group processes.
Third, aggregating expert discourse, that is, discussion of the evidence, can showcase how reasoned argument between scientists informs scientific knowledge. This can be as critical as the evidence itself, especially in crisis situations where action must be taken as evidence emerges. New digital tools for judgment aggregation in the civic participation sphere provide comprehensive packages for debating, proposing and voting on initiatives and data (e.g., Po.lis, PSi, Loomio, Consul, Decidim). These could be leveraged for communicating scientific discourse.
There are of course costs to setting up aggregation systems. To aggregate data and evidence, protocols must be developed and shared with participating researchers. Evidence quality must also be assessed to avoid undermining the accumulated knowledge base with the inclusion of unreliable data (Royal Society, 2018). When aggregating judgments and discourse, the expertise of those who are contributing needs to be verified and contributors should be representative of their collective field of research, to avoid those with vested interests gaming the power of scientific consensus (Cook et al., 2018).
Despite the costs, aggregation is highly beneficial. Communicating in terms of the “collective accumulated evidence” shifts the message toward what the best available evidence indicates. This can help resist arguments that science has not “proved” an effect (Oreskes & Conway, 2010). It is also harder for those interested in discrediting science to carry out ad hominem attacks on collective evidence from a group of scientists (Mann, 2015). Furthermore, accumulated evidence can make a scientific consensus more visible, which is important because well-communicated scientific consensus has influenced decision-making, shifted the public’s attitudes, and strengthened calls for policy action across various domains (e.g., Bartoš et al., 2022; Budescu & Chen, 2014; Kerr & van der Linden, 2022), even for partisan individuals or those who tend to be predisposed toward rejecting scientific evidence (Lewandowsky et al., 2012). In areas where consensus has yet to form, aggregation can advance science by exposing areas in which further evidence is needed (Minas & Jorm, 2010).
Involving a More Diverse Group of Individuals
To optimize the quality of aggregated evidence and a scientific consensus, collective intelligence should increase the diversity of contributions. First, diversity in ideas (e.g., epistemic diversity) tends to invite greater scrutiny, increasing the robustness of scientific inquiry (Pesonen, 2022). Involving more diverse perspectives may help scientists challenge cognitive biases when seeking or interpreting evidence. Second, diversity in representation can boost the reach and effectiveness of science communication, especially when it comes to producing messages that the public trusts. Historically, a lack of diversity in science and research has perpetuated inequalities and contributed to the marginalization of voices from groups, such as women, minority groups, and citizens of countries in the Global South (Mertkan et al., 2017). This can undermine trust in science, especially among communities that experienced discrimination in the past (Woolf et al., 2021).
Diversity needs to be deliberately engineered because biases can easily be overlooked when values and norms are embedded into contemporary society. It is necessary to review processes, such as consensus-building, information gatekeeping, and sensemaking, and establish transparent frameworks to incorporate diversity in these processes (Thapar-Björkert & Farahani, 2019). For example, frameworks for inclusion can specify how experts will be invited or selected to contribute (e.g., by issuing invitations to all identified experts in the domain, regardless of their opinions on an issue). Although frameworks do not guarantee diversity, they make the lack of diverse representation more noticeable. A transparent framework for inclusion that discloses who the experts are and why they were chosen can also help verify expertise and avoid a “manufactured” collective scientific position from non-experts (e.g., Cook et al., 2018).
Designing for diversity in the scientific collective also requires constructive spaces for deliberation, critique, and debate—discourse that is essential to knowledge-building—which support diverse participation. These spaces should be built around critiquing ideas rather than individuals, with recognized codes of conduct for respectful engagement. They should encourage scholars with opposing perspectives to collaborate rather than compete. Although there is no existing platform yet that promotes such behavior in online academic discourse, some researchers are considering how older methods to elicit, aggregate and discuss expert opinions could be harnessed as a model for shaping scientific discourse among diverse experts. Tools to scale up such processes could soon provide online infrastructure to visualize and convey the inputs to and outcomes of the consensus.
Increasing Public Participation
By definition, collective intelligence is participatory, leveraging the involvement of many individuals to produce outputs. Thus far, we have discussed the participatory input of experts in generating scientific knowledge that underpins science messages. However, science communication should also be informed by the people it will impact (Priest, 2018). Participatory input from citizens can help shape research to address the needs of those affected by it (Bruin & Bostrom, 2013). It can also generate interest and understanding from the public in how the research is conducted and evaluated (Bonney et al., 2015), thereby building trust in scientific messages (Bedessem et al., 2021). Increasingly, technological interfaces allow the public to participate in many ways. Participation can be active, for example, by acting as “citizen scientists” (Silvertown, 2009) or a mass monitoring system. The public can also passively inform scientists through their collective online discourse: such “social listening” has enabled science communicators to tackle misinformation outbreaks by targeting information provision to the public’s needs (World Health Organization, 2021).
The accessibility of scientific findings is a precondition to harvest some of the benefits of public participation, such as a more knowledgeable citizenry. Accessibility can mean making research available. Researchers are increasingly doing so through “pre-prints,” that is, draft-level papers submitted to a publicly accessible server. In theory, this gives the public early sight of findings, but pre-prints can be confused for scientific fact or weaponized to support a certain narrative (Bajak & Howeve, 2020). Hence, they should only be considered as emerging evidence in an aggregated system, and this needs to be clearly indicated on the pre-print platforms and papers. Accessibility also means making research comprehensible. Openly published articles (pre-prints or otherwise) often remain inaccessible to the public because of their technical language and general level of complexity, limiting informed discussion of these to scientists and small parts of the public (e.g., science journalists, think-tanks, and policymakers). Increasing accessibility could involve writing plain language summaries of papers (Stoll et al., 2022). It could involve supporting citizens’ skills to engage with information, identify good quality evidence, and spot misleading argumentation (Brodsky et al., 2021; Roozenbeek et al., 2022). Scientific publications could even be augmented with technological tools that indicate how findings correspond to the broader literature or how samples should be structured for this kind of research. Accessibility could also be enhanced with collective projects to communicate the state of the evidence in comprehensible language. Ultimately, scientists have a duty to make research available and comprehensible to the public that provides them with funding and academic freedom (Greenwood & Riordan, 2001).
Improving Responsiveness
It can be difficult to identify relevant evidence and judge its quality at a given point in time when it can emerge rapidly, especially during a crisis situation where scientists may accelerate research production and dissemination (Fraser et al., 2021). Collective intelligence can leverage technology to enable real-time information monitoring, thereby enhancing the responsiveness of science communication to updates and changes. Traditional evidence syntheses are lengthy processes that often exclude the most recent studies that were not published by the time the research was conducted. In contrast, AI can enable a dynamic evidence synthesis, with some promising examples already emerging across different domains. In such systems, after having established the criteria for subsequent studies to be included, researchers can regularly monitor new publications and update their syntheses in real time.
Collective intelligence could also increase the responsiveness of evaluating new information. Emerging scientific papers typically undergo independent critique, or “peer review,” but this process is notoriously slow. During the Covid-19 pandemic, researchers collectively responded by accelerating some peer review processes and, more commonly, openly sharing early-stage research as pre-prints. Not all rapid publication was helpful to the pandemic response, but some did provide valuable updates to inform decision-making (Fraser et al., 2021). Identifying and accelerating the review of better quality pre-prints could thus improve the responsiveness of science in times of crisis. A collective intelligence system could organize and support scientific evaluation of pre-prints, for example, by identifying potential reviewers through network analysis (Rodriguez & Bollen, 2008), or detecting information manipulation and erroneous statistical analyses (Henman, 2020).
However, AI cannot fully replace the human contributions needed for quality assurance. AI-supported tools to facilitate quicker pre-print or post-publication review by the scientific collective exist, but sustaining motivation to contribute collectively to this work over the longer term is difficult. This may in part be due to a lack of incentives. For example, academics often cite lack of time as the main reason for declining reviews, but it takes much less time to review a manuscript (typically hours) than to produce a new piece of research (typically months). Despite the critical contribution of peer review to the scientific process, it is not incentivised in the publication structure, nor by most employers. The same goes for maintaining contributions to consensus-building and communicating consensus. The recent Covid-19 crisis provided a glimpse of how a motivated scientific collective could produce, evaluate, and communicate research in a highly responsive fashion. However, this effort has been hard to sustain 2 years later. Harnessing the ability of collective intelligence in responding to crises and fast-paced research thus needs an overall structural change within the scientific community to better reward collective knowledge processes over individual efforts.
Implementing Collective Intelligence in Science Communication: An Example
This commentary is itself a product of our experience harnessing collective intelligence processes to create a “Manifesto for Science Communication as Collective Intelligence.” We used group discussions and interactive online discourse via the tool pol.is to gather insights from attendees at an open virtual workshop on the topic. We then invited everyone to craft the manifesto, either as co-ordinating lead authors (“CLAs,” n = 6) or contributing authors (n = 18). CLAs collectively voted on how to organize the points raised at the workshop. Each CLA then led a group of authors to draft a section of the manifesto. The CLAs condensed this draft into its key propositions and, using pol.is, all authors voted on which propositions from the draft were critical for the manifesto. Propositions with > 60% of votes were organized into the final Manifesto, which presented eight necessary features for science communication as collective intelligence. Altogether, we engaged a diverse group of researchers, captured and aggregated their judgments and discourse in an iterative fashion, and generated a consensus for communication. The full process is shared online as part of the Manifesto (https://scibeh.org/manifesto).
Conclusion
In this commentary, we highlighted the impetus for science communication to move away from a model where scientists disseminate individual findings and adopt a collective communication program that (a) develops messages from a wider base of aggregated evidence, judgments, and discourse, (b) is informed by a diverse community, (c) involves participation from stakeholders, and (d) is responsive to ongoing changes in the state of knowledge. In the online supplementary table, we provide examples that concretize how this new program would leverage collective processes, supported by participatory technology, in pursuit of a more collaborative form of science communication. While no single example (including our own) managed to harness all the advantages we describe in this commentary, they provide a glimpse of how collective processes are already enhancing the way in which scientists gather data, reach consensus, and communicate it. We hope that in the near future, more tools and examples will emerge to support a program of science communication as collective intelligence.
Supplemental Material
sj-docx-1-scx-10.1177_10755470231162634 – Supplemental material for Science Communication as a Collective Intelligence Endeavor: A Manifesto and Examples for Implementation
Supplemental material, sj-docx-1-scx-10.1177_10755470231162634 for Science Communication as a Collective Intelligence Endeavor: A Manifesto and Examples for Implementation by Dawn Holford, Angelo Fasce, Katy Tapper, Miso Demko, Stephan Lewandowsky, Ulrike Hahn, Christoph M. Abels, Ahmed Al-Rawi, Sameer Alladin, T. Sonia Boender, Hendrik Bruns, Helen Fischer, Christian Gilde, Paul H. P. Hanel, Stefan M. Herzog, Astrid Kause, Sune Lehmann, Matthew S. Nurse, Caroline Orr, Niccolò Pescetelli, Maria Petrescu, Sunita Sah, Philipp Schmid, Miroslav Sirota and Marlene Wulf in Science Communication
Footnotes
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Holford, Fasce, Lewandowsky and Schmid were supported by the European Commission Research and Innovation Horizon 2020 Grant 964728 (JITSUVAX). Lewandowksy and Abels were supported by the European Research Council (ERC Advanced Grant 101020961 PRODEMINFO). Lewandowsky was supported by the Humboldt Foundation through a research award. Hahn was supported by the Arts and Humanities Research Council (AHRC), UKRI. Herzog was supported by Deutsche Forschungsgemeinschaft (DFG) Grant 458366841 (POLTOOLS). Lehmann was supported by the Villum Foundation (34288). Nurse was supported by the Australian Government Research Training Program (RTP) scholarship. Sirota was supported by the European Commission Research and Innovation Horizon 2020 Grant 101016967 (YUFERING).
Author Biographies
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
