Abstract
Scholars and practitioners of biosafety and biosecurity (collectively, biorisk management or BRM) have argued that life scientists should play a more proactive role in monitoring their work for potential risks, mitigating harm, and seeking help as necessary. However, most efforts to promote proactive BRM have focused on training life scientists in technical skills and have largely ignored the extent to which life scientists wish to use them (ie, their motivation). In this article, we argue that efforts to promote proactive BRM would benefit from a greater focus on life scientists' motivation. We review relevant literature on life scientists' motivation to practice BRM, offer examples of successful interventions from adjacent fields, and outline ideas for possible interventions to promote proactive BRM, along with strategies for iterative development, testing, and scaling.
Introduction
Life science research can pose risks of harm involving biological agents. These can include harms from accidental release or exposure (biosafety risks) or deliberate misuse (biosecurity risks), or from sharing “dual-use” research information that might enable others to accidentally or deliberately cause harm. 1 We will refer to these risks collectively as biorisks, and their identification, assessment, and preemptive mitigation as biorisk management (BRM).
Like any task, BRM requires both ability and motivation. By ability, we mean the knowledge and skill necessary for someone to successfully execute a behavior if they choose to attempt it. By motivation, we mean all of the internal psychological factors that lead them to make the attempt—the “why” of the behavior, rather than the “how.” 2 When we refer to life scientists being motivated to practice biorisk management, we mean to refer to the complex collection of worldview, beliefs, attitudes, and other factors that make BRM seem worth practicing. For example, motivational factors might include judgments of the tangible and intangible costs and benefits of BRM, perceptions of one's likelihood of succeeding in performing BRM, perceived social norms regarding BRM, and any underlying beliefs in premises that might justify the value of BRM. 3
Many of the motivational factors driving BRM practice rely on the existence of predefined rules enforced by external and internal oversight bodies. For example, life scientists might practice BRM to avoid penalties for failing to follow the rules. However, there are limits to external enforcement. Biorisk management sometimes depends on the voluntary choices of life scientists and staff in situations where rules cannot be realistically enforced or do not yet exist. 4
In these situations, many experts have called for life scientists to practice biorisk management proactively—that is, without the need for enforceable external oversight as a source of motivation. Consider 3 hypothetical scenarios of proactive biorisk management:
An academic life scientist is working in the laboratory alone on a weekend. There is virtually no chance that she will be observed by anyone else, but she still wears the required personal protective equipment and follows safety protocols. A researcher is leaving the laboratory after a late night of work. He is usually the last to leave, but as he starts to walk out, he notices a recently arrived visiting scientist whose work is unfamiliar to him. The visiting scientist looks stressed and is working on something at the laboratory bench. The researcher approaches the visiting scientist to say hello, learn his name, learn about his project, ask him how he's doing, and offer his support if he needs anything. The researcher asks his principal investigator about the visiting scientist the next day. As they are designing an experiment to test a novel antiviral therapy, a team of life scientists realizes that others might be able to use their technique to enhance the virulence of existing pathogens. No rule requires them to do so, but they decide to pause the work and contact the National Institutes of Health Office of Science Policy for advice about how to proceed.
By definition, proactive biorisk management requires a source of motivation beyond compliance with external oversight. However, much existing guidance and literature on promoting proactive biorisk management fails to specify this presumed source of motivation and instead focuses on training technical skills.5-15 The lack of research and guidance on motivating proactive biorisk management is particularly concerning because many life scientists may lack a source of motivation to consider the risks of their work beyond compliance with external rules.16-18
To effectively promote proactive biorisk management, we need to expand efforts beyond ability-focused concepts like “training,” “educating,” and “awareness-raising” to include concepts like “motivating,” “reframing,” “persuading,” “engaging,” and “inspiring.” This should not and does not need to be deceptive, overbearing, or unethical. Skilled teachers engage and inspire their students to develop a desire for knowledge. The same should be true for anyone seeking to engage life scientists in proactively managing biological risks. For example, training programs could include compelling explanations of the seriousness of dual-use concerns alongside examples of scientists who demonstrated foresight in anticipating those concerns.
In this article, we review arguments for the importance of proactive biorisk management, and we survey existing literature that suggests that many academic life scientists in the United States may lack strong motivation to practice it. We then contrast existing guides to biorisk management, which tend to focus on knowledge and skills, with literature from social psychology and the study of “safety culture,” which conceptualizes motivation using variables such as social norms, beliefs, and goals. We offer 3 examples of how this literature might inform solutions to concrete challenges in biorisk management. We end with reflections on implementing a research and practice program to promote proactive biorisk management at scale.
Safety and Security Challenges and Need for Proactive BRM
The challenge of biorisk management is particularly salient, given the current COVID-19 pandemic. There has been debate over whether the SARS-CoV-2 virus first infected a human in a wet market in the Wuhan province of China, a nearby research laboratory, or another location. 19 Regardless of the origins of COVID-19 and SARS-CoV-2, scholars have argued that human error and deliberate misuse in research laboratories both constitute serious pandemic threats20-22 and that dual-use research can create biosafety and biosecurity risks that are comparable in magnitude.23,24
Biorisks are managed through a “web of prevention” involving institutional, national, and international policies; tools and technical standards of practice; and social norms. 25 However, risky situations can and do arise in day-to-day laboratory work. Individual life scientists exercise a great deal of freedom in their approach to these situations because there are typically many possible experimental paths with different degrees of risk for achieving a scientific goal, and because external observers often cannot see their choices and reward or punish them accordingly. 4 This may be particularly true in cutting-edge life science research, which can create novel risks that have not yet been categorized or characterized within policies or technical standards but for which life scientists may have specialized expertise.26-30
To manage biorisks when external rules are unenforceable or nonexistent, many experts have called for life scientists to practice proactive biorisk management as part of a “safety culture” or “culture of responsibility.”27,31-36 Some have used the latter term in a way that is similar to proactive BRM.37,38 However, “culture” is an extremely broad concept that, in a biorisk management context, is often used to refer not only to life scientists' motivation, but also the formal policies and management systems that surround life scientists. 37 While these factors are certainly important, we focused our research on the factors that motivate scientists in situations where formal policies and management systems do not yet exist or are not enforceable.
This does not mean that life scientists are the sole cause of risks or that they are solely responsible for managing them. However, they are in a powerful position to take initial action. Efforts to involve frontline workers in safety risk management have shown promise in a range of fields, including aviation, healthcare, and nuclear power, by equipping and inspiring workers to speak up when they notice something unusual.23,35,39 It may be particularly valuable to engage life scientists in situations where policies do not yet exist, because their observations can inform the development of new policies and procedures.40,41
Life Scientists' Motivation to Practice Proactive BRM
Background
Biorisk management is important for preventing accidental and deliberate harm that might be caused by life science research. External oversight can motivate life scientists to practice BRM by convincing them that they will risk personal costs (such as losing their jobs) if they fail to properly manage risks, or that they will be rewarded for the proper management of risks. But in situations where external oversight rules are difficult to enforce or nonexistent, what else will motivate life scientists to practice BRM?
It can be helpful to reframe this question by asking, “What would life scientists need to believe and value in order to practice BRM proactively?” For example, they might need to believe that practicing proactive BRM will keep them or others safe from biological hazards, protect their field from bad publicity, or save time with laboratory management in the long run. 17 They might also believe that practicing proactive BRM is an interesting intellectual challenge or an expected part of being a responsible life scientist. 42
Beliefs about the perceived costs and benefits of behavior are central to most theories of motivation.3,43,44 As a simple approximation, people want to engage in a behavior when they believe that they can successfully perform it and when they believe that the benefits to them outweigh the costs relative to some valued set of goals. Psychologists have cataloged many goals that are salient in work environments, such as career progress, 45 belonging to a social group,46,47 acting in alignment with a personally valued social norm or identity,48,49 and helping others. 50 If life scientists believe that BRM serves these goals, they may be willing to practice it proactively.
However, some life scientists may also hold beliefs that make them unmotivated to practice BRM. For example, they might believe that certain risk management practices are difficult, time-consuming, unpopular, useless, or even harmful to their careers and/or society because of their perceived effects on the progress of scientific work. This is not a particularly surprising possibility; research aligns with common intuition to suggest that employees in a range of fields are often disengaged from outside efforts to promote workplace safety.51-53
We have little empirical data on life scientists' beliefs about BRM or their resulting motivation to practice it. 37 However, some initial impressions can be drawn. The literature that does exist suggests that many life scientists are uninformed, apathetic, or resistant to practicing various forms of BRM. We review literature on 3 topics: life scientists' general motivations for choosing their career as a backdrop for their potential motivation to practice BRM, their engagement with and beliefs about current biosafety rules once they have begun their research career, and their beliefs about dual-use risk management. There is a notable lack of research on life scientists' beliefs regarding the risks of deliberate theft or misuse of laboratory materials or information.
Career Motivations
In general, life scientists do not consider biorisk management as part of their motivation to enter the field. Interviews and focus group studies of life scientists have found a range of reasons why people choose to enter, including enjoyment of discovering new scientific knowledge, career mobility, enjoyment of problem-solving, and the desire for independence in their work.54,55 Once they have entered the field, academic scientists generally face strong career pressures to “publish or perish.” 56 Scientists are also somewhat motivated by the desire for prestige in their communities, 57 and scholars have noted the role of prestige in driving scientific productivity. 58 None of the motivations described here are necessarily associated with a desire to reflect on and manage potential risks.
In 2 studies,59,60 researchers found broadly similar patterns when interviewing academic life scientists about their perceptions of the societal and ethical issues involved in their work. They found that life scientists were often required to justify the societal value of their work for external grants and publications, but that many were simply unaware of potential societal concerns about their work, believed that such concerns were irrelevant to their research, believed that they were incapable of managing concerns, and/or deprioritized societal concerns to focus on conducting rigorous research that contributes new knowledge to the field.
Engagement With Laboratory Biosafety Rules
At some point, most life scientists will work in a laboratory setting where they are expected to follow safety rules. However, surveys of US laboratory scientists have found that large fractions of respondents report apathy and/or competing priorities regarding laboratory safety rules. For example, in a series of annual surveys of laboratory safety staff and scientists at a US national conference, the most commonly cited barrier to improving laboratory safety (with almost 50% of each group agreeing) was “competing priorities,” the second-most commonly cited barrier was “apathy,” and the fourth was “time and hassle factors.” 61 Schroder et al16,17 similarly found “time and hassle” and “apathy” to be the largest self-reported barriers, with one scientist writing, “If I could have selected apathy 3 times over, I would have.” They also found that 15% to 30% of laboratory scientists (the percentage varied across academia, industry, and government) believed that current safety rules negatively impacted their productivity and/or interfered with the scientific discovery process. These numbers may, in fact, be optimistic. Keiser and Payne 62 provide evidence that respondents to workplace safety surveys deliberately skew their responses to manage their social impressions.
Ultimately, apathy and competing priorities may not only affect the quality of compliance with preventive safety and security rules, but also the quality of response in the event of a laboratory accident. Across 3 studies, 25% to 38% of laboratory researchers claimed to have experienced a laboratory accident but did not report it.16,63,64 Survey findings published in 2016 flagged human error as a central cause of 78% of laboratory-acquired infections in biosafety level (BSL) 3 and 4 laboratories, closing bluntly with: “In conclusion, there is still a need to implement a culture of biosafety in the life sciences, rather than strengthen regulations.” 36
Beliefs About Dual Use
In addition to direct risks from laboratory accidents, some life science research projects could also provide others with the intellectual or physical tools to accidentally or deliberately cause harm. Identifying and managing the risks of this dual-use research is difficult, and the challenge is only increasing with the scale and sophistication of the life sciences.27,31,32,65
How do life scientists think that dual-use research should be managed, and are they motivated to manage it themselves? Answering these questions is challenging for several reasons. First, there are even fewer studies on life scientists' beliefs about dual-use issues than about biosafety, and these studies are mostly qualitative analyses of small samples.18,42,66 Second, life scientists' awareness of dual-use issues may be in a state of flux at the time of this publication as a result of recent controversies related to research being performed near the site of the first identified COVID-19 cases.19,66 Third, unlike laboratory biosafety, there is little consensus on the best practices for dual-use risk management. 32 Finally, beliefs about dual-use risks and beliefs about their proper management might be only loosely coupled. For example, life scientists might disagree about the prevalence and severity of dual-use risks, and they might independently disagree about whether those risks require more or less heavy-handed efforts at risk management or whether successful management is even possible (see Table).
Table. Common Arguments by Life Scientists Against Regulation of Life Science Research With Dual-Use Potential
Therefore, we do not make strong claims about whether life scientists are currently meeting any normative standards of engagement with dual-use concerns, and we do not assume that concern about dual-use risk translates straightforwardly into an endorsement of certain risk management strategies. Instead, we note several themes that consistently emerge across multiple studies to outline some broad claims about life scientists' opinions and feelings about dual-use risks.
The first theme is that life scientists typically express concern about the potential harms of dual-use research, but they are frequently opposed to restricting it for a variety of reasons, as summarized in the Table. The arguments on display illustrate a variety of beliefs that may affect life scientists' motivation to consider dual-use risks. We emphasize that the proper management of dual-use research is highly complex and contentious. For example, arguments about the inevitability of research, the utility of research for awareness-raising and countermeasures, and the difficulty of constraining bad actors remain topics of extensive debate.67,68 Our goal is not to evaluate the merits of each argument, but rather to contribute to a sketch of life scientists' opinions about the topic.
Second, life scientists strongly prefer to manage dual-use risks themselves, rather than to follow external oversight.42,66 There are several possible explanations for this finding. Some scientists likely support self-regulation because they believe it is more effective and/or less burdensome than risk management as performed by external regulators. Indeed, this position was taken by the authors of the original National Research Council report that laid many of the foundations for current US dual-use policy: “A system of review based in scientific self-governance can, we believe, effectively address the security risks without discouraging scientists from taking part in important biodefense research.” 27 How this occurs in practice is not well understood. In a 2007 survey, 15% of life scientists self-reported taking proactive steps to manage a potential dual-use risk of their own work, but the survey had only a 16% response rate and may have disproportionately attracted respondents who were interested in dual-use issues. 42
Some life scientists may not be particularly in favor of self-regulation or external regulation, but state a preference for self-regulation to avoid being forced to modify their work. Like other scientists, 69 life scientists also strongly support norms of open science and information sharing. They tend to strongly resist approaches to dual-use risk management that involve controlling the ability to publish.18,42,66 Indeed, some life scientists in one focus group study endorsed the idea that there is simply no research that is “inappropriate for study.” 42 Life scientists have also consistently noted in interviews that they already feel “overregulated” and that dual-use risk management tends to run counter to their own career incentives to publish high-profile research.18,42,66
In addition, virtually all the arguments listed in the Table apply equally to the self-regulation or external regulation of research. Put simply, if dual-use risks are minimal and/or unstoppable (arguments 1-4), if dual-use research has benefits that outweigh the risks (5-7 and 10), if there is no agreement about what is actually dangerous (8), or if life scientists are unlikely to be bad actors (9), then both self-regulation and external regulation appear unnecessary at best, or potentially harmful at worst.
Life scientists also frequently appear to conceptualize dual-use research as solely involving a small set of pathogens and experiments. This may be due in part to US government guidelines on “dual-use research of concern,” which defines the term broadly but flags a small set of pathogens and experiments for required review under certain conditions. 70 As a result, life scientists, even those aware of the term “dual use,” tend to believe that their own work cannot be dual use, potentially reducing their motivation to consider its risks.18,42,66 Accounts of the historic 1975 Asilomar Conference on Recombinant DNA described a similar pattern of life scientists being reticent to directly acknowledge the risks of their work and motivated to practice self-governance to avoid external oversight. 71
Taken as a whole, these initial findings paint a mixed picture of life scientists' motivation to manage the potential dual-use risks of their work. However, it is essential to recognize that the life scientists in these studies may have different ideas about what “management” entails, and that this may affect their risk perceptions. Some life scientists may see “managing” dual-use risks as a euphemism for blocking research entirely. Others may be aware of lighter-touch options, such as redesigning experimental approaches or de-emphasizing certain points in public communications. Research on the phenomenon of “solution aversion” suggests that people tend to downplay the seriousness of a problem if they believe that there is only one solution and they find it unappealing.72,73 More research is needed to untangle the dependencies between life scientists' beliefs about dual-use risk and their beliefs about the available options for managing it.
Efforts to Promote Proactive BRM Focus on Skills Over Motivation
We have argued that motivation is essential for adopting and practicing proactive BRM and have reviewed research suggesting that many life scientists may lack this motivation. This suggests that efforts to train life scientists in skills of proactive BRM should also address motivational issues. Indeed, decades of research in psychology and education have found that lessons and persuasive messages are significantly more likely to cause long-lasting changes in knowledge and behavior when the receiver is sufficiently motivated to think deeply about them.74,75
However, BRM guidebooks and articles tend to focus on training life scientists in the skills of identifying and mitigating risk, and they mostly ignore sources of motivation other than compliance with rules. For example, the US Centers for Disease Control and Prevention (CDC) authoritative biosafety manual Biosafety in Microbiological and Biomedical Laboratories 5 and the International Organization for Standardization (ISO) biorisk management standard 6 contain little mention of motivational issues. The most recent World Health Organization biosafety manual contains scattered mentions of the importance of a “safety culture” 7 and references an accompanying monograph containing only 2 pages of broad recommendations to practice “effective communication,” “active engagement,” and “encouragement” of staff. There is a similar lack of detail in guidebooks and articles describing best practices and recommendations for biosafety,8,9 biosecurity,10-12 dual-use risk management,13,14 and BRM as a whole. 15
Revealingly, a number of documents seem to mischaracterize the creation of a “culture of safety” solely as a matter of skill training, or they use the term “training” as a placeholder for a lack of detail about how exactly to engage life scientists. For example, the CDC's biosafety manual contains scattered mentions of the concept of a culture of safety, and notes that it requires “a proactive rather than a reactive approach,” but describes this “culture” as created primarily through “integrating a risk management process into daily laboratory operations,” providing “information, resources, and training,” and developing “habits and procedures through training and competency checks” among life scientist staff. 5 But it is not clear how processes and training will cause life scientists to care more about BRM or to willingly raise the alarm about novel concerns.
Similarly, the University of Texas at Arlington's biosafety manual opens with a full page dedicated to a statement from the CDC about the importance of a culture of safety, but after that it only mentions the term one more time, arguing that principal investigators “must instill a positive attitude and awareness of the Culture of Safety in their laboratory workers through training and adding discussions concerning laboratory safety in their regularly scheduled laboratory meetings.” 76 Without further explanation of their content, mandatory skill training and onerous discussions based on transparent external requirements seem just as likely to annoy life scientists as to enlighten them about the value of biorisk management.
Although less frequently emphasized than skill training, some groups have also used historical or fictional case studies to promote proactive BRM. For example, outreach efforts to life scientists from the Federal Bureau of Investigation's Weapons of Mass Destruction Directorate, Biological Countermeasures Unit, and the Engineering Biology Research Consortium use historical examples of biological catastrophes and near misses to motivate life scientists to engage with dual-use issues in their work.77,78 Similarly, Kahn 79 suggests that biosecurity could be “embedded into the culture of the life sciences” by requiring the National Institutes of Health to require grantees in relevant fields to take a course covering “a brief history of bioterrorism, the Biological Weapons Convention, and key findings and recommendations of the Fink Report.”
In theory, historical case studies could motivate life scientists by clarifying that biological catastrophes from life science research are a real and horrific possibility. However, to the best of our knowledge, no study has evaluated the effects of historical case studies or fictional scenarios on life scientists' beliefs or behavior. Given the personal motivation and career incentives that they experience to get their work done, and the relative infrequency of biological catastrophes, it is not hard to imagine that some life scientists think historical case studies do not apply to their own work, or that biological catastrophes are so rare that they are not worth worrying about.
Some biorisk management texts do at least briefly mention practices for promoting BRM that have a clearer motivational aspect. For example, the CDC's biosafety manual 5 and Higgins et al 80 argue that life science research institutions should incorporate nonpunitive mechanisms for reporting safety and security concerns, and Gaudioso et al 81 note that “From the authors' personal experience, when a director attends a biorisk management training course with the workforce, instead of just mandating it for subordinates, he or she demonstrates leadership's commitment and vision better than a memo could ever communicate.”
In our view, these examples are a start, but they merely scratch the surface of efforts to change culture. For example, leadership could:
provide more detailed information about the baseline probabilities and consequences of harmful incidents in other laboratories in order to put risks in context
prompt staff to actively elaborate on the personal and professional consequences of a serious incident, including personal injury, harm to colleagues and loved ones, disrupted scientific progress, reputational harm to the institution, and derailed career plans
actively celebrate the reporting of safety and security concerns, not just refraining from punishing it
solicit and reward ideas from staff for improving biosafety and biosecurity through engineering and/or changes to policy and practice
publicly praise their safety and security practices to external audiences
privately praise influential and respected laboratory members for their safety and security practices and ask them to continue setting an example for their peers
We discuss frameworks for considering and developing other possible approaches in more depth later, but our larger point is that there is much existing work proposing or describing the implementation of management systems and training programs, but little evaluation of how life scientists subjectively experience them or whether their motivation to practice BRM actually changes as a result. 82 A 2020 critical review of laboratory-safety research agreed: “Given the likely impact of individual biases, ensuring perfect access to information and training […] and making equipment available is not likely to change outcomes without a better understanding of the psychology of safety decision-making. […] However, to date, the champions for safety have been natural scientists and engineers whose research expertise is not in social science methodology and who may be unfamiliar with important and relevant psychological constructs.” 83 We seek to remedy this situation by bringing a more psychologically informed perspective to biorisk management.
Insights From Social Psychology and Safety-Culture Research on Motivation
Motivation is a broad and interdisciplinary concept that has been productively studied by a range of disciplines under the umbrella of the behavioral sciences, including economics, 84 sociology, 85 organizational studies, 86 neuroscience, 87 and anthropology. 88 However, we believe that for risk management professionals seeking to motivate life scientists to practice proactive BRM, one promising starting point is social psychology, which is “the study of social behavior, especially of the reciprocal influence of the individual and the group with which the individual interacts.” 89 Social psychology models human behavior using intuitively familiar concepts and then tests and applies its models with quantitative rigor and a focus on practical utility. 90 Its theories and methods are foundational in many applied psychology fields, including organizational and work psychology, marketing, aspects of educational psychology, political psychology, and health psychology. 91
In this section, we outline a simple social-psychological view of human behavior. While necessarily brief, our review highlights a host of factors that are largely neglected by current approaches to promoting proactive BRM. We expand on some of these factors in the examples in the next section.
Social psychologists tend to explain human behavior in terms of internal psychological variables such as beliefs and goals.3,92,93 Collectively, these variables can be said to constitute a person's subjective worldview—their mental models of themselves and the world around them, their own desires, and their strategies for achieving those desires. For example, a laboratory technician who voluntarily chooses to double-check a freezer inventory might do so because they believe that the personal costs of doing so are reasonable, because they believe that they will be appreciated by their peers, and because double-checking the inventory satisfies a desire to see themselves as careful and detail oriented.
As people carry out actions, they affect the social environment for others, creating the conditions for reciprocal relationships, social norms, and group coordination. For example, a principal investigator in a laboratory may choose to highlight the importance of inventory management in a laboratory meeting because they believe that laboratory staff value it and expect it to be mentioned. This might cause other laboratory members to infer that the principal investigator expects them to practice inventory management, setting a social norm that affects their behavior. In turn, this could reinforce the principal investigator's original perception that inventory management is valued in the laboratory.
In this view of the social world, people are not independent actors who will practice biorisk management once programmed with the proper skills. Instead, people actively make sense of their social environments based on (potentially mistaken) beliefs, they pursue goals in light of these beliefs, and they affect each other in complex ways that can become self-reinforcing. Social psychologists often seek to create positive feedback loops for long-term change using interviews, surveys, and randomized experiments to study and intervene in the beliefs that people infer from their environments.94,95 For example, studies have improved interpersonal relationships by inviting people with low self-esteem to reinterpret compliments from their partners as sincere 96 and have improved academic outcomes by inviting students to reinterpret the experience of academic difficulty as a normal part of learning, rather than as a sign that they are incapable of learning. 97
Insights from social psychology have already been used to promote ethical research practices in ways that might generalize to biorisk management. For example, in a study published in 2020, social psychologists sought to encourage laboratory scientists to establish formal policies for determining the order of authorship for future papers. 98 They developed a “lab-embedded discourse intervention” to promote open discussion about the topic, but they discovered in pilot work that principal investigators who were opposed to the idea could sometimes intentionally or unintentionally pressure their staff to withhold dissenting views. To disrupt this dynamic, the authors of the study recruited junior laboratory members as “peer mentors” to colead each discussion, and found large effects 4 months later on laboratory members' self-reports of using a formal authorship policy. The lessons of this study hold promise for dealing with life science principal investigators who are resistant to biosafety and biosecurity reform in their laboratories.
In the domain of biorisk management, principles and perspectives from social psychology can currently be most clearly seen in research and recommendations for promoting a safety culture.33,99 The concept of safety culture originated in models of organizational culture by Schein, 100 who was originally trained as a social psychologist. Both “organizational culture” and “safety culture” broadly refer to the artifacts, behaviors, and worldviews that are shared by members of an organization, but these constructs are defined by scholars in widely differing ways, and the field generally lacks consensus or evaluation of frameworks beyond tests of face validity. 101
Nevertheless, safety culture scholars have successfully drawn attention to the ways in which safety is valued or devalued in organizations, and they have been influential in a range of industries, such as aviation, nuclear power, industrial chemistry, and healthcare.23,102-104 Because of their focus on shared perspectives across a group, safety culture reform efforts tend to focus on promoting social norms as an intervention method. Some efforts have a more systemic, transformative character, such as attempts to create “high-reliability organizations” by intervening simultaneously in all parts of an organization. 39 They have been compared to “moving from directing one play to another”—ambitious but also difficult and expensive. 105 Other change efforts are more targeted to 1 or 2 types of individuals in an organization and tend to have a lighter touch, such as having leaders publicly and credibly endorse efforts to promote safety in their organizations.33,99,106 We discuss both approaches later, noting that social psychology offers a wider range of intervention methods that can be used to promote proactive BRM.
Psychologically Informed Approaches to Motivate Proactive BRM
In this section, we elaborate on the set of examples of proactive BRM from the Introduction. For each example, we sketch current barriers to practice and then suggest an approach inspired by existing social psychology research. When possible, we also offer potential examples of this idea already being implemented or explored. We focus here on academic life science research in the United States, although the ideas described could extend more broadly.
Example 1: Listening Tours for Proactive Biosafety
Workplace safety researchers argue that laboratory scientists and their institutional biosafety staff should maintain strong working relationships.33,99 However, scientists can sometimes be guarded about interacting with biosafety staff to avoid bureaucratic barriers to their work. Scientists can also sometimes overestimate biosafety staff departments' levels of knowledge, capacity, and influence. This can lead to several problems. First, scientists can mistakenly offload responsibility for laboratory safety onto staff and fail to be proactive about safety themselves. Second, if scientists hold unrealistically high expectations of staff, and those expectations are violated, they could unjustifiably lose respect for staff. 33
While providing more skill training and resources for biosafety staff may be necessary to improve their capacity in the long run, a social-psychological approach might focus on improving scientist–staff relations by preemptively correcting mistaken assumptions on both sides. One example approach could be for biosafety staff to conduct periodic listening tours with life science laboratories. Staff could join existing laboratory meetings to introduce themselves, assure laboratory members that they are not conducting an audit, and ask scientists to teach them about the safety risks involved in their subfield (not necessarily their particular laboratory). Staff could close the conversation by thanking the group and requesting advice on how to reduce the burdens of risk management and engage other life scientists about the importance of laboratory safety. This request is derived from a persuasion technique known as self-persuasion, which can be effective because it invites participants to actively generate a message that they themselves find persuasive.107,108
By positioning themselves as learners, biosafety staff members can accomplish several psychologically potent goals simultaneously. They can send the message that they are not omnipotent, frame scientists in a position of responsibility and authority regarding laboratory safety, and convey the potential for a friendly, collaborative relationship in the future. They also give life scientists concrete practice thinking about how their own work could be unsafe without fear of being audited, and may give staff valuable information about novel safety risks and ways of making risk management less costly.
One potential example of this approach in practice can be found at the University of Wisconsin, Madison, which oversees a large and complex life science research infrastructure. The University of Wisconsin institutional biosafety committee conducts outreach visits to life scientists with the goals of establishing caring and friendly relationships and positioning themselves as helpful supporters of scientists' own values of personal safety. According to staff reports, the upfront work of building relationships pays off later with smoother future interactions and a stronger safety culture. 109
Example 2: Social Support Services for Proactive Biosecurity
Insider threats—when laboratory staff deliberately misuse biological agents, leak information, or otherwise compromise laboratory security—are a rare but frightening occurrence. A 2015 historical analysis of 93 biosecurity incidents at research institutions in the United States estimated that the relative risk from insiders was higher than the risk from outsiders.24,110 Insiders can perpetrate crimes themselves or intentionally or unwittingly assist outsiders, and they might be scientists or other staff members who have access to laboratory facilities.
Insider threats are also difficult to combat. 111 After the 2001 US anthrax attacks, 4 national studies recommended strategies to address the problem, including increased physical security, more carefully controlled access to pathogens, and improved screening for personnel reliability.34,111-113 Although valuable, these strategies are also controversial because they each suffer from security flaws and/or impose substantial costs on the ability of life scientists to conduct research. 114
At least 2 of the 4 reports also emphasized the importance of a “culture of responsibility” in high-biocontainment government laboratories, in which life scientists proactively monitor each other for potential insider threats.34,113 However, such a culture would likely be unpopular among many life scientists. Berger 110 noted that “concerns about policing peers or having the appropriate whistleblower protections have been raised by the scientific community for more than a decade.” Few people wish to imagine the possibility that their colleagues could be potential criminals, much less level an accusation.
However, one key insight from social psychology is that the same behavior can be more or less appealing depending on the purposes or meanings in which it is framed.94,115 Life scientists might see the act of monitoring their peers' behavior as caring, rather than accusatory, if it was done in a spirit of social support for personal stressors or other difficulties. In other words, peer-led social support services could give life scientists an alternate reason to practice peer monitoring that could help prevent potential insider threats while being less uncomfortable and directly benefiting those who are in need of support.
There are many drivers of insider threats, but personal difficulties and stressors, work conflicts, and/or social isolation often appear to play amplifying roles.110,116 Personal stressors also amplify biosafety risks. 80 Institutional social support services such as mentorship and counseling may help staff manage these stressors, which could support staff wellbeing as well as biosafety and biosecurity. BSL-3 and BSL-4 laboratories at the US National Biodefense Analysis and Countermeasures Center and the National Institutes of Health already use such services as part of their personnel reliability programs,80,117 and Gelles et al 116 argue that they play a critical role in mitigating insider threats.
Social support services are also scalable. Berger 110 notes that “most, if not all, major research institutions in the United States” already have some employee and student assistance programs. There is an increasing recognition that the competitive and stressful work environment of academia contributes to widespread mental health issues, 118 and many schools are already increasing their institutional support for mental health with on-campus resources. 119 Even so, more work is needed to integrate peer monitoring into social support services at research institutions. For example, scientists and staff might receive brief reminders of the existence of social support services along with messages dispelling any stigma around using or recommending them.
Example 3: Social Norm Approaches for Proactive Dual-Use Research Risk Management
Life scientists may have a valuable role to play in the “web of prevention” by proactively reflecting on whether their work creates novel dual-use risks. However, as noted earlier, surveys and interviews with life scientists in the United States, United Kingdom, and Switzerland suggest that many life scientists have never considered that their own research might be dual use, are leery of additional regulatory burdens slowing down their research, and are skeptical of attempts to control the advancement of science.18,42,66 Representatives from academic institutions at a 2017 stakeholder workshop held by the National Institutes of Health on the 2014 US Dual Use Research of Concern policy concurred, noting that it was rare in practice for life scientists to proactively flag their own research as concerning. 109
Indeed, the request being made of life scientists in managing dual use is substantial: proactively consider whether your own research projects could be misused to cause broad potential harm, and if so, submit them for further examination by an external body that may delay or even block your work. Nevertheless, there are scattered examples of life scientists proactively reporting dual-use concerns.109,120,121 What can be done to encourage more life scientists to adopt this responsibility?
One promising set of possibilities involves advocating for a social norm in which the evaluation of dual-use concerns is seen as part of what it means to be a responsible and respected life scientist. Norms are among the oldest and most well-studied constructs in social psychology,122-124 and a number of conceptual tools and methods can be brought to bear on the topic. We review 3 approaches, emphasizing that none are sufficient and that more research is needed to evaluate their potential.
One approach may involve a choice of language. Husbands 125 has noted that dual-use concerns are often framed through a lens of “security,” a term that is alien to the lived experience of many life scientists, when they could also be viewed as “responsible science,” or a more intrinsic part of the duty of a scientist. Social psychologists have found examples of small differences in language framing that can have surprisingly large effects on behavior via the subjective meanings that people make of their experiences.96,126,127
A second approach may involve pluralistic ignorance, the state in which individuals underestimate the extent to which their privately held views are shared within their group.124,128 Research has found that when people are mistaken about the popularity of a belief or behavior that they privately hold, simply informing them of the true distribution of opinions can sometimes shift their perception of the norm and thereby change behavior. 129 Many life scientists already appear to endorse the idea of self-regulating dual-use concerns in surveys and interviews,18,42,66 but it is unclear whether they are aware that other scientists like them share their views. Surveys could collect and share opinions about dual-use issues among life scientists to point out that support for self-regulation is more widely supported than people might individually expect.
A final approach might involve linking to existing norms of rigorous research. Scientists already strongly endorse a norm of “organized skepticism.”69,130 They recognize that even though they likely believe in the quality of their own work, science as a collective enterprise functions more effectively when they submit their work for peer review. By a similar logic, it might be argued that the life sciences will function more effectively when scientists avoid catastrophic dual-use risks, but that they need to submit their work to risk analysis by their peers to do this effectively. Program developers could use these arguments to convince life scientists to participate in networks of dual-use peer evaluation and to conduct novel research to inform biorisk management efforts.
The 3 approaches described could be operationalized in the form of brief, scalable online reading and writing activities modeled after the interventions used by social psychologists to trigger lasting behavior change. 94 For example, life scientists might be invited by their academic institution or professional society to complete survey questions about how they think dual-use concerns should be handled, and then shown empirical distributions of the answers of their peers. They might also be invited to write a response to a question like “How are dual-use concerns related to what it means to you to be a life scientist?” and shown vivid and representative examples of other responses from members of their field—an approach that has been successful at shifting beliefs at scale. 131
Implementation
This section offers several suggestions for implementing a program of applied research and practice to promote proactive biorisk management among life scientists. These suggestions are intended to complement existing guidelines for creating a safety culture.37,99 While these guides may offer helpful broad advice for quickly taking action across an institution, they tend to lack mechanisms for testing and feedback, which are critical for ensuring that programs are compelling for life scientists.
Specify Target Behaviors
To leverage the strengths of social psychology most effectively, it is important to clarify target behavioral outcomes. Biorisk management scholars and professionals should be more specific about operationalizing a “culture of responsibility” in terms of a set of broadly helpful behaviors, who should perform them, and under what circumstances. This is particularly true for managing the risks of novel and potentially risky research. As biorisk management scholar Sonia Ben Ouaghram-Gormley noted in 2020, “That should probably be the next priority for the National Academies in cooperation with their European and Asian counterparts: identify a decision tree to help scientists decide what to do when they hear about dangerous research.” 132
Seek to Understand the Social Environment With Surveys, Interviews, and Focus Groups
Context matters. The social environments and subcultures of life science departments vary, and the history of behavior change and culture change efforts is littered with unexpected flops and failed replications.115,133-135 Effective interventions tend to be psychologically precise in their meaning and in the time and place of their delivery. They might look insignificant, but can feel significant because they resonate with people's lived experiences. 136
Creating effective interventions therefore requires site-specific qualitative and quantitative data. As noted earlier, there is a lack of research on the beliefs of life science stakeholders regarding risks and risk management. New research will require funding and partnerships among social scientists, life scientists, and biorisk professionals. Researchers should observe, interview, and survey life scientists and neighboring groups, such as funders, publishers, regulatory staff, and institutional leadership. Researchers should pay particular attention to groups that are most important for establishing incentives and social norms for life scientists, such as funders, publishers, and professional societies. Researchers should also look for key moments in life scientists' careers that tend to establish perceptions of social norms, such as onboarding, laboratory meetings, biosafety officer interactions, conferences, and initial interactions with funders and publishers. These moments can create first impressions that persist over time, making them key sites for intervention. 46
Start Small, Then Develop and Deploy Changes With Rigor and Sensitivity
Using their initial findings, researchers and program developers should rigorously pilot test interventions on a small scale and engage in dialogue with stakeholders throughout to avoid unintended meanings, backfiring, or other downsides. 115 Intervention development also need not be centralized or siloed. One promising approach may be for biosafety offices and life science departments to partner with social science departments in universities to develop and test more compelling biosafety training modules and then share their efforts through broader networks. For example, Arizona State University has been developing and freely sharing biosafety training videos in partnership with biorisk management scholars and the Association for Biosafety and Biosecurity. 137
Offer Resources and Support to Sustain Existing Interventions and Create Conditions for New Interventions
Social-psychological interventions have been compared to seeds that need the right soil to thrive. 138 Even if an intervention resonates with someone personally, they might need additional resources or social support from the environment for it to stick. In a supportive environment, the right message can have outsized self-reinforcing effects, but in an inhospitable environment or from the wrong source, the most compelling message will have little effect. 139 Once again, program developers should be sensitive to differences in context. For example, a well-designed biosafety onboarding module might fall flat in a laboratory where the principal investigator ignores safety rules, and it may not be actionable in a laboratory that cannot afford appropriate personal protective equipment. This is one rationale for the more systemic change efforts recommended by safety culture guides.33,99
In addition to supporting existing interventions, resources and social support can also create the conditions for new intervention ideas to emerge. Funders can play an important role in this process. For example, they might follow the reproducible-science community model by partnering with publishers to release calls for applied biorisk management research that pairs life scientists and social scientists. 140 Research projects that qualify can be rewarded with higher chances of funding and/or publication. The International Genetically Engineered Machine (iGEM) synthetic biology competition can also serve as an example of a testbed for providing incentives for engagement with safety and security and encouraging interdisciplinary collaboration. 141
Conclusion
As life science research and technologies continue to advance, the risks from accidental or deliberate misuse become more complex and difficult to anticipate, and the need for life scientists to practice proactive biorisk management only increases. We have argued that life scientists need to be motivated to practice proactive biorisk management and that social psychology offers some useful tools for conceptualizing and cultivating this motivation. These tools could help meet calls for a culture of responsibility that have only intensified over the last several decades, and could offer a productive complement to policy-focused approaches to biorisk management.
Footnotes
Acknowledgments
The authors would like to thank Rob MacCoun, Rosanna Guadagno, Connor Hoffmann, Melissa Salm, Kathryn Brink, Ariel Zucker, Geetha Jeyapragasan, Jonas Sandbrink, Jessica Martin, Mildred Cho, David Manheim, Jonathan Klane, and our anonymous reviewers for their valuable feedback on this project.
