Abstract
When announcing the Sveriges Riksbank Prize in Economic Sciences in Memory of Alfred Nobel 2021, the Royal Swedish Academy emphasized how conclusions about cause and effect can be drawn from natural experiments. But what can dental research learn from this? The economist’s toolbox provides a number of methods for causal inference from observational data such as instrumental variables, regression discontinuity designs, or difference-in-differences analyses. Although the relevance of improving causal inference in dental research has repeatedly been highlighted in recent years, dental research still seems to reveal major room for improvement in the application of such methods. First, there seems to be an absence of causal literature on key essential research questions for oral health. Second, the diversity and diffusion of causal inferential methods in the dental literature seem very limited so far. Third, while dental research has widely been promoting the use of directed acyclic graphs (DAGs) to help conceptualize causal thinking, comparably little attention seems to have been paid to choosing and applying appropriate data-analytic approaches for causal inference. Fourth, similar to other fields of medicine, confusion seems to persist within the dental research community as to the use of causal language. If dental research is to secure a robust evidence base for promoting effective oral health interventions, we argue that dental research needs to move beyond its current methodological echo chamber and embrace a radically different approach to causal inference. We call for editors, reviewers, and authors to embrace a much more critically reflective approach to causal inference.
Keywords
When announcing the Sveriges Riksbank Prize in Economic Sciences in Memory of Alfred Nobel 2021, the Royal Swedish Academy emphasized that “this year’s Laureates—David Card, Joshua Angrist and Guido Imbens—have . . . shown what conclusions about cause and effect can be drawn from natural experiments. Their approach has spread to other fields and revolutionised empirical research” (Royal Swedish Academy of Sciences 2021). But what can dental research learn from this?
Causal inference refers to the process of drawing a conclusion that a specific treatment (i.e., intervention) was a “cause” of some observed “effect” or outcome (Gelman and Imbens 2013). Causal inference is highly relevant for dental research as it concerns the deciphering of mechanisms through which oral health can be influenced and mechanisms through which oral health affects people’s health and well-being. This is essential for the development, implementation, and evaluation of oral health interventions and programs.
The economist’s toolbox provides a number of methods for causal inference from observational data such as instrumental variables (IVs), regression discontinuity designs (RDDs), or difference-in-differences (DiD) analyses (Angrist and Pischke 2008; Cunningham 2021; Huntington-Klein 2021). Natural experiments can be particularly useful in the absence of randomized trials (Listl et al. 2016; Hernán and Robins 2020). Although the relevance of improving causal inference in dental research has repeatedly been highlighted in recent years, dental research still seems to reveal major room for improvement in the application of such methods (Listl et al. 2016; Raittio and Farmer 2021).
First, there seems to be an absence of causal literature on key research questions for oral health. For example, methodological gaps in studying the oral–systemic disease connection have recently been highlighted (Seitz et al. 2019; Raittio and Farmer 2021). Another example can be found in research on oral health inequalities. A large amount of evidence stems from descriptive analyses of social inequalities in oral health, and this is, of course, highly relevant to inform health policy. However, evidence on the causal mechanisms of such inequalities is also very much needed. Yet studies on causal effects of socioeconomic status on oral health remain sparse (Matsuyama et al. 2019).
Second, the diversity and diffusion of methods for causal inference from observational data seem relatively limited within the existing dental literature. It is also worth mentioning that a considerable amount of causal analyses on oral health have previously been published outside the dental literature (e.g., Shungin et al. 2015; Matsuyama et al. 2021). Fortunately, some recent applications of Mendelian randomization (a specific type of IV; see Davey Smith and Hemani 2014), propensity score matching (Rosenbaum and Rubin 1983), marginal structural models (Vanderweele et al. 2010), and g-computation (Naimi et al. 2017) seem to mirror a tendency toward more diversity in causal inferential methods in dental research (e.g., Nascimento et al. 2017; Baumeister et al. 2021; Souza et al. 2021; Wu et al. 2021). There is growing use of causal methods in the dental research field, but a wider range of alternative methodological perspectives should be welcomed and encouraged. In particular, as also highlighted by the 2021 Nobel Prize in Economics, alternative methods exist to leverage natural experiments for causal inference from observational data.
Third, while dental research has been promoting the use of directed acyclic graphs (DAGs) to help conceptualize causal thinking (Akinkugbe et al. 2016), comparably little attention seems to have been paid to choosing and applying appropriate data-analytic approaches for causal inference. Make no mistake: there is nothing wrong with promoting and using DAGs to guide causal inference. However, the use of a DAG alone does not in and by itself turn a publication into causal evidence unless supported by a suitable data-analytic approach. Vice versa, the unwary use of complex statistical methods without careful consideration of theoretical pathways to causality is equally misleading. A causal roadmap, a framework for having the question and interpretation clear, includes steps for defining the causal question, assessing its identifiability, estimating statistical parameters, and interpreting them (Petersen et al. 2014; Ahern 2018).
Fourth, similar to other fields of medicine, confusion seems to persist within the dental research community as to the use of causal language (Hernán 2018). To avoid ambiguity and confusion, causal language should be used where appropriate, and it should not be used where inappropriate. Causal and noncausal terms should not be mixed up with each other. The term causal effect should explicitly be used when a study has been using robust methods for causal inference. Where causal language is appropriate, editors and reviewers should not urge authors to remove causal wording from publications. Also note that some researchers consider the term causal association to be misleading and argue that the word association should exclusively be reserved to reporting on noncausal evidence.
If dental research is to secure a robust evidence base for promoting effective oral health interventions, we argue that dental research needs to move beyond its current methodological confinements and expand further toward a much more pluralistic approach to causal inference. Other fields of medicine increasingly exploit the benefits of causal inference from natural experiments, and dentistry cannot lag behind such methodological innovations. This is particularly relevant for complex population-level interventions that are often unamenable for testing via human-made trials.
We call for editors, reviewers, and authors as well as dental academic institutions, dental education associations, the International Association for Dental Research, and other dental research organizations to embrace a much more critically reflective approach to causal inference. Proficiency in causal inference and the appropriate use of causal language should be a core competency for every researcher. Informed by clearly formulated research questions, the full potential of available methods for causal inference should be used (including through observational data), thereby being cognizant of sound theoretical frameworks and pertinent data-analytic approaches.
Author Contributions
S. Listl, contributed to conception, design, and data interpretation, drafted and critically revised the manuscript; Y. Matsuyama, H. Jürges, contributed to conception, design, and data interpretation, critically revised the manuscript. All authors gave final approval and agree to be accountable for all aspects of the work.
Footnotes
Declaration of Conflicting Interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The authors received no financial support for the research, authorship, and/or publication of this article.
