Abstract
Preregistrations—records made a priori about study designs and analysis plans and placed in open repositories—are thought to strengthen the credibility and transparency of research. Different authors have put forth arguments in favor of introducing this practice in qualitative research and made suggestions for what to include in a qualitative preregistration form. The goal of this study was to gauge and understand what parts of preregistration templates qualitative researchers would find helpful and informative. We used an online Delphi study design consisting of two rounds with feedback reports in between. In total, 48 researchers participated (response rate: 16%). In round 1, panelists considered 14 proposed items relevant to include in the preregistration form, but two items had relevance scores just below our predefined criterion (68%) with mixed argument and were put forth again. We combined items where possible, leading to 11 revised items. In round 2, panelists agreed on including the two remaining items. Panelists also converged on suggested terminology and elaborations, except for two terms for which they provided clear arguments. The result is an agreement-based form for the preregistration of qualitative studies that consists of 13 items. The form will be made available as a registration option on Open Science Framework (osf.io). We believe it is important to assure that the strength of qualitative research, which is its flexibility to adapt, adjust and respond, is not lost in preregistration. The preregistration should provide a systematic starting point.
Introduction
Preregistrations—records made a priori about study designs and analysis plans and placed in (open) repositories—are thought to strengthen the credibility and transparency of research. Firstly, the openness of this information about the study encourages the researcher to carefully reflect on different study aspects and to systematically report on their design and analysis choices, including those made as the study progresses (Kern & Gleditsch, 2017; Wagenmakers & Dutilh, 2016). Secondly, the records about the study design and analysis plan help the reviewer or user of the study in understanding and assessing the study’s findings, because the preregistration provides a structured insight into how the study was thought out and set up (Haven & van Grootel, 2019; Nosek et al., 2018).
The importance of preregistration is recognized by different stakeholders, such as journals, funders, and research communities. This is reflected in increasing journals and funders attention toward the topic (Arnold Foundation, 2020; Nosek et al., 2015; Open Science Framework, 2020). In addition, different research communities have advocated for preregistration (Bonniaud et al., 2018; Glasziou et al., 2014; Humphreys et al., 2013; Ioannidis, 2005; Miguel et al., 2014; Munafò et al., 2017; Nosek et al., 2018; Picciotto, 2018) and developed registries based on the type of study (e.g., clinicaltrials.gov for randomized clinical trials, PROSPERO for systematic reviews) or a field of study (e.g., EGAP for governance and politics, RIDIE for development-impact studies). Different forms are used for different types of studies and some platforms welcome preregistrations from different fields using the same form (e.g., aspredicted.org) or have multiple forms available for researchers to choose from (e.g., osf.io).
Preregistering qualitative research may seem counterintuitive as one of the functions of preregistration is to distinguish exploratory and confirmatory research, and qualitative research is often exploratory by nature (Humphreys et al., 2013; Nosek et al., 2018). Indeed, preregistration may not be useful for all forms of qualitative research. Nonetheless, the usefulness of preregistration may extend more widely than is immediately apparent. For one thing, preregistration may be helpful for forms of qualitative research that involve an element of testing. 1 The posting of predictions in a timestamped registration allows the researcher to credibly communicate to audiences that the hypotheses being assessed were specified in advance of seeing the evidence, and to clearly identify deviations from the plan as exploratory undertakings. Yet preregistration can also bring greater clarity to research that operates on a more exploratory or iterative logic. Even when exploring, scholars usually bring prior knowledge or theoretical preconceptions to the study, and preregistration can help distinguish which aspects of a study’s findings drew on—and which diverged from—those initial beliefs or expectations. For both sides of the spectrum, preregistration intends to make visible the connections between analytical assumptions, evidence, and decisions that form a particular interpretation of the data.
As the idea of preregistration has been extended to qualitative research, different authors have put forth suggestions for what to include in a qualitative preregistration form (Haven & van Grootel, 2019; Jacobs, 2020; Kern & Gleditsch, 2017; Piñeiro & Rosenblatt, 2016). These suggestions built on existing strategies to foster the credibility of qualitative research, such as the audit trail (Cohen & Crabtree, 2006; Lincoln & Guba, 1985; Miles & Huberman, 1994; Miller, 1997; Schwandt & Halpern, 2011), the decision trial (Koch, 2006), and reporting guidelines (Malterud, 2001; O’Brien et al., 2014; Tong et al., 2007).
However, the proposed items for qualitative preregistration thus far are not based on systematic empirical investigations within the qualitative research community. If the preregistration form is to benefit the qualitative community, it should reflect what they as a community consider relevant, comprehensive, and comprehensible for qualitative study preregistration (Terwee et al., 2018). Tools to foster transparent reporting in qualitative research, such as reporting guidelines (e.g., COREQ, JARS-APA), are often leveraged when the study is ready to undergo peer review. What is distinctive about preregistration is that it involves the researcher declaring plans for a study design, data collection strategy, and set of empirical expectations in advance, thus allowing for a clearer and more credible delineation between ex ante beliefs and plans, on the one hand, and actual study implementation and ex post conclusions, on the other hand. Unlike an ex post checklist, preregistration makes more transparent to readers how a study unfolded and beliefs shifted over time. In this sense, preregistration can be thought of as a systematic start of a research log (Piñeiro & Rosenblatt, 2016). As such, preregistration acts as a complement to reporting guidelines, although it is unclear what aspects of existing reporting guidelines can be amended for a preregistration form. Hence, we consider it pivotal to develop a preregistration form in close consultation with the qualitative community.
The goal of this study is to gauge and understand what parts of preregistration templates qualitative researchers would find helpful and informative. Given the diversity of epistemological starting points and methods used in qualitative research, we do not expect to be able to develop a form suited for all qualitative scholarship. We tried to accommodate this diversity by inviting qualitative researchers across a broad range of disciplines and approaches to qualitative research, ranging from social constructivism to (post) positivism. We employ the Delphi method to develop a template on which there is high agreement among these surveyed participants, while still retaining meaningful structure. At the same time, the Delphi method can only achieve agreement among participating experts; thus, we can make no claims about how generally useful the form might be to qualitative research communities as a whole. The preregistration form is intended to function as a tool that will aid research planning and communication, not a prescription that hampers the flexible nature of qualitative research (Wagenmakers & Dutilh, 2016). This form will be made available with other preregistration forms available on OSF to foster the credibility and transparency of qualitative inquiry.
Methods
Ethics
The Institutional Review Board for the Social and Behavioral Sciences at the University of Virginia approved our protocol (# 3397).
Study Design
We used an online Delphi study design. Delphi uses repeated questionnaires with feedback reports in between to attain agreement on a particular topic (Okoli & Pawlowski, 2004; Rowe et al., 1991). The Delphi design was appropriate because it provided structured group communication in order to find a set of items where there has so far been little convergence on applied principles—the Delphi terminology refers to this as “agreement” of the surveyed participants (Linstone & Turoff, 2002; Powell, 2003). In addition, Delphi’s online nature allowed us to reach out to a dispersed group of international researchers and the pseudo-anonymized form precludes particular people from dominating the debate (Graham et al., 2003).
Our Delphi consisted of two rounds with feedback reports, see Figure 1. Panelists’ feedback was pseudo-anonymized and panelists received no financial compensation.

Flowchart of study design.
Panelists
We invited 295 active qualitative researchers and qualitative methodologists to participate. As alluded to above, we wanted to represent this group as broadly as possible and to attain this, we used four different recruitment methods.
First, the steering committee and research team members suggested qualitative researchers or qualitative methodologists from their own networks (
Steering Committee
The steering committee consisted of the day-to-day team as well as advisory committee members. The day-to-day team implemented the survey (TH with support from WM and TME) in close association with the advisory committee members (KSG, LvG, AJ, FK, BN, RP and FR). The members of the steering committee were trained in different disciplines that make use of qualitative research methodology. The steering committee guided the execution of the study. They helped in selecting potential panelists, advised on the design and content of the questionnaires, and on the analysis of the responses. They did not take part in the Delphi. However, if the panel did not converge after round 2, the steering committee made the final decision about the item.
Creation of the Questionnaire
Our strategy to create the list of proposed items was fourfold. First, we integrated suggestions from existing works on preregistration of qualitative research (Haven & van Grootel, 2019; Jacobs, 2020; Kern & Gleditsch, 2017) and pre-analysis plans of qualitative research (Piñeiro & Rosenblatt, 2016). Some authors’ suggestions are influenced by the logic of preregistration for quantitative work (Jacobs, 2020; Kern & Gleditsch, 2017), others are influenced by existing strategies to foster the credibility of qualitative research such as a decision trail or audit trail (Denzin & Lincoln, 1998; Koch, 2006).
Second, we searched in PubMed and PsycINFO using the terms “qualitative research” and “preregistration” at the end of January 2020. We included studies if they regarded views on qualitative preregistration and excluded studies that only mentioned preregistration as an example. Articles had to be written in English. PubMed returned 88 hits. After screening the titles and abstracts, we identified one conference abstract that was relevant (Bowers, 2019). One paper discussed open science practices in the field of aging sciences and used the publication by Haven & van Grootel (Haven & van Grootel, 2019) as an example (Isaacowitz & Lind, 2019). PsycINFO returned 10 hits where after screening of the titles and abstracts, 1 paper was relevant that discussed open science in Sport Psychology (Tamminen & Poucher, 2018) focusing on qualitative inquiry.
Third, we downloaded 21 reporting guidelines that EQUATOR listed as bearing relevance to qualitative research (see https://www.equator-network.org/reporting-guidelines-study-design/qualitative-research/?post_type=eq_guidelines&eq_guideli 2 . We included these because preregistrations and reporting guidelines are complementary. Whereas preregistration is intended to help design a study in the most rigorous and transparent way, reporting checklists are intended to write up a study in the most understandable and transparent way. Hence, it is desirable that there is some overlap. Additionally, we hand-searched reporting guidelines for qualitative research in psychology and social sciences and screened the resulting Journal Article Reporting Guidelines for Qualitative research (Levitt et al., 2018) as well as guidelines for qualitative research submitted to the Journal of the Society for Social Work and Research (Wu et al., 2016).
Finally, the above strategies resulted in a document with 32 proposed items that was put forth to the steering committee for review. Steering committee members suggested combining of similar issues (nine items), removal of items that were considered off the subject (e.g., item that regarded ethical testing or reporting) (two items), and rephrasing of particular items for clarity. The list of 21 remaining proposed items can be found in Supplemental Appendix 1 and formed the basis for the first Delphi questionnaire. We used Qualtrics (Qualtrics, Provo, UT, USA) to create and disseminate the survey.
Procedure
Round 1
We sent an invitation email to panelists in February 2020, the email linked to the questionnaire, the study protocol, and the project’s website. The questionnaire for round 1 took between 40 and 60 minutes to complete.
Panelists were directed to a personal online survey where they received additional information on the background and content of the questionnaire. They then had to indicate informed consent before starting the actual questionnaire. After the digitally written informed consent, they read the questionnaire instructions.
The questionnaire comprised 21 proposed items (a proposed item consisted of a proposed term and a proposed elaboration, e.g. “Sampling Rationale—Please describe your sampling rationale, typical strategies are maximum variation, purposive, theoretical, convenience, snowball, random or mixed,” see Supplemental Appendix 1). These proposed items were split between four parts labeled with the headings: Study information (six items), Design plan (four items), Data collection (six items), and Analysis plan (five items).
Per proposed item, participants answered four questions: (1) to what extent do you agree with the suggested term, (2) to what extent do you agree with the elaboration of the term, (3) to what extent do you agree that the topic is relevant for qualitative preregistration, and (4) under which heading should the topic be placed?
Questions 1 and 2 regarded comprehensibility and question 3 regarded the relevance of the proposed items. Finally, after having read all the proposed items that belonged to one of the major headings, we enquired whether there was any other topic that they would consider relevant (e.g., “Is there any topic you consider relevant for qualitative preregistration that you would like to add under ‘study information’?”) to assess comprehensiveness.
Questions 1–3 were answered using a 5-points Likert scale with response options “Totally agree,” “Somewhat agree,” “Neither agree nor disagree,” “Somewhat disagree,” and “Strongly disagree.” Questions had a text box underneath where participants were encouraged to give arguments for their ratings. For every question, panelists could indicate “no expertise” as an option. When panelists indicated “no expertise,” their vote did not count for the calculated percentage of agreement (this option was eventually used only twice).
The answer options for question 4 listed the four headings (Study information, Design Plan, Data collection, and Analysis plan) and a fifth option “Miscellaneous.” Again, participants could indicate “no expertise” if they felt unable to answer the question.
Finally, we enquired panelists’ professional background by asking about their disciplinary field, type of qualitative data they work with, the paradigm that guides their work, and the extent to which they saw the potential of qualitative preregistration, respectively. The full questionnaire for round 1 can be found https://osf.io/9vefh/
Round 2
The same 295 participants received the invitational email to participate in the second round in March 2020. Attached to this email they received a feedback report of round 1 with all ratings and arguments from this round provided. In round 2, we had compiled all responses and focused on issues left from round 1 where panelists did not agree on the best phrasings. We revised or removed items in accordance with participants’ ratings and arguments. In the questionnaire, we briefly summarized why and how particular arguments influenced revisions. Panelists were then invited to rate and comment on the suggested revised items. The questionnaire for round 2 took between 15 and 25 minutes to complete and can be found https://osf.io/8u25m/.
Finally, participants received a feedback report with all ratings and arguments from round 2. The feedback reported listed how many panelists chose a particular answer and included all pseudo-anonymized arguments. If panelists had indicated that they wanted to receive the final preregistration form, this was sent to them after final analyses.
Analyses
We preregistered our protocol on OSF (see https://osf.io/en3qc ). Results (i.e., ratings and arguments) were analyzed pseudo-anonymously, meaning that we removed all identifying information from the arguments that panelists provided. Rating data are open and can be found https://osf.io/2m9t7/files/.
We defined “agreement” as 68% or more of the participating panelists agreeing. This meant that if 68% or more rated “strongly agree” or “somewhat agree” on the 5-point Likert scale, we treated that as an item on which participating panelists converged. We chose 68% as an indication that a substantial group of panelists was on board with what is proposed, while retaining room for panelists with dissenting views. There is no agreement as to what is the “best” criterion to use, as these cut-offs are fundamentally arbitrary (Diamond et al., 2014). We reflect on the robustness of our agreement in the discussion.
We revised proposed terms and elaborations that the panel did not agree on based on the arguments provided. Items that had a relevance score below the cut-off were omitted from the next round, as it was our goal to make a concise preregistration form. TH read all arguments and then provided the steering committee with proposed suggestions based on these arguments that committee members commented on.
Pilot-Testing of the Surveys
For round 1, we asked researchers (
Results
Response Rate
From the 295 invitees in total, 48 participated in our Delphi (response rate: 16%). We invited all eligible panelists for both rounds. In round 1, 35 researchers completed the survey (and 4 partially). In round 2, 31 researchers completed the survey (and 4 partially). There were 25 panelists that participated in both rounds.
From the 48 panelists, 19 were researchers from our own network, 3 were authors of reporting guidelines, 23 were actively publishing qualitative researchers, and 3 were suggested by invitees. Demographic characteristics of the respondents appear in Table 1.
Demographics of the Panelists.a
a Panelists could indicate working in multiple disciplines, employing different types of research or using different paradigms, therefore the totals differ and exceed the number of panelists. Categories are displayed in alphabetical order.
Relevance of the Items
In round 1 panelists considered 14 of the 21 proposed items as relevant to include in the preregistration form. The panel indicated seven items were not relevant enough, of which two proposed items (related to the duration of the study and the researcher’s background) had relevance scores just below the cut-off value of 68% (65% and 66%, respectively) with mixed arguments. We revised these items in accordance with panelists arguments and they were put forth again in round 2 (see Supplemental Appendix 2), preceded by a summary of panelists’ arguments. At times, panelists suggested for different topics to be merged, as they were tapping into similar issues. We combined three items, namely “sort of sample,” “sampling strategy,” and “sampling rationale,” into a single revised item called “case selection strategy.” Two items, “origin of data” and “type of data” were revised into a single item called “data source and type.” This resulted in 11 items instead of 14 that were presented in round 2.
The remaining five items had relevance scores between 29% and 60%. Panelists argued that these proposed items were only relevant for a subset of qualitative research that explicitly investigates hypotheses such as process tracing (Collier, 2011). Therefore, we either removed these items or removed them as a stand-alone item and folded them into other elaborations where possible and in accordance with panelists’ views (e.g., “hypotheses” got folded into “research questions” to accommodate those who would want to preregister hypotheses).
In round 2, panelists converged on including the two terms that had a 65% or 66% relevance score in round 1. In their revised version, “researcher(s) positionality and potential biases” and “anticipated duration” obtained relevance scores of 77 and 87%, respectively. Participants argued that while it can be hard to indicate precisely how long a project will take, knowing the anticipated duration (and time when it was conducted) can be helpful. These were added to the form, so the final form consists of 13 items, see Figure 2.

Overview of items through the two rounds.
Terms & Elaborations
Panelists suggested alternative terms for the different topics addressed. They argued that our initially proposed terminology was either too quantitatively-oriented (e.g., “data validation”) or not in accordance with the qualitative canon (e.g., “sampling”). Panelists emphasized that our terminology should not clash with existing discipline-specific reporting guidelines for qualitative research and suggested small changes.
Panelists added clarifications or new examples to our suggested elaborations and at times pointed out that the elaborations were somewhat repetitive. In addition, the use of the word “typical” in the elaborations was criticized as it had normative connotations to it. Instead, panelists recommended to use a more refined phrase, such as “examples include, but are not restricted to…”.
In round 1, there were 11 proposed items (out of the 13 relevant items) where participants agreed with the suggested terminology. Despite this convergence of opinions, we carefully reviewed all arguments, substantially revising terms (
At the end of round 1, panelists also proposed new items. They consistently pointed out that items regarding ethical review, or ethics or ethical approval should be included in a preregistration, as well as funding. The Center for Open Science was working on integrating these items as meta-data fields into OSF registrations, so these were not asked in round 2, but panelists received a note explaining that these items will be added as meta-data fields to the form.
In round 2, panelists converged on 11 of our 13 revised terms, meaning they broadly agreed (agreement ranged from 68% to 100%). Panelists were dissatisfied with the term “termination criteria” (65%). We received many clear arguments to get rid of the “termination” part and to use words such as “stopping” or “stop” instead.
Panelists also argued that our revised term, “case selection strategy,” in round 2 was unclear (68% agreement) as the term conflated recruitment (where a researcher can use strategies like snowballing) and case selection (where the researcher can use selection criteria, such as typical case, or most deviant case (George & Bennett, 2005)). Furthermore, different panelists suggested to include “sampling” into the term, because “sampling” is commonly used to denote the process of selecting participants. In the previous round, panelists argued that participants are selected in variation to the phenomenon in question and not selected to representative in the statistical sense. The term now reads “Sampling & Case Selection” strategies, see Figure 3.

Visualization of the Delphi feedback process.
Finally, the revised elaborations yielded high agreement among the panel (most agreement ranged from 77% to 97%) in round 2. Yet, just 71% of the panelists agreed with the elaboration for the term “study design.” Panelists argued that the elaboration’s examples were insufficiently precise to get an indication of the level of detail with which one should respond to the item. We refined the examples and adjusted the elaboration after round 2, focusing on examples of designs that are more broadly used (e.g., “case study”). The final version of the preregistration form as it will be made available on OSF appears in Table 2.
Final Preregistration Form.
Discussion
We developed a form for the preregistration of qualitative studies in consultation with a panel of qualitative researchers. The form consists of 13 items spread over four headings and will be made available as a registration option on OSF to enable it to be broadly accessible to be used and piloted.
Comparison With Existing Literature
The preregistration form presented here differs from existing suggestions for preregistration forms (see here) for qualitative preregistration in two ways. First, it can be used for a broader range of qualitative approaches, as it has been designed with experts from different disciplines and traditions. Second, it uses the same overall structure of four headings as the existing OSF preregistration forms, although the heading labels are slightly different.
The proposed preregistration form overlaps with parts of reporting guidelines for qualitative research, such as the Consolidated Criteria for Reporting Qualitative Research (COREQ) in biomedical research (Tong et al., 2007) and the Journal Article Reporting Standards (JARS) from the American Psychological Association (Levitt et al., 2018). To illustrate, COREQ enquires whether there was an established relationship between the researcher and participants prior to the study (COREQ, p. 352, Table 1), and we pose a similar question when asking the user to reflect on their relation to or association with the participants or phenomenon they will study. Similar to JARS, we ask the user to provide an overview of the research design and a rationale for applying this particular design (JARS, p. 34, Table 1). These similarities will be useful as ideally study preregistration facilitates reporting of the final findings in accordance with reporting guidelines.
Limitations
Our response rate of 16% might seem too low to base any broad agreement-related conclusions on; however, Boulkedid and colleagues found that the median number of Delphi panelists was 17 (Boulkedid et al., 2011), whereas ours was nearly double. One reason for this low response rate is that we combined invitation and inquiry of interest, whereas some Delphi authors first poll their potential invitees for interest in participating and then only invite those who expressed interest, leading to a higher response rate.
We did not reach our predefined criterion (68%) for one proposed item. Some may argue that staying true to the Delphi approach would involve a third round. However, we only planned two rounds and it is known that response rates tend to decline after more than two rounds (Hasson et al., 2000). We felt confident that a third round was not necessary because panelists provided very clear arguments in round 2 for how the term needed to be revised that we implemented accordingly.
Strengths
We attained high agreement for the items in round 2, despite the fact that our panel was very diverse (i.e., they worked in a broad range of disciplines and employed different approaches in their research, and ground their work in widely differing epistemological perspectives and paradigms (Creswell, 2014)) and slightly skewed toward non-positivists. We regard the latter as an advantage, given that the discussion about preregistration tends to involve empiricists interested in generating production transparency, and thus critical insights from non-positivists may broaden the appeal of the preregistration form. That said, we acknowledge that the panel setup may also have led to the exclusion of some more positivist terms.
This is the first preregistration form that was developed in direct consultation with the potential user-community that can now be piloted. Other preregistration templates were developed either by the registries themselves or by small research teams that (informed by the literature) published their suggested templates, such as van’t Veer & Giner-Sorolla preregistration template for social psychology (van’t Veer & Giner-Sorolla, 2016) or Brandt and colleagues replication recipe preregistration template (Brandt et al., 2014). These templates were peer reviewed, so to some extent the community was involved in their development, but their views were not explicitly solicited in the template’s design.
Future Research
Note that a majority agreement-based form is not per se the best form (Jünger et al., 2017). In a similar vein, the qualitative study preregistration form proposed here may not be the best fit for a specific study. Future research could look into “modules” that could be added to this form and that would be relevant for their specific type of inquiry, but may not be relevant for qualitative inquiry across the board. For example, more deductively oriented approaches such as process tracing may be able to define specific criteria for how to “weigh” evidence before commencing data collection.
There is currently no empirical evidence to show that preregistering qualitative research will improve the quality of the work. A first step would be collecting good examples of qualitative study preregistration, two early examples of qualitative preregistration can be found here and here. Crucially, the developed form here will have to be piloted to test individual parts and modules for applicability, and to understand where the form needs to be expanded or reformulated.
Implications
It is pivotal to assure that the strength of qualitative research, which is its flexibility to adapt, adjust and respond, is not lost in preregistration (Tong et al., 2007). Analogous to quantitative preregistration, a preregistration is a plan, not a prison (DeHaven, 2017). The preregistration should provide a systematic starting point that can be updated as the study evolves. Although further investigation is needed, the use of this qualitative preregistration form can facilitate more transparent and credible research for various types of qualitative research.
Conclusion
We showed that despite the large variety of qualitative research methodology, it is possible to develop a template where the participating qualitative researchers converged on an opinion of what to include in a qualitative study preregistration template. This template can now be piloted, taking an important step concerning the place of qualitative research in the broader open science debate.
Supplemental Material
Supplementary_material - Preregistering Qualitative Research: A Delphi Study
Supplementary_material for Preregistering Qualitative Research: A Delphi Study by Tamarinde L. Haven, Timothy M. Errington, Kristian Skrede Gleditsch, Leonie van Grootel, Alan M. Jacobs, Florian G. Kern, Rafael Piñeiro, Fernando Rosenblatt and Lidwine B. Mokkink in International Journal of Qualitative Methods
Footnotes
Acknowledgments
We are deeply grateful for the contributions of the panelists, specifically Alessandra N. Bazzano, Anne Højager Nielsen, Ahtisham Younas, Birgith Pedersen, Stefan Bösner, Britt Marie Lindgren, Carolyn Tarrant, Crystal N. Steltenpohl, Damien Riggs, Domingo Palacios-Ceña, Gabriel Vommaro, Hillel David Soifer, Ingo Rohlfing, Ines Testoni, Jennifer Bussell, Jennifer Cyr, Juan Masullo, Kate Seers, Lynn Monrouxe, Marjan J. Westerman, María José Álvarez Rivadulla, Nicholas Weller, Peter Sainsbury, Maryam Rassouli, Suzanne Roggeveen, Siun Gallagher, Sebastian Karcher, Stina Lou, Timothy C. Guetterman, Virginia Braun, Edward B Davis and all other anonymous panel members. We would also like to acknowledge Brian Nosek (BN) in his role as helpful steering committee member.
Declaration of Conflicting Interests
The author(s) declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: T.M.E. is employed by the nonprofit Center for Open Science that has a mission to increase openness, integrity, and reproducibility of research and which maintains the open-source Open Science Framework (OSF).
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Tamarinde Haven was given an Early Career Fellowship (#0194) for guiding this project from CLUE+, the interfaculty Research Institute for Culture, Cognition, History and Heritage of the Vrije Universiteit Amsterdam (
). Fernando Rosenblatt was supported by Fondecyt #1190072 and by ANID - Millennium Science Initiative Program - Code ICN17_002.
Supplemental Material
Supplemental material for this article is available online.
Notes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
