Abstract
Fraudulent participation is a growing challenge in digital health research, particularly in online studies where duplicate identities, automated responses, and coordinated sign-ups can distort recruitment, compromise validity, and divert resources. Safeguards intended to prevent fraud might also risk excluding legitimate participants, raising concerns about sample representativeness and study generalizability. Although a wide range of technical and behavioral strategies exists, guidance is lacking on how to organize these methods and report outcomes consistently across studies. To address this gap, we introduce the Configure, Assess, Triage, Corroborate, and Hone (CATCH) framework, a hybrid fraud detection–mitigation model with actionable recommendations for investigators. CATCH begins with pre-study
Fraudulent participation in online research
Remote online studies and decentralized clinical trials are growing increasingly popular in human subjects research. Fully online approaches help address difficulties in meeting proposed samples in large-scale studies, recruiting nationally representative participant pools, and identifying individuals with rare or stigmatized conditions. 1 Social distancing restrictions due to the coronavirus disease 2019 (COVID-19) pandemic further propelled the growth of remote research procedures, 2 and meta-analytic work confirms that online recruitment is significantly more efficient and less expensive than in-person approaches. 3 Alongside these benefits, online methods have introduced new challenges: bots, duplicate participants, and fraudulent reporting have emerged as serious threats to research integrity, 4 affecting not only recruitment but also downstream study procedures, where verifying participant identity or characteristics is more difficult when all interactions occur online.
Reviews of online studies using fraud detection suggest this phenomenon is common across a range of content areas, 5 including qualitative and quantitative research paradigms. When undetected, impostors can distort findings and misdirect practice or policy. 6 Even when impostors are detected, the costs in lost funds, staff time, and slowed recruitment are substantial.7,8 For example, one study of patient perspectives of their cancer care received 256 fraudulent study sign-ups (of 271 total; 94.5%) in just 7 h of recruitment. 9 In another study, a team studying performance feedback for medical staff received 268 study incentive requests from just three email addresses. 10 While financial incentives are often assumed to be a primary driver of fraudulent participation, studies show that deception can occur even without monetary rewards, suggesting additional motives such as boredom, curiosity, or ideological intent to disrupt research. 11
Fraudulent participation can distort recruitment sampling, misrepresent demographic distributions, and obscure relationships between groups of participants and outcomes. For example, Sharma et al. 12 in Australia reported a sample that was entirely male and 80% aboriginal, despite qualitative research in their field typically recruiting predominantly female participants. Such misrepresentation can hinder the recruitment of a representative population sample, as the presence of fraudulent participants may make groups falsely appear to be over- or under-represented. It may also lead to the exclusion of legitimate participants who share traits with fraudulent participants.
Existing methods for fraud mitigation
An emergent literature has identified procedures for detecting and deterring fraud. Early reports often took the form of case studies describing encounters with suspected participants and the ad hoc strategies used to respond.13–15 These were followed by articles deriving recommendations for future work, but were grounded mainly in small, qualitative studies.16–18 Across studies, researchers have experimented with a wide range of fraud-mitigation strategies.
4
Technical approaches include bot filtering using completely automated public Turing tests to tell computers and humans apart (CAPTCHA), internet protocol (IP) address monitoring to flag suspicious geographies, and checks for duplicate contact details or unusually fast survey completions.4,19,20 Behavioral approaches have included open-ended screening questions, consistency checks across survey or interview responses, or requiring participants to briefly enable their camera for verification.4,21 Fraud-mitigation strategies fall into two broad categories:
Safeguards intended to prevent fraud may introduce new barriers for legitimate participants, particularly those with limited resources. Recruitment through interest groups, such as online communities or associations, may reduce fraudulent entries, but risks narrowing the participant pool to those with access, interest, and the capacity to engage with such groups. 12 Adjusting participant compensation is another lever that has been used to attempt to safeguard research integrity. Although higher incentives have been linked to increased fraudulent responses, lowering them may deter genuine participants, making the sample less representative.5,23 Phone verification can filter participants but excludes those without consistent access, while video chat requirements may disadvantage individuals without cameras or private spaces. 12 Government ID checks or flags on unusual login times may disproportionately exclude underserved participants. 23 Some participants, especially in studies involving sensitive topics such as mental health, drug use, or reproductive health, may hesitate to provide accurate contact information due to privacy concerns rather than fraudulent intent. 24 There is a persistent tension between safeguarding and inclusivity; if safeguards are too lenient, ineligible participants contaminate the data and compromise validity 6 ; if safeguards are too strict, legitimate participants risk being wrongly excluded, limiting representation and generalizability. 25 No single mitigation method is sufficient on its own, highlighting the need to organize and layer different strategies in complementary ways. 8
Earlier work has provided useful examples of techniques for fraud mitigation, but what remains absent is a structured way to organize these methods, define their reporting points, and ensure comparability across studies. No general guidance exists on how investigators can structure, document, or justify fraud-mitigation practices to maximize scalability while maintaining rigor. To advance cumulative knowledge, researchers also report how many participants were screened out at each stage by each method, how many required manual corroborations, and how many were ultimately excluded or “redeemed.” Detailed documentation can help illuminate the true magnitude of fraud, clarify the practical capacity for handling inconclusive cases, and help establish benchmarks for estimating false exclusion rates.
The Configure, Assess, Triage, Corroborate, and Hone (CATCH) framework
Here we introduce the CATCH framework, a layered fraud detection–mitigation model with explicit, actionable recommendations designed to guide investigators conducting online studies. The hybrid CATCH model integrates systematic and person-led methods on an as-needed basis. Building on the existing literature and our team's experience conducting remote health studies, CATCH organizes prior recommendations into a unified, staged structure that clarifies when, how, and why each method should be used.
CATCH begins with a pre-study
Once live recruitment begins, systematic
At every stage, investigators are encouraged to document not only the actions taken (screened, flagged, excluded, and retained) but also the numbers associated with these outcomes. This dual emphasis on process and reporting highlights both how fraud can be mitigated and how comparability and transparency can be strengthened across studies. Underlying this approach is the familiar tradeoff between sensitivity and specificity: greater sensitivity to fraud increases the risk of excluding legitimate participants, while greater specificity reduces false exclusions but may allow more fraudulent participants through. Each study will need to determine how to strike this balance in line with its aims, resources, and population. The CATCH framework is illustrated in Figure 1.

The Configure, Assess, Triage, Corroborate, and Hone (CATCH) framework. The funnel begins with pre-study configuration, then proceeds through systematic risk assessment, candidate triage, and manual corroboration, with ongoing monitoring to hone strategies as new threats emerge. Investigator roles corresponding to each stage are shown on the right.
Research teams are encouraged to select and apply the CATCH approaches that are most suitable for their circumstances, timeline, study population, resources, and budget. The framework is designed to serve as a flexible approach rather than a model that is excessively prescriptive or proscriptive. To support implementation, we include a supplemental checklist that investigators can adapt to their study. The checklist, available as Supplemental Table 1, provides a structured template aligned with each CATCH stage and outlines suggested decisions, documentation points, and reporting elements to promote transparency and consistency across studies.
CATCH and Emerging Technologies
The development of mitigation frameworks such as CATCH is critical as new technologies create opportunities for increasingly sophisticated forms of fraudulent participation. Emerging tools such as artificial intelligence (AI) make fraudulent participation more scalable and harder to detect. 30 Large language models can be leveraged to generate human-like text that mimics genuine responses, enabling fraudulent actors to craft detailed answers that bypass quality control measures such as free-text justifications.31–35 Combined with text obfuscation platforms, AI-generated responses can evade automated detectors entirely, with false negative rates reaching 100% in evaluations of AI-text classifiers. 36 Beyond text, AI-generated images, video, and audio can be used to create realistic participant profiles that pass identity verification checks, complicating detection efforts across modalities.36–39 These challenges highlight the need for frameworks such as CATCH to remain adaptable. Fraud mitigation cannot be static throughout a study or across studies; configuring, assessing risk, triaging, and corroborating cases should all undergo iterative refinement, with strategies honed through ongoing monitoring as new threats emerge.
Emerging technologies also create opportunities for more sophisticated fraud detection that can be integrated into CATCH once validated. Some techniques already appear in online survey practice, such as timezone alignment checks, 40 edit-distance matching for duplicate emails, 40 and contradiction checks across survey responses. 41 In online health research, these tools can support geographic verification and help identify repeated or ineligible enrollment. Other approaches represent emerging possibilities that require further evaluation in health-research settings. Chatbots can support scalable prescreening, drawing from methods in phishing and financial fraud detection.42,43 For participants requiring ongoing monitoring, AI-driven approaches may help flag longitudinal signals of fraudulent participation, such as patterned response behavior or inconsistencies over time. 6 Ultimately, fraudulent participants themselves may be deliberately recruited as intended targets for study by researchers seeking to learn more about their methods in an effort to safeguard data collection.
Conclusion
Fraud in online research is a perpetually evolving challenge. As detection systems advance, fraudulent strategies will adapt in response. Adjustments to mitigation measures can introduce new risks of false exclusion, underscoring the need to continually reassess how safeguards are applied. The task for researchers, then, is not to aim for permanent solutions but to continually update detection and mitigation measures while reporting them transparently so that the field can learn, compare, and refine responses over time. The CATCH framework constitutes a practical starting point for this process. Through the synthesis of existing fraud-mitigation strategies into a unified, staged framework, CATCH provides investigators with practical guidance for structuring decisions, documenting actions, and balancing data integrity with inclusivity. By implementing its staged structure and documenting outcomes transparently, researchers can build a shared knowledge base that strengthens study integrity and enhances resilience against future forms of fraud.
The proposed framework aims to enhance methodological rigor and comparability across online studies by clarifying when and how different mitigation strategies can be applied. Future work should examine how the CATCH framework performs across different study designs, populations, and research contexts. Empirical applications may evaluate its impact on recruitment efficiency, false exclusion rates, and equity-related outcomes. Additional research may also explore ethical tradeoffs between fraud detection and participant privacy, particularly in digitally mediated and decentralized studies. Such efforts will be essential for refining the framework and informing best practices for online research integrity.
Supplemental Material
sj-docx-1-dhj-10.1177_20552076261418807 - Supplemental material for Practical guidance for mitigating fraud in online research: The Configure, Assess, Triage, Corroborate, and Hone (CATCH) framework
Supplemental material, sj-docx-1-dhj-10.1177_20552076261418807 for Practical guidance for mitigating fraud in online research: The Configure, Assess, Triage, Corroborate, and Hone (CATCH) framework by Maya Stemmer, Justin Tauscher, Benjamin Buck, Patrick Wedgeworth, Oliver John Bear Don’t Walk, Trevor Cohen and Dror Ben-Zeev in DIGITAL HEALTH
Footnotes
Acknowledgements
Not applicable.
Ethical considerations
Not applicable because this article does not contain any studies with human or animal subjects.
Consent to participate
Not applicable because this article does not contain any studies with human subjects.
Consent for publication
Not applicable because this article does not contain any studies with human subjects.
Author contributions
MS conceived and outlined the manuscript and integrated co-author contributions. MS, JT, BB, PW, and OJBDW reviewed the literature, contributed conceptual input, and drafted sections of the manuscript. DBZ and TT provided senior guidance on framing and organization. All authors reviewed, edited, and approved the final version of the manuscript.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: The authors are supported by a grant from the National Institute of Mental Health (U01MH135901). Dr Buck is supported by a Mentored Patient-Oriented Career Development Award from the National Institute of Mental Health (K23MH122504). The views expressed in this manuscript do not necessarily represent the views of the National Institute of Mental Health, nor did the sponsor play any role in the conception or drafting of this manuscript.
Declaration of conflicting interest
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: Dr. Ben-Zeev has financial interests in Merlin and FOCUS technology. He has provided consultation services to K Health, Boehringer Ingelheim, Deep Valley Labs, Butler Hospital, and Otsuka Pharmaceuticals.
Data availability
Not applicable.
Supplemental material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
