Abstract
Generative artificial intelligence (GenAI) has introduced a transformative approach in medical informatics and education. AI-driven video models, such as Sora, HeyGen, Synthesia, and Google Veo 3, among others, can autonomously generate realistic clinical materials, including synthetic patients and simulated scenarios. This technology system represents an emerging domain of medical learning informatics that integrates AI-generated content, simulation, and pedagogy. This scoping review, based on selected studies, identifies and synthesizes educational outcomes, highlights the methodological limitations of AI-generated videos as training tools for medical students, and explores technical and pedagogical challenges to guide future research. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews framework, the literature searches were conducted across Google Scholar, PubMed, ScienceDirect, Scopus, and gray literature sources. Studies were included if they focused on AI-generated videos as educational tools for medical education. A single reviewer conducted screening of titles, abstracts, and full texts, and data were systematically extracted using a standardized charting form, including study design, AI tool utilized, outcomes, limitations, and challenges. Of the 970 retrieved records, 8 studies met the inclusion criteria. The latter demonstrated that AI-generated videos can enhance knowledge retention, skill acquisition, and learner engagement, outperforming traditional methods of delivering practical exercises in medical education. Reported challenges included issues with accuracy, limited emotional authenticity, ethical standards, and the necessity for pedagogical consistency. AI-driven videos are a promising innovation in medical education, offering scalable, interactive, and personalized learning. However, their integration requires a solid validation framework, interdisciplinary collaboration, and governance models that guarantee ethical and pedagogically appropriate use. Additionally, long-term, cross-institutional studies are necessary to evaluate the lasting educational and clinical effects.
Introduction
Background
Advances in generative artificial intelligence (GenAI) have reshaped the boundaries of digital learning and medical simulation. Among these innovations, artificial intelligence generated videos have emerged as a distinct technological and educational framework that warrants separate consideration. Also called AI-driven videos, 1 they are synthetic audiovisual learning materials automatically produced by GenAI models. They are typically generated through a process that uses a text or scientific term prompt (simple or multimodal). 2 AI-generated videos differ conceptually from virtual patients, which are interactive, software-based clinical scenarios designed for case-based training, 3 and from traditional simulation modalities, such as real people, mannequins, or instructor-led simulation environments. 4 They also differ from other AI-driven education tools, such as adaptive quizzes 5 and intelligent tutoring systems that are similar to virtual patients. 6 While these tools emphasize real-time feedback and decision-making, AI-driven videos offer pre-made, repeatable audiovisual scenarios. This enables students to observe, reflect, and practice skills independently, serving as a complementary method within interactive learning environments.
In medical informatics, AI-driven videos offer new possibilities for creating engaging, adaptive instructional materials that surpass traditional methods through immersive simulations. Video generation platforms such as Sora by OpenAI, HeyGen, Synthesia, and other multimodal systems can generate synthetic patient interactions, clinical cases, and procedural demonstrations. Creating videos with the aforementioned tools involves using prompts that simulate various clinical narratives, 7 which are introduced in Sora or other AI video generators to produce realistic videos, 8 featuring synthetic patients. This allows medical students to engage in interactive exercises to practice patient conversations in a digital environment. 9 These text-to-video innovations 10 integrate AI-based data synthesis with educational informatics to enable the automated creation, storage, and dissemination of training materials (Figure 1).

Pathway for generating videos using an artificial intelligence driven solution.
From an informatics standpoint, AI-driven videos function as an intelligent media system that facilitates knowledge management, learner modeling, and simulation-based learning, shaping how students use tools 11 to improve their learning experiences. 12 These videos enable consistent replication of clinical encounters while minimizing reliance on human actors or expensive simulation setups. Through these scenarios, students interact with digital humans that convey realistic movements and expressions,13,14 improving understanding, retention, 15 self-confidence, and expertise. 12 This innovation is particularly relevant for medical education programs that are increasingly incorporating digital learning environments such as PCS, 16 Geeky Medics, 17 Healthy Simulation, 18 Confluent, 19 and adaptive instructional technologies.
Notably, these technologies are recognized for providing personalized, 20 self-directed, 21 secure, and effective methods for students to enhance their skills,22,23 while simultaneously delivering a realistic learning experience 24 that can strengthen confidence and foster self-efficacy.25,26 Based on medical education principles, these AI-driven platforms can assume additional roles, such as (1) virtual simulation laboratories that effectively connect traditional classroom learning with authentic clinical experiences 26 and (2) realistic simulation scenarios for practitioners to rehearse intricate procedures and develop new skills without interrupting their clinical responsibilities. 27 Both roles can help reduce medical errors and improve patient outcomes. 28
However, the use of synthetic video content raises important issues. AI-generated videos may present downsides, such as inconsistent speech, distracting gestures, and emotional biases. 15 In addition to the earlier issues, students report concerns regarding the depth of interpersonal connections, the authenticity of interactions, inaccuracies in content,29,30 limited facial expressions, gestures, and monotone vocal expressions, 31 the failure to compare with traditional methods, and varying levels of users’ technological skills. 32
To the author's knowledge, the number of primary studies on AI video generators for medical education remains limited; however, the affordability and availability of these technologies are increasing, owing to the emergence of numerous AI-driven platforms for medical training. Recent studies highlight several key reasons why AI isn’t widely used in medicine, despite its novelty, global accessibility, and low cost. 33 They highlight the existing gap between developers and end users, 34 the limited focus on AI-generated videos as teaching tools for medical students, 35 and their absence from educational curricula. 36 Together, these factors pose a challenge to the effective integration of AI-driven videos into the medical education system.
Objective
Based on the selected studies, this scoping review identifies and synthesizes educational outcomes, discusses methodological limitations associated with the use of AI-driven videos as a training tool for medical students, and explores technical and pedagogical challenges that could guide future research.
Methods
Study Design
This study was meticulously designed in accordance with the methodological framework proposed by Arksey and O’Malley, 37 and further refined by Leva et al. 38 The reporting of this study conforms to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) guidelines. 39 The completed PRISMA-ScR checklist is provided as Supplemental File 1. A formal review protocol was not registered prior to conducting this study; however, predefined eligibility criteria, search strategies, and data charting procedures were established to promote transparency and reproducibility.
Protocol and Reporting Standard
The formal protocol was not registered prior to conducting the review. However, the review adhered to established scoping review methods, with eligibility criteria and analytic procedures set beforehand and applied consistently. Reporting complied with PRISMA-ScR guidelines to ensure transparency and reproducibility.
Eligibility Criteria
This research mainly investigated the use of AI-generated videos as educational tools for medical students. While it primarily focused on these students, it also included a study involving undergraduate nursing students because it examined AI-generated videos used in similar clinical education and simulation-based learning settings. Since health professions education shares pedagogical principles and overlapping skills, this inclusion aimed to provide a broader understanding of how AI-driven video technologies are used in clinical training. The findings regarding nursing students were carefully interpreted and not generalized beyond comparable educational contexts. Studies were included if they investigated the use of these tools in undergraduate, graduate, or continuing medical education; reported learning outcomes, limitations, and implementation challenges related to medical training; and were published in English in peer-reviewed journals or reputable gray literature sources.
The excluded articles were those that either did not discuss AI-generated video, mentioned it but did not target medical students, or referenced AI models without providing any medical context for students or incorporating AI-generated video. The characteristics of the studies, including the target participants, study objectives, AI tools used, and methodologies employed, were considered for inclusion. No restrictions were placed on publication year or geographical location to ensure comprehensive coverage of emerging research.
Information Source and Search Strategy
A systematic search was performed across 4 major databases: Google Scholar, PubMed, ScienceDirect, and Scopus. In addition to bibliographic databases, gray literature sources such as conference proceedings, preprint servers (eg, medRxiv and arXiv), and institutional reports were systematically searched to reduce publication bias. The identified records were screened using the same eligibility criteria as for peer-reviewed studies. Titles and abstracts were examined for relevance, followed by full-text review when sufficient methodological and educational details were available. Documents without clear authorship, transparent methodology, or educational relevance were excluded.
The study was conducted between February and December 2025, using a combination of vocabulary terms (such as MeSH) and free-text keywords related to AI, medical education, and AI-driven videos. The representative search string was “generative AI” OR “artificial intelligence” OR “text-to-video” OR “AI-driven video,” AND “medical education” OR “medical students” OR “synthetic patients” OR “simulation training.” Additionally, references within the selected studies were manually checked to identify further relevant sources.
Selection of Source Evidence
All records retrieved from databases were imported into EndNote (version 20, Clarivate Analytics) for reference management and duplicate removal. The author (PFI) screened the titles and abstracts of all records against predefined criteria. The full texts were screened thoroughly and simultaneously, with any uncertainties or conflicts addressed afterward. The screening process was documented and displayed in a PRISMA flow diagram (Figure 2).

Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) flow diagram of study selection.
Due to resource and feasibility constraints, a single reviewer conducted the screening and study selection. This approach is acceptable in a scoping review, where the aim is to map available evidence rather than generate quantitatively pooled estimates. To reduce potential bias, predefined eligibility criteria were strictly applied, and uncertainties were reviewed multiple times during the screening process.
Data Charting and Extraction Process
Data extraction was conducted utilizing a standardized charting form specifically developed for this review. The data variables included: author(s) and publication year; study population and educational setting; type of AI-generated video tool used; study design and methodology; reported educational outcomes; and challenges and limitations as explicitly reported by the included studies. Data charting was also performed by a single reviewer based on feasibility, and the chart was checked for completeness and consistency before synthesis.
Data Synthesis and Analysis
All extracted data were organized narratively and thematically using an inductive approach. Studies were categorized based on their primary focus: (1) learning outcomes and (2) challenges and limitations issues. Descriptive analysis was used to determine the frequency of study characteristics, providing an overview of the distribution of evidence. Additionally, findings were examined to identify common themes in technological design, usability, interoperability, data ethics, and pedagogical integration. This dual perspective (educational and informatic) enables a comprehensive understanding of how AI-driven videos operate within digital learning environments in medical education.
Results
Selection of Sources of Evidence
The initial search identified 946 studies on AI-generated video in medical education, sourced from various databases, including 24 from gray literature. After removing duplicates (n = 150), 608 records were excluded based on title and abstract, leaving 212 records for eligibility assessment. Of these, 204 were excluded for specific reasons: 115 were not AI-generated video-based, 68 were not intended for medical students, and 21 were irrelevant because they referenced AI models without AI-generated video or any medical context. The remaining 8 records met the final inclusion criteria. The selection process is detailed in Figure 2, which follows the PRISMA flow diagram for scoping review.
Characteristics of Sources of Evidence
The included studies (n = 8) were published between 2022 and 2025 and researched the use of AI-driven videos in various medical education contexts. The vast majority of studies adopted quantitative methodologies31,40,41–45 employing either pre–post intervention or randomized control designs, while 1 study 9 used a qualitative design focused on experiential feedback. Regarding objectives, the included studies examined learning effectiveness using pretest and posttest40,42; assessed how tailored content captures focus 44 ; evaluated the realism of interactive communication training9,41; and compared AI-generated interactive content with expert-led instruction, 43 with human-made video animation, 45 and explored perspectives on content creation and usability. 31
The target populations included medical students,31,41,42,43,45 nursing students, 44 and trainees9,40 at various training stages: from preclinical (first and second years) to postgraduate levels. Several studies have reported the use of AI-driven video tools, including Visla, HeyGen, Synthesia, CrazyTalk 8, Neuro VR, Google Veo 3, and multimodal systems such as Midjourney (Table 1). These tools facilitated the creation of synthetic patient simulations, interactive interviews, procedural demonstrations, and customized videos. Overall, the included studies demonstrate a growing diversity in how AI-generated video systems are designed and evaluated in medical training settings.
Characteristics of the Studies Included in the Scoping Review.
Abbreviations: AI, artificial intelligence; GenAI, generative AI; VOA, virtual operative assistant.
Results of Individual Sources of Evidence and Synthesis
Educational Outcomes
As shown in Table 2, the 8 studies included reported favorable educational outcomes from the use of AI-generated videos. These tools were found to be comparable to human-created videos in instructional effectiveness,40,41,43 and were appreciated for their scalability, affordability, and learner engagement.9,31,42 Specifically, AI-driven videos are more cost-effective than human-centered materials, 40 can simulate complex clinical scenarios, 9 enhance comprehension and engagement, 34 and build medical skills in an unprecedented manner. 43 Overall, AI-driven video tools support traditional teaching methods, offering more opportunities for interactive, self-directed, and adaptive medical education.
Educational Outcomes Identified Across Studies.
Identified Challenges and Limitations Across Studies
All studies, despite their findings, highlighted significant research gaps translated into technical and pedagogical challenges, as well as methodological limitations. Key themes of technical challenges included the necessity to (1) improve realism and emotional expressiveness in AI avatars 9 ; (2) ensure AI-generated videos uphold high teaching and learning standards 31 ; (3) ensure that the created tool align with the latest clinical guideline, evidence and best practices 40 ; (4) ensure accurate understanding and response to human speech and language utilizing AI 41 ; (5) examine design factors such as video length, AI voice quality, and avatar fidelity 42 ; (6) explore hybrid methods that combine expert feedback with AI-generated content 43 ; (7) ensure that AI-driven videos are accurate and clinically reliable, aligning with international patient safety standards 44 ; and (8) develop evaluation criteria to compare AI-generated and human-created videos. 45 Furthermore, Table 3 outlines the methodological limitations and pedagogical challenges to inform future studies.
Pedagogical Challenges and Methodological Limitations to Inform Future Studies on Artificial Intelligence Driven Videos in Medical Education.
Moving forward, it will be vital to prioritize comprehensive validation, ethical considerations, and stakeholder collaboration to ensure the responsible implementation of this technology. Addressing these issues through careful instructional design, ethical review, and extensive long-term research is essential to guarantee that AI works as a valuable supplement rather than a replacement for human expertise in medical education.
Discussion
Summary of Evidence
This study identified 8 primary studies that address the use of AI-generated videos as learning tools in medical education. Collectively, these studies suggest that generative video technology may enhance learner engagement, conceptual understanding, and skill development, although current evidence remains preliminary.15,46,47 Notably, the integration of such tools reflects a broader paradigm shift in medical informatics, in which AI is increasingly used to manage, personalize, and disseminate instructional content.
The reviewed evidence suggests that AI-generated video modules may complement traditional teaching by providing standardized and repeatable simulations, which are reported as potentially effective,31,40,43 valuable,9,44 comparable, 41 and acceptable,42,45 in early studies. This can help boost learners’ autonomy and enable self-paced study.48–50 As McCloy foresaw, they may now significantly impact diagnostics and surgical solutions. 51 Additionally, from an informatics standpoint, AI-driven videos are components of intelligent educational ecosystems that combine content-generation algorithms with user data analytics in training environments, aligning with key principles of digital learning in informatics, namely efficiency, 52 accessibility, 53 and personalization.54–56
Nevertheless, the finding underscores the persistent challenges in validating, governing, and ensuring ethical interoperability of such systems. Several studies have raised concerns about the realism of emotional expressions, the accuracy of medical information, and the transparency of underlying AI algorithms. These issues mirror broader concerns in medical informatics regarding algorithmic bias, data governance, and explainability. Moreover, the absence of a unified pedagogical framework for integrating AI-generated content hinders the systematic adoption of this approach in medical education.
To bridge these gaps, medical educators and informatics experts should collaborate to codesign AI video platforms that align with learning goals, accreditation requirements, and ethical principles. In particular, informatics should prioritize developing evaluation metrics for AI-driven videos, including performance analytics, cognitive load measures, and user engagement indicators. Incorporating AI-driven video systems into institutional learning management systems and electronic health record simulators could also improve authenticity and data interoperability.
Additionally, this involves improving the stakeholders’ literacy in AI tools,41,43 simulating authentic clinical histories or lived experiences,9,44 adhering to AI innovations that may improve quality and effectiveness, 31 enhancing video features for better usability, 42 and increasing its application to boost acceptance.42,45 Moreover, reviews highlight the importance of clinical educators acquiring prior knowledge of AI innovations before integrating them into training programs. 57 This includes embedding AI into both preclinical and clinical curricula,58,59 investigating novel areas, and addressing potential risks associated with AI usage in medical education. 60 These risks include inaccuracies, overreliance, integrity issues, authenticity concerns, bias, privacy breaches, security vulnerabilities, 61 and scalability issues arising from the misuse of AI-driven tools.
Limitations
The paucity of primary studies on AI-generated videos in medical education limits the scope of this review. The included studies differed primarily in methodology, sample size, and educational context, limiting the generalizability of the findings. Rapid advances in AI technologies may render some tools or findings obsolete, necessitating ongoing updates to the literature. The search was limited to English-language sources, potentially excluding relevant studies published in other languages. Furthermore, the review does not consider long-term outcomes, such as the sustained impact of AI-generated videos on clinical performance and patient outcomes. Differences in technological literacy among medical students and educators add another layer of complexity that the studies did not thoroughly address.
A key methodological limitation is that screening and data extraction were conducted by a single reviewer due to resource constraints. This may increase the risk of selection bias, omission of relevant studies, and data extraction errors, potentially affecting the review's reliability and reproducibility. Additionally, including gray literature and preprints might lower the overall certainty of the evidence, as these sources often lack formal peer review and frequently present preliminary findings that may change. However, their inclusion was deemed important to capture emerging evidence in this rapidly developing field and to reduce publication bias. Lastly, although a protocol is not mandatory for scoping reviews, the absence of a registered protocol may decrease transparency and make reproducibility more difficult.
Conclusions
This review indicates that AI-generated videos show promise as a supportive tool in medical education, particularly for enhancing simulation-based and self-directed learning. Preliminary evidence suggests that these technologies may contribute to improved knowledge retention, skill development, and student engagement. However, the current evidence base remains limited, with a small number of heterogeneous studies employing diverse methodologies and outcome measures. As a result, findings should be interpreted cautiously. Important challenges remain, including verifying the authenticity of simulated experiences, aligning technological capabilities with educational needs, and addressing ethical issues such as data privacy, bias, and governance. Achieving the potential of an AI-generated video tool will require collaboration across disciplines among educators, technologists, and ethicists.
Future research should prioritize large-scale, methodologically rigorous, and longitudinal studies to validate educational effectiveness, develop a standardized evaluation framework, and clarify best practices for responsible integration into medical curricula.
Supplemental Material
sj-docx-1-mde-10.1177_23821205261437352 - Supplemental material for Artificial Intelligence Generated Videos as Supportive Tools in Medical Education: A Scoping Review
Supplemental material, sj-docx-1-mde-10.1177_23821205261437352 for Artificial Intelligence Generated Videos as Supportive Tools in Medical Education: A Scoping Review by Pinto Francisco Impito in Journal of Medical Education and Curricular Development
Supplemental Material
sj-docx-2-mde-10.1177_23821205261437352 - Supplemental material for Artificial Intelligence Generated Videos as Supportive Tools in Medical Education: A Scoping Review
Supplemental material, sj-docx-2-mde-10.1177_23821205261437352 for Artificial Intelligence Generated Videos as Supportive Tools in Medical Education: A Scoping Review by Pinto Francisco Impito in Journal of Medical Education and Curricular Development
Footnotes
Acknowledgements
The author extends thanks to Professor José Azevedo and Vasco Cumbe for their valuable guidance on his doctoral thesis, whose insights motivated him to conduct this research, representing a year of dedicated effort.
Author Contributions
The author conceived the study, designed the methodology, conducted the literature search and screening, performed data charting and synthesis, and drafted and revised the manuscript.
Funding
The author disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research received institutional support through b-on, funded by Fundação para a Ciência e a Tecnologia (FCT), Portugal, via Porto University, Faculty of Engineering. The funding had no involvement in the study design, data collection, analysis, interpretation, or manuscript preparation.
Declaration of Conflicting Interests
The author declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Supplemental Material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
