Abstract
OBJECTIVE
To understand fellowship program directors’ (FPDs) perspectives on facilitators and barriers to using entrustable professional activities (EPAs) in pediatric subspecialty training.
METHODS
We performed a qualitative study of FPDs, balancing subspecialty, program size, geographic region and current uses of EPAs. A study coordinator conducted 1-on-1 interviews using a semistructured approach to explore EPA use or nonuse and factors supporting or preventing their use. Investigators independently coded transcribed interviews using an inductive approach and the constant comparative method. Group discussion informed code structure development and refinement. Iterative data collection and analysis continued until theoretical sufficiency was achieved, yielding a thematic analysis.
RESULTS
Twenty-eight FPDs representing 11 pediatric subspecialties were interviewed, of whom 16 (57%) reported current EPA use. Five major themes emerged: (1) facilitators including the intuitive nature and simple wording of EPAs; (2) barriers such as workload burden and lack of a regulatory requirement; (2) variable knowledge and training surrounding EPAs, leading to differing levels of understanding; (3) limited current use of EPAs, even among self-reported users; and (4) complementary nature of EPAs and milestones. FPDs acknowledged the differing strengths of both EPAs and milestones but sought additional knowledge about the value added by EPAs for assessing trainees, including the impact on outcomes.
CONCLUSIONS
Identified themes can inform effective and meaningful EPA implementation strategies: Supporting and educating FPDs, ongoing assessment of the value of EPAs in training, and practical integration with current workflow. Generating additional data and engaging stakeholders is critical for successful implementation for the pediatric subspecialties.
Introduction
As medical education has shifted toward an outcomes-based framework of competency-based medical education (CBME), considerable attention has been focused on developing and implementing comprehensive assessment methods and tools.1-4 Within the United States Graduate Medical Education (GME) system, the current approach to assessment is milestone-based, with mandatory reporting to the Accreditation Council for Graduate Medical Education (ACGME).5,6 Milestones are skills, knowledge, and behaviors that represent performance levels along a developmental continuum. 7 However, competencies and milestones require clinical context and are not designed to be used for assessment in isolation. One way to provide a clinical context for competencies and milestones is through Entrustable Professional Activities (EPAs).8-10 EPAs, which have been developed and studied in multiple countries and for numerous specialties, 11 represent observable and measurable activities essential to the profession. 12 These activities require the successful integration of behaviors and competencies by an individual learner. Learners’ ability to successfully execute an EPA is assessed using a scale delineating required levels of supervision, with the ultimate goal of achieving unsupervised practice. 12 Many faculty find using EPAs more intuitive and meaningful than context-independent milestone-based assessments, which ultimately leads to more effective learner assessments. 13
Almost a decade ago, education leaders from the pediatric subspecialty community, in collaboration with the American Board of Pediatrics (ABP), developed a set of common EPAs for all pediatric subspecialties and a set of EPAs for each specific subspecialty. 14 The common EPAs describe activities of all pediatric subspecialists whereas subspecialty-specific EPAs describe activities essential to that particular discipline. The ABP plans to incorporate these EPAs into their initial certification process by 2028. 15 Despite validity evidence for the use of EPAs for assessment and their perceived value for assessment purposes, EPAs are not yet uniformly used across pediatric fellowship programs in the United States.16,17 A number of studies have reported on the implementation of EPAs into curricular designs in residency programs, but none have focused on implementation of EPAs for pediatric subspecialty training.18,19 Effective implementation of new or different practices within education, like EPAs, requires a thoughtful and collaborative approach with relevant stakeholders and an understanding of the facilitators and barriers to their implementation.20-22 One of the key stakeholders involved in the implementation of EPAs is the fellowship program director (FPD). FPDs are the frontline leaders tasked with local implementation of the EPAs. Thus, their perspectives may be valuable to the pediatric community.
Few studies have been published about the implementation of EPAs in the United States or abroad.18,23-25 While early adopters may champion the use of EPAs, a deeper understanding of the perceived role of EPAs within pediatric subspecialties is important as the field of pediatrics continues the transition toward broader implementation of EPA-based assessment. Despite described guidelines for successful implementation by medical education experts, 26 it is imperative to explore and evaluate how EPA implementation has been adopted and adapted in a variety of settings. 24 To advance understanding of EPA implementation within pediatric subspecialty programs, we performed a qualitative study with FPDs across the United States. The primary aims were to understand the facilitators and barriers to implementing the EPAs to assess pediatric subspecialty fellows, how EPAs were currently being used, and FPD perspectives on the value added by EPAs for fellow assessment.
Methods
Study design, strategy and data collection
We performed a qualitative study of subspecialty FPDs, using an inductive approach with a constant comparative method to yield thematic analysis. 27 We used the Consolidated Framework for Implementation Research (CFIR) as the overarching framework for the study and interview guide as we sought to understand the implementation process for EPAs. 28 We explored each of the 5 domains (intervention characteristics, characteristics of individuals involved, the outer setting, the inner setting, and the process of implementation) through the study (Figure 1).

Domains of the consolidated framework for implementation research with study-specific examples.
The study team included 8 members of the Association of Pediatric Program Directors (APPD) Subspecialty Pediatrics Investigator Network (SPIN), representing 7 different subspecialties, and a dedicated study coordinator who conducted the interviews. 29 The APPD SPIN team members all had significant knowledge and experience with training pediatric subspecialty fellows and GME but variable personal experience with EPA implementation. One member (MLL) with expertise in qualitative research led the team and provided training to both the study coordinator and other research team members.
We used purposeful and snowball sampling methods to identify FPDs within various pediatric subspecialties who were aware of and willing to discuss the EPAs. Participants had to be willing to be interviewed and recorded; there were no other exclusion criteria. Potential participants were recruited through email sent to all FPDs and in-person conversations by APPD SPIN Steering Committee members. Prior to the interview, we collected information from 42 FPD volunteers through close-ended questions about their pediatric subspecialty, number of years as FPD, geographic region, size of the fellowship program, and self-rating of EPA understanding. Participants were chosen to balance FPDs from different subspecialties, program size, geographic region and current uses of EPAs. From November 2019 to September 2020, the APPD SPIN study coordinator conducted semi-structured in-depth 1:1 interviews. Participants were divided into 2 groups based on their self-reported use of EPAs as users and nonusers. EPA nonusers were included so as not to overlook valuable insights and understanding regarding rationale for not implementing EPAs or failed implementation experiences and identify specific barriers unique to this group. For each group, a separate semi-structured interview guide was created with questions mapped to the relevant CFIR domains. (Supplemental Appendix 1) We piloted interview questions among APPD SPIN members, conducted 4 pilot interviews, and iteratively revised the interview guide as needed. 30 Data from pilot interviews were not included in the analysis. Participants who completed an interview were offered a $25 gift card for their time. We conducted interviews until we reached theoretical sufficiency, or a satisfactory conceptual depth to allow for thematic analysis. 31
We conducted interviews via a teleconferencing platform in which audio, but not video, was digitally recorded and transcribed verbatim by a professional transcription service (Rev.com). One person conducted all interviews, and this individual was located at the Lundquist Institute for Biomedical Innovation at Harbor-UCLA Medical Center. Each participant received copies of both the common and subspecialty-specific EPAs prior to the interview. We asked EPA users to describe their current use of EPAs and the implementation process, and EPA nonusers about their understanding of EPAs; both were asked about potential facilitators and barriers to EPA use. Both groups were asked about fellow evaluation processes, the concept of entrustment, and their perceptions around the value of EPAs, including the relationship to milestones. We modified the interview guide after the first and second rounds of analysis. To enhance the trustworthiness of the data, we created an audit trail to document any changes to the initial interview guide as well as the development of the coding structure.
Analyses
Interview audio-files and transcripts of the interviews served as our data sources. To avoid contamination and bias, 4 authors independently analyzed and coded data for the EPA users and 3 different authors coded data for the EPA nonusers. One additional author (MLL) coded all transcripts. We used inductive analysis grounded in the views of the participants, as opposed to testing/confirming theory.32,33 First, team members read through transcripts to gain an overall understanding. Each document was then reviewed line by line and codes (words that act as labels for important concepts) were applied to excerpts of the interview transcripts. The initial code structure for each group was created using the coding data from the first 3 transcripts. The code structure was then iteratively revised with the review of each additional transcript. Each team met to review transcripts and engaged in discussion to attain consensus in coding and assess coder reliability. 30 Related codes were combined into categories which were then developed into themes that described overarching concepts. 32 After the final code structure was established, the transcripts were reviewed and the final codes applied to the data. The teams then met to discuss comparable and contrasting concepts and themes between the 2 groups of participants, a form of investigator triangulation. 34 Dedoose, a qualitative analysis software tool, was used as an organizational tool to support the analysis of data (SocioCultural Research Consultants, LLC.).
The interview guide and qualitative analysis included secondary objectives to examine the process used by CCCs to assess their fellows and gauging opinions from FPDs about EPAs, particularly around readiness for graduation. These results have been previously published. 13
Ethical approval for this study was obtained by the Lundquist Institute for Biomedical Innovation at Harbor-UCLA Medical Center Human Subjects Committee (IRB number 18CR-31910-01, approved 09/13/2019). The Human Subjects Committee approved the use of verbal consent, and the interviewer maintained a list of those subjects who consented to participate.
Results
We interviewed 28 FPDs representing 11 different subspecialities of differing program size with varying experience as a program director. 13 Among the participants, 16 (57%) self-identified as EPA users. EPA understanding was similar to prior reports 17 ; EPA users were more likely to report an in-depth and expert understanding of EPAs, whereas nonusers reported were more likely to report either unfamiliarity or only a basic understanding of EPAs. FPDs reported a median of 3 fellows in their program and 6 years of experience as program director. Five major themes for EPA implementation emerged from analysis: (1) Facilitators to EPA implementation, (2) barriers to EPA use, (3) role of knowledge dissemination surrounding EPAs, (4) limited use of EPAs, and (5) comparisons between EPAs and milestones. Representative quotes have been included in Table 1.
Major themes and representative quotes.
EPA, Entrustable Professional activities; ACGME, Accreditation Council for Graduate Medical Education; SPIN, Subspecialty Pediatrics Investigator Network.
Facilitators of EPA implementation: Ease of use and perceived relevance of EPAs aid implementation
The overall ease of use and relevance to practice were 2 important facilitators of EPA implementation. FPDs related the intuitive nature of entrustment and the assessment of supervision level to EPA usability. The relevance to subspecialty practice and simplicity of wording for both EPA description and scale of assessment also facilitated ease of use. FPDs described certain EPAs as more highly valued, particularly the subspecialty-specific ones, whereas some common subspecialty EPAs (eg, Contribute to the Fiscally Sound, Equitable, and Collaborative Management of a Healthcare Workplace) were felt to be less important or useful. Furthermore, FPDs believed EPAs accurately reflect activities that are specific to each subspecialty.
Barriers to EPA use: Addressing the perceived value of and effort needed to implement EPAs are important barriers to overcome
The main barriers to EPA implementation included the potential additional workload involved in implementing a new method of assessment, such as the EPAs, and the uncertainty around the value added. Nonusers in particular were unsure how EPAs are distinct from milestones. They also described an unclear rationale for EPAs, both as a general concept and regarding their applicability to trainees. Some FPDs expressed the view that EPAs are not a comprehensive assessment tool and, therefore, were less likely to use them. Even among those FPDs who recognized the value of EPAs, the concern about workload for themselves and faculty dominated.
FPDs described lack of incentives, a lack of buy-in for change, and a need for training as other barriers to implementation. The time and effort involved with implementation could seem overwhelming as this involves software changes, tool development, as well as training and education for administrators, faculty, and fellows. This was particularly important given the apparent limited knowledge dissemination as described in the theme below. Since FPDs have many other requirements involved in running a program, their priorities could be shifted away from implementing new practices. Furthermore, FPDs who were new to their role and had less experience expressed less ability to innovate as they acclimated to their role and focused on fellowship recruitment, mentorship, and educational needs.
Knowledge dissemination surrounding EPAs: Training resources are needed to better understand the EPA framework and rating scales
Overall, FPDs described a general lack of knowledge and training resources on EPAs. For those familiar with EPAs, the most common sources of information were national meetings, particularly within their own subspecialties. Several also described personal involvement in the development of the EPAs. Institutional training on EPAs was uncommon. Many FPDs expressed a limited understanding of how the EPA framework can be applied and a desire for information from a trusted centralized source.
Few participants described efforts to learn more about EPAs through self-directed learning. They reported perceptions of the EPAs as an unfinished product, still being developed and evolving as a concept. FPDs expressed a desire for clarity about how to use the EPAs overall and how to use the rating scale, with some FPDs advocating for the development of a better assessment tool. Many FPDs in the nonuser group did not realize there was a published rating scale with a validity evidence to assess trainees on their performance of EPAs. FPDs also commented that the value of EPA assessment is unclear when compared to milestone-based assessment.
Limited use of EPAs: Data requests from governing bodies will prompt FPDs to use EPAs beyond research participation
The current use of EPAs was often limited to fellow assessment, and frequently in the context of research studies. Additional applications of the EPA framework reported were curriculum development, individual assessment, stimulating discussion for learning plans and remediation. Several reasons influenced the FPDs use or nonuse of the EPAs. Among EPA users, personal incentives to participate in research, a sense of good citizenship, advancement of educational methodology, or the influence of an EPA champion within their subspecialty were mentioned. Nonusers of EPAs were more likely to believe EPA assessment involved unnecessary work with little additional value. It was unclear to FPDs how additional assessment data would be used by governing bodies and future employers, impacting the perceived value of EPAs and the incentive to implement a new process to assess fellows. Indeed, many FPDs commented on the need for these data to be required or requested by an external organization (eg, ABP, ACGME) prior to initiating EPA assessments.
Comparison of EPAs to milestones: Desire for a tool that combines the best aspects of both assessment methods
The respondents discussed various strengths and weaknesses of both the EPAs and the milestones. Commonly mentioned differences between the 2 included the complexity of language and concepts, the scope of content within each component, and the user-friendliness of the assessment tools. Most EPA users appreciated the clear and simple language of EPAs and the level of supervision scale commonly used with EPA assessments compared with milestones. Additionally, the intuitive nature of entrustment and real-world applicability of EPAs made this tool seem more meaningful and often preferred to milestones.
However, some FPDs expressed concern that EPAs were less specific than milestones and often too broad for meaningful feedback. In this way, they described the potential complementary broad nature of EPAs with the more granular milestones for informing assessment and learning plans. FPDs also described how the extremes of the scale may not be applicable to all EPAs. For example, they questioned level one (Observe only) for certain EPAs such as Handovers, and level 5 (Trusted to lead at regional or national levels) for EPAs like Engage in Scholarly Activities and Lead within the Subspecialty Profession, as these are felt to be aspirational for faculty. Participants also discussed the role of governing bodies as they compared the use of EPAs and milestones. While both tools are aligned and connected to one another by way of the ACGME competencies, the requirements set forth by the ACGME are a primary driver of milestone use, making this the default assessment. FPDs expressed a desire for a single approach to assessment either by replacing milestones with EPAs or better integrating the 2 together.
Importantly, FPDs highlighted that the usefulness of any framework relied upon the ability to provide meaningful assessments, identify deficiencies, create action plans, and discriminate between individuals and assessment levels. These included assessments related to readiness for graduation from fellowship. They also discussed the value of narrative feedback as opposed to objective assessment ratings.
Discussion
Our study represents the first characterization of pediatric subspecialty FPDs’ experiences with and perspectives of EPAs. The implementation process for any new tool is complex, often requires significant time and resources, and has varying degrees of success. We used the CFIR framework to explore the different constructs involved in the implementation of the EPAs, which include domains such as the outer and inner settings, individuals, intervention, implementation, and process. Participating FPDs described several facilitators and barriers to using EPAs including a general lack of knowledge regarding their purpose, available rating scales, and integration with the milestones. Even though EPAs were created for pediatric subspecialties almost a decade ago, few FPDs use them outside of assessment and research.
A majority of participants described the role of regulatory requirements (Outer Setting) to adopt the EPAs consistently within their program, using the ACGME mandate for milestone reporting as an example. Since the completion of our study, the ABP announced a plan to require EPA assessments in the certification process for pediatric subspecialties beginning in 2028. Therefore, all pediatric fellowship programs will be required to use EPAs in this capacity within the next few years. However, the goal is not to simply create another task for program directors but rather to maintain a standard for the public and provide a better framework for ensuring pediatric subspecialty fellows graduate with the ability to perform the essential activities needed by their patients without supervision.
For more effective implementation, our results suggest the need for a broader multi-faceted approach to educating FPDs (Individuals) about the EPAs and ways to use them meaningfully. Depending upon the individual FPD's background and interests, the understanding of CBME and how the EPAs fit into the overall application of CBME may be limited. Outreach through multiple forums, as described by study participants, is likely necessary. The desire for standardized materials for training and implementation was stated as useful for ease of implementation. Having standardized resources may not only ease implementation but also help promote a shared mental model across the pediatric subspecialty educational community, a factor key to successful implementation. 35
Additionally, knowledge dissemination through existing communities of practice such as national or local educational meetings and the importance of “champions” who can support and educate less experienced FPDs (Inner Setting) were highlighted. Leveraging and reinforcing this concept of educational communities of practice may not only improve engagement in learning, particularly for the novice user, but also effectiveness of practice. 36 The positive influences of both communities of practice and champions have been demonstrated in other healthcare-related implementation efforts.37,38 Self-directed learning was less commonly reported, primarily among those who were more actively engaged in the development of EPAs. However, the role of self-directed learning may become more relevant now that the external forces for EPA implementation (ie, ABP requirement) are more prominent.
When implementing a new(er) framework such as EPAs (Intervention), there needs to be consideration of the complexity of the intervention and knowledge around the value gained. 26 FPDs familiar with EPAs describe the ease of understanding them when used for assessment because they describe real clinical tasks in their daily practice. This impression is consistent with other studies describing perceptions of EPAs by learners and faculty and the validity work on the level of supervision scales, where participants received no training prior to use.11,16,18 Furthermore, the concept of entrustment, while complex and nuanced in decision-making, can be intuitive for faculty and program directors. Thus, the scales anchored in levels of supervision translate more easily with an assessment of fellows. More importantly, entrustment can be seen as the pinnacle of training before launching into real-world practice after graduation. 39
While many FPDs recognize the value of the EPA concept for the reasons just described, they also expressed a desire for additional information about validity, application, and impact on outcomes. Although validity evidence has been generated for the assessment scales of the common pediatric subspecialty EPAs, there seemed to be a lack of awareness among FPDs, especially nonusers, which further supports the need for strong knowledge dissemination efforts. 16 Similar efforts for generating validity evidence for subspecialty-specific EPAs will also be important and are ongoing. Focusing on the subspecialty-specific EPAs, as those were identified as more meaningful, may be an effective approach during the initial stages of implementation. FPDs also raised questions about how the use of EPAs would impact progression through training, the range of the scale, determination for graduation readiness, and clinical performance postgraduation. Some work in these areas has been reported in other fields and other countries, including addressing the important issue of their use for high-stakes decision-making for graduation or certification.11,18 Recent studies also suggest FPDs may believe the level of supervision required at graduation may differ based on the specific EPA.17,40
Finally, given the current ACGME requirement for milestone reporting, FPDs queried the integration of an EPA framework into the existing system. Many recognized the complementary nature of milestones and EPAs, as previously described. 8 For example, the EPA framework for assessment may better identify a fellow's readiness to perform the essential activities of a subspecialty or their need for remediation, while the milestones provide an assessment of the specific competencies required for the execution of those activities. In this way, FPDs described using both tools to not only perform assessment but also inform individualized learning plans and curriculum development. As the ACGME, ABP, and pediatric education community work together to develop Milestones 2.0 for each subspecialty, this process presents an opportunity to educate FPDs about the connection between milestones and EPAs and continue to align the overall CBME assessment framework in pediatric subspecialties. Collaborative efforts between the ABP and APPD SPIN to directly connect subspecialty EPA supervision scale ratings to milestone subcompetencies have already begun, and similar efforts have been demonstrated in pediatric and medicine residency programs.41-43 Dissemination of these efforts in accessible and usable forms will be essential, particularly given the concern expressed by FPDs, including current EPA users, about additional workload, and integration into the current workflow. Access to common resources for training, assessment tools, and application using currently available electronic platforms can help support time-constrained FPDs during implementation.
Several limitations to our study should be acknowledged. We intentionally aimed for a diverse cohort of FPDs (experience, geography, and subspecialty) and to provide a balance of self-reported EPA users and nonusers. However, EPA use was often confined to assessment and research participation, thus, the identified factors influencing successful implementation for other applications may be different with greater understanding and use. The themes of limited knowledge and barriers will be important for education and dissemination efforts. A quantitative study of all FPDs was informed by our data and may capture additional information on our aims, as well as providing a framework for implementation. The research team comprised highly engaged educators, potentially providing a deeper/nuanced understanding of the participants’ words but the identified codes and themes may have been different with a group less involved in the broader educational community. Finally, our study was conducted prior to the ABP's notification regarding EPA assessments and certification, so we were unable to explore further FPDs reaction and anticipated needs with this change.
Conclusions
The key themes identified in this work contain important factors to inform the approach to effective and meaningful implementation: (1) supporting and educating FPDs about the EPAs and the level of supervision rating scales; (2) ongoing assessment of the value of EPAs in training and outcomes after graduation which may include patient care metrics, involvement in quality and safety programs, and leadership activities; and (3) practical integration with current systems and workflow such as dissemination of specialty-specific evaluation tools and the ability to use of EPA ratings to generate the required milestone report. Generating additional data and engaging full stakeholder input will be critical to avoid this approach being viewed as another regulatory requirement and accomplish meaningful and transformative implementation across the pediatric subspecialty community.
Supplemental Material
sj-docx-1-mde-10.1177_23821205231225011 - Supplemental material for Exploring Factors for Implementation of EPAs in Pediatric Subspecialty Fellowships: A Qualitative Study of Program Directors
Supplemental material, sj-docx-1-mde-10.1177_23821205231225011 for Exploring Factors for Implementation of EPAs in Pediatric Subspecialty Fellowships: A Qualitative Study of Program Directors by Angela S. Czaja, Richard B. Mink, Bruce E. Herman, Pnina Weiss, David A. Turner, Megan L. Curran, Diane E. J. Stafford, Angela L. Myers and Melissa L. Langhan in Journal of Medical Education and Curricular Development
Supplemental Material
sj-docx-2-mde-10.1177_23821205231225011 - Supplemental material for Exploring Factors for Implementation of EPAs in Pediatric Subspecialty Fellowships: A Qualitative Study of Program Directors
Supplemental material, sj-docx-2-mde-10.1177_23821205231225011 for Exploring Factors for Implementation of EPAs in Pediatric Subspecialty Fellowships: A Qualitative Study of Program Directors by Angela S. Czaja, Richard B. Mink, Bruce E. Herman, Pnina Weiss, David A. Turner, Megan L. Curran, Diane E. J. Stafford, Angela L. Myers and Melissa L. Langhan in Journal of Medical Education and Curricular Development
Footnotes
Acknowledgments
The authors would like to thank our study coordinator, Marzia Hazara, B.S., and our consultant, Dr Dorene Balmer, for their contributions to the study and the ABP Foundation for their financial support.
Author contributions
Angela S. Czaja, Richard B. Mink, Bruce E. Herman, PninaWeiss, David A. Turner, Megan L. Curran, Diane E. J. Stafford, Angela L. Myers, Melissa L. Langhan all participated in the analysis of data, drafting and critical review of the manuscript and final approval of the work product. Richard Mink and Melissa Langhan designed the study and obtained funding for the study.
Declaration of conflicting interests
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: David Turner is employed by the American Board of Pediatrics; the remaining authors have nothing to disclose.
Funding
The authors received the following financial support for the research, authorship, and/or publication of this article: This study was supported by a grant from the American Board of Pediatrics Foundation. This manuscript does not represent the views of the ABP or the ABP Foundation.
Supplemental material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
