Abstract
Introduction
Medical education applications (apps) play a crucial role in the learning process, serving as valuable resources for both healthcare professionals and students. Their presence and widespread adoption in medical environments have significantly enhanced the overall learning experience. 1 With a range of adaptable features, apps have proven to be highly effective tools in supporting and enriching medical education. 2 Medical apps designed to aid in education can be used for several purposes such as accessing online textbooks, medical podcasts, medical calculators, online lectures, and note-taking. 3 Some studies have explored the advantages of medical apps for medical students in a clinical setting, which included aiding decision-making, providing rapid access to reliable clinical information, and assisting with diagnosis and treatment choices such as medication dosage and drug interactions.4,5 While other studies have shown that students who utilize medical apps have reported a positive impact on their educational experience, 6 and highlighted the significant association related to improving the knowledge and skill levels of apps’ users compared to non-users. 2
Although medical apps offer numerous advantages, the vast number of available options can make it challenging for medical students to navigate and independently choose the most suitable apps for their educational needs. 7 Several studies have explored medical apps from the perspective of students and their experiences.5,8–10 However, structured evaluation methods, along with faculty guidance, are essential to help students identify and adopt the most effective apps tailored to their needs. 8 Jonas-Dwyer et al. suggested that faculty and students actively collaborate to evaluate apps and benefit from their shared experiences to enhance teaching and learning. 11 Singh et al., recommended a faculty-guided approach in choosing appropriate and cost-effective apps. The authors emphasized the vital role that universities should play in providing continuous teaching and learning initiatives for educators. These initiatives would improve collaborative, and interactive teaching techniques potentially redefining educators’ roles as facilitators in the educational process. 6
The limited number of medical app evaluation tools we found available in the literature,12–14 combined with the lack of standardized criteria for evaluating medical apps, poses challenges to their use in teaching settings. Faculty play an essential role in integrating apps into the learning framework, 15 yet incorporating educational and medical apps into medical curricula is still rare. 16 Some methods, such as elective courses, 16 daylong workshops,17,18 and hands-on training 19 have been reported to help students evaluate medical apps effectively.
As digital transformation in healthcare increasingly impacts patient care, 20 it is essential to evaluate the innovative contributions, and adoption barriers and facilitators of these tools from the perspectives of future medical physicians as well as faculty. Equipping medical students with app evaluation skills is essential in an era where mobile health applications are increasingly used to support medical education, clinical practice, patient education, and self-management. 21 These skills enable students to critically assess the quality of apps, ensuring they can integrate evidence-based technology into their education future medical practice.22,23 Sharing insights from faculty can provide valuable information to medical educators and educational institutions regarding effective strategies for integrating and evaluating medical apps. 24 To address this, we designed a workshop for third-year undergraduate medical students, with the aim to equip them with the skills to evaluate various medical apps using standardized evaluation instruments. Specifically, we utilized the the Mobile App Rubric for Learning (MARuL), a validated and reliable instrument to assess the value of apps for supporting just-in-time learning. 13 This research aims to describe the workshop and evaluate both the students’ and faculty’s experiences using the MARuL instrument.
Material and methods
Study design, setting and workshop structure
This study used a qualitative research design with a consecutive sampling method, including all third-year medical students at King Saud University. These students were chosen specifically because they were enrolled in a medical informatics course that incorporated app evaluation into the curriculum, making them the ideal population for assessing the workshop’s impact on app evaluation skills. As part of this course curriculum, students are required to attend a workshop titled “Health Apps Evaluation”. The workshop is designed to teach students how to assess medical apps using a standardized framework.
To design the workshop structure, we as Associate Professors in health informatics with extensive experience in conducting qualitative studies, held two meetings each lasting 2 hours to identify which evaluation instrument to use, familiarize ourselves with the evaluation process, and develop the student guides and instructions. The final developed workshop guide included five sections: (1) a brief slide presentation outlining the workshop’s objectives and structure, (2) a didactic presentation on standard evaluation tools, (3) a detailed review of the MARuL instrument, (4) student instructions on how to fill the evaluation tool using Google Forms, and (5) student presentations outlining the results of their evaluation. We then held a three-hour meeting with other faculty facilitators to ensure a consistent approach in delivering the workshop sessions. This meeting established a standardized method for managing each session.
We used the MARuL instrument, 13 due to the developers’ rigorous testing of the instrument across 46 medical apps from the iOS and Google Play Store, demonstrating its reliability and broad applicability. The instrument includes 26 items across four measures: Teaching and Learning, User-Centered, Professional, and Usability. The Teaching and Learning measures include 9 items that evaluate an app’s effectiveness in facilitating education including its purpose, ability to promote learning, relevance to a course or study, and instructional features. The User-Centered measures include 7 items that focus on assessing user engagement by examining subjective quality, satisfaction, perceived usefulness, importance, user experience, and the intention to reuse the app. The Professional measures include 3 items, which evaluate the app’s alignment with professional standards, the credibility of its developers, and the quality of the information it provides. Lastly, the Usability measures include 7 items, which assess the app’s design aesthetics, functionality, distinctiveness, ease of use, absence of advertisements, technical specifications, and whether it offers a better experience than its web-based counterpart. 13 The MARuL uses a 5-point Likert-type scale as the rating tool for each item, where 0 indicates the item does not fulfill requirements, 1 means it poorly fulfills requirements, 2 somewhat fulfills requirements, 3 mostly fulfills requirements, and 4 fully meets requirements. The scoring is divided into four sections: Teaching and Learning (maximum score of 36), User-Centered (maximum score of 28), Professional (maximum score of 12), and Usability (maximum score of 28). The final score is calculated by summing the scores from all four sections, resulting in an Overall Usefulness for Learning Score, with a maximum possible score of 104.
In preparation for the workshop, we designed two Google Forms. The first form included the MARuL tools’ 26 items for students to fill out during the workshop, while the second form included two open-ended questions designed to capture students’ feedback at the end of the workshop regarding their experience with using the MARuL instrument. These questions were: “What was your experience with using the MARuL instrument?” and “What were the challenges you faced in using the MARuL instrument?”(Supplemental file).
The MARuL instrument was validated by Gladman et al., the instrument’s developers. 13 While the second form was reviewed for content validity by three experts in medical informatics and e-learning. Minor adjustments in some of the wordings were made based on the experts’ feedback.
The workshop was conducted three times over three consecutive days, with each session lasting 3 hours and involving a total of 275 students, representing the entire population of third-year medical students for the year 2023. Two sessions were held face-to-face, and the third session was conducted online via the Zoom videoconferencing platform. For the face-to-face sessions, students were required to bring a laptop or tablet and connect to the college’s wireless internet.
This study was approved by the Institutional Review Board at King Saud University.
Student groups
We assigned students into groups based on their previous experience and current use of one of four medical apps for learning: AMBOSS, ISABEL, Medscape, and OSMOSIS. These apps have distinct features and functionalities, covering different types of learning.
In total, there were 31 student groups, each group had 6 to 12 students, with each group having a student leader (face to face sessions had nine groups, while the online session had 13 groups). Students were required to evaluate their assigned app individually using the MARuL instrument, and then work as a group to prepare their presentations and describe their experience and the challenges they encountered using the evaluation instrument. Each group’s results of the evaluation were presented as a presentation to the entire class. The facilitators guided the discussion, both among the individual groups as they evaluate the apps and during the large-group discussion. At the end of the workshop, we asked students to fill out the end-of-workshop feedback online form.
Faculty evaluation
After completing the three workshops, we met for 3 hours to reflect on the sessions and identify areas for improvement. Discussions focused on seven main areas: (1) the importance of app evaluation tools to guide medical students in selecting learning apps, (2) the benefits of students using the MARuL instrument to evaluate medical apps, (3) challenges in explaining the MARuL instrument to students, (4) areas for improving the MARuL instrument for student use, (5) challenges in conducting the workshop, (6) challenges in conducting the workshop virtually, and (7) areas for improving the workshop overall.
Data analysis
To analyze students’ responses to the open-ended questions about the MARuL instrument, we conducted a thematic content analysis. 25 We looked for repetition and statements relevant to the challenges of using the instrument. After individually reviewing student responses, we met to identify themes through consensus and created standardized terms and definitions for these themes. Data saturation was reached with no new themes or insights emerging from the data. To enhance the validity and comprehensiveness of the themes, we categorized the responses collaboratively to reach 100% agreement.
Results
Student evaluation
A total of 275 students participated in the workshop, consisting of 152 males (55%) and 123 females (45%), with a mean age of 21 years. The students’ evaluation results of the four apps using the MARuL instrument during the workshop are published elsewhere. 26
Feedback from students regarding the MARuL instrument.
Theme 1: Length and time-consuming nature
Many students found the evaluation process lengthy and time-consuming, with the instrument containing too many categories. One student mentioned: “It [MARuL form] was a long form” Another student stated: “The questions and answers contain a lot of details”, while another student described the amount of time it took them to complete the evaluation form: “The answers were very detailed, long and sometimes confusing… Long answers It took a lot of time to finish it”.
Theme 2: Challenging terminology
Students expressed struggling with some of the technical terms used in the MARuL instrument, finding the language sometimes difficult to understand. One student mentioned: “the vocabulary used was a bit hard to understand”. Another student noted: “Could use [the form] simpler word to understand”.
Theme 3: Limited or similar answer options
Students also commented on the answer options provided in the instrument, stating that they were often limited or too similar, causing confusion and difficulty in choosing the most appropriate response. One student noted: “The choices were similar to each other we could choose by mistake because of similarity”. Another student expressed: “Some choices are similar and need to be reviewed”.
Theme 4: scoring and interpretation
Students expressed challenges in understanding how to score the apps and interpret the results, feeling that the scoring system was not user-friendly. One student specifically noted the time-consuming nature of calculating the scores: “Summing the score for each domain was difficult”. Another student pointed out: “The presence of parts made the score difficult to calculate”.
Theme 5: absence of open-ended questions
The absence of open-ended questions was another concern, as students felt that such questions would have allowed them to provide more detailed feedback and share their perceptions more effectively. One student mentioned: “No open ended question, for example we can’t mention any specific limitation regarding the application”. Another student suggested that the form: “requires some open-ended questions”.
Theme 6: Difficulty finding required information
Other students found it challenging to locate the necessary information stated in the instrument, within the apps to complete the evaluation. One of the students stated: “…some of the questions did not apply to the software tool we were reviewing”.
Theme 7: Lack of Comparative measures
The lack of comparative measures within the instrument was also noted, with one student suggesting that the form could include ways to compare different apps directly, which would have improved the evaluation process: “There was not much comparison with other applications/websites in specific aspects eg. “Between the chosen app and others, is it the most time-efficient website you’ve used?”.
Faculty evaluation
Faculty evaluation regarding the MARuL instrument.
Faculty evaluation regarding the workshop challenges and areas for improvement.
Discussion
The Health Apps Evaluation workshops were overall well received by both students and faculty. These workshops represent a new approach in integrating the use of standardized app evaluation instruments into the medical curriculum. Our aim was to prepare third-year medical students with the critical skills necessary for evaluating medical apps, which is vital considering the increased adoption rates of medical education apps by students. 2 Our study provides valuable insights into the benefits and challenges of this approach, highlighting key areas for improvement and broader implications for medical informatics education.
A significant challenge we identified was the students’ initial unfamiliarity with structured app evaluation methods. Many students were used to informal methods of app selection, which were often based on app user ratings and personal preferences. These findings align with those reported in the study by Singh et al. 6 The students’ unfamiliarity with structured evaluation methods required extending the time we spent on the introductory part to explain the scientific basis and purpose of instruments like the MARuL. We were also faced with logistical challenges, especially the management of large group sizes, and students’ assignment to their respective groups. These factors potentially had an effect on the workshops’ time management and delivery. To address these issues, it is critical to ensure that students are given a separate lecture on health technology evaluation methods and integrating app evaluation into different teaching methods, such as problem-based learning.27,28 In doing so, this could enhance the relevance and impact of utilizing evaluation methods when choosing medical apps.
While the workshop we conducted through the Zoom platform offered time flexibility, we faced several challenges when the workshop was conducted virtually. The workshop lacked the interpersonal student engagement and interaction, which we found in the face-to-face settings. The absence of physical presence made it difficult to measure the level of student engagement. We also felt that promoting group dynamics and collaborative learning experiences among our students were not maintained, which also had an effect on peer discussions that had to occur as part of the evaluation process. Similar challenges were reported by Rajaraman et al., regarding online teaching in higher education. 29
Post-workshop feedback from students highlighted the significant educational benefits of using a standardized evaluation instrument like MARuL. Students reported that the structured approach of MARuL facilitated a more thorough and objective evaluation process compared to their common app selection methods. The structured evaluation process helped students think more critically about the educational value of the apps, which potentially enhanced their digital literacy and their ability to integrate the use of the most effective technology tools into their future medical practice. The feedback from students was found to be consistent with the findings reported by the Gladman et al. 22 MARuL’s integration of teaching and learning elements within its multidimensional quality assessment framework may make it particularly well-suited for evaluating apps focused on education and learning. 22
Feedback from both students and faculty indicated that while the MARuL is a vigorous instrument, it may require refinements to better suit the needs of all medical students. Some of our students found the terminology used in the instrument challenging and the limited answer options somewhat restrictive. In addition, the absence of open-ended questions and comparative measures limited the students’ ability to provide a comprehensive evaluation. These comments may suggest that the MARuL instrument could be made more user-friendly to fit a broad range of student needs coming from different backgrounds.
This workshop was found valuable by students and addresses gaps identified in both the literature and post-workshop feedback. Medical apps are becoming an increasingly vital part of medical practice, 21 and the skills to accurately evaluate these apps should be an integral part of the medical curriculum. 23
Our study has several limitations that should be addressed in future research. Firstly, the study was limited to third-year medical students at a single institution, which may not be representative of all medical students. Expanding the study to include students from various academic years and multiple institutions would provide more generalizable results. Secondly, the logistical challenges related to group sizes may have impacted the depth of the post-workshop feedback. Future studies should introduce several evaluation instruments to students in an effort to compare the user-friendliness, and usability of such evaluation tools by medical students.
Conclusion
Evaluating medical apps requires medical students to develop critical thinking skills to assess the app’s relevance and usefulness, along with a solid understanding of standards and measures to ensure alignment with evidence-based practices. Training medical students to apply these skills presents unique challenges, including selecting the appropriate evaluation instrument relevant to their practice and determining effective methods for teaching the use of these tools. This study highlights our experience in conducting workshops designed for medical students to assess medical apps using a standardized framework, outlining benefits and underlining areas needing improvement. Our insights gained from conducting these workshops may assist medical schools and educators on approaches to integrate medical apps evaluation methods into medical education.
Supplemental Material
Supplemental Material - Workshop on evaluating medical apps using the mobile app Rubric for learning
Supplemental material for Workshop on evaluating medical apps using the mobile app Rubric for learning by Raniah N. Aldekhyyel and Jwaher A. Almulhem in Health Informatics Journal
Supplemental Material
Supplemental Material - Workshop on evaluating medical apps using the mobile app Rubric for learning
Supplemental material for Workshop on evaluating medical apps using the mobile app Rubric for learning by Raniah N. Aldekhyyel and Jwaher A. Almulhem in Health Informatics Journal
Supplemental Material
Supplemental Material - Workshop on evaluating medical apps using the mobile app Rubric for learning
Supplemental material for Workshop on evaluating medical apps using the mobile app Rubric for learning by Raniah N. Aldekhyyel and Jwaher A. Almulhem in Health Informatics Journal
Footnotes
Acknowledgments
We would like to thank the Deanship of Scientific Research, Research Chairs Program, King Saud University for their support in conducting this study.
Ethical approval
This study was approved by the Institutional Review Board at KSU, College of Medicine (IRB# E-23-7811).
Informed consent
Written Informed consent for participation in this research was obtained from students.
Author contribution
All authors made substantial equal contributions to this work. All authors contributed to the conception and design, as well as the acquisition, analysis, and interpretation of data. Additionally, all authors participated in drafting the manuscript and both have given final approval of the version to be published and has agreed to be accountable for all aspects of the work, ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Study setting
Third-year medical students at King Saud University (KSU) take a mandatory medical informatics course.
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Data Availability Statement
The data that support the findings of this study are available on request from the corresponding author (RA).
Supplemental Material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
