Abstract
The integration of artificial intelligence (AI) in health care, particularly in colorectal cancer screening, represents a transformative advancement in medical diagnostics. This commentary explores the development and deployment of AI platforms in colonoscopy, highlighting the critical role of human–AI interaction and the valuable lessons learned throughout this journey. We discuss the challenges of integrating AI into clinical settings, including the necessity for robust AI validation, transparency, and accountability. Key insights from our experience emphasize the importance of effective human–AI collaboration to enhance diagnostic accuracy and procedure quality. This commentary aims to share critical insights that can guide future AI applications in various medical disciplines.
Introduction
The potential of artificial intelligence (AI) to revolutionize health care, particularly in cancer screening, has been increasingly recognized.1,2 Cancer screening aims to detect cancer at an early stage, but human error and variability can limit its effectiveness. AI platforms can address some of these challenges by providing consistent, accurate, and real-time analysis of medical images and videos.3,4
Colorectal cancer (CRC) screening is a prime example of how AI can make a significant impact. CRC is the third most common cancer worldwide, with precancerous polyps offering a window for early detection and intervention. However, identifying and characterizing these lesions can be challenging even for experienced endoscopists. 5 AI platforms can aid in this process by enhancing the diagnostic accuracy of colonoscopies.2,6
The development of AI platforms for polyp detections during colonoscopies represents a significant milestone in the intersection of AI and health care. 1 In the journey of developing a successful AI platform for screening colonoscopy, numerous valuable lessons have been gleaned. These lessons, derived from the intricate process of translating AI from development to deployment in health care, have significantly shaped our understanding of the challenges and opportunities in this field.
The genesis of the first AI platform for polyp detection to be cleared in the United States7–9 was rooted in the need to address unmet clinical needs in gastroenterology, specifically to reduce the adenoma miss rate and enhance overall procedure quality in colonoscopy through AI technology. This endeavor began before AI became a mainstream tool in health care, capitalizing on a unique opportunity to record and analyze colonoscopy videos to aid early polyp detection. 8
Following this initial de novo clearance by the FDA, other companies joined the AI race in gastroenterology within the United States and worldwide.2,6,10 The evolution of these AI platforms since their inception has been remarkable. These devices have been adopted globally, particularly in the United States, where they have significantly influenced the field of gastroenterology. The AI platform’s success in improving clinical outcomes has been corroborated by numerous studies and trials, affirming their utility and effectiveness in enhancing gastroenterological procedures.11–13 These findings have been instrumental in establishing the credibility of these AI platforms and highlighting their potential to revolutionize CRC screening. Although high-quality evidence 13 has confirmed the benefits of AI in terms of credibility and improved adenoma detection rates, it is also essential to acknowledge that some data suggests a lack of performance improvement. 14
Lessons Learned
The journey into developing a successful AI platform for endoscopy has been a rewarding and enlightening experience. It has offered numerous lessons, some of which were expected, whereas others were surprising and unexpected. These lessons have not only shaped the development and deployment of the AI platform but have also influenced the broader understanding of AI in health care. In this commentary, we will focus on three specific areas where we learned some lessons we had not anticipated when we embarked on this journey.
Transparency and accountability
One of the first lessons learned during our journey is that developing AI technologies, particularly in health care, is a delicate balancing act. On one hand, we have the immense potential of AI to improve patient outcomes, streamline health care processes, and reduce costs. On the other hand, we have legitimate concerns about patient safety, data privacy, and the ethical implications of AI use. Striking the right balance between these competing interests is a complex task that requires careful consideration and a nuanced understanding of the technology and its potential impacts. Differing perspectives on what constitutes “responsible AI” are often influenced by the varying priorities of stakeholders. For instance, AI developers and manufacturers might focus on innovation and economic gains, whereas health care providers and patients are likely to emphasize safety and ethical standards.
Given these complexities, various authoritative bodies—including The White House, 15 The European Union, 16 WHO, 17 and FDA 18 —have set forth principles to guide the responsible development of AI. These guidelines stress the importance of transparency and accountability in AI development to foster wider acceptance and integration within health care settings.
As we progressed with our AI platform, we found that transparency about our development processes,5,19,20 the data sources used, 8 and sharing datasets to spur further innovation 21 significantly bolstered stakeholder confidence. Engaging with the scientific community through peer-reviewed publications has allowed us to subject our work to both regulatory oversight and expert academic critique, enhancing the credibility and reliability of our AI applications.
Another lesson learned regarding transparency and accountability involves robust validation and testing procedures for AI platforms, especially those intended for clinical use. Ensuring these platforms’ safety, efficacy, and reliability is imperative and requires comprehensive testing. This includes technical validation, where the AI’s performance is tested across various datasets; 18 clinical validation, such as conducting randomized controlled trials (RCTs) to evaluate AI’s impact on patient outcomes;22,23 and real-world studies, for ensuring AI-based medical devices continue to perform safely and effectively after they have reached the market.24,25 Adhering to established guidelines for conducting and reporting on AI performance studies is crucial. Guidelines such as DECIDE-AI, 26 SPIRIT-AI, 27 CONSORT-AI, 28 STARD-AI, 29 and TRIPOD-AI 30 offer structured frameworks that ensure transparency, accountability, and efficacy throughout the different stages of AI development and implementation. It is of vital importance for both the industry and the medical community to thoroughly understand these guidelines. A more profound grasp of the scientific requirements is essential for health care providers to discern well-conducted AI studies that provide solid evidence from the plethora of less rigorous research often found in the scientific literature. By enhancing this knowledge base, clinicians can effectively evaluate AI technologies, adopting only scientifically validated and effective solutions in clinical practice. Conversely, it is equally crucial that AI engineers and medical device manufacturers understand clinical needs and respond with appropriate solutions.
Communication between stakeholders
Perhaps one of the most challenging lessons learned is the need for effective communication and collaboration between AI developers and health care professionals. The successful deployment of AI in health care hinges on clinicians’ acceptance and adoption of the technology. This requires not only educating clinicians about the benefits and limitations of AI but also actively involving them in the development process to ensure that the AI platform meets their needs and fits seamlessly into their workflow.17,18,31,32
In the health care sector, the introduction of new technologies often encounters resistance because of a variety of factors, including the fear of obsolescence, concerns about patient safety, and the perceived threat to professional autonomy. 33 The adoption of AI in health care is no exception. Despite the demonstrated potential of AI to enhance clinical decision making and improve patient outcomes, its integration into routine clinical practice has been met with mixed reactions from physicians. 34 The introduction of AI in health care represents a paradigm shift in the practice of medicine, requiring physicians to adapt to new ways of working and thinking. This change can be challenging, particularly for those unfamiliar with AI or who may have reservations about its effectiveness or ethical implications. 35 Therefore, it is essential to address these concerns through effective communication, training, and support. In our experience, when physicians were not adequately prepared for the AI or did not accept the new tool, the clinical study outcomes were suboptimal. 14
Conversely, we learned that promoting information sharing is a powerful catalyst for increasing scientific papers using our technology. By encouraging investigators to publish their study results, regardless of outcome, we facilitate knowledge sharing and accelerate innovation. In less than 4 years since FDA clearance of our AI platform, more than 30 papers have been published utilizing our AI platform, with eight being RCTs. 36 This success is a testament to the importance of collaboration in developing AI solutions for health care.
One significant lesson learned from this experience was the value of a multidisciplinary approach in addressing unmet clinical needs and improving procedure quality through AI technology. 18 Our unique partnership between a pharmaceutical company, a global med-tech giant, physicians, and AI experts brought together diverse expertise in handling high-quality data, understanding clinical and regulatory needs, marketing and distribution, and AI for medical image diagnosis. 8 This collaboration was instrumental in developing our AI platform, which has improved polyp detection and overall procedure quality in colonoscopy. Our experience with AI-driven Software as a Medical Device (SaMD) highlights the importance of effective communication and resource sharing among stakeholders to prevent redundant efforts and ensure commercial viability.
Unfortunately, fragmentation is common among AI initiatives, leading to delays and failures. To address this issue, we propose introducing specialized marketplaces for AI applications that facilitate innovation sharing. These marketplaces can streamline regulatory approval processes, accelerate technology commercialization, and provide a structured environment where stakeholders can collaborate effectively. 37 By creating these ecosystems, more AI devices can reach patients, ultimately improving health care outcomes.
Human-AI interaction
The rapid growth in information and computer technologies in the 1980s led to the recognition of the importance of studying human–computer interaction. 38 This field aims to investigate social and psychological aspects of interactions that influence the acceptability and utility of new technologies. As humans’ methods of interacting with technology have become more sophisticated, so too has the complexity of their influences on and impact from these interactions. 39 Ease of use is a crucial driver of uptake for new technological products or platforms. 40 In medical science, successful implementation of AI platforms depends not only on technical success but also on the ability to interact effectively with human operators. 41
An important lesson learned is that when the appropriate conditions are met, the potential for a synergistic relationship between AI and human expertise becomes evident. 20 This “human-AI hybrid” model has demonstrated superior accuracy in medical tasks compared with either the AI or the human practitioner alone. The key to achieving this result lies in the AI’s ability to clearly communicate the boundaries of its predictions. By doing so, the physician can effectively integrate the additional information provided by the AI while also considering their own knowledge and the clinical context.5,20,42
Achieving optimal collaboration requires careful consideration of user interfaces and communication mechanisms between AI and physicians. The AI must provide clear, concise, and actionable information that physicians can easily interpret and apply. In addition, it should transparently disclose its limitations and uncertainties to prevent overreliance on AI and ensure physicians remain in control of clinical decision-making processes. 10 This “human-AI hybrid” model represents a significant advancement in AI health care applications, highlighting AI’s potential as a complementary tool that can enhance accuracy and efficiency.
Human–AI interaction will be crucial for individual application success or failure; therefore, optimizing interface elements is essential to prevent clinician frustration with poorly designed AI platforms. 32 Historically, the development of AI platforms has focused mainly on technical outcomes in idealized settings, expecting clinicians and patients to adapt to these technologies without sufficient consideration for human preferences and cognitive biases. 41 To create more user-friendly, effective, safer, and better-value AI systems, we must reorient our focus toward developing human-centered AI and incorporating the study of human interaction at each stage of platform development. By aligning with appropriate regulation and governance, we can ensure that AI platforms are designed with humans in mind, leading to more successful implementation and adoption. 43
A Look to the Future
The rapid advancements in large language models and large multimodal models, like GPT-4, 44 PaLM, 45 and Gemini, 46 have significant implications for health care. These models can effectively encode clinical knowledge and perform impressively in medical question-answering benchmarks, even for complex cases requiring clinical reasoning and multimodal understanding. 47
However, before these models are implemented in real-world scenarios, it is crucial to address the regulatory, clinical, and equity-related risks they pose. Today, generative AI in health care is being used primarily for nondiagnostic tasks that have a lower risk profile but can still significantly improve care providers’ efficiency by alleviating administrative burdens. Nevertheless, the capabilities of large multimodal language models are ushering in a new era of possibilities for health and medicine. 47 These advancements hold the potential to accelerate biomedical discoveries, assist in health care delivery, and enhance patient experiences.
In the future, AI platforms will continue to improve, becoming increasingly time-saving technologies. By automating routine tasks, AI can free up valuable time for physicians to focus on more complex aspects of patient care. In the context of CRC screening, AI platforms have demonstrated their ability to improve the efficiency and accuracy of polyp detection, thereby enhancing the overall quality of colonoscopy procedures. However, the benefits of AI extend beyond mere diagnostic accuracy. By automating routine tasks, AI allows physicians to devote more time to patient interaction, thereby fostering a more holistic approach to patient care. This shift in focus from keyboard to the patient is not merely a return to the traditional doctor–patient relationship but a significant leap forward in patient-centered care.
However, it is essential to temper optimism with realism. AI is not a panacea but rather a tool that requires careful consideration of various factors before successful integration into the health care workflow. These factors include physician acceptance, clarity of AI predictions, regulatory landscape, and more.
Conclusions
The integration of AI into CRC screening has been transformative, driving significant advancements in medical diagnostics and patient care. This journey highlights the importance of transparency, accountability, and strong communication between developers and health care professionals to build trust and facilitate AI adoption.
As we navigate this technological landscape, a crucial question arises: How do we allocate tasks between humans and machines? The rapid growth of tasks that technology can perform has led to an ever-diminishing number of tasks reserved for human expertise. This debate is not new; researchers have long disagreed on the purpose of AI—should it replace or augment human performance? However, as AI continues to evolve at a breakneck pace, it appears that the decision will ultimately be driven by technological development rather than specific ethical considerations.
To ensure AI remains a safe and effective tool, adhering to stringent regulatory and ethical standards is crucial. As we look forward, thoughtful integration of AI into health care holds great promise—better clinical outcomes, more efficient processes, and a new era of medical technology are all within reach.
Author Disclosure Statement
A.C. is affiliated with Cosmo Intelligent Medical Devices, the developer of the GI Genius medical device.
