Abstract
Background:
Shared decision-making (SDM) in urology faces challenges including limited health literacy, language barriers, and time constraints that can compromise informed consent and treatment adherence. Generative artificial intelligence (GAI), particularly large language models, offers opportunities to personalise patient education and enhance SDM.
Objective:
To evaluate the role of GAI applications in SDM for patients with urological conditions.
Eligibility criteria:
Peer-reviewed observational studies, validation studies, or mixed-methods studies evaluating GAI (e.g., large language models, AI chatbots) in patient communication, education, counselling, or SDM for urological conditions were included. Editorials, opinion pieces, conference abstracts, and non-English language publications were excluded.
Source of evidence:
PubMed, Embase, Cochrane Library, and Web of Science databases were comprehensively searched through June 2025. Study assessments: Newcastle-Ottawa Scale, the STROBE or the AGREE II as per study type.
Charting methods:
Charting methods was performed by using a standardised form. Outcomes of interest included accuracy of GAI-generated information, patient understanding, satisfaction, and decisional conflict.
Results:
Following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews guidelines, 18 observational studies (2023–2025) were included, comprising 310 patients in real-world settings plus hundreds of simulated queries across diverse urological conditions. GAI demonstrated moderate to high accuracy (52%–95%) for guideline-based information, with optimal performance in disease-specific patient education. A prospective comparative study showed 27% reduction in consultation time and improved patient understanding with ChatGPT-4 assistance. Limitations emerged including poor performance in emergencies and complex oncological counselling, and readability issues with content written at a college level (mean Flesch-Kincaid Grade Level 13.5). Most studies evaluated ChatGPT versions, limiting generalizability.
Conclusions:
GAI could enhance and potentially transform SDM in urology with appropriate clinical oversight and human-in-the-loop governance. Currently, GAI is useful for consultation preparation and patient education, while maintaining physician expertise for complex scenarios. Future implementation should prioritise patient safety, equitable access, and environmental sustainability while developing speciality-specific models and clinician education programmes.
Plain language summary
We reviewed 18 studies examining how generative artificial intelligence (GAI) can help doctors explaining urological conditions and treatment options to patients. We found that these GAI tools provide accurate information for common pathologies and can help in saving consultation time. However, GAI should not be used alone for complex scenarios, where urologists’ expertise remains essential for patient safety.
Keywords
Introduction
The landscape of patient-physician communication in urology has evolved dramatically over the past decade, driven by advances in digital health technologies and changing patient expectations for personalised healthcare. Shared decision-making (SDM), characterised by a collaborative diagnostic and therapeutic planning between patients and physicians, has emerged as a milestone of patient-centred urological care.1,2 SDM is fundamental, particularly where treatment pathways significantly affect quality of life, such as in prostate cancer, kidney transplantation, or sexual health. However, traditional counselling approaches often fail to adequately address individual patients’ needs, and other factors like limited patients’ health literacy, language differences, and time constraints can hinder patient understanding, potentially compromising the quality of informed consent and treatment adherence.3–11
In this scenario, generative artificial intelligence (GAI), especially large language models (LLMs) such as ChatGPT (OpenAI, San Francisco, CA, USA), Claude (Anthropic, San Francisco, CA, USA) or Gemini (Google DeepMind, London, UK), has demonstrated remarkable capabilities in language processing and information synthesis. These technologies offer unprecedented opportunities to personalise patient education materials, providing 24/7 access to medical information, and enhancing the efficiency of patient–physician interactions.12,13
Previous studies have demonstrated the effectiveness of the integration of AI into urological practice through multimedia presentations, 3D anatomical models, and virtual reality platforms in improving patients’ understanding and satisfaction.14–16 However, these approaches often lack the personalisation and adaptability that GAI can provide. Additionally, significant questions remain regarding accuracy, reliability, and clinical integration and the real-world evidence for LLMs’ role in supporting SDM in urology is limited and often speculative, with studies yielding to mixed results.17–19
This scoping review aims to evaluate the feasibility and safety of using GAI applications as a supportive tool in facilitating shared decision-making for patients with urological conditions, to evaluate their effectiveness compared to traditional approaches, and provide guidance for future research and clinical implementation with emphasis on human–AI collaboration and appropriate governance framework.
Evidence acquisition
Search strategy
This scoping review was conducted and reported in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) guidelines (Supplemental Table 1 shows the complete checklist). 20 A comprehensive search was conducted on PubMed, Embase, Cochrane Library, and Web of Science databases up to June 2025. The search strategy combined terms related to: artificial intelligence and machine learning; patient counselling and education; shared decision-making; and urology or urological procedures. Boolean operators (AND, OR) were used to refine the search. The complete search string is provided in Supplemental Material 1. No language or date restrictions were applied during the initial search. Additional studies were identified through manual review of reference lists and citation tracking from published reviews.
Study selection
Inclusion criteria encompassed the following criteria; (1) Types of participants: patients receiving urological care, healthcare providers in urology, or simulated patient scenarios involving urological conditions; (2) Concept: use of GAI (e.g., LLMs, AI chatbots) in patient communication, education, counselling, or SDM; (3) Comparator: standard care, other digital tools, or none; (4) Context: accuracy of information, patient understanding, satisfaction, decisional conflict, usability, accessibility; (5) Type of evidence sources: observational studies (cohort, cross-sectional, case-control), validation studies, or mixed-methods studies. Exclusion criteria comprised (1) Non-generative AI interventions (e.g., predictive algorithms only); (2) Studies in non-urological settings; (3) Editorials, opinion pieces, letters, conference abstracts without full data; (4) Non-peer-reviewed studies; (5) Non-English language publications. The primary objective was to assess whether a generative AI tool improves patients’ understanding of treatment options and increases engagement in SDM. Secondary objectives included: to compare AI-assisted consultations with standard of care in terms of patient satisfaction, decisional conflict, and communication quality; to evaluate the readability, cultural sensitivity, and accuracy of LLM-generated outputs across literacy levels and languages; to identify patients’ perceptions, trust, and ability of experiences with GAI in the SDM process.
Charting methods, quality and risk of bias assessment
Title and abstract screening were performed independently by two reviewers (CC and FR). Full-text articles were subsequently assessed for eligibility. Charting methods were performed independently by using a standardised form that included: study characteristics (design, sample size, setting, country), population demographics, GAI intervention details (platform, version, prompting strategy), outcome measures, and study quality indicators. Discrepancies were resolved through consultation with a third researcher (BKS).
Study quality and risk of bias were assessed using the Newcastle-Ottawa Scale (NOS) for observational and cross-sectional studies, 21 the STROBE for observational reporting quality, 22 or the AGREE II for guidelines adherence evaluations. 23 Risk of bias was assessed independently by two reviewers (CC and FR), with disagreements resolved by consensus or consultation with a third reviewer (BKS). Studies were categorised as high quality (green), medium quality (yellow), or low quality (red; Supplemental Table 2). As per risk of bias, one study showed low risk of bias, demonstrating only minor concerns and strong methodology; medium risk studies had limitations in two or more domains that might affect the reliability of their conclusions, while studies with high bias for comparability and/or another domain were deemed at high risk of bias (Supplemental Table 3). Consistent with the scoping review methodology, meta-analysis was not performed given the heterogeneity in study design, interventions, comparators, and outcome measures across the included studies, which is expected in an emerging field.
Evidence synthesis
Study characteristics
Figure 1 presents the PRISMA-ScR flow diagram detailing the study selection process. The initial search identified 1426 articles. Following duplicates removal (n = 23), title and abstract screening excluded 1370 studies. Full-text assessment of the 33 studies led to exclusion of 15 papers focused on traditional technology studies and clinical algorithm studies, yielding 18 papers for inclusion in the final review.17–19,24–38

The PRISMA-ScR flow chart for the inclusion of studies evaluating generative AI in urological patient counselling and shared decision-making.
All the included studies were published from 2023 onwards, reflecting the recent emergence and increasing interest of GAI in clinical practice. Table 1 shows a summary of the included studies. Studies originated from multiple countries, including the United States (n = 5), Europe (n = 3), China (n = 1), Canada (n = 1), Turkey (n = 3), Australia (n = 1), and multinational collaborations (n = 4). The total patient population directly evaluated across all the studies included 310 patients from real-world clinical settings, with additional evaluation of hundreds of simulated patients queries and clinical scenarios.
Characteristics of the GAI studies included in the review.
GAI, Generative artificial intelligence; LLM, large language model.
The GAI platforms evaluated were predominantly ChatGPT versions 3.4 and 4.0, with additional assessment of Perplexity, Microsoft Bing AI, ChatSonic, and NeevaAI. Studies focused on diverse urological conditions, including prostate cancer (n = 5), bladder cancer (n = 2), urolithiasis (n = 2), general urological malignancies (n = 2), paediatric urology (n = 1), vasectomy counselling (n = 1), and mixed urological conditions (n = 5). Most studies employing validated assessment tools, including DISCERN scores, Brief DISCERN questionnaires, PEMAT-P tools, and readability formulae.18,19,31,35
Generative AI application in urological patient counselling
GAI has been utilised across different SDM settings, from initial patient education to supporting complex treatment decisions. With respect to information exchange (the first element of SDM), studies identified three primary applications. First, disease-specific patient education showed moderate to high performance with accuracy rates ranging from 52% to 95% depending on condition complexity. Gabriel et al. found 92.9% global accuracy for robotic prostatectomy patient education, 31 while another study reported 94.6% accuracy for urolithiasis FAQs. 32 Second, treatment counselling applications demonstrated variable success, with better performance for structured procedural information than complex medical decision-making.34,35 Third, one study showed a good generation of accessible medical literature summaries for lay audiences. 30
The most comprehensive real-world GAI utilisation was described in a prospective study of 292 patients interacting directly with a GPT-4-powered chatbot in clinical settings, where patients valued AI as a powerful complement to human expertise, particularly appreciating the additional consultation time it provided. 24 Notably, age was not associated with confidence in AI tools, suggesting thus universal accessibility across patient demographics, which is important for equitable SDM implementation.
Emergency urology scenarios revealed critical limitations for SDM applications. Davis et al. found very low appropriate responses for emergency conditions, 18 highlighting current GAI unsuitability for time-sensitive clinical decisions requiring immediate intervention, where rapid accurate SDM is essential.
Patient outcomes and satisfaction
Direct patient interaction studies revealed nuanced perspective on GAI integration in SDM process. The prospective clinical trial by Rodler et al. of 466 patients found a good patients’ comfort with the technology, with AI trust correlating weakly but significantly with overall technology acceptance (r = 0.094, p = 0.04). Traditional factors like age, education, and illness perception showed no significant relationship with AI trust. However, patients demonstrated significantly higher trust in physicians compared to AI when it came to individualised diagnosis communication (a key SDM component) and understandable explanation. Critically for SDM governance, patients’ trust in AI-generated diagnoses more than doubled when physician-controlled (4.31 ± 0.88) versus unsupervised AI (1.75 ± 0.93), underscoring the importance of human-in-the loop approaches. 27 The single randomised prospective comparative trial by Chung et al. 38 examining pre-vasectomy counselling demonstrated measurable clinical benefits for SDM efficiency. The intervention group receiving ChatGPT-4 assistance showed significantly higher provider perception of patient understanding (8.8 ± 1.0 vs 6.7 ± 2.8, p = 0.04) and 27% reduction in consultation time (7.7 ± 2.3 vs 10.6 ± 3.4 min, p = 0.05). Patient satisfaction scores for AI-assisted counselling were high (quality of information: 8.3 ± 1.9; ease of use: 9.1 ± 1.5 on a 10-point scale), though privacy concerns scored 4.9 ± 2.9, indicating moderate apprehension requiring attention in SDM implementation.
Readability emerged as a possible barrier to effective SDM. Even though Eppler et al. demonstrated ChatGPT’s ability to create accurate summaries of urological research for general population without medical background when properly prompted, 30 several studies found AI-generated content exceeded the recommended sixth grade reading levels, with responses usually written at college level (mean Flesch-Kincaid Grade Level 13.5 ± 1.72). 18 Musheyev et al. evaluated quality, understandability, actionability, misinformation and readability across multiple platforms like ChatGPT, Perplexity, Chat Sonic, and Microsoft Bing AI. While AI chatbots responses had moderate to high information quality with minimal misinformation, understandability was moderate (median patient Education Material Assessment Tool for Printable Materials (PEMAT-P) with understandability 66.7% – range 44.4%–90.9%) and actionability was moderate to poor with responses being written at a difficult reading level to be actioned. 33 These findings suggest that literacy level, rather than age, may potentially limit GAI utility and access of diverse populations in SDM contexts.
Clinician perspective and AI governance
Healthcare professionals evaluating GAI responses demonstrated cautious optimism and some concerns. A blinded assessment by Ayers et al. found that healthcare professionals preferred chatbot-generated responses over physician responses for quality and empathy, 26 suggesting potential for AI to enhance patient communication when properly implemented. Minolitsis et al. demonstrated that training with specialty-specific guidelines provided more precise and prompt answers, highlighting the value of domain-specific training and effective prompt engineering. 28 Szczesniewski et al. queried ChatGPT 4.0 on five common urological conditions (prostate cancer, bladder cancer, renal cancer, BPH and urolithiasis). Two independent urologists deemed the overall information well-balanced including anatomical location, affected population, and symptom descriptions, but assigned moderate quality scores (DISCERN 3/5) for treatment-related answers, which are central to SDM. 35 Caglar et al. questioned ChatGPT about the 137 most asked inquiries related to paediatric urology; 92% of the answers were judged completely correct, with a concordance of 93.6% with the EAU Guidelines strong recommendations. 36
Clinicians consistently emphasised the need for professional oversight and human-in-the-loop governance, with multiple studies recommending AI use only under direct supervision. Whiles et al. found concerning variability, with 25% sets showing discordant appropriateness between repeated prompts, 19 raising reliability concerns for SDM application where consistency is paramount. The consensus emerging from these studies positions GAI as a drafting and preparatory tool for physicians rather than an autonomous patient interaction system, with clinicians serving as essential intermediaries who contextualise and validate AI-generated information.
Accuracy and safety considerations
Accuracy varied substantially by clinical context and question complexity. Overall appropriateness ranged from 52% to 95%, with structured, guideline-based questions achieving higher accuracy than open-ended clinical scenarios.18,19,31,32 Prostate cancer information showed particularly diverse results concerning for SDM reliability. Coskun et al. reported F1 scores of only 0.426 (range 0–1) and precision of 0.349 (range 0–1), 29 indicating substantial inaccuracy risk for complex oncological counselling. Zhu et al. found above 90% accuracy for basic information questions with definite answers for most of the LLMs’ analysed responses; however, accuracy decreased for questions requiring synthesis and analysis (e.g. Why the PSA is still high after surgery?) and responses did not exhibit humanistic care elements essential for SDM. 34 Guo et al. identified concerns regarding bladder cancer information completeness, particularly for tumour cure rate according to stage and the treatment-related side effects related, reinforcing the importance of urologist involvement in SDM. 25 Safety assessments revealed multiple concerns relevant to patient counselling, Lack of clear information sources emerged as a universal issue, with several platforms failing to disclose references, preventing verification of medical claims critical for informed decision-making.19,35 Thalyshinskii et al. found incorrect data in 8 of 11 clinical answers despite all containing some EAU guideline-corresponding information. 37 Emergency scenario performance was particularly concerning, with one study reporting of missed diagnoses of acute retention in classic presentation, representing potential patient harm if left without clinical oversight, 18 and with another cross-sectional study scoring only 11.1% appropriateness for emergency clinical scenarios. 17
Discussion
This scoping review provides evidence that GAI holds potential as supportive tool for urological counselling and SDM, though with important limitations regarding implementation (Figure 2). The evidence base reveals a technology positioned between promising potential and significant limitations that must be addressed through appropriate governance frameworks.

Summary of benefits and limitations of generative AI in urological shared decision-making.
A critical finding across studies is the necessity for human-in-the-loop governance GAI deployment for SDM. 39 This framework positions clinicians as essential intermediaries who contextualise, validate and communicate AI-generated information. The substantial increase in patient trust when physicians control AI input 15 provides empirical support for this governance model. Implementation requires establishing clear protocols identifying when AI assistance is appropriate versus when direct physician expertise is essential, particularly for complex oncological decisions and emergency scenarios where GAI performance is inadequate.7,8,13
Equally important is the need for clinician training in context prompt engineering. The finding that speciality-specific training improved GAI response quality 16 highlights how carefully constructed prompts can enhance AI utility. In SDM contexts, effective prompts must incorporate patient-specific factors including medical history, treatment preferences, cultural background, and literacy level. This represents a new competency for urologists that training programmes like the EAU Talent Incubator Programme are beginning to address, where urologists are trained as ‘knowledge brokers’ capable of effectively using AI tools while maintaining clinical judgement. 40 The EAU has recognised GAI’s potential through initiatives including the EAU Patient chatBOT integrated with ChatGPT capabilities for patients and family education.41,42 Importantly, all the information that is released through the EAU Patient Offices website is reviewed by experts in the field, either guideline panellists or members of the Young Academic Urology.
This review identifies several evidence-supported integration pathways for GAI in urological SDM (Figure 3). Pre-consultation preparation emerged as an optimal, with patients using AI to formulate questions and understand basic concepts before appointments.12,26 This approach supports rather than replaces the physician’s role in SDM while improving consultation efficiency and quality. The 27% reduction in consultation time demonstrated in the vasectomy study suggests significant workflow optimisation potential 38 when AI handles routine information provision, freeing clinicians for complex decision-making and relationship building. Post-consultation reinforcement represents another valuable application, with AI helping patients recall and understand discussed information, addressing the documented problem of patients retaining only 40%–80% of verbally provided medical information. 43

Implementation, challenges, integration and future directions.
Conversely, GAI is currently unsuitable for emergency scenarios, complex oncological counselling requiring nuanced risk-benefit discussions, and situations requiring understanding of individual patient values and preferences. In these contexts, direct physician expertise remains essential for safe and effective SDM.
The development of specialty-specific AI models trained on urological guidelines and literature shows promise for improving accuracy and clinical relevance. 28 GAI might help in workflow redesign opportunities, including AI-assisted documentation where chatbots draft consultation summaries for clinician review reducing administrative burden, structured patient history taking using AI gathering past medical histories before appointments improving consultation efficiency, multi-language and multi-cultural supports through translation and language adaptation improving care accessibility for diverse populations and level of educations. Integration with electronic health records could enable personalised AI responses based on individual patient histories and treatment plans, though privacy considerations require careful attention. GAI development should address multi-language and multi-cultural support, as GAI’s translation and adaptation capabilities could improve SDM accessibility for diverse populations. 22 Integration with electronic health records could enable personalised AI responses based on individual patient histories, though privacy considerations require careful attention.
Based on our evidence synthesis, therefore we would recommend GAI implementation for pre-consultation patient education with appropriate content at accessible reading levels; mandatory clinician oversight for all patient-facing GAI applications; explicit warnings against GAI use for emergency scenarios; institutional development of prompt engineering guidelines; regular accuracy auditing against current guidelines; and patient education about limitations and continued primary physician expertise in SDM.
Future research should prioritise head-to-head comparative studies evaluating traditional counselling and informed consent against GAI-assisted approaches across diverse urological conditions, incorporating validated patient-reported outcome measures such as decisional conflict scales and knowledge assessments to establish clinical effectiveness and cost-benefit profiles. Long-term studies examining patient outcomes, satisfaction, and potential harms from AI-assisted counselling are needed to complement the short-term efficacy data currently available. Research into optimal human-AI collaboration models should identify which tasks benefit from AI assistance versus human expertise. Health equity research must determine whether GAI narrows or widens existing disparities in access to medical information, particularly across populations with varying digital literacy levels. 34 Finally, the environmental impact of GAI deployment represents an increasingly urgent consideration, with the computational demands of LLMs contributing substantial CO2 emissions44,45; healthcare systems must balance AI’s clinical benefits against environmental costs, potentially restricting use to high-value applications. Given the rapid pace of technological evolution, living review methodologies and continuous outcome monitoring will be essential to keep evidence aligned with current model capabilities.
Limitations
This scoping review comes with and highlights some limitations that should be considered when interpreting findings. First, restriction to English language publications may have excluded relevant studies from non-English speaking regions where GAI implantation may differ. Second, exclusion of grey literature may have introduced publication bias toward positive results. Third, most included studies evaluate ChatGPT 3.5 and 4.0, limiting generalisability to other platforms or future interactions. Fourth, the rapid evolution of AI technology means published studies may not reflect current capabilities, with newer models potentially addressing identified limitation.13,46 Fifth, most studies employed simulated queries rather than patient interactions, potentially missing important contextual factors affecting real-world performance. 26 Sixth, while heterogeneity in study design, interventions, and outcomes is expected and appropriate within a scoping review framework, it precluded quantitative synthesis. Finally, the limited number of prospective studies evaluating actual SDM process rather than isolated GAI capabilities constrains conclusions about real-world effectiveness.
Conclusion
This scoping review provides comprehensive evidence that GAI holds transformative potential for urological counselling and SDM, with greatest utility as complementary tool enhancing rather than replacing human clinical expertise. GAI demonstrates strength in structured patient education and information exchange but shows limitations for complex deliberation and emergency scenarios. Effective implementation requires human-in-the-loop governance, clinician training in prompt engineering, and careful attention to readability and accessibility. Future directions should prioritise patients’ safety and health equity while addressing environmental sustainability concerns. Our evidence highlights the need for continued research to optimise human–AI collaboration in patient-centred care.
Supplemental Material
sj-docx-1-tau-10.1177_17562872261441968 – Supplemental material for Generative AI in urology: rethinking patient counselling and shared decision-making – a scoping review from the European Association of Urology Patient Office
Supplemental material, sj-docx-1-tau-10.1177_17562872261441968 for Generative AI in urology: rethinking patient counselling and shared decision-making – a scoping review from the European Association of Urology Patient Office by Clara Cerrato, Francesco Ripa, Michael R van Balken, Eamonn T. Rogers and Bhaskar K. Somani in Therapeutic Advances in Urology
Footnotes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
