Abstract
Objective
This systematic review evaluates the performance and limitations of AI-based models for Degenerative cervical diseases (DCD) diagnosis using MRI.
Methods
A comprehensive literature search was conducted in three databases—PubMed, Embase, and Web of Science—covering studies published between January 2010 and March 2024. Studies were included if they employed AI techniques for the diagnosis or prognosis of DCD using MRI. Key performance metrics, methodological details, and limitations were extracted and analyzed.
Results
Eleven studies met the inclusion criteria, with AI models showing high diagnostic performance. Accuracy ranged from 81.58% to 98%, sensitivities from 84% to 98%, specificities from 90% to 100%, and AUC values reached up to 0.97. Convolutional neural networks (CNN) were the most frequently used models (four studies), followed by support vector machines (three studies). Comparative analysis revealed that CNN-based approaches showed consistently high performance in ossification of the posterior longitudinal ligament detection, while traditional machine learning methods demonstrated varying effectiveness in cervical spondylotic myelopathy classification. Sample sizes varied significantly, ranging from 28 to 900 patients. MRI protocols also differed across studies, with variations in field strengths, slice thicknesses, and sequences used. Seven studies assessed inter-rater reliability. Most studies lacked external validation, which raises concerns about the generalizability of the models. Additionally, hardware configurations were inconsistently reported, and data augmentation techniques were underutilized, limiting the robustness of the models in smaller datasets.
Conclusion
While AI models for DCD diagnosis using MRI show high diagnostic potential, methodological weaknesses such as insufficient external validation and small sample sizes hinder broader clinical adoption. Future research should focus on larger, standardized, multi-center studies to improve the robustness and clinical relevance of AI-driven tools for DCD diagnosis.
Keywords
Introduction
Degenerative cervical diseases (DCD), including cervical spondylotic myelopathy (CSM), ossification of the posterior longitudinal ligament (OPLL), and cervical spinal stenosis, are common age-related disorders that lead to considerable neurological disabilities and a decline in health-related quality of life. 1 Magnetic resonance imaging (MRI) is the primary imaging modality for diagnosing and assessing the severity of DCD due to its superior ability to provide soft tissue contrast and detailed visualization of spinal structures. 2 However, traditional MRI interpretation in DCD diagnosis faces several significant challenges. Inter-observer variability is a major concern, as different radiologists may interpret the same MRI images differently, leading to inconsistent diagnoses. 3 The complexity of multi-level pathologies in DCD often makes it difficult to accurately assess the overall severity and pinpoint the primary problematic areas. 4 Furthermore, subtle changes in early-stage DCD can be easily overlooked, potentially delaying crucial early interventions. 5 The detailed examination of multi-sequence MRI scans for DCD is also time-consuming, which can delay diagnosis and treatment. 6 These challenges, combined with the difficulty in early-stage diagnosis, can lead to delayed or inaccurate diagnoses, potentially compromising patient care and outcomes.
Artificial intelligence (AI) models are designed to address these issues by providing quantitative and objective assessments, thereby enhancing diagnostic consistency. 7 AI, particularly deep learning and machine learning, has developed very fast in recent years and has been in the vanguard of enhancement in the accuracy and effectiveness of medical imaging-based diagnostics. 8 AI-based techniques are highly successful in various medical image analyses, such as segmentation, classification, and the detection of pathological features from multiple imaging modalities.9,10 Recently, researchers have focused on developing AI models that can predict the diagnosis or severity of DCD based on MRI analysis of the cervical spine. 11 The present research was focused on the performance of AI-driven models in the workflow of the radiologist and clinician, facilitating timely and accurate diagnoses of DCD and enabling early intervention through personalized treatment strategies. 12 Additionally, they provide objective and quantitative descriptions of the disease and its stages, complementing the subjective and qualitative assessments of human experts for improved readability.13,14
Given the increasing focus on AI applications in DCD diagnosis and prognosis, it is essential that these models are both explainable and interpretable so that human experts can understand and trust the decision-making processes of these models. This interpretability is crucial for fostering effective collaboration between AI systems and healthcare professionals in patient care. 15 Despite this growing interest, there remains a significant lack of comprehensive evaluations of existing AI models in this specific domain, which limits the identification of effective models and the understanding of challenges that must be addressed for clinical translation. To address this, our systematic review aims to critically evaluate the performance, methodological quality, and limitations of current AI-based diagnostic models for DCD using MRI data. Using the PROBAST tool, 16 it will assess specific diseases, modeling techniques, and performance metrics while evaluating the methodological quality and risk of bias (RoB) of the included studies. By synthesizing available evidence, this review seeks to fill a crucial gap in the literature, providing clinicians and researchers with valuable insights that will advance the development of clinically applicable AI tools, improve patient outcomes, and inform future research directions in enhancing diagnostic accuracy and patient care in DCD.
Materials and methods
Literature search
This systematic review did not involve direct patient participation or intervention. All data were extracted from previously published studies that had obtained appropriate ethical approvals. Therefore, ethical approval and patient consent were not required for this review. The scope of this review was confined to studies focusing on the diagnostic performance of AI models in DCD using MRI. The review was conducted according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. 17 Studies were included based on the following criteria: Published in English between 1 January 2010, and 30 March 2024. Employed AI techniques for the assisted diagnosis or prognosis of DCD using MRI. AI techniques considered included, but were not limited to, machine learning algorithms (e.g., support vector machines, random forests), deep learning models (e.g., convolutional neural networks, recurrent neural networks), and other computational intelligence methods. Utilized MRI data for analysis. Acceptable MRI parameters included various sequences (e.g., T1-weighted, T2-weighted, STIR) and field strengths (e.g., 1.5T, 3T), without restrictions on specific imaging protocols. Focused on DCD, including but not limited to CSM, OPLL, and cervical disc degeneration. Reported quantitative performance metrics for the AI models (e.g., accuracy, sensitivity, specificity, AUC).
To identify relevant studies, we conducted comprehensive searches of Web of Science, PubMed, and Embase databases, focusing on four groups of search terms: (a) AI techniques (e.g., artificial intelligence, deep learning), (b) diagnosis-related terms (e.g., segmentation, prediction), (c) MRI terminology (e.g., T1-weighted, diffusion-weighted imaging), and (d) degenerative cervical disease terms (e.g., cervical spondylosis, ossification of the posterior longitudinal ligament). These groups ensured a broad capture of literature addressing AI applications in MRI-based DCD diagnosis. The complete search query used in the different databases is shown in Supplementary File 1.
After exclusion of duplicates, the titles and abstracts were screened, and only relevant publications proceeded to full-text screening. The decision as to whether a study met the inclusion criteria of the review was performed by two authors (D.Q. and S.X.X.) without the use of automated tools. A third author (C.G.R.) acted as a referee in case of a potential disagreement between the two authors responsible for screening. All articles that did not focus on the use of AI techniques for aided diagnosis in patients with DCD were excluded at the full-text screening stage.
Data extraction
Two authors (D.Q. and S.X.X.) independently extracted data and discussed any discrepancies. Data were extracted with regard to:
Study and clinical parameters: authors, title, year, design, number of patients in training/test set, development/Test Split, ground truth, inter-/intra-rater variability, task, specific diseases, conflict of interest, sources of funding. Imaging parameters: MRI machine, number of images, field strength, slice thickness, sequences. AI parameters: algorithm, dimensionality, training duration and hardware, libraries/frameworks/packages, data augmentation, performance measures, explainability/interpretability features, code/data availability.
Quality assessment
PROBAST is a comprehensive tool designed to assess the RoB and applicability concerns in diagnostic model studies. 16 It is structured around four main domains: Participants: Evaluates the selection of study participants; Predictors: Assesses how predictors are managed and measured; Outcome: Looks at how outcomes are defined and determined; Analysis: Reviews the statistical analysis approaches used. Each domain is evaluated through several signaling questions, which guide the user in identifying potential issues. The RoB for each article is classified as Low RoB, High RoB, or Unclear RoB. The assessment is based on the following criteria:
Low Rob: If no relevant shortcomings were identified in the RoB assessment—that is, all domains had low RoB.
High RoB: If a model is developed without any external validation on different participants, downgrading to high RoB should still be considered even if all four domains had low RoB, unless the model development was based on a very large data set or included some form of internal validation.
Unclear RoB: If unclear RoB was noted in at least 1 domain and all other domains had low RoB.
Two authors (D.Q. and S.X.X.) independently assessed the RoB concerns of the included studies using the PROBAST tool. Their independent results were then compared, and any discrepancies were resolved through discussion. A third author (C.G.R.) acted as a referee in case of a potential disagreement between the two authors responsible for the quality assessment. The quality assessment process was conducted without the use of automated tools to ensure a thorough and comprehensive evaluation of each included study.
Results
The inclusion workflow is illustrated in Figure 1. The query yielded 69, 37, and 70 publications from the Web of Science, PubMed, and Embase databases, respectively. After removing 63 duplicate records, 113 publications remained. During the screening of records, 102 articles were excluded. A comprehensive list of the excluded articles and their respective reasons for exclusion is provided in Supplementary File 2. Specifically, three and eight articles were excluded as they were literature reviews and conference reviews, respectively.18–28 Additionally, 85 publications were excluded due to their irrelevance to the topic,29–113 while two articles were excluded because they were retracted.114,115 Furthermore, two animal studies were also excluded, along with one study investigating multiple anatomical sites and 1 study involving cadavers.116–119 After exclusion, 11 articles are remained and the key data are shown in Table 1.120–130 Detailed extracted data from the 11 included articles can be found in Supplementary File 3a, b and c.

Inclusion workflow diagram according to PRISMA 2020. 17
Key characteristics of the selected research articles.
Abbreviations: OPLL, Ossification of the posterior longitudinal ligament; CSM, Cervical spondylotic myelopathy; CI:, Confidence interval; DTI, Diffusion Tensor Imaging; LR, Logistic regression; SVMs, Support Vector Machines; DT, Decision Tree; RF, Random Forest; AUC, Area under the receiver operating characteristic curve; STMs, Support Tensor Machine; NB, Naive Bayes; CNNs, Convolutional Neural Networks; NPV, Negative Predictive Value; PPV, Positive Predictive Value; CycleGAN, Cycle Generative Adversarial Network; DNN, Deep Neural Network.
Study characteristics
Among the 11 included studies, 10 were retrospective, while one was a post hoc pilot study. 120 Sample sizes varied widely, from 28 patients to 900 patients.120,127 Eight studies were published after 2020, indicating a recent surge in research interest. The studies covered various DCDs, including CSM,120,122,126,128,129 OPLL,124,127 spinal canal stenosis,121,123 and cervical disc degeneration.125,130
Diagnostic performance of AI models
AI models demonstrated high diagnostic performance across studies, although the performance metrics varied (Table 1). Six studies reported accuracy,120,122,124,126,128–130 with values ranging from 81.58% 120 to 98%. 127 Area under the receiver operating characteristic curve (AUC) values were reported in five studies,120,122,124,129,130 ranging from 0.85 to 0.971. Sensitivity and specificity were reported in seven studies,122–124,126–129 with sensitivity ranging from 84.62% to 98.83%, and specificity from 90% to 100%. Other metrics, such as precision, recall, and F1-score, were only reported in three studies.122,125,130
CSM
Five studies focused on CSM.120,122,126,128,129 Accuracy was reported in four studies, ranging from 80% to 95.73%. AUC values were reported in three studies, with the highest being 0.947. 120 Sensitivity and specificity varied across studies, with the highest sensitivity being 93.41%, and specificity reaching 98.64%. 126 Only one study assessed inter-rater reliability. 129
OPLL
Two studies focused on OPLL detection.124,127 Qu et al. 124 reported an accuracy of 97.66%, while Shemesh et al. 127 reported 98%. AUC was only reported by Qu et al. 124 (0.971). Sensitivity and specificity ranged from 85% 127 to 100%.120,124 Both the studies reported inter-rater reliability.
Cervical disc degeneration
Two studies addressed cervical disc degeneration.125,130 Xie et al. 130 reported an accuracy of 89.51% and an AUC of 0.95. Niemeyer et al. 125 did not report accuracy or AUC but provided Cohen's kappa for classification tasks, achieving 0.722. Neither study provided sensitivity or specificity, and only Niemeyer et al. assessed inter-rater reliability.
Spinal canal stenosis
Two studies focused on spinal canal stenosis.121,123 Neither reported accuracy or AUC values. Both assessed inter-rater reliability, with Jardon et al. 121 reporting an increase in Cohen's kappa from 0.76 to 0.81, and Kim et al. 123 showing improvements in weighted kappa values from 0.48–0.71 to 0.61–0.75.
Comparative analysis of AI models across different DCD types
A comparative analysis of different AI models across DCD types is shown in Figure 2, with detailed performance metrics presented in Table 1. Among the CSM studies, Wang et al. 126 and Wang et al. 129 conducted classification tasks using different machine learning approaches. The traditional machine learning combination (SVM/STM/NB) 126 achieved higher accuracy, sensitivity, and specificity compared to the newer ensemble method, 129 which reported accuracy and AUC values. In OPLL detection studies, Qu et al. 124 systematically compared different CNN architectures, showing that deeper networks (ResNet-101) achieved better performance in all metrics (accuracy, sensitivity, specificity, and AUC) than shallower ones (ResNet-34). Additionally, Shemesh et al. 127 employed VGG16 network, achieving comparable high performance (accuracy: 98%, sensitivity: 85%, specificity: 98%). For cervical disc degeneration, two studies focused on different aspects with distinct evaluation metrics. Xie et al. 130 employed a combination of traditional machine learning methods (DT/RF/SVM/XGBoost) for general degeneration classification, reporting accuracy of 89.51% and AUC of 0.95. While Niemeyer et al. 125 used CNN for specific degenerative phenotype classification, evaluating performance with Cohen's kappa values for different phenotypes (ranging from 0.271 to 0.741). Unlike other DCD types, spinal canal stenosis studies by Jardon et al. 121 and Kim et al. 123 focused primarily on improving inter-rater reliability, using kappa and weighted kappa values respectively to assess their approaches.

Performance comparison of AI models across different DCD types. Performance metrics include accuracy, sensitivity, specificity, and area under the curve (AUC). In CSM classification, traditional machine learning methods were evaluated. OPLL detection employed both CNN-based approaches (ResNet-34, ResNet-101) and VGG16 network. For disc degeneration, two different approaches were used: combined machine learning methods for general classification and CNN for specific phenotype classification with different evaluation metrics (Cohen's kappa). For spinal canal stenosis studies, different evaluation metrics (kappa values) were used due to their focus on inter-rater reliability improvement.
Rob assessment
A wide range of AI methodologies was employed in the included studies, reflecting the exploratory nature of AI applications in diagnosing DCD (Table 1). Convolutional neural networks (CNNs) were the most commonly used method, appearing in four studies,122,124,125,127 followed by support vector machines (SVMs), which were utilized in three studies.126,128,129 Additionally, some studies employed a combination of multiple machine learning algorithms, such as SVM, STM, Naive Bayes, and XGBoost, indicating a trend toward more diverse approaches in three studies.126,128,129 Generative adversarial networks (GANs) were used in one study, 123 and support tensor machines (STMs) were implemented in two studies.126,128 Regarding hardware configurations, only four studies specified the computational resources used.121,123–125 According to the PROBAST tool, six studies were assessed as having a low RoB,122,124–126,129,130 while four studies had an unclear RoB,121,123,127,128 and one study was rated as having a high RoB (Figure 3). 120 Detailed assessments are provided in Supplementary File 4.

This chart provides a comprehensive summary of the risk of bias evaluations across the 11 studies. The left bar chart illustrates the overall risk of bias judgement and the right chart depicts the domain-specific risk of bias assessments.
Analysis of data heterogeneity across studies
Significant heterogeneity was observed in data acquisition and processing across studies (Supplementary File 3a, b, and c). MRI field strengths varied, with studies using either 1.5T,120,124,125 3.0T,121,122,126–130 or both.123,124 Slice thickness ranged from 0.7 121 to 7.0 mm,125,128 with most studies using 3.0–4.0mm. Imaging protocols differed, from single T2-weighted sequences121,122,125,127 to multiple sequences including T1-weighted, T2-weighted, and DTI.126,128
Sample sizes varied substantially (28 120 to 900 127 patients), as did the number of analyzed images (up to 9737 123 ). Most studies employed 2D analysis,120,122–124,126,128,130 while some utilized 3D121,129 or combined approaches. 125 Training and testing data distribution also showed considerable variation across studies, with different splitting ratios adopted for model development and validation.
Ground truth establishment
Ground truth labels were defined by experienced clinicians or radiologists in nine studies
Limitations identified
A key limitation observed was the small sample sizes, such as Hopkins et al. 120 with 28 patients, which may impact generalizability. Only one study 122 made code and data publicly available, limiting reproducibility. MRI protocol differences (machines, field strengths, slice thicknesses, and sequences) may have introduced variability, affecting image quality and model results. Most studies did not incorporate explainability features, and performance metric inconsistencies further hampered comparison across studies.
Summary of key findings
In summary, our review reveals that AI models demonstrated high diagnostic performance across studies, with accuracies ranging from 80% 120 to 98%, 127 and AUC values up to 0.971. 124 However, we also identified common limitations such as small sample sizes (ranging from 28 to 900 patients), lack of external validation in most studies, and inconsistent reporting of performance metrics. These findings highlight both the potential of AI in DCD diagnosis and the need for more robust, standardized research methodologies.
Discussion
Sample size and data splitting challenges
AI techniques for analyzing DCD using MRI data show significant promise, with sample sizes ranging from 28 to 900 patients. Data splitting approaches varied, including random splitting, cross-validation, and cases where no separate test set was used. Cross-validation offers a rigorous model performance estimate but is computationally expensive, particularly for larger datasets. Some studies applied stratified sampling to balance disease severity subtypes across training and test sets, though this approach was inconsistent across studies.131,132 Stratified sampling remains crucial for balancing datasets, ensuring robust model evaluation. 133
Ground truth labeling and standardization
Variability in ground truth labeling—from expert annotations to automated methods—introduces trust and consistency issues during model training. Disagreements among clinicians highlight the need for standardized protocols and guidelines to improve label consistency and model reliability in cervical spine MRI. 134 Multiple expert annotations, adjudication steps, and interrater reliability checks can help ensure dependable ground truth data. However, the lack of standardized evaluation protocols and benchmark datasets hinders objective comparisons. Publicly available benchmark datasets and shared tasks would foster a more competitive and replicable research landscape. Collaboration among clinicians, radiologists, and AI developers is essential to ensure clinical relevance, interpretability, and trust in AI systems.
AI techniques and comparative analysis
A variety of AI methodologies were used, with CNNs being the most common,122,124,125,127 followed by SVMs,126,128,129 GANs, 123 and decision trees. In CSM classification, traditional machine learning approaches showed varying performance levels,126,129 which might be attributed to differences in sample sizes and feature selection strategies. For OPLL detection, CNN-based approaches consistently demonstrated high performance, with both deeper architectures (ResNet-101) and VGG16 network showing superior results,124,127 suggesting that hierarchical feature learning is particularly suited for detecting complex anatomical changes. The diversity in approaches for cervical disc degeneration and spinal canal stenosis reflects the complexity of these conditions,121,123,125,130 highlighting the importance of matching model architecture to specific diagnostic requirements.
The generalizability of these findings is primarily limited by methodological considerations. Most studies focused on diagnostic accuracy without considering other important clinical factors such as inference time and model interpretability. Moreover, the lack of standardized evaluation metrics makes it challenging to conduct comprehensive model comparisons across studies.
Impact of data heterogeneity
Technical heterogeneity in data acquisition poses unique challenges for AI applications in DCD diagnosis. Specific variations in MRI parameters, particularly field strengths (1.5T vs. 3.0T) and slice thickness (ranging from 0.7 to 7.0mm), directly affect image quality and feature representation.131,132 This imaging protocol diversity necessitates careful consideration in model development and validation processes.
The diversity in study scale also significantly impacts model development. While some studies utilized large datasets with thousands of images,123,127 others were limited by smaller sample sizes. 120 These differences in data availability influence not only model training strategies but also the reliability of performance evaluations. Future multi-center studies with standardized imaging protocols are essential to establish more reliable benchmarks for model performance.133,134
Clinical relevance and implementation challenges
Our review reveals promising potential for AI applications in DCD diagnosis using MRI, with several studies reporting high diagnostic accuracies. There is a predominant focus on CSM,120,122,126,128,129 reflecting its clinical significance, while other conditions like OPLL also show promising results.124,127 Qu et al. 124 achieved 97.66% accuracy in OPLL detection, while Shemesh et al. 127 reported 98% accuracy. This focus on CSM, while important, also highlights a research gap for other DCD subtypes. AI models have demonstrated potential in diagnosis, offering objective and quantitative assessments. However, their current clinical relevance is limited by the variability in methodologies, which affects comparability between studies. To truly impact clinical practice, AI models must be further refined and standardized.
Methodological quality and external validation
The PROBAST RoB assessment provides valuable insights into the methodological quality and limitations of the included studies. It highlights common study design weaknesses and areas for improvement. Notably, some diagnosis models that were deemed low risk across all elements lacked external validation, underscoring the importance of externally validating models to ensure their performance and generalizability. This finding reinforces the need for future studies to follow established reporting guidelines, such as the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) statement, to enhance transparency and reproducibility while minimizing potential bias due to incomplete reporting. 135 In the medical domain, clinician trust in AI models is paramount. Statistical correlations and models alone are insufficient if clinicians do not trust the outputs. For AI models to be trusted and adopted, they must provide transparency, allowing clinicians to understand why certain outputs are generated.136,137 Techniques such as computer-aided transfer learning can highlight regions of anatomy that are crucial to the model's decision-making process, giving clinicians insights into areas that may warrant further review.138,139 However, most studies in this review did not incorporate explainability features, which remains a significant limitation in delaying clinical adoption. The development of interpretable AI systems will be critical for fostering clinician trust and ensuring these tools are effectively integrated into decision-making processes.
Limitations and potential biases
Variations in AI algorithms, MRI protocols, performance metrics, and patient populations complicate direct comparisons and limit the potential for meta-analysis. This diversity reflects the early stage of AI applications in DCD diagnosis using MRI and highlights the urgent need for standardized methodologies. The wide range of sample sizes across studies (28‒900 patients) also introduces potential selection bias, with smaller studies risking underpowered results and larger ones facing computational challenges and class imbalance issues.131,132 Publication bias is another concern, as studies with positive results are more likely to be published, potentially leading to an overestimation of AI's effectiveness in DCD diagnosis. Inconsistencies in ground truth labeling and the lack of standardized evaluation protocols further hinder objective comparisons across studies. Additionally, many studies lacked external validation, raising concerns about the generalizability of AI models.133,134 Despite these limitations, this systematic review provides a valuable overview of the current research landscape. It highlights existing gaps, methodological challenges, and the critical need for standardization, offering a foundation for future studies in this promising field.
Future research directions
This review identifies critical gaps in AI applications for DCD diagnosis that require urgent attention. There is a pressing need for standardized protocols for data collection, ground truth labeling, and model evaluation, along with the creation of public benchmark datasets. Many studies also lack robust external validation, raising concerns about the generalizability of AI models; future research should prioritize this aspect.133,134
While CSM has been well-studied,120,122,126,128,129 other DCD subtypes require further exploration. Enhancing model interpretability is crucial for fostering clinician trust and adoption, as most studies have not incorporated explainability features.136–139 Additionally, integrating MRI data with other clinical information and biomarkers through multimodal approaches could enhance the accuracy of AI models.140,141 The development of AI systems for early detection and diagnosis remains a priority, as this could significantly impact patient outcomes through timely intervention. Future research may also explore the potential of AI in understanding disease patterns and subtypes, which could enhance our understanding of DCD progression. By addressing these areas systematically, the field can advance toward more reliable and clinically applicable AI tools for DCD diagnosis.
Conclusions
AI has significant potential to enhance DCD diagnosis using MRI, particularly for early detection and personalized treatment. However, challenges such as the need for larger, multi-center studies, robust external validation, and improved model interpretability hinder widespread adoption. Addressing these issues through collaboration between AI developers and clinicians is essential for ensuring the clinical relevance and usability of these models. By focusing on these areas, the integration of AI into clinical practice can be more effectively achieved.
Supplemental Material
sj-docx-1-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-1-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Supplemental Material
sj-docx-2-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-2-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Supplemental Material
sj-docx-3-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-3-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Supplemental Material
sj-docx-4-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-4-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Supplemental Material
sj-docx-5-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-5-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Supplemental Material
sj-docx-6-dhj-10.1177_20552076241311939 - Supplemental material for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models
Supplemental material, sj-docx-6-dhj-10.1177_20552076241311939 for Artificial intelligence in degenerative cervical disease: A systematic review of MRI-based diagnostic models by Qian Du, Xinxin Shao, Minbo Zhang and Guangru Cao in DIGITAL HEALTH
Footnotes
Acknowledgments
The authors express my sincere gratitude for the unwavering support from my family members.
Declaration of conflicting interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by the Science and Technology Plan Projects of Zunyi City, the Science and Technology Fund Project of Guizhou Provincial Health Commission, the Basic Research Program of Guizhou Provincial Department of Science and Technology (grant numbers: Zunshi Kehe HZ [2024] No. 432, gzwkj2022-480, Qiankehe Foundation-ZK [2024] General-347).
Supplementary materials
The Supplementary Materials are uploaded with the submission.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
