Next Article in Journal
Is Autophagy Targeting a Valid Adjuvant Strategy in Conjunction with Tyrosine Kinase Inhibitors?
Next Article in Special Issue
Enhancing Surgical Guidance: Deep Learning-Based Liver Vessel Segmentation in Real-Time Ultrasound Video Frames
Previous Article in Journal
Significance of P53-Binding Protein 1 as a Novel Molecular Histological Marker for Hypopharyngeal Squamous Neoplasms
Previous Article in Special Issue
Outcome Prediction Using Multi-Modal Information: Integrating Large Language Model-Extracted Clinical Information and Image Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Oncologic Applications of Artificial Intelligence and Deep Learning Methods in CT Spine Imaging—A Systematic Review

1
Department of Diagnostic Imaging, National University Hospital, 5 Lower Kent Ridge Rd, Singapore 119074, Singapore
2
Department of Diagnostic Radiology, Yong Loo Lin School of Medicine, National University of Singapore, 10 Medical Drive, Singapore 117597, Singapore
3
National University Spine Institute, Department of Orthopaedic Surgery, National University Health System, 1E, Lower Kent Ridge Road, Singapore 119228, Singapore
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2024, 16(17), 2988; https://doi.org/10.3390/cancers16172988
Submission received: 10 July 2024 / Revised: 14 August 2024 / Accepted: 26 August 2024 / Published: 28 August 2024

Abstract

:

Simple Summary

In recent years, advances in deep learning have transformed the analysis of medical imaging, especially in spine oncology. Computed Tomography (CT) imaging is crucial for diagnosing, planning treatment, and monitoring spinal tumors. This review aims to comprehensively explore the current uses of deep learning tools in CT-based spinal oncology. Additionally, potential clinical applications of AI designed to address common challenges in this field will also be addressed.

Abstract

In spinal oncology, integrating deep learning with computed tomography (CT) imaging has shown promise in enhancing diagnostic accuracy, treatment planning, and patient outcomes. This systematic review synthesizes evidence on artificial intelligence (AI) applications in CT imaging for spinal tumors. A PRISMA-guided search identified 33 studies: 12 (36.4%) focused on detecting spinal malignancies, 11 (33.3%) on classification, 6 (18.2%) on prognostication, 3 (9.1%) on treatment planning, and 1 (3.0%) on both detection and classification. Of the classification studies, 7 (21.2%) used machine learning to distinguish between benign and malignant lesions, 3 (9.1%) evaluated tumor stage or grade, and 2 (6.1%) employed radiomics for biomarker classification. Prognostic studies included three (9.1%) that predicted complications such as pathological fractures and three (9.1%) that predicted treatment outcomes. AI’s potential for improving workflow efficiency, aiding decision-making, and reducing complications is discussed, along with its limitations in generalizability, interpretability, and clinical integration. Future directions for AI in spinal oncology are also explored. In conclusion, while AI technologies in CT imaging are promising, further research is necessary to validate their clinical effectiveness and optimize their integration into routine practice.

1. Introduction

Spinal malignancy encompasses both primary spine tumors and secondary spine metastases, with the latter being more prevalent [1,2]. The repercussions of spinal malignancy on quality of life are profound, stemming from complications like pain resulting from fractures (mechanical), spinal cord compression, and neurological deficits [3]. Moreover, it can lead to diminished mobility, bone marrow aplasia, and hypercalcemia, which manifest as symptoms such as constipation, polyuria, polydipsia, fatigue, and potentially life-threatening conditions like cardiac arrhythmias and acute renal failure [4,5]. Therefore, early detection, accurate diagnosis, and effective treatment of spinal malignancies, particularly metastases, are imperative to minimize complications and enhance patients’ overall well-being [6].
Artificial intelligence (AI) has made significant strides in the field of spinal oncology, offering radiologists valuable assistive tools to enhance their diagnostic capabilities and streamline workflows. In several applications, the findings in spinal oncology are either promising or have already surpassed previous benchmarks. For instance, the deep learning algorithm developed by Matohashi et al. [7] showed comparable sensitivity in detecting spinal metastases compared to orthopedic and radiology experts, and improved the sensitivity of in-training clinicians. AI has the potential to have a substantial impact on each step of the imaging value chain [8]. At this early stage in the integration of AI into clinical radiology, several studies have demonstrated its potential and utility. In recent years, there has been a growing interest in various AI applications, particularly in automating the detection and classification of spinal lesions [9]. CT, being a reliable imaging modality for assessing osseous involvement and the degree of destruction in spine abnormalities [10,11], has been widely studied for these reasons.
The purpose of this article is to introduce artificial intelligence techniques through a review of recent research, emphasizing their use at various phases of spinal oncology CT imaging, including image production and utilization. This narrative review aims to evaluate the current achievements of artificial intelligence (AI) and its possible applications to spinal oncology for both experts in the field and non-specialist readers. We hope to provide an overview of the most clinically pertinent applications of machine learning, including radiomics and convolutional neural networks (CNNs), in spinal oncology CT imaging, and to discuss potential future innovations in its use.

2. Technical Aspects

2.1. AI, ML, and DL

While “AI”, “machine learning” (ML), and “deep learning” (DL) are often used interchangeably, they actually represent distinct concepts. “AI” encompasses methods that enable computers to mimic intelligent human behavior [12,13]. “ML” is a specialized field within AI, utilizing statistical, mathematical, and computer science tools to enhance machine performance through experience [14,15,16]. “DL” further narrows down ML (Figure 1), focusing on the utilization of deep neural networks to analyze large datasets [17]. This particular domain is even more specialized compared to conventional machine learning, and leverages artificial neural networks, which involve several layers, to solve complex medical imaging challenges [18,19].
The complex layered structure allows the deep learning model or algorithm to learn information directly from imaging datasets and predict fundamental diagnostic patterns and features that surpass human capabilities [20,21]. These advanced deep learning techniques can then be used in accurate image classification (detecting the presence or absence of disease, as well as classifying disease severity), segmentation (based on individual pixels), and detection. Unlike traditional machine learning approaches that rely on manually engineered feature extraction from input images, deep learning methods autonomously learn image features by directly analyzing input images using multi-layer neural networks [22,23], which include convolutional neural networks (CNNs). This process transforms input images into valuable outputs. Deep learning systems not only excel at mapping image features to desired outputs but also possess the capability to discern intrinsic image details that often surpass human perception [24]. This advancement has significantly influenced the emergence of radiomics, a specialized field within medical imaging. Radiomics leverages various machine learning techniques, including deep learning, to extract and analyze a vast array of quantitative features from medical images. By utilizing these extracted features, radiomics enhances the identification, differentiation, and prognosis of lesions, enabling more precise differentiation between pathological and histological tumor subtypes. This capability underscores the pivotal role of deep learning in advancing radiomics, providing a deeper understanding of complex medical images and improving diagnostic accuracy [25].

2.2. Radiomics

Radiomics, an emerging field within AI and machine learning, involves converting radiological images containing important tumor-related information into quantifiable data [25]. These data assist clinicians in evaluating tumors beyond subjective visual interpretation, offering insights into tumor behavior and pathophysiology, including subtyping and grading [26,27]. When combined with clinical and qualitative imaging data, radiomics enhances medical decision-making processes, aiding disease prediction, prognosis assessment, and treatment response monitoring [28,29].
The workflow for building a radiomics model [30] comprises several steps: image acquisition, image processing (including segmentation and region of interest (ROI) selection), image feature extraction within the ROI, exploratory analysis and feature selection, and finally model building and classification (Figure 2). Factors such as the imaging modality, acquisition techniques, software, segmentation methods, and algorithm structure impact the process and vary accordingly [31,32]. Machine learning techniques like random decision forests can then be employed to validate and enhance the classification accuracy of predictors [33,34]. Ultimately, these tools can then be applied in the clinical setting to improve diagnostic accuracy and prognosis prediction [28,35,36,37,38].
Radiomic techniques consist of two main approaches: handcrafted-feature and deep learning-based analysis [39]. Handcrafted-feature radiomics [40,41] involves extracting numerical features from segmented regions of interest, categorized into shape, first-order statistics, textural features, and higher-order statistical features. Machine learning models, such as a Cox proportional hazards model, single vector machine (SVM), decision trees, and regression, are then developed to make clinical predictions based on these features [42,43] and validated for efficiency and sensitivity. In contrast, deep learning techniques use neural network architectures like CNNs to automatically extract important features from radiological images without prior descriptions [44,45,46]. These features undergo further processing for analysis or are used to generate machine learning models, which are then validated on larger datasets before they are applied in a clinical setting [47,48].
Although still in the early stage of development, the advent of radiomics has the potential to revolutionize medical imaging by offering rapid and comprehensive tumor characterization [49,50]. Radiomics holds the promise of enhancing traditional clinical testing and prognostication, making it a highly anticipated development for the future [51].

3. Materials and Methods

3.1. Literature Search Strategy

We systematically searched major electronic databases (PubMed, MEDLINE, Web of Science, clinicaltrials.gov) following PRISMA guidelines [52]. The search strategy involved using a combination of keywords and Medical Subject Headings (MeSH) to ensure comprehensive coverage. Specifically, we used the following query: (“Artificial intelligence” OR “AI” OR “deep learning” OR “machine learning” OR “convolutional neural network” OR “neural network” OR “radiomics”) AND (“spine” OR “spinal” OR “vertebral”) AND (“CT” OR “CT imaging”) AND (“malignancy” OR “metastasis” OR “cancer” OR “tumor” OR “oncology”). This query was applied to the title and abstract fields of the articles to ensure the retrieval of studies most relevant to our review topic. Two authors (W.O. and A.L.) independently reviewed the collected references and selected studies for detailed full-text screening. The literature search was conducted up to 31 March 2024.

3.2. Study Screening and Selection Criteria

In conducting this systematic review, we adopted an extensive search strategy to ensure a comprehensive evaluation of the literature. Our search process involved no specific restrictions, aiming to encompass a wide range of relevant studies. We focused on identifying scientific research that utilized radiomic techniques, artificial intelligence (AI), or deep learning methods within the context of CT imaging for spinal oncology.
To be included in the review, studies had to meet several criteria. First, the research needed to involve CT imaging analysis, specifically applying radiomics, AI, or deep learning techniques to the study of spinal oncology. Second, the studies had to assess the predictive and diagnostic capabilities of these technologies, with a particular emphasis on their integration into clinical practice and their potential to enhance diagnostic accuracy or provide prognostic insights. Third, only studies involving human subjects were considered to ensure the applicability of findings to real-world clinical settings. Finally, we limited our review to publications in English to maintain consistency and facilitate a thorough evaluation of the literature.
We excluded studies that did not align with these criteria. Specifically, case reports, editorial correspondence (such as letters, commentaries, and opinion pieces), and review articles were not included, as these do not present original research findings. Additionally, we excluded publications that focused on non-imaging radiomic techniques or did not utilize AI for the clinical analysis of CT images.
To ensure a thorough and exhaustive search, we concluded our literature review by examining the bibliographies of the selected publications. This additional step was undertaken to identify any relevant studies that might have been missed during the initial search, thereby enhancing the comprehensiveness of our review on the oncologic applications of AI and deep learning methods in CT spine imaging.

3.3. Data Extraction and Reporting

All chosen research articles were collected and organized into a spreadsheet using Microsoft Excel (Microsoft Corporation, Redmond, WA, USA). Information gathered from each research article included:
  • Details of the research article: authors, publication date, journal name;
  • Primary clinical application: detection, classification, segmentation, treatment, or prognosis prediction;
  • Study specifics: study type, patient characteristics or imaging modalities, body parts scanned, and specific bone areas segmented for analysis (e.g., internal or external datasets);
  • Machine learning methodologies utilized: radiomics, artificial neural networks, convolutional neural networks, etc.

4. Results

4.1. Search Results

The initial search across major electronic medical databases (Figure 3) identified a total of 226 relevant articles, which were screened based on the specified criteria. Articles were excluded if they were published more than 15 years ago, not written in English, were non-journal articles, or were conference abstracts/articles. Based on these exclusion criteria, 18 publications were initially excluded, leaving 208 articles for further full-text analysis to determine inclusion. For the remaining 208 articles, full-text reviews were performed, and a total of 177 publications were excluded as they focused on cancer sites other than the spine or did not utilize imaging data. An additional two articles were included after manually reviewing the bibliographies of selected articles. This process resulted in a total of 33 articles (Figure 1) selected for thorough analysis. Key findings from these studies were compiled (Table 1) and summarized in this review, with the main tasks of the AI models categorized into four broad categories: (1) Detection, (2) Classification, (3) Prognosis, and (4) Treatment Planning [53]. Most studies did not provide sufficient data to construct 2 × 2 contingency tables, precluding a formal meta-analysis.
Our search revealed that 12/33 studies (36.4%) focused on using AI techniques for detecting spinal malignancy, while 11/33 studies (33.3%) concentrated on classification. Additionally, 6/33 studies (18.2%) utilized radiomics for prognostication, 3/33 studies (9.1%) focused on treatment planning, and 1/33 study (3.0%) addressed both detection and classification. Among the studies focusing on classifying spinal malignancy using AI, 7/33 studies (21.2%) employed machine learning methods to distinguish between benign and malignant lesions, while 3/33 studies (9.1%) evaluated the stage or grade of spinal malignancy, and 2/33 studies (6.1%) classified the presence of certain tumor biomarkers using radiomic applications. Of the papers focusing on predicting prognosis, 3/33 studies (9.1%) dealt with predicting complications such as adverse outcomes or pathological fractures, while 3/33 studies (9.1%) aimed to predict treatment outcomes.

4.2. Performance Assessment

It is important to be aware of the various methodologies used to assess the performance of AI systems [54]. The most common key metrics used to assess the performance of AI systems include sensitivity, specificity, accuracy, precision, Area Under the Curve (AUC), Figure of Merit (FOM), F1-score [55,56], and kappa. Each of these metrics provides different insights into the performance of AI algorithms and their clinical relevance, with values closer to 1.0 indicating good performance:
  • Sensitivity (Recall) and Specificity: sensitivity (recall) measures the proportion of true positives (spinal tumors correctly identified by the AI system) out of all actual positives (all spinal tumors present in the CT images) [57]. Specificity, on the other hand, quantifies the ability of the AI system to correctly identify true negatives (normal spinal conditions) out of all actual negatives (all non-pathological conditions). These metrics are essential to assessing how well AI algorithms detect both positive and negative cases in spinal oncology;
  • Accuracy and Precision: accuracy indicates the overall correctness of the AI system’s predictions, measuring the ratio of correctly predicted cases (both true positives and true negatives) to the total number of cases evaluated. Precision, meanwhile, focuses on the AI system’s ability to accurately identify positive cases among all predicted positive instances, minimizing false positives. These metrics provide a comprehensive view of the AI algorithm’s reliability and correctness in clinical diagnosis;
  • Area Under the Curve (AUC): AUC evaluates the performance of AI models in binary classification tasks [58,59], such as distinguishing between diseased and healthy spinal conditions based on CT imaging features. A high AUC value indicates that the AI model effectively ranks diseased cases higher than healthy ones, demonstrating its discriminatory power in spinal oncology diagnostics;
  • Figure of Merit (FOM) and F1-score [60]: Figure of Merit encompasses a range of metrics including sensitivity, specificity, accuracy, and precision, tailored to the specific diagnostic challenges presented by spinal tumors. F1-score, a harmonic mean of precision and recall, balances the trade-off between these metrics and is particularly useful in scenarios where there is an imbalance between positive and negative cases in the dataset [61]. A model with a high F1 score indicates both good precision and recall, reflecting a robust model;
  • The kappa statistic [62], often denoted as κ, is a measure used to assess the level of agreement between two or more raters or classifiers beyond what would be expected by chance alone. It is particularly useful in the context of classification tasks, where it evaluates how well the AI system’s predictions align with the true classifications compared to random chance [63]. Kappa values range from −1 to 1. A kappa value of 1 indicates perfect agreement between the raters, 0 indicates agreement no better than chance, and negative values suggest worse-than-chance agreement. High kappa values indicate that the AI system’s predictions are consistently aligned with expert assessments or ground truth, which is vital for the system’s clinical applicability and trustworthiness.
These metrics provide a framework for assessing the clinical relevance and performance of AI algorithms in interpreting CT scans. By quantifying these metrics, clinicians and researchers can effectively evaluate the diagnostic accuracy, reliability, and potential impact of these AI-driven approaches in enhancing spinal tumor detection and management.
Table 1. Key characteristics of the selected articles.
Table 1. Key characteristics of the selected articles.
AuthorsAI MethodPublication YearMain ObjectivesJournalMain TaskSample Size
(No. of CTs/Patients)
Performance of AI Model
Tatjana W. et al. [64]SVM classifier2012Using supervised learning methods to detect sclerotic bone metastases in CT imaging. SPIE Medical ImagingDetection 22 Sensitivity: 71.2–87.0%
Hammon M. et al. [65]CADe (RF classifier) 2013 Automated detection of osteolytic and osteoblastic spine metastases on CT.European RadiologyDetection 134Sensitivity 83.0% (lytic), 88.0% (blastic)
Burns J. et al. [66]CADe (SVM classifier) 2013Automated detection of sclerotic metastases in the thoracolumbar spine on CT Radiology Detection59Sensitivity 79.0–90.0%
Roth. H. et al. [67] CNN (DropConnect)2014 Using Deep CNN methods to detect sclerotic spinal metastases on CT imaging. Recent Advances in Computational Methods and Clinical Applications for Spine ImagingDetection59Sensitivity: 79.0%; AUC 0.834
Masoudi S et al. [68]ResNet-50 with DC-GAN augmentation 2020 Detect and classify bone lesions in CT images of prostate cancer patients. Journal of Clinical Oncology Detection and Classification (Benign vs. malignant)56Sensitivity: 81.0%; accuracy 89.0%
Fan X et al. [69]AlexNet2021 Using Deep Learning Methods to Identify Spinal Metas in Lung Cancer using CT Images.HindawiDetection 36 Sensitivity: 66.0–81.4%
Noguchi S. et al. [70] 2D U-Net, 3D U-Net, and ResNet2022Deep learning to improve radiologists’ performance in spinal metastases detection on CT European Radiology Detection 732Sensitivity: 82.4–89.6%; improved radiologist’s sensitivity by 15.3%; reduced mean interpretation time by 83 s (p < 0.05).
Hoshiai S. et al. [71]DL Model 2022Detecting vertebral bone metastases on CTEuropean Journal of Radiology Detection 130FOM from 0.848 to 0.876 (radiologist) and from 0.752 to 0.799 (resident), (p < 0.05 for both)
Musa A. et al. [72] PyRadiomics 2022Detecting prostate cancer bone/spinal metastases invisible in CTCurrent Medical ImagingDetection 53Accuracy 85.0%; sensitivity 78.0–91.0%; specificity: 88.0–93.0%
Gilberg L. et al. [73]DLA (U-Net)2023 Deep learning to improve Radiologist’s detection of spinal malignancies in CT imaging. Applied SciencesDetection 32Sensitivity: 75.0%;
improved radiologist’s sensitivity by 20.8%
Huo T. et al. [74]3D U-Net (DCNN) 2023Deep learning to improve radiologists’ performance in lung cancer spinal metastases detection on CTFrontiers in OncologyDetection 126Sensitivity: 89.4%; improved radiologist’s sensitivity by 22.2% and accuracy of 26.2%
Koike Y. et al. [75]YOLOv5m and InceptionV3 2023 AI-aided lytic spinal bone metastasis detection on CT scans International Journal of Computer Assisted Radiology and Surgery Detection 2125Accuracy 87.2%; precision 94.8%; recall: 74.1%; F1-score 74.1%; AUC 0.940
Motohasi M. et al. [7]U-Net (DeepLabv3+)2024Using Deep Learning Algorithm to Detect Spinal Metastases on CT Images. SpineDetection 435Sensitivity: 75.0–78.0%;
precision: 36.0–68.0%;
F1 score: 0.48–0.72
Chmelik J. et al. [76]CNN2018Segmentation and classification of metastatic spinal lesions in 3D CT data Medical Image AnalysisClassification (Benign vs. malignant)31AUC 0.780–0.800;
sensitivity: 71.0–74.0%; specificity: 82.0–88.0%
Li Y. et al. [77] ResNet50 2021Differentiate benign and malignant vertebral fracture on CT using deep learning European Radiology Classification (Benign vs. malignant)433Sensitivity: 95.0%; specificity: 80.0%; accuracy: 88.0%
Masoudi S. et al. [78]2D ResNet-50, ResNeXt-50, 3D ResNet-18, 3D ResNet-50 2021Differentiate benign versus malignant spinal lesions on CT. IEEE Access Classification (Benign vs. malignant)114Accuracy: 79.4–92.2%; F1-Score: 0.755–0.923
Hallinan J. et al. [79] R-CNN (ResNet50)2022 Deep learning algorithm for grading cord compression secondary to spinal metastasis/epidural disease on CTCancersClassification (Stage/Grade)444kappas (κ: 0.873–0.911); AUC: 0.953–0.971; sensitivity: 92.6–98.0%; specificity: 94.8–99.8%
Naseri H. et al. [80]PyRadiomics 2022 Radiomics-based machine learning models to distinguish between metastatic and healthy bone. Scientific ReportClassification (Benign vs. malignant)170 AUC 0.640–0.950
Park, T. et al. [81]U-Net (CNN) 2022 Automated segmentation of the fractured vertebrae on CT and using radiomics to predict benign versus malignant. NatureClassification (Benign vs. malignant)158Dice similarity coefficient: 0.930–0.940; AUC 0.800–0.930
Wang, Q. et al. [82]Research Portal V1.1 2022Using Machine learning techniques to predict RANKL expression of Spinal GCTB. CancersClassification (Predict biomarkers) 107AUC: 0.658–0.880
sensitivity: 65.7–97.9%; specificity: 23.3–71.9%; accuracy: 64.6–80.2%
Wang, Q. et al. [83] PyRadiomics 2022Clinical and CT-Based Radiomics based techniques to predict p53 and VEGF expression in Spinal GCBT.Frontiers in OncologyClassification (Predict biomarkers) 80AUC 0.790–0.880
Hallinan J. et al. [84]R-CNN (ResNet50) 2023Deep learning method to diagnose epidural spinal cord compression using thoracolumbar CT.European Spine JournalClassification (Stage/Grade) 223kappa (κ  = 0.879); sensitivity: 91.8%; specificity: 92.0%; AUC: 0.919
Hallinan J. et al. [85]R-CNN (ResNeXt50)2023 Assess for metastatic spinal cord compression (mainly epidural extension) on CT imaging with external validation Frontiers in OncologyClassification (Stage/Grade)420Almost-perfect inter-rater agreement (κ = 0.813); sensitivity: 94.0%
Duan S. et al. [86]Inception_V3 2023Differentiating benign and malignant vertebral compression fracture on spinal CT imaging. European Journal of RadiologyClassification (Benign vs. malignant)280AUC: 0.890–0.990; accuracy: 88.0–99.0%
Gui C. et al. [87]PyRadiomics 2021 Radiomic modeling to predict risk of vertebral compression fracture after SBRT for spinal metastases Journal of NeurosurgeryPrognosis (Predicting complications) 74Sensitivity: 84.4%; specificity 80.0%, AUC 0.844–0.878 of 0.844, specificity of 0.800
Wang, Q. et al. [88]PyRadiomics 2021 Using Radiomics-based technique to predict recurrence in spinal GCBT from pre-operative CT imaging. Journal of Bone OncologyPrognosis (Predict treatment outcome) 62Accuracy: 89.0%; AUC 0.780 (predicting recurrence)
Massaad E. et al. [89] CNN (Densenet and U-Net)2022Using machine learning methods to derive body composition analysis to predict complications in spine tumor surgery. Journal of Neurosurgery SpinePrognosis (Predict treatment outcome) 484Body composition analysis using machine learning help predict risk for inferior outcomes.
Seou Y. et al. [90] PyRadiomics 2023Radiomics-based prediction of vertebral compression fracture prior to spinal SBRT from planning CT.European Spine Journal Prognosis (Predicting complications)85Accuracy: 78.8–82.9%; precision: 60.0–62.5%; F1 score: 0.573–0.650
Delrieu L. et al. [91] U-Net 2024Automated body composition analysis using L3 as reference on CT scans to predict treatment outcome in cancer patients. Frontiers in Nuclear MedicinePrognosis (Treatment Outcome) 352 DICE similarity coefficient: 0.850–0.940
Khalid S. et al. [92] DL Model2024Detection of sarcopenic obesity and association with adverse outcomes in patients undergoing surgical treatment for spinal metastases Journal of Neurosurgery SpinePrognosis (Predicting complications) 62 DL detected sarcopenia patients increased odds of non-home discharge, readmission, and postoperative mortality.
Sebastiaan et al. [93] DeepMedic2022Clinical Utility of CNN for treatment planning in spinal metastases.Physics and Imaging in Radiation OncologyTreatment Planning782DSC 96.7% HD: 3.6 mm.
Acceptable: 77.0%
Netherton T. et al. [94]CNN (X-Net) 2022 Automating Treatment Planning for Spinal Radiation Therapy using CT ImagingInternational Journal of Oncology, Biology and Physics Treatment Planning220Dice-similarity coefficient: 0.850 (cervical), 0.903 (thoracic), 93.7 (lumbar); AUC: 0.820; end-to-end treatment planning time <8 min
Hernandez S. et al. [95]nn-UNet 2023 Automating Treatment Planning for Paediatric Craniospinal Radiation Therapy using CT. Paediatric Blood CancerTreatment Planning 143Dice similarity coefficient: 0.650–0.980; end-to-end treatment planning time: 3.5 ± 0.4 min
Single Vector Machine (SVM), Convolutional Neural Network (CNN), Deep Learning Algorithm (DLA), Random Forest (RF), Computer-Aided Detection (CADe), Deep Learning (DL), Dice Similarity Coefficient (DSC), Hausdorff Distance (HD), Figure of Merit (FOM), Area Under Curve (AUC), Receiver Operating Characteristic (ROC), Giant Cell Bone Tumour (GCBT).

4.3. Applications

4.3.1. Detection of Spinal Lesions

Early identification and diagnosis of spinal lesions is crucial in clinical practice [96,97]. This plays a pivotal role in determining the disease stage in patients with malignancy, which significantly influences the treatment strategies and prognosis [98]. Spinal lesions, including metastases, are associated with increased morbidity, with over half of patients needing radiotherapy or invasive procedures due to complications such as spinal cord or nerve root compression [99,100,101]. Therefore, early diagnosis and targeted treatment before permanent neurological and functional deficits develop, is critical to achieve favorable outcomes [102,103,104].
However, detecting spinal lesions manually via various imaging techniques is often time-consuming, laborious, and challenging [105] due to factors such as overlapping imaging features with other pathologies [106] and variations in imaging quality, resulting in artifacts which can obscure lesions and hinder their detection [10,107]. Automated lesion detection is widely recognized for its potential to enhance radiologists’ sensitivity in identifying osseous metastases. Computer-aided detection (CADe) software and AI models have been shown to be as effective, if not more effective, than manual detection by radiologists.
The first CADe study using CT for spinal metastasis detection was conducted by O’Connor et al. in 2007 [108], specifically targeting osteolytic spinal metastases. This pioneering work laid the foundation for subsequent CADe research, which has been expanded to include other types of spinal metastases, such as osteoblastic and mixed lesions. Recently, advancements in artificial intelligence, particularly in DL and CNNs, have significantly enhanced the accuracy of CADe in detecting spinal metastases. These improvements have led to notable reductions in both false positive and false negative rates, thereby increasing the reliability of automated detection across different imaging techniques [67,69,109,110].
Several efforts employing diverse methods have been undertaken to detect spinal lesions in CT scans using AI techniques. For instance, Burns et al. [66] used a combination of watershed segmentation and a support vector machine classifier, achieving sensitivity of 79.0–90.0% in detecting osteoblastic spinal metastases. Hammon et al. [65] employed sequences of random forest classifiers that analyze local image features to evaluate regions for spinal metastases, attaining sensitivity of 83.0% for detecting osteolytic lesions and 88.0% in detecting osteoblastic lesions. Matohashi et al. [7] utilized deep learning methods using the “Deep Lab V3+” segmentation model, attaining sensitivity of 75.0–78.0% in detecting osteolytic lesions in thoracolumbar spine CT, which is comparable to that of orthopedic or radiology experts, and superior to that of in-training clinicians.
Several other techniques have been utilized to further improve sensitivity for detecting spinal metastases from CT. For instance, Hoshiai et al. [71] investigated the application of temporal subtraction CT for bone segmentation, using a multi-atlas-based method combined with spatial registration of two images via advanced computational techniques and deep learning methods showing improved overall field of merit (FOM) compared to both a board-certified radiologist and resident groups (p < 0.05). Both Noguchi et al. [9] and Huo et al. [59] developed deep convolutional neural network (DCNN) models that showed statistically significant improvements (p < 0.05) in the figure of merit (FOM) from 0.746 to 0.899 for radiologists, and in the detection accuracy and sensitivity of in-training radiologists, while also reducing the average interpretation time per case.
AI has already demonstrated its potential in oncology for improving the detection of various cancers. For example, in breast cancer screening, AI systems can accurately identify suspicious lesions on mammograms [111,112,113], reducing the workload for radiologists while maintaining high sensitivity and lowering recall rates [114,115]. In lung cancer, AI algorithms analyze CT scans to detect early-stage nodules with precision, facilitating timely intervention [116,117,118]. Additionally, AI tools in prostate cancer aid the detection of malignant tissues on MRI, improving diagnostic accuracy [119] and aiding treatment planning [119,120,121]. Although AI applications for detecting spinal lesions on CT scans are still in their infancy, their potential integration into clinical practice holds promise for helping alleviate radiologists’ workload [122,123] and reducing error rates [16]. By serving as a secondary reviewer, AI can enhance diagnostic accuracy and efficiency, providing a dependable safety net. Additionally, with improved detection accuracy and sensitivity, AI has the potential to minimize the necessity for further imaging modalities such as MRI or bone scans [124,125], potentially curbing healthcare costs.

4.3.2. Classification of Spinal Lesions

AI has shown promise in differentiating benign versus malignant lesions across various cancers, leveraging advanced algorithms to improve diagnostic accuracy [126,127,128,129]. In the context of spinal lesions, machine learning has been applied to differentiate between various spinal pathologies, such as metastases, by identifying key radiomic features in vertebral lesions and incorporating these into diverse machine learning models [130]. Spinal lesion classification has been predominantly studied on MRI due to its superior soft tissue contrast and various sequences allowing for more precise characterization of lesion morphology and signal characteristics [131,132,133,134]. For example, Liu et al. [135] and Xiong et al. [136] utilized MRI-based radiomics to differentiate between multiple myeloma and spinal metastases. Yin et al. [137] came out with a radiomics model for the differentiation of primary Chordoma, Giant Cell Tumor, and metastatic sacral tumors.
Distinguishing benign from malignant spinal lesions on CT remains challenging. The complexity of spinal anatomy, the variability of lesion appearance, and the overlapping features between benign and malignant lesions pose significant hurdles [138,139]. AI models can potentially overcome these challenges to provide reliable differentiation through extensive training on diverse datasets and continuous refinement to improve accuracy and reduce false positives. For example, the deep learning method developed by Masoudi et al. [78], using lesion-based average 2D ResNet-50 and 3D ResNet-18 architectures with texture, volumetric, and morphologic information, achieved an accuracy of 92.2% for the classification of benign versus malignant sclerotic bony lesions in patients with prostate cancer. Chmelik et al. [76] developed a deep CNN-based segmentation and classification of difficult-to-define spinal lesions using 3D CT data, achieving an AUC of 0.780–0.800 in distinguishing metastatic (both osteoblastic and osteolytic) bone lesions versus benign bone lesions.
Differentiating vertebral fractures due to benign causes, such as osteoporosis, from those caused by malignancy on CT imaging poses significant challenges. Both types of fracture can present with similar radiologic features, including vertebral body collapse and cortical disruption, complicating accurate diagnosis [140]. Often, CT imaging alone fails to reliably distinguish these conditions, necessitating further work-up such as contrast-enhanced MRI or isotope bone scans [141,142]. AI, leveraging radiomics and deep learning techniques, offers a promising solution by analyzing complex patterns in CT images that are not easily discernible to the human eye [143]. Radiomics and deep learning models can be trained to identify subtle differences in image characteristics, such as bone density and lesion shape, and by integrating these advanced analytics AI can improve the differentiation between benign osteoporotic fractures and pathological vertebral fractures.
For example, Li Y. et al. [77] developed a deep learning model using a ResNet 50 network with ten-fold cross-validation to achieve 85.0–88.0% accuracy in distinguishing benign from malignant compression fractures. Duan S. et al. [86] came up with radiomics and deep learning methods using Inception_V3 to differentiate these two entities through CT features and clinical characteristics, achieving an AUC of up to 0.990. Both studies also concluded that visual features such as the presence of a soft tissue mass and bone destruction were highly suggestive of malignancy, while the presence of a transverse fracture line is highly suggestive of a benign fracture—consistent with the literature [140,144,145].
AI applications go beyond mere tumor detection and differentiation, and they can also automatically generate significant parameters, such as grading spinal lesions and their complications. For example, Hallinan et al. [79] employed a deep learning model to automate the classification of metastatic epidural disease and/or spinal cord compression on CT scans based on the Bilsky classification [101,103]. The model showed near perfect agreement when compared to trained radiologists, with kappas of 0.873–0.911 (p < 0.001). Subsequent studies by the same author [85] showed that their deep learning algorithm for metastatic spinal cord compression on CT showed superior performance to the CT report with almost-perfect inter-rater agreement (κ = 0.813) and high sensitivity (94.0%) as compared to CT reports issued by experienced radiologists which had only slight inter-rater agreement (κ = 0.027) and low sensitivity (44.0%) (p < 0.001). Precise and consistent classification of metastatic epidural spinal cord compression will help clinicians determine whether initial treatment should involve radiotherapy or surgical intervention [146]. These studies demonstrate the potential of deep learning to assist clinicians in the early diagnosis and grading of metastatic spinal disease, ensuring that appropriate treatment is delivered promptly.
Radiogenomics, which combines “Radiomics” and “Genomics”, entails utilizing imaging traits or substitutes to identify genomic patterns and advanced bio-markers within tumors [147,148]. These markers subsequently inform clinical decisions, encompassing the prognosis, diagnosis, and predictive accuracy of tumor subtypes [149]. The typical workflow of a radiogenomics study (Figure 4) encompasses five key stages: image acquisition and pre-processing, feature extraction and selection from both medical imaging and genotype data, association of radiomic and genomic features, data analysis utilizing machine learning models, and the establishment of a final radiogenomics outcome model [150].
In the realm of oncologic spinal CT imaging, the focus of radiogenomics research has predominantly been on giant cell tumor of the bone (GCTB). Wang et al. [82,83] pioneered the application of machine learning and radiomic techniques to forecast the presence of key biomarkers—RANKL (receptor activator of the nuclear factor kappa B ligand), p53, and VEGF (vascular endothelial growth factor)—in spinal GCTB, achieving an AUC of 0.658–0.880. Elevated levels of RANKL, p53, and VEGF expression [151] in GCTB have been correlated with more aggressive tumor behavior and increased recurrence risk [152,153,154]. These biomarkers also form the basis of targeted molecular therapies, such as Denosumab, a monoclonal antibody targeting RANKL, used to treat challenging-to-operate spinal GCTB cases [155,156]. Identification of these biomarkers aids in prognosis prediction and facilitates the selection of optimal disease management strategies. However, their assessment typically necessitates invasive tissue biopsies. The ability to predict these biomarkers non-invasively offers a quantitative and convenient approach to support predictive decision-making, ultimately enhancing patient outcomes [157].

4.3.3. Prognosis and Predicting Complications

The integration of artificial intelligence (AI), deep learning, and radiomic models into CT spinal imaging holds promise for predicting outcomes and anticipating treatment complications in oncologic scenarios [158,159]. In addition to tumor characteristics, the assessment of sarcopenia and body composition plays a crucial role in the prognosis of patients with spinal metastases [160,161,162]. However, manual evaluation of these parameters can be tedious and exhausting, often requiring significant time and expertise [163]. AI and deep learning algorithms offer a transformative solution by automating the analysis of body composition from CT scans, providing rapid and accurate assessments [164]. These technologies can quantify muscle mass, adipose tissue distribution, and other relevant parameters with high precision, facilitating risk stratification and treatment planning.
For instance, Delrieu L et al. [91] developed a deep learning algorithm to automatically detect the L3 vertebra and segment body tissues to evaluate body composition and sarcopenia, achieving a median DICE similarity coefficient of up to 0.940 in relation to the sarcopenia metrics of the patient pool. Khalid et al. [92] came out with a machine learning-based technique for detection of sarcopenic obesity (SO) using CT in patients undergoing surgery for spinal metastases. Their algorithm effectively identified patients with SO, who exhibited increased odds of non-home discharge, re-admission, and post-operative mortality. These advancements enable clinicians to potentially recognize patients requiring nutritional interventions prior to invasive procedures, thereby enhancing overall treatment outcomes [165].
In addition to assessing body composition and sarcopenia, artificial intelligence (AI) holds promise for predicting the risk of recurrence, a crucial aspect in determining treatment outcomes for patients with spinal metastases. By analyzing radiomic features extracted from CT spinal imaging data, AI algorithms can identify subtle patterns and biomarkers associated with tumor aggressiveness and likelihood of recurrence [166,167]. For example, Wang Q et al. [88] developed a deep learning model to predict the risk of recurrence of spinal GCTB using images from pre-operative CT, achieving an accuracy of 89.0% and an AUC of 0.780. These predictive models leverage deep learning techniques to process vast amounts of imaging data and generate personalized risk assessments for individual patients [168]. Early identification of patients at high risk of recurrence enables clinicians to tailor treatment strategies, such as implementing adjuvant therapies or closer surveillance protocols, to mitigate the risk of disease progression and improve long-term outcomes [169].
Furthermore, AI-driven predictive models provide valuable insights into the underlying biological mechanisms driving tumor recurrence, facilitating the development of targeted therapeutic interventions aimed at preventing disease relapse and enhancing patient survival rates [170]. For instance, Gong et al. [171] came out with a radiomics model to predict pre-operative expression of PD-1 and PD-L1 in hepatocellular carcinoma (HCC), which was associated with more aggressive tumor behavior in the form of increased recurrence and distal metastatic risk [172]. The ability to use such an imaging biomarker may potentially help identify patients who will benefit from immune checkpoint inhibitor (ICI)-based treatment before surgery. Similarly, Bove et al. [173] developed a predictive model using radiomic features extracted from pre-treatment CT scans to forecast the likelihood of recurrence in patients with non-small cell lung cancer (NSCLC). Their findings demonstrated that specific radiomic signatures from the peri-tumor region correlated significantly with higher recurrence rates. By harnessing the power of AI-driven predictive models in spinal oncology, there is potential for clinicians to better stratify patients based on their risk of recurrence, tailor treatment strategies accordingly, and ultimately improve long-term survival rates.
AI and deep learning methodologies offer promising avenues to enhance the identification of patients at high risk of radiation-induced vertebral compression fracture (VCF) in the context of stereotactic body radiation therapy (SBRT) for spinal metastases. By extracting radiomic features from pre-treatment CT imaging data, these advanced techniques can provide more nuanced insights into the underlying tissue characteristics and microenvironment, enabling more accurate prediction of VCF risk [174]. Gui et al. [87] developed a machine learning model based on clinical characteristics and radiomic features from pre-treatment CT imaging to predict the risk of vertebral fracture following radiation therapy. Their model achieved sensitivity of 84.4%, specificity of 80.0%, and an area under the receiver operating characteristic (ROC) curve (AUC) of 0.878. The model developed by Seol et al., using similar methods, achieved an accuracy of 81.8% with an AUC of 0.870 in predicting VCF prior to spinal SBRT using pre-treatment planning CT. Integrating these AI-driven predictive models into clinical practice could potentially assist clinicians in implementing timely prophylactic measures to mitigate the occurrence of VCF and minimize associated morbidities before treatment [175]. This personalized approach to patient care will help optimize treatment strategies and improve overall treatment outcomes in individuals undergoing SBRT for spinal metastases.

4.3.4. Treatment Planning

Accurate delineation of spinal lesions is paramount for the effective planning of radiotherapy for spinal neoplasms and metastases [176,177,178]. However, manual outlining often proves time-consuming and susceptible to variability between observers. Deep learning methods present a transformative approach by automating the identification and segmentation of spinal metastases and surrounding anatomical structures [179]. This automation not only facilitates precise radiation dose calculation and treatment plan optimization but also significantly streamlines the workflow [180,181]. By reducing manual intervention and mitigating inter-observer variability, deep learning expedites the treatment planning process, offering potential benefits in clinical settings [182,183].
For instance, Hernandez et al. [95] developed a comprehensive automated contour and planning tool for 3D-conformal craniospinal irradiation therapy using CT images for pediatric patients with medulloblastoma. Their model achieved remarkable DICE similarity coefficients (DSC) ranging from 0.650 to 0.980, with an average end-to-end treatment planning time of 3.5 ± 0.4 min. Similarly, Sebastiaan et al. [93] devised a CNN-based model for automating segmentation and delineation of vertebral bodies in radiotherapy treatment planning for spinal metastases. Their model exhibited an average computational time of less than 5 min compared to an average of 20 min for manual contouring, with DSC scores ranging from 0.944 to 0.967, and 77.0% of cases deemed clinically acceptable. Although still in its infancy and not yet fully validated in clinical settings, these studies underscore the potential of AI and deep learning methods for enhancing the efficiency of radiotherapy workflows without compromising accuracy, ultimately improving patient care.
In current clinical practice, deep learning tools such as RayStation by RaySearch Labs are actively used for tumor segmentation in gamma radiotherapy planning. For instance, Riguad et al. [184] utilized RayStation for optimizing dose planning in cervical cancer treatment, while Almberg et al. [185] applied it in breast cancer treatment, demonstrating significant improvements in dose optimization. These examples highlight the real-world effectiveness of advanced segmentation technologies. Additionally, other commercial solutions, such as Varian’s Eclipse [186] and Elekta’s Monaco [187], also employ sophisticated deep learning algorithms for radiotherapy planning.

5. Discussion

5.1. Interpretation and Implications of Findings

In our systematic review of oncologic applications of artificial intelligence (AI) and deep learning methods in CT spine imaging, we observed a broad spectrum of applications including detection, classification, prognosis, and treatment planning. Our review indicated that AI models show considerable promise in these areas, demonstrating relatively reliable performance and considerable improvement in clinical practice. However, several critical insights and interpretations emerged from our findings.
AI models have proven effective in enhancing the detection and classification of spinal lesions. Despite these advancements, the performance metrics reported across studies exhibit notable variability, reflecting a significant limitation. This variability is partly due to the fact that many studies were constrained by small sample sizes and were conducted at single centers. Such limitations raise concerns about the generalizability of the results [188]. Specifically, while AI models may perform well within specific study settings, their effectiveness and reliability across diverse clinical environments remain uncertain without broader validation [189]. In the realm of prognosis and treatment planning, AI models have demonstrated substantial potential by improving the prediction of patient outcomes and aiding in the formulation of treatment strategies. Nonetheless, translating these capabilities into routine clinical practice requires further validation. The current evidence base, constrained by the limited scope of existing studies, necessitates larger, multi-center trials to confirm the models’ effectiveness and ensure their applicability to a wider range of patient populations.
Overall, while our review underscores the potential of AI and deep learning in advancing CT spine imaging, addressing the identified limitations through comprehensive validation and integration efforts is crucial for realizing their full clinical potential. These steps will ensure that AI technologies can provide reliable, actionable insights, and support the management of spinal malignancies.

5.2. Integration into Clinical Practice

The integration of AI models into clinical workflows presents several significant challenges that must be carefully addressed. Current research often suffers from a lack of external validation and does not fully tackle the practicalities of implementing these models in routine clinical settings [190,191]. To address these barriers, future research should prioritize the standardization of imaging protocols to reduce variability [192], expand datasets through collaborative and multi-center studies, and develop AI systems that seamlessly integrate into existing clinical practices [193]. Additionally, obtaining FDA and CE approvals for these tools is a critical step, ensuring they meet rigorous safety and efficacy standards before widespread adoption. To date, only a limited number of AI applications in CT spinal oncology have received such approvals, with most focusing on treatment planning [185] in general, and not specific to spinal oncology. However, we anticipate that more AI tools will achieve regulatory clearance in the future as the technology matures and more evidence of its clinical utility emerges.
Furthermore, the integration of AI and machine learning (ML) into healthcare introduces a range of ethical concerns. These include issues related to privacy and data security, the risk of biases in AI algorithms, and the challenges of ensuring transparency and explainability of AI systems [194]. Additional concerns involve patient consent, autonomy, and the equitable access to these technologies [195]. The complexity of these issues is compounded by problems such as limited data availability, data drift, and the need for ongoing retraining and regulatory updates [196]. Ensuring the safety and clinical validation of AI tools is essential to prevent misdiagnoses, avoid inappropriate treatments, and mitigate potential health disparities. Therefore, a balanced approach is crucial to leverage the benefits of AI while addressing these significant ethical and practical challenges effectively.

5.3. Other Potential Applications

5.3.1. Improving Image Quality

Oncologic spinal imaging poses significant challenges, particularly in complex cases such as post-surgical patients with metallic implants, which induce artifacts and degrade image quality, thereby impeding accurate evaluation of spinal lesions. These artifacts can obscure critical details (e.g., mass effect on the spinal cord or collections), complicating diagnosis, and treatment planning. However, deep learning reconstruction methods have emerged as valuable tools for enhancing CT imaging quality in oncology. Numerous studies have demonstrated their efficacy in mitigating artifacts and improving image clarity. For instance, Arabi et al. [197] developed a deep learning-based reconstruction algorithm that effectively reduced metal artifacts in PET/CT scans of post-surgical spinal patients, enhancing visualization of adjacent structures and facilitating more accurate lesion assessment. Similarly, Rui et al. [198] employed a deep learning-based metal artifact correction (MAC) algorithm, achieving significantly higher subjective scores as compared to conventional MAC and virtual monochromatic imaging (VMI) techniques.
Building upon these successes, there is potential for Generative Adversarial Networks (GANs) to further advance the field. GANs offer the ability to generate realistic and artifact-corrected images by learning from large datasets, potentially providing a solution to the challenges posed by various artifacts in CT spinal imaging. For instance, Lu et al. [199] utilized GAN to correct motion artefacts on CT Coronary Angiogram images. Their GAN-generated images showed statistically significant improvement in motion artifact alleviation score (4/5 vs. 1/5, p < 0.001) and overall image quality score (4/5 vs. 1/5, p < 0.001), with high accuracy in identifying stenosis (81.0% vs. 66.0%) in the mid-right coronary artery as compared to motion-affected images. Goli et al. [200] utilized GAN to improve CT images of the head and neck affected by metallic artefacts from dental implants, achieving 16.8% improvement in assessing the oral cavity region, which is important for treatment planning. While traditional non-AI-based metal artifact correction methods, such as i-MAR used in Siemens CT scanners [201,202], are commonly employed in clinical settings, deep learning reconstruction methods offer several significant advantages. These methods enhance image quality by more effectively reducing metal artifacts and improving the visualization of adjacent structures [203]. They provide greater adaptability to new data and imaging modalities, reduce the need for manual adjustments, and often result in faster processing times and better quantitative accuracy [204]. Consequently, deep learning techniques offer potentially more robust solutions for addressing metallic artifacts in CT imaging, leading to more reliable and accurate diagnostic outcomes [205]. By synthesizing artifact-free images, GANs hold promise for improving diagnostic accuracy and enhancing oncologic evaluations in patients with spinal lesions.

5.3.2. Predicting Primary Malignancy from Spinal Metastases

Spinal metastasis originating from an unidentified primary tumor presents a common clinical challenge, affecting up to 30% of patients on initial presentation [206,207]. While conventional CT scans can accurately detect vertebral metastases, distinguishing between cancers of various origins can be challenging as they often appear similar. Consequently, these patients often have to undergo additional PET/CT imaging for primary cancer diagnosis and comprehensive whole-body staging scans prior to treatment initiation [208]. In rare cases, even with further imaging, the primary tumor remains unidentified, necessitating an invasive biopsy to determine the probable primary tumor site and to explore more targeted treatment options.
Several studies have demonstrated the effectiveness of deep learning models in predicting primary tumor sites in patients with spinal metastases using MRI-derived imaging features. For example, Liu et al. [209] explored the feasibility of a ResNet-50 convolutional neural network model for this purpose, achieving an AUC–ROC of 0.770 and 53.0% accuracy in classifying spinal metastases originating from the lung, kidney, prostate, breast, and thyroid. Lang et al. [210] utilized radiomics and deep learning techniques to distinguish spinal metastases from lung cancer and other origins using dynamic contrast-enhanced (DCE) sequences from a spinal MRI database. Their findings indicated that the DCE kinetic measurement of the washout slope from a hotspot within the spinal metastatic lesion was the most reliable parameter for diagnosing primary lung cancer versus other tumor types, achieving accuracies of up to 81.0%.
Cao et al. [211] investigated radiomics for distinguishing primary tumors from brain metastases using CT images. Shang et al. [184] developed a similar model for identifying primary tumor types from lung metastases in thoracic CT scans. Despite bone metastases being common, no studies have specifically addressed primary tumor differentiation in spinal metastases using CT imaging. This gap presents a significant opportunity for future research in spinal oncology. With promising results seen in MRI studies, applying deep learning techniques to CT spinal imaging shows potential. CT scans offer widespread availability and detailed anatomical information, making them valuable for assessing primary malignancies. By training deep learning models on large datasets of CT spinal images, robust algorithms could potentially differentiate and identify primary malignancies solely from CT scans. Leveraging radiomic approaches from other metastatic contexts could assist with developing predictive models specific to spinal metastases. Advanced imaging analytics applied to CT scans could reveal distinct patterns and biomarkers for different primary tumors, enhancing diagnostic accuracy and treatment planning. These advancements may provide clinicians with a non-invasive method to efficiently diagnose and plan treatments for spinal metastases.

5.3.3. Quantifying Tumor Burden to Predict Treatment Response

Artificial intelligence (AI) applications hold promise for enhancing the measurement of tumor burden for assessing treatment response and monitoring tumor progression. AI-assisted segmentation allows for precise lesion and/or tumor volumetry, facilitating more accurate evaluations. For instance, Goehler et al. [212] implemented a deep learning method to estimate overall tumor burden for neuroendocrine neoplasia on MRI, achieving a concordance of 91% with manual clinician assessment and DSC of up to 0.81. Belal et al. [213] developed a fully automated CNN-based model to calculate skeletal tumor burden in patients with prostate cancer on PET/CT, achieving a moderately strong PET index correlation with that estimated by the physician (mean r = 0.69).
Assessing tumor burden and volume in spinal metastases presents challenges due to variations in vertebral shape and involvement across multiple levels [185] Manual volumetric assessment is not only time-consuming but also prone to variability among different observers, and even within the same observer. In clinical practice, evaluating radiological images by oncology specialists and radiologists is hindered by the labor-intensive nature of manual analysis, the absence of standardized quantification, and challenges related to reproducibility in both evaluation and measurement. Despite that, assessing tumor burden in spinal metastases is pivotal for predicting prognosis and treatment efficacy [214]. While there are studies which have looked into using deep learning to assess bone tumor burden in PET/CT [213,215] and bone scintigraphy [216], the current literature lacks studies exploring the use of CT imaging to comprehensively evaluate spinal tumor burden. This gap is significant, especially given the frequent use of whole-body CT scans in oncology for monitoring treatment response. Spinal metastases, including those affecting the spine, necessitate a holistic assessment of the entire skeleton due to the systemic nature of metastatic disease.
Recent advancements in deep learning applied to CT imaging have automated the segmentation of bone metastases, including those affecting the spine. Studies such as those by Motohashi et al. [7] and Saeed et al. [217] have shown deep learning algorithms’ effectiveness in identifying and quantifying metastatic lesions across skeletal regions. Despite these capabilities in segmenting bone tumors, there remains a lack of research specifically focused on assessing tumor burden. These AI-driven approaches could streamline the volumetric evaluation of spinal metastases and pave the way for improved clinical management by providing a comprehensive view of tumor burden throughout the body. Integrating AI into the assessment of bone metastases using CT holds the potential to significantly enhance the timely and accurate monitoring of treatment response. By automating tumor burden quantification from whole-body CT scans, AI can potentially guide oncologists to more effectively evaluate overall disease burden, which is crucial for assessing treatment efficacy and making informed decisions in patient care [218,219].

5.4. Study Limitations

Due to the heterogeneity among the studies reviewed and insufficient data, a formal meta-analysis could not be conducted. Consequently, our review is presented as a descriptive analysis rather than as a quantitative synthesis. This limitation affects our ability to perform statistical aggregation of results and derive more generalized conclusions. Additionally, the quality of data in the individual studies was not analyzed in detail, which could influence the robustness of the findings. However, this does not diminish the value of our review. Despite these limitations, our descriptive review provides valuable insights into the current state of oncologic applications of artificial intelligence and deep learning methods in CT spine imaging. It highlights key trends, identifies research gaps, and offers a comprehensive overview of technological advancements and their clinical implications. By synthesizing and summarizing findings from a diverse range of studies, our review contributes significantly to the understanding of these emerging technologies and provides a foundation for future research directions and clinical applications in the field. The qualitative synthesis presented serves as a useful resource for researchers and practitioners, guiding further investigation and application of AI and deep learning in this domain.
Furthermore, our review acknowledges a notable gap in addressing perspectives from patients and clinicians regarding the use of AI in spinal oncology. Incorporating feedback from these key stakeholders is essential for understanding the real-world implications of AI applications. Patients’ experiences [220] and clinicians’ insights [221,222] are critical for evaluating the practical benefits and limitations of AI technologies, as well as their acceptance and integration into routine clinical practice. Additionally, our review did not extensively cover the long-term impact and cost-effectiveness of AI applications [223]. Evaluating these factors is important for assessing the overall value of AI to spinal oncology, as it can influence decision-making processes, resource allocation, and the sustainability of these technologies in clinical settings. Future research should address these aspects to provide a more comprehensive evaluation of AI’s role in spinal oncology, ensuring that the benefits of AI applications are balanced with considerations of their economic and long-term impacts.

6. Conclusions

In conclusion, the integration of deep learning techniques with computed tomography (CT) imaging in spinal oncology presents a promising avenue for enhancing diagnostic accuracy, treatment planning, and improving patient outcomes. This review has demonstrated the diverse applications of artificial intelligence (AI) in CT imaging of spinal metastases, including detection, classification, grading, and treatment planning. AI technologies have demonstrated notable performance across these domains, offering the potential to support clinicians by improving workflow efficiency and minimizing complications. However, despite promising findings, additional research is warranted to validate the clinical effectiveness of these AI tools and streamline their integration into everyday clinical workflows. Ultimately, the continued exploration and refinement of AI applications in CT spinal imaging holds immense promise for advancing the field of spinal oncology and improving patient care.

Author Contributions

Conceptualization, methodology, supervision, and writing: A.L., W.O., W.C.T., X.Z.L., D.D.L., S.G. and J.T.P.D.H. Investigation and project administration: W.O., A.L., K.T.D.F., Y.L.T., A.M., Y.H.T. and J.T.P.D.H. Resources and software: W.O., A.L., J.H.T., N.K. and J.T.P.D.H. Formal analysis and validation: W.O., A.L., W.C.T., X.Z.L., S.G., K.T.D.F., Y.L.T., S.J.O. and J.T.P.D.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was directly funded by MOH/NMRC, Singapore. Specifically, this study received support from the Singapore Ministry of Health National Medical Research Council under the NMRC Clinician Innovator Award (CIA). The grant was awarded for the project titled “Deep learning pipeline for augmented reporting of MRI whole spine” (Grant ID: CIAINV23jan-0001, MOH-001405, J.T.P.D.H).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mundy, G.R. Metastasis to bone: Causes, consequences and therapeutic opportunities. Nat. Rev. Cancer 2002, 2, 584–593. [Google Scholar] [CrossRef] [PubMed]
  2. Witham, T.F.; Khavkin, Y.A.; Gallia, G.L.; Wolinsky, J.P.; Gokaslan, Z.L. Surgery insight: Current management of epidural spinal cord compression from metastatic spine disease. Nat. Clin. Pract. Neurol. 2006, 2, 87–94. [Google Scholar] [CrossRef] [PubMed]
  3. Coleman, R.E. Metastatic bone disease: Clinical features, pathophysiology and treatment strategies. Cancer Treat. Rev. 2001, 27, 165–176. [Google Scholar] [CrossRef]
  4. Cecchini, M.G.; Wetterwald, A.; van der Pluijm, G.; Thalmann, G.N. Molecular and Biological Mechanisms of Bone Metastasis. EAU Update Ser. 2005, 3, 214–226. [Google Scholar] [CrossRef]
  5. Cuccurullo, V.; Cascini, G.L.; Tamburrini, O.; Rotondo, A.; Mansi, L. Bone metastases radiopharmaceuticals: An overview. Curr. Radiopharm. 2013, 6, 41–47. [Google Scholar] [CrossRef] [PubMed]
  6. Yu, H.H.; Tsai, Y.Y.; Hoffe, S.E. Overview of diagnosis and management of metastatic disease to bone. Cancer Control 2012, 19, 84–91. [Google Scholar] [CrossRef] [PubMed]
  7. Motohashi, M.; Funauchi, Y.; Adachi, T.; Fujioka, T.; Otaka, N.; Kamiko, Y.; Okada, T.; Tateishi, U.; Okawa, A.; Yoshii, T.; et al. A New Deep Learning Algorithm for Detecting Spinal Metastases on Computed Tomography Images. Spine 2024, 49, 390–397. [Google Scholar] [CrossRef]
  8. Cui, Y.; Zhu, J.; Duan, Z.; Liao, Z.; Wang, S.; Liu, W. Artificial Intelligence in Spinal Imaging: Current Status and Future Directions. Int. J. Environ. Res. Public Health 2022, 19, 11708. [Google Scholar] [CrossRef]
  9. Martín-Noguerol, T.; Oñate Miranda, M.; Amrhein, T.J.; Paulano-Godino, F.; Xiberta, P.; Vilanova, J.C.; Luna, A. The role of Artificial intelligence in the assessment of the spine and spinal cord. Eur. J. Radiol. 2023, 161, 110726. [Google Scholar] [CrossRef]
  10. Kim, G.U.; Chang, M.C.; Kim, T.U.; Lee, G.W. Diagnostic Modality in Spine Disease: A Review. Asian Spine J. 2020, 14, 910–920. [Google Scholar] [CrossRef]
  11. Ariyaratne, S.; Jenko, N.; Iyengar, K.P.; James, S.; Mehta, J.; Botchu, R. Primary Osseous Malignancies of the Spine. Diagnostics 2023, 13, 1801. [Google Scholar] [CrossRef] [PubMed]
  12. Thrall, J.H.; Li, X.; Li, Q.; Cruz, C.; Do, S.; Dreyer, K.; Brink, J. Artificial Intelligence and Machine Learning in Radiology: Opportunities, Challenges, Pitfalls, and Criteria for Success. J. Am. Coll. Radiol. 2018, 15, 504–508. [Google Scholar] [CrossRef] [PubMed]
  13. Oren, O.; Gersh, B.J.; Bhatt, D.L. Artificial intelligence in medical imaging: Switching from radiographic pathological data to clinically meaningful endpoints. Lancet Digit. Health 2020, 2, e486–e488. [Google Scholar] [CrossRef]
  14. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine Learning for Medical Imaging. Radiographics 2017, 37, 505–515. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, S.; Summers, R.M. Machine learning and radiology. Med. Image Anal. 2012, 16, 933–951. [Google Scholar] [CrossRef]
  16. Najjar, R. Redefining Radiology: A Review of Artificial Intelligence Integration in Medical Imaging. Diagnostics 2023, 13, 2760. [Google Scholar] [CrossRef]
  17. Cheng, P.M.; Montagnon, E.; Yamashita, R.; Pan, I.; Cadrin-Chênevert, A.; Romero, F.P.; Chartrand, G.; Kadoury, S.; Tang, A. Deep Learning: An Update for Radiologists. RadioGraphics 2021, 41, 1427–1445. [Google Scholar] [CrossRef]
  18. Zaharchuk, G.; Gong, E.; Wintermark, M.; Rubin, D.; Langlotz, C.P. Deep Learning in Neuroradiology. Am. J. Neuroradiol. 2018, 39, 1776–1784. [Google Scholar] [CrossRef]
  19. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef]
  20. Jussupow, E.; Spohrer, K.; Heinzl, A. Radiologists’ Usage of Diagnostic AI Systems. Bus. Inf. Syst. Eng. 2022, 64, 293–309. [Google Scholar] [CrossRef]
  21. Kaka, H.; Zhang, E.; Khan, N. Artificial intelligence and deep learning in neuroradiology: Exploring the new frontier. Can. Assoc. Radiol. J. 2021, 72, 35–44. [Google Scholar] [CrossRef] [PubMed]
  22. Daugaard Jørgensen, M.; Antulov, R.; Hess, S.; Lysdahlgaard, S. Convolutional neural network performance compared to radiologists in detecting intracranial hemorrhage from brain computed tomography: A systematic review and meta-analysis. Eur. J. Radiol. 2022, 146, 110073. [Google Scholar] [CrossRef]
  23. Mutasa, S.; Sun, S.; Ha, R. Understanding artificial intelligence based radiology studies: CNN architecture. Clin. Imaging 2021, 80, 72–76. [Google Scholar] [CrossRef]
  24. Manakitsa, N.; Maraslidis, G.S.; Moysis, L.; Fragulis, G.F. A Review of Machine Learning and Deep Learning for Object Detection, Semantic Segmentation, and Human Action Recognition in Machine and Robotic Vision. Technologies 2024, 12, 15. [Google Scholar] [CrossRef]
  25. van Timmeren, J.E.; Cester, D.; Tanadini-Lang, S.; Alkadhi, H.; Baessler, B. Radiomics in medical imaging—”How-to” guide and critical reflection. Insights Imaging 2020, 11, 91. [Google Scholar] [CrossRef] [PubMed]
  26. Shur, J.D.; Doran, S.J.; Kumar, S.; ap Dafydd, D.; Downey, K.; O’Connor, J.P.B.; Papanikolaou, N.; Messiou, C.; Koh, D.-M.; Orton, M.R. Radiomics in Oncology: A Practical Guide. RadioGraphics 2021, 41, 1717–1732. [Google Scholar] [CrossRef] [PubMed]
  27. Aerts, H.J. The Potential of Radiomic-Based Phenotyping in Precision Medicine: A Review. JAMA Oncol. 2016, 2, 1636–1642. [Google Scholar] [CrossRef]
  28. Rogers, W.; Thulasi Seetha, S.; Refaee, T.A.G.; Lieverse, R.I.Y.; Granzier, R.W.Y.; Ibrahim, A.; Keek, S.A.; Sanduleanu, S.; Primakov, S.P.; Beuque, M.P.L.; et al. Radiomics: From qualitative to quantitative imaging. Br. J. Radiol. 2020, 93, 20190948. [Google Scholar] [CrossRef]
  29. Cobo, M.; Menéndez Fernández-Miranda, P.; Bastarrika, G.; Lloret Iglesias, L. Enhancing radiomics and Deep Learning systems through the standardization of medical imaging workflows. Sci. Data 2023, 10, 732. [Google Scholar] [CrossRef]
  30. Avery, E.; Sanelli, P.C.; Aboian, M.; Payabvash, S. Radiomics: A Primer on Processing Workflow and Analysis. Semin. Ultrasound CT MR 2022, 43, 142–146. [Google Scholar] [CrossRef]
  31. Sharma, N.; Aggarwal, L.M. Automated medical image segmentation techniques. J. Med. Phys. 2010, 35, 3–14. [Google Scholar] [CrossRef]
  32. Rana, M.; Bhushan, M. Machine learning and deep learning approach for medical image analysis: Diagnosis to detection. Multimed. Tools Appl. 2023, 82, 26731–26769. [Google Scholar] [CrossRef] [PubMed]
  33. Valladares, A.; Beyer, T.; Rausch, I. Physical imaging phantoms for simulation of tumor heterogeneity in PET, CT, and MRI: An overview of existing designs. Med. Phys. 2020, 47, 2023–2037. [Google Scholar] [CrossRef]
  34. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Rizzo, S.; Botta, F.; Raimondi, S.; Origgi, D.; Fanciullo, C.; Morganti, A.G.; Bellomi, M. Radiomics: The facts and the challenges of image analysis. Eur. Radiol. Exp. 2018, 2, 36. [Google Scholar] [CrossRef] [PubMed]
  36. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef] [PubMed]
  37. Larsen, M.; Olstad, C.F.; Lee, C.I.; Hovda, T.; Hoff, S.R.; Martiniussen, M.A.; Mikalsen, K.; Lund-Hanssen, H.; Solli, H.S.; Silberhorn, M.; et al. Performance of an Artificial Intelligence System for Breast Cancer Detection on Screening Mammograms from BreastScreen Norway. Radiol. Artif. Intell. 2024, 6, e230375. [Google Scholar] [CrossRef]
  38. Khawaja, A.; Bartholmai, B.J.; Rajagopalan, S.; Karwoski, R.A.; Varghese, C.; Maldonado, F.; Peikert, T. Do we need to see to believe?—Radiomics for lung nodule classification and lung cancer risk stratification. J. Thorac. Dis. 2020, 12, 3303–3316. [Google Scholar] [CrossRef] [PubMed]
  39. Astaraki, M.; Yang, G.; Zakko, Y.; Toma-Dasu, I.; Smedby, Ö.; Wang, C. A Comparative Study of Radiomics and Deep-Learning Based Methods for Pulmonary Nodule Malignancy Prediction in Low Dose CT Images. Front. Oncol. 2021, 11, 737368. [Google Scholar] [CrossRef]
  40. Liu, Z.; Jiang, Z.; Meng, L.; Yang, J.; Liu, Y.; Zhang, Y.; Peng, H.; Li, J.; Xiao, G.; Zhang, Z.; et al. Handcrafted and Deep Learning-Based Radiomic Models Can Distinguish GBM from Brain Metastasis. J. Oncol. 2021, 2021, 5518717. [Google Scholar] [CrossRef]
  41. Salmanpour, M.R.; Rezaeijo, S.M.; Hosseinzadeh, M.; Rahmim, A. Deep versus Handcrafted Tensor Radiomics Features: Prediction of Survival in Head and Neck Cancer Using Machine Learning and Fusion Techniques. Diagnostics 2023, 13, 1696. [Google Scholar] [CrossRef]
  42. Huang, Y.; Li, J.; Li, M.; Aparasu, R.R. Application of machine learning in predicting survival outcomes involving real-world data: A scoping review. BMC Med. Res. Methodol. 2023, 23, 268. [Google Scholar] [CrossRef] [PubMed]
  43. Xiao, J.; Mo, M.; Wang, Z.; Zhou, C.; Shen, J.; Yuan, J.; He, Y.; Zheng, Y. The Application and Comparison of Machine Learning Models for the Prediction of Breast Cancer Prognosis: Retrospective Cohort Study. JMIR Med. Inform. 2022, 10, e33440. [Google Scholar] [CrossRef] [PubMed]
  44. Zhang, Z.; Sejdić, E. Radiological images and machine learning: Trends, perspectives, and prospects. Comput. Biol. Med. 2019, 108, 354–370. [Google Scholar] [CrossRef]
  45. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
  46. Puttagunta, M.; Ravi, S. Medical image analysis based on deep learning approach. Multimed. Tools Appl. 2021, 80, 24365–24398. [Google Scholar] [CrossRef]
  47. Arbet, J.; Brokamp, C.; Meinzen-Derr, J.; Trinkley, K.E.; Spratt, H.M. Lessons and tips for designing a machine learning study using EHR data. J. Clin. Transl. Sci. 2020, 5, e21. [Google Scholar] [CrossRef]
  48. Zhang, A.; Xing, L.; Zou, J.; Wu, J.C. Shifting machine learning for healthcare from development to deployment and from models to data. Nat. Biomed. Eng. 2022, 6, 1330–1345. [Google Scholar] [CrossRef]
  49. Liu, Z.; Wang, S.; Dong, D.; Wei, J.; Fang, C.; Zhou, X.; Sun, K.; Li, L.; Li, B.; Wang, M.; et al. The Applications of Radiomics in Precision Diagnosis and Treatment of Oncology: Opportunities and Challenges. Theranostics 2019, 9, 1303–1322. [Google Scholar] [CrossRef]
  50. Limkin, E.J.; Sun, R.; Dercle, L.; Zacharaki, E.I.; Robert, C.; Reuzé, S.; Schernberg, A.; Paragios, N.; Deutsch, E.; Ferté, C. Promises and challenges for the implementation of computational medical imaging (radiomics) in oncology. Ann. Oncol. 2017, 28, 1191–1206. [Google Scholar] [CrossRef]
  51. Li, S.; Zhou, B. A review of radiomics and genomics applications in cancers: The way towards precision medicine. Radiat. Oncol. 2022, 17, 217. [Google Scholar] [CrossRef]
  52. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
  53. Pinto-Coelho, L. How Artificial Intelligence Is Shaping Medical Imaging Technology: A Survey of Innovations and Applications. Bioengineering 2023, 10, 1435. [Google Scholar] [CrossRef] [PubMed]
  54. Tanguay, W.; Acar, P.; Fine, B.; Abdolell, M.; Gong, B.; Cadrin-Chênevert, A.; Chartrand-Lefebvre, C.; Chalaoui, J.; Gorgos, A.; Chin, A.S.-L.; et al. Assessment of Radiology Artificial Intelligence Software: A Validation and Evaluation Framework. Can. Assoc. Radiol. J. 2023, 74, 326–333. [Google Scholar] [CrossRef] [PubMed]
  55. Handelman, G.S.; Kok, H.K.; Chandra, R.V.; Razavi, A.H.; Huang, S.; Brooks, M.; Lee, M.J.; Asadi, H. Peering into the Black Box of Artificial Intelligence: Evaluation Metrics of Machine Learning Methods. Am. J. Roentgenol. 2019, 212, 38–43. [Google Scholar] [CrossRef]
  56. McKinney, S.M.; Sieniek, M.; Godbole, V.; Godwin, J.; Antropova, N.; Ashrafian, H.; Back, T.; Chesus, M.; Corrado, G.S.; Darzi, A.; et al. International evaluation of an AI system for breast cancer screening. Nature 2020, 577, 89–94. [Google Scholar] [CrossRef] [PubMed]
  57. Hajian-Tilaki, K. Receiver Operating Characteristic (ROC) Curve Analysis for Medical Diagnostic Test Evaluation. Casp. J. Intern. Med. 2013, 4, 627–635. [Google Scholar]
  58. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef]
  59. Obuchowski, N.A. Receiver operating characteristic curves and their use in radiology. Radiology 2003, 229, 3–8. [Google Scholar] [CrossRef]
  60. Powers, D. Evaluation: From Precision, Recall and F-Factor to ROC, Informedness, Markedness & Correlation. Int. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar] [CrossRef]
  61. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  62. Czodrowski, P. Count on kappa. J. Comput.-Aided Mol. Des. 2014, 28, 1049–1055. [Google Scholar] [CrossRef]
  63. Yang, F.; Zamzmi, G.; Angara, S.; Rajaraman, S.; Aquilina, A.; Xue, Z.; Jaeger, S.; Papagiannakis, E.; Antani, S.K. Assessing Inter-Annotator Agreement for Medical Image Segmentation. IEEE Access 2023, 11, 21300–21312. [Google Scholar] [CrossRef]
  64. Wiese, T.; Yao, J.; Burns, J.E.; Summers, R.M. Detection of sclerotic bone metastases in the spine using watershed algorithm and graph cut. In Proceedings of the SPIE Medical Imaging 2012, San Diego, CA, USA, 4–9 February 2012; SPIE: Bellingham, WA, USA, 2012; Volume 831512. [Google Scholar] [CrossRef]
  65. Hammon, M.; Dankerl, P.; Tsymbal, A.; Wels, M.; Kelm, M.; May, M.; Suehling, M.; Uder, M.; Cavallaro, A. Automatic detection of lytic and blastic thoracolumbar spine metastases on computed tomography. Eur. Radiol. 2013, 23, 1862–1870. [Google Scholar] [CrossRef] [PubMed]
  66. Burns, J.E.; Yao, J.; Wiese, T.S.; Muñoz, H.E.; Jones, E.C.; Summers, R.M. Automated detection of sclerotic metastases in the thoracolumbar spine at CT. Radiology 2013, 268, 69–78. [Google Scholar] [CrossRef] [PubMed]
  67. Roth, H.R.; Yao, J.; Lu, L.; Stieger, J.; Burns, J.E.; Summers, R.M. Detection of Sclerotic Spine Metastases via Random Aggregation of Deep Convolutional Neural Network Classifications. In Recent Advances in Computational Methods and Clinical Applications for Spine Imaging; Yao, J., Glocker, B., Klinder, T., Li, S., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 3–12. [Google Scholar]
  68. Masoudi, S.; Mehralivand, S.; Harmon, S.; Walker, S.; Pinto, P.A.; Wood, B.J.; Citrin, D.E.; Karzai, F.; Gulley, J.L.; Madan, R.A.; et al. Artificial intelligence assisted bone lesion detection and classification in computed tomography scans of prostate cancer patients. J. Clin. Oncol. 2020, 38, e17567. [Google Scholar] [CrossRef]
  69. Fan, X.; Zhang, X.; Zhang, Z.; Jiang, Y. Deep Learning-Based Identification of Spinal Metastasis in Lung Cancer Using Spectral CT Images. Sci. Program. 2021, 2021, 2779390. [Google Scholar] [CrossRef]
  70. Noguchi, S.; Nishio, M.; Sakamoto, R.; Yakami, M.; Fujimoto, K.; Emoto, Y.; Kubo, T.; Iizuka, Y.; Nakagomi, K.; Miyasa, K.; et al. Deep learning-based algorithm improved radiologists’ performance in bone metastases detection on CT. Eur. Radiol. 2022, 32, 7976–7987. [Google Scholar] [CrossRef]
  71. Hoshiai, S.; Hanaoka, S.; Masumoto, T.; Nomura, Y.; Mori, K.; Okamoto, Y.; Saida, T.; Ishiguro, T.; Sakai, M.; Nakajima, T. Effectiveness of temporal subtraction computed tomography images using deep learning in detecting vertebral bone metastases. Eur. J. Radiol. 2022, 154, 110445. [Google Scholar] [CrossRef]
  72. Musa Aguiar, P.; Zarantonello, P.; Aparisi Gómez, M.P. Differentiation between Osteoporotic and Neoplastic Vertebral Fractures: State of the Art and Future Perspectives. Curr. Med. Imaging 2022, 18, 187–207. [Google Scholar] [CrossRef]
  73. Gilberg, L.; Teodorescu, B.; Maerkisch, L.; Baumgart, A.; Ramaesh, R.; Ataide, E.; Koç, A. Deep Learning Enhances Radiologists’ Detection of Potential Spinal Malignancies in CT Scans. Appl. Sci. 2023, 13, 8140. [Google Scholar] [CrossRef]
  74. Huo, T.; Xie, Y.; Fang, Y.; Wang, Z.; Liu, P.; Duan, Y.; Zhang, J.; Wang, H.; Xue, M.; Liu, S.; et al. Deep learning-based algorithm improves radiologists’ performance in lung cancer bone metastases detection on computed tomography. Front. Oncol. 2023, 13, 1125637. [Google Scholar] [CrossRef] [PubMed]
  75. Koike, Y.; Yui, M.; Nakamura, S.; Yoshida, A.; Takegawa, H.; Anetai, Y.; Hirota, K.; Tanigawa, N. Artificial intelligence-aided lytic spinal bone metastasis classification on CT scans. Int. J. Comput. Assist. Radiol. Surg. 2023, 18, 1867–1874. [Google Scholar] [CrossRef] [PubMed]
  76. Chmelik, J.; Jakubicek, R.; Walek, P.; Jan, J.; Ourednicek, P.; Lambert, L.; Amadori, E.; Gavelli, G. Deep convolutional neural network-based segmentation and classification of difficult to define metastatic spinal lesions in 3D CT data. Med. Image Anal. 2018, 49, 76–88. [Google Scholar] [CrossRef] [PubMed]
  77. Li, Y.; Zhang, Y.; Zhang, E.; Chen, Y.; Wang, Q.; Liu, K.; Yu, H.J.; Yuan, H.; Lang, N.; Su, M.Y. Differential diagnosis of benign and malignant vertebral fracture on CT using deep learning. Eur. Radiol. 2021, 31, 9612–9619. [Google Scholar] [CrossRef]
  78. Masoudi, S.; Mehralivand, S.; Harmon, S.A.; Lay, N.; Lindenberg, L.; Mena, E.; Pinto, P.A.; Citrin, D.E.; Gulley, J.L.; Wood, B.J.; et al. Deep Learning Based Staging of Bone Lesions From Computed Tomography Scans. IEEE Access 2021, 9, 87531–87542. [Google Scholar] [CrossRef]
  79. Hallinan, J.T.; Zhu, L.; Zhang, W.; Kuah, T.; Lim, D.S.; Low, X.Z.; Cheng, A.J.L.; Eide, S.E.; Ong, H.Y.; Muhamat Nor, F.E.; et al. Deep Learning Model for Grading Metastatic Epidural Spinal Cord Compression on Staging CT. Cancers 2022, 14, 3219. [Google Scholar] [CrossRef]
  80. Naseri, H.; Skamene, S.; Tolba, M.; Faye, M.D.; Ramia, P.; Khriguian, J.; Patrick, H.; Andrade Hernandez, A.X.; David, M.; Kildea, J. Radiomics-based machine learning models to distinguish between metastatic and healthy bone using lesion-center-based geometric regions of interest. Sci. Rep. 2022, 12, 9866. [Google Scholar] [CrossRef]
  81. Park, T.; Yoon, M.A.; Cho, Y.C.; Ham, S.J.; Ko, Y.; Kim, S.; Jeong, H.; Lee, J. Automated segmentation of the fractured vertebrae on CT and its applicability in a radiomics model to predict fracture malignancy. Sci. Rep. 2022, 12, 6735. [Google Scholar] [CrossRef]
  82. Wang, Q.; Chen, Y.; Qin, S.; Liu, X.; Liu, K.; Xin, P.; Zhao, W.; Yuan, H.; Lang, N. Prognostic Value and Quantitative CT Analysis in RANKL Expression of Spinal GCTB in the Denosumab Era: A Machine Learning Approach. Cancers 2022, 14, 5201. [Google Scholar] [CrossRef]
  83. Wang, Q.; Zhang, Y.; Zhang, E.; Xing, X.; Chen, Y.; Nie, K.; Yuan, H.; Su, M.Y.; Lang, N. A Multiparametric Method Based on Clinical and CT-Based Radiomics to Predict the Expression of p53 and VEGF in Patients with Spinal Giant Cell Tumor of Bone. Front. Oncol. 2022, 12, 894696. [Google Scholar] [CrossRef]
  84. Hallinan, J.; Zhu, L.; Tan, H.W.N.; Hui, S.J.; Lim, X.; Ong, B.W.L.; Ong, H.Y.; Eide, S.E.; Cheng, A.J.L.; Ge, S.; et al. A deep learning-based technique for the diagnosis of epidural spinal cord compression on thoracolumbar CT. Eur. Spine J. 2023, 32, 3815–3824. [Google Scholar] [CrossRef]
  85. Hallinan, J.T.P.D.; Zhu, L.; Zhang, W.; Ge, S.; Muhamat Nor, F.E.; Ong, H.Y.; Eide, S.E.; Cheng, A.J.L.; Kuah, T.; Lim, D.S.W.; et al. Deep learning assessment compared to radiologist reporting for metastatic spinal cord compression on CT. Front. Oncol. 2023, 13, 1151073. [Google Scholar] [CrossRef]
  86. Duan, S.; Hua, Y.; Cao, G.; Hu, J.; Cui, W.; Zhang, D.; Xu, S.; Rong, T.; Liu, B. Differential diagnosis of benign and malignant vertebral compression fractures: Comparison and correlation of radiomics and deep learning frameworks based on spinal CT and clinical characteristics. Eur. J. Radiol. 2023, 165, 110899. [Google Scholar] [CrossRef]
  87. Gui, C.; Chen, X.; Sheikh, K.; Mathews, L.; Lo, S.L.; Lee, J.; Khan, M.A.; Sciubba, D.M.; Redmond, K.J. Radiomic modeling to predict risk of vertebral compression fracture after stereotactic body radiation therapy for spinal metastases. J. Neurosurg. Spine 2021, 36, 294–302. [Google Scholar] [CrossRef] [PubMed]
  88. Wang, Q.; Zhang, Y.; Zhang, E.; Xing, X.; Chen, Y.; Su, M.Y.; Lang, N. Prediction of the early recurrence in spinal giant cell tumor of bone using radiomics of preoperative CT: Long-term outcome of 62 consecutive patients. J. Bone Oncol. 2021, 27, 100354. [Google Scholar] [CrossRef] [PubMed]
  89. Massaad, E.; Bridge, C.P.; Kiapour, A.; Fourman, M.S.; Duvall, J.B.; Connolly, I.D.; Hadzipasic, M.; Shankar, G.M.; Andriole, K.P.; Rosenthal, M.; et al. Evaluating frailty, mortality, and complications associated with metastatic spine tumor surgery using machine learning-derived body composition analysis. J. Neurosurg. Spine 2022, 37, 263–273. [Google Scholar] [CrossRef]
  90. Seol, Y.; Song, J.H.; Choi, K.H.; Lee, Y.K.; Choi, B.O.; Kang, Y.N. Predicting vertebral compression fracture prior to spinal SBRT using radiomics from planning CT. Eur. Spine J. 2023, 33, 3221–3229. [Google Scholar] [CrossRef]
  91. Delrieu, L.; Blanc, D.; Bouhamama, A.; Reyal, F.; Pilleul, F.; Racine, V.; Hamy, A.S.; Crochet, H.; Marchal, T.; Heudel, P.E. Automatic deep learning method for third lumbar selection and body composition evaluation on CT scans of cancer patients. Front. Nucl. Med. 2024, 3, 1292676. [Google Scholar] [CrossRef]
  92. Khalid, S.I.; Massaad, E.; Kiapour, A.; Bridge, C.P.; Rigney, G.; Burrows, A.; Shim, J.; de la Garza Ramos, R.; Tobert, D.G.; Schoenfeld, A.J.; et al. Machine learning-based detection of sarcopenic obesity and association with adverse outcomes in patients undergoing surgical treatment for spinal metastases. J. Neurosurg. Spine 2024, 40, 291–300. [Google Scholar] [CrossRef]
  93. Arends, S.R.S.; Savenije, M.H.F.; Eppinga, W.S.C.; van der Velden, J.M.; van den Berg, C.A.T.; Verhoeff, J.J.C. Clinical utility of convolutional neural networks for treatment planning in radiotherapy for spinal metastases. Phys. Imaging Radiat. Oncol. 2022, 21, 42–47. [Google Scholar] [CrossRef]
  94. Netherton, T.J.; Nguyen, C.; Cardenas, C.E.; Chung, C.; Klopp, A.H.; Colbert, L.E.; Rhee, D.J.; Peterson, C.B.; Howell, R.; Balter, P.; et al. An Automated Treatment Planning Framework for Spinal Radiation Therapy and Vertebral-Level Second Check. Int. J. Radiat. Oncol. Biol. Phys. 2022, 114, 516–528. [Google Scholar] [CrossRef] [PubMed]
  95. Hernandez, S.; Nguyen, C.; Parkes, J.; Burger, H.; Rhee, D.J.; Netherton, T.; Mumme, R.; Vega, J.G.; Duryea, J.; Leone, A.; et al. Automating the treatment planning process for 3D-conformal pediatric craniospinal irradiation therapy. Pediatr. Blood Cancer 2023, 70, e30164. [Google Scholar] [CrossRef]
  96. Mok, S.; Chu, E.C. The Importance of Early Detection of Spinal Tumors Through Magnetic Resonance Imaging in Chiropractic Practices. Cureus 2024, 16, e51440. [Google Scholar] [CrossRef] [PubMed]
  97. Costelloe, C.M.; Rohren, E.M.; Madewell, J.E.; Hamaoka, T.; Theriault, R.L.; Yu, T.-K.; Lewis, V.O.; Ma, J.; Stafford, R.J.; Tari, A.M.; et al. Imaging bone metastases in breast cancer: Techniques and recommendations for diagnosis. Lancet Oncol. 2009, 10, 606–614. [Google Scholar] [CrossRef]
  98. Curtin, M.; Piggott, R.P.; Murphy, E.P.; Munigangaiah, S.; Baker, J.F.; McCabe, J.P.; Devitt, A. Spinal Metastatic Disease: A Review of the Role of the Multidisciplinary Team. Orthop. Surg. 2017, 9, 145–151. [Google Scholar] [CrossRef]
  99. Tomita, K.; Kawahara, N.; Kobayashi, T.; Yoshida, A.; Murakami, H.; Akamaru, T. Surgical strategy for spinal metastases. Spine 2001, 26, 298–306. [Google Scholar] [CrossRef]
  100. Delank, K.S.; Wendtner, C.; Eich, H.T.; Eysel, P. The treatment of spinal metastases. Dtsch. Arztebl. Int. 2011, 108, 71–79. [Google Scholar] [CrossRef]
  101. Barzilai, O.; Fisher, C.G.; Bilsky, M.H. State of the Art Treatment of Spinal Metastatic Disease. Neurosurgery 2018, 82, 757–769. [Google Scholar] [CrossRef]
  102. Clemons, M.; Gelmon, K.A.; Pritchard, K.I.; Paterson, A.H. Bone-targeted agents and skeletal-related events in breast cancer patients with bone metastases: The state of the art. Curr. Oncol. 2012, 19, 259–268. [Google Scholar] [CrossRef] [PubMed]
  103. Bilsky, M.H.; Lis, E.; Raizer, J.; Lee, H.; Boland, P. The diagnosis and treatment of metastatic spinal tumor. Oncologist 1999, 4, 459–469. [Google Scholar] [CrossRef]
  104. Chen, B.; Cai, L.; Zhou, F. Management of acute spinal cord compression in multiple myeloma. Crit. Rev. Oncol. Hematol. 2021, 160, 103205. [Google Scholar] [CrossRef]
  105. Cardobi, N.; Dal Palù, A.; Pedrini, F.; Beleù, A.; Nocini, R.; De Robertis, R.; Ruzzenente, A.; Salvia, R.; Montemezzi, S.; D’Onofrio, M. An Overview of Artificial Intelligence Applications in Liver and Pancreatic Imaging. Cancers 2021, 13, 2162. [Google Scholar] [CrossRef]
  106. Ong, W.; Zhu, L.; Zhang, W.; Kuah, T.; Lim, D.S.W.; Low, X.Z.; Thian, Y.L.; Teo, E.C.; Tan, J.H.; Kumar, N.; et al. Application of Artificial Intelligence Methods for Imaging of Spinal Metastasis. Cancers 2022, 14, 4025. [Google Scholar] [CrossRef] [PubMed]
  107. Hallinan, J.T.P.D.; Ge, S.; Zhu, L.; Zhang, W.; Lim, Y.T.; Thian, Y.L.; Jagmohan, P.; Kuah, T.; Lim, D.S.W.; Low, X.Z.; et al. Diagnostic Accuracy of CT for Metastatic Epidural Spinal Cord Compression. Cancers 2022, 14, 4231. [Google Scholar] [CrossRef]
  108. O’Connor, S.D.; Yao, J.; Summers, R.M. Lytic metastases in thoracolumbar spine: Computer-aided detection at CT—Preliminary study. Radiology 2007, 242, 811–816. [Google Scholar] [CrossRef] [PubMed]
  109. Wang, J.; Fang, Z.; Lang, N.; Yuan, H.; Su, M.Y.; Baldi, P. A multi-resolution approach for spinal metastasis detection using deep Siamese neural networks. Comput. Biol. Med. 2017, 84, 137–146. [Google Scholar] [CrossRef] [PubMed]
  110. Chang, C.Y.; Buckless, C.; Yeh, K.J.; Torriani, M. Automated detection and segmentation of sclerotic spinal lesions on body CTs using a deep convolutional neural network. Skeletal Radiol. 2022, 51, 391–399. [Google Scholar] [CrossRef]
  111. Yoen, H.; Jang, M.J.; Yi, A.; Moon, W.K.; Chang, J.M. Artificial Intelligence for Breast Cancer Detection on Mammography: Factors Related to Cancer Detection. Acad. Radiol. 2024, 31, P2239–P2247. [Google Scholar] [CrossRef]
  112. Dembrower, K.; Crippa, A.; Colón, E.; Eklund, M.; Strand, F. Artificial intelligence for breast cancer detection in screening mammography in Sweden: A prospective, population-based, paired-reader, non-inferiority study. Lancet Digit. Health 2023, 5, e703–e711. [Google Scholar] [CrossRef]
  113. Díaz, O.; Rodríguez-Ruíz, A.; Sechopoulos, I. Artificial Intelligence for breast cancer detection: Technology, challenges, and prospects. Eur. J. Radiol. 2024, 175, 111457. [Google Scholar] [CrossRef]
  114. Kim, Y.S.; Jang, M.J.; Lee, S.H.; Kim, S.Y.; Ha, S.M.; Kwon, B.R.; Moon, W.K.; Chang, J.M. Use of Artificial Intelligence for Reducing Unnecessary Recalls at Screening Mammography: A Simulation Study. Korean J. Radiol. 2022, 23, 1241–1250. [Google Scholar] [CrossRef] [PubMed]
  115. Lee, S.E.; Kim, G.R.; Yoon, J.H.; Han, K.; Son, W.J.; Shin, H.J.; Moon, H.J. Artificial intelligence assistance for women who had spot compression view: Reducing recall rates for digital mammography. Acta Radiol. 2023, 64, 1808–1815. [Google Scholar] [CrossRef] [PubMed]
  116. Zhang, Y.; Jiang, B.; Zhang, L.; Greuter, M.J.W.; de Bock, G.H.; Zhang, H.; Xie, X. Lung Nodule Detectability of Artificial Intelligence-assisted CT Image Reading in Lung Cancer Screening. Curr. Med. Imaging. 2022, 18, 327–334. [Google Scholar] [CrossRef] [PubMed]
  117. Hendrix, W.; Hendrix, N.; Scholten, E.T.; Mourits, M.; Trap-de Jong, J.; Schalekamp, S.; Korst, M.; van Leuken, M.; van Ginneken, B.; Prokop, M.; et al. Deep learning for the detection of benign and malignant pulmonary nodules in non-screening chest CT scans. Commun. Med. 2023, 3, 156. [Google Scholar] [CrossRef] [PubMed]
  118. Goncalves, S.; Fong, P.C.; Blokhina, M. Artificial intelligence for early diagnosis of lung cancer through incidental nodule detection in low- and middle-income countries-acceleration during the COVID-19 pandemic but here to stay. Am. J. Cancer Res. 2022, 12, 1–16. [Google Scholar] [PubMed]
  119. Twilt, J.J.; van Leeuwen, K.G.; Huisman, H.J.; Fütterer, J.J.; de Rooij, M. Artificial Intelligence Based Algorithms for Prostate Cancer Classification and Detection on Magnetic Resonance Imaging: A Narrative Review. Diagnostics 2021, 11, 959. [Google Scholar] [CrossRef]
  120. Bhattacharya, I.; Khandwala, Y.S.; Vesal, S.; Shao, W.; Yang, Q.; Soerensen, S.J.C.; Fan, R.E.; Ghanouni, P.; Kunder, C.A.; Brooks, J.D.; et al. A review of artificial intelligence in prostate cancer detection on imaging. Ther. Adv. Urol. 2022, 14, 17562872221128791. [Google Scholar] [CrossRef]
  121. Li, H.; Lee, C.H.; Chia, D.; Lin, Z.; Huang, W.; Tan, C.H. Machine Learning in Prostate MRI for Prostate Cancer: Current Status and Future Opportunities. Diagnostics 2022, 12, 289. [Google Scholar] [CrossRef]
  122. Schalekamp, S.; van Leeuwen, K.; Calli, E.; Murphy, K.; Rutten, M.; Geurts, B.; Peters-Bax, L.; van Ginneken, B.; Prokop, M. Performance of AI to exclude normal chest radiographs to reduce radiologists’ workload. Eur. Radiol. 2024, 1–9. [Google Scholar] [CrossRef]
  123. Raya-Povedano, J.L.; Romero-Martín, S.; Elías-Cabot, E.; Gubern-Mérida, A.; Rodríguez-Ruiz, A.; Álvarez-Benito, M. AI-based Strategies to Reduce Workload in Breast Cancer Screening with Mammography and Tomosynthesis: A Retrospective Evaluation. Radiology 2021, 300, 57–65. [Google Scholar] [CrossRef] [PubMed]
  124. Seah, J.; Brady, Z.; Ewert, K.; Law, M. Artificial intelligence in medical imaging: Implications for patient radiation safety. Br. J. Radiol. 2021, 94, 20210406. [Google Scholar] [CrossRef]
  125. van Leeuwen, K.G.; de Rooij, M.; Schalekamp, S.; van Ginneken, B.; Rutten, M. How does artificial intelligence in radiology improve efficiency and health outcomes? Pediatr. Radiol. 2022, 52, 2087–2093. [Google Scholar] [CrossRef]
  126. Hu, H.T.; Wang, W.; Chen, L.D.; Ruan, S.M.; Chen, S.L.; Li, X.; Lu, M.D.; Xie, X.Y.; Kuang, M. Artificial intelligence assists identifying malignant versus benign liver lesions using contrast-enhanced ultrasound. J. Gastroenterol. Hepatol. 2021, 36, 2875–2883. [Google Scholar] [CrossRef]
  127. Anaya-Isaza, A.; Mera-Jiménez, L.; Verdugo-Alejo, L.; Sarasti, L. Optimizing MRI-based brain tumor classification and detection using AI: A comparative analysis of neural networks, transfer learning, data augmentation, and the cross-transformer network. Eur. J. Radiol. Open 2023, 10, 100484. [Google Scholar] [CrossRef]
  128. Wan, Y.L.; Wu, P.W.; Huang, P.C.; Tsay, P.K.; Pan, K.T.; Trang, N.N.; Chuang, W.Y.; Wu, C.Y.; Lo, S.B. The Use of Artificial Intelligence in the Differentiation of Malignant and Benign Lung Nodules on Computed Tomograms Proven by Surgical Pathology. Cancers 2020, 12, 2211. [Google Scholar] [CrossRef]
  129. Ong, W.; Zhu, L.; Tan, Y.L.; Teo, E.C.; Tan, J.H.; Kumar, N.; Vellayappan, B.A.; Ooi, B.C.; Quek, S.T.; Makmur, A.; et al. Application of Machine Learning for Differentiating Bone Malignancy on Imaging: A Systematic Review. Cancers 2023, 15, 1837. [Google Scholar] [CrossRef] [PubMed]
  130. Filograna, L.; Lenkowicz, J.; Cellini, F.; Dinapoli, N.; Manfrida, S.; Magarelli, N.; Leone, A.; Colosimo, C.; Valentini, V. Identification of the most significant magnetic resonance imaging (MRI) radiomic features in oncological patients with vertebral bone marrow metastatic disease: A feasibility study. Radiol. Med. 2019, 124, 50–57. [Google Scholar] [CrossRef] [PubMed]
  131. Chung, W.J.; Chung, H.W.; Shin, M.J.; Lee, S.H.; Lee, M.H.; Lee, J.S.; Kim, M.J.; Lee, W.K. MRI to differentiate benign from malignant soft-tissue tumours of the extremities: A simplified systematic imaging approach using depth, size and heterogeneity of signal intensity. Br. J. Radiol. 2012, 85, e831–e836. [Google Scholar] [CrossRef]
  132. van Rijswijk, C.S.; Kunz, P.; Hogendoorn, P.C.; Taminiau, A.H.; Doornbos, J.; Bloem, J.L. Diffusion-weighted MRI in the characterization of soft-tissue tumors. J. Magn. Reson. Imaging 2002, 15, 302–307. [Google Scholar] [CrossRef]
  133. Vande Berg, B.C.; Malghem, J.; Lecouvet, F.E.; Maldague, B. Classification and detection of bone marrow lesions with magnetic resonance imaging. Skeletal Radiol. 1998, 27, 529–545. [Google Scholar] [CrossRef]
  134. Davies, A.M.; Sundaram, M.; James, S.L. Imaging of Bone Tumors and Tumor-like Lesions: Techniques and Applications; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  135. Liu, J.; Guo, W.; Zeng, P.; Geng, Y.; Liu, Y.; Ouyang, H.; Lang, N.; Yuan, H. Vertebral MRI-based radiomics model to differentiate multiple myeloma from metastases: Influence of features number on logistic regression model performance. Eur. Radiol. 2022, 32, 572–581. [Google Scholar] [CrossRef]
  136. Xiong, X.; Wang, J.; Hu, S.; Dai, Y.; Zhang, Y.; Hu, C. Differentiating Between Multiple Myeloma and Metastasis Subtypes of Lumbar Vertebra Lesions Using Machine Learning-Based Radiomics. Front. Oncol. 2021, 11, 601699. [Google Scholar] [CrossRef] [PubMed]
  137. Yin, P.; Mao, N.; Zhao, C.; Wu, J.; Chen, L.; Hong, N. A Triple-Classification Radiomics Model for the Differentiation of Primary Chordoma, Giant Cell Tumor, and Metastatic Tumor of Sacrum Based on T2-Weighted and Contrast-Enhanced T1-Weighted MRI. J. Magn. Reson. Imaging 2019, 49, 752–759. [Google Scholar] [CrossRef]
  138. Sciubba, D.M.; Petteys, R.J.; Dekutoski, M.B.; Fisher, C.G.; Fehlings, M.G.; Ondra, S.L.; Rhines, L.D.; Gokaslan, Z.L. Diagnosis and management of metastatic spine disease. A review. J. Neurosurg. Spine 2010, 13, 94–108. [Google Scholar] [CrossRef] [PubMed]
  139. Andreula, C.; Murrone, M. Metastatic disease of the spine. Eur. Radiol. 2005, 15, 627–632. [Google Scholar] [CrossRef]
  140. Mauch, J.T.; Carr, C.M.; Cloft, H.; Diehn, F.E. Review of the Imaging Features of Benign Osteoporotic and Malignant Vertebral Compression Fractures. AJNR Am. J. Neuroradiol. 2018, 39, 1584–1592. [Google Scholar] [CrossRef] [PubMed]
  141. Abowarda, M.H.; Abdel-Rahman, H.M.; Taha, M.M. Differentiation of acute osteoporotic from malignant vertebral compression fractures with conventional MRI and diffusion MR imaging. Egypt. J. Radiol. Nucl. Med. 2017, 48, 207–213. [Google Scholar] [CrossRef]
  142. Uetani, M.; Hashmi, R.; Hayashi, K. Malignant and benign compression fractures: Differentiation and diagnostic pitfalls on MRI. Clin. Radiol. 2004, 59, 124–131. [Google Scholar] [CrossRef]
  143. Tang, X. The role of artificial intelligence in medical imaging research. BJR Open 2020, 2, 20190031. [Google Scholar] [CrossRef]
  144. Laredo, J.D.; Lakhdari, K.; Bellaïche, L.; Hamze, B.; Janklewicz, P.; Tubiana, J.M. Acute vertebral collapse: CT findings in benign and malignant nontraumatic cases. Radiology 1995, 194, 41–48. [Google Scholar] [CrossRef]
  145. Kubota, T.; Yamada, K.; Ito, H.; Kizu, O.; Nishimura, T. High-resolution imaging of the spine using multidetector-row computed tomography: Differentiation between benign and malignant vertebral compression fractures. J. Comput. Assist. Tomogr. 2005, 29, 712–719. [Google Scholar] [CrossRef]
  146. Perry, J.; Chambers, A.; Laperriere, N. Systematic Review of the Diagnosis and Management of Malignant Extradural Spinal Cord Compression: The Cancer Care Ontario Practice Guidelines Initiative‘s Neuro-Oncology Disease Site Group. J. Clin. Oncol. Off. J. Am. Soc. Clin. Oncol. 2005, 23, 2028–2037. [Google Scholar] [CrossRef]
  147. Liu, Z.; Duan, T.; Zhang, Y.; Weng, S.; Xu, H.; Ren, Y.; Zhang, Z.; Han, X. Radiogenomics: A key component of precision cancer medicine. Br. J. Cancer 2023, 129, 741–753. [Google Scholar] [CrossRef] [PubMed]
  148. Shui, L.; Ren, H.; Yang, X.; Li, J.; Chen, Z.; Yi, C.; Zhu, H.; Shui, P. The Era of Radiogenomics in Precision Medicine: An Emerging Approach to Support Diagnosis, Treatment Decisions, and Prognostication in Oncology. Front. Oncol. 2020, 10, 570465. [Google Scholar] [CrossRef] [PubMed]
  149. Rabasco Meneghetti, A.; Zwanenburg, A.; Linge, A.; Lohaus, F.; Grosser, M.; Baretton, G.B.; Kalinauskaite, G.; Tinhofer, I.; Guberina, M.; Stuschke, M.; et al. Integrated radiogenomics analyses allow for subtype classification and improved outcome prognosis of patients with locally advanced HNSCC. Sci. Rep. 2022, 12, 16755. [Google Scholar] [CrossRef] [PubMed]
  150. Singh, G.; Manjila, S.; Sakla, N.; True, A.; Wardeh, A.H.; Beig, N.; Vaysberg, A.; Matthews, J.; Prasanna, P.; Spektor, V. Radiomics and radiogenomics in gliomas: A contemporary update. Br. J. Cancer 2021, 125, 641–657. [Google Scholar] [CrossRef]
  151. Ferrara, N. Vascular endothelial growth factor: Basic science and clinical progress. Endocr. Rev. 2004, 25, 581–611. [Google Scholar] [CrossRef]
  152. Quattrini, I.; Pollino, S.; Pazzaglia, L.; Conti, A.; Novello, C.; Ferrari, C.; Pignotti, E.; Picci, P.; Benassi, M.S. Prognostic role of nuclear factor/IB and bone remodeling proteins in metastatic giant cell tumor of bone: A retrospective study. J. Orthop. Res. 2015, 33, 1205–1211. [Google Scholar] [CrossRef]
  153. Li, B.; Li, J.; Miao, W.; Zhao, Y.; Jiao, J.; Wu, Z.; Yang, X.; Wei, H.; Xiao, J. Prognostic Analysis of Clinical and Immunohistochemical Factors for Patients with Spinal Schwannoma. World Neurosurg. 2018, 120, e617–e627. [Google Scholar] [CrossRef]
  154. Ghani, S.A.; Wan Ismail, W.F.; Md Salleh, M.S.; Yahaya, S.; Syahrul Fitri, Z.M. The Values of Receptor Activator Nuclear Kappa-B Ligand Expression in Stage III Giant Cell Tumor of the Bone. Indian J. Orthop. 2018, 52, 31–34. [Google Scholar] [CrossRef] [PubMed]
  155. Tsukamoto, S.; Mavrogenis, A.F.; Kido, A.; Errani, C. Current Concepts in the Treatment of Giant Cell Tumors of Bone. Cancers 2021, 13, 3647. [Google Scholar] [CrossRef]
  156. Bukata, S.V.; Blay, J.Y.; Rutkowski, P.; Skubitz, K.; Henshaw, R.; Seeger, L.; Dai, T.; Jandial, D.; Chawla, S. Denosumab Treatment for Giant Cell Tumor of the Spine Including the Sacrum. Spine 2021, 46, 277–284. [Google Scholar] [CrossRef] [PubMed]
  157. Darvish, L.; Bahreyni-Toossi, M.-T.; Roozbeh, N.; Azimian, H. The role of radiogenomics in the diagnosis of breast cancer: A systematic review. Egypt. J. Med. Hum. Genet. 2022, 23, 99. [Google Scholar] [CrossRef]
  158. Zhang, B.; Shi, H.; Wang, H. Machine Learning and AI in Cancer Prognosis, Prediction, and Treatment Selection: A Critical Approach. J. Multidiscip. Healthc. 2023, 16, 1779–1791. [Google Scholar] [CrossRef]
  159. Bera, K.; Braman, N.; Gupta, A.; Velcheti, V.; Madabhushi, A. Predicting cancer outcomes with radiomics and artificial intelligence in radiology. Nat. Rev. Clin. Oncol. 2022, 19, 132–146. [Google Scholar] [CrossRef]
  160. Tan, H.; Gao, X.; Li, X.; Huang, Y.; Cao, Q.; Wan, T. Sarcopenia in Patients With Spinal Metastasis: A Systematic Review and Meta-Analysis of Retrospective Cohort Studies. Front. Oncol. 2022, 12, 864501. [Google Scholar] [CrossRef]
  161. Nishigori, T.; Obama, K.; Sakai, Y. Assessment of body composition and impact of sarcopenia and sarcopenic obesity in patients with gastric cancer. Transl. Gastroenterol. Hepatol. 2020, 5, 22. [Google Scholar] [CrossRef]
  162. Bongers, M.E.R.; Groot, O.Q.; Buckless, C.G.; Kapoor, N.D.; Twining, P.K.; Schwab, J.H.; Torriani, M.; Bredella, M.A. Body composition predictors of mortality on computed tomography in patients with spinal metastases undergoing surgical treatment. Spine J. 2022, 22, 595–604. [Google Scholar] [CrossRef]
  163. Xie, W.-Q.; Xiao, G.-L.; Hu, P.-W.; He, Y.-Q.; Lv, S.; Xiao, W.-F. Possible sarcopenia: Early screening and intervention-narrative review. Ann. Palliat. Med. 2020, 9, 4283–4293. [Google Scholar] [CrossRef]
  164. Pickhardt, P.J.; Summers, R.M.; Garrett, J.W. Automated CT-Based Body Composition Analysis: A Golden Opportunity. Korean J. Radiol. 2021, 22, 1934–1937. [Google Scholar] [CrossRef] [PubMed]
  165. Wall, J.; Paul, M.; Phillips, B.E. Nutritional interventions in prehabilitation for cancer surgery. Curr. Opin. Clin. Nutr. Metab. Care 2023, 26, 497–507. [Google Scholar] [CrossRef] [PubMed]
  166. Cepeda, S.; Luppino, L.T.; Pérez-Núñez, A.; Solheim, O.; García-García, S.; Velasco-Casares, M.; Karlberg, A.; Eikenes, L.; Sarabia, R.; Arrese, I.; et al. Predicting Regions of Local Recurrence in Glioblastomas Using Voxel-Based Radiomic Features of Multiparametric Postoperative MRI. Cancers 2023, 15, 1894. [Google Scholar] [CrossRef]
  167. Libling, W.A.; Korn, R.; Weiss, G.J. Review of the use of radiomics to assess the risk of recurrence in early-stage non-small cell lung cancer. Transl. Lung Cancer Res. 2023, 12, 1575–1589. [Google Scholar] [CrossRef]
  168. Dixon, D.; Sattar, H.; Moros, N.; Kesireddy, S.R.; Ahsan, H.; Lakkimsetti, M.; Fatima, M.; Doshi, D.; Sadhu, K.; Junaid Hassan, M. Unveiling the Influence of AI Predictive Analytics on Patient Outcomes: A Comprehensive Narrative Review. Cureus 2024, 16, e59954. [Google Scholar] [CrossRef] [PubMed]
  169. Liao, J.; Li, X.; Gan, Y.; Han, S.; Rong, P.; Wang, W.; Li, W.; Zhou, L. Artificial intelligence assists precision medicine in cancer treatment. Front. Oncol. 2022, 12, 998222. [Google Scholar] [CrossRef]
  170. Bhinder, B.; Gilvary, C.; Madhukar, N.S.; Elemento, O. Artificial Intelligence in Cancer Research and Precision Medicine. Cancer Discov. 2021, 11, 900–915. [Google Scholar] [CrossRef]
  171. Gong, X.-Q.; Liu, N.; Tao, Y.-Y.; Li, L.; Li, Z.-M.; Yang, L.; Zhang, X.-M. Radiomics models based on multisequence MRI for predicting PD-1/PD-L1 expression in hepatocellular carcinoma. Sci. Rep. 2023, 13, 7710. [Google Scholar] [CrossRef]
  172. Li, Q.; Han, J.; Yang, Y.; Chen, Y. PD-1/PD-L1 checkpoint inhibitors in advanced hepatocellular carcinoma immunotherapy. Front. Immunol. 2022, 13, 1070961. [Google Scholar] [CrossRef]
  173. Bove, S.; Fanizzi, A.; Fadda, F.; Comes, M.C.; Catino, A.; Cirillo, A.; Cristofaro, C.; Montrone, M.; Nardone, A.; Pizzutilo, P.; et al. A CT-based transfer learning approach to predict NSCLC recurrence: The added-value of peritumoral region. PLoS ONE 2023, 18, e0285188. [Google Scholar] [CrossRef]
  174. Yoo, G.S.; Park, H.C.; Yu, J.I.; Lim, D.H.; Cho, W.K.; Lee, E.; Jung, S.H.; Han, Y.; Kim, E.S.; Lee, S.H.; et al. Stereotactic ablative body radiotherapy for spinal metastasis from hepatocellular carcinoma: Its oncologic outcomes and risk of vertebral compression fracture. Oncotarget 2017, 8, 72860–72871. [Google Scholar] [CrossRef] [PubMed]
  175. Al Taha, K.; Lauper, N.; Bauer, D.E.; Tsoupras, A.; Tessitore, E.; Biver, E.; Dominguez, D.E. Multidisciplinary and Coordinated Management of Osteoporotic Vertebral Compression Fractures: Current State of the Art. J. Clin. Med. 2024, 13, 930. [Google Scholar] [CrossRef]
  176. Hammoudeh, L.; Abunimer, A.M.; Lee, H.Y.; Dee, E.C.; Brennan, S.V.; Yaguang, P.; Shin, K.Y.; Chen, Y.H.; Huynh, M.A.; Spektor, A.; et al. Spinal Cord Delineation Based on Computed Tomography Myelogram Versus T2 Magnetic Resonance Imaging in Spinal Stereotactic Body Radiation Therapy. Adv. Radiat. Oncol. 2023, 8, 101158. [Google Scholar] [CrossRef] [PubMed]
  177. Sacino, A.N.; Chen, H.; Sahgal, A.; Bettegowda, C.; Rhines, L.D.; Maralani, P.; Redmond, K.J. Stereotactic body radiation therapy for spinal metastases: A new standard of care. Neuro-Oncology 2024, 26, S76–S87. [Google Scholar] [CrossRef]
  178. Kuah, T.; Vellayappan, B.; Makmur, A.; Nair, S.; Song, J.; Tan, J.H.; Kumar, N.; Quek, S.; Hallinan, J. State-of-the-Art Imaging Techniques in Metastatic Spinal Cord Compression. Cancers 2022, 14, 3289. [Google Scholar] [CrossRef] [PubMed]
  179. McGrath, H.; Li, P.; Dorent, R.; Bradford, R.; Saeed, S.; Bisdas, S.; Ourselin, S.; Shapey, J.; Vercauteren, T. Manual segmentation versus semi-automated segmentation for quantifying vestibular schwannoma volume on MRI. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1445–1455. [Google Scholar] [CrossRef]
  180. Matoska, T.; Patel, M.; Liu, H.; Beriwal, S. Review of Deep Learning Based Autosegmentation for Clinical Target Volume: Current Status and Future Directions. Adv. Radiat. Oncol. 2024, 9, 101470. [Google Scholar] [CrossRef]
  181. Lastrucci, A.; Wandael, Y.; Ricci, R.; Maccioni, G.; Giansanti, D. The Integration of Deep Learning in Radiotherapy: Exploring Challenges, Opportunities, and Future Directions through an Umbrella Review. Diagnostics 2024, 14, 939. [Google Scholar] [CrossRef]
  182. Samarasinghe, G.; Jameson, M.; Vinod, S.; Field, M.; Dowling, J.; Sowmya, A.; Holloway, L. Deep learning for segmentation in radiation therapy planning: A review. J. Med. Imaging Radiat. Oncol. 2021, 65, 578–595. [Google Scholar] [CrossRef]
  183. Wang, M.; Zhang, Q.; Lam, S.; Cai, J.; Yang, R. A Review on Application of Deep Learning Algorithms in External Beam Radiotherapy Automated Treatment Planning. Front. Oncol. 2020, 10, 580919. [Google Scholar] [CrossRef]
  184. Rigaud, B.; Anderson, B.M.; Yu, Z.H.; Gobeli, M.; Cazoulat, G.; Söderberg, J.; Samuelsson, E.; Lidberg, D.; Ward, C.; Taku, N.; et al. Automatic Segmentation Using Deep Learning to Enable Online Dose Optimization During Adaptive Radiation Therapy of Cervical Cancer. Int. J. Radiat. Oncol. Biol. Phys. 2021, 109, 1096–1110. [Google Scholar] [CrossRef]
  185. Almberg, S.S.; Lervåg, C.; Frengen, J.; Eidem, M.; Abramova, T.M.; Nordstrand, C.S.; Alsaker, M.D.; Tøndel, H.; Raj, S.X.; Wanderås, A.D. Training, validation, and clinical implementation of a deep-learning segmentation model for radiotherapy of loco-regional breast cancer. Radiother. Oncol. 2022, 173, 62–68. [Google Scholar] [CrossRef]
  186. Li, J.; Anne, R. Comparison of Eclipse Smart Segmentation and MIM Atlas Segment for liver delineation for yttrium-90 selective internal radiation therapy. J. Appl. Clin. Med. Phys. 2022, 23, e13668. [Google Scholar] [CrossRef]
  187. Buchanan, L.; Hamdan, S.; Zhang, Y.; Chen, X.; Li, X.A. Deep learning-based prediction of deliverable adaptive plans for MR-guided adaptive radiotherapy: A feasibility study. Front. Oncol. 2023, 13, 939951. [Google Scholar] [CrossRef] [PubMed]
  188. Eche, T.; Schwartz, L.H.; Mokrane, F.-Z.; Dercle, L. Toward Generalizability in the Deployment of Artificial Intelligence in Radiology: Role of Computation Stress Testing to Overcome Underspecification. Radiol. Artif. Intell. 2021, 3, e210097. [Google Scholar] [CrossRef]
  189. Khan, B.; Fatima, H.; Qureshi, A.; Kumar, S.; Hanan, A.; Hussain, J.; Abdullah, S. Drawbacks of Artificial Intelligence and Their Potential Solutions in the Healthcare Sector. Biomed. Mater. Devices 2023, 1, 731–738. [Google Scholar] [CrossRef] [PubMed]
  190. Petersson, L.; Larsson, I.; Nygren, J.M.; Nilsen, P.; Neher, M.; Reed, J.E.; Tyskbo, D.; Svedberg, P. Challenges to implementing artificial intelligence in healthcare: A qualitative interview study with healthcare leaders in Sweden. BMC Health Serv. Res. 2022, 22, 850. [Google Scholar] [CrossRef]
  191. Wubineh, B.Z.; Deriba, F.G.; Woldeyohannis, M.M. Exploring the opportunities and challenges of implementing artificial intelligence in healthcare: A systematic literature review. Urol. Oncol. Semin. Orig. Investig. 2024, 42, 48–56. [Google Scholar] [CrossRef] [PubMed]
  192. Hagiwara, A.; Fujita, S.; Ohno, Y.; Aoki, S. Variability and Standardization of Quantitative Imaging: Monoparametric to Multiparametric Quantification, Radiomics, and Artificial Intelligence. Investig. Radiol. 2020, 55, 601–616. [Google Scholar] [CrossRef]
  193. Kotter, E.; Ranschaert, E. Challenges and solutions for introducing artificial intelligence (AI) in daily clinical workflow. Eur. Radiol. 2021, 31, 5–7. [Google Scholar] [CrossRef]
  194. Brady, A.P.; Neri, E. Artificial Intelligence in Radiology-Ethical Considerations. Diagnostics 2020, 10, 231. [Google Scholar] [CrossRef]
  195. Acar, E.; Leblebici, A.; Ellidokuz, B.E.; Basbinar, Y.; Kaya, G.C. Machine learning for differentiating metastatic and completely responded sclerotic bone lesion in prostate cancer: A retrospective radiomics study. Br. J. Radiol. 2019, 92, 20190286. [Google Scholar] [CrossRef] [PubMed]
  196. Joshi, G.; Jain, A.; Araveeti, S.R.; Adhikari, S.; Garg, H.; Bhandari, M. FDA-Approved Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices: An Updated Landscape. Electronics 2024, 13, 498. [Google Scholar] [CrossRef]
  197. Arabi, H.; Zaidi, H. Deep learning-based metal artefact reduction in PET/CT imaging. Eur. Radiol. 2021, 31, 6384–6396. [Google Scholar] [CrossRef]
  198. Guo, R.; Zou, Y.; Zhang, S.; An, J.; Zhang, G.; Du, X.; Gong, H.; Xiong, S.; Long, Y.; Ma, J. Preclinical validation of a novel deep learning-based metal artifact correction algorithm for orthopedic CT imaging. J. Appl. Clin. Med. Phys. 2023, 24, e14166. [Google Scholar] [CrossRef] [PubMed]
  199. Zhang, L.; Jiang, B.; Chen, Q.; Wang, L.; Zhao, K.; Zhang, Y.; Vliegenthart, R.; Xie, X. Motion artifact removal in coronary CT angiography based on generative adversarial networks. Eur. Radiol. 2023, 33, 43–53. [Google Scholar] [CrossRef]
  200. Khaleghi, G.; Hosntalab, M.; Sadeghi, M.; Reiazi, R.; Mahdavi, S.R. Metal artifact reduction in computed tomography images based on developed generative adversarial neural network. Inform. Med. Unlocked 2021, 24, 100573. [Google Scholar] [CrossRef]
  201. Lim, S.-W.; Choi, W.-K.; Kim, S. Evaluation of the Effects of Differences in Metal Artifact Type and Location on Image Quality in Computed Tomography Scans. J. Med. Phys. 2023, 48, 80–84. [Google Scholar] [PubMed]
  202. Schmitt, N.; Weyland, C.S.; Wucherpfennig, L.; Herweh, C.; Bendszus, M.; Möhlenbruch, M.A.; Vollherbst, D.F. Iterative Metal Artifact Reduction (iMAR) of the Non-adhesive Liquid Embolic Agent Onyx in Computed Tomography: An Experimental Study. Clin. Neuroradiol. 2022, 32, 695–703. [Google Scholar] [CrossRef]
  203. Koetzier, L.R.; Mastrodicasa, D.; Szczykutowicz, T.P.; van der Werf, N.R.; Wang, A.S.; Sandfort, V.; van der Molen, A.J.; Fleischmann, D.; Willemink, M.J. Deep Learning Image Reconstruction for CT: Technical Principles and Clinical Prospects. Radiology 2023, 306, e221257. [Google Scholar] [CrossRef]
  204. Njiti, M.M.; Osman, N.D.; Mansor, M.S.; Rabaiee, N.A.; Abdul Aziz, M.Z. Potential of Metal Artifact Reduction (MAR) and Deep Learning-based Reconstruction (DLR) algorithms integration in CT Metal Artifact Correction: A review. Radiat. Phys. Chem. 2024, 218, 111541. [Google Scholar] [CrossRef]
  205. Selles, M.; van Osch, J.A.C.; Maas, M.; Boomsma, M.F.; Wellenberg, R.H.H. Advances in metal artifact reduction in CT images: A review of traditional and novel metal artifact reduction techniques. Eur. J. Radiol. 2024, 170, 111276. [Google Scholar] [CrossRef] [PubMed]
  206. Paholpak, P.; Sirichativapee, W.; Wisanuyotin, T.; Kosuwon, W.; Jeeravipoolvarn, P. Prevalence of known and unknown primary tumor sites in spinal metastasis patients. Open Orthop. J. 2012, 6, 440–444. [Google Scholar] [CrossRef]
  207. Piccioli, A.; Maccauro, G.; Spinelli, M.S.; Biagini, R.; Rossi, B. Bone metastases of unknown origin: Epidemiology and principles of management. J. Orthop. Traumatol. 2015, 16, 81–86. [Google Scholar] [CrossRef]
  208. Lawrenz, J.M.; Gordon, J.; George, J.; Haben, C.; Rubin, B.P.; Ilaslan, H.; Mesko, N.W.; Nystrom, L.M. Does PET/CT Aid in Detecting Primary Carcinoma in Patients with Skeletal Metastases of Unknown Primary? Clin. Orthop. Relat. Res. 2020, 478, 2451–2457. [Google Scholar] [CrossRef] [PubMed]
  209. Liu, K.; Qin, S.; Ning, J.; Xin, P.; Wang, Q.; Chen, Y.; Zhao, W.; Zhang, E.; Lang, N. Prediction of Primary Tumor Sites in Spinal Metastases Using a ResNet-50 Convolutional Neural Network Based on MRI. Cancers 2023, 15, 2974. [Google Scholar] [CrossRef]
  210. Lang, N.; Zhang, Y.; Zhang, E.; Zhang, J.; Chow, D.; Chang, P.; Yu, H.J.; Yuan, H.; Su, M.Y. Differentiation of spinal metastases originated from lung and other cancers using radiomics and deep learning based on DCE-MRI. Magn. Reson. Imaging 2019, 64, 4–12. [Google Scholar] [CrossRef]
  211. Cao, G.; Zhang, J.; Lei, X.; Yu, B.; Ai, Y.; Zhang, Z.; Xie, C.; Jin, X. Differentiating Primary Tumors for Brain Metastasis with Integrated Radiomics from Multiple Imaging Modalities. Dis. Markers 2022, 2022, 5147085. [Google Scholar] [CrossRef]
  212. Goehler, A.; Harry Hsu, T.-M.; Lacson, R.; Gujrathi, I.; Hashemi, R.; Chlebus, G.; Szolovits, P.; Khorasani, R. Three-Dimensional Neural Network to Automatically Assess Liver Tumor Burden Change on Consecutive Liver MRIs. J. Am. Coll. Radiol. 2020, 17, 1475–1484. [Google Scholar] [CrossRef] [PubMed]
  213. Lindgren Belal, S.; Larsson, M.; Holm, J.; Buch-Olsen, K.M.; Sörensen, J.; Bjartell, A.; Edenbrandt, L.; Trägårdh, E. Automated quantification of PET/CT skeletal tumor burden in prostate cancer using artificial intelligence: The PET index. Eur. J. Nucl. Med. Mol. Imaging 2023, 50, 1510–1520. [Google Scholar] [CrossRef]
  214. Oprea-Lager, D.E.; Cysouw, M.C.F.; Boellaard, R.; Deroose, C.M.; de Geus-Oei, L.F.; Lopci, E.; Bidaut, L.; Herrmann, K.; Fournier, L.S.; Bäuerle, T.; et al. Bone Metastases Are Measurable: The Role of Whole-Body MRI and Positron Emission Tomography. Front. Oncol. 2021, 11, 772530. [Google Scholar] [CrossRef]
  215. Moreau, N.; Rousseau, C.; Fourcade, C.; Santini, G.; Ferrer, L.; Lacombe, M.; Guillerminet, C.; Campone, M.; Colombié, M.; Rubeaux, M.; et al. Deep learning approaches for bone and bone lesion segmentation on 18FDG PET/CT imaging in the context of metastatic breast cancer. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 1532–1535. [Google Scholar]
  216. Liu, S.; Feng, M.; Qiao, T.; Cai, H.; Xu, K.; Yu, X.; Jiang, W.; Lv, Z.; Wang, Y.; Li, D. Deep Learning for the Automatic Diagnosis and Analysis of Bone Metastasis on Bone Scintigrams. Cancer Manag. Res. 2022, 14, 51–65. [Google Scholar] [CrossRef] [PubMed]
  217. Saeed, M.U.; Dikaios, N.; Dastgir, A.; Ali, G.; Hamid, M.; Hajjej, F. An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images. Diagnostics 2023, 13, 2658. [Google Scholar] [CrossRef]
  218. Schutte, K.; Brulport, F.; Harguem-Zayani, S.; Schiratti, J.-B.; Ghermi, R.; Jehanno, P.; Jaeger, A.; Alamri, T.; Naccache, R.; Haddag-Miliani, L.; et al. An artificial intelligence model predicts the survival of solid tumour patients from imaging and clinical data. Eur. J. Cancer 2022, 174, 90–98. [Google Scholar] [CrossRef] [PubMed]
  219. Gålne, A.; Enqvist, O.; Sundlöv, A.; Valind, K.; Minarik, D.; Trägårdh, E. AI-based quantification of whole-body tumour burden on somatostatin receptor PET/CT. Eur. J. Hybrid Imaging 2023, 7, 14. [Google Scholar] [CrossRef]
  220. Moy, S.; Irannejad, M.; Manning, S.J.; Farahani, M.; Ahmed, Y.; Gao, E.; Prabhune, R.; Lorenz, S.; Mirza, R.; Klinger, C. Patient Perspectives on the Use of Artificial Intelligence in Health Care: A Scoping Review. J. Patient Cent. Res. Rev. 2024, 11, 51–62. [Google Scholar] [CrossRef]
  221. Alanazi, A. Clinicians’ Views on Using Artificial Intelligence in Healthcare: Opportunities, Challenges, and Beyond. Cureus 2023, 15, e45255. [Google Scholar] [CrossRef]
  222. Daniyal, M.; Qureshi, M.; Marzo, R.R.; Aljuaid, M.; Shahid, D. Exploring clinical specialists’ perspectives on the future role of AI: Evaluating replacement perceptions, benefits, and drawbacks. BMC Health Serv. Res. 2024, 24, 587. [Google Scholar] [CrossRef] [PubMed]
  223. Khanna, N.N.; Maindarkar, M.A.; Viswanathan, V.; Fernandes, J.F.E.; Paul, S.; Bhagawati, M.; Ahluwalia, P.; Ruzsa, Z.; Sharma, A.; Kolluri, R.; et al. Economics of Artificial Intelligence in Healthcare: Diagnosis vs. Treatment. Healthcare 2022, 10, 2493. [Google Scholar] [CrossRef]
Figure 1. Figure depicting the hierarchical arrangement of artificial intelligence (AI). Machine learning (ML), a subset of AI, aims to empower computers to learn independently without explicit programming. Deep learning (DL), a specialized field within machine learning, involves the computation of neural networks comprising multiple layers.
Figure 1. Figure depicting the hierarchical arrangement of artificial intelligence (AI). Machine learning (ML), a subset of AI, aims to empower computers to learn independently without explicit programming. Deep learning (DL), a specialized field within machine learning, involves the computation of neural networks comprising multiple layers.
Cancers 16 02988 g001
Figure 2. Diagram illustrating the basic framework and essential stages of radiomics, including image acquisition, image processing (including segmentation), feature extraction within specified regions of interest (ROIs), feature selection, exploratory analysis, and subsequent modeling.
Figure 2. Diagram illustrating the basic framework and essential stages of radiomics, including image acquisition, image processing (including segmentation), feature extraction within specified regions of interest (ROIs), feature selection, exploratory analysis, and subsequent modeling.
Cancers 16 02988 g002
Figure 3. PRISMA flowchart (adapted from PRISMA group, 2020) outlining the process of selecting pertinent articles for analysis.
Figure 3. PRISMA flowchart (adapted from PRISMA group, 2020) outlining the process of selecting pertinent articles for analysis.
Cancers 16 02988 g003
Figure 4. Diagram illustrating a typical radiogenomics process, encompassing image acquisition and pre-processing, feature extraction and selection from medical imaging and genotype data, correlation between radiomic and genomic features, data analysis using machine learning models, and the determination of the final radiogenomics outcome.
Figure 4. Diagram illustrating a typical radiogenomics process, encompassing image acquisition and pre-processing, feature extraction and selection from medical imaging and genotype data, correlation between radiomic and genomic features, data analysis using machine learning models, and the determination of the final radiogenomics outcome.
Cancers 16 02988 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ong, W.; Lee, A.; Tan, W.C.; Fong, K.T.D.; Lai, D.D.; Tan, Y.L.; Low, X.Z.; Ge, S.; Makmur, A.; Ong, S.J.; et al. Oncologic Applications of Artificial Intelligence and Deep Learning Methods in CT Spine Imaging—A Systematic Review. Cancers 2024, 16, 2988. https://doi.org/10.3390/cancers16172988

AMA Style

Ong W, Lee A, Tan WC, Fong KTD, Lai DD, Tan YL, Low XZ, Ge S, Makmur A, Ong SJ, et al. Oncologic Applications of Artificial Intelligence and Deep Learning Methods in CT Spine Imaging—A Systematic Review. Cancers. 2024; 16(17):2988. https://doi.org/10.3390/cancers16172988

Chicago/Turabian Style

Ong, Wilson, Aric Lee, Wei Chuan Tan, Kuan Ting Dominic Fong, Daoyong David Lai, Yi Liang Tan, Xi Zhen Low, Shuliang Ge, Andrew Makmur, Shao Jin Ong, and et al. 2024. "Oncologic Applications of Artificial Intelligence and Deep Learning Methods in CT Spine Imaging—A Systematic Review" Cancers 16, no. 17: 2988. https://doi.org/10.3390/cancers16172988

APA Style

Ong, W., Lee, A., Tan, W. C., Fong, K. T. D., Lai, D. D., Tan, Y. L., Low, X. Z., Ge, S., Makmur, A., Ong, S. J., Ting, Y. H., Tan, J. H., Kumar, N., & Hallinan, J. T. P. D. (2024). Oncologic Applications of Artificial Intelligence and Deep Learning Methods in CT Spine Imaging—A Systematic Review. Cancers, 16(17), 2988. https://doi.org/10.3390/cancers16172988

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop