Next Article in Journal
Thai Guideline for Nuclear Medicine Investigations of Neurocognitive Disorders: Nuclear Medicine Society of Thailand, the Neurological Society of Thailand, and Thai Medical Physicist Society Collaboration
Previous Article in Journal
Impact of Cardiac Surgery Scar on Heart Rupture Following a Fall from Height
Previous Article in Special Issue
Radiomics Based on Single-Phase CTA for Distinguishing Left Atrial Appendage Thrombus from Circulatory Stasis in Patients with Atrial Fibrillation before Ablation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Decoding Radiomics: A Step-by-Step Guide to Machine Learning Workflow in Hand-Crafted and Deep Learning Radiomics Studies

by
Maurizio Cè
1,
Marius Dumitru Chiriac
2,
Andrea Cozzi
3,
Laura Macrì
1,
Francesca Lucrezia Rabaiotti
1,
Giovanni Irmici
4,
Deborah Fazzini
5,
Gianpaolo Carrafiello
6,7 and
Michaela Cellina
8,*
1
Postgraduation School in Radiodiagnostics, Università degli Studi di Milano, Via Festa del Perdono 7, 20122 Milan, Italy
2
Politecnico di Milano, Piazza Leonardo da Vinci 32, 20133 Milan, Italy
3
Imaging Institute of Southern Switzerland (IIMSI), Ente Ospedaliero Cantonale (EOC), Via Tesserete 46, 6900 Lugano, Switzerland
4
Breast Imaging Department, Fondazione IRCCS Istituto Nazionale dei Tumori, Via Giacomo Venezian 1, 20133 Milan, Italy
5
Radiology Department, Centro Diagnostico Italiano, Via Saint Bon 20, 20147 Milan, Italy
6
Radiology Department, Fondazione IRCCS Cà Granda Ospedale Maggiore Policlinico, Via Francesco Sforza 35, 20122 Milan, Italy
7
Department of Oncology and Hematology-Oncology, Università degli Studi di Milano, Via Festa del Perdono 7, 20122 Milan, Italy
8
Radiology Department, ASST Fatebenefratelli Sacco, Piazza Principessa Clotilde 3, 20121 Milan, Italy
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(22), 2473; https://doi.org/10.3390/diagnostics14222473
Submission received: 4 September 2024 / Revised: 25 October 2024 / Accepted: 28 October 2024 / Published: 5 November 2024
(This article belongs to the Special Issue Machine Learning in Radiomics: Opportunities and Challenges)

Abstract

:
Although radiomics research has experienced rapid growth in recent years, with numerous studies dedicated to the automated extraction of diagnostic and prognostic information from various imaging modalities, such as CT, PET, and MRI, only a small fraction of these findings has successfully transitioned into clinical practice. This gap is primarily due to the significant methodological challenges involved in radiomics research, which emphasize the need for a rigorous evaluation of study quality. While many technical aspects may lie outside the expertise of most radiologists, having a foundational knowledge is essential for evaluating the quality of radiomics workflows and contributing, together with data scientists, to the development of models with a real-world clinical impact. This review is designed for the new generation of radiologists, who may not have specialized training in machine learning or radiomics, but will inevitably play a role in this evolving field. The paper has two primary objectives: first, to provide a clear, systematic guide to radiomics study pipeline, including study design, image preprocessing, feature selection, model training and validation, and performance evaluation. Furthermore, given the critical importance of evaluating the robustness of radiomics studies, this review offers a step-by-step guide to the application of the METhodological RadiomICs Score (METRICS, 2024)—a newly proposed tool for assessing the quality of radiomics studies. This roadmap aims to support researchers and reviewers alike, regardless of their machine learning expertise, in utilizing this tool for effective study evaluation.

1. Introduction

In recent years, radiomics has emerged as a prominent topic in medical literature, highlighting the growing importance of artificial intelligence applications in medicine, particularly in radiology [1,2,3]. Radiomics refers to the extraction and analysis of large sets of quantitative features from medical images [4,5]. From now on, unless otherwise specified, we will use the term “radiomics” in its broadest sense, including all AI-driven applications aimed at automating the extraction of diagnostic information from imaging data.
The development of radiomics studies requires a deep understanding of the complex machine learning (ML) pipeline, which often goes beyond the expertise of a single radiologist. However, considering that the integration of AI in radiology is expected to have significant professional implications [6,7,8], it is essential for radiologists to acquire foundational skills in this field, which will allow them to critically evaluate the impact of these technologies in daily clinical practice and to strengthen their collaborative role with data scientists [9].
While many comprehensive reviews provide an introductory overview of radiomics aims and applications [10,11,12,13], very few have ever addressed the mechanisms and challenges of the ML workflow in a unified framework [8,14,15,16,17,18,19].
Therefore, this paper pursues two key objectives. First, it aims to offer an introductory but rigorous guide to the core ML concepts applied to radiomics. Second, it aims to provide a roadmap for navigating the recently proposed Methodological Radiomics Score (METRICS, 2024) tool [20] for evaluating the quality of radiomics studies, helping researchers and reviewers adopt it for study assessment. METRICS offers expert opinion-based importance weights for categories and items, marking the first time a scoring tool has adopted such a transparent methodology. The tool is versatile, covering both hand-crafted radiomics and fully deep learning-based pipelines, adapting to various use cases [20].
The structure of this article is as follows: Section 2 presents a brief overview of ML and deep learning (DL) definitions; Section 3 introduces the bias–variance trade-off, the cornerstone of any model optimization process; Section 4 dissects the various stages of the radiomics study pipeline, including study design, image preprocessing, feature selection, model training, validation, and performance assessment; and Section 5 discusses some of the latest developments and highlights potential challenges in the evolution and implementation of these tools.
In general, a radiomic approach can follow one of three different pathways: (1) hand-crafted radiomics, (2) deep radiomics, or (3) end-to-end deep learning (Figure 1).
The METRICS tool incorporates conditional items designed to cover all of these pathways [13,14,15]. A detailed, step-by-step roadmap is presented in Figure 2, including subtasks specific to each phase and the corresponding items of the METRICS tool. Table 1 lists METRICS items and the corresponding paragraphs in which the topic is discussed.
In hand-crafted radiomics, mathematically designed imaging features are extracted from the segmented region or volume of interest and presented as a tabular dataset for modeling using traditional statistical ML models or neural networks. In contrast, deep radiomics employs DL models, particularly convolutional neural networks, to automatically learn and extract features from images. The end-to-end DL approach integrates the entire image processing pipeline into a DL model, directly predicting outcomes from raw images without manual intervention.

2. Machine Learning and Deep Learning

Radiomics extensively utilizes ML models [4,8,21]. These models leverage computational power to automate algorithms that learn from data, mimicking human learning processes and improving accuracy over time through experience [22,23,24,25]. More specifically, ML algorithms encompass a wide range of approaches for estimating functions that map the relationship between a set of inputs, X, and an output variable, Y. According to the context, input variables are also referred to as predictors or independent variables, while output variables can be called the dependent variable, target, label, or response [25]. In radiomics, the predictors always include imaging features and may also incorporate clinical, demographic, or laboratory variables. Conversely, the target variable can be either quantitative (e.g., months of survival) or qualitative (e.g., histological subtype or recurrence prediction).
ML algorithms can be divided into traditional (“statistics-based”) ML models and neural networks (Figure 3).
Traditional ML models can be seen as the automation of classical statistical approaches and their derivatives, which are typically represented by mathematical formulas [25], such as the regression line of linear regression: (f(X) = β0 + β1X1 + β2X2 + … + βpXp) [20]. Other examples include logistic regression, decision tree (DT), support vector machines (SVM), naive Bayes, and k-nearest neighbor (KNN). In general, traditional ML models require less training data and offer greater interpretability and transparency, providing clearer insights into how input variables influence the output [26]. However, their capacity to capture complicated patterns is limited [25,26,27,28].
On the other hand, neural networks follow a completely different paradigm, inspired by the biological structure of the human brain [27,28]. These models consist of layers of nodes (neurons) that process data through weights and activation functions [27,29]. When neural networks contain a large number of intermediate layers, they are referred to as deep neural networks [27,29]. Deep learning (DL) employs these deep neural networks, allowing for more complex non-linear representations at the cost of a more difficult interpretation of their internal organization, i.e., the “black box” phenomenon [27,28,29].
Convolutional neural networks (CNNs), a specific type of deep neural networks designed for image analysis, use convolutional layers that compare overlapping regions of the input with small weight matrices, known as kernels or filters, to encode image features. While increasing the number of layers enhances their ability to learn complex patterns, it also makes these DL models harder to train [27,28]. This has led to architectural refinements, such as skip connections and bottleneck blocks (as seen in ResNet), multibranch convolutions (GoogLeNet), and ensemble methods [29]. CNNs have been widely adopted in radiomic studies, and they can be used in an end-to-end fashion or to support specific subtasks of the radiomics pipeline, such as data augmentation, feature extraction, image segmentation, classification, prediction, or multimodal integration.

3. Bias–Variance Trade-Off

The trade-off between bias and variance is key in ML, as many of the challenges and methodological issues in ML ultimately aim to optimize this balance [25,30]. Bias measures the average difference between the predicted values and the true values, reflecting how accurately a model can predict on the training dataset. High bias indicates that the model is not capturing the underlying patterns of the data. On the other hand, variance measures the variability of model predictions across different datasets, indicating how well the model can generalize to new, unseen data. High variance means the model is overly sensitive to the training data, resulting in more errors on the testing set (i.e., new, unseen data). Essentially, bias and variance inversely represent model accuracy for training and test sets, respectively.
As shown in Figure 4A, as the complexity of the model increases, the bias decreases while the variance increases. Although developers aim to reduce both, it is often not possible to do so simultaneously, leading to the necessity for regularization, which decreases the model variance at the cost of increased bias.
The concepts of bias and variance are strictly related to overfitting and underfitting (Figure 4B) [14,25]. Overfitting occurs when the model has low bias but high variance, capturing noise and fluctuations in the training data rather than the underlying pattern. By introducing regularization and feature selection, the model’s variance increases and the model’s performance on unseen data improves, at the cost of slightly increasing the model’s bias. Underfitting occurs when a model has both high bias and high variance, resulting in a poor performance for both training and testing data. This often happens when the model is too simple or the training data are insufficient.
Constructing models with good performance necessitates a large number of predictors [15,25,30,31,32]. However, this complexity can lead to overfitting, creating a disadvantageous loop where the model fits the training data too closely and performs poorly on unseen data. Overfitting can be mitigated by several methods (Figure 5), discussed in the following paragraphs.

4. Step-by-Step Radiomic Workflow

4.1. Study Design and Data Collection

When formulating a radiomics study, adherence to established guidelines and checklists—such as CLAIM [33,34] and CLEAR [27]—is crucial for ensuring rigorous design and reporting standards (METRICS item 1 [20]).

4.1.1. Eligibility

Eligibility criteria must be meticulously defined to select a representative sample of the population of interest. Therefore, inclusion and exclusion criteria should be explicitly outlined to minimize bias (METRICS item 2 [20]).

4.1.2. Reference Standard

A high-quality reference standard is essential for robust outcome measures (METRICS item 3 [20]) [28]. This standard should align with current clinical practices and reliable methods, such as histopathology, well-established clinical scores, genomic markers, prognostic tools, or consensus-based guidelines and expert opinions [20].
Furthermore, the time interval between the examination and the acquisition of the reference standards must accurately reflect the presence or absence of target conditions at the time of the diagnostic exams (METRICS item 7 [20]) [11,15]. If there is a significant delay, changes in the patient’s condition or disease progression could undermine the validity of the radiomic analysis.

4.1.3. Monocentric Versus Multicenter

Radiomic analysis relies heavily on the quality and consistency of diagnostic imaging data [14,16,35]. Indeed, the involvement of various institutions (METRICS item 4 [20]) is beneficial to enhance the generalizability and robustness of the radiomic models. However, a multicenter design does not necessarily imply external validation. These are two separate design choices, even if one (availability of different data sources) is a prerequisite of the other (external validation/testing). Of note, a multicenter setting implies that the variability of imaging protocols between different institutions must be carefully managed to prevent distortions [14,15,36,37].

4.1.4. Imaging Protocol

The adherence to established guidelines for imaging acquisition protocol is essential for ensuring consistency and comparability across different imaging studies, as well as clinical translatability (METRICS item 5 [20]) [11]. For example, in clinical practice for prostate cancer, the PI-RADS guidelines provide specific recommendations for MRI acquisition parameters, ensuring that obtained images are of a high quality and suitable for accurate assessment of prostate lesions [38,39]. Therefore, acquisition protocols should be clearly reported (METRICS item 6 [20]), as standardized acquisition protocols reduce the variability in image quality and technical parameters, which can significantly impact the extracted radiomic features. Whenever possible, preference should be given to single imaging sets (such as a single MRI sequence) over multi-parametric imaging to avoid unnecessary data complexity and overfitting [20].

4.2. Image Preprocessing

Preprocessing medical images is a crucial step that impacts the accuracy and reliability of feature extraction (METRICS item 11 [20]). Preprocessing is essential for reducing variability and making radiomic features comparable across different scans and patients [40,41,42], especially considering the diverse imaging modalities (e.g., PET, MRI, and CT) and the specific requirements of different feature extraction techniques (e.g., 2D versus 3D). Image preprocessing encompasses several procedures that are outlined in the following subsections [43].

4.2.1. Normalization and Standardization

In image preprocessing, normalization and standardization are often confused but are distinct processes. Normalization (or min–max scaling) scales data to a specific range, like [0, 1] or [−1, 1], while standardization (or Z-score normalization) adjusts data to have a mean of 0 and a standard deviation of 1, transforming them to a normal distribution [44], allowing a better understanding of the data’s spread and variability [14,16]. Both are used to prepare data for algorithms, facilitating optimal comparisons across data acquisition methods and texture instances [45].
Normalization entails dividing each pixel’s value by the maximum possible value for that pixel (255 for an 8 bit image, 4095 for a 12 bit image, and 65,535 for a 16 bit image). For instance, CT images are primarily encoded in 12 bits with gray levels; normalizing a CT image involves dividing each pixel value by 4095 to achieve a range between 0 and 1 [46].
Standardization is especially recommended in MRI since, unlike a CT, units of signal intensity are arbitrary [16].
Different normalization methods impact feature extraction, and the choice of method depends on dataset characteristics and desired feature types [47].

4.2.2. Discretization

Discretization involves modifying images to adhere to a common format or scale and is crucial for MRI scans, where intensity values can differ significantly across different scanners or scanning protocols [48,49,50,51,52]. Gray-level discretization consists of converting continuous intensity values (or density for CT) into discrete bins, simplifying the intensity distribution and facilitating texture analysis. Discretization involves three key parameters: the range of the data, the number of bins, and the width of each bin, but only two can be independently controlled. The range is typically preserved from the original data, although it may be adjusted in certain contexts, such as when comparing it to a reference dataset. The optimal bin number in discretization is crucial, as too few bins can obscure features, while too many can amplify noise, and finding the right balance depends on both data acquisition parameters and feature content [15]. Fixing the bin number, as often done in MRI, helps normalize image intensities and improve reproducibility across different samples. In contrast, fixing the bin size, as often done in PET, allows for a direct relationship between the bins and the original intensity scale, aiding in the comparison of data with different ranges [15]. The parameters for discretization, such as bin width and the resulting gray-level range, or bin count, should be thoroughly reported [48].

4.2.3. Co-Registration

Image co-registration is necessary when a single mask is to be applied to all image sets (e.g., multiple contrast phases or sequences), ensuring that the anatomical structures are correctly aligned [53]. This process involves aligning images from different time points, scanners, or imaging modalities to a common coordinate system. In neuroimaging studies, the co-registration of images may also involve aligning images from different sequences or modalities to a common reference frame using anatomical atlases, like the one from the Montreal Neurological Institute, used to map the location of brain structures, regardless of individual differences in the overall brain size and shape [54,55].

4.2.4. Resampling

Upsampling and downsampling are techniques used to adjust the image resolution [15]. In cases of a large slice thickness (e.g., ≥5 mm), extreme upsampling to a very fine resolution (e.g., 1 × 1 × 1 mm3) might introduce artifacts and inaccuracies [56]. Instead, using 2D feature extraction techniques that ensure in-plane isotropy of the pixels can be more appropriate [20,56]. Conversely, for 3D feature extraction, achieving isotropic voxel values is essential for ensuring the rotational invariance of texture features [57].

4.2.5. Image Filtering and Enhancement Techniques

Image filtering and enhancement techniques are used to highlight specific features or to reduce noise [14,16,58]. Image filtration can be used before the extraction of features as a preprocessing step to highlight particular image properties. For instance, wavelet decomposition uses low-pass filters to capture low-frequency components (basic structures of the image) and a high-pass filter to capture high-frequency components (fine details, such as edges) [59]. Moreover, Laplacian of Gaussian filtering, with specified σ values, can enhance edges and make the features more distinct [60]. It is important to document the type of filters used and their parameters, as these can significantly affect the extracted features.

4.3. Segmentation

In computer vision, image segmentation involves clustering together parts of an image that belong to the same object class, thus creating a region of interest (ROI) in 2D images or a volume of interest (VOI) in 3D images [14,15,16,61,62]. Indeed, this process enhances the relevance of the segmented areas, making them easier to interpret and analyze.
The METRICS tool encompasses a wide definition of segmentation, including (1) fine delineation of a ROI or a VOI; (2) rough delineation with bounding boxes; or, (3) cropping the image around a ROI (METRICS condition 1 [20]), the last two mostly applying to deep radiomics. In radiomics studies, the ROI/VOI generally coincides with a mass or nodule (for example, a lung tumor), but it can also be a tissue or an organ [62,63].
There are several segmentation methods, each with its advantages and limitations [64]. Based on the level of human intervention, they can be classified as manual, semi-automatic, and automatic (METRICS condition 2 [20]). According to the METRICS tool, any manual adjustments to the annotation, in terms of area, volume, or predefined parameters, classify the technique as semi-automatic [20]. The segmentation procedure should be explicitly detailed (METRICS item 8 [20]), as accurate tumor segmentation is one of the primary challenges for ensuring the reliability of radiomic features, particularly when using manual or semiautomatic methods.
Fully manual segmentation, where the human annotator manually outlines the ROI (Figure 6A), represents the simplest but most time-consuming option, and is typically performed when dealing with small datasets that can be annotated by expert radiologists [14]. Its other main limitation lies in the frequent absence of standardized segmentation protocols, with high intra- and inter-observer variabilities [14,16].
Currently, several open source and proprietary software solutions allow for the automatic or semi-automatic segmentation of radiological images. Among the most popular are 3D Slicer, MITK, ITK-SNAP, LifEx, and ImageJ [15,16,66].
Semi-automatic segmentation typically involves utilizing customized algorithms to segment images based on predefined parameters, after which manual verification and adjustments are made by an expert to maximize accuracy [67]. Semi-automated segmentation techniques include thresholding, edge detection, and region-based segmentation [64]. Thresholding techniques classify pixels based on their intensity values; they are simple and fast, but may struggle with images that have poor contrast (Figure 6B) [64,65,68]. Edge detection methods identify boundaries within an image, providing a clear delineation of structures, though they can be more sensitive to noise [69]. Region-based segmentation groups neighboring pixels with similar properties, resulting in more homogeneous regions, but potentially missing fine details. Clustering algorithms, such as k-means or hierarchical clustering, categorize pixels into distinct clusters based on their attributes (Figure 6C) [70].
Deep neural networks are extremely powerful tools used to perform automatic segmentation and can reduce the workload and increase reproducibility in the setting of hand-crafted radiomics [71,72]. For example, encoder–decoder architecture, pioneered by the fully convolutional network, is commonly used for segmentation tasks [27]: the encoder downsamples the image through convolutional layers, while the decoder upsamples the feature maps to produce per-pixel labeled outputs. U-Net, a popular architecture for semantic segmentation, employs a symmetric U-shaped design with skip connections to facilitate upsampling. For the so-called “instance segmentation”, which combines object detection and semantic segmentation, the Mask R-CNN architecture is prominent: it extends two-stage detection models by adding a branch to predict binary masks for each object category, enabling instance segmentation [71]. The primary limitations of DL-based automatic methods stem from their requirement for large, labeled datasets to train accurate models, as well as concerns about the generalizability of these algorithms, which may perform poorly when applied to datasets different from those used during training [14].
If an entirely automated segmentation technique is used, examples of the results should be provided, and a formal accuracy assessment should be included in the study, comparing the results with those of expert annotators (for example, using the DICE score or Jaccard index against a radiologist’s semantic annotation) [73,74]. This requirement also applies to the use of segmentation models that have been previously validated on other datasets (METRICS item 9 [20]).
In any case, a clear statement should be provided about whether the final segmentation in the test set is produced by a single reader (manually or with a semi-automated tool) or an entirely automated tool (METRICS item 10 [20]).

4.4. Feature Extraction

Radiomic features are quantitative characteristics extracted from medical images and represent the main input of the radiomic model [4,5,16,75]. These features represent the radiological “signature” of phenotypes or biological characteristics and can be used in ML modeling for diagnostic, prognostic, and predictive purposes [76,77,78,79]. Radiomic features can be categorized into two main types: handcrafted features and deep features [14,15,20,78].

4.4.1. Hand-Crafted Features

Radiomic features are termed “hand-crafted” because they are generated by algorithms designed or selected by data scientists, rather than being learned directly from images as in deep learning (METRICS condition 3 [16,20]).
Hand-crafted features are traditionally divided into first-order, second-order, and higher-order features (Table 2) [80,81,82]. As the order increases, they become less interpretable, meaning it becomes more difficult to visually understand what they represent in the images. First-order features describe the distribution of values of individual voxels disregarding their spatial relationship (for example, mean, median standard deviation, kurtosis, etc.), or basic geometric properties of the ROI/VOI, such as volume and maximum surface area [81,83]. Second-order features, for example those derived from the gray-level co-occurrence matrix, capture textural information by examining the spatial relationship between pixel pairs [84]. Higher-order features provide more complex textural patterns and spatial relationships within the image [80]. For a list of hand-crafted features, see [19]. There is no general consensus on the definition, giving rise to problems in comparing different radiomics studies [19]. Some authors categorize radiomic features into different classes based on their significance [81]. These include intensity-based measures (primarily first-order features), heterogeneity and texture (second-order features), shape and volume (second-order features such as volume, sphericity, compactness, and surface-to-volume ratio), peritumoral radiomics (which assesses structural heterogeneity in the peritumoral region surrounding a tumor, including stroma, lymph nodes, and potential metastatic sites), and tumor vascularity radiomics (vessel tortuosity and structural organization).
The feature classes and the number of extracted features for each class should be clearly reported (METRICS item 13 [20,34]).
The Image Biomarker Standardisation Initiative (IBSI) provides guidelines for the extraction, definition, and validation of hand-crafted radiomic features [85]. This standardization ensures reproducibility and comparability across different studies and institutions, thereby enhancing the reliability and clinical utility of radiomic analyses. Whether the feature extraction was conducted according to a compliant standard is the subject of the METRICS item 12 [20].

4.4.2. Deep Features

On the other hand, deep features are automatically extracted using DL methods, capturing patterns and abstractions without explicit human intervention [21,27]. Since DL methods utilize CNNs, their deterministic output helps eliminate variations both within and between observers [14,21]. Compared to hand-crafted features, deep feature representation works at the expense of less interpretability of the geometric and radiological meaning of individual features [27,86]. In the case of DL, the architecture of the neural network should be described along with all operations on the image (METRICS item 13 [20]).
Deep features can be “exposed” to the user and available for processing as tabular data.

4.5. Tabular Data

After the extraction of hand-crafted features from medical images, the results are organized into a tabular data structure, often referred to as a “tabular dataset”. This step applies to both hand-crafted radiomics and deep radiomics, but not to the end-to-end deep learning pathway (METRICS condition 4 [20]). This tabular dataset consists of rows and columns, where each row represents an observation (ROI) and each column represents a radiomic feature or other clinical and demographic variables used as predictors [15]. A tabular structure improves data manipulation and analysis, allowing researchers to perform better data cleaning and statistical computations [87].

4.6. Data Preparation: Missing Values, Data Scarcity, Confounding Factors, and Class Imbalance Problems

The accuracy of radiomic models can be undermined by data scarcity, confounding factors and class imbalance problems, which can introduce biases and lead to unreliable predictions [88,89,90]. Handling confounding factors and class imbalance issues can span across both data preparation and modeling; however, the primary efforts to address these issues occur during data preparation.

4.6.1. Missing Values

Handling missing values is crucial for maintaining the integrity and validity of analyses [34]. Missing values can arise from several sources, such as image acquisition errors, technical problems, or simply from variability in clinical practices [91]. Strategies to handle missing data include removing records with missing values, simple imputation methods in which missing values are replaced by statistics, like the mean or median, and advanced imputation techniques exploiting ML approaches [90,92]. Additionally, analyzing missing data patterns to determine if they are random or systematic can prevent bias and ensure robustness and reliability [93].

4.6.2. Data Scarcity

Radiomic datasets can be subject to limited sizes and a scarcity of data that easily results in overfitting. This problem can be addressed through several approaches. Data augmentation consists of applying a series of transformations to the original images to obtain new data [93]. Generative adversarial networks (GANs) are a peculiar type of DL network particularly suitable for this purpose, capable of creating “fake” images from the originals [26]. Conditional GANs introduce additional input information to guide the generation process, while CycleGANs translate images from one domain to another, useful for tasks such as medical image synthesis and translation [94,95]. These approaches can also be used to evaluate the robustness of radiomics features in different conditions.

4.6.3. Confounding Factors

Confounding factors are variables that independently affect both the predictors (radiomics features) and the target variables (disease outcomes), misleading the model into attributing effects to the radiomic features that are actually due to the confounders [89,96]. For instance, in a study aiming to predict disease outcomes based on certain imaging features, age and socioeconomic status might act as confounders if not appropriately controlled. Since the variability of imaging exams is typically controlled in early stages (imaging protocol and preprocessing), confounding factors primarily rely on different distributions of demographic or clinical features (e.g., sex, stage, or lesion grade) across sites or scanners. Their presence can lead to spurious associations and erroneous conclusions; therefore, the proper handling of confounding factors is the key to developing a robust radiomic pipeline (METRICS item 19 [20]).
To mitigate the impact of confounding factors in clinical studies, several strategies can be employed, like stratification, matching, statistical adjustment, and randomization [89]. However, in retrospective contexts, as is generally the case of radiomic studies, researchers must rely on confounding control techniques during data analysis, rather than during collection. These include statistical methods such as multivariate analysis, propensity score matching, and others [97,98]. However, these techniques can only partially mitigate the impact of confounders, and their success depends on the availability and quality of information about potential confounders.

4.6.4. Class Imbalance Problems

Class imbalance occurs when the number of instances in one class significantly outnumbers the instances in other classes [99,100]. This imbalance can skew the performance of ML models, leading to biased predictions that favor the majority class. For example, if a dataset contains 95% healthy cases and only 5% diseased cases, a model that predicts every case to be healthy would achieve high overall accuracy, despite being practically useless. To address class imbalance problems oversampling, undersampling, cost-sensitive learning, algorithmic approaches, and ensemble methods [100].
Addressing both confounding factors and class imbalance often requires an integrated approach. For instance, while stratifying the data to control for confounders, one might also need to apply resampling techniques within each stratum to address class imbalance. Additionally, preprocessing steps should be carefully designed to ensure that solutions to one problem do not exacerbate the other.

4.7. Features Robustness

Feature robustness refers to the stability (or reliability) of radiomic features when subjected to variations in imaging conditions, such as different scanners, acquisition parameters, and image preprocessing techniques [101,102,103]. When evaluating the robustness of radiomics features, the aim is to determine how much of the feature variability is attributable to the intrinsic properties of the imaged object (such as a tissue or a tumor) and how much is due to the methods of image acquisition.
Methods for testing robustness include test–retest analysis, where images of the same subject are acquired at different times, and image perturbation, where the original image is subjected to a series of transformations, such as adding noise, translation, rotation, and others (Figure 7) [94,104].
Factors that can affect robustness include noise, resolution, segmentation variability, ROI size, but also image preprocessing steps, such as normalization and resampling [50,105]. Ensuring robustness is crucial for the clinical applicability of radiomic features (METRICS item 14 [20]) [20,106,107].
The Intraclass Correlation Coefficient (ICC) and the Concordance Correlation Coefficient (CCC) are two common metrics used for assessing the robustness of radiomic features (Table 3) [20,108,109]. The ICC and CCC offer a comprehensive evaluation of feature robustness, assisting researchers with identifying features that are less influenced by technical and biological variations. Features with a high CCC (i.e., >0.9) and ICC (i.e., >0.75) demonstrate strong test–retest reliability and inter-observer agreement, respectively, ensuring high reproducibility [14,20,21,110]. These features are subsequently retained for further analysis, while features falling below these thresholds are discarded [14,20,21,110].
Assessing the consistency of performances in an end-to-end DL pipeline is essential for ensuring its robustness and clinical reliability. This can be achieved in practice by taking advantage of some of the approaches discussed above, including test–retest settings, such as scan–rescan studies, segmentations by different readers, or stability analysis involving image perturbations. These evaluations help determine whether the model maintains high performance, despite variations in the imaging conditions or reader interpretations (METRICS item 17 [20]).

4.8. Feature Selection and Regularization

4.8.1. The Need for Feature Selection and Regularization

Not all extracted features are useful for modeling, hence the need for feature selection techniques (METRICS item 15 [20]) [111]. A critical problem in radiomics studies is the high dimensionality of datasets which leads to: (1) a large computational demand and (2) an excessive number of features that can cause a reduction in model performance due to overfitting. Therefore, feature selection methods are used to simplify the model by removing redundant and irrelevant ones [84]. The subset of features obtained through feature selection should be minimal but effective in accurately identifying the target and, ideally, should improve the accuracy of the prediction model [84]. In the process of mapping high-dimensional data into a low-dimensional space, however, there is a risk of losing some important information; for this reason, the choice of the method is a crucial step in the pipeline of a radiomics study [112]. While in traditional radiomics feature selection is an explicit and crucial phase, in deep radiomics it is often incorporated into the learning process of the model itself.

4.8.2. Filter, Wrapper, and Embedded Methods

Feature selection methods can be distinguished based on the relationship with the learning model in filter, wrapper, and embedded methods (Figure 8 and Figure 9) [100].
In filter methods, features are evaluated and ranked based on statistical criteria using predefined thresholds for the chosen statistical measure, independently of any learning algorithm [112,113,114,115] (Figure 9 and Figure 10A, Table 4). This evaluation is performed directly from the tabular data using measures such as correlation coefficients, mutual information, and the χ2 test [25,113,114,115]. When the correlation coefficients are close to or equal to 0 with the target variable, the feature is excluded, as this indicates that it does not significantly contribute to explaining the variations in the target variable. The main advantage of this method is that it is not biased toward any specific learner model and maintains a simple structure.
Wrapper methods use feedback from the learning algorithms to guide feature selection (Figure 10B) [84]. First, specific rules are applied to generate candidate feature subsets. Then, the optimal subset is selected by comparing the performance of models trained on different subsets. Wrapper methods include best subset selection, stepwise model selection, and others (Table 5) [116,117]. The best subset of features can be achieved through an exhaustive search that consists of testing all 2p possible models containing subsets of the p predictors. The main disadvantage of this method is that the increase of combinations leads to an exponential increase in computational costs [25]. Forward and backward stepwise selections are both computationally efficient alternatives (Figure 10B).
Embedded methods combine the advantages of filter and wrapper methods by integrating feature selection directly into the model construction process. This approach eliminates the need for multiple runs of the learning model [113]. Shrinkage methods are a type of embedded method that fit a model using all predictors, but the estimated coefficients are progressively shrunken toward zero depending on a penalty parameter. This process, also known as regularization, reduces variance at the cost of slightly increasing the bias, with the result of significantly improving the model’s performance for new unseen data [76]. According to the type of shrinkage performed, some coefficients may be estimated as exactly zero, allowing these methods to also perform variable selection. In this case, after the feature selection process is finished, the model is trained. A common embedded method is the least absolute shrinkage and selection operator (LASSO) [118], a type of linear regression that incorporates regularization to enhance the model’s prediction accuracy and interpretability (Figure 10C). It achieves this by adding a penalty term to the ordinary least squares objective function, which is the sum of the absolute values of the model coefficients. This penalty term encourages the coefficients of less important features to shrink toward zero. As a result, LASSO regression not only performs regularization, but also variable selection, making the model more interpretable by reducing the number of features. LASSO regression is particularly useful when dealing with high-dimensional data, where the number of predictors can be large. It helps prevent overfitting by reducing model complexity and can improve the model’s prediction accuracy for new, unseen data. However, choosing the appropriate value for λ is crucial and is typically achieved through cross-validation.

4.8.3. Dimensionality Reduction

The optimal number of the subset of features is controversial and depends on the context; generally, the “rule of thumb” of a maximum of ten features for each instance is considered (METRICS item 16 [16,20,34]).
Dimensionality reduction techniques should be mentioned, the best known being principal component analysis (PCA) [20]. PCA maps the original feature space to new coordinates, called principal components, which are linear combinations of the original features, generating a new data representation that preserves the maximum possible variance. Therefore, PCA does not select specific features from the original ones; rather, it transforms the entire set of features into a set of principal components representing the directions of maximum variance in the data [112].

4.9. Modeling

4.9.1. Supervised Versus Unsupervised Learning

Supervised learning is a type of ML in which the model is trained on a labeled dataset, which means that each observation is associated with a known value of the output [25,30]. For example, each subset of features representing a lesion comes with an associated label (malignant/benign) (Figure 11A). The supervised model learns to map from inputs to outputs in order to correctly predict the labels for new unseen data. In practice, the label of the target variable can be assigned by the radiologist, if the imaging diagnosis is adopted as the reference standard, or by the pathologist, if the histopathological diagnosis is adopted. Most radiomics studies fall into this paradigm, and many ML models work in a supervised framework.
On the other hand, unsupervised learning describes a more challenging situation in which the output associated to each observation is unknown [25,30]. In this scenario, for example, it is not possible to fit a linear regression model, since there is no output variable to predict. Unsupervised algorithms extract knowledge directly from the input data, for example by identifying some appropriate classifications between groups of data based on common characteristics (clustering). In this way, the model itself generates the possible values of Y and this can be used to make predictions about the new inputs (Figure 11B).

4.9.2. Regression, Classification, and Clustering Problems

The class of the target variable divides supervised learning problems into two main tasks: problems with a quantitative response are usually referred to as regression problems, and problems involving a qualitative response are often referred to as classification problems (the class being either binary or multi-level) [25,30]. This distinction is key because it guides the choice of the model (or models) to be fitted (Figure 12). Some models are only suitable for a specific task: for example, linear regression is a popular regression technique, while others can be used for both quantitative and qualitative responses [25,30].
Concerning unsupervised learning, the task of main interest in the radiomics field is clustering [79,119]. Clustering models are ML techniques used to group a set of data into homogeneous subgroups called clusters [30]; these models aim to identify structures and patterns within the data, ensuring that elements within a cluster are more similar to each other than to elements in other clusters [30].
Most ML algorithms have one or more adjustable settings, known as hyperparameters, which can be modified to influence the model’s performance. Hyperparameters are parameters external to the model itself that influence its behavior during training, such as the depth of a decision tree or the learning rate of a neural network. In practice, hyperparameter adjustment is shaped by optimizing the model’s performance across multiple subsets of data, often through cross-validation, to find the best configuration that balances accuracy and generalization.

4.9.3. Model Selection

There is no one-size-fits-all ML model (Figure 13). Each algorithm has advantages and disadvantages, and its performance is determined by the type of data used for training and the characteristics of the problem being solved (Figure 14).

4.9.4. Data Partition for Training, Validation, and Test Phases

A standard ML workflow includes training, validation, and testing phases, which require appropriate data partition (Figure 15) [25,30]. In the training phase, the model learns patterns from the data, adjusting its parameters to minimize predictive errors. In the validation phase, the model’s performance is evaluated on a separate dataset, which is used to optimize hyperparameters and prevent overfitting. In radiomics and ML pipelines, the concept of “validation” can often be misunderstood, hence terms like “validation set” and “test set” are used interchangeably in some studies, despite their distinct purposes. The training set is used to develop the model, while the validation set is used to tune parameters and prevent overfitting during model development [14]. In the test phase, the final performance of the model is measured on an independent dataset that was not used during training or validation, to assess its ability to generalize to new data [14,25,30].
In radiomics studies, the level at which data are split is a crucial aspect of data preparation and significantly impacts the validity and generalizability of the results. The data can be split at various levels, such as patient-wise, image-wise, study-wise, scanner-wise, or institution-wise, each having distinct implications. In radiomics studies, it is crucial to split data at the patient-level, ensuring that all information for a single patient is contained within the same partition (METRICS item 18 [20]). This approach prevents biases in model integrity and avoids overfitting, which can compromise performance evaluation.

4.9.5. Validation

Validation consists in the process of fine-tuning the model’s architecture by adjusting hyperparameters to prevent overfitting [14,25,30]. It is performed using an independent dataset, distinct from the training set, or through resampling methods, like cross-validation (CV), where the model is run multiple times on different subsets of the training data. This process ensure that the model does not become too closely aligned with any specific patterns in the training data, thereby enhancing its generalizability to unseen data.
Each validation method (Figure 16) has advantages and disadvantages (Table 6) [14,25,30,120]. The simplest approach consists of evaluating the model on a subset of data left aside specifically for this purpose (hold-out dataset). In practice, to avoid sampling issues, resampling methods are commonly preferred: they consist of repeatedly taking samples from a training set and re-fitting a model of interest on each sample in order to obtain additional information about the fitted model.
Once the training and validation datasets are defined, it is important to verify that their feature distributions are similar. This ensures that patterns observed in the training data are likely present in the validation data. Common univariate tests, such as the Mann–Whitney U, Kolmogorov–Smirnov, or Shapiro–Wilk tests, are used to compare medians or distributions. These tests are unsupervised, meaning they do not use outcome data and do not violate the rule of reserving the validation set for model testing [14,16].

4.10. Model Testing and Performance Metrics

After a model has been trained and validated, testing is a crucial phase to provide a reliable estimate of its performances and generalization capability (METRICS items 26 and 27 [20]) [20,121,122].
Typically the test set is a partition of the initial dataset, kept separate and blind from the training and validation data (internal test set). Ideally, to avoid random sampling issues, an external test set acquired from a different center should be used to better assess the model’s performance in a real-world scenario (METRICS 27 [20]). Additionally, it is also possible to acquire new data with a temporal gap after model development; this “prospective test set” allows for an ongoing evaluation of the model’s performance as new data becomes available.
The use of appropriate performance metrics is necessary to correctly evaluate models’ performances (Table 7) (METRICS item 20 [20]) [122].

4.11. Model Uncertainty Assessment and Calibration

Uncertainty assessment (METRICS item 21 [20]) and model calibration (METRICS item 22 [20]) are crucial steps for enhancing the reliability and interpretability of ML models, especially in high-stakes fields, like medical imaging and radiomics [20].
The uncertainty assessment involves quantifying the confidence intervals of a model’s predictions, which can be achieved through techniques such as Bayesian inference, bootstrapping, or ensemble methods [25,123,124]. Bootstrapping is one of the most popular techniques and involves running multiple tests by resampling with replacement from the testing dataset [124].
Calibration, on the other hand, focuses on aligning the predicted probabilities with actual outcomes [20,125,126]. A well-calibrated model produces probability estimates that reflect true likelihoods, which is vital for risk assessment and decision-making processes. Calibration techniques, such as Platt scaling or isotonic regression, adjust the model’s output probabilities to better match the observed frequencies [125].

4.12. Model Comparison

Many radiomic studies develop several models to proof the added value of a radiomics approach, incorporating different types of predictors, such as clinical predictors only, radiomic features only, or a combination of radiomic and clinical variables. Therefore, model comparison is essential to identify the most effective predictive models and assess the superiority of the radiomics approach to traditional ones.
In multi-parametric studies, such as those involving different MRI sequences, like T2-weighted and diffusion-weighted images, it is important to perform also uni-parametric evaluations to demonstrate the added value and provide a clear justification for adopting a multi-parametric model (METRICS item 23 [20]).
Moreover, studies should compare their methods to non-radiomic approaches commonly used in clinical practice, such as those including semantic features, RADS or RECIST scoring, and simple volume or size evaluations (METRICS item 24 [20]) [39,127]. If non-radiomic methods are not available, proof of improved diagnostic accuracy or patient outcomes, such as improved radiologist performance or overall survival, should be evaluated.
Additionally, studies should include a comparison with a simple baseline reference model, such as a Zero Rules/No Information Rate classifier, to justify the use of more complex ML methods by demonstrating increased performance (METRICS item 25 [20]) [128,129,130].
A comparison of different models should use appropriate statistical methods. The DeLong test is a widely used non-parametric approach to compare the areas under the ROC (receiver operating characteristic) curves of two or more classification models [131,132]. The McNemar’s test is used for paired nominal data to determine whether there are differences in a dichotomous dependent variable between two related groups and is often applied to the confusion matrices directly by evaluating the differences in their misclassification rates [133]. Other methods include decision curve analysis for net benefit comparison, and the Net Reclassification Index [134,135,136]. To address the issue of multiple comparisons, which can inflate the risk of type I errors, multiple testing correction methods, such as Bonferroni, Benjamini–Hochberg, or Tukey, are applied [136].

5. Challenges and Future Perspectives: Collaborative Science and Clinical Translatability

In this review, we outlined a comprehensive guide that covers every stage of the machine learning workflow in radiomics studies, with specific references to the METRICS tool [20].
In recent decades, there has been an ongoing debate about the reliability of scientific publications, with concerns that a significant amount of research may be at least inadequately supported by the data and methodologies [137]. Radiomics is not exempt from this scrutiny, and a clear concern is the discrepancy between the large number of published radiomics studies and the limited clinical implementation of these models [138,139]. Consequently, a rigorous evaluation of the methodology in radiomics studies is essential to generate more robust evidence for AI’s efficacy and clinical application (METRICS item 30 [20]). However, it is important to remember that the challenges extend beyond the methodology; economic, organizational, educational, and regulatory issues must also be addressed [6,9,140,141].
In 2017, the Radiomics Quality Score (RQS) was introduced to assess the quality of the radiomics workflow across steps such as image protocol, feature selection, model building, validation, and cost-effectiveness, providing a cumulative percentage score for evaluating radiomics studies [10,14,142]. In 2020, the USA-based CLAIM and MINIMAR checklists were introduced to define the best AI practices, particularly stressing the use of independent datasets for model training and validation. More recently, the METRICS tool was developed through a large international expert panel using a modified Delphi protocol to assess the quality of radiomics research [20]. Guidelines like SPIRIT-AI, CONSORT-AI, and DECIDE-AI offer recommendations for evaluating AI tools in trials [143,144,145].
Some authors believe that the main obstacle to clinical implementation is the lack of model-, code-, and data-sharing practices [138]. Transparency is essential for ensuring reproducibility and comparability across studies, which is a cornerstone of robust scientific research. For this reason, the availability of imaging data (METRICS item 28 [20]) and code (METRICS item 29 [20]) is crucial in radiomics research.
Open code and software enable researchers to replicate studies, independently validate findings, and build upon existing work [146,147]. By making code publicly available, researchers can identify and correct errors, optimize algorithms, and enhance the overall robustness of radiomics models. The introduction of large, open source image repositories offers access to diverse and extensive datasets, addressing the key challenges of radiomics related to reproducibility and standardization [148,149,150]. Recently, RadiomicsHub was proposed as a comprehensive public repository, featuring radiomics data derived from a systematic review of public cancer imaging datasets [151]. The repository includes 29 datasets with segmentations and labels for health outcomes, tumor pathology, and imaging-based scores, encompassing 10,354 patients and 49,515 scans. Of these, 15 datasets are under Creative Commons licenses, while the rest have custom or restricted licensing.
Finally, open datasets also contribute to democratizing research, allowing smaller institutions and less-funded researchers to contribute to and benefit from the collective knowledge base. Ultimately, collaborative science is essential to overcome the need for multidisciplinary expertise in radiomics [131,132,133]. This collective cross-fertilization approach of engineers, physicians, and biostatisticians accelerates the advancement and clinical implementation of radiomics models. The use of cloud-based infrastructures stands as the new frontier for implementing and sharing transparent, reproducible AI-based radiology pipelines, and can further enhanced radiomics research [147]. By providing a consistent computational environment, streamlining data exploration and access, and enabling the storage and sharing of code and results, these platforms support the creation and dissemination of fully reproducible AI pipelines that enhance original research publications [147].

Author Contributions

Conceptualization, M.C. (Maurizio Cè) and M.C. (Michaela Cellina); methodology, M.D.C. and M.C. (Maurizio Cè); validation, M.C. (Maurizio Cè), M.D.C. and A.C.; formal analysis, M.D.C.; writing—original draft preparation, L.M., F.L.R., G.I. and M.C. (Michaela Cellina); writing—review and editing, M.C. (Maurizio Cè) and A.C.; visualization, M.C. (Maurizio Cè); supervision, G.C.; project administration, D.F. and G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ding, H.; Wu, C.; Liao, N.; Zhan, Q.; Sun, W.; Huang, Y.; Jiang, Z.; Li, Y. Radiomics in Oncology: A 10-Year Bibliometric Analysis. Front. Oncol. 2021, 11, 689802. [Google Scholar] [CrossRef] [PubMed]
  2. Volpe, S.; Mastroleo, F.; Krengli, M.; Jereczek-Fossa, B.A. Quo vadis Radiomics? Bibliometric analysis of 10-year Radiomics journey. Eur. Radiol. 2023, 33, 6736–6745. [Google Scholar] [CrossRef] [PubMed]
  3. Polidori, T.; De Santis, D.; Rucci, C.; Tremamunno, G.; Piccinni, G.; Pugliese, L.; Zerunian, M.; Guido, G.; Pucciarelli, F.; Bracci, B.; et al. Radiomics applications in cardiac imaging: A comprehensive review. Radiol. Med. 2023, 128, 922–933. [Google Scholar] [CrossRef]
  4. Lambin, P.; Rios-Velazquez, E.; Leijenaar, R.; Carvalho, S.; van Stiphout, R.G.P.M.; Granton, P.; Zegers, C.M.L.; Gillies, R.; Boellard, R.; Dekker, A.; et al. Radiomics: Extracting more information from medical images using advanced feature analysis. Eur. J. Cancer 2012, 48, 441–446. [Google Scholar] [CrossRef]
  5. Mayerhoefer, M.E.; Materka, A.; Langs, G.; Häggström, I.; Szczypiński, P.; Gibbs, P.; Cook, G. Introduction to Radiomics. J. Nucl. Med. 2020, 61, 488–495. [Google Scholar] [CrossRef]
  6. Cè, M.; Ibba, S.; Cellina, M.; Tancredi, C.; Fantesini, A.; Fazzini, D.; Fortunati, A.; Perazzo, C.; Presta, R.; Montanari, R.; et al. Radiologists’ perceptions on AI integration: An in-depth survey study. Eur. J. Radiol. 2024, 177, 111590. [Google Scholar] [CrossRef]
  7. Pesapane, F.; Codari, M.; Sardanelli, F. Artificial intelligence in medical imaging: Threat or opportunity? Radiologists again at the forefront of innovation in medicine. Eur. Radiol. Exp. 2018, 2, 35. [Google Scholar] [CrossRef]
  8. Castiglioni, I.; Rundo, L.; Codari, M.; Di Leo, G.; Salvatore, C.; Interlenghi, M.; Gallivanone, F.; Cozzi, A.; D’Amico, N.C.; Sardanelli, F. AI applications to medical images: From machine learning to deep learning. Phys. Medica 2021, 83, 9–24. [Google Scholar] [CrossRef]
  9. Sardanelli, F.; Colarieti, A. Open issues for education in radiological research: Data integrity, study reproducibility, peer-review, levels of evidence, and cross-fertilization with data scientists. Radiol. Med. 2022, 128, 133–135. [Google Scholar] [CrossRef]
  10. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.T.H.M.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef]
  11. Hatt, M.; Le Rest, C.C.; Tixier, F.; Badic, B.; Schick, U.; Visvikis, D. Radiomics: Data Are Also Images. J. Nucl. Med. 2019, 60, 38S–44S. [Google Scholar] [CrossRef] [PubMed]
  12. Li, S.; Zhou, B. A review of radiomics and genomics applications in cancers: The way towards precision medicine. Radiat. Oncol. 2022, 17, 217. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, Z.; Wang, S.; Dong, D.; Wei, J.; Fang, C.; Zhou, X.; Sun, K.; Li, L.; Li, B.; Wang, M.; et al. The Applications of Radiomics in Precision Diagnosis and Treatment of Oncology: Opportunities and Challenges. Theranostics 2019, 9, 1303–1322. [Google Scholar] [CrossRef] [PubMed]
  14. Stanzione, A.; Cuocolo, R.; Ugga, L.; Verde, F.; Romeo, V.; Brunetti, A.; Maurea, S. Oncologic Imaging and Radiomics: A Walkthrough Review of Methodological Challenges. Cancers 2022, 14, 4871. [Google Scholar] [CrossRef] [PubMed]
  15. Van Timmeren, J.E.; Cester, D.; Tanadini-Lang, S.; Alkadhi, H.; Baessler, B. Radiomics in medical imaging-”how-to” guide and critical reflection. Insights Imaging 2020, 11, 91. [Google Scholar] [CrossRef]
  16. Shur, J.D.; Doran, S.J.; Kumar, S.; Ap Dafydd, D.; Downey, K.; O’connor, J.P.B.; Papanikolaou, N.; Messiou, C.; Koh, D.-M.; Orton, M.R. Radiomics in Oncology: A Practical Guide. RadioGraphics 2021, 41, 1717–1732. [Google Scholar] [CrossRef]
  17. Kocak, B.; Durmaz, E.S.; Ates, E.; Kilickesmez, O. Radiomics with artificial intelligence: A practical guide for beginners. Diagn. Interv. Radiol. 2019, 25, 485–495. [Google Scholar] [CrossRef]
  18. Zitnik, M.; Nguyen, F.; Wang, B.; Leskovec, J.; Goldenberg, A.; Hoffman, M.M. Machine learning for integrating data in biology and medicine: Principles, practice, and opportunities. Inf. Fusion 2019, 50, 71–91. [Google Scholar] [CrossRef]
  19. Scapicchio, C.; Gabelloni, M.; Barucci, A.; Cioni, D.; Saba, L.; Neri, E. A deep look into radiomics. Radiol. Med. 2021, 126, 1296–1311. [Google Scholar] [CrossRef]
  20. Kocak, B.; D’antonoli, T.A.; Mercaldo, N.; Alberich-Bayarri, A.; Baessler, B.; Ambrosini, I.; Andreychenko, A.E.; Bakas, S.; Beets-Tan, R.G.H.; Bressem, K.; et al. METhodological RadiomICs Score (METRICS): A quality scoring tool for radiomics research endorsed by EuSoMII. Insights Imaging 2024, 15, 8. [Google Scholar] [CrossRef]
  21. Majumder, S.; Katz, S.; Kontos, D.; Roshkovan, L. State of the art: Radiomics and radiomics-related artificial intelligence on the road to clinical translation. BJR|Open 2023, 6, tzad004. [Google Scholar] [CrossRef] [PubMed]
  22. Hsiao, J.H. Understanding Human Cognition Through Computational Modeling. Top. Cogn. Sci. 2024, 16, 349–376. [Google Scholar] [CrossRef] [PubMed]
  23. Kufel, J.; Bargieł-Łączek, K.; Kocot, S.; Koźlik, M.; Bartnikowska, W.; Janik, M.; Czogalik, A.; Dudek, P.; Magiera, M.; Lis, A.; et al. What Is Machine Learning, Artificial Neural Networks and Deep Learning?—Examples of Practical Applications in Medicine. Diagnostics 2023, 13, 2582. [Google Scholar] [CrossRef]
  24. Spicer, J.; Sanborn, A.N. What does the mind learn? A comparison of human and machine learning representations. Curr. Opin. Neurobiol. 2019, 55, 97–102. [Google Scholar] [CrossRef]
  25. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning: With Applications in R; Springer: New York, NY, USA, 2013. [Google Scholar]
  26. Lu, S.-C.; Swisher, C.L.; Chung, C.; Jaffray, D.; Sidey-Gibbons, C. On the importance of interpretable machine learning predictions to inform clinical decision making in oncology. Front. Oncol. 2023, 13, 1129380. [Google Scholar] [CrossRef]
  27. Cheng, P.M.; Montagnon, E.; Yamashita, R.; Pan, I.; Cadrin-Chênevert, A.; Romero, F.P.; Chartrand, G.; Kadoury, S.; Tang, A. Deep Learning: An Update for Radiologists. RadioGraphics 2021, 41, 1427–1445. [Google Scholar] [CrossRef]
  28. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2020, 23, 18. [Google Scholar] [CrossRef]
  29. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  30. Tan, P.-N.; Steinbach, M.; Karpatne, A.; Kumar, V. Introduction to Data Mining, 2nd ed.; Pearson Education: London, UK, 2019. [Google Scholar]
  31. Larrañaga, P.; Calvo, B.; Santana, R.; Bielza, C.; Galdiano, J.; Inza, I.; Lozano, J.A.; Armañanzas, R.; Santafé, G.; Pérez, A.; et al. Machine learning in bioinformatics. Brief. Bioinform. 2006, 7, 86–112. [Google Scholar] [CrossRef]
  32. Jiang, T.; Gradus, J.L.; Rosellini, A.J. Supervised Machine Learning: A Brief Primer. Behav. Ther. 2020, 51, 675–687. [Google Scholar] [CrossRef]
  33. Mongan, J.; Moy, L.; Kahn, C.E. Checklist for Artificial Intelligence in Medical Imaging (CLAIM): A Guide for Authors and Reviewers. Radiol. Artif. Intell. 2020, 2, e200029. [Google Scholar] [CrossRef] [PubMed]
  34. Kocak, B.; Baessler, B.; Bakas, S.; Cuocolo, R.; Fedorov, A.; Maier-Hein, L.; Mercaldo, N.; Müller, H.; Orlhac, F.; dos Santos, D.P.; et al. CheckList for EvaluAtion of Radiomics research (CLEAR): A step-by-step reporting guideline for authors and reviewers endorsed by ESR and EuSoMII. Insights Imaging 2023, 14, 75. [Google Scholar] [CrossRef] [PubMed]
  35. Cui, Y.; Yin, F.-F. Impact of image quality on radiomics applications. Phys. Med. Biol. 2022, 67, 15TR03. [Google Scholar] [CrossRef] [PubMed]
  36. Orlhac, F.; Lecler, A.; Savatovski, J.; Goya-Outi, J.; Nioche, C.; Charbonneau, F.; Ayache, N.; Frouin, F.; Duron, L.; Buvat, I. How can we combat multicenter variability in MR radiomics? Validation of a correction procedure. Eur. Radiol. 2021, 31, 2272–2280. [Google Scholar] [CrossRef]
  37. Campello, V.M.; Martín-Isla, C.; Izquierdo, C.; Guala, A.; Palomares, J.F.R.; Viladés, D.; Descalzo, M.L.; Karakas, M.; Çavuş, E.; Raisi-Estabragh, Z.; et al. Minimising multi-centre radiomics variability through image normalisation: A pilot study. Sci. Rep. 2022, 12, 12532. [Google Scholar] [CrossRef]
  38. Weinreb, J.C.; Barentsz, J.O.; Choyke, P.L.; Cornud, F.; Haider, M.A.; Macura, K.J.; Margolis, D.; Schnall, M.D.; Shtern, F.; Tempany, C.M.; et al. PI-RADS Prostate Imaging—Reporting and Data System: 2015, Version 2. Eur. Urol. 2016, 69, 16–40. [Google Scholar] [CrossRef]
  39. An, J.Y.; Unsdorfer, K.M.L.; Weinreb, J.C. BI-RADS, C-RADS, CAD-RADS, LI-RADS, Lung-RADS, NI-RADS, O-RADS, PI-RADS, TI-RADS: Reporting and Data Systems. RadioGraphics 2019, 39, 1435–1436. [Google Scholar] [CrossRef]
  40. Park, D.; Oh, D.; Lee, M.; Lee, S.Y.; Shin, K.M.; Jun, J.S.; Hwang, D. Importance of CT image normalization in radiomics analysis: Prediction of 3-year recurrence-free survival in non-small cell lung cancer. Eur. Radiol. 2022, 32, 8716–8725. [Google Scholar] [CrossRef]
  41. Moradmand, H.; Aghamiri, S.M.R.; Ghaderi, R. Impact of image preprocessing methods on reproducibility of radiomic features in multimodal magnetic resonance imaging in glioblastoma. J. Appl. Clin. Med. Phys. 2020, 21, 179–190. [Google Scholar] [CrossRef]
  42. Um, H.; Tixier, F.; Bermudez, D.; Deasy, J.O.; Young, R.J.; Veeraraghavan, H. Impact of image preprocessing on the scanner dependence of multi-parametric MRI radiomic features and covariate shift in multi-institutional glioblastoma datasets. Phys. Med. Biol. 2019, 64, 165011. [Google Scholar] [CrossRef]
  43. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 4th ed.; Pearson: London, UK, 2018. [Google Scholar]
  44. Ellingson, B.M.; Zaw, T.; Cloughesy, T.F.; Naeini, K.M.; Lalezari, S.; Mong, S.; Lai, A.; Nghiemphu, P.L.; Pope, W.B. Comparison between intensity normalization techniques for dynamic susceptibility contrast (DSC)-MRI estimates of cerebral blood volume (CBV) in human gliomas. J. Magn. Reson. Imaging 2012, 35, 1472–1477. [Google Scholar] [CrossRef] [PubMed]
  45. Depeursinge, A.; Al-Kadi, O.; Mitchell, J. Biomedical Texture Analysis: Fundamentals, Tools and Challenges; Academic Press: London, UK, 2017. [Google Scholar]
  46. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python library for efficient loading, preprocessing, augmentation and patch-based sampling of medical images in deep learning. Comput. Methods Programs Biomed. 2021, 208, 106236. [Google Scholar] [CrossRef] [PubMed]
  47. Li, Y.; Tan, G.; Vangel, M.; Hall, J.; Cai, W. Influence of feature calculating parameters on the reproducibility of CT radiomic features: A thoracic phantom study. Quant. Imaging Med. Surg. 2020, 10, 1775–1785. [Google Scholar] [CrossRef] [PubMed]
  48. Duron, L.; Balvay, D.; Perre, S.V.; Bouchouicha, A.; Savatovsky, J.; Sadik, J.-C.; Thomassin-Naggara, I.; Fournier, L.; Lecler, A. Gray-level discretization impacts reproducible MRI radiomics texture features. PLoS ONE 2019, 14, e0213459. [Google Scholar] [CrossRef]
  49. Loi, S.; Mori, M.; Palumbo, D.; Crippa, S.; Palazzo, G.; Spezi, E.; Del Vecchio, A.; Falconi, M.; De Cobelli, F.; Fiorino, C. Limited impact of discretization/interpolation parameters on the predictive power of CT radiomic features in a surgical cohort of pancreatic cancer patients. Radiol. Med. 2023, 128, 799–807. [Google Scholar] [CrossRef]
  50. Liberini, V.; De Santi, B.; Rampado, O.; Gallio, E.; Dionisi, B.; Ceci, F.; Polverari, G.; Thuillier, P.; Molinari, F.; Deandreis, D. Impact of segmentation and discretization on radiomic features in 68Ga-DOTA-TOC PET/CT images of neuroendocrine tumor. EJNMMI Phys. 2021, 8, 21. [Google Scholar] [CrossRef]
  51. Larue, R.T.H.M.; van Timmeren, J.E.; de Jong, E.E.C.; Feliciani, G.; Leijenaar, R.T.H.; Schreurs, W.M.J.; Sosef, M.N.; Raat, F.H.P.J.; van der Zande, F.H.R.; Das, M.; et al. Influence of gray level discretization on radiomic feature stability for different CT scanners, tube currents and slice thicknesses: A comprehensive phantom study. Acta Oncol. 2017, 56, 1544–1553. [Google Scholar] [CrossRef]
  52. Carré, A.; Klausner, G.; Edjlali, M.; Lerousseau, M.; Briend-Diop, J.; Sun, R.; Ammari, S.; Reuzé, S.; Andres, E.A.; Estienne, T.; et al. Standardization of brain MR images across machines and protocols: Bridging the gap for MRI-based radiomics. Sci. Rep. 2020, 10, 12340. [Google Scholar] [CrossRef]
  53. Stefano, A.; Leal, A.; Richiusa, S.; Trang, P.; Comelli, A.; Benfante, V.; Cosentino, S.; Sabini, M.G.; Tuttolomondo, A.; Altieri, R.; et al. Robustness of PET Radiomics Features: Impact of Co-Registration with MRI. Appl. Sci. 2021, 11, 10170. [Google Scholar] [CrossRef]
  54. Jiao, F.; Wang, M.; Sun, X.; Ju, Z.; Lu, J.; Wang, L.; Jiang, J.; Zuo, C. Based on Tau PET Radiomics Analysis for the Classification of Alzheimer’s Disease and Mild Cognitive Impairment. Brain Sci. 2023, 13, 367. [Google Scholar] [CrossRef]
  55. Mandal, P.K.; Mahajan, R.; Dinov, I.D. Structural Brain Atlases: Design, Rationale, and Applications in Normal and Pathological Cohorts. J. Alzheimer’s Dis. 2012, 31, S169–S188. [Google Scholar] [CrossRef] [PubMed]
  56. Bleker, J.; Roest, C.; Yakar, D.; Huisman, H.; Kwee, T.C. The Effect of Image Resampling on the Performance of Radiomics-Based Artificial Intelligence in Multicenter Prostate MRI. J. Magn. Reson. Imaging 2023, 59, 1800–1806. [Google Scholar] [CrossRef] [PubMed]
  57. Schick, U.; Lucia, F.; Dissaux, G.; Visvikis, D.; Badic, B.; Masson, I.; Pradier, O.; Bourbonne, V.; Hatt, M. MRI-derived radiomics: Methodology and clinical applications in the field of pelvic oncology. Br. J. Radiol. 2019, 92, 20190105. [Google Scholar] [CrossRef] [PubMed]
  58. Demircioğlu, A. The effect of preprocessing filters on predictive performance in radiomics. Eur. Radiol. Exp. 2022, 6, 40. [Google Scholar] [CrossRef]
  59. Lo, S.-C.; Li, H.; Freedman, M. Optimization of wavelet decomposition for image compression and feature preservation. IEEE Trans. Med. Imaging 2003, 22, 1141–1151. [Google Scholar] [CrossRef]
  60. Lunscher, W.H.H.J.; Beddoes, M.P. Optimal Edge Detector Design I: Parameter Selection and Noise Effects. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 164–177. [Google Scholar] [CrossRef]
  61. Kumar, A. Study and analysis of different segmentation methods for brain tumor MRI application. Multimed. Tools Appl. 2023, 82, 7117–7139. [Google Scholar] [CrossRef]
  62. du Plessis, T.; Ramkilawon, G.; Rae, W.I.D.; Botha, T.; Martinson, N.A.; Dixon, S.A.P.; Kyme, A.; Sathekge, M.M. Introducing a secondary segmentation to construct a radiomics model for pulmonary tuberculosis cavities. Radiol. Med. 2023, 128, 1093–1102. [Google Scholar] [CrossRef]
  63. Kim, D.H.; Kim, Y.J.; Kim, K.G.; Jeon, J.Y. Automated Vertebral Segmentation and Measurement of Vertebral Compression Ratio Based on Deep Learning in X-Ray Images. J. Digit. Imaging 2021, 34, 853–861. [Google Scholar] [CrossRef]
  64. Yu, Y.; Wang, C.; Fu, Q.; Kou, R.; Huang, F.; Yang, B.; Yang, T.; Gao, M. Techniques and Challenges of Image Segmentation: A Review. Electronics 2023, 12, 1199. [Google Scholar] [CrossRef]
  65. Wang, Z.; Gu, S.; Leader, J.K.; Kundu, S.; Tedrow, J.R.; Sciurba, F.C.; Gur, D.; Siegfried, J.M.; Pu, J. Optimal threshold in CT quantification of emphysema. Eur. Radiol. 2013, 23, 975–984. [Google Scholar] [CrossRef] [PubMed]
  66. Fedorov, A.; Beichel, R.; Kalpathy-Cramer, J.; Finet, J.; Fillion-Robin, J.-C.; Pujol, S.; Bauer, C.; Jennings, D.; Fennessy, F.; Sonka, M.; et al. 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn. Reson. Imaging 2012, 30, 1323–1341. [Google Scholar] [CrossRef] [PubMed]
  67. Dionisio, F.C.F.; Oliveira, L.S.; Hernandes, M.d.A.; Engel, E.E.; de Azevedo-Marques, P.M.; Nogueira-Barbosa, M.H. Manual versus semiautomatic segmentation of soft-tissue sarcomas on magnetic resonance imaging: Evaluation of similarity and comparison of segmentation times. Radiol. Bras. 2021, 54, 155–164. [Google Scholar] [CrossRef] [PubMed]
  68. Sharma, S.R.; Alshathri, S.; Singh, B.; Kaur, M.; Mostafa, R.R.; El-Shafai, W. Hybrid Multilevel Thresholding Image Segmentation Approach for Brain MRI. Diagnostics 2023, 13, 925. [Google Scholar] [CrossRef]
  69. Amer, G.M.H.; Abushaala, A.M. Edge detection methods. In Proceedings of the 2015 2nd World Symposium on Web Applications and Networking (WSWAN), Sousse, Tunisia, 21–23 March 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–7. [Google Scholar]
  70. Sammouda, R.; El-Zaart, A. An Optimized Approach for Prostate Image Segmentation Using K-Means Clustering Algorithm with Elbow Method. Comput. Intell. Neurosci. 2021, 2021, 1–13. [Google Scholar] [CrossRef]
  71. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 2021, 1. [Google Scholar] [CrossRef]
  72. Tajbakhsh, N.; Jeyaseelan, L.; Li, Q.; Chiang, J.N.; Wu, Z.; Ding, X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med. Image Anal. 2020, 63, 101693. [Google Scholar] [CrossRef]
  73. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  74. Zijdenbos, A.; Dawant, B.; Margolin, R.; Palmer, A. Morphometric analysis of white matter lesions in MR images: Method and validation. IEEE Trans. Med. Imaging 1994, 13, 716–724. [Google Scholar] [CrossRef]
  75. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images Are More than Pictures, They Are Data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef]
  76. O’Connor, J.P.B.; Aboagye, E.O.; Adams, J.E.; Aerts, H.J.W.L.; Barrington, S.F.; Beer, A.J.; Boellaard, R.; Bohndiek, S.E.; Brady, M.; Brown, G.; et al. Imaging biomarker roadmap for cancer studies. Nat. Rev. Clin. Oncol. 2017, 14, 169–186. [Google Scholar] [CrossRef] [PubMed]
  77. Tomaszewski, M.R.; Gillies, R.J. The Biological Meaning of Radiomic Features. Radiology 2021, 298, 505–516. [Google Scholar] [CrossRef] [PubMed]
  78. Demircioğlu, A. Deep Features from Pretrained Networks Do Not Outperform Hand-Crafted Features in Radiomics. Diagnostics 2023, 13, 3266. [Google Scholar] [CrossRef] [PubMed]
  79. Parmar, C.; Leijenaar, R.T.H.; Grossmann, P.; Velazquez, E.R.; Bussink, J.; Rietveld, D.; Rietbergen, M.M.; Haibe-Kains, B.; Lambin, P.; Aerts, H.J. Radiomic feature clusters and Prognostic Signatures specific for Lung and Head & Neck cancer. Sci. Rep. 2015, 5, srep11044. [Google Scholar] [CrossRef]
  80. van Griethuysen, J.J.M.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.H.; Fillion-Robin, J.-C.; Pieper, S.; Aerts, H.J.W.L. Computational Radiomics System to Decode the Radiographic Phenotype. Cancer Res. 2017, 77, e104–e107. [Google Scholar] [CrossRef]
  81. Aerts, H.J.W.L.; Velazquez, E.R.; Leijenaar, R.T.H.; Parmar, C.; Grossmann, P.; Carvalho, S.; Bussink, J.; Monshouwer, R.; Haibe-Kains, B.; Rietveld, D.; et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 2014, 5, 4006. [Google Scholar] [CrossRef]
  82. Grossmann, P.; Stringfield, O.; El-Hachem, N.; Bui, M.M.; Velazquez, E.R.; Parmar, C.; Leijenaar, R.T.; Haibe-Kains, B.; Lambin, P.; Gillies, R.J.; et al. Defining the biological basis of radiomic phenotypes in lung cancer. eLife 2017, 6, 23421. [Google Scholar] [CrossRef]
  83. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural Features for Image Classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef]
  84. Zhang, W.; Guo, Y.; Jin, Q. Radiomics and Its Feature Selection: A Review. Symmetry 2023, 15, 1834. [Google Scholar] [CrossRef]
  85. Zwanenburg, A.; Vallières, M.; Abdalah, M.A.; Aerts, H.J.W.L.; Andrearczyk, V.; Apte, A.; Ashrafinia, S.; Bakas, S.; Beukinga, R.J.; Boellaard, R.; et al. The Image Biomarker Standardization Initiative: Standardized Quantitative Radiomics for High-Throughput Image-based Phenotyping. Radiology 2020, 295, 328–338. [Google Scholar] [CrossRef]
  86. Li, Y.; Chen, C.-Y.; Wasserman, W.W. Deep Feature Selection: Theory and Application to Identify Enhancers and Promoters. J. Comput. Biol. 2016, 23, 322–336. [Google Scholar] [CrossRef] [PubMed]
  87. Tripathi, S.; Fritz, B.A.; Abdelhack, M.; Avidan, M.S.; Chen, Y.; King, C.R. Multi-view representation learning for tabular data integration using inter-feature relationships. J. Biomed. Inform. 2024, 151, 104602. [Google Scholar] [CrossRef] [PubMed]
  88. Smith, S.M.; Nichols, T.E. Statistical Challenges in “Big Data” Human Neuroimaging. Neuron 2018, 97, 263–268. [Google Scholar] [CrossRef] [PubMed]
  89. Jager, K.; Zoccali, C.; MacLeod, A.; Dekker, F. Confounding: What it is and how to deal with it. Kidney Int. 2008, 73, 256–260. [Google Scholar] [CrossRef]
  90. Rios, R.; Miller, R.J.; Manral, N.; Sharir, T.; Einstein, A.J.; Fish, M.B.; Ruddy, T.D.; Kaufmann, P.A.; Sinusas, A.J.; Miller, E.J.; et al. Handling missing values in machine learning to predict patient-specific risk of adverse cardiac events: Insights from REFINE SPECT registry. Comput. Biol. Med. 2022, 145, 105449. [Google Scholar] [CrossRef]
  91. Heymans, M.W.; Twisk, J.W. Handling missing data in clinical research. J. Clin. Epidemiol. 2022, 151, 185–188. [Google Scholar] [CrossRef]
  92. Ahmadian, M.; Bodalal, Z.; van der Hulst, H.J.; Vens, C.; Karssemakers, L.H.; Bogveradze, N.; Castagnoli, F.; Landolfi, F.; Hong, E.K.; Gennaro, N.; et al. Overcoming data scarcity in radiomics/radiogenomics using synthetic radiomic features. Comput. Biol. Med. 2024, 174, 108389. [Google Scholar] [CrossRef]
  93. Park, C.J.; Park, Y.W.; Ahn, S.S.; Kim, D.; Kim, E.H.; Kang, S.-G.; Chang, J.H.; Kim, S.H.; Lee, S.-K. Quality of Radiomics Research on Brain Metastasis: A Roadmap to Promote Clinical Translation. Korean J. Radiol. 2022, 23, 77–88. [Google Scholar] [CrossRef]
  94. Zwanenburg, A.; Leger, S.; Agolli, L.; Pilz, K.; Troost, E.G.C.; Richter, C.; Löck, S. Assessing robustness of radiomic features by image perturbation. Sci. Rep. 2019, 9, 614. [Google Scholar] [CrossRef]
  95. Guo, K.; Chen, J.; Qiu, T.; Guo, S.; Luo, T.; Chen, T.; Ren, S. MedGAN: An adaptive GAN approach for medical image generation. Comput. Biol. Med. 2023, 163, 107119. [Google Scholar] [CrossRef]
  96. van Stralen, K.; Dekker, F.; Zoccali, C.; Jager, K. Confounding. Nephron Clin. Pract. 2010, 116, c143–c147. [Google Scholar] [CrossRef] [PubMed]
  97. Chyzhyk, D.; Varoquaux, G.; Milham, M.; Thirion, B. How to remove or control confounds in predictive models, with applications to brain biomarkers. GigaScience 2022, 11, 14. [Google Scholar] [CrossRef] [PubMed]
  98. Spisak, T. Statistical quantification of confounding bias in machine learning models. GigaScience 2022, 11, 82. [Google Scholar] [CrossRef]
  99. Qu, W.; Balki, I.; Mendez, M.; Valen, J.; Levman, J.; Tyrrell, P.N. Assessing and mitigating the effects of class imbalance in machine learning with application to X-ray imaging. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 2041–2048. [Google Scholar] [CrossRef]
  100. Thölke, P.; Mantilla-Ramos, Y.-J.; Abdelhedi, H.; Maschke, C.; Dehgan, A.; Harel, Y.; Kemtur, A.; Berrada, L.M.; Sahraoui, M.; Young, T.; et al. Class imbalance should not throw you off balance: Choosing the right classifiers and performance metrics for brain decoding with imbalanced data. NeuroImage 2023, 277, 120253. [Google Scholar] [CrossRef]
  101. Hajianfar, G.; Hosseini, S.A.; Bagherieh, S.; Oveisi, M.; Shiri, I.; Zaidi, H. Impact of harmonization on the reproducibility of MRI radiomic features when using different scanners, acquisition parameters, and image pre-processing techniques: A phantom study. Med. Biol. Eng. Comput. 2024, 62, 2319–2332. [Google Scholar] [CrossRef]
  102. Fornacon-Wood, I.; Mistry, H.; Ackermann, C.J.; Blackhall, F.; McPartlin, A.; Faivre-Finn, C.; Price, G.J.; O’connor, J.P.B. Reliability and prognostic value of radiomic features are highly dependent on choice of feature extraction platform. Eur. Radiol. 2020, 30, 6241–6250. [Google Scholar] [CrossRef]
  103. Sanchez, L.E.; Rundo, L.; Gill, A.B.; Hoare, M.; Serrao, E.M.; Sala, E. Robustness of radiomic features in CT images with different slice thickness, comparing liver tumour and muscle. Sci. Rep. 2021, 11, 8262. [Google Scholar] [CrossRef]
  104. Alomar, K.; Aysel, H.I.; Cai, X. Data Augmentation in Classification and Segmentation: A Survey and New Strategies. J. Imaging 2023, 9, 46. [Google Scholar] [CrossRef]
  105. Jensen, L.J.; Kim, D.; Elgeti, T.; Steffen, I.G.; Hamm, B.; Nagel, S.N. Stability of Radiomic Features across Different Region of Interest Sizes—A CT and MR Phantom Study. Tomography 2021, 7, 238–252. [Google Scholar] [CrossRef]
  106. Zhang, J.; Zhan, C.; Zhang, C.; Song, Y.; Yan, X.; Guo, Y.; Ai, T.; Yang, G. Fully automatic classification of breast lesions on multi-parameter MRI using a radiomics model with minimal number of stable, interpretable features. Radiol. Med. 2023, 128, 160–170. [Google Scholar] [CrossRef] [PubMed]
  107. Gitto, S.; Bologna, M.; Corino, V.D.A.; Emili, I.; Albano, D.; Messina, C.; Armiraglio, E.; Parafioriti, A.; Luzzati, A.; Mainardi, L.; et al. Diffusion-weighted MRI radiomics of spine bone tumors: Feature stability and machine learning-based classification performance. Radiol. Med. 2022, 127, 518–525. [Google Scholar] [CrossRef] [PubMed]
  108. Xue, C.; Yuan, J.; Lo, G.G.; Chang, A.T.Y.; Poon, D.M.C.; Wong, O.L.; Zhou, Y.; Chu, W.C.W. Radiomics feature reliability assessed by intraclass correlation coefficient: A systematic review. Quant. Imaging Med. Surg. 2021, 11, 4431–4460. [Google Scholar] [CrossRef] [PubMed]
  109. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef]
  110. Van Timmeren, J.E.; Leijenaar, R.T.H.; van Elmpt, W.; Wang, J.; Zhang, Z.; Dekker, A.; Lambin, P. Test–Retest Data for Radiomics Feature Stability Analysis: Generalizable or Study-Specific? Tomography 2016, 2, 361–365. [Google Scholar] [CrossRef]
  111. Remeseiro, B.; Bolon-Canedo, V. A review of feature selection methods in medical applications. Comput. Biol. Med. 2019, 112, 103375. [Google Scholar] [CrossRef]
  112. Jia, W.; Sun, M.; Lian, J.; Hou, S. Feature dimensionality reduction: A review. Complex Intell. Syst. 2022, 8, 2663–2693. [Google Scholar] [CrossRef]
  113. Stańczyk, U. Feature Evaluation by Filter, Wrapper, and Embedded Approaches; Springer: Berlin/Heidelberg, Germany, 2015; pp. 29–44. [Google Scholar] [CrossRef]
  114. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  115. Yu, L.; Lu, H. Feature selection for high-dimensional data: A fast correlation-based filter solution. In Proceedings of the 20th International Conference on Machine Learning (ICML-03), Washington, DC, USA, 21–24 August 2003. [Google Scholar]
  116. Theodoridis, S.; Pikrakis, A.; Koutroumbas, K.; Cavouras, D. Introduction to Pattern Recognition; Elsevier: Amsterdam, The Netherlands, 2010. [Google Scholar]
  117. Mustafa, S. Feature selection using sequential backward method in melanoma recognition. In Proceedings of the 2017 13th International Conference on Electronics, Computer and Computation (ICECCO), Abuja, Nigeria, 28–29 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–4. [Google Scholar]
  118. Tibshirani, R. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  119. Tan, H.Q.; Cai, J.; Tay, S.H.; Sim, A.Y.; Huang, L.; Chua, M.L.; Tang, Y. Cluster-based radiomics reveal spatial heterogeneity of bevacizumab response for treatment of radiotherapy-induced cerebral necrosis. Comput. Struct. Biotechnol. J. 2024, 23, 43–51. [Google Scholar] [CrossRef]
  120. Tougui, I.; Jilbab, A.; El Mhamdi, J. Impact of the Choice of Cross-Validation Techniques on the Results of Machine Learning-Based Diagnostic Applications. Health Inform. Res. 2021, 27, 189–199. [Google Scholar] [CrossRef] [PubMed]
  121. Garau, N.; Paganelli, C.; Summers, P.; Choi, W.; Alam, S.; Lu, W.; Fanciullo, C.; Bellomi, M.; Baroni, G.; Rampinelli, C. External validation of radiomics-based predictive models in low-dose CT screening for early lung cancer diagnosis. Med. Phys. 2020, 47, 4125–4136. [Google Scholar] [CrossRef] [PubMed]
  122. Decoux, A.; Duron, L.; Habert, P.; Roblot, V.; Arsovic, E.; Chassagnon, G.; Arnoux, A.; Fournier, L. Comparative performances of machine learning algorithms in radiomics and impacting factors. Sci. Rep. 2023, 13, 14069. [Google Scholar] [CrossRef] [PubMed]
  123. Haukoos, J.S.; Lewis, R.J. Advanced Statistics: Bootstrapping Confidence Intervals for Statistics with “Difficult” Distributions. Acad. Emerg. Med. 2005, 12, 360–365. [Google Scholar] [CrossRef] [PubMed]
  124. Henderson, A.R. The bootstrap: A technique for data-driven statistics. Using computer-intensive analyses to explore experimental data. Clin. Chim. Acta 2005, 359, 1–26. [Google Scholar] [CrossRef]
  125. Huang, Y.; Li, W.; Macheret, F.; Gabriel, R.A.; Ohno-Machado, L. A tutorial on calibration measurements and calibration models for clinical prediction models. J. Am. Med. Inform. Assoc. 2020, 27, 621–633. [Google Scholar] [CrossRef]
  126. Bella, A.; Ferri, C.; Hernández-Orallo, J.; Ramírez-Quintana, M.J. Calibration of Machine Learning Models. In Handbook of Research on Machine Learning Applications and Trends; IGI Global: Hershey, PA, USA, 2010; pp. 128–146. [Google Scholar]
  127. Schwartz, L.H.; Litière, S.; de Vries, E.; Ford, R.; Gwyther, S.; Mandrekar, S.; Shankar, L.; Bogaerts, J.; Chen, A.; Dancey, J.; et al. RECIST 1.1—Update and clarification: From the RECIST committee. Eur. J. Cancer 2016, 62, 132–137. [Google Scholar] [CrossRef]
  128. Schweitzer, A.; Chiang, G.; Ivanidze, J.; Baradaran, H.; Young, R.; Zimmerman, R. Regarding “Computer-Extracted Texture Features to Distinguish Cerebral Radionecrosis from Recurrent Brain Tumors on Multiparametric MRI: A Feasibility Study”. Am. J. Neuroradiol. 2016, 38, E18–E19. [Google Scholar] [CrossRef]
  129. Tiwari, P.; Prasanna, P.; Wolansky, L.; Pinho, M.; Cohen, M.; Nayate, A.; Gupta, A.; Singh, G.; Hatanpaa, K.; Sloan, A.; et al. Computer-Extracted Texture Features to Distinguish Cerebral Radionecrosis from Recurrent Brain Tumors on Multiparametric MRI: A Feasibility Study. Am. J. Neuroradiol. 2016, 37, 2231–2236. [Google Scholar] [CrossRef]
  130. Colby, J.B. Radiomics Approach Fails to Outperform Null Classifier on Test Data. Am. J. Neuroradiol. 2017, 38, E92–E93. [Google Scholar] [CrossRef]
  131. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the areas under two or more correlated receiver operating characteristic curves: A nonparametric approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef] [PubMed]
  132. Park, S.H.; Goo, J.M.; Jo, C.-H. Receiver Operating Characteristic (ROC) Curve: Practical Review for Radiologists. Korean J. Radiol. 2004, 5, 11–18. [Google Scholar] [CrossRef] [PubMed]
  133. Wu, Y. Joint comparison of the predictive values of multiple binary diagnostic tests: An extension of McNemar’s test. J. Biopharm. Stat. 2023, 33, 31–42. [Google Scholar] [CrossRef] [PubMed]
  134. Piovani, D.; Sokou, R.; Tsantes, A.G.; Vitello, A.S.; Bonovas, S. Optimizing Clinical Decision Making with Decision Curve Analysis: Insights for Clinical Investigators. Healthcare 2023, 11, 2244. [Google Scholar] [CrossRef]
  135. Kerr, K.F.; Wang, Z.; Janes, H.; McClelland, R.L.; Psaty, B.M.; Pepe, M.S. Net Reclassification Indices for Evaluating Risk Prediction Instruments. Epidemiology 2014, 25, 114–121. [Google Scholar] [CrossRef]
  136. Armstrong, R.A. When to use the Bonferroni correction. Ophthalmic Physiol. Opt. 2014, 34, 502–508. [Google Scholar] [CrossRef]
  137. Ioannidis, J.P.A. Why Most Published Research Findings Are False. PLoS Med. 2005, 2, e124. [Google Scholar] [CrossRef]
  138. Akinci D’Antonoli, T.; Cuocolo, R.; Baessler, B.; dos Santos, D.P. Towards reproducible radiomics research: Introduction of a database for radiomics studies. Eur. Radiol. 2023, 34, 436–443. [Google Scholar] [CrossRef]
  139. Zhong, J.; Lu, J.; Zhang, G.; Mao, S.; Chen, H.; Yin, Q.; Hu, Y.; Xing, Y.; Ding, D.; Ge, X.; et al. An overview of meta-analyses on radiomics: More evidence is needed to support clinical translation. Insights Imaging 2023, 14, 111. [Google Scholar] [CrossRef]
  140. Goisauf, M.; Cano Abadía, M. Ethics of AI in Radiology: A Review of Ethical and Societal Implications. Front. Big Data 2022, 14, 850383. [Google Scholar] [CrossRef]
  141. Hillis, J.M.; Visser, J.J.; Cliff, E.R.S.; Aspers, K.v.d.G.; Bizzo, B.C.; Dreyer, K.J.; Adams-Prassl, J.; Andriole, K.P. The lucent yet opaque challenge of regulating artificial intelligence in radiology. NPJ Digit. Med. 2024, 7, 69. [Google Scholar] [CrossRef] [PubMed]
  142. Park, J.E.; Kim, D.; Kim, H.S.; Park, S.Y.; Kim, J.Y.; Cho, S.J.; Shin, J.H.; Kim, J.H. Quality of science and reporting of radiomics in oncologic studies: Room for improvement according to radiomics quality score and TRIPOD statement. Eur Radiol. 2020, 30, 523–536. [Google Scholar] [CrossRef] [PubMed]
  143. Vasey, B.; Nagendran, M.; Campbell, B.; Clifton, D.A.; Collins, G.S.; Denaxas, S.; Denniston, A.K.; Faes, L.; Geerts, B.; Ibrahim, M.; et al. Reporting guideline for the early stage clinical evaluation of decision support systems driven by artificial intelligence: DECIDE-AI. BMJ 2022, 377, e070904. [Google Scholar] [CrossRef] [PubMed]
  144. Rivera, S.C.; Liu, X.; Chan, A.-W.; Denniston, A.K.; Calvert, M.J. Guidelines for clinical trial protocols for interventions involving artificial intelligence: The SPIRIT-AI extension. Nat. Med. 2020, 26, 1351–1363. [Google Scholar] [CrossRef]
  145. Liu, X.; Rivera, S.C.; Moher, D.; Calvert, M.J.; Denniston, A.K. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: The CONSORT-AI extension. Nat. Med. 2020, 26, 1364–1374. [Google Scholar] [CrossRef]
  146. Korte, J.C.; Cardenas, C.; Hardcastle, N.; Kron, T.; Wang, J.; Bahig, H.; Elgohari, B.; Ger, R.; Court, L.; Fuller, C.D.; et al. Radiomics feature stability of open-source software evaluated on apparent diffusion coefficient maps in head and neck cancer. Sci. Rep. 2021, 11, 17633. [Google Scholar] [CrossRef]
  147. Bontempi, D.; Nuernberg, L.; Pai, S.; Krishnaswamy, D.; Thiriveedhi, V.; Hosny, A.; Mak, R.H.; Farahani, K.; Kikinis, R.; Fedorov, A.; et al. End-to-end reproducible AI pipelines in radiology using the cloud. Nat. Commun. 2024, 15, 6931. [Google Scholar] [CrossRef]
  148. Zaffino, P.; Marzullo, A.; Moccia, S.; Calimeri, F.; De Momi, E.; Bertucci, B.; Arcuri, P.P.; Spadea, M.F. An Open-Source COVID-19 CT Dataset with Automatic Lung Tissue Classification for Radiomics. Bioengineering 2021, 8, 26. [Google Scholar] [CrossRef]
  149. Prior, F.; Smith, K.; Sharma, A.; Kirby, J.; Tarbox, L.; Clark, K.; Bennett, W.; Nolan, T.; Freymann, J. The public cancer radiology imaging collections of The Cancer Imaging Archive. Sci. Data 2017, 4, 170124. [Google Scholar] [CrossRef]
  150. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  151. Woznicki, P.; Laqua, F.C.; Al-Haj, A.; Bley, T.; Baeßler, B. Addressing challenges in radiomics research: Systematic review and repository of open-access cancer imaging datasets. Insights Imaging 2023, 14, 216. [Google Scholar] [CrossRef]
Figure 1. Overview of the radiomics framework.
Figure 1. Overview of the radiomics framework.
Diagnostics 14 02473 g001
Figure 2. The main steps, the specific sub-tasks for each phase, and the related METRICS items.
Figure 2. The main steps, the specific sub-tasks for each phase, and the related METRICS items.
Diagnostics 14 02473 g002
Figure 3. Machine learning and deep learning.
Figure 3. Machine learning and deep learning.
Diagnostics 14 02473 g003
Figure 4. Bias–variance trade-off, overfitting, and underfitting. (A) Bias–variance trade-off. As the complexity of the model increases, the bias decreases but the variance increases. More complex models can capture intricate patterns in the data, better fitting the training dataset and reducing systematic errors (bias), but they also become more sensitive to noise and specific data points, leading to higher variability in model predictions (variance). (B) Overfitting and underfitting. Overfitting occurs when a model captures noise and fluctuations in the training data rather than the underlying patterns, resulting in an excellent performance on the training set (low bias) but poor generalizability to new data (high variance). Underfitting happens when a model is too simple to detect the underlying patterns, leading to a poor performance on both the training (high bias) and testing datasets (high variance).
Figure 4. Bias–variance trade-off, overfitting, and underfitting. (A) Bias–variance trade-off. As the complexity of the model increases, the bias decreases but the variance increases. More complex models can capture intricate patterns in the data, better fitting the training dataset and reducing systematic errors (bias), but they also become more sensitive to noise and specific data points, leading to higher variability in model predictions (variance). (B) Overfitting and underfitting. Overfitting occurs when a model captures noise and fluctuations in the training data rather than the underlying patterns, resulting in an excellent performance on the training set (low bias) but poor generalizability to new data (high variance). Underfitting happens when a model is too simple to detect the underlying patterns, leading to a poor performance on both the training (high bias) and testing datasets (high variance).
Diagnostics 14 02473 g004aDiagnostics 14 02473 g004b
Figure 5. Strategies for addressing the problem of model overfitting.
Figure 5. Strategies for addressing the problem of model overfitting.
Diagnostics 14 02473 g005
Figure 6. Examples of automated and semi-automated segmentation methods. (A) Fully manual segmentation. (B) Threshold segmentation. For example, a threshold of −950 HU is set to automatically segment (in red) the areas of emphysema [65]. The human annotator checks for the accuracy of the segmentation process and might adjust for inaccuracies; in this case, the process is called semi-automatic. (C) Clustering-based segmentation. Unsupervised clustering algorithm (k-means) is used to segment the signal associated with lymphedema and to differentiate the lymphedema from the signal of other types of tissue.
Figure 6. Examples of automated and semi-automated segmentation methods. (A) Fully manual segmentation. (B) Threshold segmentation. For example, a threshold of −950 HU is set to automatically segment (in red) the areas of emphysema [65]. The human annotator checks for the accuracy of the segmentation process and might adjust for inaccuracies; in this case, the process is called semi-automatic. (C) Clustering-based segmentation. Unsupervised clustering algorithm (k-means) is used to segment the signal associated with lymphedema and to differentiate the lymphedema from the signal of other types of tissue.
Diagnostics 14 02473 g006
Figure 7. Examples of data augmentation techniques.
Figure 7. Examples of data augmentation techniques.
Diagnostics 14 02473 g007
Figure 8. Overview of feature selection methods: filter, wrapper, and embedded.
Figure 8. Overview of feature selection methods: filter, wrapper, and embedded.
Diagnostics 14 02473 g008
Figure 9. Mechanism of feature selection methods, advantages, and disadvantages.
Figure 9. Mechanism of feature selection methods, advantages, and disadvantages.
Diagnostics 14 02473 g009
Figure 10. Examples of feature selection methods. (A) Filter method: correlation coefficient. (B) Wrapper method: forward subset selection. It starts with 1 predictor and gradually adds one predictor at a time until the optimal model is achieved. The estimated test error can be calculated indirectly by adjusting the training error to account for the bias due to overfitting, or directly for example through cross-validation. In this example, the minimum estimate is achieved using 4 predictors. (C) Embedded method: least absolute shrinkage and selection operator (LASSO) regression. It is a type of linear regression that incorporates regularization to enhance the model’s prediction accuracy and interpretability. It achieves this by adding a penalty term to the ordinary least squares (OLS) objective function, which is the sum of the absolute values of the model coefficients. This penalty term encourages the coefficients of less important features to shrink toward zero. As a result, lasso regression not only performs regularization but also variable selection, making the model more interpretable by reducing the number of features. Lasso regression is particularly useful when dealing with high-dimensional data, where the number of predictors can be large. It helps prevent overfitting by reducing model complexity and can improve the model’s prediction accuracy for new, unseen data. However, choosing the appropriate value for λ is crucial, typically achieved through cross-validation.
Figure 10. Examples of feature selection methods. (A) Filter method: correlation coefficient. (B) Wrapper method: forward subset selection. It starts with 1 predictor and gradually adds one predictor at a time until the optimal model is achieved. The estimated test error can be calculated indirectly by adjusting the training error to account for the bias due to overfitting, or directly for example through cross-validation. In this example, the minimum estimate is achieved using 4 predictors. (C) Embedded method: least absolute shrinkage and selection operator (LASSO) regression. It is a type of linear regression that incorporates regularization to enhance the model’s prediction accuracy and interpretability. It achieves this by adding a penalty term to the ordinary least squares (OLS) objective function, which is the sum of the absolute values of the model coefficients. This penalty term encourages the coefficients of less important features to shrink toward zero. As a result, lasso regression not only performs regularization but also variable selection, making the model more interpretable by reducing the number of features. Lasso regression is particularly useful when dealing with high-dimensional data, where the number of predictors can be large. It helps prevent overfitting by reducing model complexity and can improve the model’s prediction accuracy for new, unseen data. However, choosing the appropriate value for λ is crucial, typically achieved through cross-validation.
Diagnostics 14 02473 g010aDiagnostics 14 02473 g010b
Figure 11. (A) Supervised learning. (B) Unsupervised learning.
Figure 11. (A) Supervised learning. (B) Unsupervised learning.
Diagnostics 14 02473 g011
Figure 12. Regression, classification and clustering tasks.
Figure 12. Regression, classification and clustering tasks.
Diagnostics 14 02473 g012
Figure 13. Examples of machine learning models.
Figure 13. Examples of machine learning models.
Diagnostics 14 02473 g013
Figure 14. Criteria for model selection. The choice of the best model depends on different factors: the size of the dataset (A), the main task that should be executed by the model itself (B), the relationship between the variables (C), and the interpretability (D).
Figure 14. Criteria for model selection. The choice of the best model depends on different factors: the size of the dataset (A), the main task that should be executed by the model itself (B), the relationship between the variables (C), and the interpretability (D).
Diagnostics 14 02473 g014aDiagnostics 14 02473 g014b
Figure 15. Dataset partition for training, validation, and testing. See also paragraphs 9.4.5 and 9.5.
Figure 15. Dataset partition for training, validation, and testing. See also paragraphs 9.4.5 and 9.5.
Diagnostics 14 02473 g015
Figure 16. Validation methods. (A) Holdout validation. The training/validation set is randomly divided into two parts: a training set and a validation set or hold-out set. The model is fit on the training set, and the fitted model is used to predict the responses for the observations in the validation set. The resulting validation set error rate provides an estimate of the test error rate. The split ratio is usually 70/30 or 80/20. (B) Cross-validation. The dataset is divided into k equal-sized subsets (folds, usually k = 5 or 10). The model is trained k times, each time using k-1 folds for training and the remaining fold for validation. The model’s performance is then averaged over the k iterations to obtain a more robust and reliable estimate. Stratified k-fold cross-validation is a variant that preserves the proportion of classes in the folds, particularly useful in cases of class imbalance. (C) Leave-one-out cross-validation: an extreme form of cross-validation, where k is equal to the number of observations in the dataset. Each observation is used once as the validation set, while the remaining n-1 observations are used for training. (D) Bootstrap. From the original dataset, n observations are randomly selected with replacement to produce B bootstrap datasets (B usually from 100 to 1000). The model is trained on each bootstrap sample and validated on the data not included in that sample (out-of-bag samples). The model’s performance is averaged over the various bootstrap cycles to provide an estimate of its generalization.
Figure 16. Validation methods. (A) Holdout validation. The training/validation set is randomly divided into two parts: a training set and a validation set or hold-out set. The model is fit on the training set, and the fitted model is used to predict the responses for the observations in the validation set. The resulting validation set error rate provides an estimate of the test error rate. The split ratio is usually 70/30 or 80/20. (B) Cross-validation. The dataset is divided into k equal-sized subsets (folds, usually k = 5 or 10). The model is trained k times, each time using k-1 folds for training and the remaining fold for validation. The model’s performance is then averaged over the k iterations to obtain a more robust and reliable estimate. Stratified k-fold cross-validation is a variant that preserves the proportion of classes in the folds, particularly useful in cases of class imbalance. (C) Leave-one-out cross-validation: an extreme form of cross-validation, where k is equal to the number of observations in the dataset. Each observation is used once as the validation set, while the remaining n-1 observations are used for training. (D) Bootstrap. From the original dataset, n observations are randomly selected with replacement to produce B bootstrap datasets (B usually from 100 to 1000). The model is trained on each bootstrap sample and validated on the data not included in that sample (out-of-bag samples). The model’s performance is averaged over the various bootstrap cycles to provide an estimate of its generalization.
Diagnostics 14 02473 g016
Table 1. METRICS items and corresponding paragraphs.
Table 1. METRICS items and corresponding paragraphs.
Item/Condition Paragraph
Item 1Adherence to checklists4.1 Checklist and guidelines
Item 2Eligibility criteria4.1.2 Eligibility
Item 3High-quality reference standard4.1.3 Reference standard
Item 4Multi-centric4.1.4 Monocentric vs. multicentric
Item 5Standardized imaging protocol4.1.5 Imaging protocol
Item 6Acquisition parameters4.1.5 Imaging protocol
Item 7Time interval imaging-ref.std.4.1.3 Reference standard
Condition 1Segmentation?4.3 Segmentation
Condition 2Fully automated segmentation?4.3 Segmentation
Item 8Segmentation method4.3 Segmentation
Item 9Formal evaluation segm. meth.4.3 Segmentation
Item 10Test segmentation4.3 Segmentation
Condition 3Hand-crafted features?4.4.1 Hand-crafted features
Item 11Image preprocessing4.2 Image preprocessing
Item 12Standardized feat. extraction soft.4.4.1 Hand-crafted features
Item 13Extraction parameters4.4.1 Hand-crafted features
Condition 4Tabular data?4.5 Tabular data
Condition 5End-to-end deep learning?1. Introduction, Figure 1
Item 14Removal non-robust features4.7 Features robustness
Item 15Removal redundant features4.8 Features selection and regularization
Item 16Dimensionality compared to data size4.8 Features selection and regularization
Item 17Robustness E2E DL pipeline4.7 Features robustness
Item 18Data partitioning (train./val./test.)4.9 Data partition for training, validation, and test
Item 19Confounding factors4.6 Confounding factors
Item 20Appropriate performance metrics4.10 Model testing and performance metrics
Item 21Uncertainty assessment4.11 Model uncertainty assessment and calibration
Item 22Calibration4.11 Model uncertainty assessment and calibration
Item 23Uni-parametric or proof of added value4.12 Model comparison
Item 24Comparison with non-radiomics4.12 Model comparison
Item 25Comparison with classic stat. model4.12 Model comparison
Item 26Internal testing4.10 Model testing and performance metrics
Item 27External testing4.10 Model testing and performance metrics
Item 28Data availability4.13 Challenges and future perspectives
Item 29Code availability4.13 Challenges and future perspectives
Item 30Model availability4.13 Challenges and future perspectives
Table 2. Hand-crafted features (adapted from [19]).
Table 2. Hand-crafted features (adapted from [19]).
Feature CategoriesExample Radiomic FeaturesDescription
First-orderMean, median
Max/mean/min intensity
10–90th percentile
Skewness, Kurtosis
Range, Variance
Root Mean Squared (RMS)
Standard Deviation (SD)
Mean Absolut Deviation (MAD)

Area
Volume
Maximum 3D diameter
Major axis length
Minor axis length
Surface area
Elongation
Flatness
Sphericity
First-order features include basic statistics on the distribution of the values of individual voxels, disregarding spatial relationship, or shape-based features.
Second-orderGray level co-occurrence matrix
Gray-level run length matrix
Gray-level size zone matrix
Neighboring Gray Tone Difference Matrix
Gray Levele Dependence Matrix
Second-order features describe the statistical relationships between pixels or voxels.
High-orderAutoregressive model
Haar wavelet
High-order features are usually based on matrices that consider relationships between three or more pixels or voxels.
Table 3. Intraclass Correlation Coefficient (ICC) and the Concordance Correlation Coefficient (CCC).
Table 3. Intraclass Correlation Coefficient (ICC) and the Concordance Correlation Coefficient (CCC).
Subset Search ProcessWhat It RepresentsWhen It Can Be Used
Intraclass Correlation Coefficient (ICC)The ICC is useful for evaluating the reproducibility or reliability of measurements between different repetitions or between different assessments made by different observers.
A high ICC indicates that most variability is due to genuine differences between subjects, suggesting feature robustness.
If a test–retest is performed (the same image from the same patient and scanner obtained a few minutes apart), then the ICC can be calculated.
Concordance Correlation Coefficient (CCC)It combines measures of precision and accuracy to assess how well bivariate pairs of observations conform relative to a gold standard or another set.
It is valuable for comparing the agreement between features extracted with different imaging techniques or acquisition parameters.
If multiple phantom images for the same and different scanners can be acquired, then the CCC for each feature can be calculated.
Table 4. Statistical-based filtering methods.
Table 4. Statistical-based filtering methods.
Statistical MethodExclusion Criteria
Missing PercentageDisproportionate share of missing samples and difficult to fill
VarianceVariance close to or equal to 0
FrequencyFeatures excessively concentrated in one category of values
Correlation Coefficients (Spearman, Pearson, and Kendall)Correlation coefficients close to or equal to 0
Analysis of Variance (ANOVA)Too-low F-value or excluded features with a p-value < 0.05
χ2 TestToo-low χ2 value or p-value < 0.05
Mutual InformationMutual information close to or equal to 0
mRMR (Minimum Redundancy Maximum Relevance)Features with the minimum correlation and maximum redundancy
Fisher ScoreLarge intraclass distances and small interclass distances
Table 5. Wrapper methods.
Table 5. Wrapper methods.
Subset Search ProcessSubset Search MethodCriteria
Complete searchBreadth First Search
Best First Search
Iterate through all possible combinations of feature subsets, then select the feature subset with the best model score. High computational cost
Heuristic searchSequential Forward Selection
Sequential Backward Selection
Bidirectional Search
Plus-L Minus-R Selection
Sequential Floating Selection
Decision Tree Method
Uses rules or guided search strategies to find a good subset of features, without necessarily guaranteeing the optimal solution
Random searchRandom Generation plus Sequential Selection
Simulated Annealing
Genetic
A random subset of features is generated and then these feature subsets are evaluated; does not guarantee optimality or computational efficiency
Table 6. Comparison of different validation methods.
Table 6. Comparison of different validation methods.
Validation MethodsAdvantagesDisadvantages
Holdout dataset
  • Computationally feasible with respect to LOOCV
  • Lower variance with respect to LOOCV
  • Higher bias with respect to LOOCV
Cross-validation (CV)
  • Computationally feasible with respect to LOOCV
  • Lower variance with respect to LOOCV
  • Higher bias with respect to LOOCV
Leave-one-out CV (LOOCV)
  • Useful for a small dataset
  • Lower bias compared to CV
  • Computationally intensive
  • Lower variance compared to CV
Bootstrapping
  • Can be applied in almost all situations
  • Makes the most of available data without the need for a separate validation set
  • Can also be used to quantify the uncertainty associated with a given estimator or statistical learning method (i.e., to calculate the confidence intervals)
  • Can lead to an overestimation of the model’s performance
Table 7. Main performance metrics and explanations.
Table 7. Main performance metrics and explanations.
Performance MetricsMeaningDescription
Regression taskResidual standard error (RSE)It represents the standard deviation of the residuals, which are the differences between the observed values and the values predicted by the model. It provides an estimate of the typical size of the prediction errors.It is calculated as the square root of the sum of squared residuals divided by the degrees of freedom (n − p − 1), where n is the number of observations and p is the number of predictors.
R2 statisticIt represents the proportion of the variance in the dependent variable that is predictable from the independent variables. R2 values range from 0 to 1, with 0 indicating that the model does not explain any of the variability in the response data around its mean and 1 indicating that the model explains all the variability.It is calculated as 1 minus the ratio of the sum of squared residuals to the total sum of squares.
F-statisticIt compares a model with no predictors (intercept only) to the model being evaluated. A higher F-statistic indicates that the model provides a better fit to the data than a model without any predictors.It is calculated as the ratio of the mean regression sum of squares to the mean error sum of squares, and its significance is evaluated using the F-distribution.
Classification taskSensitivitySensitivity, or true-positive rate, is the ability of a test to correctly identify positive cases.It is calculated as the ratio of true positives to the sum of true positives and false negatives (TP/(TP + FN))
SpecificitySpecificity, or true-negative rate, is the ability of a test to correctly identify negative cases.It is calculated as the ratio of true negatives to the sum of true negatives and false positives (TN/(TN + FP)).
AccuracyAccuracy is the overall correctness of a test.It is calculated as the ratio of the number of correct predictions (true positives and true negatives) to the total number of cases examined ((TP + TN)/(TP + TN + FP + FN)).
PrecisionPrecision, or positive predictive value, indicates the proportion of correct positive results among all those identified as positive.It is calculated as the ratio of true positives to the sum of true positives and false positives (TP/(TP + FP)).
RecallRecall is synonymous with sensitivity.
F1 scoreThe F1 score is the harmonic mean of precision and recall, providing a single measure that balances both aspects.It is calculated as 2 × (Precision × Recall)/(Precision + Recall).
Receiver operating characteristic (ROC) curves
and area under the curve (AUC)
Depict the trade-off between the true-positive rate (sensitivity) and the false-positive rate (1-specificity) across various classification thresholds. A higher area under the ROC curve signifies a better discrimination capability, with an AUC of 1 indicating a perfect classifier.
Confusion matricesThey present the counts of true-positive, true-negative, false-positive, and false-negative predictions, providing insights into the model’s ability to correctly classify instances from each class. From the confusion matrix, derived metrics, such as accuracy, precision, recall (sensitivity), specificity, and F1-score, can be calculated.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cè, M.; Chiriac, M.D.; Cozzi, A.; Macrì, L.; Rabaiotti, F.L.; Irmici, G.; Fazzini, D.; Carrafiello, G.; Cellina, M. Decoding Radiomics: A Step-by-Step Guide to Machine Learning Workflow in Hand-Crafted and Deep Learning Radiomics Studies. Diagnostics 2024, 14, 2473. https://doi.org/10.3390/diagnostics14222473

AMA Style

Cè M, Chiriac MD, Cozzi A, Macrì L, Rabaiotti FL, Irmici G, Fazzini D, Carrafiello G, Cellina M. Decoding Radiomics: A Step-by-Step Guide to Machine Learning Workflow in Hand-Crafted and Deep Learning Radiomics Studies. Diagnostics. 2024; 14(22):2473. https://doi.org/10.3390/diagnostics14222473

Chicago/Turabian Style

Cè, Maurizio, Marius Dumitru Chiriac, Andrea Cozzi, Laura Macrì, Francesca Lucrezia Rabaiotti, Giovanni Irmici, Deborah Fazzini, Gianpaolo Carrafiello, and Michaela Cellina. 2024. "Decoding Radiomics: A Step-by-Step Guide to Machine Learning Workflow in Hand-Crafted and Deep Learning Radiomics Studies" Diagnostics 14, no. 22: 2473. https://doi.org/10.3390/diagnostics14222473

APA Style

Cè, M., Chiriac, M. D., Cozzi, A., Macrì, L., Rabaiotti, F. L., Irmici, G., Fazzini, D., Carrafiello, G., & Cellina, M. (2024). Decoding Radiomics: A Step-by-Step Guide to Machine Learning Workflow in Hand-Crafted and Deep Learning Radiomics Studies. Diagnostics, 14(22), 2473. https://doi.org/10.3390/diagnostics14222473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop