Next Article in Journal
The Efficacy of MRI-Based ADC Measurements in Detecting Axillary Lymph Node Metastasis: Evaluation of a Prospective Study
Previous Article in Journal
Cutting Through History: The Evolution of Glioblastoma Surgery
Previous Article in Special Issue
Artificial Intelligence in Head and Neck Cancer: Innovations, Applications, and Future Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Breast Cancer Relapse from Histopathological Images with Ensemble Machine Learning Models

by
Ghanashyam Sahoo
1,
Ajit Kumar Nayak
2,
Pradyumna Kumar Tripathy
3,
Amrutanshu Panigrahi
1,
Abhilash Pati
1,*,
Bibhuprasad Sahu
4,
Chandrakanta Mahanty
5,* and
Saurav Mallik
6,7
1
Department of Computer Science and Engineering, Siksha ‘O’ Anusandhan (Deemed to Be University), Bhubaneswar 751030, India
2
Department of Computer Science and Information Technology, Siksha ‘O’ Anusandhan (Deemed to Be University), Bhubaneswar 751030, India
3
Department of Computer Science and Engineering, Silicon University, Bhubaneswar 751024, India
4
Department of Information Technology, Vardhaman College of Engineering (Autonomous), Hyderabad 501218, India
5
Department of Computer Science and Engineering, GITAM Deemed to Be University, Visakhapatnam 530045, India
6
Department of Environmental Health, Harvard T.H. Chan School of Public Health, Boston, MA 02115, USA
7
Department of Pharmacology & Toxicology, University of Arizona, Tucson, AZ 85721, USA
*
Authors to whom correspondence should be addressed.
Curr. Oncol. 2024, 31(11), 6577-6597; https://doi.org/10.3390/curroncol31110486
Submission received: 10 September 2024 / Revised: 21 October 2024 / Accepted: 22 October 2024 / Published: 24 October 2024

Abstract

:
Relapse and metastasis occur in 30–40% of breast cancer patients, even after targeted treatments like trastuzumab for HER2-positive breast cancer. Accurate individual prognosis is essential for determining appropriate adjuvant treatment and early intervention. This study aims to enhance relapse and metastasis prediction using an innovative framework with machine learning (ML) and ensemble learning (EL) techniques. The developed framework is analyzed using The Cancer Genome Atlas (TCGA) data, which has 123 HER2-positive breast cancer patients. Our two-stage experimental approach first applied six basic ML models (support vector machine, logistic regression, decision tree, random forest, adaptive boosting, and extreme gradient boosting) and then ensembled these models using weighted averaging, soft voting, and hard voting techniques. The weighted averaging ensemble approach achieved enhanced performances of 88.46% accuracy, 89.74% precision, 94.59% sensitivity, 73.33% specificity, 92.11% F-Value, 71.07% Mathew’s correlation coefficient, and an AUC of 0.903. This framework enables the accurate prediction of relapse and metastasis in HER2-positive breast cancer patients using H&E images and clinical data, thereby assisting in better treatment decision-making.

1. Introduction

Globally, breast cancer is still among the most common malignancies affecting women. Among them, HER2-positive breast cancer stands out as a subtype defined by the overexpression of the HER2 protein. When cancer cells overexpress the HER2 protein, it can lead to HER2-positive breast cancer, a subtype of the disease that can grow more aggressively in tumors. Using histological photographs stained with H&E, one may physically inspect the tissue samples and aid in identifying and classifying the cancer cells [1,2]. Information like this might be useful for researchers in various ways, such as monitoring the progression of an illness, discovering novel therapeutic approaches, and evaluating the effectiveness of current therapies. Even with advancements in treatment methods, a major challenge in clinical practice is predicting the likelihood of sickness return [3,4]. Pathologists may learn a lot about the cellular architecture and structure of breast tumors from images captured after staining tissue samples with hematoxylin and eosin (H&E). Researchers have sought to develop deep learning (DL) and machine learning (ML) models for predicting breast cancer relapse using the mountain of data available in databases such as The Cancer Genome Atlas (TCGA). The most recent worldwide statistics from 2020 show that, surpassing even lung cancer, breast cancer has the highest incidence rate. Although targeted therapy has greatly enhanced the survival rates of breast cancer patients, around 30–40% of patients still face the challenge of relapse, and 10–15% ultimately succumb to cancer metastasis or relapse [5]. Timely identification of the reappearance and spread of cancer increases the effectiveness of treatment and substantially reduces the death rate. Due to the fast progress of AI and the increasing volume of multi-modal medical data, there is now a chance to detect and predict the relapse and spread of cancer at an early stage [6]. The prediction models should be further optimized to improve the discrimination ability and benefit, and prospective external validation studies should also be carried out before they are applied to clinical practice. Researchers conducted a comprehensive study using a network pharmacology-based strategy to predict therapy targets of Ecliptae Herba in treating breast cancer [7,8].

1.1. Existing Works

Sakri et al. [9] used the Breast Cancer Wisconsin Prognostic Breast Cancer (WPBC) Datasets for breast cancer relapse prediction. They employed ML techniques such as the Reduced Error Pruning Tree (REPTree), Naïve Bayes (NB), and K-Nearest Neighbour (KNN)-IBK in conjunction with particle swarm optimization (PSO). The results showed an accuracy rate of 81.3%, a precision of 88.3%, a recall of 93.4%, an F-score of 87.7%, and an area under the curve (AUC) of 0.820. Alom et al. [10] used the Breast Cancer Dataset to achieve a classification rate of 100% accuracy, 100% sensitivity, 100% specificity, and 1.00 AUC in breast cancer classification from histopathological images using an inception recurrent residual convolutional neural network (IRRCNN) that took into account DL techniques such as the Inception Network (Inception-v4), Residual Network (ResNet), and Recurrent Convolutional Neural Network (RCNN). Hong et al. [11] utilized a panel of eight microRNAs to predict the return of triple-negative breast cancer. They used ML techniques, specifically Logistic Regression (LR) and Gaussian mixture analyses, on the TCGA_TNBC, GEOD-40525, GSE40049, and GSE19783 datasets. The results showed an AUC of 0.89 for GSE40049 and 0.90 for GSE19783. Yan et al. [12] developed a prediction model in a primary cohort consisting of 498 patients with invasive breast cancer, and the data were gathered from July 2016 to September 2018 to predict pathological complete response and tumor shrinkage size in breast cancer considering an ML approach, i.e., LR, on Breast Cancer Cinoma (BCC) using the Harbin Medical University Cancer Hospital (HMUCH) dataset and resulted in a nomogram for HER2-positive and triple-negative breast cancer (TNBC) with an AUC of 0.820 and 0.785, respectively. Mosayebi et al. [13] proposed a new model for the prediction of relapse of breast cancer considering ML approaches, i.e., RF, Learning Vector Quantization (LVQ), NB, C5.0 DT, Multilayer Perceptron (MLP), Kernel Principal Component Analysis-SVM (KPCA-SVM), and SVM using Ministry of Health and Medical Education and the Iran Cancer Research Centre (MHME-ICRC) datasets. They resulted in 81.9% accuracy, 86.9% sensitivity, 77.7% specificity, 82.2% for the geometric mean of sensitivity and specificity, 76.9% PPV, 87.4% NPV, 82.0% for the geometric mean of PPV and NPV, 81.6% for the F-Value, and 77.4% AUC. Comes et al. [14] utilized DL approaches, specifically a Convolutional Neural Network (CNN) and Support Vector Machine (SVM), on DCE-MRI exams from the public datasets I-SPY1 TRIAL and BREAST-MRI-NACT-Pilot. The results showed an accuracy rate of 85.2%, a sensitivity rate of 84.6%, and an AUC of 0.83. Sanyal et al. [15] developed a weakly supervised temporal model that could predict distant breast cancer relapses using DL techniques (e.g., LSTM and XGBoost) on datasets curated manually and by natural language processing experts. The model achieved an AUC of 0.94, a sensitivity of 89.0%, and a specificity of 84.0%. The HEROHE challenge, organized by Conde-Sousa et al. [16] as a side event of the 16th European Congress on Digital Pathology, attempted to determine the HER2 status in breast cancer using solely H&E-stained tissue samples. The results showed 79% precision, 100% recall, a 79% F-Value, and 0.88 AUC. Multimodal prediction of five-year breast cancer relapse in women who receive neoadjuvant chemotherapy was proposed by Rabinovici-Cohen et al. [17]. Using a real-world retrospective dataset, the authors achieved 57% specificity, a 56% F-Value, 72% balanced accuracy, 41% PPV, 96% NPV, 93% sensitivity, and 0.75 AUC. The study also considered the DL approach known as CNN. Liu et al. [18] explored the use of color and texture features in histopathological images to predict the likelihood of breast cancer recurrence and metastasis. They used ML techniques such as RF, LR, and XGBoost on both a clinical dataset and formalin-fixed, paraffin-embedded (FFPE) samples from the TCGA. The results showed that the clinical dataset achieved 80% accuracy, 40% precision, 75% recall, a 50% F-Value, and 0.75 AUC. In contrast, the TCGA dataset achieved 73% accuracy, 33% precision, 60% recall, 42% F-Value, and 0.72 AUC. Yang et al. [19] suggested using DL to predict the risk of metastasis and relapse for HER2-positive breast cancer using CNN and ResNet50 on the datasets of the Cancer of the Chinese Academy of Medical Sciences (CAMS) and TCGA, respectively, using histopathological images and clinical data, with AUCs of 0.76 and 0.72. Lu et al. [20] presented a new model called Slide-Graph + that uses graph neural networks (GNNs) to forecast HER2 status using whole-slide images of routinely stained slides. The model was tested on the TCGA and two independent test datasets, HER2C and Nott-HER2, and achieved AUC values of 0.75 and 0.80, respectively. Su et al. [21] created a DL framework called the Breast Cancer Relapse Network (BCR-Net). The goal was to use the CNN DL approach on adjacent pairs of H&E and Ki67 breast cancer resection tissues for 50 patients. The results showed an impressive 80% accuracy, 79.2% F-Value, and 0.811 AUC. Liu et al. [22] developed and validated ML models to predict changes in visual acuity and keratometry two years after corneal crosslinking (CXL) for progressive keratoconus. Using patient demographics, visual acuity, spherical equivalence, and Pentacam parameters from 277 eyes in the Collected dataset from Aier Eye Hospital of Wuhan University (AEHWU, Wuhan, Hubei province, China), the study found that the XGBoost model produced the most accurate predictions for corrected distance visual acuity (CDVA) and maximum keratometry (Kmax) changes, with R-squared values of 0.9993 and 0.9888 for the testing set and 0.8956 and 0.8382 for the validation set. This approach significantly improved prediction accuracy by incorporating key baseline metrics such as CDVA and keratometry ratios. Botlagunta et al. [23] used ML techniques such as LR, KNN, DT, RF, SVM, GB, and XGBoost to classify and diagnose breast cancer metastasis on clinical data. They achieved 83% accuracy, 83% precision, 100% recall, an 85% F-Value, and 0.87 AUC on medical Metastatic Breast Cancer (MBC) data. Dammu et al. [24] utilized DL techniques, specifically CNN, to predict pathological complete response, residual cancer burden, and progression-free survival in breast cancer patients. The model achieved 81% accuracy, 68% sensitivity, 97% specificity, a 76% F-Value, and 0.83 AUC on the I-SPY-1 TRIAL dataset. Table 1 provides a summary of the examined state-of-the-art works.

1.2. Research Gap and Motivation

Relapse prediction using histopathological images and ML/ensemble learning (EL) is very important in the setting of HER2-positive breast cancer diagnosis. By applying ML algorithms to these pictures, important features may be extracted, enabling the detection of subtle patterns and traits linked to the development and return of diseases. By analyzing many image attributes, ML algorithms can shed light on the biological mechanisms that cause cancer to relapse [25,26,27]. EL offers a solid structure for combining many ML models to improve the precision of forecasts. When trying to predict when breast cancer may recur, EL approaches can be employed to merge many models that were trained on different datasets or using various algorithms. The predictive model’s capacity to generalize is improved, and the risk of overfitting is decreased by employing several modeling approaches. Using EL approaches, which combine the insights of many models to provide more accurate predictions, can enhance clinical decision-making in treating HER2-positive breast cancer.
This study aims to address the challenge of predicting breast cancer relapse in HER2-positive patients by employing multimodal analysis of histopathology images. Histological examinations are the gold standard for detecting breast cancer. The conventional, labor-intensive, and error-prone histopathology method relies heavily on histopathologists’ subjective expertise. Building a reliable computer-assisted system capable of autonomously analyzing histological images for cancer diagnosis and prognosis is of the utmost importance. According to the researchers, we may be able to glean insights from images of tissues that reveal the disease’s progression or return by merging deep neural networks with ML algorithms. We aim to develop prediction models that can accurately identify patients with a higher likelihood of experiencing a relapse by analyzing the TCGA database with HER2-positive diagnosis. By bringing together researchers from different fields, we hope to increase the accuracy of breast cancer management predictions and pave the way for more individualized approaches to therapy.

1.3. Objective and Contribution

It is critical to use basic numerical characteristics that accurately characterize histological pictures without changing their visual appearance, which might lead pathologists to make incorrect diagnoses to save computing costs. We overcame these challenges by detecting the color features of histopathological pictures using a Wavelet Multi-sub-band Co-occurrence Matrix and extracting texture information from low-order color moments. This research aims to develop a model for the early detection of breast cancer relapse using ML and DL techniques. The study aims to accurately characterize histopathological images by extracting color and texture features without altering their visual appearance, thereby minimizing the risk of misdiagnosis and reducing computational costs. Two stages of experiments are carried out in this study. In stage-1, six basic ML approaches, including SVM, LR, decision tree (DT), random forest (RF), adaptive boosting (AdaBoost), and extreme gradient boosting (XGBoost), are employed, whereas, in stage-2, these six ML approaches are ensembled with weighted averaging, soft voting, and hard voting to achieve the targeted improved predictive outcomes in terms of accuracy, precision, etc. These are our contributions to this study:
  • The introduction of an ensemble ML approach for enhanced breast cancer relapse prediction of histopathological images.
  • The creation of an ensemble model that reduces training time and optimizes performance.
  • The consideration of various approaches for better image pre-processing and extracting more relevant features for prediction.
  • An improvement in patient outcomes by assisting doctors in making better treatment decisions.

1.4. Paper Structure

The following is the structured analysis of this proposed study: Section 2 details the methods used in this study, including the suggested dataset. The design, flowchart, block diagrams, and working principle of the proposed work are covered in Section 3, which addresses the architectural aspect of the research. In contrast to the relevant outcomes examined in this study, Section 4 details the fruitful examination of the proposed work. Section 5 concludes the analysis by outlining a practical expansion of the suggested effort.

2. Materials and Methods

The research methodology and materials employed in this study are thoroughly elucidated in this section. The dataset, pre-processing procedures, feature extraction employing diverse methodologies, and the different ML and EL algorithms employed may be located in the respective subsections.

2.1. Dataset Description

A total of 123 individuals with HER2-positive breast cancer diagnoses were included in this research [28]. The Cancer Genome Atlas (TCGA) database was used to obtain the histology images stained with hematoxylin and eosin (H&E). It was shown that 26% of the people whose lymph nodes tested positive for breast cancer were younger than 50 years old. There is a total of 88 H&E-stained whole slide images (WSIs) preserved in the formalin-fixed paraffin-embedded (FFPE) format and 53 flash-frozen WSIs of breast cancer. There were five cases of relapse or metastasis among these WSIs. Patients with relapse or metastasis are represented by positive samples (labeled 1), whereas patients who did not have these problems are represented by negative samples (labeled 0). This makes it easier to explain and calculate what follows. To address the significant imbalance between relapse (positive) and non-relapse (negative) samples, we applied the Synthetic Minority Over-sampling Technique (SMOTE) to the training set. SMOTE generates synthetic samples for the minority class to balance the dataset, which helps in improving the model’s specificity and overall predictive performance. Table 2 provides a summary of the TCGA dataset that was evaluated.

2.2. Methodology

The current work utilizes the Contrast Limited Adaptive Histogram Equalization (CLAHE) method to enhance picture contrast. Subsequently, the Two-Dimensional Wavelet Packet Transform (2DWPT) integrated Multi-sub-bands Co-occurrence Matrix (WMCM) is used to extract texture characteristics from the images. Next, the collected features are evaluated using six distinct ML classifiers: SVM, LR, RF, DT, AdaBoost, and XGBoost. These classifiers are used as the base learners to make the first predictions. Three ensemble strategies, weighted averaging, hard voting, and soft voting are used to make the final prediction based on the initial predictions obtained using these ML classifiers.

2.2.1. Preprocessing: CLAHE

Splitting an image into tiny, non-overlapping tiles is what contrast-limited adaptive histogram equalization (CLAHE) does to boost the local contrast [29,30]. CLAHE is used to enhance the contrast of histopathological images. This technique is chosen because it improves the local contrast and enhances the visibility of features without amplifying noise, making it suitable for medical images where subtle differences are crucial. Each tile’s pixel intensity values are used to create a histogram. In order to keep noise levels from getting too high, the cumulative distribution function of the histogram for each tile is clipped at a certain limit (CP Limit). Using clipping ensures that the brightness of each pixel stays below a certain limit. After that, we use the clipped CDF to build a lookup table that converts the raw intensity values to their histogram-equalized equivalents. Finally, the image is transformed pixel-by-pixel using the lookup table, which improves the local contrast without changing the image’s overall quality. One may control the amount of augmentation and the dimensions of the local regions by adjusting parameters like the tile size and clipping limit. Because of this quality, CLAHE is very effective at improving the aesthetic appeal of photographs with dynamic contrast and lighting. The detailed working of CLAHE is as follows:
  • Divide the original image into small, non-overlapping tiles.
  • For each tile, compute the cumulative distribution function (C) as in Equation (1):
    C i = k = 0 i H ( k ) M × N
    where C i is the cumulative distribution function at intensity level i, H ( k ) is the histogram value at intensity level k, and M × N is the dimension of the image. H k can be defined as Equation (2).
    H k = m = 1 M n = 1 N 1 ,   i f   t h e   p i x e l   i n t e n s i t y   a t   p o s i t i o n   m , n = k 0 ,   O t h e r w i s e
  • Clip C to limit the contrast level for the current intensity level k as in Equation (3).
    C c l i p p e d k = C l i p ( C i , 0 , C l i p m a x )
    where C l i p ( C i ,   0 ,   C l i p m a x ) is the method used to limit the cumulative distribution function ( C i ) in the range of [0, C l i p m a x ]. C l i p m a x is the controlling parameter for controlling the maximum contrast enhancement.
  • Calculate the Histogram equalization Lookup Table (ELT) for the calculated C c l i p p e d k by rounding the value to the nearest integer as in Equation (4)
    E L T k = R o u n d ( C c l i p p e d k × 255 )
  • Apply E L T to each individual pixel to find the enhanced pixel value.
CLAHE is a potent technique that effectively enhances the contrast of an image while preserving its original local characteristics. One may control the degree of enlargement and the size of the specific regions being analyzed by modifying parameters such as the tile size and clipping limit.

2.2.2. Feature Extraction: 2D-WPT

The current work employs the 2D wavelet packet transformation (2D-WPT)-motivated Wavelet Multi-sub-bands Co-occurrence Matrix (WMCM) to extract the texture-based features from the images [21,31]. 2DWPT is used to decompose images into multiple sub-bands, allowing for capturing both spatial and frequency information. This method is selected because it effectively captures texture variations at different scales, which is important for histopathological images. WMCM is employed to extract texture features from the decomposed sub-bands. This approach is chosen because it captures the spatial relationships of pixel intensities, providing rich texture information that aids in accurate classification. The feature extraction procedure using the WMCM has many successive processes. The input picture is first wavelet-transformed into many sub-bands. The several sub-bands, each corresponding to a distinct frequency component, represent details in the picture. After that, the Co-occurrence Matrix (CM) is calculated for every sub-band, which records the spatial correlations between pixel intensities across different dimensions and orientations. Within each sub-band, the CM measures the frequency of various pixel value combinations. The WMCM unifies texture information across several frequency domains by combining the CMs from all sub-bands. It is possible to extract a collection of texture characteristics from the WMCM. Some of these properties may be statistical measurements that tell us much about the image’s texture, such as energy, correlation, contrast, and entropy. Feature extraction is a strong tool for image analysis and pattern recognition tasks. It uses the multi-scale and multi-directional information stored in the WMCM to capture texture features. The working of the WMCM model is shown below.
  • Step-1: Creating the WMCM
Initially, the 2D WPT model is executed to divide the input image into multiple 256 × 256 pixels sub-bands. The sub-bands are High–High (HH), High–Low (HL), Low–High (LH), and Low–Low (LL) bands. HH illustrates the sub-band with high frequency in both horizontal and vertical directions, HL shows the high-frequency information in the horizontal direction and low-frequency information in the vertical direction, LH shows the noise, HL represents the horizontal detailed sub-images, and LH represents the low- and high-frequency information in the horizontal and vertical directions, respectively. Similarly, LL represents the low-frequency information in both directions. These four bands can be found using Equations (5)–(8).
H H = ( P φ d ( x , y ) )
H L = ( P φ v ( x , y ) )
L H = ( P φ h ( x , y ) )
L L = ( P ( x , y ) )
where ∗ is the convolution operation, is the scaling function, φ is the wavelet function, x is the scaling parameter, and y is the translation parameter. φ v , φ h , and φ d are wavelet functions in the vertical (v), horizontal (h), and diagonal (d) directions, respectively.
  • Calculate the Gray-Level Co-Occurrence Matrix (GLCM) for each pixel (i,j) sub-bands (SBs) for HH, HL, LH, LL using Equation (9) with H a , b as the intensity value at any position (a, b). M and N are the dimensions of the image.
    G L C M S B i , j = m = 1 M n = 1 N 1 ,   i f   H a , b = i   , H a , b = j 0 ,   O t h e r w i s e
    a = a
    b = b
    where the is the distance from the initial position (a, b), and SB is the sub-band, which can be represented as S B { L L , L H , H L , H H } .
  • Combine the GLCM values to form the WMCM using Equation (12) with K as the total number of sub-bands available for the corresponding image.
    W M C M ( i , j ) = k = 1 K G L C M S B k i , j
  • Step-2: WMCM-based Texture Feature Extraction Process
WMCM-based texture features can be calculated using eleven different parameters including Small Gray Level with Small and Big Detail advantage denoted as SGSDA and SGBDA, respectively, Gray-Level Average (GLA), Detail-Level Average (DLA), Gray-Level and Detail-level Mean Square Error denoted as GLMSE and DLMSE, respectively, Correlation (C), Energy (E), Entropy (En), Contrast (Co), and Homogeneity. These parameters can be calculated using Equations (13)–(25) with (i,j) as the intensity level, H(i,j) as the value at intensity (i,j), ϑ i   a n d   ϑ ( j ) as the marginal distribution along rows and columns, respectively, and L as the total number of intensity levels, which can be 8, 16, or 32. In the current study, the total intensity level is 16.
S G S D A = i = 1 L j = 1 L H i , j , ( i 2 . j 2 )
S G B D A = i = 1 L j = 1 L H i , j . ( j 2 i 2 )
G L A = i = 1 L i . ϑ ( i )
D L A = j = 1 L i . ϑ ( j )
ϑ i = j = 1 L G L C M ( i , j ) i = 1 L j = 1 L G L C M ( i , j )
ϑ j = i = 1 L G L C M ( i , j ) i = 1 L j = 1 L G L C M ( i , j )
G L M S E = j = 1 L ( i G L A ) 2 . ϑ i i = 1 L ϑ i
D L M S E = j = 1 L ( i D L A ) 2 . ϑ j i = 1 L ϑ j
C = i , j = 1 L i . j . W M C M i , j μ a μ b σ a . σ b
C o = i , j = 1 L i j 2 . W M C M ( i , j )
E = i , j = 1 L ( W M C M i , j ) 2
E n = i , j = 1 L W M C M i , j log W M C M i , j  
H = i , j = 1 L 1 1 + ( i j ) W M C M i , j
After calculating the parameters above, the corresponding intensity level’s texture feature (TF) can be calculated by normalizing the parameters. The TF for each intensity can be calculated using Equation (26).
T F i , j = ω 1 . S G S D A + ω 2 . S G B D A + ω 3 . G L A + ω 4 . D L A + ω 5 . G L M S E + ω 6 . D L M S E + ω 7 . C + ω 8 . C o + ω 9 . E + ω 10 . E n + ω 11 . H
ω 1 + ω 2 + ω 3 + ω 4 + ω 5 + ω 6 + ω 7 + ω 8 + ω 9 + ω 10 + ω 11 = 1
Based on the TF for each intensity (i,j), the final texture feature for the entire image sized M × N can be quantified as in Equation (28):
T F = T F i , j T F i , M T F i , N T F N , M

2.2.3. Employed ML Approaches

Supervised learning methods such as SVM frequently identify the optimal hyperplane to separate classes in high-dimensional space in classification challenges. A logistic function is a binary dependent variable in LR, a statistical model used to estimate the probability of a particular event occurring. DTs are representations of the chronological and logical progression of healthcare issues. RF constructs and amalgamates several DTs to enhance classification accuracy and mitigate overfitting. XGBoost is a scalable and efficient implementation of a gradient-boosting framework designed to optimize computation time and enhance predictive accuracy. AdaBoost focuses on misclassified samples to combine the outputs of weak classifiers, generating a strong classifier.
Six ML classifiers are chosen for their diverse strengths in handling different types of data and their proven effectiveness in medical image analysis: SVM, effective in high-dimensional spaces and robust against overfitting, particularly with clear margins of separation; LR, a simple yet powerful classifier for binary outcomes, known for its interpretability and effectiveness when the relationship between features and the target variable is approximately linear; RF, an ensemble method that builds multiple DTs to achieve more accurate and stable predictions, chosen for its robustness and high accuracy; DT, a non-parametric model that is easy to interpret and understand, selected for its simplicity and effectiveness in capturing non-linear relationships; AdaBoost, which combines weak classifiers to create a strong classifier, improving model accuracy by focusing on difficult-to-classify samples; and XGBoost, a powerful gradient-boosting algorithm that is efficient, scalable, and capable of capturing complex patterns. Combining these techniques ensures a robust methodology that enhances image quality, extracts meaningful features, and employs diverse classifiers to handle various aspects of the data.

2.2.4. Employed EL Approaches

EL integrates the outputs of several separate models to forge a more robust forecast. The ensemble strategies are used to improve the overall predictive performance by combining the strengths of individual classifiers, thereby reducing the risk of overfitting and improving generalization to new data. Three ensemble strategies are employed to enhance predictive performance: weighted averaging, soft voting, and hard voting. Weighted averaging combines the predictions of multiple classifiers by assigning different weights to each classifier based on their performance, leveraging their strengths to produce more accurate predictions. Soft voting averages the predicted probabilities of each classifier to make the final prediction, effectively combining probabilistic outputs for better performance. Hard voting aggregates the majority class predicted by each classifier to make the final decision, chosen for its simplicity and effectiveness in scenarios where the majority rule is reliable [32,33]. The goal of weighted averaging is to integrate the predictions of several models into a single score, with the weights assigned to each model based on their performance or other criteria. The ensemble prediction p ^ is mathematically calculated as shown in Equation (29).
p ^ = k = 0 n W t k p ^ k
Here, n is the number of models, p ^ k is the prediction of the kth model, and W t k is the weight assigned to the kth model.
In soft voting, the predictions of several models are combined by averaging their anticipated class probabilities, with the confidence level of each model being used to weight the probability estimations. When all of the models’ class probabilities are averaged, the ensemble prediction results. Predictions using hard voting are based on majority voting, with the class with the most votes from the individual models being the final forecast. The mathematical mean (or most frequent class) of all the individual model predictions is the ensemble prediction. Conventional ML techniques like DTs, LR, SVMs, etc., may all benefit from these ensemble methods. Improved predictive performance and resilience are common outcomes of EL, combining many models’ predictions by weighted averaging, soft voting, or hard voting.

3. Proposed Model

The suggested model for accurate cancer detection is built using a set of well-planned stages that use ML methods to boost precision. As the first step in data preparation, we remove photos with a blank ratio higher than 30%. Images with a lot of empty space are not going to be considered for categorization until this requirement is met. This phase greatly improves the accuracy of the following predictions by keeping just the most pertinent and instructive pictures. Next, the photos’ contrast is improved using the CLAHE method, which stands for Contrast-Limited Adaptive Histogram Equalization. By adjusting the intensity levels, CLAHE makes key characteristics more visible, which is essential for correct categorization and extensive texture analysis.
The feature extraction process involves capturing complex texture patterns from the improved pictures using the Wavelet Multi-Channel Method (WMCM). In order to extract texture-based elements that are crucial for differentiating between non-cancerous and malignant pictures, WMCM, a variation of the 2D Wavelet Packet Transform (2DWPT), decomposes the image into several frequency sub-bands. SVM, LR, DT, RF, AdaBoost, and XGBoost are six classifiers that are used throughout the model training phase. Based on the retrieved textural information, each classifier acts as a base learner, first predicting cancer status. Improving prediction accuracy and resolving any discrepancies among these fundamental classifiers are achieved through the use of ensemble machine learning techniques. One of these methods is weighted averaging, which takes into account the relative performance of each base classifier and uses that information to combine their predictions. Another is hard voting, which uses the aggregated class labels to choose the most frequently predicted class, and the third is soft voting, which uses the average of the probability estimates to obtain the final prediction.
The balanced dataset is split into two halves, the training set and the test set, according to an 80:20 ratio. About 102 samples do not experience relapse, whereas 101 samples do (using synthetic samples produced by SMOTE), for a total of 203 samples in the training set. The test set includes 51 samples, out of which 26 were from relapses (including manufactured samples) and 25 were from non-relapses. A complete assessment of the model’s efficacy may be achieved thanks to its balanced distribution, guaranteeing that both sets have a representative mix of classes. Accuracy, precision, recall, and F-Value are some of the measures used to evaluate the model’s performance in properly classifying relapse and non-relapse instances. A comprehensive evaluation follows the suggested model’s robust training using various classifiers and ensemble approaches, which begins with a comprehensive preprocessing phase. The model next undergoes sophisticated feature extraction. Figure 1 shows the proposed model’s workflow, and Algorithm A1 (in Appendix A) provides the pseudocode for the operations’ sequence.

4. Empirical Analysis

Several assumptions are incorporated into the evaluation of this proposed ensemble ML method. A new ensemble ML method was developed to improve the evaluation metrics by applying six traditional ML techniques to the data generated via wavelet decomposition. The proposed idea was developed using two different computer systems. The main computer, System-1(S1), is a desktop PC with various specs, including an Intel Core i7 CPU, 32 GB of RAM, 1 TB of solid-state drive storage, and Ubuntu 20.04. The secondary system, System-2 (S2), has an Intel Core i5 CPU, 8 GB of RAM, 256 GB of primary SSD, 512 GB of secondary SSD, and runs Windows 10. A thorough empirical analysis of the collected data should be conducted as part of any endeavor. Using a systematic experimental approach, these metrics aim to construct a class confusion matrix that compares the actual results to predicted ones. The confusion matrix uses the letters TA and TB to represent true positives and negatives and FA and FB to represent false positives and false negatives. The study’s performance indicators for categorization include MCC, F-Value, Accuracy, Precision, Sensitivity, and Specificity. Equations (30)–(35) detail these metrics’ definitions [34,35].
A c c u r a c y = T A + T B T A + T B + F A + F B
P r e c i s i o n = T A T A + F A
S e n s i t i v i t y = T A T A + F B
S p e c i f i c i t y = T B T B + F A
F - V a l u e = 2 × T A 2 × T A + F B + F B
M C C = T A + T B F A + F B T A + F A T A + F B T B + F A T B + F B
This research used a two-stage experimental design. Stage-1 uses the following six ML methods: SVM, LR, DT, RF, AdaBoost, and XGBoost. In stage-2, these six ML methods are combined with weighted averaging, soft voting, and hard voting to create a more effective breast cancer relapse dataset model. Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 and Table 3 display the outcomes. The XGBoost method achieves an accuracy of 84.62% at its peak, as seen in Figure 2. In Figure 3, we can see the precision values that different ML methods achieved; XGBoost achieved the greatest, at 86.84%. Figure 4 shows the sensitivity values of several ML methods, with XGBoost achieving the highest value of 91.67%. The specificity values produced by different ML techniques are shown in Figure 5, with RF achieving the best value of 70.59%. Figure 6 shows the F-Values produced by several ML methods, with XGBoost achieving the best at 89.19%. Several ML methods are shown in Figure 7, with XGBoost producing the highest MCC value of 62.87%.
Then, in stage-2, we used weighted averaging, soft voting, and hard voting—three EL techniques—to improve the predictive outcomes. Table 4 and Figure 8 show the percentage of findings that were achieved. Compared to other ensemble ML techniques, ML approaches ensembled using weighted averaging achieve superior results (88.46% accuracy, 89.74% precision, 94.59% sensitivity, 73.33% specificity, a 92.11% F-Value, and 71.7% MCC). A phrase was coined to describe these improved results: the suggested ensemble ML strategy is based on the weighted average of several ML methods.
Additionally, using a receiver operating characteristics (ROC) curve plot, we have verified that the basic classifiers using weighted averaging as the EL method are part of the ensemble ML strategy we suggested. A ROC curve shows how well a classification model does on every level [23]. Two measures, TPR and FPR, are shown on the graph. There is a ROC curve that shows the level of categorical uncertainty. As a result of lowering the classification criteria, the number of false and true positives rises. The AUC or area under the ROC curve indicates general efficacy. A model’s AUC indicates its propensity to assign a higher score to positive examples than negative ones. The ROC curve remains within the (0, 0) to (1, 1) range. The ROC curve and AUC value for this proposed hybrid method are shown in Figure 9. Figure 9 shows that the suggested ensemble ML method uses the ROC curve to achieve the best AUC of 0.903 out of the three EL techniques.

4.1. Critical Analysis

The proposed ensemble approaches are compared in contrast to the state-of-the-art ML classifiers to show the efficacy of the proposed work. From Table 3 and Table 4, the improvement of the proposed work can be clearly observed:
  • While comparing the proposed weighted averaging of ML approaches in contrast to the ML classifiers, it outperforms the SVM, LR, DT, RF, AdaBoost, and XGBoost in terms of accuracies by ~12.19%, ~15%, ~9.52%, ~6.98%, ~6.98%, and ~4.54%, respectively.
  • In the case of soft voting, it outperforms the considered ML classifiers, including SVM, LR, DT, RF, AdaBoost, and XGBoost, in terms of accuracies by ~7.32%, ~10.01%, ~4.77%, ~2.33%, ~2.33%, and ~0%, respectively.
  • Comparing the hard voting ensemble technique with the ML classifiers, it outperforms SVM, LR, DT, RF, AdaBoost, and XGBoost in terms of accuracies by ~9.75 ~12.51%, ~7.14%, ~4.66%, ~4.66%, and ~2.27%, respectively.
  • From the above-stated comparison, it can be clearly observed that all ensemble techniques outperform the ML classifies more or less. However, three ensemble techniques were compared to find the best-fit ensemble technique for the current work. From the results obtained, it can be concluded that the weighted averaging outperforms soft voting and hard voting techniques by ~4.54% and ~2.22%, respectively, which states that the reported weighted averaging is the best-fitting ensemble technique for the current work.
  • Figure 10 presents a percentage-based comparison of results, illustrating the performance differences between the best EL approach and the best ML approach across various evaluation metrics:
    Weighted averaging achieves an accuracy of 88.46%, outperforming XGBoost, which obtained 84.62%. This highlights a notable improvement of ~3.84%.
    The precision for weighted averaging is 89.74%, compared to 86.84% for XGBoost, indicating an increase of ~2.9% in precision.
    In terms of sensitivity, weighted averaging reaches 94.59%, surpassing XGBoost’s sensitivity of 91.67% by ~2.92%. This suggests that weighted averaging is better at correctly identifying positive cases.
    Weighted averaging achieves 73.33% specificity, which is significantly higher than XGBoost’s 68.75%, resulting in an improvement of ~4.58% in identifying negative cases.
    The F-Value for weighted averaging is 92.11%, compared to 89.19% for XGBoost, showing an enhancement of ~2.92%.
    For MCC, weighted averaging scores 71.07%, significantly better than XGBoost’s 62.87%, demonstrating an improvement of ~8.2%, which reflects the overall quality of binary classifications.

4.2. Comparative Analysis

In addition to the above comparison among the ensemble techniques and ML classifier, we have also compared the results of using the proposed model with the weighted averaging technique with some of the existing literature, regardless of the dataset used for the work, as shown in Table 5, to support the novelty of the proposed work. When comparing the proposed work to existing literature, it becomes evident that in most situations, the proposed work outperforms the latter using all of the evaluated factors. As shown in Table 5, when comparing the proposed method with the works from the references, our proposed model outperforms most of them, particularly when using the TCGA dataset, as demonstrated in references [18,19,20]. This consistent improvement across multiple performance parameters, including AUC, clearly shows the effectiveness and novelty of our approach.
It could be noted that not all referenced studies provided comparable data for all metrics, such as MCC or specificity. However, we included these studies in the comparison table to provide a comprehensive view of how our model compares with various approaches used in breast cancer diagnosis research. This allows for a broader understanding of the performance of our method across different contexts and datasets. While direct comparisons were made for studies using the same dataset (TCGA), we also sought to provide context by including studies that used different datasets but evaluated similar performance parameters. This further reinforces the robustness and adaptability of our proposed method.

5. Conclusions

It is essential to accurately forecast the likelihood of relapse and metastasis in each patient to choose the most effective adjuvant treatment and initiate early intervention strategies since the relapse of breast cancer is directly associated with death. Patients who have had surgery for breast cancer are attracting more and more attention from researchers interested in computer-assisted diagnosis and prognosis prediction using H&E pathology images. To assess the possibility of breast cancer metastasis and relapse using histopathological images, we developed an innovative structure in this study that uses EL, ML, and image features. It is essential to employ fundamental numerical features that properly characterize histological images without altering their visual appearance to avoid misleading pathologists into making erroneous diagnoses to cut computer expenses. We overcame these obstacles by extracting texture information from low-order color moments and utilizing a Wavelet Multi-sub-band Co-occurrence Matrix to detect the color characteristics of histological images. We selected the most effective prediction models and extracted the most valuable characteristics using cross-validation. This research used a two-stage experimental design. SVM, LR, DT, RF, AdaBoost, and XGBoost are the six ML techniques used in stage-1. In stage-2, these techniques are combined with weighted averaging, soft voting, and hard voting to produce better results. Using the TCGA data warehouse, we were able to compile information on 123 breast cancer patients who had HER2-positive results. All of these patients’ relapse and metastasis statuses were known, and they had easy access to their H&E images. After two rounds of testing, the improved outcomes of an AUC of 0.903, an F-Value of 92.11%, an MCC of 71.07%, a sensitivity of 94.59%, a specificity of 73.33%, and an accuracy of 88.46% are all achieved by the suggested ensembled ML technique.
The advantages and disadvantages of each study are inevitable. The following are the limitations of the study. The findings may not be generalizable to larger groups because of the study’s restricted sample size (123 HER2-positive breast cancer patients). Previous research has examined the impact of sample size on the dependability of machine learning models. Although SMOTE addressed the class imbalance between relapse and non-relapse samples, it is crucial to acknowledge that synthetic data may introduce noise or inadequately capture the complexities of real-world data, thus affecting model performance. The histopathological images utilized in the investigation were derived solely from a single dataset (TCGA), perhaps failing to encompass the full spectrum of breast cancer manifestations across diverse populations or contexts. Research has shown that external validation across several datasets is crucial for validating prediction models. Although ensemble learning models yield superior predictions, they often lack the interpretability of simpler models, complicating their application in clinical decision-making.
Furthermore, with the use of H&E images, clinical data, and sophisticated DL algorithms, patients with HER2-positive breast cancer may have their chances of relapse and metastasis assessed. For example, our multimodal technique might be expanded in future work to include data from histopathological imaging and gene expression analysis. We aim to employ larger cohorts from more locations to improve our training data and achieve greater generalization. This will help us validate our models on a larger scale. Future work will also validate the model’s performance on external datasets and real-world clinical samples to ensure its robustness and applicability in diverse settings.

Author Contributions

Conceptualization, G.S., A.P. (Amrutanshu Panigrahi), and A.P. (Abhilash Pati); methodology, G.S.; software, G.S.; validation, A.K.N. and P.K.T.; formal analysis, C.M. and B.S.; investigation, A.K.N. and P.K.T.; resources, S.M.; data curation, G.S.; writing—original draft preparation, G.S., A.P. (Amrutanshu Panigrahi), and A.P. (Abhilash Pati); writing—review and editing, A.K.N., P.K.T., B.S., C.M., and S.M.; visualization, C.M.; supervision, A.K.N. and P.K.T.; project administration, A.P. (Abhilash Pati) and A.P. (Amrutanshu Panigrahi). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: The dataset included for this study can be found in The Cancer Genome Atlas (TCGA) [Link: https://portal.gdc.cancer.gov/ (accessed on 22 July 2023)].

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Algorithm A1: Pseudocode of the Proposed Model Working Principle.
Input: Input image set (D), total pixel (TP), Gray Level intensity (8/16/24), number of images (I)
Output: Cancer Relapse, Cancer Non-Relapse
Procedure:
Initiate the image preprocessing phase-1 to calculate the blank ratio of an image from the dataset.
For t 1←I
Find the number of Blank Pixels (B)
  Find the TP
  Find the Blank Ratio (BR) as B T P × 100 %
  If (BR > 30%)
    Discard the image
  Else
    Keep the image
  EndIf
  Return D’ with a set of images having BR < 30%
EndFor
Initiate the preprocessing phase-2 (CLAHE) to D’ for enhancing the intensity level contrast
Divide the original image into small, non-overlapping tiles (ST)
For t 1←ST
  Calculate the Cumulative distribution function (C) using Equation (1).
  Clip the C to limit ( C l i p m a x ) the contrast level for the current intensity
  Return image i
EndFor
Form update the dataset D’ with enhanced Contrast level
Apply WMCM to D’.
For t ← 1 to I
Divide the image into sub-bands S B { L L , L H , H L ,   H H }
For k ← 1 to |SB|
  Find GLCM for each intensity (i,j) of the image
     G L C M S B i , j = m = 1 M n = 1 N 1 ,   i f   H a , b = i   , H a , b = j 0 ,   O t h e r w i s e
  EndFor
   W M C M ( i , j ) = k = 1 K G L C M S B k i , j
EndFor
Initiate Texture feature extraction from WMCM
For t ← 1 to I
  Calculate S G S D A ,   S G B D A ,   G L A ,   D L A ,   G L M S E ,   D L M S E ,   C ,   C o ,   E ,   E n , H
  Calculate Texture Feature ( T F i , j ) for intensity (i,j)
EndFor
Find TF by combining calculated T F i , j
Split the TF with test size 0.2
Apply Base Learners SVM, LR, DT, RF, AdaBoost, and XGBoost to form an initial prediction.
Apply Weighted Averaging, Hard Voting, and Soft Voting to form ensemble models.
Evaluate the trained ensemble model over Test Data
Cancer Classification.

References

  1. Liu, H.; Qiu, C.; Wang, B.; Bing, P.; Tian, G.; Zhang, X.; Ma, J.; He, B.; Yang, J. Evaluating DNA methylation, gene expression, somatic mutation, and their combinations in inferring tumor tissue-of-origin. Front. Cell Dev. Biol. 2021, 9, 619330. [Google Scholar] [CrossRef]
  2. Ahmad, A. Current updates on trastuzumab resistance in HER2 overexpressing breast cancers. In Breast Cancer Metastasis and Drug Resistance: Challenges and Progress; Springer: Cham, Switzerland, 2019; pp. 217–228. [Google Scholar] [CrossRef]
  3. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  4. He, B.; Lang, J.; Wang, B.; Liu, X.; Lu, Q.; He, J.; Gao, W.; Bing, P.; Tian, G.; Yang, J. TOOme: A novel computational framework to infer cancer tissue-of-origin by integrating both gene mutation and expression. Front. Bioeng. Biotechnol. 2020, 8, 394. [Google Scholar] [CrossRef] [PubMed]
  5. Farahmand, S.; Fernandez, A.I.; Ahmed, F.S.; Rimm, D.L.; Chuang, J.H.; Reisenbichler, E.; Zarringhalam, K. Deep learning trained on hematoxylin and eosin tumor region of Interest predicts HER2 status and trastuzumab treatment response in HER2+ breast cancer. Mod. Pathol. 2022, 35, 44–51. [Google Scholar] [CrossRef] [PubMed]
  6. Ji, M.Y.; Yuan, L.; Jiang, X.D.; Zeng, Z.; Zhan, N.; Huang, P.X.; Lu, C.; Dong, W.G. Nuclear shape, architecture and orientation features from H&E images are able to predict recurrence in node-negative gastric adenocarcinoma. J. Transl. Med. 2019, 17, 92. [Google Scholar] [CrossRef] [PubMed]
  7. Dang, X.; Xiong, G.; Fan, C.; He, Y.; Sun, G.; Wang, S.; Liu, Y.; Zhang, L.; Bao, Y.; Xu, J.; et al. Systematic external evaluation of four preoperative risk prediction models for severe postpartum hemorrhage in patients with placenta previa: A multicenter retrospective study. J. Gynecol. Obstet. Hum. Reprod. 2022, 51, 102333. [Google Scholar] [CrossRef]
  8. Li, H.; Shi, W.; Shen, T.; Hui, S.; Hou, M.; Wei, Z.; Qin, S.; Bai, Z.; Cao, J. Network pharmacology-based strategy for predicting therapy targets of Ecliptae Herba on breast cancer. Medicine 2023, 102, e35384. [Google Scholar] [CrossRef]
  9. Sakri, S.B.; Rashid, N.B.; Zain, Z.M. Particle swarm optimization feature selection for breast cancer recurrence prediction. IEEE Access 2018, 6, 29637–29647. [Google Scholar] [CrossRef]
  10. Alom, M.Z.; Yakopcic, C.; Nasrin, M.S.; Taha, T.M.; Asari, V.K. Breast cancer classification from histopathological images with inception recurrent residual convolutional neural network. J. Digit. Imaging 2019, 32, 605–617. [Google Scholar] [CrossRef]
  11. Hong, H.C.; Chuang, C.H.; Huang, W.C.; Weng, S.L.; Chen, C.H.; Chang, K.H.; Liao, K.W.; Huang, H.D. A panel of eight microRNAs is a good predictive parameter for triple-negative breast cancer relapse. Theranostics 2020, 10, 8771. [Google Scholar] [CrossRef]
  12. Yan, S.; Wang, W.; Zhu, B.; Pan, X.; Wu, X.; Tao, W. Construction of nomograms for predicting pathological complete response and tumor shrinkage size in breast cancer. Cancer Manag. Res. 2020, 20, 8313–8323. [Google Scholar] [CrossRef] [PubMed]
  13. Mosayebi, A.; Mojaradi, B.; Bonyadi Naeini, A.; Khodadad Hosseini, S.H. Modeling and comparing data mining algorithms for prediction of recurrence of breast cancer. PLoS ONE 2020, 15, e0237658. [Google Scholar] [CrossRef]
  14. Comes, M.C.; La Forgia, D.; Didonna, V.; Fanizzi, A.; Giotta, F.; Latorre, A.; Martinelli, E.; Mencattini, A.; Paradiso, A.V.; Tamborra, P.; et al. Early prediction of breast cancer recurrence for patients treated with neoadjuvant chemotherapy: A transfer learning approach on DCE-MRIs. Cancers 2021, 13, 2298. [Google Scholar] [CrossRef] [PubMed]
  15. Sanyal, J.; Tariq, A.; Kurian, A.W.; Rubin, D.; Banerjee, I. Weakly supervised temporal model for prediction of breast cancer distant recurrence. Sci. Rep. 2021, 11, 9461. [Google Scholar] [CrossRef]
  16. Conde-Sousa, E.; Vale, J.; Feng, M.; Xu, K.; Wang, Y.; Della Mea, V.; La Barbera, D.; Montahaei, E.; Baghshah, M.; Turzynski, A.; et al. HEROHE challenge: Predicting HER2 status in breast cancer from hematoxylin–eosin whole-slide imaging. J. Imaging 2022, 8, 213. [Google Scholar] [CrossRef]
  17. Rabinovici-Cohen, S.; Fernández, X.M.; Grandal Rejo, B.; Hexter, E.; Hijano Cubelos, O.; Pajula, J.; Pölönen, H.; Reyal, F.; Rosen-Zvi, M. Multimodal prediction of five-year breast Cancer recurrence in women who receive Neoadjuvant chemotherapy. Cancers 2022, 14, 3848. [Google Scholar] [CrossRef]
  18. Liu, X.; Yuan, P.; Li, R.; Zhang, D.; An, J.; Ju, J.; Liu, C.; Ren, F.; Hou, R.; Li, Y.; et al. Predicting breast cancer recurrence and metastasis risk by integrating color and texture features of histopathological images and machine learning technologies. Comput. Biol. Med. 2022, 146, 105569. [Google Scholar] [CrossRef]
  19. Yang, J.; Ju, J.; Guo, L.; Ji, B.; Shi, S.; Yang, Z.; Gao, S.; Yuan, X.; Tian, G.; Liang, Y.; et al. Prediction of HER2-positive breast cancer recurrence and metastasis risk from histopathological images and clinical information via multimodal deep learning. Comput. Struct. Biotechnol. J. 2022, 20, 333–342. [Google Scholar] [CrossRef]
  20. Lu, W.; Toss, M.; Dawood, M.; Rakha, E.; Rajpoot, N.; Minhas, F. SlideGraph+: Whole slide image level graphs to predict HER2 status in breast cancer. Med. Image Anal. 2022, 80, 102486. [Google Scholar] [CrossRef]
  21. Su, Z.; Niazi, M.K.; Tavolara, T.E.; Niu, S.; Tozbikian, G.H.; Wesolowski, R.; Gurcan, M.N. BCR-Net: A deep learning framework to predict breast cancer recurrence from histopathology images. PLoS ONE 2023, 18, e0283562. [Google Scholar] [CrossRef]
  22. Liu, Y.; Shen, D.; Wang, H.Y.; Qi, M.Y.; Zeng, Q.Y. Development and validation to predict visual acuity and keratometry two years after corneal crosslinking with progressive keratoconus by machine learning. Front. Med. 2023, 10, 1146529. [Google Scholar] [CrossRef] [PubMed]
  23. Botlagunta, M.; Botlagunta, M.D.; Myneni, M.B.; Lakshmi, D.; Nayyar, A.; Gullapalli, J.S.; Shah, M.A. Classification and diagnostic prediction of breast cancer metastasis on clinical data using machine learning algorithms. Sci. Rep. 2023, 13, 485. [Google Scholar] [CrossRef] [PubMed]
  24. Dammu, H.; Ren, T.; Duong, T.Q. Deep learning prediction of pathological complete response, residual cancer burden, and progression-free survival in breast cancer patients. PLoS ONE 2023, 18, e0280148. [Google Scholar] [CrossRef]
  25. Bayramoglu, N.; Kannala, J.; Heikkilä, J. Deep learning for magnification independent breast cancer histopathology image classification. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2440–2445. [Google Scholar] [CrossRef]
  26. Rawat, R.R.; Ortega, I.; Roy, P.; Sha, F.; Shibata, D.; Ruderman, D.; Agus, D.B. Deep learned tissue “fingerprints” classify breast cancers by ER/PR/Her2 status from H&E images. Sci. Rep. 2020, 10, 7275. [Google Scholar] [CrossRef]
  27. Pati, A.; Panigrahi, A.; Parhi, M.; Giri, J.; Qin, H.; Mallik, S.; Pattanayak, S.R.; Agrawal, U.K. Performance assessment of hybrid machine learning approaches for breast cancer and recurrence prediction. PLoS ONE 2024, 19, e0304768. [Google Scholar] [CrossRef]
  28. The Cancer Genome Atlas (TCGA). Available online: https://portal.gdc.cancer.gov/ (accessed on 22 July 2023).
  29. Alwazzan, M.J.; Ismael, M.A.; Ahmed, A.N. A hybrid algorithm to enhance colour retinal fundus images using a Wiener filter and CLAHE. J. Digit. Imaging 2021, 34, 750–759. [Google Scholar] [CrossRef]
  30. Hayati, M.; Muchtar, K.; Maulina, N.; Syamsuddin, I.; Elwirehardja, G.N.; Pardamean, B. Impact of CLAHE-based image enhancement for diabetic retinopathy classification through deep learning. Procedia Comput. Sci. 2023, 216, 57–66. [Google Scholar] [CrossRef]
  31. Srivastava, R.; Kumar, P. Deep-GAN: An improved model for thyroid nodule identification and classification. Neural Comput. Appl. 2024, 36, 7685–7704. [Google Scholar] [CrossRef]
  32. Pati, A.; Panigrahi, A.; Nayak, D.S.; Sahoo, G.; Singh, D. Predicting pediatric appendicitis using ensemble learning techniques. Procedia Comput. Sci. 2023, 218, 1166–1175. [Google Scholar] [CrossRef]
  33. Panigrahi, A.; Pati, A.; Sahu, B.; Das, M.N.; Nayak, D.S.; Sahoo, G.; Kant, S. En-MinWhale: An ensemble approach based on MRMR and Whale optimization for Cancer diagnosis. IEEE Access 2023, 11, 113526–113542. [Google Scholar] [CrossRef]
  34. Sahoo, G.; Nayak, A.K.; Tripathy, P.K.; Tripathy, J. A novel machine learning based hybrid approach for breast cancer relapse prediction. Indones. J. Electr. Eng. Comput. Sci. (IJEECS) 2023, 32, 1655–1663. [Google Scholar] [CrossRef]
  35. Pati, A.; Panigrahi, A.; Sahu, B.; Sahoo, G.; Dash, M.; Parhi, M.; Pattanayak, B.K. FOHC: Firefly Optimizer Enabled Hybrid approach for Cancer Classification. Int. J. Recent Innov. Trends Comput. Commun. 2023, 11, 118–125. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed model.
Figure 1. Block diagram of the proposed model.
Curroncol 31 00486 g001
Figure 2. Accuracies obtained in % employing various ML approaches.
Figure 2. Accuracies obtained in % employing various ML approaches.
Curroncol 31 00486 g002
Figure 3. Precisions obtained in % employing various ML approaches.
Figure 3. Precisions obtained in % employing various ML approaches.
Curroncol 31 00486 g003
Figure 4. Sensitivities obtained in % employing various ML approaches.
Figure 4. Sensitivities obtained in % employing various ML approaches.
Curroncol 31 00486 g004
Figure 5. Specificities obtained in % employing various ML approaches.
Figure 5. Specificities obtained in % employing various ML approaches.
Curroncol 31 00486 g005
Figure 6. F-Values obtained in % employing various ML approaches.
Figure 6. F-Values obtained in % employing various ML approaches.
Curroncol 31 00486 g006
Figure 7. MCCs obtained in % employing various ML approaches.
Figure 7. MCCs obtained in % employing various ML approaches.
Curroncol 31 00486 g007
Figure 8. Findings acquired in % using various EL methods.
Figure 8. Findings acquired in % using various EL methods.
Curroncol 31 00486 g008
Figure 9. AUC values with ROC curves for the ensemble ML approaches.
Figure 9. AUC values with ROC curves for the ensemble ML approaches.
Curroncol 31 00486 g009
Figure 10. Percentage comparison of findings: best EL approach vs. best ML approach.
Figure 10. Percentage comparison of findings: best EL approach vs. best ML approach.
Curroncol 31 00486 g010
Table 1. Summary of the considered state-of-the-art works.
Table 1. Summary of the considered state-of-the-art works.
ReferenceMethodologiesDataset(s)Outcomes
Sakri et al. [9]REPTree, NB, and KNN-IBK with PSOWPBCAccuracy: 81.3%, Precision: 88.3%, Recall: 93.4%, F-Score: 87.7%, AUC: 0.820
Alom et al. [10]Inception-v4, ResNet, and RCNNBreakHis and BCDAccuracy: 100%, Sensitivity: 100%, Specificity: 100%, AUC: 1.0
Hong et al. [11]LR and Gaussian mixture TCGA_TNBC, GEOD-40525, GSE40049 and GSE19783AUC (GSE40049): 0.89, AUC (GSE19783): 0.90
Yan et al. [12]LRBCC at HMUCH AUC (HER2-Positive): 0.820, AUC (TNBC): 0.785
Mosayebi et al. [13]RF, LVQ, NB, C5.0 DT, MLP, KPCA-SVM, and SVMMHME-ICRCAccuracy: 81.9%, Sensitivity: 86.9%, Specificity:77.7%, F-Value: 81.6%, AUC: 0.774
Comes et al. [14]CNN and SVMI-SPY1 TRIAL and BREAST-MRI-NACT-PilotAccuracy: 85.2%, Sensitivity: 84.6%, AUC: 0.83
Sanyal et al. [15]LSTM and XGBoostManually and NLP-curatedSensitivity: 89.0%, Specificity: 84.0%, AUC: 0.94
Conde-Sousa et al. [16]DL HEROHEPrecision: 79%, Recall: 100%, F-Value: 79%, AUC: 0.88
Rabinovici-Cohen et al. [17]CNNReal-World Retrospective DatasetSpecificity: 57%, F-Value: 56%, Balanced Accuracy: 72%, PPV: 41%, NPV: 96%, Sensitivity: 93%, AUC: 0.75
Liu et al. [18]RF, LR, and XGBoostClinical dataset Accuracy: 80%, Precision: 40%, Recall: 75%, F-Value: 50%, AUC: 0.75
TCGAAccuracy: 73%, Precision: 33%, Recall: 60%, F-Value: 42%, AUC: 0.72
Yang et al. [19]CNN and ResNet50CAMSAUC: 0.76
TCGAAUC: 0.72
Lu et al. [20]GNN-based Slide Graph TCGAAUC: 0.75
HER2C and Nott-HER2AUC: 0.80
Su et al. [21]CNNH&E and Ki67Accuracy: 80%, F-Value: 79.2%, AUC: 0.811
Liu et al. [22]XGBoostAEHWUR-Square (CVDA): 0.9993, R-Square (Kmax): 0.9888
Botlagunta et al. [23]LR, KNN, DT, RF, SVM, GB, and XGBoostMedical data on MBCAccuracy: 83%, Precision: 83%, Recall: 100%, F-Value: 85%, AUC: 0.87
Dammu et al. [24]CNNI-SPY-1 TRIALAccuracy: 81%, Sensitivity: 68%, Specificity: 97%, F-Value: 76%, AUC: 0.83
Table 2. An overview of the considered TCGA dataset.
Table 2. An overview of the considered TCGA dataset.
DatasetParameters
Stages of TumorERPRLymph Node StatusAge (Years)OutcomeTotal
IIIIIIER−ER+PR−PR+LMN−LMN+<50≥50RNR
TCGA14773233905073586534895118123
Table 3. Results obtained by employing various ML approaches.
Table 3. Results obtained by employing various ML approaches.
ML
Approaches
Results Obtained (in %)
AccuracyPrecisionSensitivitySpecificityF-ValueMCC
SVM78.8583.3385.7164.7184.5151.25
LR76.9280.5685.2961.1182.8647.83
DT80.7786.1186.1168.7586.1154.86
RF82.6986.1188.5770.5987.3260.13
AdaBoost82.6986.4988.8968.7587.6758.72
XGBoost84.6286.8491.6768.7589.1962.87
Table 4. Results obtained in % by employing various EL approaches.
Table 4. Results obtained in % by employing various EL approaches.
EL
Approaches
Results Obtained (in %)
AccuracyPrecisionSensitivitySpecificityF-ValueMCC
Weighted
Averaging
88.4689.7494.5973.3392.1171.07
Soft Voting84.6286.1191.1872.2288.5765.35
Hard Voting 86.5486.4994.1272.2290.1469.66
Table 5. Comparison of the proposed work with some state-of-the-art works.
Table 5. Comparison of the proposed work with some state-of-the-art works.
Ref.Dataset(s)Comparison Parameters
Accuracy (%)Precision (%)Sensitivity (%)Specificity (%)F-Value (%)MCC (%)AUC
[9]WPBC81.3 88.393.4-87.7-0.820
[10]BreakHis and BCD100 -100100--1.0
[11]GSE40049------0.89
GSE19783------0.90
[12]BCC at HMUCH------0.820
[13]MHME-ICRC81.9 -86.977.781.6-0.774
[14]I-SPY1 TRIAL, BREAST-MRI-NACT-Pilot85.2-84.6---0.83
[15]Manually and NLP-curated- -89.084.0--0.94
[16]HEROHE-79100-79-0.88
[17]Real-World Retrospective Dataset--935756-0.75
[18]Clinical dataset 80 4075-50-0.75
TCGA733360-42-0.72
[19]CAMS------0.76
TCGA------0.72
[20]TCGA------0.75
HER2C and Nott-HER2------0.80
[21]H&E and Ki6780---79.2-0.811
[22]AEHWU-------
[23]Medical data on MBC83 83100-85-0.87
[24]I-SPY-1 TRIAL81 -689776-0.83
[Proposed]TCGA88.4689.7494.5973.3392.1171.070.903
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sahoo, G.; Nayak, A.K.; Tripathy, P.K.; Panigrahi, A.; Pati, A.; Sahu, B.; Mahanty, C.; Mallik, S. Predicting Breast Cancer Relapse from Histopathological Images with Ensemble Machine Learning Models. Curr. Oncol. 2024, 31, 6577-6597. https://doi.org/10.3390/curroncol31110486

AMA Style

Sahoo G, Nayak AK, Tripathy PK, Panigrahi A, Pati A, Sahu B, Mahanty C, Mallik S. Predicting Breast Cancer Relapse from Histopathological Images with Ensemble Machine Learning Models. Current Oncology. 2024; 31(11):6577-6597. https://doi.org/10.3390/curroncol31110486

Chicago/Turabian Style

Sahoo, Ghanashyam, Ajit Kumar Nayak, Pradyumna Kumar Tripathy, Amrutanshu Panigrahi, Abhilash Pati, Bibhuprasad Sahu, Chandrakanta Mahanty, and Saurav Mallik. 2024. "Predicting Breast Cancer Relapse from Histopathological Images with Ensemble Machine Learning Models" Current Oncology 31, no. 11: 6577-6597. https://doi.org/10.3390/curroncol31110486

APA Style

Sahoo, G., Nayak, A. K., Tripathy, P. K., Panigrahi, A., Pati, A., Sahu, B., Mahanty, C., & Mallik, S. (2024). Predicting Breast Cancer Relapse from Histopathological Images with Ensemble Machine Learning Models. Current Oncology, 31(11), 6577-6597. https://doi.org/10.3390/curroncol31110486

Article Metrics

Back to TopTop