Next Article in Journal
Clinical Outcomes of Immune Checkpoint Inhibitors in Unique Cohorts Underrepresented in Clinical Trials
Next Article in Special Issue
Early 2-Factor Transcription Factors Associated with Progression and Recurrence in Bevacizumab-Responsive Subtypes of Glioblastoma
Previous Article in Journal
Ex Vivo Fluorescence Confocal Microscopy for Intraoperative Examinations of Lung Tumors as Alternative to Frozen Sections—A Proof-of-Concept Study
Previous Article in Special Issue
Radiotherapy Plan Quality Assurance in NRG Oncology Trials for Brain and Head/Neck Cancers: An AI-Enhanced Knowledge-Based Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Deep Learning-Based Image Classification for Breast Cancer Subtype and Invasiveness Diagnosis from Whole Slide Image Histopathology

by
Aadhi Aadhavan Balasubramanian
1,
Salah Mohammed Awad Al-Heejawi
2,
Akarsh Singh
2,
Anne Breggia
3,
Bilal Ahmad
4,
Robert Christman
4,
Stephen T. Ryan
4 and
Saeed Amal
5,*
1
Khoury College of Computer Sciences, Northeastern University, Boston, MA 02115, USA
2
College of Engineering, Northeastern University, Boston, MA 02115, USA
3
MaineHealth Institute for Research, Scarborough, ME 04074, USA
4
Maine Medical Center, Portland, ME 04102, USA
5
The Roux Institute, Department of Bioengineering, College of Engineering, Northeastern University, Boston, MA 02115, USA
*
Author to whom correspondence should be addressed.
Cancers 2024, 16(12), 2222; https://doi.org/10.3390/cancers16122222
Submission received: 9 May 2024 / Revised: 7 June 2024 / Accepted: 10 June 2024 / Published: 14 June 2024
(This article belongs to the Collection Artificial Intelligence in Oncology)

Abstract

:

Simple Summary

Breast cancer is a significant cause of female cancer-related deaths in the US. Checking how severe the cancer is helps in planning treatment. Modern AI methods are good at grading cancer, but they are not used much in hospitals yet. We developed and utilized ensemble deep learning algorithms for addressing the tasks of classifying (1) breast cancer subtype and (2) breast cancer invasiveness from whole slide image (WSI) histopathology slides. The ensemble models used were based on convolutional neural networks (CNNs) known for extracting distinctive features crucial for accurate classification. In this paper, we provide a comprehensive analysis of these models and the used methodology for breast cancer diagnosis tasks.

Abstract

Cancer diagnosis and classification are pivotal for effective patient management and treatment planning. In this study, a comprehensive approach is presented utilizing ensemble deep learning techniques to analyze breast cancer histopathology images. Our datasets were based on two widely employed datasets from different centers for two different tasks: BACH and BreakHis. Within the BACH dataset, a proposed ensemble strategy was employed, incorporating VGG16 and ResNet50 architectures to achieve precise classification of breast cancer histopathology images. Introducing a novel image patching technique to preprocess a high-resolution image facilitated a focused analysis of localized regions of interest. The annotated BACH dataset encompassed 400 WSIs across four distinct classes: Normal, Benign, In Situ Carcinoma, and Invasive Carcinoma. In addition, the proposed ensemble was used on the BreakHis dataset, utilizing VGG16, ResNet34, and ResNet50 models to classify microscopic images into eight distinct categories (four benign and four malignant). For both datasets, a five-fold cross-validation approach was employed for rigorous training and testing. Preliminary experimental results indicated a patch classification accuracy of 95.31% (for the BACH dataset) and WSI image classification accuracy of 98.43% (BreakHis). This research significantly contributes to ongoing endeavors in harnessing artificial intelligence to advance breast cancer diagnosis, potentially fostering improved patient outcomes and alleviating healthcare burdens.

1. Introduction

Breast cancer, constituting 31% of all female cancers, remains a significant health concern, necessitating precise diagnostic measures [1]. Estimations by the American Cancer Society (ACS) project 297,790 newly diagnosed cases of invasive breast cancer among women in 2024 [2]. Early detection and accurate diagnosis are crucial for effective treatment and better patient outcomes. Histopathology, considered the gold standard for breast cancer diagnosis, involves microscopic examination of tissue specimens to identify cancerous abnormalities. However, manual interpretation of histopathological images is time-consuming, prone to inter-observer discrepancies, and reliant on pathologist expertise [3,4,5]. In response, computational methodologies, particularly deep learning-driven image analysis, offer promising avenues to enhance diagnostic precision and efficiency.
Recent studies have investigated various methodologies to advance breast cancer histopathology image classification using the same datasets. Notably, research efforts have centered around three themes based on our detailed review: ensemble techniques, the direct utilization of specific convolutional neural network (CNN) architectures, and the integration of transfer learning with novel model designs. In this research, several studies are reviewed to identify gaps in the research, thereby paving the way for innovations aimed at improving diagnostic accuracy and efficiency through more robust AI-driven algorithms and state-of-the-art techniques in clinical practice.
To explore ensemble techniques, ref. [6] used ResNet [7] models, emphasizing fine-tuning and downsampling strategies to address computational complexities, achieving commendable accuracies of 86% on the BACH (Grand Challenge on Breast Cancer Histology images) test set [8] and 97.2% on the Bioimaging 2015 challenge dataset. Ref. [9] employed ResNet-101 and DenseNet-161 [10] networks, showcasing a notable accuracy of 87% on BACH, albeit with varying sensitivities across classes. Both studies elucidated the efficacy of ensemble approaches incorporating pre-trained CNNs. Similarly, ref. [11] introduced a hybrid CNN architecture integrating local and global model branches, achieving competitive accuracies of 87.5% on the BACH dataset and 85.2% on the Breast Cancer Histopathological Image Classification (BreakHis) dataset [12]. These studies demonstrated the potential of ensemble techniques to improve classification accuracy and robustness. Nevertheless, there are opportunities to enhance consistency in sensitivity across different classes and further optimize the handling of computational complexities. In our study, these ensemble techniques were enhanced by incorporating advanced optimization strategies and adaptive ensemble methods to ensure consistent performance across all classes. Additionally, efficient computational resource management techniques were employed to address the computational challenges more effectively.
Multiple studies have underscored the efficacy of the direct application of established CNN architectures. Ref. [13] demonstrated the utility of VGG16 [14] in breast cancer grading, achieving a commendable accuracy of 83% on the BACH test set. Similarly, ref. [15] pioneered a hybrid convolutional and recurrent deep learning method, yielding notable accuracies of 82.1% and 91.3% at patch and image levels, respectively. These studies highlighted the effectiveness of using well-established CNN architectures for breast cancer histopathology image classification. However, they often did not explore the full potential of integrating different CNN architectures to further improve performance.
Moreover, numerous studies emphasized the integration of transfer learning with innovative model architectures. Ref. [16] delineated a pipeline for breast cancer histopathology image classification, leveraging a two-stage neural network approach and achieving a particularly high image-level accuracy of 97.5%. For classification between benign and malignant, ref. [17] proposed DenTnet, a deep learning model that achieved an accuracy of 99.28% on the BreakHis dataset. Comparably, ref. [18] introduced BreastNet, an end-to-end model that incorporates a convolutional block attention module (CBAM), dense block, residual block, and hypercolumn technique. Ref. [19] developed DenseNet121-AnoGAN, which consists of two components: an unsupervised anomaly detection segment using generative adversarial networks to screen mislabeled patches and a densely connected CNN (DenseNet) for extracting multi-layered features from discriminative patches. Additionally, refs. [20,21] utilized transfer learning on the BreakHis dataset, employing ResNet-18 and Inception-V3, respectively. Further advancements were made by ref. [22], who proposed an enhanced autoencoder network utilizing a Siamese framework, and ref. [23], who designed DRDA-Net, a model featuring dense attention and dual-shuffle attention mechanisms to guide deep learning processes. Concurrently, ref. [24] introduced a hierarchical structure utilizing ResNeXt50 models, achieving notable accuracies of 81% on the BACH challenge test set. These studies showed the potential of transfer learning and novel model designs to achieve high accuracy and effective classification in breast cancer histopathology images. However, there was often a lack of generalizability across different datasets and for practical application in diverse clinical settings. In our study, we improved upon these approaches by developing more generalizable models that can be effectively applied across a wide range of datasets and clinical environments.
As a summary, the collective endeavors in breast cancer histopathology image classification underscore the significance of ensemble strategies [25,26,27,28], direct application of CNN architectures [29,30,31], and the fusion of transfer learning with innovative model designs [32,33,34]. Inspired by these pioneering works, our study aimed to build upon existing methodologies and enhance performance on the aforementioned datasets through innovative techniques and meticulous experimentation. This research addresses histopathology image analysis limitations through advanced deep learning algorithms, focusing on breast cancer classification using VGG16 and ResNet50 architectures. These CNN models were selected for their ability to extract distinctive features crucial for accurate classification. Employing an image patching technique, each image was divided into smaller patches for focused analysis, thus managing computational demands. Initial experiments achieved a promising 89.84% accuracy in patch-level classification. This study provides a comprehensive investigation into deep learning for breast cancer histopathology classification, detailing the methodology, including image patching and CNN implementation, alongside ensemble learning techniques. The results, analysis, and discussions highlight the approach’s efficacy in surpassing the performance of existing studies on the two specified datasets and potential implications for breast cancer diagnosis improvement.

2. Materials and Methods

2.1. Dataset and Implementation Details

The BACH dataset utilized in this study comprises hematoxylin and eosin (H&E)-stained breast histology microscopy images and whole-slide images. Hematoxylin and eosin (H&E)-stained breast histology microscopy images were valuable in this study because they provided detailed visual information regarding the cellular structure, morphology, and spatial distribution of tissues within the breast, aiding in the interpretation of pathological features relevant to breast cancer research.
In the BACH dataset, we focused on the microscopy images, which are meticulously labeled by medical experts based on the predominant cancer type present in each image. Specifically, images are categorized into four classes: Normal, Benign, In Situ Carcinoma, and Invasive Carcinoma (see Figure 1). The annotation process involved the assessment of each image by two medical experts, with any disagreements resulting in the exclusion of the respective image from the dataset. In total, the microscopy image dataset consisted of 400 images, evenly distributed across the four classes, with each class containing 100 images. These images were stored in .tiff format and possessed specific characteristics: a color model of red–green–blue (RGB), dimensions of 2048 × 1536 pixels, and a pixel scale of 0.42 µm × 0.42 µm. Additionally, the memory space required for each image ranged from 10 to 20 MB, approximately. Notably, the labeling of the dataset was image-wise, providing comprehensive information about the cancer type represented in each image. This meticulously curated dataset served as a valuable resource for training and evaluating machine learning models for breast cancer histopathology image analysis.
The BreakHis dataset consists of 9109 microscopic images of breast tumor tissue. These images were collected from 82 patients and captured at various magnification factors (40×, 100×, 200×, and 400×). The dataset includes 2480 benign and 5429 malignant samples. Each sample is a 700 × 460 pixel, 3-channel RGB image with an 8-bit depth in each channel, stored in PNG format. The dataset categorizes tumors into benign and malignant classes, each with several subclasses. For benign tumors, these subclasses include Adenosis, Fibroadenoma, Tubular adenoma, and Phyllodes adenoma. For malignant tumors, the subclasses include Ductal carcinoma, Lobular carcinoma, Mucinous carcinoma, and Papillary carcinoma. Figure 2 displays eight images, with the first column representing four types of malignant tissues and the second column representing four types of benign tissues.

2.2. Data Preprocessing

Breast cancer histopathology images, often characterized by large dimensions and high-resolution details, present challenges in computational analysis. To address this, an image patching technique was used on the BACH dataset to segment the images into smaller, more manageable patches. Each original image, with dimensions of 2048 × 1536 pixels and a pixel scale of 0.42 µm × 0.42 µm, underwent partitioning into six smaller images. This approach effectively reduced the computational burden and memory requirements, which typically ranged from 10 to 20 MB per image. Moreover, employing image-wise labeling facilitated the categorization of patches into distinct classes, including Benign, In Situ, Invasive, and Normal. The patching algorithm involved dividing the image into two halves horizontally and further segmenting each half into three parts vertically. Subsequently, each patch was resized to a target size of 256 × 256 pixels and saved individually. This patching strategy not only facilitated efficient processing but also allowed for focused analysis on specific regions of interest within the histopathology images, thereby enhancing classification accuracy.
In the BreakHis dataset, no preprocessing was applied. The data, which included four scales (40×, 100×, 200×, and 400×), was used to train a single model using multiscale magnification. The data were divided into five folds, with each fold containing all four scale magnifications. There were around 39,545 images of all magnifications, each with a size of 3 × 700 × 460, divided into five folds.

2.3. Implementation of VGG16 and ResNet50 Architectures

For our image classification tasks, two powerful convolutional neural network (CNN) models, VGG16 and ResNet50, were utilized. These models have been proven effective in various computer vision tasks and provided a solid foundation for our classification efforts.
The VGG16 architecture is a well-known CNN model comprising multiple convolutional and max-pooling layers, followed by fully connected layers. To expedite our implementation, we employed transfer learning by utilizing pre-trained weights from the ImageNet dataset. We initialized the VGG16 model without its top classification layers and froze the weights of the convolutional layers to preserve the learned features. Custom fully connected layers were then added on top of the VGG16 architecture for classification purposes. During training, we fine-tuned these custom layers while keeping the convolutional layers frozen. This approach facilitated efficient training on our breast cancer histopathology dataset, resulting in accurate classification results.
In addition to VGG16, we incorporated the ResNet50 architecture, another prominent CNN model, into our classification framework. ResNet50 is characterized by its deep structure with skip connections, which helped alleviate the vanishing gradient problem and enabled the training of very deep neural networks. Similar to our approach with VGG16, we loaded the pre-trained ResNet50 model without its classification layers and froze the weights of the convolutional layers to retain learned features. Custom fully connected layers were then added for classification purposes. By leveraging ResNet50, we aimed to capture intricate features from the breast cancer histopathology images, potentially enhancing classification performance compared to shallower architectures.

2.4. Implementation of Ensemble Learning

To enhance predictive performance, an ensemble learning approach was adopted by combining predictions from both the VGG16 and ResNet50 models. The ensemble model leveraged the predicted probabilities generated by the individual models. The implementation involved loading the pre-trained VGG16 and ResNet50 models and creating a new model that combined their outputs using averaging. The ensemble model was then compiled with appropriate loss and optimization functions. By harnessing the complementary strengths of VGG16 and ResNet50, our ensemble model aimed to achieve enhanced classification accuracy and robustness, contributing to more reliable breast cancer histopathology image analysis.
Figure 3 presents a schematic of the proposed ensemble model. The model configurations for ResNet34 and VGG16 were as follows: both models were pre-trained and designed for input image size (256 × 256 × 3) to obtain N classes. Specifically, in ResNet34, the last fully connected layer was modified to obtain N features from the 512 input features (ResNet34 pre-trained model features), whereas the last classifier layer in VGG 16 was modified to obtain N features from the 1000 input features (VGG16 pre-trained model features). Both models utilized the CrossEntropyLoss function and the SGD optimizer with a learning rate of 0.001 and momentum of 0.9.

3. Results

In this section, the classification outcomes obtained from our study on two distinct datasets are presented. The performance of machine learning algorithms, including VGG16, ResNet, and EfficientNet, is evaluated in comparison with our proposed novel model in this research.
On the BACH dataset, three architectures, VGG16, ResNet, and an ensemble model, were trained and evaluated. Their performance was assessed using various metrics, including training and validation accuracies, obtained through both single-model and ensemble approaches. Table 1 provides a comprehensive summary of the training and validation accuracies, as well as the losses for 5-fold cross-validation, achieved by each architecture on the BACH dataset.
The VGG16 architecture demonstrated commendable results, achieving a training accuracy of 98.72% and a validation accuracy of 79.95%, as depicted in Figure 4a. Similarly, the ResNet architecture exhibited strong performance, with a training accuracy of 99.05% and a validation accuracy of 80.47%, as illustrated in Figure 4b. Notably, the ensemble model, trained with 5-fold cross-validation, showcased superior performance, boasting a training accuracy of 99.84% and a validation accuracy of 92.58%, as indicated in Figure 4c. Furthermore, the ensemble model trained without cross-validation attained the highest accuracy, with a training accuracy of 99.78% and a validation accuracy of 95.31%, as illustrated in Figure 4d.
The notable improvement observed with the ensemble model compared to individually modified CNN architectures can be attributed to its ability to combine the strengths of individual architectures while mitigating their weaknesses. By leveraging the diversity among models generated through different initialization and training strategies, the ensemble model achieved robust predictions. Additionally, the incorporation of cross-validation helped generalize the model, reducing overfitting and enhancing its performance on unseen data. These results underscore the efficacy of ensemble learning in enhancing classification accuracy, particularly in the context of medical image analysis, where subtle variations in image features can significantly impact diagnosis. Several studies have implemented machine learning algorithms for the BACH dataset for breast cancer histology image classification, achieving varying levels of accuracy. For instance, the top submissions in the BACH 2018 Grand Challenge used convolutional neural networks (CNNs) and reported an accuracy of 87% [8]. Another study by Song et al. (2020) applied a DenseNet model combined with data augmentation techniques and reported an accuracy of 88.7% [35]. Similarly, Araujo et al. (2019) employed a CNN-based approach, achieving an accuracy of 85.6% [36]. These studies highlight the effectiveness of deep learning methods in improving the accuracy of breast cancer histology classification using the BACH dataset.
On the BreakHis dataset, the ensemble model demonstrated superior performance compared to state-of-the-art algorithms in terms of accuracy and Jaccard index for five-fold cross-validation. Table 2 presents a comparative analysis of various machine learning models, including EfficientNet b0, VGGNet 16, ResNet 34, ResNet 50, and an ensemble model. These models were evaluated based on their training and validation performance metrics such as accuracy, Jaccard index, and loss. Remarkably, the ensemble model outperformed all other models, with the highest average validation accuracy of 98.43%, indicating its superior predictive capability on unseen data. Various studies have implemented machine learning algorithms for the BreaKHis dataset, showcasing a range of accuracies. Simonyan, K et al. proposed a three-path ensemble architecture combining VGG19, MobileNetV2, and DenseNet201, which resulted in a 97% classification accuracy [14]. Additionally, Howard et al. proposed a stacked ensemble learning framework and achieved 94.35% accuracy, while a traditional SVM-based method reported an 81.2% accuracy [37]. Also, the recent work presented in İrem Sayın et al.’s comparison study [38] achieved an accuracy of 89% using the Xception model. Both the Inception and InceptionResNet models achieved an accuracy of 87%. These results demonstrate the effectiveness of different machine learning models on the BreaKHis dataset, highlighting significant advancements in breast cancer classification accuracy.
Table 3 presents an analysis of the Ensemble model designed for the classification of breast histopathological images from the BreakHis dataset into benign and malignant categories. This model achieved an exceptional average validation accuracy of 99.72%, significantly surpassing the performance of all prior models referenced in Table 4. This can be seen with BreastNet [18], which achieved an overall accuracy of 97.56%. Similarly, DenseNet121-AnoGAN [19], an unsupervised anomaly detection model using generative adversarial networks, recorded an accuracy of 91.44%. Transfer learning applications on the BreakHis dataset with ResNet-18 [20] and Inception-V3 [21] demonstrated accuracies of 92.15% and 98.97%, respectively. Additionally, ref. [22] proposed an enhanced autoencoder network utilizing a Siamese framework, achieving an accuracy of 96.97%. DRDA-Net [23], a model featuring dual-shuffle attention mechanisms, attained an accuracy of 96.10%, and DenTnet obtained an accuracy of 99.28%. Figure 5 shows the ROC curve of the ensemble model for binary classification on the BreakHis dataset. The ROC curve shows an excellent performance of the proposed model for classifying benign and malignant blocks.
The success of the ensemble model on the BreakHis dataset can be attributed to its capacity to capture intricate patterns and features present in histopathology images, which are crucial for accurate diagnosis. Furthermore, the ensemble approach helps alleviate the challenges posed by dataset heterogeneity and variability, leading to robust and reliable predictions across diverse tissue samples. These findings highlight the potential of ensemble learning to address the complexities inherent in medical image classification tasks, thereby facilitating more accurate and timely diagnoses.
Overall, our results demonstrate the effectiveness of deep learning techniques, particularly ensemble learning, in accurately classifying breast cancer histopathology images. These findings underscore the potential of artificial intelligence to enhance breast cancer diagnosis and management, paving the way for improved patient outcomes and reduced healthcare burdens. The training and validation were conducted using an Nvidia Tesla V100 graphics card. The ensemble learning architecture was trained over 20 epochs, with each epoch comprising 48 steps. On average, each step of training took approximately 16 s. Additionally, our study highlights the importance of exploring novel approaches and ensemble strategies in medical image analysis, offering promising avenues for future research and clinical applications.

4. Discussion

The results obtained from our comprehensive study on breast cancer histopathology image classification underscore the pivotal role of advanced computational techniques, particularly deep learning algorithms, in improving diagnostic precision and efficiency. Various aspects, including computational efficiency, algorithmic advancements, and future directions for leveraging these methodologies in clinical practice and beyond, were discussed.

4.1. Computational Efficiency

One of the key challenges in histopathology image analysis lies in handling large-scale datasets with high-resolution images, which often pose computational burdens. To address this challenge, innovative techniques such as image patching, downsampling, and ensemble learning were employed. The approach of segmenting high-resolution images into smaller patches enabled focused analysis of localized regions, reducing computational demands while preserving diagnostic accuracy. Additionally, pre-trained convolutional neural network (CNN) models like VGG16 and ResNet50 were leveraged to harness learned features, thereby optimizing computational efficiency. Furthermore, the ensemble learning strategy, which combined predictions from multiple models, not only enhanced classification accuracy but also mitigated computational costs by leveraging existing model architectures.

4.2. Algorithmic Advancements

This study contributes to the ongoing advancements in algorithmic design for breast cancer histopathology image classification. By exploring diverse CNN architectures and ensemble learning techniques, the efficacy of leveraging deep learning for accurate classification across multiple datasets was demonstrated. The utilization of pre-trained models, fine-tuned strategies, and ensemble approaches enabled the extraction of intricate features from histopathology images, leading to enhanced classification performance. Moreover, the investigation into multiscale analysis, particularly for the BreakHis dataset, highlighted the importance of integrating information from different magnification levels to capture diverse pathological structures and improve diagnostic accuracy. This approach enhanced the performance of image classification by integrating information at different magnifications, thereby improving the recognition of various pathological structures that are more discernible at certain scales. It also increased the efficiency and accuracy of diagnosis by reducing operator variability. Moreover, multiscale analysis can assist in understanding disease progression for the development of targeted therapeutics. Automated systems that employ this technique can reduce the time required for image analysis. Consequently, multiscale magnification and can offer a comprehensive view of histopathological images and facilitate a more precise and efficient disease diagnosis. These algorithmic advancements also hold promise for accelerating the research and development of personalized treatment strategies.

4.3. Future Directions

Looking ahead, several avenues warrant exploration to further enhance the utility of deep learning in breast cancer diagnosis and management. Firstly, the development of user-friendly web applications incorporating our classification algorithms could empower healthcare professionals with accessible tools for rapid and accurate histopathology image analysis. Promising studies that validate the efficacy of such AI-aided tools would be preferred to create viable applied solutions to enhance healthcare workflows [41,42]. Such applications could streamline clinical workflows, facilitate remote consultations, and improve patient care outcomes. Additionally, extending our algorithmic framework to encompass other types of cancer, such as lung cancer or prostate cancer, may hold promise for broadening the scope of AI-driven diagnostic solutions. Additional imaging modalities could also be leveraged to encompass end-to-end care paths aided by AI solutions, as explained by the authors of a review study [43]. By leveraging transfer learning and dataset adaptation techniques, our methodologies could be adapted to address diverse oncological challenges, thereby advancing cancer research and clinical practice. Furthermore, integrating multimodal data, such as genomic profiles or clinical metadata, into our classification models could enrich diagnostic insights and enable more comprehensive patient stratification.

5. Conclusions

In conclusion, our research represents a significant stride towards leveraging advanced computational techniques, particularly deep learning algorithms, to enhance breast cancer diagnosis and management through histopathology image analysis. By focusing on two prominent datasets, BACH and BreakHis, and employing innovative methodologies, a substantial contribution is made to the growing body of knowledge aimed at improving diagnostic precision and efficiency in breast cancer detection. Through the meticulous experimentation and implementation of deep learning models, including VGG16 and ResNet50, the efficacy of this approach in accurately classifying breast cancer histopathology images was demonstrated. The proposed ensemble learning strategy, which combines predictions from multiple models, showcased superior performance compared to state-of-the-art algorithms, underscoring the importance of leveraging the complementary strengths of different CNN architectures.
Furthermore, our research addressed the computational challenges inherent in histopathology image analysis by introducing techniques such as image patching and multiscale analysis. These strategies enhance computational efficiency and facilitate focused analysis on specific regions of interest within the histopathology images, improving classification accuracy.

Author Contributions

Conceptualization, S.M.A.A.-H. and S.A.; Methodology, A.B., S.M.A.A.-H. and S.A.; Software, A.A.B., S.M.A.A.-H. and S.A.; Validation, A.A.B., S.M.A.A.-H., A.B., B.A., R.C., S.T.R. and S.A.; Formal analysis, S.A.; Investigation, S.M.A.A.-H., A.S., A.B., B.A., R.C., S.T.R. and S.A.; Writing—original draft, A.A.B.; Writing—review & editing, S.M.A.A.-H., A.S. and S.A.; Visualization, S.A.; Supervision, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The data presented in this study are available in this article.

Acknowledgments

We would like to express our deepest gratitude to the Roux Institute, the IEAI, and the Alfond Foundation for their invaluable support and contributions. Additionally, we extend our sincere appreciation to Isha Hemant Arora for her valuable contributions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Wagle, N.S.; Jemal, A. Cancer statistics, 2023. CA Cancer J. Clin. 2023, 73, 17–48. [Google Scholar] [CrossRef] [PubMed]
  2. American Cancer Society. Cancer Facts & Figures 2024; American Cancer Society: New York, NY, USA, 2024. [Google Scholar]
  3. Alyami, W.; Kyme, A.; Bourne, R. Histological validation of MRI: A review of challenges in registration of imaging and whole-mount histopathology. J. Magn. Reson. Imaging 2022, 55, 11–22. [Google Scholar] [CrossRef] [PubMed]
  4. Çinar, U. Integrating Hyperspectral Imaging and Microscopy for Hepatocellular Carcinoma Detection from H&E Stained Histopathology Images. Ph.D. Thesis, Middle East Technical University, Ankara, Turkey, 2023. [Google Scholar]
  5. Panayides, A.S.; Amini, A.; Filipovic, N.D.; Sharma, A.; Tsaftaris, S.A.; Young, A.; Foran, D.; Do, N.; Golemati, S.; Kurc, T.; et al. AI in medical imaging informatics: Current challenges and future directions. IEEE J. Biomed. Health Inform. 2020, 24, 1837–1857. [Google Scholar] [CrossRef] [PubMed]
  6. Brancati, N.; Frucci, M.; Riccio, D. Multi-classification of breast cancer histology images by using a fine-tuning strategy. In Proceedings of the Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 771–778. [Google Scholar]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  8. Aresta, G.; Araújo, T.; Kwok, S.; Chennamsetty, S.S.; Safwan, M.; Alex, V.; Marami, B.; Prastawa, M.; Chan, M.; Donovan, M. BACH: Grand challenge on breast cancer histology images. Med. Image Anal. 2019, 56, 122–139. [Google Scholar] [CrossRef] [PubMed]
  9. Chennamsetty, S.S.; Safwan, M.; Alex, V. Classification of breast cancer histology image using ensemble of pre-trained neural networks. In Proceedings of the Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 804–811. [Google Scholar]
  10. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  11. Zhu, C.; Song, F.; Wang, Y.; Dong, H.; Guo, Y.; Liu, J. Breast cancer histopathology image classification through assembling multiple compact CNNs. BMC Med. Inform. Decis. Mak. 2019, 19, 1–17. [Google Scholar] [CrossRef] [PubMed]
  12. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A dataset for breast cancer histopathological image classification. IEEE Trans. Biomed. Eng. 2015, 63, 1455–1462. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, Y.; Sun, L.; Ma, K.; Fang, J. Breast cancer microscope image classification based on CNN with image deformation. In Proceedings of the Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 845–852. [Google Scholar]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Yan, R.; Ren, F.; Wang, Z.; Wang, L.; Zhang, T.; Liu, Y.; Rao, X.; Zheng, C.; Zhang, F. Breast cancer histopathological image classification using a hybrid deep neural network. Methods 2020, 173, 52–60. [Google Scholar] [CrossRef] [PubMed]
  16. Bagchi, A.; Pramanik, P.; Sarkar, R. A multi-stage approach to breast cancer classification using histopathology images. Diagnostics 2022, 13, 126. [Google Scholar] [CrossRef]
  17. Wakili, M.A.; Shehu, H.A.; Sharif, M.H.; Sharif, M.H.U.; Umar, A.; Kusetogullari, H.; Ince, I.F.; Uyaver, S. Classification of breast cancer histopathological images using DenseNet and transfer learning. Comput. Intell. Neurosci. 2022, 2022, 8904768. [Google Scholar] [CrossRef]
  18. Toğaçar, M.; Özkurt, K.B.; Ergen, B.; Cömert, Z. BreastNet: A novel convolutional neural network model through histopathological images for the diagnosis of breast cancer. Phys. A Stat. Mech. Its Appl. 2020, 545, 123592. [Google Scholar] [CrossRef]
  19. Man, R.; Yang, P.; Xu, B. Classification of breast cancer histopathological images using discriminative patches screened by generative adversarial networks. IEEE Access 2020, 8, 155362–155377. [Google Scholar] [CrossRef]
  20. Boumaraf, S.; Liu, X.; Zheng, Z.; Ma, X.; Ferkous, C. A new transfer learning based approach to magnification dependent and independent classification of breast cancer in histopathological images. Biomed. Signal Process. Control. 2021, 63, 102192. [Google Scholar] [CrossRef]
  21. Soumik, M.F.I.; Aziz, A.Z.B.; Hossain, M.A. Improved transfer learning based deep learning model for breast cancer histopathological image classification. In Proceedings of the 2021 International Conference on Automation, Control and Mechatronics for Industry 4.0 (ACMI), Rajshahi, Bangladesh, 8–9 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–4. [Google Scholar]
  22. Liu, M.; He, Y.; Wu, M.; Zeng, C. Breast histopathological image classification method based on autoencoder and siamese framework. Information 2022, 13, 107. [Google Scholar] [CrossRef]
  23. Chattopadhyay, S.; Dey, A.; Singh, P.K.; Sarkar, R. DRDA-Net: Dense residual dual-shuffle attention network for breast cancer classification using histopathological images. Comput. Biol. Med. 2022, 145, 105437. [Google Scholar] [CrossRef]
  24. Koné, I.; Boulmane, L. Hierarchical ResNeXt models for breast cancer histology image classification. In Proceedings of the Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 796–803. [Google Scholar]
  25. Ray, R.K.; Linkon, A.A.; Bhuiyan, M.S.; Jewel, R.M.; Anjum, N.; Ghosh, B.P.; Mia, M.T.; Badruddowza; Sarker, M.S.U.; Shaima, M. Transforming Breast Cancer Identification: An In-Depth Examination of Advanced Machine Learning Models Applied to Histopathological Images. J. Comput. Sci. Technol. Stud. 2024, 6, 155–161. [Google Scholar]
  26. Addo, D.; Zhou, S.; Sarpong, K.; Nartey, O.T.; Abdullah, M.A.; Ukwuoma, C.C.; Al-antari, M.A. A hybrid lightweight breast cancer classification framework using the histopathological images. Biocybern. Biomed. Eng. 2024, 44, 31–54. [Google Scholar] [CrossRef]
  27. Sahran, S.; Qasem, A.; Omar, K.; Albashih, D.; Adam, A.; Abdullah, S.N.H.S.; Abdullah, A.; Hussain, R.I.; Ismail, F.; Abdullah, N.; et al. Machine learning methods for breast cancer diagnostic. In Breast Cancer Surgery; IntechOpen: London, UK, 2018; pp. 57–76. [Google Scholar]
  28. Alirezazadeh, P.; Dornaika, F.; Moujahid, A. Chasing a Better Decision Margin for Discriminative Histopathological Breast Cancer Image Classification. Electronics 2023, 12, 4356. [Google Scholar] [CrossRef]
  29. Bayramoglu, N.; Kannala, J.; Heikkilä, J. Deep learning for magnification independent breast cancer histopathology image classification. In Proceedings of the 2016 23rd International conference on pattern recognition (ICPR), Cancun, Mexico, 4–8 December 2016; IEEE: Piscataway, NJ, USA; pp. 2440–2445. [Google Scholar]
  30. Gour, M.; Jain, S.; Sunil Kumar, T. Residual learning based CNN for breast cancer histopathological image classification. Int. J. Imaging Syst. Technol. 2020, 30, 621–635. [Google Scholar] [CrossRef]
  31. Titoriya, A.; Sachdeva, S. Breast cancer histopathology image classification using AlexNet. In Proceedings of the 2019 4th International Conference on Information Systems and Computer Networks (ISCON), Mathura, UP, India, 21–22 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 708–712. [Google Scholar]
  32. Deniz, E.; Şengür, A.; Kadiroğlu, Z.; Guo, Y.; Bajaj, V.; Budak, Ü. Transfer learning based histopathologic image classification for breast cancer detection. Health Inf. Sci. Syst. 2018, 6, 1–7. [Google Scholar] [CrossRef]
  33. Vesal, S.; Ravikumar, N.; Davari, A.; Ellmann, S.; Maier, A. Classification of breast cancer histology images using transfer learning. In Proceedings of the Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 812–819. [Google Scholar]
  34. Ahmad, N.; Asghar, S.; Gillani, S.A. Transfer learning-assisted multi-resolution breast cancer histopathological images classification. Vis. Comput. 2022, 38, 2751–2770. [Google Scholar] [CrossRef]
  35. Song, Y.; Zhang, L.; Chen, S.; Ni, D.; Lei, B.; Wang, T. Accurate automated detection of invasive ductal carcinoma in whole-slide images of breast cancer tissues using deep learning. BMC Bioinform. 2020, 21, 1–12. [Google Scholar]
  36. Araujo, T.; Aresta, G.; Castro, E.; Rouco, J.; Aguiar, P.; Eloy, C.; Polonia, A.; Campilho, A. Classification of breast cancer histology images using Convolutional Neural Networks. PLoS ONE 2019, 14, e0218677. [Google Scholar] [CrossRef] [PubMed]
  37. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  38. Sayın, İ.; Soydaş, M.A.; Mert, Y.E.; Yarkadaş, A.; Ergün, B.; Sözen Yeh, S.; Üvet, H. Comparative Analysis of Deep Learning Architectures for Breast Cancer Diagnosis Using the BreaKHis Dataset. arXiv, 2023; arXiv:2309.01007. [Google Scholar]
  39. Parvin, F.; Hasan, M.A.M. A comparative study of different types of convolutional neural networks for breast cancer histopathological image classification. In Proceedings of the 2020 IEEE Region 10 Symposium (TENSYMP), Dhaka, Bangladesh, 5–7 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 945–948. [Google Scholar]
  40. Zerouaoui, H.; Idri, A. Deep hybrid architectures for binary classification of medical breast cancer images. Biomed. Signal Process. Control. 2022, 71, 103226. [Google Scholar] [CrossRef]
  41. Singh, A.; Randive, S.; Breggia, A.; Ahmad, B.; Christman, R.; Amal, S. Enhancing Prostate Cancer Diagnosis with a Novel Artificial Intelligence-Based Web Application: Synergizing Deep Learning Models, Multimodal Data, and Insights from Usability Study with Pathologists. Cancers 2023, 15, 5659. [Google Scholar] [CrossRef]
  42. Singh, A.; Wan, M.; Harrison, L.; Breggia, A.; Christman, R.; Winslow, R.L.; Amal, S. Visualizing Decisions and Analytics of Artificial Intelligence based Cancer Diagnosis and Grading of Specimen Digitized Biopsy: Case Study for Prostate Cancer. In Proceedings of the 28th International Conference on Intelligent User Interfaces, New York, NY, USA, 27–31 March 2023; pp. 166–170. [Google Scholar]
  43. Milosevic, M.; Jin, Q.; Singh, A.; Amal, S. Applications of AI in multi-modal imaging for cardiovascular disease. Front. Radiol. 2024, 3, 1294068. [Google Scholar] [CrossRef]
Figure 1. Exemplary microscopy images demonstrating the four classes in the BACH dataset (on 40× magnification).
Figure 1. Exemplary microscopy images demonstrating the four classes in the BACH dataset (on 40× magnification).
Cancers 16 02222 g001
Figure 2. Representative microscopy images of malignant and benign breast tissues from the BreakHis dataset (on 40× magnification).
Figure 2. Representative microscopy images of malignant and benign breast tissues from the BreakHis dataset (on 40× magnification).
Cancers 16 02222 g002
Figure 3. Schematic of ensemble model.
Figure 3. Schematic of ensemble model.
Cancers 16 02222 g003
Figure 4. The average accuracy on 5-fold cross-validation on BACH dataset for the following architectures: (a) VGG 16 model, (b) ResNet 50 model, (c) ResNet 34 model, and (d) ensemble model.
Figure 4. The average accuracy on 5-fold cross-validation on BACH dataset for the following architectures: (a) VGG 16 model, (b) ResNet 50 model, (c) ResNet 34 model, and (d) ensemble model.
Cancers 16 02222 g004
Figure 5. ROC curve of the ensemble model for five-fold cross-validation.
Figure 5. ROC curve of the ensemble model for five-fold cross-validation.
Cancers 16 02222 g005
Table 1. Comparative performance analysis of machine learning models for breast cancer diagnosis on BACH dataset.
Table 1. Comparative performance analysis of machine learning models for breast cancer diagnosis on BACH dataset.
ModelAverage Train AccuracyAverage Validation AccuracyTrain LossAverage Validation Loss
VGGNet 1698.7279.950.09660.7399
ResNet 5099.0580.470.05640.4489
Ensemble Model (5-fold cross-validation)99.8492.580.05910.2793
Table 2. Comparative performance analysis of machine learning models for breast cancer diagnosis on BreakHis dataset.
Table 2. Comparative performance analysis of machine learning models for breast cancer diagnosis on BreakHis dataset.
ModelTrain AccuracyTrain Jaccard IndexVal AccuracyVal Jaccard IndexAverage Accuracy
EfficientNet B079.694020.67309784.956380.7573584.95
VGGNet 1696.911750.93752497.382730.95530997.38
ResNet 3497.758880.96284998.224810.9694598.22
ResNet 5097.47440.95703897.304340.95392197.30
Ensemble Models (ours)98.419520.97319398.429640.97364298.43
Table 3. Binary classification performance of ensemble models on BreakHis dataset.
Table 3. Binary classification performance of ensemble models on BreakHis dataset.
ModelTrain AccuracyTrain Jaccard IndexVal AccuracyVal Jaccard IndexAverage Accuracy
Ensemble Models (ours)99.768870.99561699.719310.99453299.71930712
Table 4. Binary classification performance on BreakHis dataset.
Table 4. Binary classification performance on BreakHis dataset.
YearMethodAccuracy
2020Togacar et al. [18]97.56
Parvin et al. [39]91.25
Man et al. [19]91.44
2021Boumaraf et al. [20]92.15
Soumik et al. [21]98.97
2022Liu et al. [22]96.97
Zerouaoui and Idri [40]93.85
Chattopadhyay et al. [23]96.10
DenTnet [17]99.28
2024Ensemble Models (ours)99.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Balasubramanian, A.A.; Al-Heejawi, S.M.A.; Singh, A.; Breggia, A.; Ahmad, B.; Christman, R.; Ryan, S.T.; Amal, S. Ensemble Deep Learning-Based Image Classification for Breast Cancer Subtype and Invasiveness Diagnosis from Whole Slide Image Histopathology. Cancers 2024, 16, 2222. https://doi.org/10.3390/cancers16122222

AMA Style

Balasubramanian AA, Al-Heejawi SMA, Singh A, Breggia A, Ahmad B, Christman R, Ryan ST, Amal S. Ensemble Deep Learning-Based Image Classification for Breast Cancer Subtype and Invasiveness Diagnosis from Whole Slide Image Histopathology. Cancers. 2024; 16(12):2222. https://doi.org/10.3390/cancers16122222

Chicago/Turabian Style

Balasubramanian, Aadhi Aadhavan, Salah Mohammed Awad Al-Heejawi, Akarsh Singh, Anne Breggia, Bilal Ahmad, Robert Christman, Stephen T. Ryan, and Saeed Amal. 2024. "Ensemble Deep Learning-Based Image Classification for Breast Cancer Subtype and Invasiveness Diagnosis from Whole Slide Image Histopathology" Cancers 16, no. 12: 2222. https://doi.org/10.3390/cancers16122222

APA Style

Balasubramanian, A. A., Al-Heejawi, S. M. A., Singh, A., Breggia, A., Ahmad, B., Christman, R., Ryan, S. T., & Amal, S. (2024). Ensemble Deep Learning-Based Image Classification for Breast Cancer Subtype and Invasiveness Diagnosis from Whole Slide Image Histopathology. Cancers, 16(12), 2222. https://doi.org/10.3390/cancers16122222

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop