Next Article in Journal
The Effect of Magnetic Resonance Imaging Based Radiomics Models in Discriminating stage I–II and III–IVa Nasopharyngeal Carcinoma
Next Article in Special Issue
5G-Based Telerobotic Ultrasound System Improves Access to Breast Examination in Rural and Remote Areas: A Prospective and Two-Scenario Study
Previous Article in Journal
Styloid Jugular Nutcracker: The Possible Role of the Styloid Process Spatial Orientation—A Preliminary Morphometric Computed Study
Previous Article in Special Issue
Immunohistochemical HER2 Recognition and Analysis of Breast Cancer Based on Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Separable Convolution Network for Breast Cancer Histopathological Identification

by
Grace Ugochi Nneji
1,2,*,
Happy Nkanta Monday
1,2,*,
Goodness Temofe Mgbejime
3,
Venkat Subramanyam R. Pathapati
4,
Saifun Nahar
5 and
Chiagoziem Chima Ukwuoma
6
1
Department of Computing, Oxford Brookes College of Chengdu University of Technology, Chengdu 610059, China
2
Deep Learning and Intelligent Computing Lab, HACE SOFTTECH, Lagos 102241, Nigeria
3
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
4
Department of Cybersecurity, University of Missouri St. Louis, St. Louis, MO 63121, USA
5
Department of Information System and Technology, University of Missouri St. Louis, St. Louis, MO 63121, USA
6
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(2), 299; https://doi.org/10.3390/diagnostics13020299
Submission received: 30 November 2022 / Revised: 8 January 2023 / Accepted: 10 January 2023 / Published: 13 January 2023
(This article belongs to the Special Issue Breast Cancer Imaging: Successes and Challenges)

Abstract

:
Breast cancer is one of the leading causes of death among women worldwide. Histopathological images have proven to be a reliable way to find out if someone has breast cancer over time, however, it could be time consuming and require much resources when observed physically. In order to lessen the burden on the pathologists and save lives, there is need for an automated system to effectively analysis and predict the disease diagnostic. In this paper, a lightweight separable convolution network (LWSC) is proposed to automatically learn and classify breast cancer from histopathological images. The proposed architecture aims to treat the problem of low quality by extracting the visual trainable features of the histopathological image using a contrast enhancement algorithm. LWSC model implements separable convolution layers stacked in parallel with multiple filters of different sizes in order to obtain wider receptive fields. Additionally, the factorization and the utilization of bottleneck convolution layers to reduce model dimension were introduced. These methods reduce the number of trainable parameters as well as the computational cost sufficiently with greater non-linear expressive capacity than plain convolutional networks. The evaluation results depict that the proposed LWSC model performs optimally, obtaining 97.23% accuracy, 97.71% sensitivity, and 97.93% specificity on multi-class categories. Compared with other models, the proposed LWSC obtains comparable performance.

1. Introduction

Majority of women are affected by breast cancer, which is one of the deadliest and most common type of cancer. In the world of today, about 1 in 8 women maybe diagnosed of breast cancer [1]. In order to fight the disease, it is essential to predict breast cancer risk, and there exist two kinds of breast cancer risk [1]. According to the first category, a person may develop breast cancer within a certain time frame [1]. The likelihood of a high-risk gene mutation is implied by the second type [2]. Breast tumors are unnatural expansion of breast tissue that might show themselves as discharge from a lump or nipple or as a change in the skin’s texture near the nipple. Cancers are uncontrolled cell growth that can spread to other parts of the body through the lymphatic and circulatory systems [3]. After lung cancer, breast cancer is the second most common kind of cancer and the leading cause of cancer mortality in women [3]. The condition has gained in notoriety over the past 50 years, and its prevalence has risen recently.
There are currently few standards for breast cancer screening. The Iranian preventive services working group suggests screening for females between the ages of 45 and 70 but makes no firm recommendations for females older than this. Breast cancer risk assessment can be used to both motivate high-risk women who have not yet had screening and to ensure that those who would not otherwise comply with screening requirements do so. Building cancer prevention and risk reduction strategies can also make use of a statistical model that predicts the risk of breast cancer [4]. The contribution of this study is in two-folds. The first-fold handles the problem of low quality attribute of the image by extracting the visual trainable features of the histopathological image using a contrast enhancement and edge detector algorithms. The second-fold implements a deep learning model of separable convolution layers stacked in parallel with multiple filters of different sizes in order to obtain wider receptive fields. Additionally, the factorization and the utilization of bottleneck convolution layers to reduce model dimension are introduced. This proposed model aim to reduce the number of trainable parameters as well as the computational cost sufficiently with greater non-linear expressive capacity than plain convolutional networks.
This paper will be fashioned with the following sections. Section 2 comprises the literature review of the breast cancer diagnosis; Section 3 details the proposed method and the dataset that will be used to validate the framework. Section 4 presents the experimental findings and model evaluation, and the conclusion is written in Section 5.

2. Literature Review

In several earlier researches, the Gail model [5], which is a statistical model that calculates the probability of developing breast cancer has been used. The model integrated different breast cancer risk rates, and different inputs are used to forecast a woman’s risk of acquiring breast cancer using logistics regression [5]. The researchers in [6] utilized the Breast Cancer Risk Assessment Tool also known as BCRAT with the six traditional Gail model inputs plus a typical hyperplasia personal data history for predicting the risk of breast cancer, however this approach has been shown to be ineffective in several populations. Another demonstration by Hart et al. [7] was examined using machine learning algorithm which is capable of predicting the risk of breast cancer and achieved increased prediction accuracy.
Belsare et al. [8] describes a computer-aided concept system that uses a space-color partition of graphical illustrations, tissue property extraction using tools like the gray-level co-occurrence matrix (GLCM), grassland-based ruminant livestock models (GRLM), and Euler methods, and classification using linear discriminative analysis. With 100% accuracy, the algorithm classified 70 histopathology and mammography pictures. Vu et al. [9] employed a feature-oriented vocabulary learning system used on a data set of human intraocular lesions and animal diagnostic laboratories to reach an accuracy of 97.75%. It has been demonstrated that identifying histopathological and mammographic data sets from 70 photos using a computer-aided design technique based on the extraction of morphological characteristics is 85.7% accurate [10]. Mouelhi et al. [11] employed enhanced morphology and adaptive local thresholding algorithms to extract and segment mammography and histochemistry pictures and achieved 98% accuracy.
Khalilabad et al. [12] developed an automated system to analyze how to identify breast cancer masses using micro array pictures into three-fold of preprocessing of images, data mining and detection of the disease and achieved accuracy of 95.45%. Kaymak et al. [13] proposed back propagation neural network as a tool for analyzing and diagnosing breast cancer masses and further improved the technique with radial basis neural network and achieved 59% for the former and 70.4% for the latter. An evolutionary state of the art decision-making approach based on regression and evolutionary approaches has been utilized by Wang et al. [14] to identify breast cancer from mammograms. Mohebian et al. [15] presents a computer-aided diagnostic method for the prediction of breast cancer recurrence that uses optimal ensemble learning or a hybrid approach.
However, researchers highly respect deep learning-based feature extraction techniques for categorizing mammography and histopathology pictures. Also, it is said to outperformed older feature-engineered histopathology analysis techniques in spotting breast cancer metastases in lymph nodes, displaying extremely skilled performance [16]. Wang et al. [17] looked at the usage of an extreme learning machine using deep convolution features for breast cancer diagnosis and classification and achieved remarkable performance of 98.18% classification accuracy. Kumar et al. [18] introduced canine mammary tumors dataset and employed a framework based on VGGNet with different classifiers and achieved mean accuracy of 93%. A support vector machine approach and a deep neural network were combined by Kaur et al. [19] with the goal of categorizing mammographic pictures to find cancerous tumors. This method’s accuracy ranged from 94% to 98%, depending on the different data sets.
Ting et al. [20] proposed a convolutional neural network, CNN-based algorithm for the improvement of breast cancer classification and achieved sensitivity, accuracy and specificity of 89.47%, 90.50% and 90.71% respectively. Li et al. [21] employed an effective and accurate classification of benign and malignant mammography images using an improved DenseNet II model achieving average accuracy of 94.55%. Shen et al. [22] developed a deep neural network algorithm to accurately detect breast cancer on screening mammograms achieving sensitivity and specificity of 86.7% and 96.1% respectively. Saha et al. [23] employed semantic segmentation and classification of breast cancer masses. In this technique, cell membranes and nuclei in the breast region are segmented and categorized using deep learning algorithms such as the human epidermal growth factor receptor-2 deep neural network (Her2Net) and trapezoidal long short-term memory (TLSTM) and achieved accuracy rate of 98.3%.
Rustam et al. [24] used different evaluation metrics like accuracy, sensitivity, specificity, and F1-score to compare linear discriminant analysis with support vector machine (SVM). The outcome demonstrates that the DVM outperforms the linear discriminant analysis in terms of overall performance with 98.8% accuracy. Khan et al. [25] utilized different pre-trained CNNs for the low level features which are fed into a fully connected layer using average pooling for the recognition of malignant and benign cells. Although several deep learning models have been proposed for breast cancer classification and identification, the computational cost and model complexity that may affect the efficacy of breast cancer classification have not been explored thoroughly. Spanhol et al. [26] analyzed six different feature descriptors with several classifiers and achieved 80% to 85% within the magnification factors. In similarity to Spanhol et al. [26], Bayramoglu et al. [27] used an independent magnification factors for the classification of breast cancer and achieved good result. Spanhol et al. [28] presented random-patches of the images for training and testing and achieved an increment of 84% to 91%. Han et al. [29] explored different deep learning models and achieved an average accuracy of 93.2% for patient-level BC classification. Alom et al. [30] proposed an adaptive sparse support vector machine with L1 weighted norm achieving accuracy of 94.97% for 40x magnification factor. A lightweight convolution network is suggested for the identification of breast cancer in line with this perspective.
This paper focuses on the downside of model complexity and computational cost. The novelty of the proposed model is the parallel stacking of separable convolution layers with multiple filters of different sizes with bottleneck convolution layers to shrink feature maps as a technique to reduce dimensionality. The public dataset belonging to Kaggle is used to evaluate the performance of this research.

3. Materials and Methods

Data acquisition, data pre-processing, and network training and testing are the phases of the proposed technique. Each step of the proposed approach is detailed in the subsequent headings.

3.1. Datasets

BreaKHis dataset is an open-source dataset obtained from Kaggle repository to trained the proposed lightweight convolution network. Figure 1 shows both the benign and malignant BreakHis images with different magnifying factors. The BreakHis dataset consists of microscopic histopathological images of breast cancer composed of over 2400 benign and 5400 malignant samples obtained from over 80 patients utilizing different magnifying factors (40×, 100×, 200×, and 400×). The image is 3-channel RGB of 8-bit depth with the resolution of 700 × 460 pixels in PNG format [26]. Table 1 shows the dataset distribution based on the binary class and magnifying factors while Table 2 shows the multi-class description of the BreakHis dataset.

3.2. Data Pre-Processing

Data pre-processing is a technique often used by deep learning practitioners to enhance the visual characteristics of images. This article adopts contrast enhancement and edge detection techniques in order to enhance the visual trainable features of the histopathological breast cancer images as presented in Figure 2.

Contrast Enhancement (CE) and Edge Detection (ED) Images

This paper employed both contrast enhancement [31] and edge detection [32] pre-processing techniques for a better enhancement of the images. First, the application of contrast enhancement makes it more realistic in appearance amongst its histogram equalisation-based member as displayed in Figure 2. In contrast limited histogram equalization, the histogram is cut at some threshold and then equalization is applied. The contrast of an image is enhanced by applying contrast algorithm on small regions called tiles rather than the entire image. The resulting neighboring tiles are then stitched back seamlessly using bilateral interpolation. The contrast in the homogeneous region can be limited so that noise amplification can be avoided. Uniform distribution is used as the basis for creating the contrast transform function. the expression of the modified chrominance channel tile with uniform distribution is given in Equation (1)
I c _ o u t = [ I c _ m a x I c _ m i n ] K j ( I c _ i n ) + I c _ m i n
where I c _ m i n and I c _ m a x are the minimum and maximum permissible intensity levels respectively. K j ( I c _ i n ) is the cumulative distribution function for input contextual tile I c _ i n .
The edge detection technique applies Gaussian smoothing for noise reduction while enhancing the morphological image contrast and from the determination of the intensity of the image, hysteresis thresholding is apply for the detection of the edges from the image as depicted in Figure 2. Additionally, less important regions are discarded so the visual characteristics can be detected. The image f is first smoothed with a Gaussian filter to reduce noise. This is done by using a convolution with a Gaussian kernel g to obtain an image z = f × g . The gradient of the image is calculated in terms of amplitude and angle as seen in Equations (2) and (3). Non-maxima are removed from the amplitude. This means that excessively large outlines are replaced by thinner outlines.
M = ( h x z ) 2 + ( h y z ) 2
A = ( h y z h x z )
where h x and h y represents the pixels in M which is the gradient of the image.

3.3. Lightweight Separable Convolution

The novelty behind the proposed method in this paper is in two phases. The first phase involves the implementation of separable convolution layers stacked in parallel with multiple filters of different sizes in order to obtain wider receptive fields as well as achieving wider rather than deeper network. The second phase involves the factorization of convolution layers and the utilization of bottleneck convolution layers to reduce model dimensionality. These methods sufficiently decrease the amount of trainable parameters as well as the computational cost, allowing the network to be developed much deeper and with greater non-linear expressive capacity than plain convolution networks. The proposed architecture consists of three parts, the ‘input head’, ‘separable block layer’, and ‘classification head’. The input head consists of two separable convolution layers followed by a max pool layer before a bottleneck convolution layer and then a separable convolution layer followed by a max pool layer. The separable block layer is divided into ‘Block A’ and ‘Block B’ stacked on top of one another. On the one hand, ‘Block A’ consists of separable convolution layers of 3 × 3 with bottleneck layers and factorization layers of 1 × 3 and 3 × 1 kernel sizes. On the other hand, ’Block B’ consists of separable convolution layers of 5 × 5 with bottleneck layers and factorization layers of 1 × 5 and 5 × 1 kernel sizes as presented in Figure 3. Each separable block has 10 convolution layers and 1 max pooling layer. The proposed method reduced computational cost and achieved reduction in feature dimensionality during the low-level feature extraction and overall network depth. In the classification head, the regular fully connected layer is replaced with average pooling of 8 × 8 to flatten the feature vector and finally, one dense layer having its dimension set as 1 × 512 is adopted as seen in Figure 3. To overcome over-fitting, dropout layers were added.

4. Experimental Results

This section presents the evaluation results of the proposed architecture both on binary and multi-class categories. The effect of data pre-processing on the performance of the proposed model is also detailed in this section. The standard metrics utilized to examine the diagnostic performance of the proposed LWSC are specificity (SPE), accuracy (ACC), precision (PRE), and sensitivity (SEN). The numerical expression for each metric is presented in Equations (4)–(8) [13,14].
A c c u r a c y = T P + T N T P + T N + F P + F N
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
F 1 - s c o r e = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l
TN denotes true negative, TP stands for true positive, FP depicts false positive and FN denotes false negative.

Evaluation of the Lightweight Separable Convolution

Extensive study was conducted to evaluate the influence of the proposed lightweight separable convolution network to the identification performance in terms of accuracy on histopathological breast cancer benchmark dataset known as BreakHis. The first study considered the original BreakHis images for the identification of histopathological breast cancer. The second study considered the edge enhanced BreakHis images for the identification of histopathological breast cancerand finally, the third study considered the contrast enhanced BreakHis images for the identification of histopathological breast cancer. The proposed LWSC model presented in Figure 3 clearly revealed that the multiple receptive fields is capable of handling low quality images in histopathological breast cancer identification obtaining better recognition performance on both binary and multi-class categories. Figure 4a shows the test accuracy curves for the different magnifying factors on the binary category indicating that the model obtains high accuracy of 93.12% on the 40 × magnifying factor of the contrast enhanced image while Figure 4b represents the test accuracy curves for the different magnifying factors on the multi-class category indicating that the model obtains high accuracy of 97.23% on the 40 × magnifying factor of the contrast enhanced image.
Table 3 illustrates the recognition accuracy for the original raw image and the pre-processed images for both contrast and edge enhancements. The LWSC + CE-based images denotes the contrast enhanced image with LWSC, the LWSC + ED-based images denotes the edge enhanced image with LWSC, and the LWSC + Original images denotes the raw original image with LWSC for the identification of histopathological breast cancer. From all indications, the proposed lightweight separable convolution model with contrast enhanced image outweighs both the original raw and the edge enhanced images achieving 93.12% accuracy on the binary class category of magnifying factor of 40 × and 97.23% accuracy on the eight class category of magnifying factor of 40 × . The proposed LWSC with CE outperforms the other combinations on all evaluation metrics.
The classification performance of LWSC with contrast enhanced image is higher than that of edge enhanced image which suggest that the contribution of contrast enhanced histopathological images in breast cancer identification is greater than that of edge enhanced histopathological images. The proposed LWSC is further evaluated in terms of SPE, SEN, PRE, and AUC on both categories of class labels with the magnifying factor of 40 × as depicted in Table 3. It is observed that the proposed LWSC performs better on multi-class category with the magnifying factor of 40 × achieving 97.23% accuracy, 97.71% sensitivity, 97.93% specificity, and 98.11% precision.

5. Discussion

The efficacy of the proposed method in identifying breast cancer DR in histopathological images with different magnifying factors on both binary and multi-class categories has been presented and the classification result is presented in Table 3. As denoted by the above mentioned results, the proposed LWSC can efficiently classify the different breast cancer types for the multi-class category. It is important to note that the proposed LWSC indicates better generalization ability with the contrast enhanced histopathological images with a commendable computational efficiency of 5.9 min training time. The proposed method is further compared with some up-to-date methods using BreakHis dataset and other benchmark dataset. Table 4 indicates that the proposed LWSC obtains satisfactory performance in sensitivity, specificity, and F1 score of 97.71%, 97.93%, and 97.98% respectively. Rustam et al. [24] achieved the highest accuracy value of 98.77%. Kaur et al. [19] obtained the highest AUC value of 99.0%.
The proposed LWSC obtains the highest accuracy score of 97.23% indicating the superiority of the proposed method for histopathological breast cancer identification. The competitive merit of the proposed model is attributed to the wider receptive fields from the different filter sizes. It is well known that different deep learning architecture will show different behaviours for different conditions. In order to select what number of separable blocks and its combination will produce the best result for the proposed lightweight separable convolution, ablation study is conducted.
Table 5 shows the experimental results obtained by the proposed architecture in comparison with various pre-trained networks on the BreakHis dataset using the same computing device. From the result, VGG-19 model shows better performance, achieving 96.4% accuracy, 97.1% sensitivity, and 96.0% specificity. Inception V3 model indicates significant improvement in performance, achieving 95.8% accuracy, 96.0% sensitivity, and 95.3% specificity.
In general, AlexNet gave the least performance across all evaluation metrics followed by Xception model. Considering a sensitive condition like histopathological breast cancer, it is imperative to adopt the ROC curve as a method to examine the total accuracy and the precision-recall curve to examine the average precision of the proposed lightweight separable convolution. The precision-recall curve is presented in Figure 5a while the ROC curve for the proposed LWSC model on binary category is presented in Figure 5b. Similarly, Figure 6a,b present the precision-recall curve and the ROC curve for the proposed LWSC model on multi-class category respectively.It is worthy to mention that the pre-trained models and the proposed LWSC model are trained using the same computing resource and dataset for fair comparison.
Additionally, some of the histopathological images were blurred which could have prevented the proposed LWSC model from training useful features. The advantage of improving the visual trainable features of the histopathological images using contrast enhancement and edge detection pre-processing techniques is to characterize distinctive representation features of the histopathological images with viable trainable details. The proposed LWSC obtained a significant performance in classifying histopathological breast cancer.
From all indications, the proposed LWSC outperforms the other networks in the perspective of precision-recall and ROC especially in handling low quality histopathological images. The precision-recall graphs shows that the curve of the proposed LWSC model is nearest to the upper right corner of the graph which implies that the proposed LWSC model has high precision associated with high sensitivity. Similarly, the ROC graphs depicts that the curve of the proposed model is nearest to the upper left corner of the graph which implies that it has high sensitivity associated with high specificity. Importantly, the obtained result presented by ROC and precision-recall curves can help expert histopathologist in maintaining a balance between accuracy and precision.
This study has achieved a significant degree of accuracy in classifying histopathological breast cancer however, there are certain setbacks. The level of accuracy obtained on histopathological breast cancer dataset might not be the same for another medical dataset. The singular reason is due to the fact that images of different dataset differ owing to different factors such as labeling, noise, image collection method, and location. Aside the non-uniformity of data, the partitioning of the data category is also paramount. The differences in class weight has a negative effect on training. The classification accuracy is also affected by the various data augmentation techniques adopted to correct class weight imbalance. In light of these constraints, studies will be carried out in the future to accommodate a wider range of dataset and possibly utilize different hyper-parameter tuning techniques.

6. Conclusions

This manuscript proposed a technique of identifying histopathological breast cancer using lightweight separable convolution neural network trained on BreakHis dataset. Image contrast enhancement and edge detection were implemented as pre-processing steps to extract visual trainable characteristics in order to achieve high classification accuracy. The proposed LWSC model implements a separable convolution layers stacked in parallel with multiple filters of different sizes in order to obtain wider receptive fields as well as achieving wider rather than deeper network. Factorization of convolution layers and the utilization of bottleneck convolution layers to reduce model dimensionality were introduced in the proposed LWSC model. The proposed work sufficiently decrease the amount of trainable parameters as well as the computational cost, allowing the network to be developed much deeper and with greater non-linear expressive capacity than plain convolution networks. The proposed LWSC model outperforms several state-of-the-art models. The evaluation results depict that the proposed LWSC model performs optimally obtaining 93.12% accuracy, 93.61% sensitivity and 94.07% specificity on binary category while on multi-class category, the proposed LWSC obtained 97.23% accuracy, 97.71% sensitivity, and 97.93% specificity. From the comparative results of the other established techniques, it is confirmed that the proposed LWSC model obtained state-of-the-art classification accuracy which makes it an efficient solution for breast cancer diagnosis. These findings could efficiently help expert histopathologist in maintaining a balance between accuracy and precision while saving time.

Author Contributions

G.U.N. and H.N.M.: Conceptualization, Methodology, Resources, Data curation, Formal analysis, Writing—original draft, Funding acquisition, Project administration, Supervision. G.T.M.: Investigation, Software, Validation, Visualization, Writing—review and editing. V.S.R.P., S.N., C.C.U.: Software, Validation, Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to the reason that this study only makes use of publicly available data.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets as cited within the text as reference [26] used in this study are available at: https://www.kaggle.com/datasets/ambarish/breakhis (accessed on: 23 June 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Evans, D.G.R.; Howell, A. Breast cancer risk-assessment models. Breast Cancer Res. 2007, 9, 1–8. [Google Scholar] [CrossRef] [Green Version]
  2. Mehta, J.M.; MacLaughlin, K.L.; Millstine, D.M.; Faubion, S.S.; Wallace, M.R.; Shah, A.A.; Fields, H.E.; Ruddy, B.E.; Bryan, M.J.; Patel, B.; et al. Breast cancer screening: Women’s attitudes and beliefs in light of updated united states preventive services task force and american cancer society guidelines. J. Women’s Health 2019, 28, 302–313. [Google Scholar] [CrossRef] [PubMed]
  3. Pfeiffer, R.M.; Park, Y.; Kreimer, A.R.; Lacey, J.V., Jr.; Pee, D.; Greenlee, R.T.; Buys, S.S.; Hollenbeck, A.; Rosner, B.; Gail, M.H.; et al. Risk prediction for breast, endometrial, and ovarian cancer in white women aged 50 y or older: Derivation and validation from population-based cohort studies. PLoS Med. 2013, 10, e1001492. [Google Scholar] [CrossRef] [PubMed]
  4. Stark, G.F.; Hart, G.R.; Nartowt, B.J.; Deng, J. Predicting breast cancer risk using personal health data and machine learning models. PLoS ONE 2019, 14, e0226765. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Sa-nguanraksa, D.; Sasanakietkul, T.; Charoenrat, C.O.; Kulprom, A.; Charoenrat, P.O. Gail model underestimates breast cancer risk in Thai population. Asian Pac. J. Cancer Prev. 2019, 20, 2385–2389. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Clendenen, T.V.; Ge, W.; Koenig, K.L.; Afanasyeva, Y.; Agnoli, C.; Brinton, L.A.; Darvishian, F.; Dorgan, J.F.; Eliassen, A.H.; Falk, R.T.; et al. Breast cancer risk prediction in women aged 35–50 years: Impact of including sex hormone concentrations in the Gail model. Breast Cancer Res. 2019, 21, 1–12. [Google Scholar] [CrossRef] [Green Version]
  7. Hart, G.R.; Nartowt, B.; Muhammad, W.; Liang, Y.; Huang, G.; Deng, J. Endometrial cancer risk prediction and stratification using personal health data. Int. J. Radiat. Oncol. Biol. Phys. 2019, 105, E134–E135. [Google Scholar] [CrossRef]
  8. Belsare, A.D.; Mushrif, M.M.; Pangarkar, M.A.; Meshram, N. Classification of breast cancer histopathology images using texture feature analysis. In Proceedings of the TENCON 2015-2015 IEEE Region 10 Conference, Macao, China, 1–4 November 2015; pp. 1–5. [Google Scholar] [CrossRef]
  9. Vu, T.H.; Mousavi, H.S.; Monga, V.; Rao, G.; Rao, U.K.A. Histopathological image classification using discriminative feature-oriented dictionary learning. IEEE Trans. Med. Imaging 2016, 35, 738–751. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Anuranjeeta, A.; Shukla, K.K.; Tiwari, A.; Sharma, S. Classification of histopathological images of breast cancerous and non cancerous cells based on morphological features. Biomed. Pharmacol. J. 2017, 10, 353–366. [Google Scholar] [CrossRef]
  11. Mouelhi, A.; Rmili, H.; Ali, J.B.; Sayadi, M.; Doghri, R.; Mrad, K. Fast unsupervised nuclear segmentation and classification scheme for automatic allred cancer scoring in immunohistochemical breast tissue images. Comput. Methods Programs Biomed. 2018, 165, 37–51. [Google Scholar] [CrossRef]
  12. Khalilabad, N.D.; Hassanpour, H. Employing image processing techniques for cancer detection using microarray images. Comput. Biol. Med. 2017, 81, 139–147. [Google Scholar] [CrossRef] [PubMed]
  13. Kaymak, S.; Helwan, A.; Uzun, D. Breast cancer image classification using artificial neural networks. Procedia Comput. Sci. 2017, 120, 126–131. [Google Scholar] [CrossRef]
  14. Wang, F.; Zhang, S.; Henderson, L.M. Adaptive decision-making of breast cancer mammography screening: A heuristic based regression model. Omega 2018, 76, 70–84. [Google Scholar] [CrossRef]
  15. Mohebian, M.R.; Marateb, H.R.; Mansourian, M.; Mañanas, M.A.; Mokarian, F. A hybrid computer-aided-diagnosis system for prediction of breast cancer recurrence (HPBCR) using optimized ensemble learning. Comput. Struct. Biotechnol. J. 2017, 15, 75–85. [Google Scholar] [CrossRef] [PubMed]
  16. Monday, H.N.; Li, J.; Nneji, G.U.; Hossin, M.A.; Nahar, S.; Jackson, J.; Chikwendu, I.A. WMR-DepthwiseNet: A Wavelet Multi-Resolution Depthwise Separable Convolutional Neural Network for COVID-19 Diagnosis. Diagnostics 2022, 12, 765. [Google Scholar] [CrossRef]
  17. Wang, P.; Song, Q.; Li, Y.; Lv, S.; Wang, J.; Li, L.; Zhang, H. Cross-task extreme learning machine for breast cancer image classification with deep convolutional features. Biomed. Signal Process. Control 2020, 57, 101789. [Google Scholar] [CrossRef]
  18. Kumar, A.; Singh, S.K.; Saxena, S.; Lakshmanan, K.; Sangaiah, A.K.; Chauhan, H.; Shrivastava, S.; Singh, R.K. Deep feature learning for histopathological image classification of canine mammary tumors and human breast cancer. Inf. Sci. 2020, 508, 405–421. [Google Scholar] [CrossRef]
  19. Kaur, P.; Singh, G.; Kaur, P. Intellectual detection and validation of automated mammogram breast cancer images by multi class SVM using deep learning classification. Inform. Med. Unlocked 2019, 16, 100151. [Google Scholar] [CrossRef]
  20. Ting, F.F.; Tan, Y.J.; Sim, K.S. Convolutional neural network improvement for breast cancer classification. Expert Syst. Appl. 2019, 120, 103–115. [Google Scholar] [CrossRef]
  21. Li, H.; Zhuang, S.; Li, D.; Zhao, J.; Ma, Y. Benign and malignant classification of mammogram images based on deep learning. Biomed. Signal Process. Control 2019, 51, 347–354. [Google Scholar] [CrossRef]
  22. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep learning to improve breast cancer detection on screening mammography. Sci. Rep. 2019, 9, 1–12. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Saha, M.; Chakraborty, C. Her2Net: A deep framework for semantic segmentation and classification of cell membranes and nuclei in breast cancer evaluation. IEEE Trans. Image Process. 2018, 27, 2189–2200. [Google Scholar] [CrossRef]
  24. Rustam, Z.; Amalia, Y.; Hartini, S.; Saragih, G.S. Linear discriminant analysis and support vector machines for classifying breast cancer. IAES Int. J. Artif. Intell. 2021, 10, 253–256. [Google Scholar] [CrossRef]
  25. Khan, S.; Islam, N.; Jan, Z.; Din, I.U.; Rodrigues, J.J.P.C. A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 2019, 125, 1–6. [Google Scholar] [CrossRef]
  26. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A dataset for breast cancer histopathological image classification. IEEE Trans. Biomed. Eng. 2015, 63, 1455–1462. [Google Scholar] [CrossRef] [PubMed]
  27. Bayramoglu, N.; Kannala, J.; Heikkila, J. Deep learning for magnification independent breast cancer histopathology image classification. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2440–2445. [Google Scholar]
  28. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. Breast cancer histopathological image classification using convolutional neural networks. In Proceedings of the 2016 International Joint Conference on Neural Networks OJCN, Vancouver, BC, Canada, 24–29 July 2016; pp. 2560–2567. [Google Scholar]
  29. Han, Z.; Wei, B.; Zheng, Y.; Yin, Y.; Li, K.; Li, S. Breast cancer multi-classification from histopathological images with structured deep learning model. Sci. Rep. 2017, 7, 1–10. [Google Scholar] [CrossRef]
  30. Alom, M.K.; Al-Hayani, W.; Algamal, Z.Y. Classification of breast cancer histopathology images based on adaptive sparse support vector machine. Appl. Math. Bioinform. 2017, 7, 49. [Google Scholar]
  31. Nneji, G.U.; Cai, J.; Deng, J.; Monday, H.N.; James, E.C.; Ukwuoma, C.C. Multi-Channel Based Image Processing Scheme for Pneumonia Identification. Diagnostics 2022, 12, 325. [Google Scholar] [CrossRef]
  32. Mgbejime, G.T.; Hossin, M.A.; Nneji, G.U.; Monday, H.N.; Ekong, F. Parallelistic Convolution Neural Network Approach for Brain Tumor Diagnosis. Diagnostics 2022, 12, 2484. [Google Scholar] [CrossRef]
  33. Hao, Y.; Zhang, L.; Qiao, S.; Bai, Y.; Cheng, R.; Xue, H.; Zhang, G. Breast cancer histopathological images classification based on deep semantic features and gray level co-occurrence matrix. PLoS ONE 2022, 17, e0267955. [Google Scholar] [CrossRef]
Figure 1. Illustration of the histopathological images of breast cancer from the BreaKHis database.
Figure 1. Illustration of the histopathological images of breast cancer from the BreaKHis database.
Diagnostics 13 00299 g001
Figure 2. Histopathological breast cancer images. The first row represents the original images. The second row depicts the edge detected images. The third row presents the contrast-enhanced images.
Figure 2. Histopathological breast cancer images. The first row represents the original images. The second row depicts the edge detected images. The third row presents the contrast-enhanced images.
Diagnostics 13 00299 g002
Figure 3. Structure of the lightweight separable convolution.
Figure 3. Structure of the lightweight separable convolution.
Diagnostics 13 00299 g003
Figure 4. Accuracy performance report of the BC classification for both two classes and eight classes. (a) Test accuracy curve for different magnification factors for two classes of BC classification using the proposed model with contrast-enhanced images, (b) Test accuracy curve for different magnification factors for eight classes of BC classification using the proposed model with contrast-enhanced images.
Figure 4. Accuracy performance report of the BC classification for both two classes and eight classes. (a) Test accuracy curve for different magnification factors for two classes of BC classification using the proposed model with contrast-enhanced images, (b) Test accuracy curve for different magnification factors for eight classes of BC classification using the proposed model with contrast-enhanced images.
Diagnostics 13 00299 g004
Figure 5. Comparison report of the BC classification for two classes. (a) Precision–recall curve for different magnification factors for two classes of BC classification; (b) ROC curve for different magnification factors for two classes BC classification.
Figure 5. Comparison report of the BC classification for two classes. (a) Precision–recall curve for different magnification factors for two classes of BC classification; (b) ROC curve for different magnification factors for two classes BC classification.
Diagnostics 13 00299 g005
Figure 6. Comparison report of the BC classification for eight classes. (a) Precision–recall curve for different magnification factors for eight classes BC classification; (b) ROC curve for different magnification factors for eight classes BC classification.
Figure 6. Comparison report of the BC classification for eight classes. (a) Precision–recall curve for different magnification factors for eight classes BC classification; (b) ROC curve for different magnification factors for eight classes BC classification.
Diagnostics 13 00299 g006
Table 1. Description of the BreakHis dataset for Binary class.
Table 1. Description of the BreakHis dataset for Binary class.
Magnifying FactorBenignMalignantTotal
40×62513701995
100×64414372081
200×62313902013
400×58812321820
Sum Total248054297909
Number of Patients245882
Table 2. Description of the BreakHis dataset for Multi-class.
Table 2. Description of the BreakHis dataset for Multi-class.
ClassesSub-ClassesPatients Count40×100×200×400×Total
Adenosis (A)4114113111106444
Fibroadenoma (F)102532602642371014
BenignTubular Adenoma (TA)3109121108115453
Phyllodes Tumor (PT)7149150140130569
Ductal Carcinoma (DC)388649038967883451
MalignantLobular Carcinoma (LC)5156170163137626
Mucinous Carcinoma (MC)9205222196169792
Papillary Carcinoma (PC)6145142135138560
Total 8219952081201318207909
Table 3. Comparison of the proposed model with different pre-processing images.
Table 3. Comparison of the proposed model with different pre-processing images.
ModelBinary ClassMulti-Class
ACC (%)SEN (%)SPE (%)PRE (%)AUC (%)Time (min)ACC (%)SEN (%)SPE (%)PRE (%)AUC (%)Time (min)
LWSC + Original images90.1190.0791.7691.7892.457.392.1391.7890.6691.9292.417.5
LWSC + ED-based images91.7692.1890.8291.6592.555.893.7192.6291.0992.8593.165.9
LWSC + CE-based images93.1293.6194.0793.3494.215.897.2397.7197.9398.1198.025.9
Table 4. A comparison analysis of the proposed model with state-of-the-art models on BreakHis dataset.
Table 4. A comparison analysis of the proposed model with state-of-the-art models on BreakHis dataset.
State-of-the-Art ModelACC (%)AUC (%)SPE (%)SEN (%)F1-Score (%)
Kumar et al. [18]93.095.0NRNR94
Kaur et al. [19]92.099.090.093.096.0
Ting et al. [20]90.5090.1090.7189.4NR
Rustam et al. [24]98.77NR97.6496.4499.0
Spanhol et al. [26]85.086.1NRNRNR
Bayramoglu et al. [27]83.25NRNRNRNR
Alom et al. [30]97.6598.9197.5292.9NR
Hao et al. [33]96.75NR96.997.18NR
LWSC + CE-based images97.2396.5397.9397.7197.98
Table 5. Fair comparison of the proposed model on BreakHis dataset using different pre-trained models on the same computing device.
Table 5. Fair comparison of the proposed model on BreakHis dataset using different pre-trained models on the same computing device.
ModelBinary CategoryMulti-Class Category
ACCSENSPEACCSENSPE
AlexNet89.591.688.088.988.087.2
VGG-1695.296.394.996.497.196.0
ResNet-10193.994.594.194.595.093.2
DenseNet-12194.593.792.294.294.694.9
Inception V397.397.897.095.896.095.3
Xception91.992.689.691.392.890.7
LWSC + CE-based images93.1293.6194.0797.2397.7197.93
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nneji, G.U.; Monday, H.N.; Mgbejime, G.T.; Pathapati, V.S.R.; Nahar, S.; Ukwuoma, C.C. Lightweight Separable Convolution Network for Breast Cancer Histopathological Identification. Diagnostics 2023, 13, 299. https://doi.org/10.3390/diagnostics13020299

AMA Style

Nneji GU, Monday HN, Mgbejime GT, Pathapati VSR, Nahar S, Ukwuoma CC. Lightweight Separable Convolution Network for Breast Cancer Histopathological Identification. Diagnostics. 2023; 13(2):299. https://doi.org/10.3390/diagnostics13020299

Chicago/Turabian Style

Nneji, Grace Ugochi, Happy Nkanta Monday, Goodness Temofe Mgbejime, Venkat Subramanyam R. Pathapati, Saifun Nahar, and Chiagoziem Chima Ukwuoma. 2023. "Lightweight Separable Convolution Network for Breast Cancer Histopathological Identification" Diagnostics 13, no. 2: 299. https://doi.org/10.3390/diagnostics13020299

APA Style

Nneji, G. U., Monday, H. N., Mgbejime, G. T., Pathapati, V. S. R., Nahar, S., & Ukwuoma, C. C. (2023). Lightweight Separable Convolution Network for Breast Cancer Histopathological Identification. Diagnostics, 13(2), 299. https://doi.org/10.3390/diagnostics13020299

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop