Next Article in Journal
Multiplex RT Real-Time PCR Based on Target Failure to Detect and Identify Different Variants of SARS-CoV-2: A Feasible Method That Can Be Applied in Clinical Laboratories
Next Article in Special Issue
Controversies in the Staging of Patients with Locally Advanced Cervical Cancer
Previous Article in Journal
Primary Lymphoproliferative Lung Diseases: Imaging and Multidisciplinary Approach
Previous Article in Special Issue
ColpoClassifier: A Hybrid Framework for Classification of the Cervigrams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min–Max Neural Network for Cervical Cancer Diagnosis

1
Department of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune 411044, India
2
Faculty of Electrical Engineering and Information Technology, University of Oradea, 410087 Oradea, Romania
3
Department of ECE, Karunya Institute of Technology and Sciences, Coimbatore 641114, India
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(7), 1363; https://doi.org/10.3390/diagnostics13071363
Submission received: 13 February 2023 / Revised: 27 March 2023 / Accepted: 2 April 2023 / Published: 6 April 2023
(This article belongs to the Special Issue Imaging of Cervical Cancer)

Abstract

:
Medical image analysis and classification is an important application of computer vision wherein disease prediction based on an input image is provided to assist healthcare professionals. There are many deep learning architectures that accept the different medical image modalities and provide the decisions about the diagnosis of various cancers, including breast cancer, cervical cancer, etc. The Pap-smear test is the commonly used diagnostic procedure for early identification of cervical cancer, but it has a high rate of false-positive results due to human error. Therefore, computer-aided diagnostic systems based on deep learning need to be further researched to classify the pap-smear images accurately. A fuzzy min–max neural network is a neuro fuzzy architecture that has many advantages, such as training with a minimum number of passes, handling overlapping class classification, supporting online training and adaptation, etc. This paper has proposed a novel hybrid technique that combines the deep learning architectures with machine learning classifiers and fuzzy min–max neural network for feature extraction and Pap-smear image classification, respectively. The deep learning pretrained models used are Alexnet, ResNet-18, ResNet-50, and GoogleNet. Benchmark datasets used for the experimentation are Herlev and Sipakmed. The highest classification accuracy of 95.33% is obtained using Resnet-50 fine-tuned architecture followed by Alexnet on Sipakmed dataset. In addition to the improved accuracies, the proposed model has utilized the advantages of fuzzy min–max neural network classifiers mentioned in the literature.

1. Introduction

Cervical cancer is a type of cancer that develops in the cells of the cervix, which is the lower part of the uterus that connects to the vagina. Cervical cancer is usually caused by a human papillomavirus (HPV) infection, which is a sexually transmitted infection. HPV is a very common virus that can cause abnormal changes in the cells of the cervix, which can eventually lead to cancer if left untreated [1].
Cervical carcinoma is the most prevalent cancer diagnosed in 23 countries and the primary cause of mortality in 36 nations [1,2]. Furthermore, 85 percent of cervical cancers were encountered in the late stages. It is the fourth most frequent cancer in women as well as the leading cause of death, with an approximate 604,000 reported incidents and 342,000 deaths worldwide in 2020 [1]. Figure 1 depicts the mortality age-standardized rates and region-specific incidence for cervical cancer in 2020. The (W) world age standardized incidence rate is shown in descending order, and the highest national age-standardized incidence and mortality rates are overlaid. In such areas, it is critical to ensure that resource-intensive vaccination and screening programs are carried out to improve the situation [2].
Pap smear, liquid based cytology, and colposcopy are the main screening methods for cervical cancer diagnosis. In a Pap-smear test, cell samples are collected from the transformation zone of the cervix, and for abnormalities, it is examined under the microscope. The colposcopy examination deals with examining abnormalities in the cervix with the help of the colposcope; it is a direct visual examination done by gynecologists [3]. Regular screening of women over 30 years of age is advisable for early detection and treatment.
The human-based smear analysis is difficult, laborious, time consuming, costly, and prone to errors since each smear slide consists of approximately 3 million cells with varying overlapping and orientation, necessitating the development of a computerized system capable of analyzing the Pap smear effectively and efficiently [4]. Extensive research has been conducted to assist pathologists in tracking cervical cancer with the development of computer-aided diagnostic (CAD) systems. This type of system consists of different steps, including image preprocessing, segmentation, feature extraction, feature selection, and classification. To enhance the image quality, filtering-based preprocessing is carried out. Much work is carried out to segment the nucleus and cytoplasm using different image-processing techniques [5]. The images are used to extract texture, morphological, and color metric features. The feature selection techniques are applied for the identification of the most discriminant features, and then, classifiers are designed to classify the cervical cytology cell images [6].
The above mentioned workflow necessitates multiple steps for processing the data. The handcrafted features lack the guarantee superior classification performance, highlighting the inadequacy of automatic learning. Deep learning methods have demonstrated success in a variety of applications over the last decade, including object recognition, natural language processing, signal processing, image classification, segmentation, and so on [7,8,9,10]. The deep network architecture has the ability to learn features automatically based on the spatial relationships among the pixels. The multiple layers with simple nonlinear activation functions are used to transform input data from abstract to specific at multiple levels of feature representation.
The network can learn such hierarchical feature representations from a large scale of training data in an unsupervised or supervised manner. In many practical applications, such learned hierarchical features have outperformed handcrafted designs [11].
Lotfi A. Zadeh [12] proposed a fuzzy logic data analysis approach and an engineering approach. Fuzzy set theory is the basis for fuzzy logic which deals with reasoning that is approximate rather than precise in classical two-valued logic. As a result, it is a technique for formalizing the human capacity for imprecise reasoning. Such reasoning exemplifies the human ability to reason roughly and make decisions in the face of uncertainty [12]. Fuzzy set theory is considered a good framework for classification problems because of the inherent fuzziness in the cluster. FMMN has been used in many applications, including fault detection, lung cancer detection, breast cancer detection, medical data analysis, etc. [13,14,15].
This paper presents a hybrid method for the classification of cytology Pap-smear images into abnormal and normal. The machine learning classifiers and fuzzy min–max neural network are trained for two-class problems using the features to extract by fine tuning the deep learning pre-trained models. The following are the main contributions of the proposed work.
(1) Presents a novel and hybrid approach by leveraging the strengths of pre-trained deep learning models with machine learning classifiers and fuzzy min–max neural networks.
(2) Fine tunes the pretrained CNN architectures, including Alexnet, ResNet-18, ResNet-50, and GoogleNet, to overcome the dataset limitations.
(3) Extracts the learned and specific features from Pap-smear images, which are proven to be more effective than handcrafted features and classify by using different machine learning classifiers and enhancing the classification performance using fuzzy min-max neural network.
(4) Provides improved accuracy with the advantages of different properties of the fuzzy min–max neural network classifier given by Simpson [16].

2. Literature Review

To classify the cervical cytology images, various deep learning and machine learning-based techniques are used, for example, researchers in [17,18] make use of local binary pattern, texture, histogram features, local binary pattern, and grey level features. The features are then given as input to a hybrid classifier system that combines SVM and a neuro-fuzzy for classification of the cervical images [19].
Jyothi Priyankaa et al. (2021) [20] consider Pap smear test images for cancerous cell prediction combined with deep learning techniques for more efficient results. The ResNet50 pre-trained model of convolutional neural networks (CNNs) for the prediction of cancerous cells produces accurate results. Except for the final layer, which is trained according to the requirements, all the layers in the proposed work are considered as they are. This methodology correctly classifies all classes with 74.04 percent accuracy.
Deep transfer learning was used by Anurag Tripathi et al. (2021) [21] to aid in the diagnosis of cervical cancer. They used the SIPAKMED dataset for this purpose. Dyskeratotic, koilocytotic, metaplastic, parabasal, and superficial intermediate were the five classes used. The testing accuracy of ResNet50 is 93.87 percent. The ResNet-152 model achieved an accuracy of 94.89 percent. VGG-16 performed best with parabasal cells, achieving the lowest accuracy of all four models at 92.85 percent. The testing accuracy of VGG-19 was slightly higher than that of VGG-16, which was 94.38 percent.
Wafa Mousser et al. (2019) [22] used deep neural networks and optimized MLP classifiers for the classification of Herlev Pap-smear images. Feature extraction is done using deep neural networks and classification using optimized MLP classifiers. The ability of feature extraction from four different pre-trained models to classify Pap-smear images was investigated. The comparisons concluded that ResNet50 outperforms the VGGs and the InceptionV3 by 15% in Pap-smear image classification.
Kurnianingsih et al. (2019) [23] applied mask R-CNN to the whole slide cell image, outperforming the previous segmentation method in precision, recall, and ZSI. For classification, a VGG-like net is used on whole segmented cells. Results shown for binary classification problem had 98.1% accuracy and for the seven-class problem accuracy of 95.9% is obtained.
Sornapudi et al. (2019) [24] proposed a method for automatically classifying cervical cell images by generating labelled patch data, fine-tuning convolutional neural networks for the extraction of deep hierarchical features and the novel graph-based cell detection approach for cellular level evaluation. The results demonstrated that the proposed pipeline could classify images of single cells as well as overlapping cells. The VGG-19 model performed accurately at classifying cervical cytology patch data, with a precision-recall curve of 95%.
The deep learning approach reviewed in Swati Shinde et al. (2022) [25] can directly process raw images and offers automated learning of features based on specific objective functions, such as detection, segmentation, and classification. Different existing pre-trained models, such as ResNet-50, ResNet-152, and VGG are used in the literature for the classification of Pap-smear images for the diagnosis of cervical cancer. Table 1 shows the summarization of the different papers studied and analyzed.

3. Proposed Methodology

In this paper, a hybrid convolutional neural network classification technique is proposed to classify the cervical cytology images into abnormal and normal. Figure 2 shows the block diagram of the proposed work. The offered hybrid CNN framework is divided into two major phases. In the first phase, a pre-trained deep learning model for feature extraction is used. Successive layers, such as FC6 and FC7, are used to extract features. In the second phase, machine learning classifiers and fuzzy min–max neural network is used for the classification process [27].

3.1. Module 1

3.1.1. Feature Extraction Using Pre-Trained Models

For medical image analysis, deep learning architecture is most prevalent. To train a convolutional neural network, a massive quantity of data and high computational resources are required, as well as a longer training time. Transfer learning (TL) is a solution to this problem because it aids in the creation of an accurate model by beginning to learn from previous patterns of knowledge on solving various problems instead of starting from scratch [28,29]. As a result, TL is a technique in artificial intelligence that allows us to transfer knowledge from one model to another [30]. A TL process consists of two steps.
Step 1: Choose a pre-trained model that is trained on large-scale data that is relevant to the problem at hand.
Step 2: Fine-tune a pre-trained model based on the similarity of our dataset.
AlexNet, GoogleNet, ResNet-18, and ResNet-50 are different pre-trained deep learning architectures that have been experimented with using the proposed hybrid technique. AlexNet, GoogleNet, ResNet-18, and ResNet-50 networks are utilized in the transfer learning process, with the weights pre-trained on the ImageNet dataset [31]. ImageNet is made up of 1 million training images, 50,000 validation images, and 100,000 testing images from 1000 different classes. The earlier layers of the pre-trained models are frozen, which capture more low-level features. Alexnet fc7 layer, ResNet-18 pool 5 layer, ResNet-50 fc1000 layer, and Googlenet loss3-classifier layer are used as features. Figure 2 shows the overall process carried out where feature extraction is carried out using AlexNet. Similarly, GoogleNet, ResNet-18, and ResNet-50 are used. For the machine learning classifiers in Module 2, the number of features is fed for training and testing, as mentioned in Table 2.

3.1.2. Min–Max Normalization

Along with the various machine learning algorithms, the fuzzy min–max neural network is also tested. For classification, the features are normalized and fed into a fuzzy min–max neural network. One of the most common methods for normalizing data is min–max normalization. For each feature, the minimum value is converted to 0, the maximum value is converted to 1, and all other values are converted to a decimal between 0 and 1. The following equation is used to normalize the features [32].
X n e w = X X m i n / X m a x X m i n  
where X is the set is of feature values obtained,
  • X m i n is minimum value in X , and
  • X m a x is maximum value in X .

3.2. Module 2

3.2.1. Machine Learning Classifiers

Classification is a machine learning method that determines which class a new object belongs to based on a set of predefined classes. There are numerous classifiers that can be used to classify data, including decision trees, bays, functions, rules, lazy, meta, and so on. In this work we used different classifiers belonging to the different families, and performance comparison is to evaluate the best classifier. We experimented with the BayesNet, Naive Bayes, random forest, random tree, decision table and part machine learning classifiers.

3.2.2. Fuzzy Min–Max Neural Network

Simpson pioneered the hyperboxes for pattern classification [16]. FMM learns using a hyperbox fuzzy set. An expansion parameter theta (θ) controls the size of the hyperbox; in this case the theta (θ) ranges from values 0 to 1. The maximum (max) and minimum (min) points in a hyperbox are used to measure how a training sample accommodates in the hyperbox from a fuzzy membership function [31].
Equation (2) defines a hyperbox fuzzy logic with maximum (HW), minimum (HV), and unit hypercube In points. Figure 3 depicts a 3-D hyperbox with its maximum point (HWj) and minimum point (HVj).
Fuzzy logic Hj can be used to represent each hyperbox as follows [16]:
H j = H A h ,   H V j ,   H W j ,   f H A h , H V j ,   H W j                 H A h   ϵ   I n
where hth represents the input pattern as HAh = (ah1, ah2, …, ahn). jth hyperbox minimum and maximum points are represented as HVj = (hvj1, hvj2, …, hvhn) and HWj = (hwj1, hwj2, …, hwhn) respectively.
Fuzzy min–max classifier is made up of three layers. The first is input feature vectors (FA), the second is the fuzzy hyperbox sets (FB), and the third is the classification nodes (FC). The fuzzy membership computes the input pattern for various hyperboxes and determines the pattern’s class label. The feature vector obtained from the feature extraction step is provided to the input layer, FA. For hyperboxes, the membership function is evaluated by the nodes (bj) in the fuzzy hyperbox set layer (FB). V and W represent the weights of connections between layers FA and FB, which are a set of min and max points of hyperboxes, respectively. The FMMN expansion process [16] is used to update these parameters. U stores the weights between the nodes in the middle and third layers. Equation (3) shows the U is computed.
u j k = 1               i f   H j   i s   h y p e r b o x   f o r   c l a s s   C K 0                     o t h e r w i s e                                                                          
FMMN calls the membership function when a new input sample is provided. Equation (4) is used to calculate the membership value.
H j H A h     1 2 n i = 1 n   m a x 0 , 1 m a x 0 ,   γ   m i n 1 ,   a h i h w j i + m a x 0 , 1 m a x 0 ,   γ   m i n 1 ,   h v j i a h i    
where Hj denotes the membership of jth hyper box, HAh is the hth input data, HWji is the maximum point of Hj, HVji is the minimum point of Hj, and γ indicates the sensitivity parameter which controls the decrease in speed of membership value as the gap between HAh and Hj rises. The FMMN classification method is primarily based on expansion test, overlap test, and contraction test.

Expansion

To include a new input pattern, HAh, in the hyperbox, the following equation is used to determine if a hyperbox can be expanded.
n θ i = 1 n max h w j i , a h i   min h v j i , a h i  

Overlap Test

If a hyperbox is chosen for expansion, an overlap test is run to determine whether there is any overlapping between two or more hyperboxes caused by the expansion. If any of the following conditions are met, overlapping of hyperboxes will occur.
  • Case 1
    HV ji < HV ki < HW ji < HW ki δ n   =   min ( HW ji HV ki ,   δ o )
  • Case 2
    HV hi < HV ki < HW ki < HW ji δ n   =   min ( HW ki HV ji ,   δ o )
  • Case 3
    HV ji < HV ki < HW ki < HW ji δ n   =   min ( min ( HW ji HV ki , HW ki HV ji ) ,   δ o )
  • Case 4
    HV ki < HV ji < HW ji < HW ki δ n   =   min ( min ( HW ji HV ki , HW ki HV ji ) ,   δ o )

Contraction

A suitable contraction rule is applied to eliminate the overlap between the hyperboxes if the overlap is detected. The corresponding contraction rules are shown in the following equations with respect to the overlap test rules as stated in the overlap test.
  • Case 1
    HV j Δ < HV k Δ < HW j Δ < HW k Δ < HW j Δ new = HW k Δ new = ( HW k Δ old + HW j Δ old ) / 2
  • Case 2
    HV k Δ < HV j Δ < HW k Δ < HW j Δ < HW k Δ new = HV j Δ new = ( HW k Δ old + HW j Δ old ) / 2
  • Case 3(a)
    HV j Δ < HV k Δ < HW k Δ < HW j Δ   and   ( HW k Δ HV j Δ ) < ( HW j Δ HV k Δ ) ,   HV j Δ new + HW k Δ old
  • Case 3(b)
    HV j Δ < HV k Δ < HW k Δ < HW j Δ   and   ( HW k Δ HV j Δ ) < ( HW j Δ HV k Δ ) ,   HW j Δ new + HV k Δ old
  • Case 4(a)
    HV k Δ < HV j Δ < HW j Δ < HW k Δ   and   ( HW k Δ HV j Δ ) < ( HW j Δ HV k Δ ) ,   HW k Δ new = HV j Δ old
  • Case 4(b)
    HV k Δ < HV j Δ < HW j Δ < HW jk Δ   and   ( HW k Δ HV j Δ ) > ( HW j Δ HV k Δ ) ,   HV j Δ new = HW k Δ old
The training process is completed after successful completion of the preceding three processes, which results in a list of hyperboxes to represent the FMM network.

3.3. Algorithm 1

The algorithm for the proposed work is as follows:
Algorithm 1: Algorithm for cervical cancer classification
  Input:  Herlev dataset, Sipakmed dataset of Pap-smear images
  Output: Prediction of classes—normal or abnormal
  Begin
  Step 1: Pre-process the images
  Step 2: Split the dataset into training and testing datasets
  Step 3: Pre-trained models= {AlextNet, GoogleNet, ResNet18, ResNet50}
  Step 4: For each model in Step 3
    Train the model
    Extract the feature vector
  Step 5: Classifiers = {{machine learning classifiers: simple logistic, Naive Bays, Bayes Net, decision table, random forest, random tree, PART}, {fuzzy min–max neural network}}
  Step 6: For each classifier in Step 5
    Train with the feature vector
    Evaluate with Testing Set
  End

4. Experimentation Environment

The proposed technique is implemented using Matlab software with Intel core i5 processor and 4 GB RAM. To investigate the effectiveness of the proposed techniques, it is applied to two different standard datasets, namely the Herlev dataset and the Sipakmed dataset. Both the datasets are rearranged into two classes, normal and abnormal, and the proposed techniques are used to solve binary classification. The dataset is split into training and testing.

4.1. Herlev Dataset

It consists of 917 single cell images. Seven classes are converted to normal and abnormal. The normal class contains 242 images, while 675 images belong to the malignant class. Table 3 shows the cell distribution of the dataset and Figure 4 shows sample images from the Herlev dataset [33].

4.2. Sipakmed

The Sipakmed dataset consists of 4049 images. There are five categories for classification of the Sipakmed dataset: dyskeratotic, metaplastic, koilocytotic, parabasal, and superficial-intermediate [34]. The Sipakmed dataset samples are shown in Figure 5. Table 4 shows the cell distribution of the dataset.

4.3. Performance Measures

Choosing an appropriate evaluation metric is critical for overcoming bias among the various algorithms. Accuracy, sensitivity, specificity, precision and F1 Score are different performance metrics to evaluate the classification performance. True positive (TP) is the number of correctly labelled positive samples, true negative (TN) is the number of correctly classified negative samples, false positive (FP) is the number of negative samples classified as positive, and false negative (FN) is the number of positive instances predicted as negative (FN) [35]. Table 5 shows the formula of evaluation metrics.

5. Experiments and Results

The results of an experiment carried out when the AlexNet pretrained model is used as a feature extractor are shown in Table 6. From the results it can be analyzed that the highest classification testing accuracy of 88.6% is given by the simple logistic classifier on the Herlev dataset. With the Sipakmed dataset, 95.14% highest classification accuracy is given by the simple logistic classifier. Hence, the combination of Alexnet with a simple logistic classifier among the experimentations has the best performance.
Experimentation carried out with the GoogleNet pre-trained model results are demonstrated in the following Table 7. Highest testing classification accuracy on Herlev dataset is obtained with simple logistic of 87.32%. On the Sipakmed dataset, the highest accuracy obtained is 92.21% with simple logistic classifiers. With the Googlenet also, the simple logistic is outperforming the other classifiers.
Experimentation carried out with the ResNet-18 pre-trained model results are accumulated in the following Table 8. The highest testing classification accuracies of 88.76% and 93.85% are obtained with the simple logistic classifier on the Herlev and Sipakmed datasets, respectively.
Table 9 shows the experiment carried out when the ResNet-50 pre-trained model is used as a feature extractor. From the results it can be analyzed that the highest classification testing accuracies of 92.03% and 93.60% are given by the simple logistic classifier on the Herlev and Sipakmed datasets, respectively.
Binary classification of cervical cytology images is performed using the pre-trained models, and fuzzy min–max neural networks are elaborated further. Table 10 shows the results of the AlexNet pre-trained model used as a feature extractor. From the tables it can be observed that the highest classification accuracy on the Herlev dataset is 90.22% and good sensitivity of 95% with θ 0.3, whereas the 95.33% is the highest classification accuracy on the Sipakmed dataset and good sensitivity of 95% with θ 0.5. Along with the accuracy, sensitivity, specificity, precision, and F1 score are calculated and presented in the table.
Table 11 represents the results of the Googlenet pre-trained model. From the tables it can be observed that highest classification accuracy on the Herlev dataset is 89.49% and good sensitivity of 97% with θ 0.6, whereas 92.13% is the highest classification accuracy on the Sipakmed dataset and good sensitivity of 91% with θ 0.3.
The results of the RestNet-18 model are shown in Table 12. The highest classification accuracy on the Herlev dataset is 91.67% and good sensitivity of 99% with θ 0.5, whereas 92.87% is the highest classification accuracy on the Sipakmed dataset and good sensitivity of 93% with θ 0.4.
The results of the RestNet-50 model are shown in Table 13. The highest classification accuracy on the Herlev dataset is 88.77% and good sensitivity of 91%, whereas 95.33% is the highest classification accuracy on the Sipakmed dataset and good sensitivity of 95% with 0 and 0.5, respectively.

Performance Analysis

The result analysis discussed above shows that the proposed techniques give overall good classification accuracy. Comparing the performance of the different pretrained models, the best classification accuracy obtained by the experimented pre-trained models is shown in Figure 6. The performance comparison demonstrated with the best classification accuracy, RestNet-50 followed by Alexnet, has performed better than other models with best accuracies of 95.33% and 95.32%, respectively.
The performance comparison between the machine learning classifiers and the FMMN for classification shows that overall, the performance of the FMMN outperforms the machine learning classifier. Table 14 shows the comparative analysis.
Comparing the two datasets with the classification accuracy obtained, it can be observed from Figure 7 that the Sipakmed dataset average classification accuracy with all the pre-trained models have outperformed over the Herlev dataset. As mentioned, the convolutional neural networks need large amounts of data to train the models, and the Sipakmed dataset has a considerably large number of images as compared to the Herlev dataset. Table 15 shows the comparative study outcomes with the results of the existing studies on cervical cancer diagnosis that uses Pap-smear images using computer-aided applications.
The advantage of the proposed method is it has given a significant good accuracy and sensitivity for the cervical cancer image classification compared with the existing methods. However, the limitation is FMMN is a complex architecture that requires a significant amount of computational resources and training data.

6. Conclusions

A novel hybrid deep learning technique is proposed to solve the problem of cervical cytology image classification to aid pathologists to carry out the smear test with good accuracy and less time. The proposed hybrid technique is based on deep learning pretrained models, transfer learning, machine learning classifiers, and fuzzy min–max neural network. Attempts are made to compare the performance of different deep learning models. The highest classification accuracy is given by the ResNet-50 classifier of 95.33% with theta value 0.5. Experimentation is performed on two different datasets to evaluate the performance. Results obtained on the Sipakmed dataset were better than those obtained on the Herlev dataset.
The future scope is to use the modified versions of the fuzzy min–max neural network to improve the classification accuracy. The seven-class, five-class problem for classification can be experimented with the proposed techniques to evaluate the performance for multiclass classification problem.

Author Contributions

Conceptualization, S.S.; Methodology, M.K.; Investigation, M.K.; Writing—Original draft, M.K.; Writing—Review & editing, S.S., D.E.P. and D.J.H.; Supervision, D.E.P. and D.J.H.; Funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Department of Science and Technology Ministry of Science and Technology, India, grant number TDP/BDTD/29/2021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be available on request to authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. Vaccarella, S.; Laversanne, M.; Ferlay, J.; Bray, F. Cervical cancer in Africa, Latin America and the Caribbean and Asia: Regional inequalities and changing trends. Int. J. Cancer 2017, 141, 1997–2001. [Google Scholar] [CrossRef] [Green Version]
  3. Chandran, V.; Sumithra, M.G.; Karthick, A.; George, T.; Deivakani, M.; Elakkiya, B.; Subramaniam, U.; Manoharan, S. Diagnosis of Cervical Cancer based on Ensemble Deep Learning Network using Colposcopy Images. BioMed Res. Int. 2021, 2021, 5584004. [Google Scholar] [CrossRef]
  4. Gençtav, A.; Aksoy, S.; Önder, S. Unsupervised segmentation and classification of cervical cell images. Pattern Recognit. 2012, 45, 4151–4168. [Google Scholar] [CrossRef] [Green Version]
  5. Jantzen, J.; Norup, J.; Dounias, G.; Bjerregaard, B. Pap-smear benchmark data for pattern classification. In Proceedings of the NiSIS 2005: Nature Inspired Smart Information Systems (NiSIS), EU Co-Ordination Action, Albufeira, Portugal, 1 January 2005; pp. 1–9. [Google Scholar]
  6. Marinakis, Y.; Dounias, G.; Jantzen, J. Pap smear diagnosis using a hybrid intelligent scheme focusing on genetic algorithm based feature selection and nearest neighbor classification. Comput. Biol. Med. 2009, 39, 69–78. [Google Scholar] [CrossRef] [PubMed]
  7. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  9. Sutskever, I.; Vinyals, O.; Le, Q. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27, 3104–3112. [Google Scholar] [CrossRef]
  10. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.E.; Mohamed, A.-R.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.N.; et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  11. Yan, Z.; Zhan, Y.; Zhang, S.; Metaxas, D.; Zhou, X.S. Chapter 4—Multi-Instance Multi-Stage Deep Learning for Medical Image Recognition, Deep Learning for Medical Image Analysis; Academic Press: Cambridge, MA, USA, 2017; pp. 83–104. ISBN 9780128104088. [Google Scholar] [CrossRef]
  12. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef] [Green Version]
  13. Ahmed, A.A.; Mohammed, M.F. SAIRF: A similarity approach for attack intention recognition using fuzzy min-max neural network. J. Comput. Sci. 2018, 25, 467–473. [Google Scholar] [CrossRef]
  14. Deshmukh, S.; Shinde, S. Diagnosis of lung cancer using pruned fuzzy min-max neural network. In Proceedings of the 2016 International Conference on Automatic Control and Dynamic Optimization Techniques ICACDOT, Pune, India, 9–10 September 2016; pp. 398–402. [Google Scholar] [CrossRef]
  15. Quteishat, A.; Lim, C.P. Application of the fuzzy min-max neural networks to medical diagnosis. In Proceedings of the 12th International Conference on Knowledge-Based Intelligent Information and Engineering Systems, Part III, Zagreb, Croatia, 3–5 September 2008; pp. 548–555. [Google Scholar] [CrossRef]
  16. Simpson, P. Fuzzy min-max neural networks. I. Classification. IEEE Trans. Neural Netw. 1992, 3, 776–786. [Google Scholar] [CrossRef]
  17. Sukumar, P.; Gnanamurthy, R. Computer aided detection of cervical cancer using pap smear images based on hybrid classifiers. Int. J. Appl. Eng. Res. Res. India Publ. 2015, 10, 21021–21032. [Google Scholar]
  18. Alaslani, M.G.; Elrefaei, L. Convolutional Neural Network Based Feature Extraction for IRIS Recognition. Int. J. Comput. Sci. Inf. Technol. 2018, 10, 65–78. [Google Scholar] [CrossRef] [Green Version]
  19. Athinarayanan, S.; Srinath, M.V.; Kavitha, R. Multi Class Cervical Cancer Classification by using ERSTCM, EMSD & CFE methods based Texture Features and Fuzzy Logic based Hybrid Kernel Support Vector Machine Classifier. IOSR J. Comput. Eng. 2017, 19, 23–34. [Google Scholar] [CrossRef]
  20. Priyankaa, J.; Bhadri Rajub, M.S.V.S. Machine Learning Approach for Prediction of Cervical Cancer. Turk. J. Comput. Math. Educ. 2021, 12, 3050–3058. [Google Scholar]
  21. Tripathi, A. Classification of cervical cancer using Deep Learning Algorithm. In Proceedings of the Fifth International Conference on Intelligent Computing and Control Systems (ICICCS 2021), Madurai, India, 6–8 May 2021. pp. 1210–1218. [CrossRef]
  22. Mousser, W.; Ouadfel, S. Deep Feature Extraction for Pap-Smear Image Classification: A Comparative Study. In Proceedings of the ICCTA 2019: 2019 5th International Conference on Computer and Technology Applications, Istanbul, Turkey, 16–17 April 2019; pp. 6–10. [Google Scholar] [CrossRef]
  23. Kurnianingsih; Allehaibi, K.H.S.; Nugroho, L.E.; Widyawan; Lazuardi, L.; Prabuwono, A.S.; Mantoro, T. Segmentation and Classification of Cervical Cells Using Deep Learning. IEEE Access 2019, 7, 116925–116941. [Google Scholar] [CrossRef]
  24. Sornapudi, S.; Brown, G.T.; Xue, Z.; Long, R.; Allen, L.; Antani, S. Comparing Deep Learning Models for Multi-cell Classification in Liquid- based Cervical Cytology Image. AMIA Annu. Symp. Proc. AMIA Symp. 2020, 2019, 820–827. [Google Scholar]
  25. Shinde, S.; Kalbhor, M.; Wajire, P. DeepCyto: A hybrid framework for cervical cancer classification by using deep feature fusion of cytology images. Math. Biosci. Eng. 2022, 19, 6415–6434. [Google Scholar] [CrossRef]
  26. Kalbhor, M.; Shinde, S.V.; Jude, H. Cervical cancer diagnosis based on cytology pap smear image classification using fractional coefficient and machine learning classifiers. TELKOMNIKA Telecommun. Comput. Electron. Control 2022, 20, 1091–1102. [Google Scholar] [CrossRef]
  27. Kalbhor, M.; Shinde, S. ColpoClassifier: A Hybrid Framework for Classification of the Cervigrams. Diagnostics 2023, 13, 1103. [Google Scholar] [CrossRef] [PubMed]
  28. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. Int. J. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  29. Raghu, M.; Zhang, C.; Kleinberg, J.; Bengio, S. Transfusion: Understanding transfer learning for medical imaging. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; pp. 3347–3357. [Google Scholar] [CrossRef]
  30. Lin, H.; Hu, Y.; Chen, S.; Yao, J.; Zhang, L. Fine-Grained Classification of Cervical Cells Using Morphological and Appearance Based Convolutional Neural Networks. IEEE Access 2019, 7, 71541–71549. [Google Scholar] [CrossRef]
  31. Kalbhor, M.; Shinde, S.; Joshi, H.; Wajire, P. Pap smear-based cervical cancer detection using hybrid deep learning and performance evaluation. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2023, 19, 6415–6434. [Google Scholar] [CrossRef]
  32. Mbaga, A.H.; Zhijun, P. Pap Smear Images Classification for Early Detection of Cervical Cancer. Int. J. Comput. Appl. 2015, 118, 10–16. [Google Scholar] [CrossRef]
  33. Shanthi, P.B.; Hareesha, K.S.; Kudva, R. Automated Detection and Classification of Cervical Cancer Using Pap Smear Microscopic Images: A Comprehensive Review and Future Perspectives. Eng. Sci. 2022, 19, 20–41. [Google Scholar] [CrossRef]
  34. Plissiti, M.E.; Dimitrakopoulos, P.; Sfikas, G.; Nikou, C.; Krikoni, O.; Charchanti, A. Sipakmed: A new dataset for feature and image based classification of normal and pathological cervical cells in Pap smear images. IEEE Int. Conf. Image Process. 2018, 2018, 3144–3148. [Google Scholar] [CrossRef]
  35. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
Figure 1. Mortality Age-Standardized Rates and Region-Specific Incidence for Cervical Cancer in 2020. Reprinted with permission from Ref. [1]. Copyright 2020 IARC/WHO.
Figure 1. Mortality Age-Standardized Rates and Region-Specific Incidence for Cervical Cancer in 2020. Reprinted with permission from Ref. [1]. Copyright 2020 IARC/WHO.
Diagnostics 13 01363 g001
Figure 2. Block diagram of proposed work.
Figure 2. Block diagram of proposed work.
Diagnostics 13 01363 g002
Figure 3. Hyperbox.
Figure 3. Hyperbox.
Diagnostics 13 01363 g003
Figure 4. Sample images from Herlev dataset.
Figure 4. Sample images from Herlev dataset.
Diagnostics 13 01363 g004
Figure 5. Sample images from Sipakmed dataset.
Figure 5. Sample images from Sipakmed dataset.
Diagnostics 13 01363 g005
Figure 6. Performance comparison of pre-trained models.
Figure 6. Performance comparison of pre-trained models.
Diagnostics 13 01363 g006
Figure 7. Performance comparison with datasets.
Figure 7. Performance comparison with datasets.
Diagnostics 13 01363 g007
Table 1. Summarization of Prevailing Research Work.
Table 1. Summarization of Prevailing Research Work.
Paper Data SetPre-ProcessingFeature Extraction/
Classification
Results
[20]Herlev University HospitalResize, Color to Grey, Expansion of dimensionsRESNET-50Accuracy 74.04%
[21]SIPAKMEDResize 244 × 244RESNET-50, RESNET-152, VGG-16, VGG-19Highest 94.89% accuracy was obtained with ResNet-152
[22]Herlev University HospitalData AugmentationVGG16. InceptionV3
VGG19, ResNet50
Classification—MLP classifier
ResNet-50 89%
[23]Herlev University HospitalData Augmentation
Segmentation—Mask R-CNN
VGGNetMask R-CNN segmentation produces
the best average performance, i.e., 0.92 ± 0.06 precision, 0.91 ± 0.05 recall and 0.91 ± 0.04 ZSI and 0.83 ± 0.10
Binary classification problem 98.1% accuracy
Seven-class problem high accuracy of 95.9%
[24]Herlev University HospitalSubtraction of blue color space from red color space, skeletonizing and refining boundariesVGG-19, ResNet-50, DenseNet-120, and Inception_v3VGG-19—88% Accuracy
[25]Herlev University Hospital, SIPAKMED,
LBC
Data AugmentationXceptionNet,
VGGNet, ResNet50 and Ensemble of classifiers
Accuracy 97%, 99%, and 100%
[26]Herlev University HospitalResize 256 × 256DCT and Haar transformHighest 81.11% accuracy was obtained with DCT
Table 2. Number of Features Extracted from Pre-Trained Models.
Table 2. Number of Features Extracted from Pre-Trained Models.
Pre-Trained ModelAlexnetGooglenetResnet-18Resnet-50
Number of Features409610005121000
Table 3. Herlev dataset.
Table 3. Herlev dataset.
Cell Category Number of Cells
Normal squamousNormal74
Intermediate squamous70
Columnar98
Mild dysplasiaAbnormal182
Moderate dysplasia146
Severe dysplasia197
Carcinoma in situ150
Total 917
Table 4. Sipakmed dataset.
Table 4. Sipakmed dataset.
Cell Category Number of Cells
SuperficialNormal831
Parabasal787
KoilocytoticAbnormal825
Dyskeratotic813
MetaplasticBenign793
Total4049
Table 5. Performance Evaluation Metric.
Table 5. Performance Evaluation Metric.
Assessments Formula
Accuracy TP   +   TN TP   +   TN   +   FP   +   FN
Sensitivity/Recall   TP TP   +   FN
Specificity TN   TN   +   FP
Precision TP   TP   +   FP
F1 Score 2 × Pricision   ×   Recall   Precision   +   Recall
Table 6. Classification accuracy of Alexnet model with machine learning classifiers.
Table 6. Classification accuracy of Alexnet model with machine learning classifiers.
AlexNet
DatasetClassifierBayes NetNavie BayesRandom ForestRandom TreeDecision TablePartSimple Logistic
HerlevTesting Accuracy (%)83.3382.2487.6881.888.0486.5988.6
Sipakmed91. 291.691.290.7093.2389.595.14
Table 7. Classification accuracy of Googlenet model with machine learning classifiers.
Table 7. Classification accuracy of Googlenet model with machine learning classifiers.
GoogleNet
DatasetClassifierBayeNetNavie BayesRandom ForestRandom TreeDecision TablePartSimple Logistic
HerlevTesting Accuracy (%)83.7082.9786.9681.8884.0686.5987.32
Sipakmed87.3785.2490.2483.1187.6289.7592.21
Table 8. Classification accuracy of ResNet-18 model with machine learning classifiers.
Table 8. Classification accuracy of ResNet-18 model with machine learning classifiers.
ResNet-18
DatasetClassifierBayeNetNaive BayesRandom ForestRandom TreeDecision TablePartSimple Logistic
HerlevTesting Accuracy (%)86.5986.5987.6882.684.4279.7188.76
Sipakmed90.989.2688.3680.4984.7588.4293.85
Table 9. Classification accuracy of ResNet-50 model with machine learning classifiers.
Table 9. Classification accuracy of ResNet-50 model with machine learning classifiers.
ResNet-50
DatasetClassifierBayeNetNaive BayesRandom ForestRandom TreeDecision TablePartSimple Logistic
HerlevTesting Accuracy (%)88.0489.1388.0478.6286.2381.8892.03
Sipakmed89.6788.1989.8381.884.759093.60
Table 10. Performance Evaluation of Alexnet Pre-Trained Model with Fuzzy Min–Max Neural Network.
Table 10. Performance Evaluation of Alexnet Pre-Trained Model with Fuzzy Min–Max Neural Network.
Theta00.10.20.30.40.50.60.70.80.91
AlexnetHerlev DatasetAccuracy87.3284.0684.0690.2282.9784.7885.1488.0484.7839.8634.78
Sensitivity0.900.940.860.950.850.900.910.970.910.190.11
Specificity0.810.580.780.770.770.700.700.640.680.991.00
Precision0.930.860.920.920.910.890.890.880.890.971.00
F1 Score0.910.900.890.930.880.900.900.920.900.310.20
Sipakmed DatasetAccuracy92.6293.2095.0895.0093.9395.3394.9293.6990.8280.6680.00
Sensitivity0.950.930.940.940.930.950.950.940.950.990.99
Specificity0.900.930.960.970.950.960.950.930.850.540.52
Precision0.930.950.970.980.970.970.970.950.900.760.76
F1 Score0.940.940.960.960.950.960.960.950.930.860.86
Table 11. Performance Evaluation of Googlenet Pre-Trained Model with Fuzzy Min–Max Neural Network.
Table 11. Performance Evaluation of Googlenet Pre-Trained Model with Fuzzy Min–Max Neural Network.
Theta00.10.20.30.40.50.60.70.80.91
GooglenetHerlev DatasetAccuracy82.2586.2383.7084.7886.9688.4189.4988.0486.9682.2582.25
Sensitivity0.870.930.890.890.920.980.970.970.950.870.87
Specificity0.680.670.700.740.740.620.700.630.640.700.70
Precision0.890.890.890.900.910.880.900.880.880.890.89
F1 Score0.880.910.890.900.910.930.930.920.910.880.88
Sipakmed DatasetAccuracy89.3490.6690.6692.1391.1591.8091.1588.5285.1683.0382.79
Sensitivity0.910.910.920.910.890.910.900.860.860.960.93
Specificity0.860.900.890.940.940.920.930.920.840.640.68
Precision0.910.930.930.960.960.950.950.940.890.800.81
F1 Score0.910.920.920.930.920.930.920.900.870.870.87
Table 12. Performance Evaluation of ResNet-18 Pre-Trained Model with Fuzzy Min–Max Neural Network.
Table 12. Performance Evaluation of ResNet-18 Pre-Trained Model with Fuzzy Min–Max Neural Network.
Theta00.10.20.30.40.50.60.70.80.91
ResNet-18HerlevAccuracy88.7775.0089.4989.1391.3091.6788.0486.9686.2386.9686.96
Sensitivity0.920.920.910.910.970.990.970.940.940.950.95
Specificity0.810.270.860.850.750.730.640.670.640.660.66
Precision0.930.780.950.940.920.910.880.890.880.880.88
F1 Score0.920.840.930.920.940.950.920.910.910.910.91
SipakmedAccuracy91.4890.8291.3192.7992.8793.7790.9086.8081.7277.2172.46
Sensitivity0.930.920.920.920.930.930.930.920.910.930.96
Specificity0.890.880.900.940.930.950.870.790.670.530.36
Precision0.930.920.930.960.950.960.920.870.810.750.70
F1 Score0.930.920.930.940.940.950.930.890.860.830.81
Table 13. Performance Evaluation of ResNet-50 Pre-Trained Model with Fuzzy Min–Max Neural Network.
Table 13. Performance Evaluation of ResNet-50 Pre-Trained Model with Fuzzy Min–Max Neural Network.
Theta00.10.20.30.40.50.60.70.80.91
ResNet50HerlevAccuracy88.7786.2387.3288.0487.3287.3285.8787.3286.9682.2581.88
Sensitivity0.910.930.910.900.900.930.890.930.910.830.85
Specificity0.840.680.780.820.790.730.770.730.770.810.73
Precision0.940.890.920.930.920.900.910.900.920.920.90
F1 Score0.920.910.910.920.910.910.900.910.910.870.87
SipakmedAccuracy92.0592.6292.7094.1895.2595.3394.1889.1084.0280.8272.70
Sensitivity0.930.930.940.950.940.950.940.850.820.950.99
Specificity0.900.920.910.930.970.960.950.960.870.600.32
Precision0.930.950.940.950.980.970.960.970.910.780.69
F1 Score0.930.940.940.950.960.960.950.900.860.860.81
Table 14. Best Classification Accuracy (%) of Two Datasets.
Table 14. Best Classification Accuracy (%) of Two Datasets.
AlexNetGoogleNetResNet18ResNet50
Herlev90.22 (FMMN)89.49 (FMMN)91.67 (FMMN)92.03 (Simple logistic)
Sipakmed95.32 (FMMN)92.21 (Simple logistic)93.85 (Simple logistic)95.33 (FMMN)
Table 15. Comparison Between the Proposed Method with the Existing Studies.
Table 15. Comparison Between the Proposed Method with the Existing Studies.
ApproachAccuracy
Deep Learning (Resnet-50) [20]74.04%
ResNet-152 [21]94.89%
ResNet-50 [22]89%
VGG-19 [24]88%
Proposed Model [Hybrid CNN] ResNet5095.33%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kalbhor, M.; Shinde, S.; Popescu, D.E.; Hemanth, D.J. Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min–Max Neural Network for Cervical Cancer Diagnosis. Diagnostics 2023, 13, 1363. https://doi.org/10.3390/diagnostics13071363

AMA Style

Kalbhor M, Shinde S, Popescu DE, Hemanth DJ. Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min–Max Neural Network for Cervical Cancer Diagnosis. Diagnostics. 2023; 13(7):1363. https://doi.org/10.3390/diagnostics13071363

Chicago/Turabian Style

Kalbhor, Madhura, Swati Shinde, Daniela Elena Popescu, and D. Jude Hemanth. 2023. "Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min–Max Neural Network for Cervical Cancer Diagnosis" Diagnostics 13, no. 7: 1363. https://doi.org/10.3390/diagnostics13071363

APA Style

Kalbhor, M., Shinde, S., Popescu, D. E., & Hemanth, D. J. (2023). Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min–Max Neural Network for Cervical Cancer Diagnosis. Diagnostics, 13(7), 1363. https://doi.org/10.3390/diagnostics13071363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop