Next Article in Journal
Impact of the COVID-19 Pandemic on Psychosocial Well-being and Adaptation in Children and Adolescents with Asperger’s Syndrome
Next Article in Special Issue
A Machine Learning Approach for Monitoring and Classifying Healthcare Data—A Case of Emergency Department of KSA Hospitals
Previous Article in Journal
The Effect of Farmland Transfer on Agricultural Green Total Factor Productivity: Evidence from Rural China
Previous Article in Special Issue
Infant Low Birth Weight Prediction Using Graph Embedding Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Detection of Oral Squamous Cell Carcinoma from Histopathological Images of Oral Mucosa Using Deep Convolutional Neural Network

1
Department of Computer Application, Siksha ‘O’ Anusandhan Deemed to be University, Bhubaneswar 751030, India
2
Department of Computer Science and Engineering, Siksha ‘O’ Anusandhan Deemed to be University, Bhubaneswar 751030, India
3
Department of Computer Science and Engineering, SRM University-AP, Guntur 522240, India
*
Authors to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2023, 20(3), 2131; https://doi.org/10.3390/ijerph20032131
Submission received: 8 December 2022 / Revised: 19 January 2023 / Accepted: 20 January 2023 / Published: 24 January 2023
(This article belongs to the Special Issue Machine Learning for Healthcare Applications)

Abstract

:
Worldwide, oral cancer is the sixth most common type of cancer. India is in 2nd position, with the highest number of oral cancer patients. To the population of oral cancer patients, India contributes to almost one-third of the total count. Among several types of oral cancer, the most common and dominant one is oral squamous cell carcinoma (OSCC). The major reason for oral cancer is tobacco consumption, excessive alcohol consumption, unhygienic mouth condition, betel quid eating, viral infection (namely human papillomavirus), etc. The early detection of oral cancer type OSCC, in its preliminary stage, gives more chances for better treatment and proper therapy. In this paper, author proposes a convolutional neural network model, for the automatic and early detection of OSCC, and for experimental purposes, histopathological oral cancer images are considered. The proposed model is compared and analyzed with state-of-the-art deep learning models like VGG16, VGG19, Alexnet, ResNet50, ResNet101, Mobile Net and Inception Net. The proposed model achieved a cross-validation accuracy of 97.82%, which indicates the suitability of the proposed approach for the automatic classification of oral cancer data.

1. Introduction

In India, the rate of oral cancer, specifically OSCC, is increasing, and the main reason behind this rise is alcohol consumption and tobacco chewing, etc. The death rate of oral cancer is also high in India. There are commercial advertisements regarding the disadvantage of tobacco and alcohol consumption; however, due to the lack of understanding and inadequate knowledge, people are still in the habit of alcohol and tobacco, which causes an increase in the number of oral cancer patients [1,2]. One of the international agencies for research conducted a survey and predicted that the number of cancer patients will increase from 1 million in 2012 to more than 1.7 million by 2035. This implies the death rate will also increase from 680,000 to 1–2 million by 2035 [3]. Hence, it is utmost important to detect the OSCC at its early stage so that the treatment can be started as early as possible and the death rate due to oral cancer can be reduced. With the advancement of technology, continuous research has been performed by many researchers for the early detection of oral cancer and, at the same time, enormous amounts of oral cancer data are collected and made available for research purposes. One of the challenging tasks for doctors is to correctly predict the type and stage of cancer. The oldest method used by doctors is the physical screening in the first stage and, after that, for the conformity purpose, biopsy is used. With the improvement in computer technology, many machine learning techniques along with image processing techniques are used by researchers to predict the type as well as stage of cancer, and that can help the doctors to give a better treatment to the OSCC patient. Among different imaging techniques, the histopathological imaging technique is more suitable for OSCC diagnosis. Thus, in this work, the author has focused on histopathological images of OSCC for the detection and classification problem.
Certain features are seen in the cell morphology, which can also be found in digital histopathological images of OSCC. Thus, it is the fundamental task of various machine learning algorithms to extract those features that can be mapped with the original cell morphology and predict the presence or absence of cancer cells accurately. A few successful supervised and unsupervised machine learning algorithms are support vector machines, neural networks, decision tree, fuzzy, genetic algorithms, k-NN, kernel PCA, etc., used for medical histopathological image classification problems [4,5,6,7,8,9]. However, in the current era, deep learning techniques, specifically convolutional neural networks (CNN), have become more popular among researchers for the image analysis problem. It is already proven that CNN outperforms for different computer vision problems mostly for object detection; however, in recent studies, it has shown that CNN can also outperform for different low-level image processing problems, such as restoration, denoising and mitosis detection, and these are the general problem in histopathological image analysis. CNN is a special type of artificial neural network (ANN). The ANN was inspired by biological neural structures and it has a learning ability like biological neurons. Because of these characteristics, a neural network can generate new dataset from the existing ones and can discover optimal outputs. CNN can observe new samples and after generalization can produce new learning rules from the sample, and this learning rule can be used to decide the output when an unseen dataset is used. The recent challenges in the area of medical imaging are to use these CNN approaches to handle the classification problem. The efficiency of deep learning, mainly in CNN, is to find out the architecture, which meets the issues of medical image classification, so that the predictable outcome can be improved [10,11].

1.1. Contribution

The application of deep learning techniques is quite high in several types of cancer; however, limited research analysis has been done using histopathological OSCC images. This research deliberates the classification of oral cancer data samples using histopathological images. The classification outcome can be used as the input for other tasks like the extraction of the nucleus feature classification and predicting various stages of cancer. Our main task is to design an optimal CNN model which can achieve the automatic detection of oral cancer from the histopathological images.

1.2. Organization

The rest of the paper is arranged as follows: related work is discussed in Section Two. The methods and techniques used are discussed in Section Three. In Section Four, the proposed methodology along with the proposed CNN model are highlighted. Experimental studies are discussed in Section Five; performance evaluation and result analysis are completed in Section Six. Finally, in the Section Seven, the conclusion is elaborated.

2. Related Work

Machine learning algorithms became highly successful in disease prediction models; however, the performance of the models degrades in the case of an insufficient and imbalance dataset. Models developed using insufficient dataset face the problem of overfitting. An overfitted model’s prediction is either underestimated or overestimated, which leads to poor generalization in real time of the dataset. To address the problem of model overfitting, the author of [12] used the penalized regression method for the risk prediction model. Similarly, to deal with insufficient data, in the work [13], the author discusses a novel age adoption algorithm based on feature compensation and soft decision threshold adjustment for diabetes mellitus risk prediction. In [14], a detailed survey is conducted on a few shot learning techniques to deal with the insufficient dataset. Various image data augmentation techniques are discussed in [15] to deal with the insufficient dataset problem.
Research on the use of machine learning algorithms for the detection and classification of cancer almost covers all the varieties of cancer, which include brain cancer, skin cancer, breast cancer, lung cancer, prostate cancer, oral cancer, etc. Here, the author presents a comprehensive study on the area of oral cancer.
In the literature, for oral cancer classification and for its early detection, different machine learning and deep learning techniques are widely used. Most of the research is done on the efficiency of different machine learning algorithms for the prediction of oral sub-mucous fibrosis (OSF) [16,17,18]. The advanced stage of OSF can lead to oral cancer.
A detailed study was done on the classification of OSCC by T.Y Rahaman et al. [19,20]. Rahaman et al., used Support Vector Machine (SVM) and Linear Discriminant Classifier (LDA) for the binary classification of the OSCC images. In both SVM and LDA, they used texture, color and shape features for the classification. In 2017, they used two sets of datasets, which were collected from Ayursundra healthcare Pvt. Ltd. and Dr. Borooah Cancer Research Institute, respectively, for the experiment. In the first dataset, there were a total of 110 normal images and 113 malignant images. In the second dataset, 86 normal and 88 malignant images were there. In the research work, Rahaman et al., applied SVM on 400× magnification images and achieved acceptable accuracy [19].
In 2012, Muthu et al. [21] applied a few classifiers, namely fuzzy, decision tree, Gaussion mixture model (GMM), K-nearest neighbor and probabilistic neural network, for the classification of OSF, and they found that the fuzzy classifier gave the highest accuracy of 95.7% as compared to other classifiers.
Muthu et al. also considered morphological and texture features in another study, for the classification of OSF, and in that study they applied SVM and GMM and found out that SVM gives better results compared to the GMM model [22].
Anuradha. K. et al. reflect the use of distinctive features, such as energy, entropy, contrast, correlation and homogeneity, along with the SVM classifier for the classification of the oral tumor as malignant or benign [23].
T. Belvin et al. used 16 malignant images with 192 patches, which were collected from the Himalayan Institute of Medicine, Deradun, for the multiclass classification purpose, and they applied a backpropagation-based artificial neural network model (ANN) [24].
D. Devkumar et al., in 2015, considered 10 numbers of OSCC patients and collected 30 labeled images from Tata medical center, Kolkata. Keratinization and keratin pearls were cast off as features for the keratinization index to grade the OSCC images into three grades, namely grade I, grade II and grade III. This keratinization index is used when the image magnification is as low as ( 4 X ) [25].
In 2018, D. Devkumaret et al., collected 126 images with the size 2048 × 1536 from Barsat Cancer Research and Welfare Center, West Bengal, and they identified various layers, such as epithelial, subepithelial, keratin region and keratin pearls, by applying CNN with the Gober filter and Random Forest, and they achieved an accuracy of 96.88% [26].
In the above studies, the research for the classification of oral cancer using machine learning techniques is stated. From the literature, it is claimed that the prediction of oral cancer is effective for early detection and diagnosis of the disease, and it is also found that the SVM classifier gives comparatively better results than other machine learning algorithms such as decision tree, ANN, Random Forest and KNN. From the literature, it can be seen that most authors have used 10-fold cross-validation method for the evaluation process and different measures, such as accuracy, sensitivity, specificity and ROC curve, are also used.
The selection of the most appropriate algorithm depends on many factors such as the size of the dataset, the image magnification, the noise in the image, etc. The frequently appearing issue in the above literature is the small size of the dataset. For the classification algorithm, the proper amount of training and testing the dataset is required so that the validation accuracy can achieve a promising percentage. Along with the smaller dataset, the other limitations are poor dataset quality and inappropriate feature selection. For the automatic classification of a cancer dataset, it is of the utmost important to use the best feature for the classification purpose, which is a very promising task. In this scenario, deep learning algorithms achieved superior classification accuracy without manual feature extraction requirements.
There are many papers available in the literature for oral cancer prediction, but according to our best knowledge, a limited amount of work has been done using histopathological images along with deep learning advancement. Due to the availability of high-performance GPUs, deep learning algorithms can be used for histopathological images, and in the case of deep learning algorithms, manual feature extraction of the input histopathological images is not required [27,28]. For the classification of histopathological oral cancer images, Navarun et al. [29] compared the use of transfer learning with a CNN model for the classification of histopathological images of oral cancer. The transfer learning is used with pretrained CNN models, namely Alexnet, VGG-16, VGG-19 and ResNet-50. ResNet-50 outperform the other model with an accuracy of 92.15%, whereas Navarun et al. proposed a CNN model and applied it to the same dataset and achieved an accuracy of 97.5%. In the study [29], it is concluded that the CNN model outperforms the four pretrained transfer learning models.
Santisudha et al. suggested deep learning approaches [30,31] whose accuracy is comparable with other state-of-the-art models. In this work [30], the authors suggested three variants of improved ResNet-based models. Out of three models, the best one was chosen for oral cancer image classification. In the work [31], the authors suggested a modified capsule network for the classification task of oral histopathological images and achieved an accuracy of 97.35%.
There are many studies found in the literature that are associated with deep learning for biopsy image classification [32,33]. Thus, deep learning approaches are predicted to outperform other machine learning approaches in the classification problem, without manual feature extraction requirements. However, to our best knowledge, limited work has been done in the domain of histopathological images of oral cancer cells using deep learning approaches. The literature reveals the importance of deep learning methods for classification purposes.
Inspired by the above literature analysis, this proposed study considers the binary classification of oral histopathological images using CNN architecture of different variations and layers. The outcomes of all variations are compared, and the best-performing model for classification is determined.

3. Methods and Techniques Used

In traditional pattern recognition problems, relevant features are extracted by the export of the area manually, and then those features are submitted to the simple neural network for the classification task [34]. However, in deep learning, the relevant features are automatically extracted and used for the solution of the problem. Deep learning is a type of neural network, which takes input in the form of metadata and process these data through several layers to compute the output [35].
In this research work, one of the deep learning techniques, CNN, is used for the classification of oral cancer histopathological images. CNN mainly consists of six layers: (i) input layer, (ii) convolution layer, (iii) pooling layer, (iv) flattening layer, (v) fully connected layer and (vi) output layer [36,37,38].
Input Layer: In this layer, the input is given, and it is converted into a matrix of pixels, where the input is an image.
Convolution Layer: The convolution layer is used to extract features by using a weight matrix in the form of filter. In each layer of convolution, one set of feature maps is generated according to Equation (1).
Y p   r = f [ o N p ( Y o r 1 × M o p r + a p r ) ]
Here, N p represents the input image set, and Y p   r is the pth feature set of the rth layer. represents convolution function. Y o r 1 represents the oth feature map of the r 1   layer. M o p r represents the filter connecting the pth feature map of the r layer and the oth feature map of the r 1 layer. a p r is the bias. f (   ) represents the nonlinear activation functions used such as ReLU, sigmoid and tanh.
Pooling layer: The pooling layer is used to reduce the dimension of the feature map. The redundant and unnecessary features are discarded here. The pooling layer is used in between the subsequent convolution layers [39]. In max pooling methodology, the maximum value of patch is considered as the output and a reduced feature map is generated. In the case of average pooling instead of maximum value, the average value of the pixels is considered for each block. Equation (2) is the mathematical expression for the max pooling.
Y p   r = f [ β p r d o w n c a s t   ( Y o r 1 ) + a p r ]
Here, d o w n c a s t   (   ) is the subsampling function and β represents the subsampling coefficient’ Y p   r represents the pth feature set of the rth layer. Y o r 1 represents the oth feature map of the r 1   layer. a p r is the bias. and f (   ) represents the nonlinear activation function.
Flatten layer: In the flatten layer, the pooled featured map generated in the last pooling layer is converted to a one-dimensional feature map, and this one-dimensional featured map is the input to the fully connected layer.
Fully connected layer: The fully connected layer combines the features transmitted in the previous layer to achieve accuracy in the classification problem.
Output layer: The input image in the form of a set of features that are finally passed over to the output layer after crossing over the convolution, pooling and fully connected layers. The output layer is the classification layer in the form of probability. In the case of binary classifiers, a logistic regression model is used in the output layer; however, for a multiclass classification problem, the softmax classifier is commonly used [38,39]. The softmax function is a normalized exponential function. In the input dataset, let the training dataset consist of n number of tag samples such that { ( y ( 1 ) , z ( 1 ) ) , ( y ( 2 ) , z ( 2 ) ) ( y ( n ) , z ( n ) ) } ; here, z ( i ) { 1 , 2 .. k } . Let y be the input data given, then the probability of j for each category p ( z = j | y ) can be determined using the hypothesis function given in Equation (3).
h o ( y ( i ) ) = [ p ( z ( i ) = 1 y ( i ) ; θ ) p ( z ( i ) = 2 y ( i ) ; θ ) p ( z ( i ) = k y ( i ) ; θ ) ] = 1 j = 1 k e θ j T y ( i ) [ e θ 1 T y ( i ) e θ 2 T y ( i ) e θ k T y ( i ) ]
Here, θ 1 , θ 2 , θ k ϵ R m + 1 is the parameter and 1 j = 1 k e θ j T y ( i ) is used for normalization, and it ensures the probability of all classes is 1.
The loss function is given in Equation (4).
L o s s ( θ ) = 1 n [ i = 1 n j = 1 n 1 { z i = j } log e θ j T y ( i ) l = 1 k e θ j T y ( i ) ]
Here, the function 1 {   } is defined as
1 { i f   t h e   e x p r e s s i o n   v a l u e   i s   t r u e   } = 1 ,
1 { i f   t h e   e x p r e s s i o n   v a l u e   i s   f a l s e   } = 0 ,
The value of j in the 1 {   } is j = { 1 , 2 , k   } . The gradient equation of the loss function is given in Equation (5).
θ j L o s s ( θ ) = 1 n i = 1 n [ y ( i ) ( 1 { z ( i ) = j } p ( z ( i ) = j | y ( i ) ; θ ) )   ]    
Equation (6) is used to update θ j .
θ j = θ j α θ j L o s s ( θ )
The probability function, used to determine the input y, belongs to class j and is given in Equation (7).
p ( z ( i ) = j | y ( i ) ; θ ) = e θ j T y ( i ) l = 1 k e θ j T y ( i )
In this research work, the above general architecture is used to design a customized CNN model for the classification of the histopathological images of oral cancer. The proposed CNN model is compared with a few state-of-the-art models, such as VGG16, VGG19, Alex net, ResNet50, ResNet101, Mobile Net and Inception Net.

3.1. Dataset Used

The present study was conducted for the classification of the oral database into binary classes, such as benign and malignant. The oral dataset for this study was collected from a repository of normal oral cavities and OSCC [40].The dataset consists of 1224 histopathological images of an oral cavity. The total images available in this dataset are of two different categories. In the first category, all the images are of 100 × magnification . In the second category, all the images are of 400 × m a g n i f i c a t i o n . There are 528 images in the first category, out of which 439 images are of oral cancer and 89 are of normal oral images. In the second category, all the images are of 400 × m a g n i f i c a t i o n . Here, a total of 696 images is available, out of which 495 images are of the malignant or cancerous type, and the rest of the 201 images belong to the normal or benign type. The detailed distribution of images available in the dataset is given in Table 1. Considering both categories, there are 1224 total images, out of which there are 290 total normal oral images and 934 total cancerous images. Figure 1 represents the sample images of each category. In the current study, all 1224 images in the category are considered for the histopathological oral cancer image classification purpose.

3.2. Image Preprocessing:

In the current study, the images are collected from the specified repository [40], which we discuss in Section 3.1. The images are of different quality and dimensions. Some images are perfect, whereas some images are with a noisy background. Preprocessing of images is thus required to reduce the impurities and for noise removal. We have used Gaussian blur for the removal of noise from the images and it also smooths the edges. The Gaussian filter is a low pass filter, which reduces noise or high-frequency components from the images [41]. Figure 2 shows an original histopathological oral image and the resultant image after the Gaussian filter is applied to the same image. After the Gaussian filter is applied to all the images, in the next step, all the images are converted to a fixed size of 128 × 128 as the images are of diverse sizes. After the preprocessing of the images, the images are used for the image augmentation process.

3.3. Image Augmentation

The classification task of medical images is getting cutting edge accuracy through the CNN approach; however, there are a few challenges to consider. One of the biggest challenges in medical image classification is the inadequate amount of training and testing data. The deep learning models require a huge dataset to overcome the problem of overfitting and network generalization. Generally, a deep learning model performs better with the balanced and large dataset. In this study, we considered the histopathological oral images dataset, which consists of a total of 1224 images of two classes, namely cancerous and non-cancerous. The dataset is also imbalanced. Thus, image augmentation technique is used to generate balance and an adequate dataset. The augmentation technique is used to generate new dataset from the existing one, without losing any important feature of the image [42]. In this study, we used rotation with a range of 40, height and width shift with a range of 0.2, and zooming and shirring with a range of 0.2. Using all these techniques for data augmentation leads to the generation of 8199 images out of 1224 images. The detail of the images generated is given in Table 2. From Table 2 it is observed that the imbalanced dataset is balanced with a nearly equal number of OSCC and normal images after the augmentation process is carried out. After augmentation, 4119 images are generated in OSCC class, and 4080 images are generated for the normal class. This new dataset with 8199 images is used for the classification task in the proposed CNN model.

3.4. Data Partition

The total number of images generated from Table 2 is 8199, and all these images are split into training and testing dataset with a ratio of 75% and 25% based on the train test split strategy [43]. After the train and test split, the number of images in the training dataset is 6149 and it is 2050 in the testing dataset. While splitting, the dataset to train and the test set, along with an original image and all its augmented images, are kept in the same folder, i.e., either in the train or test set, which ensure the train and test set to be disjointed. The training dataset is used for the initial training of the model by initializing the weights and it is also used for fine tuning of hyperparameters to improve the accuracy of the model [44,45]. The hyperparameters are selected once the model is trained using the train dataset; after that, the test dataset is used to evaluate the predictive accuracy of the model [46]. In this work, we use the mentioned train test split ratio 75% and 25% of the total 8199 images in the proposed CNN model as well as on predefined models, namely VGG16, VGG19, Alexnet, ResNet50, ResNet101, Mobile Net and Inception Net.

4. Proposed Methodology

In the present study, the classification task of histopathological oral images is carried out using the proposed methodology given in Figure 3. The proposed methodology consists of six stages: (1) dataset collection, which is already discussed in Section 3.1; (2) preprocessing of images, which is discussed in Section 3.2; (3) data augmentation, which is discussed in Section 3.3; (4) data partition for the training and testing set, as discussed in Section 3.4; (5) classification of images using proposed CNN and for the comparison purpose various predefined state-of-the-art models like VGG16, VGG19, Alexnet, ResNet50, ResNet101, Mobile Net and Inception Net are also used for the classification task; and finally (6) performance analysis of classification result is carried out over the output of proposed model with all other models considered. The detailed structure of the proposed CNN model used in stage 5 is given in Section 4.1.

4.1. Proposed 10-Layer CNN Architecture

The classification of histopathological images of oral cancer problem can be solved by using the proposed CNN architecture shown in the figure. As discussed in Section 3, the proposed CNN model also consists of six basic layers: (i) input layer, (ii) Convolution layer, (iii) pooling layer, (iv) flattening layer, (v) fully connected layer and (vi) output layer. In the input layer, the histopathological oral images are taken as input images. These images are converted to matrix of pixels. The size of the input layer is 128 × 128 × 3   pixels. In the proposed model, a total of 10 layers is considered, specifically eight convolution layers, one dropout layer, one flatten layer and two fully connected layers. Figure 4 represents the architecture of the proposed 10-layer CNN model. In the proposed model, six pooling layers and six batch normalization layers are also used. In each layer of convolution, one set of the feature map is generated using Equation (1).
In the first convolution layer, 32 filters are used with kernel size of   3 × 3 . The second convolution layer also uses 32 filters with a 3 × 3 kernel size. The 3rd and 4th convolution layers use 64 filters with a kernel size of 3 × 3 . The 5th and 6th convolution layers use 128 filters with a 3 × 3 kernel size. The 7th and 8th convolution layers use 256 filters with the same 3 × 3 kernel size. All the eight convolution layers use rectified linear (ReLu) activation function. The output of one layer is given as input to the next layer. After the 1st convolution layer, a set of one max pooling layer followed by a batch normalization layer is used, and the 1st set of the reduced feature map is generated. Equation (2) is used for the max pooling calculation. After the 2nd convolution layer, again one set of a max pooling and batch normalization layer is used. After the 3rd convolution layer, only one max pooling layer is used. The 4th and 5th convolution layer are connected parallelly and, in between them, there is no other layer. After that one set of a max pooling and batch normalization layer is used. The 6th convolution layer is followed by only one max pooling layer. Similarly, only one batch normalization layer is used after the 7th convolution layer. The 8th convolution layer is followed by global average pooling and one batch normalization layer. After all the convolution layers, a dense layer is used with sigmoid activation function followed by a flatten layer.
The flatten layer is used to convert the resultant output from the previous layer to a single dimension. After the flatten layer one dropout layer is used with a 30% dropout mechanism to overcome the problem of overfitting. Finally, in the output layer, one dense layer is used with the softmax activation function. The proposed model is trained using the ‘Adam’ optimizer and categorical cross entropy loss function is used as the model is proposed for binary classification task. The mathematical expression for the categorial cross entropy loss function is given in Equation (4), which is used in this output layer of the proposed model. Table 3 shows the summary of parameters used for the proposed model in detail.

5. Experimental Studies

In this work for the binary classification of the histopathological image of oral cavity, the proposed methodology along with the 10-layer CNN model is implemented using python on google colab platform. The google colab is a freely available platform and it runs entirely on cloud [47]. The Keras library and Tensorflow framework are used to implement the proposed work. For the execution of the code, the graphics processing unit (GPU) is used. The GPU is also freely available on the google colab platform. For the comparison analysis, along with the customized 10-layer CNN model, seven different types of predefined models, such as VGG16, VGG19, Alexnet, ResNet50, ResNet101, Mobile Net and Inception Net, are also implemented using python on the google colab platform.

Hyper Parameter Setting of the 10-Layer CNN

In this section experimentally the idle hyperparameter is decided for the proposed 10-layer CNN model by considering the validation accuracy as a measure. The proposed model is trained using 100 epochs in the google colab platform, and it took 2 h to complete the training process. The 10-layer CNN is executed several times with different hyperparameter settings, to find out the idle hyperparameters for the model. The performance of the proposed model under different hyperparameter settings is given in Table 4.
The proposed configurations given in the Table 4 and highlighted in bold letters gave best result with the highest accuracy 0.9782 on the test dataset. From Table 4, it is observed that the proposed CNN model gives its best result with eight convolution layers, max pooling through the kernel size 3 × 3 and by using the dropout rate 0.3.

6. Performance Evaluation and Result Analysis

The performance evaluation and results are discussed using three various aspects: (1) statistical measures of confusion metrics, which are used for the evaluation of cross validation accuracy; (2) performance measure graph, which is specifically used for accuracy, and the loss graph, which is used for result analysis; and finally (3) comparison with other models available in the literature is deliberated to prove the superiority of the proposed model.

6.1. Performance Evaluation Using Statistical Measure

The performance evaluation of the proposed model is done using five statistical measures such as precision, recall, specificity, F-measure and, most importantly, accuracy. The mathematical notation for precision is given in Equation (8); similarly, Equations (9)–(12) are used for the mathematical notation of recall or sensitivity, specificity, F-measure and accuracy, respectively.
P r e c i s i o n = T P T P + F P
R e c a l l   O r   S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
A c c u r a c y = T P + T N T P + T N + F P + F N
Here, T P indicates true positive and T N indicates true negative; similarly, FP is false positive and F N is false negative. For the performance evaluation of the proposed 10-layer CNN model, the error rate is also calculated, which is the complementary to the accuracy measure of the model. Equation (13) is used for mathematical notation of the error rate. The error rate gives the misclassification measure [48].
E r r o r   r a t e = 1 A c c u a r c y
To prove the superiority of the proposed 10-layer CNN, all the performance matrices mentioned above are evaluated for the proposed model as well as for the comparative models, namely VGG16, VGG19, Alexnet, ResNet50, ResNet101, Mobile Net and Inception Net. Figure 5 shows the performance evaluation matrix of all the comparative models along with the proposed model. From the figure it is observed that ResNet101, ResNet50, Alexnet, Mobile Net and Inception Net performed better as compared to VGG16 and VGG19. However, the proposed CNN model outperforms all the considered models with accuracy of 0.97, F-measure of 0.97, recall of 0.98, precision of 0.97 and specificity of 0.97.
The performance of the proposed model is also presented in terms of bar chart of error rate and accuracy. In Figure 6, the error rate and accuracy are shown for all the considered models. From this figure, it can be seen that the proposed model achieved the highest accuracy of 0.97 with lowest error rate of 0.03, whereas VGG19 achieved the lowest accuracy of 0.71 with the highest error rate of 0.29. The various performance measures of all the comparative models along with the proposed CNN model are listed out in Table 5.
From the comparative analysis of various performance measures, mentioned in Table 5, it can be justified that the proposed 10-layer CNN model outperforms the other comparative model with the highest accuracy of 0.9782. However, the performance of Alexnet, Resnet50, ResNet101, Mobile Net and Inception Net is promising with accuracy of 0.88, 0.91, 0.89, 0.93 and 0.92, respectively. VGG16 and VGG19 achieved the lowest performance with an accuracy 0.74 and 0.71, respectively.

6.2. Result Analysis Using Performance Measure Graph

In this section, the result analysis of the proposed model is done using a performance measure graph, specifically the accuracy and loss graph, generated during the training and validation process for all the comparative models. A model with minimum loss signifies the best result. The minimum loss indicates that the model learns from the training and validation phase with a lower error rate. On the other hand, the maximum accuracy value indicates optimal results for the model [49].
Figure 7 highlights the accuracy and loss graph of VGG16. From Figure 7, it can be observed that VGG16 showed an average validation accuracy of 0.74 and validation loss comes to a constant value of 0.63.
The accuracy and loss graph of VGG19 is given in Figure 8. In Figure 8, VGG19 showed an improved accuracy with the range of 0.65 to 0.74; however, the validation loss is constant at 0.62. The results of VGG16 and VGG19 confirmed that these two models are not suitable for the considered dataset.
Similarly, Resnet50 shows an accuracy improvement within the range of 0.52 to 0.91 and inversely the validation loss decreases to 0.25. From this study, it is observed that the Alexnet and ResNet50 models show promising performance. The accuracy and loss graphs of Alexnet and ResNet50 are represented in Figure 9 and Figure 10, respectively. Alexnet shows the validation accuracy with a range from 0.65 to 0.88 within 100 epochs, and the validation loss also decreases from 2.31 to 0.31.
The accuracy and loss graph of ResNet101 is given in Figure 11. From the figure, it can be observe that the validation accuracy reached 0.89, whereas the training accuracy reached up to 0.99. The gap between the training and validation accuracy is more as compared to the Rasnet50 and Alexnet. For an optimum model, the gap between training and validation accuracy should be minimum.
The accuracy and loss graph of Inception Net is represented in Figure 12. From the graph, it can be observed that the model achieved a training accuracy of 0.99 and validation accuracy of 0.92. The figure shows the gap between training and validation accuracy is less as compared to Resnet101.
The accuracy and loss graph of the Mobile Net model is given in Figure 13. From the figure, it can be observed that the model achieved a validation accuracy of 0.93 and training accuracy of 0.99. From this observation, it can be concluded that Mobile Net is a better model as compared to the other models discussed above.
However, the proposed 10-layer CNN model achieved the highest validation accuracy of 0.97 with a validation loss decrease from 1.5 to 0.06. Figure 14 shows the proposed 10-layer CNN model’s accuracy and loss graph. In our study, none of the models shows the overfitting problem, which occurs when the training and validation loss are decreasing, but after a certain point, suddenly the validation loss increases. Similarly, none of the models undergoes the problem of underfitting, which occurs when there is a decrease in the training loss until the last epoch is reached. The proposed CNN model can be considered as a good fit model for our dataset as the training and validation loss had a minimal difference, and there is also a smaller gap between training and validation accuracy. Thus, the proposed model is considered as a suitable model for the dataset used.

6.3. Comparative Analysis with Various Models Available in Literature

To show the competence of the proposed model, the performance in terms of accuracy of the model is compared with the existing models available in the literature. Table 6 deliberates the comparative analysis of the various methods available in the literature with the proposed model. For the comparison, we chose only the models where deep learning approaches are considered and the dataset used for the classification task is restricted to histopathological images of oral cavity. By analyzing Table 6, it can be perceived that the proposed model is comparable with the recent literature. The proposed model achieved an accuracy of 0.9782 with an improvement of 0.32 as compared to the accuracy of model proposed by Navarun et al. (2020) [29]. As revealed from the past literature, limited work has been done in the area of classification of histopathological images of oral cavity. In this scenario, our proposed model would also be an added advantage in the particular field due to its high performance.

7. Conclusions

There are numerous codes of behavior for the detection and diagnosis of oral cancer by doctors. The detailed screening of the histopathological biopsy image is one of the major components to understand the diseases and for a better treatment. For the qualitative evaluation of the biopsy image, a skilled pathologist is required who can minutely differentiate between a healthy cell and a cancerous cell from the histopathological biopsy image of an oral cell. This process of qualitative and minute evaluation of biopsy image by the pathologist is a time-consuming process that results in a delay in disease detection and, hence, there will be a delay in treatment. In this aspect, there is a need for automated detection of OSCC to ensure a quick and correct diagnosis. In this current work, the authors propose the use of a deep learning models for the automatic detection of oral cancer from the oral biopsy histopathological image. The proposed 10-layer CNN model outperforms with the highest accuracy of 97.82% as compared to other state-of-the-art models, namely VGG16, VGG19, Alexnet, ResNet50, ResNet101, Inception Net and Mobile Net. The performance of the 10-layer CNN is also compared and analyzed with some of the recent work presented in the literature, and it is found that its performance is strong compared with the recent work done. From the above analysis, it concluded that the proposed 10-layer CNN model can be used as an automated tool to identify oral cancer and can help doctors as a supportive measure for the identification and treatment planning of oral cancer. For a future perspective, the proposed model can be extended to detect distinct stages of oral cancer, which can help both the patient and doctor to defeat cancer.

Author Contributions

Conceptualization, M.D. and R.D.; methodology, M.D. and R.D.; software, M.D. and R.D.; validation, R.D. and S.K.M.; formal analysis, M.D., R.D. and S.K.M.; investigation, R.D. and S.K.M.; resources, M.D.; data curation, M.D. and R.D.; writing—original draft preparation, M.D., R.D. and S.K.M.; writing—review and editing, R.D. and S.K.M.; visualization, R.D. and S.K.M.; supervision, R.D.; project administration, R.D. and S.K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Borse, V.; Konwar, A.N.; Buragohain, P. Oral cancer diagnosis and perspectives in India. Sens. Int. 2020, 1, 100046. [Google Scholar] [CrossRef] [PubMed]
  2. Markopoulos, A.K. Current aspects on oral squamous cell carcinoma. Open Dent. J. 2012, 6, 126. [Google Scholar] [CrossRef] [Green Version]
  3. Bray, F.; Ren, J.-S.; Masuyer, E.; Ferlay, J. Global estimates of cancer prevalence for 27 sites in the adult population in 2008. Int. J. Cancer 2013, 132, 1133–1145. [Google Scholar] [CrossRef]
  4. Ojansivu, V.; Linder, N.; Rahtu, E.; Pietikäinen, M.; Lundin, M.; Joensuu, H.; Lundin, J. Automated classification of breast cancer morphology in histopathological images. Diagn. Pathol. 2013, 8 (Suppl. S1), S29. [Google Scholar] [CrossRef]
  5. Sertel, O.; Kong, J.; Shimada, H.; Catalyurek, U.; Saltz, J.; Gurcan, M. Computer-aided prognosis of neuroblastoma on whole-slide images: Classification of stromal development. Pattern Recognit. 2009, 42, 1093–1103. [Google Scholar] [CrossRef] [Green Version]
  6. Lim, L.A.G.; Maguib, R.N.; Dadios, E.P.; Avila, J.M.C.; Naguib, R.N.G. Implementation of GA-KSOM and ANFIS in the classification of colonic histopathological images. In TENCON 2012 IEEE Region 10 Conference; IEEE: New York, NY, USA, 2012; pp. 1–5. [Google Scholar]
  7. Li, C.; Zhang, S.; Zhang, H.; Pang, L.; Lam, K.; Hui, C.; Zhang, S. Using the K-nearest neighbor algorithm for the classification of lymph node metastasis in gastric cancer. Comput. Math. Methods Med. 2012, 2012, 876545. [Google Scholar] [CrossRef] [Green Version]
  8. Hilado, S.D.F.; Lim, L.A.G.; Naguib, R.N.; Dadios, E.P.; Avila, J.M.C. Implementation of wavelets and artificial neural networks in colonic histopathological classification. J. Adv. Comput. Intell. Intell. Inform. 2014, 18, 792–797. [Google Scholar] [CrossRef]
  9. Deif, M.A.; Attar, H.; Amer, A.; Issa, H.; Khosravi, M.R.; Solyman, A.A.A. A New Feature Selection Method Based on Hybrid Approach for Colorectal Cancer Histology Classification. Wirel. Commun. Mob. Comput. 2022, 2022, 7614264. [Google Scholar] [CrossRef]
  10. Chen, G.; Zhang, J.; Zhuo, D.; Pan, Y.; Pang, C. Identification of pulmonary nodules via CT images with hierarchical fully convolutional networks. Med. Biol. Eng. Comput. 2019, 57, 1567–1580. [Google Scholar] [CrossRef]
  11. Cireşan, D.C.; Giusti, A.; Gambardella, L.M.; Schmidhuber, J. Mitosis detection in breast cancer histology images with deep neural networks. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2013; pp. 411–418. [Google Scholar]
  12. Pavlou, M.; Ambler, G.; Seaman, S.; Guttmann, O.; Elliott, P.; King, M.; Omar, R.Z. How to develop a more accurate risk prediction model when there are few events. BMJ 2015, 351, h3868. [Google Scholar] [CrossRef]
  13. Su, Y.; Huang, C.; Yin, W.; Lyu, X.; Ma, L.; Tao, Z. Diabetes Mellitus risk prediction using age adaptation models. Biomed. Signal Process. Control. 2023, 80, 104381. [Google Scholar] [CrossRef]
  14. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. 2020, 53, 1–34. [Google Scholar] [CrossRef]
  15. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar]
  16. Krishnan, M.M.R.; Acharya, U.R.; Chakraborty, C.; Ray, A.K. Automated diagnosis of oral cancer using higher order spectra features and local binary pattern: A comparative study. Technol. Cancer Res. Treat. 2011, 10, 443–455. [Google Scholar] [CrossRef] [PubMed]
  17. Patra, R.; Chakraborty, C.; Chatterjee, J. Textural analysis of spinous layer for grading oral submucous fibrosis. Int. J. Comput. Appl. 2012, 47, 975–8887. [Google Scholar] [CrossRef]
  18. Driemel, O.; Kunkel, M.; Hullmann, M.; Eggeling, F.V.; Müller-Richter, U.; Kosmehl, H.; Reichert, T.E. Diagnosis of oral squamous cell carcinoma and its precursor lesions. JDDG J. Der Dtsch. Dermatol. Ges. 2007, 5, 1095–1100. [Google Scholar] [CrossRef]
  19. Rahman, T.; Mahanta, L.; Chakraborty, C.; DAS, A.; Sarma, J. Textural pattern classification for oral squamous cell carcinoma. J. Microsc. 2018, 269, 85–93. [Google Scholar] [CrossRef] [PubMed]
  20. Rahman, T.Y.; Mahanta, L.B.; Choudhury, H.; Das, A.K.; Sarma, J.D. Study of morphological and textural features for classification of oral squamous cell carcinoma by traditional machine learning techniques. Cancer Rep. 2020, 3, e1293. [Google Scholar] [CrossRef] [PubMed]
  21. Krishnan, M.M.R.; Venkatraghavan, V.; Acharya, U.R.; Pal, M.; Paul, R.R.; Min, L.C.; Ray, A.K.; Chatterjee, J.; Chakraborty, C. Automated oral cancer identification using histopathological images: A hybrid feature extraction paradigm. Micron 2012, 43, 352–364. [Google Scholar] [CrossRef]
  22. Krishnan, M.M.R.; Chakraborty, C.; Paul, R.R.; Ray, A.K. Hybrid segmentation, characterization and classification of basal cell nuclei from histopathological images of normal oral mucosa and oral submucous fibrosis. Expert Syst. Appl. 2012, 39, 1062–1077. [Google Scholar] [CrossRef]
  23. Anuradha, K.; Sankaranarayanan, K. Detection of Oral Tumors using Marker Controlled Segmentation. Int. J. Comp. Appl. 2012, 52, 15–18. [Google Scholar]
  24. Thomas, B.; Kumar, V.; Saini, S. Texture analysis based segmentation and classification of oral cancer lesions in color images using ANN. In 2013 IEEE International Conference on Signal Processing, Computing and Control (ISPCC); IEEE: New York, NY, USA, 2013; pp. 1–5. [Google Scholar]
  25. Das, D.K.; Chakraborty, C.; Sawaimoon, S.; Maiti, A.K.; Chatterjee, S. Automated identification of keratinization and keratin pearl area from in situ oral histological images. Tissue Cell 2015, 47, 349–358. [Google Scholar] [CrossRef] [PubMed]
  26. Das, D.K.; Bose, S.; Maiti, A.K.; Mitra, B.; Mukherjee, G.; Dutta, P.K. Automatic identification of clinically relevant regions from oral tissue histological images for oral squamous cell carcinoma diagnosis. Tissue Cell 2018, 53, 111–119. [Google Scholar] [CrossRef] [PubMed]
  27. Shi, L.; Liu, W.; Zhang, H.; Xie, Y.; Wang, D. A survey of GPU-based medical image computing techniques. Quant. Imaging Med. Surg. 2012, 2, 188. [Google Scholar]
  28. Srinidhi, C.L.; Ciga, O.; Martel, A.L. Deep neural network models for computational histopathology: A survey. Med. Image Anal. 2021, 67, 101813. [Google Scholar] [CrossRef]
  29. Das, N.; Hussain, E.; Mahanta, L.B. Automated classification of cells into multiple classes in epithelial tissue of oral squamous cell carcinoma using transfer learning and convolutional neural network. Neural Netw. 2020, 128, 47–60. [Google Scholar] [CrossRef]
  30. Panigrahi, S.; Swarnkar, T. Automated Classification of Oral Cancer Histopathology images using Convolutional Neural Network. In 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); IEEE: New York, NY, USA, 2019; pp. 1232–1234. [Google Scholar]
  31. Panigrahi, S.; Das, J.; Swarnkar, T. Capsule network based analysis of histopathological images of oral squamous cell carcinoma. J. King Saud Univ.-Comput. Inf. Sci. 2020, 34, 4546–4553. [Google Scholar] [CrossRef]
  32. Karen, S.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Prabhakar, S.K.; Rajaguru, H. Performance analysis of linear layer neural networks for oral cancer classification. In 2017 6th ICT International Student Project Conference (ICT-ISPC); IEEE: New York, NY, USA, 2017; pp. 1–4. [Google Scholar]
  34. Xi, E. Image feature extraction and analysis algorithm based on multi-level neural network. In 2021 5th International Conference on Computing Methodologies and Communication (ICCMC); IEEE: New York, NY, USA, 2021. [Google Scholar]
  35. Shrestha, A.; Mahmood, A. Review of deep learning algorithms and architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  36. Li, J.; Song, K. Research on Image Classification Based on Deep Learning. In 2021 IEEE/ACIS 19th International Conference on Computer and Information Science (ICIS); IEEE: New York, NY, USA, 2021; pp. 132–136. [Google Scholar]
  37. Yousef, R.; Gupta, G.; Yousef, N.; Khari, M. A holistic overview of deep learning approach in medical imaging. Multimed. Syst. 2022, 28, 881–914. [Google Scholar] [CrossRef]
  38. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef] [PubMed]
  39. Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN variants for computer vision: History, architecture, application, challenges and future scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
  40. Rahman, T.Y.; Mahanta, L.B.; Das, A.K.; Sarma, J.D. Histopathological imaging database for oral cancer analysis. Data Brief 2020, 29, 105114. [Google Scholar] [CrossRef] [PubMed]
  41. Gedraite, E.S.; Hadad, M. Investigation on the effect of a Gaussian Blur in image filtering and segmentation. In Proceedings ELMAR-2011; IEEE: New York, NY, USA, 2011; pp. 393–396. [Google Scholar]
  42. Kashyap, R. Breast cancer histopathological image classification using stochastic dilated residual ghost model. Int. J. Inf. Retr. Res. (IJIRR) 2022, 12, 1–24. [Google Scholar]
  43. Nguyen, Q.H.; Ly, H.-B.; Ho, L.S.; Al-Ansari, N.; Van Le, H.; Tran, V.Q.; Prakash, I.; Pham, B.T. Influence of data splitting on performance of machine learning models in prediction of shear strength of soil. Math. Probl. Eng. 2021, 2021, 4832864. [Google Scholar] [CrossRef]
  44. Qin, X.; Ban, Y.; Wu, P.; Yang, B.; Liu, S.; Yin, L.; Liu, M.; Zheng, W. Improved Image Fusion Method Based on Sparse Decomposition. Electronics 2022, 11, 2321. [Google Scholar] [CrossRef]
  45. Hu, Z.; Zhao, T.V.; Huang, T.; Ohtsuki, S.; Jin, K.; Goronzy, I.N.; Wu, B.; Abdel, M.P.; Bettencourt, J.W.; Berry, G.J.; et al. The transcription factor RFX5 coordinates antigen-presenting function and resistance to nutrient stress in synovial macrophages. Nat. Metab. 2022, 4, 759–774. [Google Scholar] [CrossRef]
  46. Zhao, H.; Ming, T.; Tang, S.; Ren, S.; Yang, H.; Liu, M.; Tao, Q.; Xu, H. Wnt signaling in colorectal cancer: Pathogenic role and therapeutic target. Mol. Cancer 2022, 21, 1–34. [Google Scholar] [CrossRef]
  47. Canesche, M.; Bragança, L.; Neto OP, V.; Nacif, J.A.; Ferreira, R. Google colab cad4u: Hands-on cloud laboratories for digital design. In 2021 IEEE International Symposium on Circuits and Systems (ISCAS); IEEE: New York, NY, USA, 2021; pp. 1–5. [Google Scholar]
  48. Hicks, S.A.; Strümke, I.; Thambawita, V.; Hammou, M.; Riegler, M.A.; Halvorsen, P.; Parasa, S. On evaluation metrics for medical applications of artificial intelligence. Sci. Rep. 2022, 12, 1–9. [Google Scholar] [CrossRef]
  49. Gu, S.; Pednekar, M.; Slater, R. Improve image classification using data augmentation and neural networks. SMU Data Sci. Rev. 2019, 2, 1. [Google Scholar]
Figure 1. (a) Normal cell (100× magnification), (b) cancerous cell (100× magnification), (c) normal cell (400× magnification), (d) cancerous cell (400× magnification).
Figure 1. (a) Normal cell (100× magnification), (b) cancerous cell (100× magnification), (c) normal cell (400× magnification), (d) cancerous cell (400× magnification).
Ijerph 20 02131 g001
Figure 2. (a) Original histopathological oral image; (b) resultant image after Gaussian filter is applied.
Figure 2. (a) Original histopathological oral image; (b) resultant image after Gaussian filter is applied.
Ijerph 20 02131 g002
Figure 3. Proposed Methodology.
Figure 3. Proposed Methodology.
Ijerph 20 02131 g003
Figure 4. Architecture of proposed 10-layer CNN.
Figure 4. Architecture of proposed 10-layer CNN.
Ijerph 20 02131 g004
Figure 5. Performance evaluation using accuracy, F-measure, specificity, recall and precision.
Figure 5. Performance evaluation using accuracy, F-measure, specificity, recall and precision.
Ijerph 20 02131 g005
Figure 6. Comparison analysis of all the models using accuracy and error rate.
Figure 6. Comparison analysis of all the models using accuracy and error rate.
Ijerph 20 02131 g006
Figure 7. Accuracy and loss graph of VGG16.
Figure 7. Accuracy and loss graph of VGG16.
Ijerph 20 02131 g007
Figure 8. Accuracy loss graph of VGG19.
Figure 8. Accuracy loss graph of VGG19.
Ijerph 20 02131 g008
Figure 9. Accuracy and loss graph of Alexnet.
Figure 9. Accuracy and loss graph of Alexnet.
Ijerph 20 02131 g009
Figure 10. Accuracy and loss graph of ResNet50.
Figure 10. Accuracy and loss graph of ResNet50.
Ijerph 20 02131 g010
Figure 11. Accuracy and loss graph of ResNet101.
Figure 11. Accuracy and loss graph of ResNet101.
Ijerph 20 02131 g011
Figure 12. Accuracy and loss graph of Inception Net.
Figure 12. Accuracy and loss graph of Inception Net.
Ijerph 20 02131 g012
Figure 13. Accuracy and loss graph of Mobile Net.
Figure 13. Accuracy and loss graph of Mobile Net.
Ijerph 20 02131 g013
Figure 14. Accuracy and loss graph of the proposed 10-layer CNN model.
Figure 14. Accuracy and loss graph of the proposed 10-layer CNN model.
Ijerph 20 02131 g014
Table 1. Distribution of images.
Table 1. Distribution of images.
CategoryTypeTotal QuantityNormal ImagesCancerous Images
Category-1100× magnification52889439
Category-2400× magnification696201495
Total images =1224290934
Table 2. Detail of image generation from augmentation process.
Table 2. Detail of image generation from augmentation process.
CategoryNo of Original OSCC ImageNo. of New Images Generated from OSCC ImagesNo. of Original Non-Cancerous ImageNo. of New Images Generated from Non-Cancerous Image
Category-14392634891869
Category-249514852012211
Total93441192904080
Table 3. Summary of parameter used in the proposed CNN.
Table 3. Summary of parameter used in the proposed CNN.
LayerOutput ShapeNumber of Kernel/ChannelKernel SizeStridePaddingParameter Generated
Conv2D-1 128 × 128 32 3 × 3 11896
Maxpooling2D-1 64 × 64 32 3 × 3 210
Batch Normalization 64 × 64 32 3 × 3 11128
Conv2D-2 64 × 64 32 3 × 3 119248
Maxpooling2D-2 32 × 32 32 3 × 3 210
Batch Normalization 32 × 32 32 3 × 3 11128
Conv2D-3 32 × 32 64 3 × 3 1118,496
Maxpooling2D-3 64 3 × 3 210
Conv2D-4 16 × 16 64 3 × 3 1136,928
Conv2D-5 16 × 16 128 3 × 3 1173,856
Maxpooling2D-4 8 × 8 128 3 × 3 210
Batch Normalization 8 × 8 128 3 × 3 11512
Conv2D-6 8 × 8 128 3 × 3 1 147,584
Maxpooling2D-5 4 × 4 128 3 × 3 210
Conv2D-7 4 × 4 256 3 × 3 11295,168
Batch Normalization 4 × 4 256 3 × 3 111024
Conv2D-8 4 × 4 256 3 × 3 11590,080
Global Average Pooling-6 1 × 1 256 3 × 3 410
Batch normalization 1 × 1 256 3 × 3 111024
Dense(sigmoid) 1 × 1 1024 3 × 3 11263,168
Batch Normalization 1 × 1 1024 3 × 3 114096
Flatten 1 × 1 1024 3 × 3 110
Dropout 1 × 1 1024 3 × 3 110
Dense(softmax) 1 × 1 2 3 × 3 112050
Total=1,444,386
Table 4. Performance of proposed CNN under different hyperparameters.
Table 4. Performance of proposed CNN under different hyperparameters.
Number of Convolution LayersKernel SizePoolingActivation FunctionOptimizerEpochDropout RateTraining AccuracyValidation
Accuracy
5 5 × 5 AvgReLUAdam100.10.8120.782
5 4 × 4 AvgReLUAdam500.20.8120.772
5 3 × 3 MaxReLUAdam1000.30.8510.813
6 5 × 5 AvgReLUAdam100.10.8110.769
6 4 × 4 AvgReLUAdam500.20.8170.771
6 3 × 3 MaxReLUAdam1000.30.8570.825
7 5 × 5 MaxReLUAdam100.10.8870.850
7 4 × 4 MaxReLUAdam500.20.9350.892
7 3 × 3 MaxReLUAdam1000.30.9820.942
8 5 × 5 MaxReLUAdam100.10.9920.955
8 4 × 4 MaxReLUAdam500.20.9920.969
8 3 × 3 MaxReLUAdam1000.30.9930.978
Table 5. Various performance measures of different models.
Table 5. Various performance measures of different models.
VGG16VGG19ALEXNETRESNET
50
RESNET
101
INCEPTION NETMOBILENETPROPOSED 10-LAYER CNN
Precesion0.740.700.880.910.890.920.930.97
Recall/
Sensitivity
0.740.710.890.920.900.920.930.98
Specificity0.780.740.880.920.880.920.920.97
F measure0.740.700.890.910.900.920.930.97
AUC0.740.700.890.910.900.920.930.97
Accuracy0.740.710.880.910.890.920.930.97
Error rate0.260.290.120.090.110.080.070.03
Table 6. Comparison of different existing models with the proposed model.
Table 6. Comparison of different existing models with the proposed model.
Existing Model in LiteratureMethod UsedDataset UsedAccuracy in %
Navarun et al. (2020) [29]CNNHistopathological oral cavity images97.50
Santisudha et al. (2019) [30]CNNHistopathological oral cavity images96.77
Santisudha et al. (2020) [31]Capsule NetworkHistopathological oral cavity images97.35
Proposed 10-layer CNN modelCustomized CNNHistopathological oral cavity images97.82
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Das, M.; Dash, R.; Mishra, S.K. Automatic Detection of Oral Squamous Cell Carcinoma from Histopathological Images of Oral Mucosa Using Deep Convolutional Neural Network. Int. J. Environ. Res. Public Health 2023, 20, 2131. https://doi.org/10.3390/ijerph20032131

AMA Style

Das M, Dash R, Mishra SK. Automatic Detection of Oral Squamous Cell Carcinoma from Histopathological Images of Oral Mucosa Using Deep Convolutional Neural Network. International Journal of Environmental Research and Public Health. 2023; 20(3):2131. https://doi.org/10.3390/ijerph20032131

Chicago/Turabian Style

Das, Madhusmita, Rasmita Dash, and Sambit Kumar Mishra. 2023. "Automatic Detection of Oral Squamous Cell Carcinoma from Histopathological Images of Oral Mucosa Using Deep Convolutional Neural Network" International Journal of Environmental Research and Public Health 20, no. 3: 2131. https://doi.org/10.3390/ijerph20032131

APA Style

Das, M., Dash, R., & Mishra, S. K. (2023). Automatic Detection of Oral Squamous Cell Carcinoma from Histopathological Images of Oral Mucosa Using Deep Convolutional Neural Network. International Journal of Environmental Research and Public Health, 20(3), 2131. https://doi.org/10.3390/ijerph20032131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop