Next Article in Journal
Endothelial, Vascular and Sympathetic Alterations as Therapeutic Targets in Chronic Heart Failure
Next Article in Special Issue
IFIT2 Depletion Promotes Cancer Stem Cell-like Phenotypes in Oral Cancer
Previous Article in Journal
Repurposing Benztropine, Natamycin, and Nitazoxanide Using Drug Combination and Characterization of Gastric Cancer Cell Lines
Previous Article in Special Issue
Deciphering the Functions of Telomerase Reverse Transcriptase in Head and Neck Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Automated Identification and Visualization of Oral Cancer in Optical Coherence Tomography Images

1
Institute of Modern Optics, Tianjin Key Laboratory of Micro-Scale Optical Information Science and Technology, Nankai University, Tianjin 300350, China
2
Department of Oral Pathology, Tianjin Stomatological Hospital, Hospital of Stomatology, Nankai University, Tianjin 300041, China
3
Department of Oral-Maxillofacial Surgery, Tianjin Stomatological Hospital, Hospital of Stomatology, Nankai University, Tianjin 300041, China
*
Author to whom correspondence should be addressed.
Biomedicines 2023, 11(3), 802; https://doi.org/10.3390/biomedicines11030802
Submission received: 18 January 2023 / Revised: 15 February 2023 / Accepted: 4 March 2023 / Published: 6 March 2023
(This article belongs to the Special Issue Oral Cancer: From Pathophysiology to Novel Therapeutic Approaches)

Abstract

:
Early detection and diagnosis of oral cancer are critical for a better prognosis, but accurate and automatic identification is difficult using the available technologies. Optical coherence tomography (OCT) can be used as diagnostic aid due to the advantages of high resolution and non-invasion. We aim to evaluate deep-learning-based algorithms for OCT images to assist clinicians in oral cancer screening and diagnosis. An OCT data set was first established, including normal mucosa, precancerous lesion, and oral squamous cell carcinoma. Then, three kinds of convolutional neural networks (CNNs) were trained and evaluated by using four metrics (accuracy, precision, sensitivity, and specificity). Moreover, the CNN-based methods were compared against machine learning approaches through the same dataset. The results show the performance of CNNs, with a classification accuracy of up to 96.76%, is better than the machine-learning-based method with an accuracy of 92.52%. Moreover, visualization of lesions in OCT images was performed and the rationality and interpretability of the model for distinguishing different oral tissues were evaluated. It is proved that the automatic identification algorithm of OCT images based on deep learning has the potential to provide decision support for the effective screening and diagnosis of oral cancer.

1. Introduction

Oral cancer is one of the most common cancers in the head and neck [1]. In terms of the pathogenesis of oral cancer, the predominant type of oral cancer is oral squamous cell carcinoma (OSCC) with a long preclinical stage [2]. In addition, precancerous lesions (oral potentially malignant disorder), such as homogeneous leukoplakia and nonhomogeneous leukoplakia, are at risk of malignant transformation [3]. Despite the advancement in targeted cancer therapy, survival rates for oral cancer have remained flat over the last 50 years [4]. Fortunately, the patient’s survival can be improved if the OSCC can be detected and diagnosed early for appropriate treatments [5]. The study indicated that the 5-year survival rate can increase from less 30% to 83% with early detection [6]. Therefore, it is critical that oral cancer can be diagnosed and treated in the pre- or early cancerous stages.
The conventional visual examination is the most commonly screening procedure for oral lesions, but its sensitivity and specificity vary greatly [7]. Auxiliary methods, such as, toluidine blue, auto-fluorescence, or non-linear microscopy have been studied [8,9,10,11,12]. However, there are some limitations, such as the safety assessment of chemiluminescence methods, the lack of three-dimensional (3D) information of fluorescence, or the limited field of view and depth of microscopic methods. While histopathology is still the gold standard, this processing is invasive and time-consuming.
The study has shown that the thickness of oral mucosa (epithelium and lamina propria) is less than 1 mm [13]. For oral cavity imaging with microscopic techniques, their penetration depth is limited, which may not be deep enough to investigate the existence of basement membrane. In contrast, optical coherence tomography (OCT) has the advantages of high-resolution (1–20 μm), real-time and large-depth (1–2 mm) imaging which is suitable for imaging oral mucosa. OCT has been applied in biomedical fields since it was first introduced in 1991 [14], such as ophthalmology [15], cardiology [16], gastroenterology [17], and dermatology [18]. In the oral cavity, studies based on OCT have been attempted to differentiate benign and OSCC by different structural or optical indicators, including the thickness of the epithelium, the intactness of basement membrane, or optical scattering properties [19,20,21]. It has been proved that OCT can enable imaging of oral mucosa and identification of the morphological structures.
Automatic image recognition and classification play an important role in biomedicine. To identify oral lesions automatically, texture feature-based methods were proposed. Krishnan et al. made use of high-order spectra, local binary pattern and laws texture energy from histopathological images to identify oral sub-mucous fibrosis [22]. Thomas et al. used the grey level co-occurrence matrix and grey level run-length for classification of oral cancer in digital camera images [23]. Recently, our laboratory has studied the use of texture features to distinguish salivary gland tumors [24], as well as OSCC [25] in OCT images.
In addition, deep learning has been surprisingly successful in recent years [26,27,28]. In the field of biomedicine, deep learning has been developed for disease classification, object segmentation and image enhancement. Aubreville et al. presented and evaluated an automatic approach for OSCC diagnosis using deep learning on confocal laser endomicroscopy images [29]. Welikala et al. assessed two deep-learning-based computer vision approaches for the automated detection and classification of oral lesions in photographs [30]. However, there is no research on deep-learning-based automatic recognition of oral cancer in OCT images.
The goal of this study is to explore the potential of automatic recognition of oral cancer based on deep learning in OCT images and evaluate the effectiveness by identifying precancerous and cancerous tissues. In addition, feature visualization is also studied to evaluate the rationality and interpretability of the network. It has the great potential to assist clinicians in screening and diagnosis of oral cancer and precancerous lesions.

2. Materials and Methods

2.1. Sample Preparation and Data Acquisition

Fresh tissue samples investigated in this study were obtained from the Tianjin Stomatology Hospital, China. All procedures performed in this study were in accordance with the ethical standards of the Ethics Committee of Tianjin Stomatological Hospital. These samples came from 19 patients who were diagnosed with oral diseases, including leukoplakia with hyperplasia (LEH) and OSCC. The normal and diseased oral tissues were sequentially scanned, and then were fixed and stained with H&E. The slices were evaluated by an experienced pathologist. The details about the OCT system and imaging protocol were described in the previous work [31,32].

2.2. Establishment of the Data Set

Different morphological features of oral tissues were marked in Figure 1. Figure 1a,d show the OCT image and the corresponding histopathologic image of normal mucosa. It can be found that the epithelium (EP) and the lamina propria (LP) are clearly distinguishable due to the different optical scattering intensity, which corresponds well to the histopathological image (Figure 1d). The boundary of EP and LP is called the basement membrane (BM), as shown by the white dashed curve. The typical OCT image and the corresponding histopathologic image of LEH are shown in Figure 1b,e. We can see a boundary (BM) similar to that of normal mucosa from Figure 1b,e. It is worth noting that the thickness of EP is increased and the stratum corneum (SC) can also be observed. In contrast, the epithelial cells of OSCC proliferate maliciously, resulting in the destruction of BM. Moreover, due to the aggregation of cancer cells, the distribution of optical scattering signal appears as cord-like in the OCT image, as indicated by the red arrows in Figure 1c,f.
According to the above analysis of morphological characteristics of different oral tissues in OCT images, OCT images matched with the histopathological images were manually segmented to the appropriate size (256 × 256 pixels) as the regions of interest (ROIs), which contain information unique to different tissues, as shown in the green square boxes in Figure 1. After the segmentation, a total of 13,799 OCT images of ROIs were used to establish the data set. In order to avoid data bias, we randomly selected OCT images from some patients for training and others for test, as described in detail in Table 1.

2.3. CNN Architecture

Three CNNs, including LeNet-5, VGG16, and ResNet18, were used for the classification and identification of these oral tissues (Figure S1). As one of the most basic and earliest proposed deep learning networks, LeNet-5 has a simple network structure and a small number of parameters [33]. There are two convolution layers and three fully connected layers in LeNet-5. The rectified linear unit (ReLU) activation function and max pooling operation are used after each convolutional layer. VGG16 is composed of 13 convolution layers and 3 full connection layers, in which ReLU is used as activation functions after every two convolutional layer and full connected layer [34]. VGG16 changed the convolution mode, set multiple convolution kernels, increased the channel, and reduced the matrix width and height through pooling. With the deepening of layers of VGG16, the amount of computation increases. ResNet18 consists of 18 layers with weights, including the convolutional layers and the fully connected layers. ResNet18 avoided the vanishing gradient and reduced computation amount by skip connections [35].

2.4. Training and Classification

Figure 2 is the flowchart of our experiment. CNNs were firstly trained by using random initialization parameters. To reduce the risk of overfitting, 10-fold cross-validation was performed. Then, the independent test set (never seen by the network before) was used to test the classification performance of different CNNs. In this study, the CNNs were implemented under the PyTorch framework. The batch size is set to 32, the cross-entropy is used as the loss function, and Adam is used as the optimizer with a learning rate of 0.0001, momentum of 0.9.
Further, considering that CNN can perform feature extraction, we used CNN as the feature extractor and machine learning (ML) as the classifier to evaluate the effectiveness in oral tissues classification (CNN + ML). Here, the features were extracted from the last layer before the classification layer (the last fully connected layer) of the pre-trained networks. After that, the feature dimensionality reduction was carried out by the principal component analysis (PCA) algorithm [25]. Finally, three kinds of ML classifiers were used, including decision tree (DT), random forest (RF), and support vector machine (SVM) [25].
In this method, transfer learning was applied and the networks were trained on the ImageNet dataset in advance. Transfer learning can transfer the acquired powerful skills to relevant problems, thus saving time and computing costs [36]. For the classifiers we used, DT model is a kind of tree structure, which is composed of a series of nodes, and each node represents a feature. RF is an algorithm that integrates multiple decision trees through ensemble learning. The random vector is used to generate the ensemble of trees and control the growth of each tree in the ensemble, which can significantly improve the classification accuracy. The number and the depth of tree nodes were used to optimize the best results. Multi-class SVM classifiers with Gaussian radial basis function as the kernel function were employed and the non-linear decision boundary was obtained. The penalty factors C and gamma were optimized for SVM.
In order to apply this method to a common scenario, the algorithms were executed on a desktop computer with an eight-core Intel Xeon 3.5 GHz (E5-1620) processor and a 24 GB random-access memory using the Python programming software (Version 3.7.3).

2.5. Evaluation Indicators

To evaluate the performance of different CNNs and approaches in distinguishing oral tissues, four metrics including sensitivity (Sen), specificity (Spe), precision (Pre), and accuracy (Acc) were calculated.
S e n = T P T P + F N
S p e = T N T N + F P
P r e = T P T P + F P
A c c = T P + T N T P + F P + T N + F N
TP: true positives, FP: false positives, FN: false negatives, TN: true negatives
In addition, receiver operating characteristic curves (ROCs) were plotted and areas under ROC (AUCs) were also calculated. ROCs and AUCs can be used to describe the classification performance of models objectively. The degree of convergence of the networks was determined by the loss value obtained by the loss function (cross entropy loss).

2.6. Visualization

To more directly display classification performance, the predictions of the CNNs were calculated. We extracted 384 overlapping patches from each image. Each patch was input the trained network one by one, and the prediction results were visualized using pseudo-color map.
To enhance interpretability of networks, gradient weighted class activation mapping (Grad-CAM) technique was used to highlight the important regions in the OCT images of oral tissues, which creates the visual explanation for CNNs and helps determine more information about the models when performing detection or prediction work [37].

3. Results

3.1. Identification Using CNN Alone

All the CNNs were trained and tested using PyTorch, which is a deep learning framework enabling fast implementation. After 40 epochs of training, loss values of three kinds of CNNs tend to converge (Figure S2). The results of three CNNs using 10-fold cross-validation were shown in Table S1 and the accuracy of the CNN models was verified.
The performances of three kinds of CNNs were presented in Figure 3a–c, respectively. It is observed that three kinds of CNNs are capable of distinguishing each type of tissue, especially for LEH (AUC = 0.99 for all CNNs). The classification accuracies were further calculated, as shown in Figure 3d. For LeNet-5, the classification accuracies of LEH, normal mucosa, and OSCC are 99.56%, 97.51%, and 93.37%, respectively. For VGG16, the accuracy of each type of tissue is 97.87%, 99.77%, and 82.79%, respectively. For ResNet18, the accuracy of each class is 99.87%, 99.32%, and 77.01%, respectively. The overall accuracies of using LeNet-5, VGG16, and ResNet18 are 96.76%, 91.94%, and 90.43%, respectively.

3.2. Identification Using CNN + ML

Given that ML-based methods often require manual feature extraction and feature selection, it brings about contingency and inconvenience for accurate recognition. To address these issues, we used CNNs as feature extractor, and then used ML as classifier to identify different tissues in OCT images. Figure 4 shows the performance of three classifiers after feature extraction using different networks. For LeNet-5 as feature extractor, the overall accuracies of SVM, DT, and RF are 92.52%, 88.23%, and 91.53%, respectively. For VGG16 as feature extractor, the overall accuracies of SVM, DT, and RF are 91.33%, 89.42%, and 90.52%, respectively. For ResNet18 as feature extractor, the overall accuracies of SVM, DT, and RF are 89.51%, 90.12%, and 91.01%, respectively. The corresponding ROC curves show that CNNs combined with SVM can obtain the best results (Figure S3).
To further evaluate the best strategy, the comparison of using CNNs as feature extraction and SVM as classifier is shown in Figure 5. As a whole, it can be found that SVM combined with LeNet-5 achieved best results, whose overall classification accuracy is 92.52%. Accordingly, the precision, sensitivity, and specificity of identifying normal mucosa, LEH, and OSCC are shown in Table 2.

3.3. Performance Evaluation of Two Strategies

Two classification strategies including the use of CNN alone and the use of CNN combined with ML were evaluated from accuracy (Table 3). It can be found that if only CNN is used, LeNet-5 obtained the highest accuracy of 96.76%; if CNN + ML was used, LeNet-5 combined with SVM achieved the highest accuracy of 92.52%. Therefore, the evaluation between two best strategies was implemented. The confusion matrices were shown in Figure S4. In addition, the statistics analysis of two best strategies was performed (Figure S5). Based on the two-sample student’s t test, there was a statistical significance between the accuracies of LeNet-5 (CNN only) and LeNet-5 combined with SVM (CNN + ML) at p < 0.05.
In addition, the training time spent on both strategies was also assessed (Table 4). For using CNN alone, due to the difference in the number of network structure and parameters, the average time of each epoch of training LeNet-5 is much less than that of training VGG16, and ResNet18. Similarly, LeNet-5 need the least time for the network to converge. For CNN + ML, it took less time to train ML classifiers, although it required extracting features from the CNNs.

3.4. Predictive Visualization

Figure 6 shows the predictive visualization results at the junction between normal mucosa and OSCC using a trained CNN model. Figure 6a shows the imaging area in the photograph of the excised sample. According to the histopathological image (Figure 6b), the normal area and the cancerous area are located on the left and the right sides of the image, respectively. From the corresponding OCT image, as shown in Figure 6c, there is a slight distinction between normal and cancerous regions. We can see that there is a slight BM structure on the left, but not on the right. After the patches are input into the network, the differences between the left and the right of the predicted results can be clearly seen in Figure 6d and are consistent with the histopathological image.

3.5. Grad-CAM Visualization

The interpretability of neural networks using Grad-CAM was also evaluated for making efficient and confident decisions. As shown in Figure 7, different oral tissues showed different characteristics in a trained CNN model. Figure 7a is an OCT image of normal mucosa and the corresponding activation map is shown in Figure 7d. It can be found that the network primarily extracts the EP. The OCT image and corresponding activation map of LEH are shown in Figure 7b,e. It can be found that the thickened EP and LP are highlighted. The OCT image of OSCC is shown in Figure 7c. The cord-like morphological structures can be seen in the OCT image due to the accumulation of cancer nests. From Figure 7f, the highlighted area is mainly the aggregation area of cancer cells, and the neural network pays more attention to this area (the area below indicated by the yellow dotted line). These results are consistent with histological findings as described above. It demonstrated that the network learned different characteristics of oral lesions to distinguish each type of oral tissues.

4. Discussion

We studied deep-learning-based identification of oral precancerous and cancerous lesions in this paper. Firstly, three basic kinds of CNNs were trained and evaluated based on oral OCT image data sets that includes normal mucosa, LEH, and OSCC. Next, to avoid the contingency and inconvenience of traditional machine learning methods when extracting features manually, CNNs were used as the feature extractors. DT, RF, and SVM were trained by using the activations of the last layer before the classification layer. Both of two strategies obtained excellent classification results. In addition, the performance of the networks was further verified by feature visualization in OCT images.
Compared to traditional ML methods, deep learning reduces the dependence on feature extraction. The methods used in deep learning are substantially effective to describe the characteristics of images than texture features. Through comparison on the same dataset, it can be found that using CNN alone especially using LeNet-5 obtained better classification results than that of using CNN + ML, whereas the training time of the former about several hundreds of times longer than that of the latter.
To speed up network training, and expand the data set, we segmented the entire image into a certain size. The images were split according to the following segmentation criteria:
(1)
According to the requirements of the network on the input size, the size of ROI is determined as 256 × 256 pixels, which can speed up network training compared to the whole image being input.
(2)
Each ROI must contain the unique characteristics of oral tissue, for example, the epithelium and lamina propria must be included for normal tissue. It can be found that the size of 256 × 256 pixels can not only include the features of oral tissue, but also effectively reduce the interference of background area.
(3)
In order to effectively use the information in the image, we selected ROI areas in an overlapping approach.
(4)
Areas with poor image quality were discarded, such as areas without focus due to large fluctuation of tissue surface.
Similar to the data augmentation used in conventional deep learning, the ROIs we obtained were used to make the network learn invariable features and prevent the network from learning irrelevant features, thus improving network performance.
According to the high-performance model, we evaluated the oral tissue OCT images to automatically predict and visualize the lesions, which will be more in line with the actual needs and suitable for intuitive judgement between normal and cancerous areas.
In addition, neural networks are often seen as black boxes in disease screening because they provide only the final diagnosis of the subject without any details of the basis of the diagnosis, which brings a major challenge to the application of artificial intelligence in clinical devices. In our study, the Grad-CAM was used to visualize the important regions in the oral tissue OCT images. It can be found that three types of oral tissues showed different characteristics in the deep learning network. Moreover, the aggregation of characteristics can reflect the unique feature of each oral tissue. Through the feature visualization, there is reasonable basis for understanding the model classification and identification.
Although microscopic or histopathological examination of tissue is the gold standard, an accurate result of biopsy may depend on the clinician’s experience and confidence, and the selection of biopsy site. A more accurate diagnosis was achieved via multiple site biopsies and larger volume samples [38], which seems to be more important but makes patients more painful for oral precancerous lesions suspected to be malignant transformation.
In addition, intraoperative frozen section biopsy for surgical margin is a routine procedure after oral cancers are resected en bloc. Surgical margins are usually selected according to surgeons’ estimate for suspicious sites of inadequate resection, which may result in omission of positive margins.
As auxiliary tools, imaging techniques have become indispensable in clinic, where image identification algorithms play an important role [39]. This study extends our prior work in oral cancer, which demonstrated the feasibility of OCT image-based identification of OSCC and normal mucosa by using optical parameters as markers to establish the optical attenuation model and using texture-based ML models [21,25].
It is noted that the robustness of deep-learning-based identification methods is worth further exploration with different OCT systems. Deep-learning-based identification is performed by extracting differential features from OCT images of oral tissues. That is, the OCT images of different lesions contain differentiated morphological features, which are recorded by the OCT system. In this case, if different OCT systems are employed, such as different wavelengths or different bandwidths, then the acquired OCT images contain different features. Therefore, collaboration between different devices to obtain more data and conduct robust research based on deep learning is the next direction.
Fortunately, using the powerful learning capabilities of deep learning and the advantage of high-speed and high-resolution imaging of swept-source OCT system, it laid a foundation for guiding clinicians to screen and resect tumors in real time accurately.

5. Conclusions

In conclusion, the feasibility and validity of automatic recognition strategies for OSCC based on OCT and deep learning have been demonstrated. The interpretability of disease assessment was further investigated by visualizing network feature maps. It is proved that automatic identification methods combining the powerful learning capabilities of deep learning with the advantages of OCT imaging are feasible, which is expected to provide decision support for effective screening and diagnosis of oral cancer and precancerous tissues.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biomedicines11030802/s1, Table S1: The accuracy (%) of identifying oral tissues using 10-fold cross-validation of different CNN models; Figure S1: Schematic architectures of three typical CNNs; Figure S2: The training loss curves of three CNN models; Figure S3: The ROC curves for SVM, DT, and RF as classifiers and LeNet-5, VGG16, and ResNet18 as feature extractors; Figure S4: Confusion matrices of two strategies; Figure S5: Statistical analysis of two strategies based on student’s t test.

Author Contributions

Conceptualization, Z.Y. and Y.L.; methodology, Y.L.; software, Z.Y. and H.P.; validation, J.S., J.Z. and Y.L.; formal analysis, Z.Y., J.S. and J.Z.; investigation, Y.L. and J.Z.; resources, J.S. and J.Z.; data curation, Z.Y. and J.S.; writing—original draft preparation, Z.Y.; writing—review and editing, Y.L., J.S. and J.Z.; visualization, Z.Y., H.P. and J.S.; supervision, Y.L.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61875092, the Science and Technology Support Program of Tianjin, the grant number 17YFZCSY00740, and the Beijing-Tianjin-Hebei Basic Research Cooperation Special Program, grant number 19JCZDJC65300.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Tianjin Stomatological Hospital (ethical code: PH2017-B-003 and approval Date: 30 August 2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data will be available from the corresponding author upon reasonable request.

Acknowledgments

We thank all participants for their participation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Wagle, N.S.; Jemal, A. Cancer statistics, 2023. CA A Cancer J. Clin. 2023, 73, 17–48. [Google Scholar] [CrossRef]
  2. Chen, P.H.; Wu, C.H.; Chen, Y.F.; Yeh, Y.C.; Lin, B.H.; Chang, K.W.; Lai, P.Y.; Hou, M.C.; Lu, C.L.; Kuo, W.C. Combination of structural and vascular optical coherence tomography for differentiating oral lesions of mice in different carcinogenesis stages. Biomed. Opt. Express 2018, 9, 1461–1476. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Amagasa, T.; Yamashiro, M.; Uzawa, N. Oral premalignant lesions: From a clinical perspective. Int. J. Clin. Oncol. 2011, 16, 5–14. [Google Scholar] [CrossRef] [PubMed]
  4. Warnakulasuriya, S. Global epidemiology of oral and oropharyngeal cancer. Oral Oncol. 2009, 45, 309–316. [Google Scholar] [CrossRef]
  5. Tiziani, S.; Lopes, V.; Günther, U.L. Early stage diagnosis of oral cancer using 1H NMR—Based metabolomics. Neoplasia 2009, 11, 269–276, IN7–IN10. [Google Scholar] [CrossRef] [Green Version]
  6. Messadi, D.V. Diagnostic aids for detection of oral precancerous conditions. Int. J. Oral Sci. 2013, 5, 59–65. [Google Scholar] [CrossRef]
  7. Downer, M.C.; Moles, D.R.; Palmer, S.; Speight, P.M. A systematic review of test performance in screening for oral cancer and precancer. Oral Oncol. 2004, 40, 264–273. [Google Scholar] [CrossRef] [PubMed]
  8. Lingen, M.W.; Kalmar, J.R.; Karrison, T.; Speight, P.M. Critical evaluation of diagnostic aids for the detection of oral cancer. Oral Oncol. 2008, 44, 10–22. [Google Scholar] [CrossRef] [Green Version]
  9. Tsai, M.R.; Shieh, D.B.; Lou, P.J.; Lin, C.F.; Sun, C.K. Characterization of oral squamous cell carcinoma based on higher-harmonic generation microscopy. J. Biophotonics 2012, 5, 415–424. [Google Scholar] [CrossRef]
  10. Kumar, P.; Kanaujia, S.K.; Singh, A.; Pradhan, A. In vivo detection of oral precancer using a fluorescence-based, in-house-fabricated device: A Mahalanobis distance-based classification. Lasers Med. Sci. 2019, 34, 1243–1251. [Google Scholar] [CrossRef]
  11. Brouwer de Koning, S.G.; Weijtmans, P.; Karakullukcu, M.B.; Shan, C.; Baltussen, E.J.M.; Smit, L.A.; van Veen, R.L.P.; Hendriks, B.H.W.; Sterenborg, H.J.C.M.; Ruers, T.J.M. Toward assessment of resection margins using hyperspectral diffuse reflection imaging (400–1700 nm) during tongue cancer surgery. Lasers Surg. Med. 2020, 52, 496–502. [Google Scholar] [CrossRef] [PubMed]
  12. Scanlon, C.S.; Van Tubergen, E.A.; Chen, L.-C.; Elahi, S.F.; Kuo, S.; Feinberg, S.; Mycek, M.-A.; D’Silva, N.J. Characterization of squamous cell carcinoma in an organotypic culture via subsurface non-linear optical molecular imaging. Exp. Biol. Med. 2013, 238, 1233–1241. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Kurtzman, J.; Sukumar, S.; Pan, S.; Mendonca, S.; Lai, Y.; Pagan, C.; Brandes, S. The impact of preoperative oral health on buccal mucosa graft histology. J. Urol. 2021, 206, 655–661. [Google Scholar] [CrossRef] [PubMed]
  14. Huang, D.; Swanson, E.A.; Lin, C.P.; Schuman, J.S.; Stinson, W.G.; Chang, W.; Hee, M.R.; Flotte, T.; Gregory, K.; Puliafito, C.A. Optical coherence tomography. Science 1991, 254, 1178–1181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Takusagawa, H.L.; Hoguet, A.; Junk, A.K.; Nouri-Mandavi, K.; Radhakrishnan, S.; Chen, T.C. Swept-source OCT for evaluating the lamina cribrosa: A report by the american academy of ophthalmology. Ophthalmology 2019, 126, 1315–1323. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Yonetsu, T.; Bouma, B.E.; Kato, K.; Fujimoto, J.G.; Jang, I.K. Optical coherence tomography-15 years in cardiology. Circ. J. 2013, 77, 1933–1940. [Google Scholar] [CrossRef] [Green Version]
  17. Tsai, T.H.; Leggett, C.L.; Trindade, A.J.; Sethi, A.; Swager, A.F.; Joshi, V.; Bergman, J.J.; Mashimo, H.; Nishioka, N.S.; Namati, E. Optical coherence tomography in gastroenterology: A review and future outlook. J. Biomed. Opt. 2017, 22, 121716. [Google Scholar] [CrossRef] [Green Version]
  18. Olsen, J.; Holmes, J.; Jemec, G.B. Advances in optical coherence tomography in dermatology-a review. J. Biomed. Opt. 2018, 23, 040901. [Google Scholar] [CrossRef] [Green Version]
  19. Tsai, M.-T.; Lee, H.-C.; Lee, C.-K.; Yu, C.-H.; Chen, H.-M.; Chiang, C.-P.; Chang, C.-C.; Wang, Y.-M.; Yang, C.C. Effective indicators for diagnosis of oral cancer using optical coherence tomography. Opt. Express 2008, 16, 15847–15862. [Google Scholar] [CrossRef]
  20. Adegun, O.K.; Tomlins, P.H.; Hagi-Pavli, E.; McKenzie, G.; Piper, K.; Bader, D.L.; Fortune, F. Quantitative analysis of optical coherence tomography and histopathology images of normal and dysplastic oral mucosal tissues. Lasers Med. Sci. 2012, 27, 795–804. [Google Scholar] [CrossRef]
  21. Yang, Z.; Shang, J.; Liu, C.; Zhang, J.; Liang, Y. Identification of oral cancer in OCT images based on an optical attenuation model. Lasers Med. Sci. 2020, 35, 1999–2007. [Google Scholar] [CrossRef] [PubMed]
  22. Krishnan, M.M.; Venkatraghavan, V.; Acharya, U.R.; Pal, M.; Paul, R.R.; Min, L.C.; Ray, A.K.; Chatterjee, J.; Chakraborty, C. Automated oral cancer identification using histopathological images: A hybrid feature extraction paradigm. Micron 2012, 43, 352–364. [Google Scholar] [CrossRef]
  23. Thomas, B.; Kumar, V.; Saini, S. Texture analysis based segmentation and classification of oral cancer lesions in color images using ANN. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Computing and Control (ISPCC), Solan, India, 26–28 September 2013; pp. 1–5. [Google Scholar]
  24. Yang, Z.; Shang, J.; Liu, C.; Zhang, J.; Liang, Y. Classification of oral salivary gland tumors based on texture features in optical coherence tomography images. Lasers Med. Sci. 2021, 37, 1139–1146. [Google Scholar] [CrossRef] [PubMed]
  25. Yang, Z.; Shang, J.; Liu, C.; Zhang, J.; Liang, Y. Identification of oral squamous cell carcinoma in optical coherence tomography images based on texture features. J. Innov. Opt. Health Sci. 2020, 14, 2140001. [Google Scholar] [CrossRef]
  26. Azam, S.; Rafid, A.K.M.R.H.; Montaha, S.; Karim, A.; Jonkman, M.; De Boer, F. Automated detection of broncho-arterial pairs using CT scans employing different approaches to classify lung diseases. Biomedicines 2023, 11, 133. [Google Scholar] [CrossRef] [PubMed]
  27. Ali, Z.; Alturise, F.; Alkhalifah, T.; Khan, Y.D. IGPred-HDnet: Prediction of immunoglobulin proteins using graphical features and the hierarchal deep learning-based approach. Comput. Intell. Neurosci. 2023, 2023, 2465414. [Google Scholar] [CrossRef]
  28. Hassan, A.; Alkhalifah, T.; Alturise, F.; Khan, Y.D. RCCC_Pred: A novel method for sequence-based identification of renal clear cell carcinoma genes through DNA mutations and a blend of features. Diagnostics 2022, 12, 3036. [Google Scholar] [CrossRef]
  29. Aubreville, M.; Knipfer, C.; Oetter, N.; Jaremenko, C.; Rodner, E.; Denzler, J.; Bohr, C.; Neumann, H.; Stelzle, F.; Maier, A. Automatic classification of cancerous tissue in laserendomicroscopy images of the oral cavity using deep learning. Sci. Rep. 2017, 7, 11979. [Google Scholar] [CrossRef] [Green Version]
  30. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; et al. Automated detection and classification of oral lesions using deep learning for early detection of oral cancer. IEEE Access 2020, 8, 132677–132693. [Google Scholar] [CrossRef]
  31. Li, K.; Yang, Z.; Liang, W.; Shang, J.; Liang, Y.; Wan, S. Low-cost, ultracompact handheld optical coherence tomography probe for in vivo oral maxillofacial tissue imaging. J. Biomed. Opt. 2020, 25, 046003. [Google Scholar] [CrossRef]
  32. Yang, Z.; Shang, J.; Liu, C.; Zhang, J.; Hou, F.; Liang, Y. Intraoperative imaging of oral-maxillofacial lesions using optical coherence tomography. J. Innov. Opt. Health Sci. 2020, 13, 2050010. [Google Scholar] [CrossRef] [Green Version]
  33. Li, T.; Jin, D.; Du, C.; Cao, X.; Chen, H.; Yan, J.; Chen, N.; Chen, Z.; Feng, Z.; Liu, S. The image-based analysis and classification of urine sediments using a LeNet-5 neural network. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2020, 8, 109–114. [Google Scholar] [CrossRef]
  34. Krishnaswamy Rangarajan, A.; Purushothaman, R. Disease classification in eggplant using pre-trained VGG16 and MSVM. Sci. Rep. 2020, 10, 2322. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Odusami, M.; Maskeliūnas, R.; Damaševičius, R.; Krilavičius, T. Analysis of features of alzheimer’s disease: Detection of early stage from functional brain changes in magnetic resonance images using a finetuned ResNet18 network. Diagnostics 2021, 11, 1071. [Google Scholar] [CrossRef]
  36. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef] [Green Version]
  37. Panwar, H.; Gupta, P.K.; Siddiqui, M.K.; Morales-Menéndez, R.; Bhardwaj, P.; Singh, V. A deep learning and grad-CAM based color visualization approach for fast detection of COVID-19 cases using chest X-ray and CT-Scan images. Chaos Solitons Fractals 2020, 140, 110190. [Google Scholar] [CrossRef]
  38. Chen, S.; Forman, M.; Sadow, P.M.; August, M. The diagnostic accuracy of incisional biopsy in the oral cavity. J. Oral Maxillofac. Surg. 2016, 74, 959–964. [Google Scholar] [CrossRef]
  39. Bisht, S.R.; Mishra, P.; Yadav, D.; Rawal, R.; Mercado-Shekhar, K.P. Current and emerging techniques for oral cancer screening and diagnosis: A review. Prog. Biomed. Eng. 2021, 3, 042003. [Google Scholar] [CrossRef]
Figure 1. Morphological characteristics and statistical analysis of oral tissues. The representative OCT images of normal mucosa (a), LEH (b), and OSCC (c) and corresponding histopathological images (df). The ROI indicates 256 × 256 pixels.
Figure 1. Morphological characteristics and statistical analysis of oral tissues. The representative OCT images of normal mucosa (a), LEH (b), and OSCC (c) and corresponding histopathological images (df). The ROI indicates 256 × 256 pixels.
Biomedicines 11 00802 g001
Figure 2. Flowchart of the oral tissue classification experiment.
Figure 2. Flowchart of the oral tissue classification experiment.
Biomedicines 11 00802 g002
Figure 3. The classification evaluation of three kinds of CNNs. (ac) are the ROCs and AUCs of distinguishing three types of tissues using these three CNNs, respectively. (d) is the classification accuracies of three types of tissues with three types of CNNs. TPR: true positive rate, FPR: false positive rate.
Figure 3. The classification evaluation of three kinds of CNNs. (ac) are the ROCs and AUCs of distinguishing three types of tissues using these three CNNs, respectively. (d) is the classification accuracies of three types of tissues with three types of CNNs. TPR: true positive rate, FPR: false positive rate.
Biomedicines 11 00802 g003
Figure 4. The performance comparison of the three classifiers (DT, RF, and SVM) when using LeNet-5 (a), VGG16 (b), and ResNet18 (c) as a feature extractor, respectively.
Figure 4. The performance comparison of the three classifiers (DT, RF, and SVM) when using LeNet-5 (a), VGG16 (b), and ResNet18 (c) as a feature extractor, respectively.
Biomedicines 11 00802 g004
Figure 5. The performance comparison of using SVM as classifier and CNNs as feature extractor. Accuracy (a), sensitivity (b), precision (c), and specificity (d) of classification models using LeNet-5, VGG16, and ResNet18 as feature extractor, respectively.
Figure 5. The performance comparison of using SVM as classifier and CNNs as feature extractor. Accuracy (a), sensitivity (b), precision (c), and specificity (d) of classification models using LeNet-5, VGG16, and ResNet18 as feature extractor, respectively.
Biomedicines 11 00802 g005
Figure 6. OCT imaging and prediction results at the junction between normal mucosa and OSCC. (a) is a photograph of the excised tissue. (b) is the corresponding histopathological image with the normal region on the left and the cancerous region on the right. (c) is the OCT image at the black line of (a). (d) is the corresponding prediction visualization.
Figure 6. OCT imaging and prediction results at the junction between normal mucosa and OSCC. (a) is a photograph of the excised tissue. (b) is the corresponding histopathological image with the normal region on the left and the cancerous region on the right. (c) is the OCT image at the black line of (a). (d) is the corresponding prediction visualization.
Biomedicines 11 00802 g006
Figure 7. Visualization on OCT images of oral tissues using Grad-CAM. (ac) are representative OCT images of normal mucosa, LEH, and OSCC, respectively. (df) are the corresponding activation maps with unique feature aggregations, respectively.
Figure 7. Visualization on OCT images of oral tissues using Grad-CAM. (ac) are representative OCT images of normal mucosa, LEH, and OSCC, respectively. (df) are the corresponding activation maps with unique feature aggregations, respectively.
Biomedicines 11 00802 g007
Table 1. Information of the patients and the partitioning of the data set.
Table 1. Information of the patients and the partitioning of the data set.
DatasetNormal *LEHOSCCTotal
Patients’ number-51419
Age (median [range])-62 (37–73)60 (29–69)
Gender (male/female)-3/27/710/9
Training set
Patients’ number-31013
OCT images2151363939479737
Test set
Patients’ number-246
OCT images1043160114184062
* OCT images of normal mucosa were captured from the normal part of the abnormally excised tissues. The normal area was determined to be at least 1 cm away from abnormal area under the guidance of an experienced surgeon.
Table 2. The precision, sensitivity, and specificity of identifying NM, LEH, and OSCC using SVM combined with LeNet-5.
Table 2. The precision, sensitivity, and specificity of identifying NM, LEH, and OSCC using SVM combined with LeNet-5.
ParameterNMLEHOSCC
Precision (%)87.894.094.5
Sensitivity (%)90.799.586.0
Specificity (%)95.695.897.3
Table 3. Overall accuracy (%) of the two classification strategies.
Table 3. Overall accuracy (%) of the two classification strategies.
ModelClassifierLeNet-5VGG16ResNet18
CNN alone-96.7691.9490.43
CNN + MLDT87.2389.4290.51
RF91.5390.5290.01
SVM92.5291.3389.51
Table 4. Time of training CNNs and machine learning classifiers.
Table 4. Time of training CNNs and machine learning classifiers.
ModelLeNet-5VGG16ResNet18
CNN alone
Each epoch/s22828911618
Convergence/s9120115,64064,720
CNN + ML
Feature extraction/s86710481
DT/s0.577.120.88
RF/s0.271.250.29
SVM/s15221.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Pan, H.; Shang, J.; Zhang, J.; Liang, Y. Deep-Learning-Based Automated Identification and Visualization of Oral Cancer in Optical Coherence Tomography Images. Biomedicines 2023, 11, 802. https://doi.org/10.3390/biomedicines11030802

AMA Style

Yang Z, Pan H, Shang J, Zhang J, Liang Y. Deep-Learning-Based Automated Identification and Visualization of Oral Cancer in Optical Coherence Tomography Images. Biomedicines. 2023; 11(3):802. https://doi.org/10.3390/biomedicines11030802

Chicago/Turabian Style

Yang, Zihan, Hongming Pan, Jianwei Shang, Jun Zhang, and Yanmei Liang. 2023. "Deep-Learning-Based Automated Identification and Visualization of Oral Cancer in Optical Coherence Tomography Images" Biomedicines 11, no. 3: 802. https://doi.org/10.3390/biomedicines11030802

APA Style

Yang, Z., Pan, H., Shang, J., Zhang, J., & Liang, Y. (2023). Deep-Learning-Based Automated Identification and Visualization of Oral Cancer in Optical Coherence Tomography Images. Biomedicines, 11(3), 802. https://doi.org/10.3390/biomedicines11030802

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop