Next Article in Journal
The Role of Female Reproductive Hormones in the Association between Migraine and Breast Cancer: An Unanswered Question
Next Article in Special Issue
Development of New Models of Oral Mucosa to Investigate the Impact of the Structure of Transmembrane Mucin-1 on the Mucosal Pellicle Formation and Its Physicochemical Properties
Previous Article in Journal
Therapeutic Approaches in Pancreatic Cancer: Recent Updates
Previous Article in Special Issue
Neopterin, the Cell-Mediated Immune Response Biomarker, in Inflammatory Periodontal Diseases: A Narrative Review of a More than Fifty Years Old Biomarker
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review

by
Sanjeev B. Khanagar
1,2,*,
Lubna Alkadi
2,3,
Maryam A. Alghilan
2,3,
Sara Kalagi
2,3,
Mohammed Awawdeh
1,2,
Lalitytha Kumar Bijai
2,4,
Satish Vishwanathaiah
5,
Ali Aldhebaib
2,6 and
Oinam Gokulchandra Singh
2,6
1
Preventive Dental Science Department, College of Dentistry, King Saud bin Abdulaziz University for Health Sciences, Riyadh 11426, Saudi Arabia
2
King Abdullah International Medical Research Centre, Ministry of National Guard Health Affairs, Riyadh 11481, Saudi Arabia
3
Restorative and Prosthetic Dental Sciences Department, College of Dentistry, King Saud bin Abdulaziz University for Health Sciences, Riyadh 11426, Saudi Arabia
4
Maxillofacial Surgery and Diagnostic Sciences Department, College of Dentistry, King Saud bin Abdulaziz University for Health Sciences, Riyadh 11426, Saudi Arabia
5
Department of Preventive Dental Sciences, Division of Pediatric Dentistry, College of Dentistry, Jazan University, Jazan 45142, Saudi Arabia
6
Radiological Sciences Program, College of Applied Medical Sciences, King Saud bin Abdulaziz University for Health Sciences, Riyadh 11426, Saudi Arabia
*
Author to whom correspondence should be addressed.
Biomedicines 2023, 11(6), 1612; https://doi.org/10.3390/biomedicines11061612
Submission received: 27 April 2023 / Revised: 27 May 2023 / Accepted: 31 May 2023 / Published: 1 June 2023
(This article belongs to the Special Issue Models for Oral Biology Research 2.0)

Abstract

:
Oral cancer (OC) is one of the most common forms of head and neck cancer and continues to have the lowest survival rates worldwide, even with advancements in research and therapy. The prognosis of OC has not significantly improved in recent years, presenting a persistent challenge in the biomedical field. In the field of oncology, artificial intelligence (AI) has seen rapid development, with notable successes being reported in recent times. This systematic review aimed to critically appraise the available evidence regarding the utilization of AI in the diagnosis, classification, and prediction of oral cancer (OC) using histopathological images. An electronic search of several databases, including PubMed, Scopus, Embase, the Cochrane Library, Web of Science, Google Scholar, and the Saudi Digital Library, was conducted for articles published between January 2000 and January 2023. Nineteen articles that met the inclusion criteria were then subjected to critical analysis utilizing QUADAS-2, and the certainty of the evidence was assessed using the GRADE approach. AI models have been widely applied in diagnosing oral cancer, differentiating normal and malignant regions, predicting the survival of OC patients, and grading OC. The AI models used in these studies displayed an accuracy in a range from 89.47% to 100%, sensitivity from 97.76% to 99.26%, and specificity ranging from 92% to 99.42%. The models’ abilities to diagnose, classify, and predict the occurrence of OC outperform existing clinical approaches. This demonstrates the potential for AI to deliver a superior level of precision and accuracy, helping pathologists significantly improve their diagnostic outcomes and reduce the probability of errors. Considering these advantages, regulatory bodies and policymakers should expedite the process of approval and marketing of these products for application in clinical scenarios.

1. Introduction

Oral cancer (OC) is one of the most common forms of head and neck cancer and continues to have the lowest survival rates worldwide, despite advancements in research and therapy. The prognosis of OC has not significantly improved in recent years, creating a persistent challenge in the biomedical field [1,2]. Cancers of the lip, mouth, and oropharynx, collectively known as oral cancer, are the 13th most prevalent cancer globally [3]. According to the World Health Organization, it is estimated that there were 377,713 new cases and 177,757 deaths from cancers of the lip and oral cavity worldwide in 2020 [3,4]. It is estimated that over 90% of all oral cancers are oral squamous cell carcinomas (OSCCs), which are highly aggressive and have a strong propensity to spread both locally and to other parts of the body [4,5,6].
Early detection of OSCC is vital for successful therapy, increased chances of survival, and lower rates of mortality and morbidity [7]. Microscopy-based histopathological analysis of tissue samples is considered the gold standard for diagnosing and grading oral cancer. However, this approach can be slow and prone to errors, limiting its clinical usefulness [8]. Furthermore, it may lead to subjective discrepancies in interpretation and variability in results. This, in turn, may impact the treatment process [9]. Therefore, there is a need for alternative diagnostic methods that can offer greater accuracy, speed, and standardization to overcome these drawbacks. In recent times, significant efforts have been invested in investigating the potential of artificial intelligence (AI) in improving medical diagnosis. Landini and Othman pioneered and developed an automated method that utilizes morphological reconstruction to report the architectural characteristics of the epithelium for oral cancer diagnosis [10]. Additionally, digital pathology is gaining traction in quantitative analysis as an effective approach by leveraging high-performance computer technology [11,12,13].
Machine learning (ML) techniques identify distinguishable patterns from existing data but rely on human knowledge and efforts to distinguish features. Deep learning (DL), a subset of machine learning that employs artificial neural networks to imitate the human brain’s process of learning, is a recent advancement that can directly extract features from raw images. Both ML and DL algorithms improve the performance of computer-aided diagnostic systems (CAD) with more training samples [14]. As a result, researchers have integrated image processing, pattern recognition, machine learning, and deep learning methods to develop CAD for oral cancer diagnosis. The literature suggests that automated quantification of an oral cancer diagnosis reduces grading conflicts between pathologists [15].
The major advantage of AI is that it reduces the load of manual visualization of slides. It also assists pathologists in fast decision making with better accuracy. Computerized image analysis of tissue slides can obtain information that may be missed with traditional viewing of slides. Precise and accurate histological findings are necessary for early diagnosis, classification, prediction, and specific treatment planning for OC [16]. Various reports have been published describing the application of AI in the early diagnosis, prognosis, and classification of OC [4,5,6,13,14,17,18,19]. This systematic review is exclusively intended to evaluate the performance of AI in oral cancer detection, diagnosis, classification, and prediction using histopathological images.

2. Materials and Methods

2.1. Search Strategy

The authors of this systematic review followed the diagnostic test accuracy guidelines set forth in the Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension (PRISMA-DTA) [20] to ensure methodological quality. The search was conducted based on the PICO (Problem/Patient, Intervention/Indicator, Comparison, and Outcome) criteria, which are presented in Table 1.
An electronic search was conducted using several reputable databases, including PubMed, Scopus, Embase, the Cochrane Library, Web of Science, Google Scholar, and the Saudi Digital Library, for articles published between January 2000 and January 2023. The index words used for the search of the articles were artificial intelligence, automated models, histopathology, histopathological images, slides, hematoxylin and eosin-stained images, oral pathology, biopsy, computational pathology, oral cancer diagnosis, oral cancer detection, oral cancer classification, oral cancer prognosis, cancer cell detection, epithelial layer, keratin pearl, keratinization, artificial neural networks (ANN), supervised learning, unsupervised learning, machine learning, and deep learning. The article search was performed in electronic databases utilizing Boolean operators (AND, OR), along with a filter for years (2000–2023) and a language filter for English. In addition to our electronic search, we also conducted a manual search for relevant research articles and citations. This involved reviewing the reference lists of previously retrieved articles in the college’s library. The search was performed independently by two qualified authors.

2.2. Study Selection

A total of 590 articles were obtained through the electronic database search, and an additional four articles were retrieved through the manual search, resulting in a total of 594 articles for initial consideration. The initial selection of articles was based on their relevance to the research area, as well as the title and abstract. To ensure that there were no duplicated articles, two members not involved in the initial search checked all articles for duplicates, leading to the removal of 322 duplicates. Subsequently, 272 full-text articles were thoroughly reviewed for data selection, with eligibility criteria being applied at this stage.

2.3. Inclusion and Exclusion Criteria

The inclusion criteria for selecting articles were as follows: (a) the article must be original research and must report on AI technology, (b) the article should mention quantifiable values that can be evaluated/analyzed, and (c) articles should mention the data used to evaluate the AI-based models. The study design was not restricted for inclusion in this systematic review. On the other hand, the following types of articles were excluded: (a) those that did not mention AI innovation, (b) unpublished articles or conference papers uploaded online, (c) articles that did not have full-text versions available, and (d) articles available in languages other than English.

2.4. Data Extraction

Following the application of the inclusion criteria, 21 articles were initially selected for analysis. In the second phase, the journal and author details were removed from the articles, and two independent authors who were not involved in the initial search (S.B.K. and L.A.) evaluated them critically. The data from the selected articles were extracted and entered into a Microsoft Excel sheet. These data included publication year, study objectives, AI algorithm types used, and data utilized for training, validation, and testing of the model, as well as the results, conclusions, and recommendations made. However, due to insufficient data to substantiate the results and conclusions of two articles, there were contrasting opinions among the authors regarding their inclusion. After discussing the matter with another expert author (M.A.), a decision was made to exclude them. Therefore, a total of 19 articles were ultimately included for quantitative synthesis, as illustrated in Figure 1. These 19 articles were considered to be potentially eligible articles for this systematic review and were critically analyzed.
A quality assessment of the included articles was carried out using QUADAS-2. It comprises four domains that assess different aspects of study design and reporting, including patient selection, index test, reference standard, and flow and timing [21]. The reliability between the two reviewers was tested using Cohen’s kappa on a sample of articles, displaying an 88% level of agreement. By evaluating each domain for risk of bias and applicability concerns, researchers can identify potential sources of bias and assess the generalizability of the results to different clinical settings and patient populations.

3. Results

After conducting a critical analysis of the 19 articles, qualitative data were extracted. The majority of the studies, which were reported over the past seven years, revealed an increasing trend of reporting on the use of AI for OC diagnosis and prognostic prediction using histopathological images.

3.1. Qualitative Synthesis of the Included Studies

AI technology has been mainly applied for diagnosing OSCC [22,23,24,25,26,27,28], differentiating between normal and malignant conditions [29,30,31,32,33,34,35,36], forming an early diagnosis of OSCC [37], predicting the survival of patients [38,39], and grading the severity of OSCC [40]. In this systematic review, 17 studies were reported using convolutional neural networks (CNNs), while the other two were conducted utilizing capsule networks and a hybrid technique (CNN + ANN), as depicted in Table 2 [15,16].

3.2. Study Characteristics

The extracted features from the studies included information on the authors, publication year, study objectives, AI model development algorithm type, data sources utilized for model training, validation and testing, evaluation accuracy, conclusions, and recommendations.

3.3. Outcome Measures

Task performance efficiency was assessed using various outcome measures, including accuracy, sensitivity, specificity, precision, recall, receiver operating characteristic curve (ROC), area under the curve (AUC), statistical significance, F1 scores, positive predictive value (PPV), negative predictive value (NPV), discrete wavelet transform (DWT), local binary pattern (LBP), fuzzy color histogram (FCH), gray level co-occurrence matrix (GLCM), mean intersection-over-union (mIOU), Dice coefficient, and Jaccard Index.

3.4. Risk of Bias Assessment and Applicability Concern

The quality and risk of bias of the included studies were assessed using the QUADAS-2 assessment tool (Table S1). All studies utilized histopathological images as input for the neural networks, resulting in a low risk of bias for the patient-selection domain in both arms. Standardized techniques were used for data feeding in AI technology, leading to a low risk of bias for flow and timing. All the studies had implemented a highly uniform training system, resulting in a low risk of bias for the index test in both arms of QUADAS-2. Six studies [22,23,25,28,38,39] had used human observations as the reference standard, and hence 30% of the studies reported a high risk in bias assessment and applicability concern. Overall, there was a low risk of bias in both arms across all categories of the included studies. Details regarding the risk of bias assessment and the applicability of the included studies are provided in Supplementary Table S1 and Figure 2.

3.5. Assessment of the Strength of Evidence

The Grading of Recommendations Assessment Development and Evaluation (GRADE) approach was utilized to determine the certainty of the evidence in this systematic review [41]. The certainty of evidence was evaluated based on five domains: risk of bias, inconsistency, indirectness, imprecision, and publication bias. It is classified as very low, low, moderate, or high certainty of evidence. Based on this assessment, the overall included studies in this systematic review demonstrated high certainty of evidence (Table 3).

4. Discussion

The healthcare industry is experiencing the rising power and potential of AI innovations in enhancing the quality of clinical care. AI technologies have the capacity to assist clinicians in minimizing human errors and achieving more precise decision making with superior outcomes than traditional approaches [42]. Deep CNN represents a promising advancement in AI that utilizes algorithms based on neural networks that imitate human neuron mechanisms. CNNs are currently under development as tools to support clinicians in solving various challenges and improving the accuracy of disease detection in radiographic and clinical images [43]. It is important to note that AI technology is not intended to replace clinicians but to aid them in making more precise evaluations and diagnoses of patients [44].
The field of head and neck cancer diagnosis has seen a rapid influx of AI applications that have shown promising results in the preliminary interpretation of medical images [45]. Detecting tumoral changes early on is crucial to ensure timely surgical intervention, subsequent treatment, and, ultimately, increased survival rates. Additionally, this can significantly lower postsurgical morbidity rates and improve quality of life, particularly in cases of invasive and malignant tumors [46].
From its inception, OC is a disease that is often aggressive and resistant to treatment in its more advanced stages. A five-year outlook can vary significantly, with early-stage detection providing an 84% survival rate, while late detection in stages III and IV drops to a 39% survival rate [47]. Additionally, postoperative quality of life declines significantly, particularly for those in advanced stages [48].
The identification of OSCC demands rigorous histopathological investigation, which requires tissue preparation and consumes a considerable amount of time. Moreover, in the case of extensive tumors, multiple samples from different sites need to be removed, and surgeons have to scrutinize the excised margins repeatedly to ensure a clear and cancer-free space [7]. However, prompt and reliable histopathological assessment may not be feasible under certain circumstances [8]. Hence, the implementation of AI as an auxiliary screening tool presents itself as a significant opportunity to enhance diagnostic accuracy noninvasively. In this systematic review, we endeavored to appraise the efficacy of AI in the detection, diagnosis, classification, and prediction of OC from histopathological images [46].

4.1. Effectiveness of AI in the Diagnosis of Oral Cancer

Advanced machine learning algorithms are revolutionizing the field of oncology, particularly in the diagnosis of oral cancer. These models offer a swift and noninvasive method of detecting lesions at a level of accuracy that rivals leading human specialists [49]. While the oral cavity is readily accessible during routine check-ups, many cancers often go unnoticed until they reach advanced stages. Thus, the use of AI offers the potential for a solution to combat the high mortality rates associated with OC [50,51]. Of the pool of articles analyzed in our study, a total of seven studies considered the possibility of utilizing AI-based models for diagnosing OC. One noteworthy study is by Das et al., in which they developed a segmentation method that could identify keratin pearls and quantify the keratinization layer in the oral mucosa of patients with OC. The study utilized a keratinization index (CKI) measure for the automated diagnosis and grading of OSCC. The results were promising, demonstrating the potential for using AI to diagnose oral cancer through quantitative analysis of microscopic images of oral tissues at lower magnification [22].
Hameed et al. introduced a novel technique leveraging the power of machine learning to score immunohistochemistry (IHC) [23]. This involves identifying the most tightly linked feature elements, thereby automating the scoring process. The accuracy of this methodology was compared against manual IHC scores provided by two observers, which were then statistically evaluated. The experiment revealed that the automated IHC score, generated from the top 10 most interdependent feature elements out of a total of 214, has a high correlation coefficient (CC) to the manual scores provided by the observers. Therefore, this finding confirms that the proposed automated IHC scoring mechanism has promising potential in the analysis of IHC-stained tissues [23].
Deif et al. achieved a superior standard of diagnostic efficiency in OSCC patients using histopathological images, employing Inception V3 with BPSO to attain a classification accuracy of 96.3%. This approach not only enhanced accuracy but also effectively curtailed diagnostic costs. While the formidable performance of deep learning algorithms is undeniable, the authors stressed the need for further research to corroborate their efficacy on larger datasets and to compare their results to those of human experts [24].
In a recent study by Yang et al., a deep learning algorithm outperformed pathologists in accurately identifying OSCC in medical images. Moreover, when aided by an AI model, junior pathologists were able to identify OSCC in images 6.26 min faster than when working alone. The model improved the F1 score for both junior (0.922 to 0.957) and senior pathologists (0.936 to 0.946), indicating its potential for improving the accuracy of diagnoses. However, it is important to note that the algorithm was only trained and tested on images from a single institution, and further evaluations are required to determine its generalizability to other populations [25].
In a study by Das et al., a novel two-stage approach was devised to develop improved techniques for the processing of oral histology images. The first stage entails the utilization of a deep CNN comprising 12 convoluted layers of 7 × 7 × 3 channel patches that collate and segment the constituent layers. In the subsequent stage, they detect keratin pearls in the segmented keratin regions by harnessing the power of texture-based features through a Gabor filter-trained random forest. Through this pioneering methodology, they achieved a detection accuracy of 96.88% for the identification of keratin pearls [26].
The second study, conducted by Das et al. in 2019, focused on the development of a computer-aided tool for detecting and delineating nuclei from oral histopathology images to aid in OSCC screening. The authors utilized a combination of texture analysis and machine learning techniques to create an algorithm that could automatically segment nuclei from histological images. The deep learning algorithm underwent training and testing on images from a single institution, potentially limiting the generalizability of its outcomes to other populations [27].
Yoshizawa et al. devised a method using automated machine learning to distinguish OSCC cases based on the YK classification through digital images obtained from histopathological specimens. The method produced strong outcomes overall, with an F-value of 0.87. However, the authors were unable to employ H&E-stained images, even though doing so would be cost-effective and pragmatic. Deep learning could elevate classification accuracy, but acquiring an adequate number of cases remains a crucial roadblock. To augment the precision of classification via deep learning, an ample number of samples would be mandatory. Regrettably, they were unable to obtain the requisite number of cases to achieve this [28].

4.2. Effectiveness of AI in Differentiating Normal from Malignant Regions

Rahman et al. categorized histological slides of oral squamous cell carcinoma into normal (benign) or abnormal (malignant) based on microscopic images. Texture features of the images were analyzed using GLCM, and histogram techniques were used for feature extraction. Linear SVM was used for classification, resulting in 100% accuracy and satisfactory outcomes [29].
Martino et al. carried out a study to evaluate the segmentation performance of four deep networks using the mean intersection-over-union (mIOU) metric. The findings revealed that the U-Net modified with ResNet50 as an encoder performed better than the original U-Net due to its deeper structure. The authors also highlighted the potential of using an automated segmentation algorithm for oral squamous cell carcinoma [30].
According to the research conducted by Das et al., the proposed CNN exhibited superior performance in comparison to other methods in the form of the highest accuracy, precision, and recall metrics. The average accuracy of the pre-trained VGG-19 and Resnet-50 models exceeded 80%, while Alexnet and VGG-16 exhibited the poorest outcomes in terms of accuracy, precision, and recall. [31].
Fraz et al. proposed a deep learning network called Fabnet for simultaneous segmentation of microvessels and nerves in commonly used H&E-stained histology images. The study showed promising results, suggesting that Fabnet can accurately delineate microvessels and nerves, even in challenging cases, outperforming other semantic segmentation networks in terms of accuracy. This achievement may potentially reduce processing time by only segmenting the identified regions of interest. Therefore, Fabnet potentially paves the way for more efficient segmentation of histology images with microvessels and nerves [32].
The system proposed by Rahman et al. demonstrated a high level of accuracy in identifying unknown classes using color, texture, and shape features. Specifically, the system achieved 100% accuracy with color features and high accuracies of 99.4% and 100% with shape and texture features, respectively. This method offers the advantage of accurate classification and computational efficiency, making it a useful tool for automated oral cancer diagnosis or as an assistive tool for physicians to validate their findings [33].
The concatenated model developed by Amin et al. improved the performance for identifying both cancerous and normal images. Similarly, a high AUC value of 0.997 demonstrates that the concatenated model is highly capable of differentiating between the two classes [34].
Panigrahi et al. reported on classifying and grading oral histopathological images through ResNet architecture’s various forms and depths. They obtained optimal results with less computational complexity and a small dataset by using the ResNet13-A as a computer-aided automated model [35].
In another study, Panigrahi et al. used capsule networks (CapsNets), which represent novel machine learning architectures that aim to improve the modeling of hierarchical pose relationships. This is achieved through the use of capsules, which can be defined as collections of neurons that represent an object’s instantiation parameters, such as its pose and orientation. To enable effective routing of the capsule vectors in successive layers, a dynamic routing algorithm is employed. The result is a part-to-whole relationship that is not present in conventional CNNs. CapsNets have been shown to outperform CNNs on the same datasets, demonstrating their enhanced classification capabilities. Additionally, as CapsNets can handle spatial data, they provide better accuracy (97.35%) compared to CNNs (96.77%). The loss function of CapsNets is approximately 0.083 when evaluated on test datasets, while the validation loss of CNNs is 0.132 [36]. This learning curve can be used to evaluate and select the appropriate classifier for a given dataset [36].

4.3. Effectiveness of AI in Early Diagnosis of OC

In the realm of histological analysis for oral cancer detection, researchers have sought to enhance the diagnostic process by employing a hybrid approach. Fati et al. explored two innovative methodologies involving a combination of machine learning techniques, such as CNN, support vector machines (SVM), and ANN [37].
In their first approach, they implemented a two-part hybrid method utilizing the power of CNN models, such as AlexNet and ResNet-18, to extract deep features. These features underwent PCA algorithmic intervention to minimize dimensionality. SVM algorithms were tagged in the second phase of diagnosis to accurately classify these features depicting higher ratings. The researchers observed promising outcomes when they utilized this methodology to diagnose the OSCC dataset.
The second approach adopted the use of ANN techniques, which were grounded on hybrid features obtained by integrating color, texture, and shape features derived from other algorithms, such as discrete wavelet transform (DWT), local binary pattern (LBP), fuzzy color histogram (FCH), and gray level co-occurrence matrix (GLCM). This approach effectively diagnosed histological images of oral cancer cells, targeting early detection with admirable diagnosis rates. By harnessing the hybrid features of the ResNet-18, DWT, LBP, FCH, and GLCM algorithms, the ANN methodology achieved 99.3% accuracy, 99.42% specificity, 99.26% sensitivity, and 99.31% precision, with a 99.39% AUC [37]. The results attained by these novel techniques could greatly improve the current diagnostic capabilities of specialists, aiding diagnostic decision making.

4.4. Effectiveness of AI in Predicting Survival of OC Patients

Traditional statistical techniques, such as the Cox proportional hazard, have been employed to predict the survival of OC patients; however, they prove to be inadequate when it comes to such intricate conditions. A complex “dataset” for oral carcinoma necessitates an AI-based predictive system to yield promising results [17].
Lu et al. developed a classifier based on image analysis that uses quantitative histomorphometric features to assess nuclear shape, size, and texture diversity in clusters of cells from 2 mm OSCC microarray tumor sections that have been digitized using H&E slides. However, this study has some limitations. The image analysis was conducted only on tissue microarrays, which represent a small portion of the complete tumor. Therefore, it may not capture all the morphological variations that exist in the same tumor. The use of whole-slide images may provide a more complete picture of tumors. Additionally, the study’s sample size was relatively small, and some of the well-established clinical and pathological features, such as depth of invasion and nodal extracapsular extension, were not controlled for. A larger, statistically powered retrospective cohort of patients should be analyzed to validate the classifier’s effectiveness on whole-slide images while controlling for all of the established clinical and pathologic features, as well as within well-established patient outcome subgroups [38].
The digital score for TIL abundance was calculated to investigate its potential as a prognostic marker for DFS in OSCC patients in a study conducted by Shaban et al. The TIL abundance score (TILAb) was computed based on the classification of tumor and lymphocytic regions. State-of-the-art CNN-based image classifiers were employed for tissue region classification. The classifiers with the highest (TRC-5) and lowest (TRC-1) patch-level classification accuracy were chosen for further analysis of TIL detection, score computation, and survival analysis. Significant results were obtained by both classifiers [39].

4.5. Effectiveness of AI in the Grading of OC

An innovative method to grade oral tumors using fuzzy cognitive maps (FCM) was developed by Anuradha et al. The FCM model uses eight histopathological features, and an active Hebbian learning (AHL) algorithm is utilized as the supervised learning mechanism to train and improve the model’s grading system. To test the accuracy of the FCM and AHL approach, 123 cases, including 85 normal and 38 abnormal oral tumors, were assessed. The model achieved an accuracy of 90.58% for low-grade oral tumors and 89.47% for high-grade tumors, demonstrating its potential as an important tool in the effective diagnosis and grading of oral tumors. This innovative approach has the potential to improve patient outcomes and reduce morbidity and mortality rates associated with oral tumors [40].

5. Future Perspectives and Limitations

The field of oncology has made significant progress with the incorporation of deep learning algorithms. These intelligent systems assist pathologists in effectively classifying cancer across multiple categories, thereby empowering the oncology team to chart out a treatment module, reducing the operational workload and enhancing disease management. Moreover, deep learning models allow clinicians to classify patients into different risk categories for determining the most suitable treatment [31]. This approach could spare those who do not fall into the high-risk bracket from the more unpleasant side effects of intensive treatments. However, while this has the potential to pave the way for AI to be widely implemented, data privacy and confidentiality remain obstacles in applying AI to clinical oncology [52,53]. Of particular concerns are the potential interpretation errors that could arise while relying heavily on software for medical diagnoses and who should bear ultimate responsibility—the digital intelligence or the skilled doctor [18]. Additionally, AI’s introduction into oncology practice has the potential to impact the patient–doctor relationship and the patient’s autonomy. AI models are designed to assist pathologists and clinicians in clinical decision making. These models have demonstrated outstanding results in performing these tasks. However, when there is discordance between the AI models and the human experts, the latter make the final decision based on their clinical expertise [54]. AI models developed for application in histopathological diagnosis are based on ML and DL, which are subsets of AI. DL models are easy to use in comparison to ML models and have better accuracy, as they are suitable for large sets of data. Moreover, the input of the defined features is not required, as their performance continues to improve with more practice [55]. DL models have an added advantage due to their ability to work on unstructured data and to generate new features with higher quality from datasets without human interventions, which improves their accuracy in diagnosis [56].
One of the limitations of these AI applications is the problem of the interpretability and explainability of the operation of these algorithms. AI models should provide clinicians and patients with a complete understanding of their decisions. However, to date, there has been no unified method for evaluating interpretability [57]. All of these concerns require careful consideration in order to arrive at an appropriate solution.

6. Conclusions

This systematic review provides evidence in support of machine learning models and their significant potential for delivering highly accurate detection of OC with better sensitivity, specificity, and precision. This can significantly aid pathologists in improving their diagnostic outcomes and reducing the probability of errors. Artificial intelligence presents remarkable prospects for the automation of tasks by identifying intricate patterns. Therefore, it is imperative to investigate and promote the integration of AI techniques across disciplines. Such advancements could pave the way for further exploration and research in the future. The diligent scrutiny and surveillance of AI systems to ensure their security, efficiency, and equitability are of critical importance. The growing interest in the development of these advanced models requires an assessment of their quality and application to ensure their safety and cost-effectiveness before being deployed in clinical scenarios. However, considering the advantages, regulatory bodies and policymakers should expedite the process of approval and marketing of these products for application in clinical practice.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/biomedicines11061612/s1, Table S1: Assessment of risk of bias domains and applicability concerns.

Author Contributions

Conceptualization, S.B.K. and L.A.; methodology, M.A. and S.K.; software, M.A.; validation, L.K.B., A.A., and O.G.S.; formal analysis, M.A.; investigation, S.B.K.; resources, L.A. and S.K.; data curation, S.B.K. and S.V.; writing—original draft preparation, S.B.K. and S.V.; writing—review and editing, L.A.; visualization, M.A.A.; supervision, L.A.; project administration, A.A. and O.G.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. García-Pola, M.; Pons-Fuster, E.; Suárez-Fernández, C.; Seoane-Romero, J.; Romero-Méndez, A.; López-Jornet, P. Role of Artificial Intelligence in the Early Diagnosis of Oral Cancer. A Scoping Review. Cancers 2021, 13, 4600. [Google Scholar] [CrossRef] [PubMed]
  2. Rivera, C. Essentials of Oral Cancer. Int. J. Clin. Exp. Pathol. 2015, 8, 11884–11894. [Google Scholar] [PubMed]
  3. World Health Organization. Oral Health. Available online: https://www.who.int/news-room/fact-sheets/detail/oral-health (accessed on 25 April 2023).
  4. López-Cortés, X.A.; Matamala, F.; Venegas, B.; Rivera, C. Machine-Learning Applications in Oral Cancer: A Systematic Review. Appl. Sci. 2022, 12, 5715. [Google Scholar] [CrossRef]
  5. Jubair, F.; Al-karadsheh, O.; Malamos, D.; Al Mahdi, S.; Saad, Y.; Hassona, Y. A Novel Lightweight Deep Convolutional Neural Network for Early Detection of Oral Cancer. Oral Dis. 2021, 8, 1123–1130. [Google Scholar] [CrossRef]
  6. Ghosh, A.; Chaudhuri, D.; Adhikary, S.; Das, A.K.; Chatterjee, K.; Roychowdhury, A.; Das, A.K.; Barui, A. Deep Reinforced Neural Network Model for Cyto-Spectroscopic Analysis of Epigenetic Markers for Automated Oral Cancer Risk Prediction. Available online: https://www.x-mol.net/paper/article/1505978304986726400 (accessed on 22 March 2023).
  7. Chakraborty, D.; Natarajan, C.; Mukherjee, A. Advances in Oral Cancer Detection. Adv. Clin. Chem. 2019, 91, 181–200. [Google Scholar] [CrossRef]
  8. Santana, M.F.; Ferreira, L.C.L. Diagnostic Errors in Surgical Pathology. J. Bras. Patol. Med. Lab. 2017, 53, 124–129. [Google Scholar] [CrossRef]
  9. Kong, J.; Sertel, O.; Shimada, H.; Boyer, K.L.; Saltz, J.H.; Gurcan, M.N. Computer-Aided Evaluation of Neuroblastoma on Whole-Slide Histology Images: Classifying Grade of Neuroblastic Differentiation. Pattern Recognit. 2009, 42, 1080–1092. [Google Scholar] [CrossRef] [PubMed]
  10. Landini, G.; Othman, I.E. Architectural Analysis of Oral Cancer, Dysplastic, and Normal Epithelia. Cytometry 2004, 61A, 45–55. [Google Scholar] [CrossRef] [PubMed]
  11. Warnke-Sommer, J.D.; Ali, H. Evaluation of the Oral Microbiome as a Biomarker for Early Detection of Human Oral Carcinomas. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017. [Google Scholar]
  12. Lavanya, L.; Chandra, J. Oral Cancer Analysis Using Machine Learning Techniques. Int. J. Eng. Res. Technol. 2019, 12, 596–601. [Google Scholar]
  13. Adeoye, J.; Tan, J.Y.; Choi, S.-W.; Thomson, P. Prediction Models Applying Machine Learning to Oral Cavity Cancer Outcomes: A Systematic Review. Int. J. Med. Inform. 2021, 154, 104557. [Google Scholar] [CrossRef]
  14. Mahmood, H.; Shaban, M.; Indave, B.I.; Santos-Silva, A.R.; Rajpoot, N.; Khurram, S.A. Use of Artificial Intelligence in Diagnosis of Head and Neck Precancerous and Cancerous Lesions: A Systematic Review. Oral Oncol. 2020, 110, 104885. [Google Scholar] [CrossRef]
  15. Eid, R.A.; Landini, G. Quantification of the Global and Local Complexity of the Epithelial-Connective Tissue Interface of Normal, Dysplastic, and Neoplastic Oral Mucosae Using Digital Imaging. Pathol.—Res. Pract. 2003, 199, 475–482. [Google Scholar] [CrossRef] [PubMed]
  16. Kayser, K.; GĂśrtler, J.; Bogovac, M.; Bogovac, A.; Goldmann, T.; Vollmer, E.; Kayser, G. AI (Artificial Intelligence) in Histopathology—From Image Analysis to Automated Diagnosis. Folia Histochem. Cytobiol. 2010, 47, 355–361. [Google Scholar] [CrossRef] [PubMed]
  17. Khanagar, S.B.; Naik, S.; Al Kheraif, A.A.; Vishwanathaiah, S.; Maganur, P.C.; Alhazmi, Y.; Mushtaq, S.; Sarode, S.C.; Sarode, G.S.; Zanza, A.; et al. Application and Performance of Artificial Intelligence Technology in Oral Cancer Diagnosis and Prediction of Prognosis: A Systematic Review. Diagnostics 2021, 11, 1004. [Google Scholar] [CrossRef]
  18. Al-Rawi, N.; Sultan, A.; Rajai, B.; Shuaeeb, H.; Alnajjar, M.; Alketbi, M.; Mohammad, Y.; Shetty, S.R.; Mashrah, M.A. The Effectiveness of Artificial Intelligence in Detection of Oral Cancer. Int. Dent. J. 2022, 72, 436–447. [Google Scholar] [CrossRef] [PubMed]
  19. Kim, J.-S.; Kim, B.G.; Hwang, S.H. Efficacy of Artificial Intelligence-Assisted Discrimination of Oral Cancerous Lesions from Normal Mucosa Based on the Oral Mucosal Image: A Systematic Review and Meta-Analysis. Cancers 2022, 14, 3499. [Google Scholar] [CrossRef]
  20. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. Br. Med. J. 2021, 372, n71. [Google Scholar] [CrossRef]
  21. Whiting, P.F.; Rutjes, A.W.S.; Westwood, M.E.; Mallett, S.; Deeks, J.J.; Reitsma, J.B.; Leeflang, M.M.G.; Sterne, J.A.C.; Bossuyt, P.M.M.; QUADAS-2 Group. QUADAS-2: A Revised Tool for the Quality Assessment of Diagnostic Accuracy Studies. Ann. Intern. Med. 2011, 155, 529–536. [Google Scholar] [CrossRef]
  22. Das, D.K.; Chakraborty, C.; Sawaimoon, S.; Maiti, A.K.; Chatterjee, S. Automated Identification of Keratinization and Keratin Pearl Area from in Situ Oral Histological Images. Tissue Cell 2015, 47, 349–358. [Google Scholar] [CrossRef]
  23. Hameed, K.A.S.; Banumathi, A.; Ulaganathan, G. Cell Nuclei Classification and Immunohistochemical Scoring of Oral Cancer Tissue Images: Machine-Learning Approach. Asian J. Res. Soc. Sci. Humanit. 2016, 6, 732. [Google Scholar] [CrossRef]
  24. Deif, M.A.; Attar, H.; Amer, A.; Elhaty, I.A.; Khosravi, M.R.; Solyman, A.A.A. Diagnosis of Oral Squamous Cell Carcinoma Using Deep Neural Networks and Binary Particle Swarm Optimization on Histopathological Images: An AIoMT Approach. Comput. Intell. Neurosci. 2022, 2022, 6364102. [Google Scholar] [CrossRef] [PubMed]
  25. Yang, S.Y.; Li, S.H.; Liu, J.L.; Sun, X.Q.; Cen, Y.Y.; Ren, R.Y.; Ying, S.C.; Chen, Y.; Zhao, Z.H.; Liao, W. Histopathology-Based Diagnosis of Oral Squamous Cell Carcinoma Using Deep Learning. J. Dent. Res. 2022, 101, 1321–1327. [Google Scholar] [CrossRef] [PubMed]
  26. Das, D.K.; Bose, S.; Maiti, A.K.; Mitra, B.; Mukherjee, G.; Dutta, P.K. Automatic Identification of Clinically Relevant Regions from Oral Tissue Histological Images for Oral Squamous Cell Carcinoma Diagnosis. Tissue Cell 2018, 53, 111–119. [Google Scholar] [CrossRef]
  27. Das, D.; Koley, S.; Bose, S.; Maiti, A.; Mitra, B.; Mukherjee, G.; Dutta, P. Computer Aided Tool for Automatic Detection and Delineation of Nucleus from Oral Histopathology Images for OSCC Screening. Appl. Soft Comput. 2019, 83, 105642. [Google Scholar] [CrossRef]
  28. Yoshizawa, K.; Ando, H.; Kimura, Y.; Kawashiri, S.; Yokomichi, H.; Moroi, A.; Ueki, K. Automatic Discrimination of Yamamoto-Kohama Classification by Machine Learning Approach for Invasive Pattern of Oral Squamous Cell Carcinoma Using Digital Microscopic Images: A Retrospective Study. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2022, 133, 441–452. [Google Scholar] [CrossRef] [PubMed]
  29. Rahman, T.Y.; Mahanta, L.B.; ChakrabortyH, C.; Das, A.K.; Sarma, J.D. Textural Pattern Classification for Oral Squamous Cell Carcinoma. J. Microsc. 2017, 269, 85–93. [Google Scholar] [CrossRef]
  30. Martino, F.; Bloisi, D.D.; Pennisi, A.; Fawakherji, M.; Ilardi, G.; Russo, D.; Nardi, D.; Staibano, S.; Merolla, F. Deep Learning-Based Pixel-Wise Lesion Segmentation on Oral Squamous Cell Carcinoma Images. Appl. Sci. 2020, 10, 8285. [Google Scholar] [CrossRef]
  31. Das, N.; Hussain, E.; Mahanta, L.B. Automated Classification of Cells into Multiple Classes in Epithelial Tissue of Oral Squamous Cell Carcinoma Using Transfer Learning and Convolutional Neural Network. Neural Netw. 2020, 128, 47–60. [Google Scholar] [CrossRef]
  32. Fraz, M.M.; Khurram, S.A.; Graham, S.; Shaban, M.; Hassan, M.; Loya, A.; Rajpoot, N.M. FABnet: Feature Attention-Based Network for Simultaneous Segmentation of Microvessels and Nerves in Routine Histology Images of Oral Cancer. Neural Comput. Appl. 2019, 32, 9915–9928. [Google Scholar] [CrossRef]
  33. Rahman, T.Y.; Mahanta, L.B.; Das, A.K.; Sarma, J.D. Automated Oral Squamous Cell Carcinoma Identification Using Shape, Texture and Color Features of Whole Image Strips. Tissue Cell 2020, 63, 101322. [Google Scholar] [CrossRef]
  34. Amin, I.; Zamir, H.; Khan, F.F. Histopathological Image Analysis for Oral Squamous Cell Carcinoma Classification Using Concatenated Deep Learning Models. medRxiv 2021. [Google Scholar] [CrossRef]
  35. Panigrahi, S.; Bhuyan, R.; Kumar, K.; Nayak, J.; Swarnkar, T. Multistage Classification of Oral Histopathological Images Using Improved Residual Network. Math. Biosci. Eng. 2021, 19, 1909–1925. [Google Scholar] [CrossRef]
  36. Panigrahi, S.; Das, J.; Swarnkar, T. Capsule Network Based Analysis of Histopathological Images of Oral Squamous Cell Carcinoma. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 4546–4553. [Google Scholar] [CrossRef]
  37. Fati, S.M.; Senan, E.M.; Javed, Y. Early Diagnosis of Oral Squamous Cell Carcinoma Based on Histopathological Images Using Deep and Hybrid Learning Approaches. Diagnostics 2022, 12, 1899. [Google Scholar] [CrossRef] [PubMed]
  38. Lu, C.; Lewis, J.S.; Dupont, W.D.; Plummer, W.D.; Janowczyk, A.; Madabhushi, A. An Oral Cavity Squamous Cell Carcinoma Quantitative Histomorphometric-Based Image Classifier of Nuclear Morphology Can Risk Stratify Patients for Disease-Specific Survival. Mod. Pathol. 2017, 30, 1655–1665. [Google Scholar] [CrossRef] [PubMed]
  39. Shaban, M.; Khurram, S.A.; Fraz, M.M.; Alsubaie, N.; Masood, I.; Mushtaq, S.; Hassan, M.; Loya, A.; Rajpoot, N.M. A Novel Digital Score for Abundance of Tumour Infiltrating Lymphocytes Predicts Disease Free Survival in Oral Squamous Cell Carcinoma. Sci. Rep. 2019, 9, 13341. [Google Scholar] [CrossRef]
  40. Anuradha, K.; Uma, K.P. Histological Grading of Oral Tumors Using Fuzzy Cognitive Map. Biomed. Pharmacol. J. 2017, 10, 1695–1700. [Google Scholar] [CrossRef]
  41. Granholm, A.; Alhazzani, W.; Møller, M.H. Use of the GRADE Approach in Systematic Reviews and Guidelines. Br. J. Anaesth. 2019, 123, 554–559. [Google Scholar] [CrossRef]
  42. McKinney, S.M.; Sieniek, M.; Godbole, V.; Godwin, J.; Antropova, N.; Ashrafian, H.; Back, T.; Chesus, M.; Corrado, G.C.; Darzi, A.; et al. International Evaluation of an AI System for Breast Cancer Screening. Nature 2020, 577, 89–94. [Google Scholar] [CrossRef]
  43. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  44. Warin, K.; Limprasert, W.; Suebnukarn, S.; Jinaporntham, S.; Jantana, P.; Vicharueang, S. AI-Based Analysis of Oral Lesions Using Novel Deep Convolutional Neural Networks for Early Detection of Oral Cancer. PLoS ONE 2022, 17, e0273508. [Google Scholar] [CrossRef] [PubMed]
  45. Derlin, T.; Grünwald, V.; Steinbach, J.; Wester, H.-J.; Ross, T.L. Molecular Imaging in Oncology Using Positron Emission Tomography. Dtsch. Aerzteblatt Online 2018, 115, 175–181. [Google Scholar] [CrossRef] [PubMed]
  46. Ramezani, K.; Tofangchiha, M. Oral Cancer Screening by Artificial Intelligence-Oriented Interpretation of Optical Coherence Tomography Images. Radiol. Res. Pract. 2022, 2022, 1614838. [Google Scholar] [CrossRef] [PubMed]
  47. Howlader, N.; Noone, A.M.; Krapcho, M.; Miller, D.; Bishop, K.; Kosary, C.L.; Yu, M.; Ruhl, J.; Tatalovich, Z.; Mariotto, A. Lewis Cancer Statistics Review, 1975–2014—SEER Statistics. Available online: https://seer.cancer.gov/csr/1975_2014 (accessed on 22 March 2023).
  48. Fu, Q.; Chen, Y.; Li, Z.; Jing, Q.; Hu, C.; Liu, H.; Bao, J.; Hong, Y.; Shi, T.; Li, K.; et al. A Deep Learning Algorithm for Detection of Oral Cavity Squamous Cell Carcinoma from Photographic Images: A Retrospective Study. EClinicalMedicine 2020, 27, 100558. [Google Scholar] [CrossRef]
  49. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; et al. Automated Detection and Classification of Oral Lesions Using Deep Learning for Early Detection of Oral Cancer. IEEE Access 2020, 8, 132677–132693. [Google Scholar] [CrossRef]
  50. Ilhan, B.; Guneri, P.; Wilder-Smith, P. The Contribution of Artificial Intelligence to Reducing the Diagnostic Delay in Oral Cancer. Oral Oncol. 2021, 116, 105254. [Google Scholar] [CrossRef]
  51. Hegde, S.; Ajila, V.; Zhu, W.; Zeng, C. Artificial Intelligence in Early Diagnosis and Prevention of Oral Cancer. Asia-Pac. J. Oncol. Nurs. 2022, 9, 100133. [Google Scholar] [CrossRef]
  52. Alabi, R.O.; Bello, I.O.; Youssef, O.; Elmusrati, M.; Mäkitie, A.A.; Almangush, A. Utilizing Deep Machine Learning for Prognostication of Oral Squamous Cell Carcinoma—A Systematic Review. Front. Oral Health 2021, 2, 44. [Google Scholar] [CrossRef]
  53. Alabi, R.O.; Vartiainen, T.; Elmusrati, M. Machine Learning for Prognosis of Oral Cancer: What Are the Ethical Challenges? Proc. Conf. Technol. Ethics 2020, 2373, 1–21. [Google Scholar]
  54. Reis-Filho, J.S.; Kather, J.N. Overcoming the Challenges to Implementation of Artificial Intelligence in Pathology. J. Natl. Cancer Inst. 2023, djad048. [Google Scholar] [CrossRef]
  55. Ahmed, A.A.; Abouzid, M.; Kaczmarek, E. Deep Learning Approaches in Histopathology. Cancers 2022, 14, 5264. [Google Scholar] [CrossRef] [PubMed]
  56. Shmatko, A.; Ghaffari Laleh, N.; Gerstung, M.; Kather, J.N. Artificial Intelligence in Histopathology: Enhancing Cancer Research and Clinical Oncology. Nat. Cancer 2022, 3, 1026–1038. [Google Scholar] [CrossRef] [PubMed]
  57. Farah, L.; Murris, J.; Borget, I.; Guilloux, A.; Martelli, N.; Katsahian, S.I. Assessment of Performance, Interpretability, and Explainability in Artificial Intelligence–Based Health Technologies: What Healthcare Stakeholders Need to Know. Mayo Clin. Proc. Digit. Health 2023, 1, 120–138. [Google Scholar] [CrossRef]
Figure 1. PRISMA 2020 flow diagram for new systematic reviews that included searches of databases, registers, and other sources.
Figure 1. PRISMA 2020 flow diagram for new systematic reviews that included searches of databases, registers, and other sources.
Biomedicines 11 01612 g001
Figure 2. QUADAS-2 assessment of the individual risk of bias domains and applicability concerns.
Figure 2. QUADAS-2 assessment of the individual risk of bias domains and applicability concerns.
Biomedicines 11 01612 g002
Table 1. Description of the PICO (P = Population, I = Intervention, C = Comparison, O = Outcome) elements.
Table 1. Description of the PICO (P = Population, I = Intervention, C = Comparison, O = Outcome) elements.
Research QuestionWhat is the Performance of the Artificial Intelligence Models That Have Been Widely Used in Oral Cancer Detection, Diagnosis, Classification, and Prediction Using Histopathological Images?
Population Patients who underwent investigations for oral cancer (histological images).
InterventionAI-based models designed for oral cancer diagnosis, classification, and prediction of prognosis.
ComparisonExpert opinions and reference standards.
OutcomeMeasurable or predictive outcomes, such as accuracy, sensitivity, specificity, precision, recall, receiver operating characteristic curve (ROC), area under the curve (AUC), statistical significance, F1 scores, positive predictive value (PPV), negative predictive value (NPV), discrete wavelet transform (DWT), local binary pattern (LBP), fuzzy color histogram (FCH), gray level co-occurrence matrix (GLCM), mean intersection-over-union (mIOU), Dice coefficient, and Jaccard index.
Table 2. Details of the studies that have used AI-based models in oral cancer detection, diagnosis, classification, and prediction using histopathological images.
Table 2. Details of the studies that have used AI-based models in oral cancer detection, diagnosis, classification, and prediction using histopathological images.
Sl NoAuthorsYear of PublicationStudy DesignAlgorithm ArchitectureObjective of the StudyNo. of Images/Photographs for TestingStudy FactorModalityComparison if anyEvaluation Accuracy/Average Accuracy/Statistical SignificanceResults
(+) Effective,
(−) Non-Effective
(N) Neutral
OutcomesAuthors’ Suggestions/Conclusions
1Das DK et al. [22]2015Observational studyCNNTo develop a computer-assisted quantitative microscopic methodology for automated identification of keratinization and keratin pearl areas from in situ oral histological images10 OSCC patients’ oral histological slidesDiagnosis of OSCCHistopathological imagesManual experts95.08% segmentation accuracy(+) EffectiveThis provides a computer-aided diagnostic framework for microscopic image-based OSCC diagnosis, which can assist clinicians/pathologists for rapid evaluation.The proposed methodology would be able to provide more robust performance when the sample size is large and can also be recommended as one of the tele-pathology applications.
2Hameed KA et al. [23]2016Observational studyCNNTo develop automatic IHC scoring of p53-immunostained tissue images of oral cancer400 sub-regions of tissue imagesDiagnosis of OSCCHistopathological imagesState-of-the-art methods, including intensity and texture features
(manual)
Classification accuracy of 96.09% achieved by the proposed method for LDA classifiers(+) EffectiveThe automatic scoring methods presented have high potential in IHC image analysis.It helps the pathologist during the diagnostic and prognostic evaluation of oral cancer.
3Deif MA et al. [24]2022Observational studyCNNTo diagnose OSCC using deep neural networksHistopathological images of 230 individualsDiagnosis of OSCCHistopathological imagesVGG16, AlexNet, ResNet50, and
Inception V3
Accuracy of 96.3% was obtained when using Inception V3 with BPSO.(+) EffectiveBest classification accuracy of 96.3% was obtained when using Inception V3 with BPSO.This approach significantly contributes to improve the diagnostic efficiency of OCSCC patients while reducing diagnostic costs.
4Yang SY et al. [25]2022Observational studyCNNTo develop a custom-made deep learning model to assist pathologists in detecting OSCC from histopathology images2025 imagesDiagnosis of OSCCHistopathological images3 junior pathologists, 3 senior pathologists, and 1 chief pathologistSensitivity of 0.98, specificity of 0.92, positive predictive value of 0.924, negative predictive value of 0.978, and F1 score of 0.951(+) EffectiveThe results demonstrated that the automated deep learning method could evaluate OSCC approximately 249 times faster than a junior pathologist.These findings indicate that deep learning can improve the accuracy and speed of OSCC diagnosis from histopathology images.
5Das DK et al. [26]2018Observational studyCNNTo identify clinically relevant regions from oral tissue histological images for OSCC diagnosis42 tissue slidesComputer-aided diagnosis and screening of oral cancersHistopathological imagesGabor filter-trained random forestEpithelial layer segmentation: 98.42% segmentation accuracy, 97.76% sensitivity, 90.63% Jaccard index, and 95.03% Dice coefficient.
Keratin pearls:
98.05% segmentation accuracy, 71.87% Jaccard
index, and 75.19% Dice coefficient.
(+) EffectiveProposed approach is good enough to extract epithelial, subepithelial, and keratin regions from oral histological images.Segmentation of epithelial and subepithelial layers and detection of keratin pearls can be utilized for oral precancerous screening and OSCC grading, respectively.
6Das DK et al. [27]2019Observational studyCNNTo develop a two-stage computational pipeline for automatic detection of nucleus and its segmentation from oral histology images42 tissue slidesComputer-aided diagnosis of OSCCHistopathological imagesChan–Vese model94.22% Dice coefficient, 89.38% Jaccard index, 88.87% recall, and 82.03% precision(+) EffectiveThe proposed segmentation methodology performed well, with 94.22% Dice coefficient, 89.38% Jaccard
index, 97.56% precision, and 91.58% recall.
This is the first attempt on oral tissue histology image computation for joint nucleus detection and segmentation to diagnose OSCC.
7Yoshizawa K et al. [28]2022Observational studyCNNTo determine the mode of invasion based on digital images of the invasive front of an OSCC.101 digitized photographic imagesDiagnosis of OSCCHistopathological imagesYamamoto–Kohama grades (1, 2, 3, 4C, 4D) determined by a human oral and maxillofacial surgeonF-measure value of 87%(+) EffectiveThese results suggest that the output of the classifier was very similar to the judgments of the clinician.This system may be valuable for diagnostic support to provide an accurate determination of the mode of invasion.
8Rahman TY et al. [29]2019Observational studyCNNTo develop a CAD system for OSCC classification using textural features on real histopathologic images134 images with
normal tissue and 135 images with malignant tissue
Differentiating normal
and malignant
Histopathological imagesLinear support vector machine (SVM)100% accuracy
AUC = 0.92
(+) EffectiveThe linear support vector machine classifier
provided 100%
accuracy for the automated diagnosis of oral cancer.
It can be used to
assist clinicians in the rapid evaluation and differentiation of
tumorous lesions and normal tissue.
9Martino F et al. [30]2020Observational studySSNsTo compare four different deep learning-based architectures for oral cancer segmentation188 imagesDifferentiating normal
and malignant areas
Histopathological imagesSegNet, U-Net, U-Net with VGG16 encoder, and U-Net with ResNet50 encodermIOU
SegNet = 0.54
U-Net = 0.57
U-Net with VGG16 encoder = 0.62
U-Net with ResNet50 encoder = 0.63
(+) EffectiveThe deeper network, U-Net modified
with ResNet50 as the encoder, performed better than the original U-Net (having a more shallow encoder).
This will help those who work in generalist diagnostic centres, not specialized in the diagnosis of an infrequent but extremely lethal disease.
10Das N et al.
[31]
2020Observational studyCNNTo classify OSCC into its four classes as per Broder’s system of histological grading156 slide imagesDifferentiation of malignant lesions in biopsy imagesHistopathological imagesAlexnet, Resnet-50, VGG
16, and VGG 19
Accuracy of 97.5%.(+) EffectiveHighest classification accuracy of 92.15% was achieved with the Resnet-50 model. The proposed CNN model outperformed the transfer learning approaches, displaying an accuracy of 97.5%.It can be concluded that the proposed CNN-based multi-class grading method of OSCC could be used for the diagnosis of patients with OSCC.
11Fraz MM et al. [32]2020Observational studyCNNTo propose a deep network for simultaneous segmentation of microvessels and nerves in routinely used H&E-stained histology images7780 imagesDifferentiating normal
from malignant areas
Histopathological imagesFCN-8, U-Net, Segnet,
and DeepLabv3
Accuracies of 96.3% and 97.05% for nerves and blood vessels(+) EffectiveThe proposed network
outperformed the current deep neural networks used for semantic segmentation.
The proposed network also provides robust segmentation performance when applied to the full digital whole slide image.
12Rahman TY et al. [33]2021Observational studyCNNTo propose an automated efficient computer-aided system to distinguish normal from malignant OSCC categories42 slidesClassification of cell nuclei into normal and malignant categoriesHistopathological imagesDecision tree classifier,
SVM, and logistic regression
99.4% accuracy using decision tree classifier, 100%
accuracy using both SVM and logistic regression, and 100% accuracy using SVM
(+) EffectiveThe in-depth analysis showed SVM and linear discriminant classifiers provided the best results for texture and color features, respectively.This system is fast, cost-effective, and accurate. Hence, physicians can use it in their daily clinical screening as an assistant diagnostic tool.
13Amin I
et al. [34]
2021Observational studyCNNTo propose an automated classification of cancerous oral histo pathological images290 normal and 934
cancerous oral histo pathological images
Differentiating normal
and malignant areas
Histopathological imagesVGG16, InceptionV3, and Resnet5096.66% accuracy,
95.16% precision, 98.33% recall, and
95.00% specificity; concatenated model AUC = 0.997
(+) EffectiveThe concatenated model yielded the best results and outperformed the individual models.These results demonstrate that the concatenated model can effectively replace the use of a single DL architecture.
14Panigrahi S et al. [35]2022Observational studyCNNTo propose three ResNet architectures for the multistage classification of OSCC into benign and malignant400 image patchesDifferentiating normal
from malignant
Histopathological imagesResNet-based model97.59% accuracy(+) EffectiveThe Optimal ResNet model (ResNet13-A) was chosen as the best model, which is an automated computer-aided method to obtain high-performance results with less computational complexity and small datasets.The proposed ResNet model is an efficient model for detecting multistage oral cancer, and it can be utilized as a diagnostic tool to help physicians in daily clinical screening.
15Panigrahi S et al. [36]2022Observational studyCapsule networkTo classify oral cancer using a deep learning technique known as capsule network to discriminate between cancerous and non-cancerous images82 malignant and 68 benign imagesDifferentiating normal
and malignant areas
Histopathological imagesRegular CNN model97.78% sensitivity,
96.92% specificity, and 97.35% accuracy
(+) EffectiveCapsule networks have better capabilities in capturing the pose information and spatial relationship and can better discriminate between cancerous and non-cancerous images compared to the CNN model.The proposed system can be extended to classify the different stages of oral cancer in the future.
16Fati SM et al. [37]2022Observational studyCNN
ANN
To achieve satisfactory results for the early diagnosis of OSCC by applying hybrid techniques based on fused features5192 imagesEarly diagnosis of OSCCHistopathological imagesCNN models (AlexNet and
ResNet-18) and
SVM algorithm
ANN models
(ResNet-18, DWT, LBP, FCH, and GLCM)
ResNet-18, DWT, LBP, FCH, and GLCM achieved an accuracy of 99.3%, specificity of 99.42%, sensitivity of 99.26%, precision of 99.71%, and AUC of 99.39%(+) EffectiveThe ANN algorithm based on hybrid features yielded promising results in histological image diagnostics for early diagnosis of OSCC.This study highlights the tremendous potential of artificial intelligence techniques to diagnose OSCC and increase cure rates among patients.
17Lu C et al. [38]2017Observational studyCNNTo construct an oral cavity histomorphometric-based image classifier for risk stratification of OSCC patientsSlides from 115 patientsTo risk stratify patients for disease-specific survivalHistopathological imageStandard clinical and pathologic parametersROC = 0.72,
hazard ratio = 11.02
(+) EffectivePatients with positive results were 11 times more likely to develop disease recurrence and die from it.Quantitative histomorphometric features of local nuclear architecture derived from digitized H&E slides of OSCCs are independently predictive of patient survival
18Shaban M et al. [39]2019Observational studyCNNTo obtain an automated TIL abundance score and explore its prognostic
significance for disease-free survival (DFS) of OSCC patients
Slides from 70 patientsPrognostic significance for disease-free survival of OSCC patientsHistopathological imagesManual TIL scoreHigh accuracy of 96.31%(+) EffectiveThe automated TILAb score had a significantly higher prognostic value than the manual TIL score (p = 0.0024).The TILAb score can be used as an independent prognostic parameter in OSCC patients.
19Anuradha K et al. [40]2017Observational studyCNNTo histologically grade oral tumors using fuzzy cognitive map (FCM)Histopathological images from 123 casesComputer-aided grading of oral tumorsHistopathological imagesActive Hebbian learning (AHL)Accuracy of 90.58% for oral tumors of low grade and 89.47% of high grade(+) EffectiveThe proposed method used an FCM grading model to categorize tumor cases into low grade and high grade. In addition, to improve the values, an active Hebbian learning algorithm was used.Features can be extracted using feature extraction methods and can be given as input to the FCM.
Footnotes: OSCC = oral squamous cell carcinoma, CNNs = convolutional neural networks, SNNs = semantic segmentation deep neural networks, ROC = receiver operating characteristic curve, AUC = area under the curve, IHC = immunohistochemical, TIL = tumor infiltrating lymphocytes, mIOU = mean intersection-over-union, DWT = discrete wavelet transform, LBP = local binary pattern, FCH = fuzzy color histogram, and GLCM = gray level co-occurrence matrix.
Table 3. Assessment of Strength of Evidence.
Table 3. Assessment of Strength of Evidence.
OutcomeInconsistencyIndirectnessImprecisionRisk of BiasPublication BiasStrength of Evidence
Application of AI in the diagnosis of OSCC [22,23,24,25,26,27,28]Not PresentNot PresentNot PresentPresentNot Present⨁⨁⨁◯
Application of AI in differentiating between normal and malignant conditions [29,30,31,32,33,34,35,36]Not PresentNot PresentNot PresentNot PresentNot Present⨁⨁⨁⨁
Application of AI in the early diagnosis of OSCC [37]Not PresentNot PresentNot PresentNot PresentNot Present⨁⨁⨁⨁
Application of AI in predicting survival of OSCC patients [38,39]Not PresentNot PresentNot PresentPresentNot Present⨁⨁⨁◯
Application of AI in severity grading of OSCC [40]Not PresentNot PresentNot PresentNot PresentNot Present⨁⨁⨁⨁
Foot Notes: ⨁⨁⨁⨁—high evidence, ⨁⨁⨁◯—moderate evidence.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khanagar, S.B.; Alkadi, L.; Alghilan, M.A.; Kalagi, S.; Awawdeh, M.; Bijai, L.K.; Vishwanathaiah, S.; Aldhebaib, A.; Singh, O.G. Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review. Biomedicines 2023, 11, 1612. https://doi.org/10.3390/biomedicines11061612

AMA Style

Khanagar SB, Alkadi L, Alghilan MA, Kalagi S, Awawdeh M, Bijai LK, Vishwanathaiah S, Aldhebaib A, Singh OG. Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review. Biomedicines. 2023; 11(6):1612. https://doi.org/10.3390/biomedicines11061612

Chicago/Turabian Style

Khanagar, Sanjeev B., Lubna Alkadi, Maryam A. Alghilan, Sara Kalagi, Mohammed Awawdeh, Lalitytha Kumar Bijai, Satish Vishwanathaiah, Ali Aldhebaib, and Oinam Gokulchandra Singh. 2023. "Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review" Biomedicines 11, no. 6: 1612. https://doi.org/10.3390/biomedicines11061612

APA Style

Khanagar, S. B., Alkadi, L., Alghilan, M. A., Kalagi, S., Awawdeh, M., Bijai, L. K., Vishwanathaiah, S., Aldhebaib, A., & Singh, O. G. (2023). Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review. Biomedicines, 11(6), 1612. https://doi.org/10.3390/biomedicines11061612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop