Next Article in Journal
Ophthalmic Evaluation of Diagnosed Cases of Eye Cystinosis: A Tertiary Care Center’s Experience
Next Article in Special Issue
Automatic Grading of Individual Knee Osteoarthritis Features in Plain Radiographs Using Deep Convolutional Neural Networks
Previous Article in Journal
Diagnostic Performance of Biomarkers Urinary KIM-1 and YKL-40 for Early Diabetic Nephropathy, in Patients with Type 2 Diabetes: A Systematic Review and Meta-Analysis
Previous Article in Special Issue
CoSinGAN: Learning COVID-19 Infection Segmentation from a Single Radiological Image
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Performance Comparison between Automated Deep Learning and Dental Professionals in Classification of Dental Implant Systems from Dental Imaging: A Multi-Center Study

1
Department of Periodontology, Daejeon Dental Hospital, Institute of Wonkwang Dental Research, Wonkwang University College of Dentistry, Daejeon 35233, Korea
2
Department of Periodontology, Ilsan Hospital, National Health Insurance Service, Goyang 10444, Korea
3
Department of Periodontology, Mokdong Hospital, Ewha Woman’s University School of Medicine, Seoul 07985, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2020, 10(11), 910; https://doi.org/10.3390/diagnostics10110910
Submission received: 15 October 2020 / Revised: 4 November 2020 / Accepted: 5 November 2020 / Published: 7 November 2020
(This article belongs to the Special Issue Deep Learning for Computer-Aided Diagnosis in Biomedical Imaging)

Abstract

:
In this study, the efficacy of the automated deep convolutional neural network (DCNN) was evaluated for the classification of dental implant systems (DISs) and the accuracy of the performance was compared against that of dental professionals using dental radiographic images collected from three dental hospitals. A total of 11,980 panoramic and periapical radiographic images with six different types of DISs were divided into training (n = 9584) and testing (n = 2396) datasets. To compare the accuracy of the trained automated DCNN with dental professionals (including six board-certified periodontists, eight periodontology residents, and 11 residents not specialized in periodontology), 180 images were randomly selected from the test dataset. The accuracy of the automated DCNN based on the AUC, Youden index, sensitivity, and specificity, were 0.954, 0.808, 0.955, and 0.853, respectively. The automated DCNN outperformed most of the participating dental professionals, including board-certified periodontists, periodontal residents, and residents not specialized in periodontology. The automated DCNN was highly effective in classifying similar shapes of different types of DISs based on dental radiographic images. Further studies are necessary to determine the efficacy and feasibility of applying an automated DCNN in clinical practice.

1. Introduction

Dental implants have become a predictable treatment alternative for patients with partial or complete edentulous conditions [1]. Over the years, this treatment modality has evolved as a standard treatment protocol for replacing missing teeth. Thus, hundreds of manufacturers worldwide are producing and distributing over 2000 different types of dental implant systems (DISs) that differ in diameter, length, shape, coating, and surface material and properties [2,3]. Therefore, clinical dental practitioners have to select the appropriate DIS for a specific clinical indication based on their personal skillset and preferences.
DISs have shown a success rate of more than 90% and long-term survival rate of more than 10 years in systematic and meta-analytic review studies, which inevitably increases with the occurrence of mechanical and biological complications, such as fixtures or screw fractures, screw loosening, veneer chipping or fractures, low implant stability, peri-implant mucositis, and peri-implantitis [4,5,6,7]. Therefore, identifying the correct DIS is very important to repair or replace the existing DIS without available information [8,9]. However, studies on methods and techniques that enable the clear identification of DISs are scarce [10,11].
A deep convolutional neural network (DCNN) is a type of artificial intelligence that uses a cascade of multiple layers of nonlinear processing units for feature extraction, transformation, and classification of high-dimensional datasets [12]. A DCNN that is specifically designed for detection, classification, and segmentation in vision tasks and practical applications has been rapidly exploited in recent years in conjunction with improvements in computer performance and deep learning techniques [12]. Particularly, DCNN has been successfully applied in several medical and dental fields, and demonstrated significant advantages in terms of diagnosis and prognosis, such as diabetic retinopathy in retinal fundus photographs, skin cancer in skin lesion photographs, periodontally compromised teeth and dental caries on dental radiographs, and oral cystic lesions on cone beam computed tomography [13,14,15,16,17].
The fine-tuning of deep learning-based algorithms requires specific technical skills and mathematical knowledge, and creating an optimized DCNN for medical and dental applications is an extremely challenging task with numerous hidden challenges [18,19]. Therefore, in recent years, an automated DCNN that regulates the entire deep learning process involved in appropriate model selection and optimized hyper-parameter tuning was developed. The effectiveness and suitability of this automated DCNN are being evaluated in medical applications [20,21].
In the dental field, it is difficult to find studies related to automated DCNN, and to the best of our knowledge, no studies have been conducted on the classification of DISs using fully automated DCNN. We hypothesized that automated DCNN is highly effective in classifying similar shapes of different types of DISs compared to most dental professionals. Therefore, the purpose of this study is to evaluate the efficacy of the automated DCNN for classifying various types of DISs and compare the performance accuracy with dental professionals using dental radiographic images.

2. Materials and Methods

2.1. Dataset

The study design was approved by the Institutional Review Board of Daejeon Dental Hospital, Wonkwang University (approval no. W2003/003-001). Anonymized raw panoramic and periapical radiographic images (DICOM format panoramic images with a pixel resolution of 2868 × 1504 and periapical images with a pixel resolution of 1440 × 1920) were collected from three multi-center investigations conducted by Daejeon Dental Hospital, Wonkwang University (WKUDH); Ilsan Hospital, National Health Insurance Service (NHIS-IH); and Mokdong Hospital, Ewha Womans University (EWU-MH). The dataset contained six different types of DISs taken between January 2006 and December 2009 at WKUDH and NHIS-IH, and from May 2009 to May 2019 at EWU-MH.

2.2. Classification of DISs

DISs were classified into six different types with a diameter of 3.3–5.0 mm and a length of 7–13 mm:
  • Astra OsseoSpeed® TX (Dentsply IH AB, Molndal, Sweden), with a diameter of 4.5–5.0 mm and a length of 9–13 mm;
  • Implantium® (Dentium, Seoul, Korea), with a diameter of 3.6–5.0 mm and a length of 8–12 mm;
  • Superline® (Dentium, Seoul, Korea), with a diameter of 3.6–5.0 mm and a length of 8–12 mm;
  • TSIII® (Osstem, Seoul, Korea), with a diameter of 3.5–5.0 mm and a length of 7–13 mm;
  • SLActive® BL (Institut Straumann AG, Basel, Switzerland), with a diameter of 3.3–4.8 mm and a length of 8–12 mm;
  • SLActive® BLT (Institut Straumann AG, Basel, Switzerland), with a diameter of 3.3–4.8 mm and a length of 8–12 mm.

2.3. Data Preparation

Images with severe noise, blur, distortion, and other conditions that impeded the clinical detection and classification of DISs were excluded from the dataset. All included DISs were then manually classified and labeled by five periodontal residents (EHJ, BRN, DHK, JWK, and KYP) who did not directly participate in this study, and confirmed by three participating board-certified periodontists (JHL, YTK, and JBL) based on annotated electronic dental and medical records. A total of 11,980 images, including Astra OsseoSpeed® TX (n = 388, 3.2%), Impantium® (n = 2512, 21.0%), Superline® (n = 2360, 19.7%), TSIII® (n = 5617, 46.9%), SLActive® BL (n = 540, 4.5%), and SLActive® BLT (n = 563, 4.7%), were extracted from 7146 (59.6%) panoramic and 4834 (40.4%) periapical radiographic images. The details and numbers of radiographic images for each DIS are listed in Table 1. The dataset was randomly divided into two groups: 9584 (80%) radiographic images selected for the training dataset and the remaining 2396 (20%) radiographic images used as the testing dataset. The dataset was resized and transformed into a pixel resolution of 112 × 224, and the brightness and contrast were normalized using the OpenCV library functions [22].

2.4. Automated DCNN

Automated DCNN using Neuro-T version 2.0.1 (Neurocle Inc., Seoul, Korea), which are specialized tools for automatic model selection and hyper-parameter optimization, were adopted for this study. During training and inference, the automated DCNN automatically creates effective deep learning models and searches the optimal hyperparameters. An Adam optimizer with L2 regularization was used for transfer learning. The batch size was set to 432, and the automated DCNN architecture consisted of 18 layers with no dropout (Figure 1).

2.5. Comparing the Performance of the Automated DCNN to that of Dental Professionals

A total of 180 radiographic images (each DIS included 30 panoramic and periapical images) were randomly selected from the test dataset using the Keras framework in Python (version 3.8, Python Software Foundation). We then compared the accuracy of the performance of 25 dental professionals (including six board-certified periodontists, eight periodontology residents, and 11 residents not specialized in periodontology, from WKUDH, NHIS-IH, and EWU-MH) to the trained automated DCNN.

2.6. Statistical Analysis

The accuracy of the automated DCNN was evaluated, and the differences between the trained automated DCNN and the dental professionals were compared using the datasets from WKUDH, NHIS-IH, and EWU-MH. For the evaluation, the following statistical parameters were taken into account: receiver operating characteristic (ROC) curve, area under the ROC curve (AUC), 95% confidence intervals (CIs), standard error (SE), Youden index (sensitivity + specificity − 1), sensitivity, and specificity, which were calculated using Neuro-T (version 2.0.1) and R statistical software (version 3.5, R Foundation for Statistical Computing, Vienna, Austria). Delong’s method was used to compare the AUCs generated from the test dataset, and the significance level was set at p < 0.05.

3. Results

3.1. Outcomes of Automated DCNN on the Test Dataset

The accuracy of the automated DCNN abased on the AUC, Youden index, sensitivity, and specificity for the 2,396 panoramic and periapical radiographic images were 0.954 (95% CI = 0.933–0.970, SE = 0.011), 0.808, 0.955, and 0.853, respectively. Using only panoramic radiographic images (n = 1429), the automated DCNN achieved an AUC of 0.929 (95% CI = 0.904–0.949, SE = 0.018, Youden index = 0804, sensitivity = 0.922, and specificity = 0.882), while the corresponding value using only periapical radiographic images (n = 967) achieved an AUC of 0.961 (95% CI = 0.941–0.976, SE = 0.009, Youden index = 0.802, sensitivity = 0.955, and specificity = 0.846). There were no significant differences in accuracy among the three ROC curves (Table 2 and Figure 2).

3.2. Outcomes for Automated DCNN Algorithm Compared to that of Dental Professionals

Using 180 panoramic and periapical radiographic images randomly selected from the testing dataset, the automated DCNN outperformed most of the participating dental professionals, including board-certified periodontists, periodontal residents, and residents in other departments, in terms of the overall sensitivity and specificity (Figure 3). In particular, the superior accuracy of the automated DCNN was distinct for Straumann SLActive® BLT (AUC = 0.981, 95% CI = 0.949–0.996, SE = 0.009, Youden index = 0.880, sensitivity = 0.900, and specificity = 0.980) and Straumann SLActive® BL (AUC = 0.974, 95% CI = 0.938–0.992, SE = 0.010, Youden index = 0.833, sensitivity = 0.967, and specificity = 0.867), as shown in Table 3.

4. Discussion

Attempts have been made to identify or classify various types of DISs in the past, but most studies have been confined to research in field trials (which use few DIS images or require additional detailed information, such as diameter, length, taper angle, type of thread, and collar shape) [23,24]. Recently, various studies were conducted to confirm the effectiveness of DCNN with respect to identifying various types of DISs [25,26]. As far as we know, this is the first study to use automated DCNN for classifying similar shapes of different types of DISs and demonstrated higher performance accuracy compared to dental professionals.
In our previous studies, we demonstrated that the pre-trained DCNN using dental radiographic images demonstrated high accuracy in identifying and classifying periodontally compromised teeth (AUC = 0.781, 95% CI = 0.650–0.87.6) and dental caries (AUC = 0.845, 95% CI = 0.790–0.901) at a level equivalent to that of experienced dental professionals [15,16]. However, an assessment of clinical parameters (including clinical attachment level, probing depth, bleeding upon probing, tooth mobility, percussion, and electric pulp test), subjective symptoms (including duration and severity of pain and swelling), and radiological interpretation are essential for accurate diagnosis and appropriate treatment. Therefore, the DCNN approach for diagnosing periodontal disease and dental caries using radiographic images has limitations in clinical practice.
In contrast, the DCNN-based approach that uses only radiographic images is very effective and considered to be quite useful in actual clinical practice as a method for classifying various types of DISs with similar diameters and lengths. Two recent studies found that pre-trained or finely tuned DCNN architectures (including VGG16, VGG19, SqueezeNet, GoogLeNet, ResNet-18, MobileNet-v2, and ResNet-50) showed a high accuracy of more than 86% for classifying similar but different types of DISs [25,26]. Our previous study also indicated that the pre-trained DCNN (GoogLeNet Inception-v3) provided reliable results and achieved a higher accuracy (AUC = 0.962, 95% CI = 0.954–0.970) than a board-certified periodontist (AUC = 0.925, 95% CI = 0.913–0.935) for classifying three types of DISs using panoramic and periapical radiographic images [27].
The results of our previous pilot study demonstrated that there is an insignificant difference in the accuracy between panoramic-only and periapical-only based datasets [27]. Moreover, the results of this study confirmed that the accuracy was not statistically or significantly different among the use of panoramic-only (AUC of 0.929, 95% CI = 0.904–0.949), periapical-only (AUC = 0.961, 95% CI = 0.941–0.976), and panoramic and periapical (AUC = 0.954, 95% CI = 0.933–0.970) datasets. Therefore, to compare the accuracy of automated DCNN with that of dental professionals, panoramic and periapical radiographic images were included in one dataset (rather than divided into separate datasets). Additionally, because each DIS used in this study had the same shape but different diameters and lengths, the DISs were not divided according to the diameter and length used to build the dataset.
The Straumann SLActive® BLT implant system has a relatively large tapered shape compared to other types of DISs. Thus, the automated DCNN (AUC = 0.981, 95% CI = 0.949–0.996) and dental professionals (AUC = 0.928, 95% CI = 0.920–0.936) achieved appropriate classifications with high AUC. However, for the Dentium Superline® and Osstem TSIII® implant systems that do not have conspicuous characteristic elements with a tapered shape, the automated DCNN classified correctly with an AUC of 0.903 (95% CI = 0.850–0.967) and 0.937 (95% CI = 0.890–0.967), whereas dental professionals showed a low AUC of 0.541 (95% CI = 0.527– 0.556) and 0.525 (95% CI = 0.510–0.540), respectively. Based on these results, the automated DCNN showed statistically significant higher classification accuracy than dental professionals, including experienced periodontists. Furthermore, it was confirmed that the automated DCNN was highly effective in classifying similar shapes of DISs based on dental radiographic images. Additionally, several previous studies reported that the professional experience of the examiner is an important factor for interpreting dental radiographs [28,29]. Contrastingly, we found that the difference in the experience level associated with DISs did not affect the classification accuracy of DISs significantly because the classification of DISs is unfamiliar regardless of their professional experience.
Nonetheless, this study has certain limitations. Although six types of DISs were selected from three different dental hospitals and categorized as a dataset, the training dataset was still insufficient for clinical practice. Therefore, it is necessary to build a high-quality and large-scale dataset containing different types of DISs. If time and cost are not limited, the automated DCNN can be continuously trained and optimized for improved accuracy. However, owing to computing power constraints, we have to compromise on optimization at the appropriate level. Additionally, the automated DCNN regulates the entire process, including appropriate model selection and optimized hyper-parameter tuning. Therefore, there is less room for human experts to manually check and intervene during the entire process of deep learning training. Cone-beam computed tomography-based three-dimensional images are widely used in the dental field. However, they were not included in the dataset of this study. The classification of DISs using three-dimensional images with less distortion than two-dimensional images is expected to improve accuracy significantly. Therefore, further research is required based on three-dimensional images.

5. Conclusions

The selection of an appropriate DCNN model with optimized hyper-parameter tuning is key to the success of deep learning research. We demonstrated that the accuracy of the automated DCNN outperformed most of the participating dental professionals. Therefore, the automated DCNN can help clinical dental practitioners to classify various types of DISs based on dental radiographic images. Nevertheless, further studies are necessary to determine the efficacy and feasibility of applying the automated DCNN in clinical practice.

Author Contributions

Conceptualization, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; methodology, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; software, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; validation, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; formal analysis, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; investigation, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; resources, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; data curation, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; writing—original draft preparation, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; writing—review and editing, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; visualization, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; supervision, J.-H.L., Y.-T.K., J.-B.L., and S.-N.J.; project administration, J.-H.L.; funding acquisition, J.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a National Research Foundation of Korea (NRF) grant funded by the Korean Government (MSIT) (No. 2019R1A2C1083978).

Acknowledgments

We are grateful to the periodontology resident dentists (Eun-Hee Jeong, Bo-Ram Nam, Do-Hyung Kim, Jung-Woong Kim, and Ki-Yeol Park) who helped prepare the dataset for this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Esposito, M.; Coulthard, P.; Thomsen, P.; Worthington, H.V. The role of implant surface modifications, shape and material on the success of osseointegrated dental implants. A cochrane systematic review. Eur. J. Prosthodont. Restor. Dent. 2005, 13, 15–31. [Google Scholar] [PubMed]
  2. Jokstad, A.; Braegger, U.; Brunski, J.B.; Carr, A.B.; Naert, I.; Wennerberg, A. Quality of dental implants. Int. Dent. J. 2003, 53, 409–443. [Google Scholar] [CrossRef] [PubMed]
  3. Esposito, M.; Ardebili, Y.; Worthington, H.V. Interventions for replacing missing teeth: Different types of dental implants. Cochrane Database Syst. Rev. 2014. [Google Scholar] [CrossRef] [PubMed]
  4. Moraschini, V.; Poubel, L.A.; Ferreira, V.F.; Barboza Edos, S. Evaluation of survival and success rates of dental implants reported in longitudinal studies with a follow-up period of at least 10 years: A systematic review. Int. J. Oral Maxillofac. Surg. 2015, 44, 377–388. [Google Scholar] [CrossRef] [PubMed]
  5. Srinivasan, M.; Meyer, S.; Mombelli, A.; Muller, F. Dental implants in the elderly population: A systematic review and meta-analysis. Clin. Oral Implants Res. 2017, 28, 920–930. [Google Scholar] [CrossRef] [PubMed]
  6. Howe, M.S.; Keys, W.; Richards, D. Long-term (10-year) dental implant survival: A systematic review and sensitivity meta-analysis. J. Dent. 2019, 84, 9–21. [Google Scholar] [CrossRef] [PubMed]
  7. Scarano, A.; Degidi, M.; Iezzi, G.; Petrone, G.; Piattelli, A. Correlation between implant stability quotient and bone-implant contact: A retrospective histological and histomorphometrical study of seven titanium implants retrieved from humans. Clin. Implant Dent. Relat. Res. 2006, 8, 218–222. [Google Scholar] [CrossRef]
  8. Jaarda, M.J.; Razzoog, M.E.; Gratton, D.G. Geometric comparison of five interchangeable implant prosthetic retaining screws. J. Prosthet. Dent. 1995, 74, 373–379. [Google Scholar] [CrossRef]
  9. Al-Wahadni, A.; Barakat, M.S.; Abu Afifeh, K.; Khader, Y. Dentists’ most common practices when selecting an implant system. J. Prosthodont. 2018, 27, 250–259. [Google Scholar] [CrossRef]
  10. Nuzzolese, E.; Lusito, S.; Solarino, B.; Di Vella, G. Radiographic dental implants recognition for geographic evaluation in human identification. J. Forensic Odontostomatol. 2008, 26, 8–11. [Google Scholar] [PubMed]
  11. Berketa, J.W.; Hirsch, R.S.; Higgins, D.; James, H. Radiographic recognition of dental implants as an aid to identifying the deceased. J. Forensic Sci. 2010, 55, 66–70. [Google Scholar] [CrossRef]
  12. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  13. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  14. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  15. Lee, J.H.; Kim, D.H.; Jeong, S.N.; Choi, S.H. Diagnosis and prediction of periodontally compromised teeth using a deep learning-based convolutional neural network algorithm. J. Periodontal Implant Sci. 2018, 48, 114–123. [Google Scholar] [CrossRef] [Green Version]
  16. Lee, J.H.; Kim, D.H.; Jeong, S.N.; Choi, S.H. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J. Dent. 2018, 77, 106–111. [Google Scholar] [CrossRef]
  17. Lee, J.H.; Kim, D.H.; Jeong, S.N. Diagnosis of cystic lesions using panoramic and cone beam computed tomographic images based on deep learning neural network. Oral Dis. 2020, 26, 152–158. [Google Scholar] [CrossRef]
  18. Rajkomar, A.; Dean, J.; Kohane, I. Machine learning in medicine. N. Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef]
  19. Stead, W.W. Clinical implications and challenges of artificial intelligence and deep learning. JAMA 2018, 320, 1107–1108. [Google Scholar] [CrossRef]
  20. Faes, L.; Wagner, S.K.; Fu, D.J.; Liu, X.; Korot, E.; Ledsam, J.R.; Back, T.; Chopra, R.; Pontikos, N.; Kern, C.; et al. Automated deep learning design for medical image classification by health-care professionals with no coding experience: A feasibility study. Lancet Digit. Health 2019, 1, e232–e242. [Google Scholar] [CrossRef] [Green Version]
  21. Waring, J.; Lindvall, C.; Umeton, R. Automated machine learning: Review of the state-of-the-art and opportunities for healthcare. Artif. Intell. Med. 2020, 104, 101822. [Google Scholar] [CrossRef]
  22. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: Cnn architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Sahiwal, I.G.; Woody, R.D.; Benson, B.W.; Guillen, G.E. Radiographic identification of nonthreaded endosseous dental implants. J. Prosthet. Dent. 2002, 87, 552–562. [Google Scholar] [CrossRef] [Green Version]
  24. Michelinakis, G.; Sharrock, A.; Barclay, C.W. Identification of dental implants through the use of implant recognition software (irs). Int. Dent. J. 2006, 56, 203–208. [Google Scholar] [CrossRef]
  25. Kim, J.E.; Nam, N.E.; Shim, J.S.; Jung, Y.H.; Cho, B.H.; Hwang, J.J. Transfer learning via deep neural networks for implant fixture system classification using periapical radiographs. J. Clin. Med. 2020, 9, 1117. [Google Scholar] [CrossRef] [Green Version]
  26. Sukegawa, S.; Yoshii, K.; Hara, T.; Yamashita, K.; Nakano, K.; Yamamoto, N.; Nagatsuka, H.; Furuki, Y. Deep neural networks for dental implant system classification. Biomolecules 2020, 10, 984. [Google Scholar] [CrossRef]
  27. Lee, J.H.; Jeong, S.N. Efficacy of deep convolutional neural network algorithm for the identification and classification of dental implant systems, using panoramic and periapical radiographs: A pilot study. Medicine (Baltimore) 2020, 99, e20787. [Google Scholar] [CrossRef] [PubMed]
  28. Anthonappa, R.P.; King, N.M.; Rabie, A.B.; Mallineni, S.K. Reliability of panoramic radiographs for identifying supernumerary teeth in children. Int. J. Paediatr. Dent. 2012, 22, 37–43. [Google Scholar] [CrossRef]
  29. Geibel, M.A.; Carstens, S.; Braisch, U.; Rahman, A.; Herz, M.; Jablonski-Momeni, A. Radiographic diagnosis of proximal caries-influence of experience and gender of the dental staff. Clin. Oral Investig. 2017, 21, 2761–2770. [Google Scholar] [CrossRef]
Figure 1. Overview of an automated deep convolutional neural network (DCNN) system.
Figure 1. Overview of an automated deep convolutional neural network (DCNN) system.
Diagnostics 10 00910 g001
Figure 2. (a) Receiver operating characteristic (ROC) curve for classification of six types of DISs in the testing dataset, which consisted of 2396 panoramic and periapical radiographic images. (b) The accuracy of the automated DCNN for the test dataset did not show a significant difference among the three ROC curves based on DeLong’s method.
Figure 2. (a) Receiver operating characteristic (ROC) curve for classification of six types of DISs in the testing dataset, which consisted of 2396 panoramic and periapical radiographic images. (b) The accuracy of the automated DCNN for the test dataset did not show a significant difference among the three ROC curves based on DeLong’s method.
Diagnostics 10 00910 g002
Figure 3. (af) Performance of the automated DCNN and comparison with dental professionals for classification of six types of DISs.
Figure 3. (af) Performance of the automated DCNN and comparison with dental professionals for classification of six types of DISs.
Diagnostics 10 00910 g003
Table 1. Number of panoramic and periapical radiographic images for each dental implant system (DIS).
Table 1. Number of panoramic and periapical radiographic images for each dental implant system (DIS).
Dataset
WKUDHNHIS-IHEWU-MH
Dental Implant SystemPanoramic images
(n = 4989)
Periapical images
(n = 3872)
Panoramic images
(n = 1120)
Periapical images
(n = 204)
Panoramic images
(n = 1037)
Periapical images
(n = 758)
Dentsply Astra OsseoSpeed TX®2471392---
Dentium Implantium®58957894414817479
Dentium Superline®1011970713220274
Osstem TSIII®2788199010023351365
Straumann SLActive® BL1028931206139
Straumann SLActive® BLT252106--104101
Dataset collected from three dental hospitals: Daejeon Dental Hospital, Wonkwang University (WKUDH), Ilsan Hospital, National Health Insurance Service (NHIS-IH), and Mokdong Hospital, Ewha Womans University (EWU-MH). All DISs consist of a diameter of 3.3–5.0 mm and length of 7–13 mm.
Table 2. Pairwise comparison of ROC curve for classification of six different types of DISs in the testing dataset.
Table 2. Pairwise comparison of ROC curve for classification of six different types of DISs in the testing dataset.
Difference between AreasSE95% CIp-Value
Panoramic and periapical images
vs. oeriapical images
0.0070.007−0.008–0.0220.365
Panoramic and periapical images
vs. panoramic images
0.0250.021−0.016–0.0670.235
Panoramic images
vs. oeriapical images
0.0320.020−0.006–0.0720.106
AUC, area under the curve; ROC, receiver operating characteristic curve; SE, standard error; CI, confidence interval; AUCs were compared using DeLong’s method for paired ROC curves; panoramic and periapical images, dataset consisting of 2396 panoramic and periapical radiographic images; panoramic images, dataset consisting of 1429 panoramic radiographic images; periapical images, dataset consisting of 967 periapical radiographic images.
Table 3. Accuracy comparison between the automated deep convolutional neural network and dental professionals for the classification of six types of DISs, based on 180 panoramic and periapical images randomly selected from the training dataset.
Table 3. Accuracy comparison between the automated deep convolutional neural network and dental professionals for the classification of six types of DISs, based on 180 panoramic and periapical images randomly selected from the training dataset.
VariablesAUC95% CISEYouden IndexSensitivitySpecificity
Dentsply Astra OsseoSpeed TX®
Automated DCNN0.9450.901–0.9730.0230.7660.9330.833
Board-certified periodontists0.8960.877–0.9140.0140.7250.7770.947
Periodontology residents0.8310.811–0.8500.0150.5170.5700.946
Residents not specialized in periodontology0.7770.758–0.7950.0140.4250.4930.931
Dentium Implantium®
Automated DCNN0.9080.856–0.9460.0260.7800.9330.847
Board-certified periodontists0.7910.766–0.8150.0130.7330.9660.766
Periodontology residents0.8060.785–0.8260.0110.6820.9120.770
Residents not specialized in periodontology0.7360.716–0.7550.0130.4650.6720.792
Dentium Superline®
Automated DCNN0.9030.850–0.9420.0410.7860.8330.954
Board-certified periodontists0.5370.507–0.5670.0160.3330.7780.588
Periodontology residents0.5340.508–0.5600.0150.3300.9450.384
Residents not specialized in periodontology0.5440.522–0.5660.0130.2920.8840.407
Osstem TSIII®
Automated DCNN0.9370.890–0.9670.0240.8130.9000.913
Board-certified periodontists0.5010.471–0.5320.0180.2980.9110.387
Periodontology residents0.5030.477–0.5290.0160.2700.1040.625
Residents not specialized in periodontology0.5560.534–0.5780.0140.2150.8210.394
Straumann SLActive® BL
Automated DCNN0.9740.938–0.9920.0100.8330.9670.867
Board-certified periodontists0.7590.732–0.7840.0150.6610.8880.772
Periodontology residents0.7530.730–0.7750.0140.6500.8700.779
Residents not specialized in periodontology0.6980.677–0.7180.0120.5070.7810.726
Straumann SLActive® BLT
Automated DCNN0.9810.949–0.9960.0090.8800.9000.980
Board-certified periodontists0.9680.955–0.9770.0110.9510.9550.995
Periodontology residents0.9150.899–0.9290.0140.8510.8660.985
Residents not specialized in periodontology0.9150.902–0.9270.0110.8520.8870.964
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, J.-H.; Kim, Y.-T.; Lee, J.-B.; Jeong, S.-N. A Performance Comparison between Automated Deep Learning and Dental Professionals in Classification of Dental Implant Systems from Dental Imaging: A Multi-Center Study. Diagnostics 2020, 10, 910. https://doi.org/10.3390/diagnostics10110910

AMA Style

Lee J-H, Kim Y-T, Lee J-B, Jeong S-N. A Performance Comparison between Automated Deep Learning and Dental Professionals in Classification of Dental Implant Systems from Dental Imaging: A Multi-Center Study. Diagnostics. 2020; 10(11):910. https://doi.org/10.3390/diagnostics10110910

Chicago/Turabian Style

Lee, Jae-Hong, Young-Taek Kim, Jong-Bin Lee, and Seong-Nyum Jeong. 2020. "A Performance Comparison between Automated Deep Learning and Dental Professionals in Classification of Dental Implant Systems from Dental Imaging: A Multi-Center Study" Diagnostics 10, no. 11: 910. https://doi.org/10.3390/diagnostics10110910

APA Style

Lee, J. -H., Kim, Y. -T., Lee, J. -B., & Jeong, S. -N. (2020). A Performance Comparison between Automated Deep Learning and Dental Professionals in Classification of Dental Implant Systems from Dental Imaging: A Multi-Center Study. Diagnostics, 10(11), 910. https://doi.org/10.3390/diagnostics10110910

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop