Next Article in Journal
Perineural Invasion and Associated Pain Transmission in Pancreatic Cancer
Next Article in Special Issue
Interpretability of a Deep Learning Based Approach for the Classification of Skin Lesions into Main Anatomic Body Sites
Previous Article in Journal
Combined Therapy of Locally Advanced Oesophageal and Gastro–Oesophageal Junction Adenocarcinomas: State of the Art and Aspects of Predictive Factors
Previous Article in Special Issue
Site-Specific Variation in Radiomic Features of Head and Neck Squamous Cell Carcinoma and Its Impact on Machine Learning Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Imaging Combined with Artificial Intelligence in the Early Detection of Esophageal Cancer

1
Division of Gastroenterology and Hepatology, Department of Internal Medicine, Ditmanson Medical Foundation Chiayi Christian Hospital, Chia Yi City 60002, Taiwan
2
Department of Mechanical Engineering, Advanced Institute of Manufacturing with High tech Innovations (AIM-HI) and Center for Innovative Research on Aging Society (CIRAS), National Chung Cheng University, 168, University Rd., Min Hsiung, Chia Yi County 62102, Taiwan
3
Department of Internal Medicine, Far Eastern Memorial Hospital, No.21, Sec. 2, Nanya S. Rd., Banciao Dist., New Taipei City 22060, Taiwan
4
Division of Gastroenterology, Department of Internal Medicine, Kaohsiung Medical University Hospital, Kaohsiung Medical University, No.100, Tzyou 1st Rd., Sanmin Dist., Kaohsiung City 80756, Taiwan
5
Department of Medicine, Faculty of Medicine, College of Medicine, Kaohsiung Medical University, No.100, Tzyou 1st Rd., Sanmin Dist., Kaohsiung City 80756, Taiwan
6
Graduate Institute of Clinical Medicine, College of Medicine, Kaohsiung Medical University, No.100, Tzyou 1st Rd., Sanmin Dist., Kaohsiung City 80756, Taiwan
7
Department of Gastroenterology, Kaohsiung Armed Forces General Hospital, 2, Zhongzheng 1st. Rd., Lingya District, Kaohsiung City 80284, Taiwan
8
Department of Nursing, Tajen University, 20, Weixin Rd., Yanpu Township, Pingtung City 90741, Taiwan
*
Authors to whom correspondence should be addressed.
Cancers 2021, 13(18), 4593; https://doi.org/10.3390/cancers13184593
Submission received: 3 August 2021 / Revised: 9 September 2021 / Accepted: 10 September 2021 / Published: 13 September 2021

Abstract

:

Simple Summary

Detection of early esophageal cancer is important to improve patient survival, however, early diagnosis of the cancer cells is difficult, even for experienced endoscopists. This article provides a new method by using hyperspectral imaging and a deep learning diagnosis model to classify and diagnose esophageal cancer using a single-shot multibox detector. The accuracy of the results when using an RGB image in WLI was 83% and while using the spectrum data the accuracy was increased to 88%. There was an increase of 5% in WLI. The accuracy of the results when using an RGB image in NBI was 86% and while using the spectrum data the accuracy was increased to 91%. There was an increase of 5% in NBI. This study proves that the accuracy of prediction when using the spectrum data has been significantly improved and the diagnosis of narrow-band endoscopy data is more sensitive than that of white-light endoscopy.

Abstract

This study uses hyperspectral imaging (HSI) and a deep learning diagnosis model that can identify the stage of esophageal cancer and mark the locations. This model simulates the spectrum data from the image using an algorithm developed in this study which is combined with deep learning for the classification and diagnosis of esophageal cancer using a single-shot multibox detector (SSD)-based identification system. Some 155 white-light endoscopic images and 153 narrow-band endoscopic images of esophageal cancer were used to evaluate the prediction model. The algorithm took 19 s to predict the results of 308 test images and the accuracy of the test results of the WLI and NBI esophageal cancer was 88 and 91%, respectively, when using the spectral data. Compared with RGB images, the accuracy of the WLI was 83% and the NBI was 86%. In this study, the accuracy of the WLI and NBI was increased by 5%, confirming that the prediction accuracy of the HSI detection method is significantly improved.

1. Introduction

Esophageal cancer can be regarded as the least researched and one of the deadliest cancers around the globe [1,2]. It is the eighth most common cancer and the sixth most common cause of cancer death [3]. At present, the endoscopists are unable to draw conclusions from the images of esophageal cancer. This also causes early symptoms to be easily overlooked. A precise assessment for individualized treatment depends on the accuracy of the initial diagnosis [4]. The images obtained by the means of general instruments will be affected by tissue secretions or instrument specifications, which may directly or indirectly lead to misjudgments in diagnosis. However, by using hyperspectral imaging technology combined with artificial intelligence deep learning methods to perform spectral data for esophageal cancer will offer a faster and more accurate diagnosis. The hyperspectral images have nanometer-level spectral intervals and hence the amount of spectrum information that can be detected is much larger than that of multispectral images [5,6,7,8,9,10,11]. The image spectrum conversion is the use of an imaging spectrometer to obtain an image that has a wide wavelength measurement range, the image is divided into a multispectral image and a hyperspectral image according to the spectral resolution. The two images are usually separated by a resolution of 10 nm. When the spectral resolution of the image is less than or equal to 10 nm, it is a hyperspectral image; when the spectral resolution of the image is greater than 10 nm, it is called a multispectral image. In addition, the spectral characteristics are unique to each substance, so it can be used to identify any object, which can be very accurately identified from the spectral characteristics or spectral response of the material or object in the obtained hyperspectral image [12,13]. In recent years, many scholars have conducted deep learning-related research on esophageal cancer which has proved to be a potential method for the early detection of esophageal cancer [14,15,16,17,18,19,20,21]. Shahidi et al. conducted research on white-light endoscopic images and narrow-band endoscopic images of gastrointestinal cancers to accurately diagnose and treat the esophagus by artificial intelligence [22]. Yoshitaka et al. conducted research on the application of convolutional neural networks in artificial intelligence to determine the invasion depth of esophageal cancer under white-light endoscopy. They also proved that the AI diagnostic system is much more effective for the diagnostic accuracy of esophageal squamous cell carcinoma (ESCC) as the depth of invasion is higher than that of endoscopists [23]. Hiromu et al. also studied the comparison between artificial intelligence and endoscopy professional physicians in real-time assisted diagnosis of ESCC and proved that the AI method had a higher detection sensitivity and accuracy of noncancerous tissue [24]. Wang et al. developed a single-shot multibox detector using a convolutional neural network for diagnosing esophageal cancer by using endoscopic images.
The main objective of this research is to use the hyperspectral conversion of the esophageal cancer image to obtain the spectrum information and then use the SSD model to perform deep learning training on the esophageal cancer spectrum information, thereby establishing the detection model for diagnosis of the stage and the location of cancerous lesions.

2. Materials and Methods

The number of endoscopic images of esophageal cancer used in this study was 1232. Out of these images, the number of WLI and NBI esophageal cancers was 620 and 612, respectively. Among the 620 white-light endoscopic images, the number of images co-relating with four levels of severity were 100, 112, 196 and 212 normal, low-grade dysplasia, high-grade dysplasia, invasive cancer, respectively. Among the 612 narrow-band esophageal cancer endoscopy images, the number of images co-relating with four stages in esophageal cancer were 100, 108, 180 and 224 normal, low-grade dysplasia, high-grade dysplasia, invasive cancer, respectively. In this study, the system was trained by 1232 images of esophageal cancer and its precursor lesions (620 WLI, 612 NBI), and then 308 images (155 WLI, 153 NBI) were evaluated with the trained system (accuracy of AI-HIS). The hyperspectral conversion technology (Smart Spectrum Cloud, Hitspectra Intelligent Technology Co., Ltd., Kaohsiung City, Taiwan) used in this study to convert the two types of endoscopic esophageal cancer images into 401 bands (380–780 nm) is shown in Figure 1.
The Deep Convolutional Neural Network (CNN) model of the Single Shot MultiBox Detector (SSD, Hitspectra Intelligent Technology Co., Ltd., Kaohsiung City, Taiwan) used in this study is a detection architecture based on the VGG-16-Atrous network [25]. SSD mainly predicts from multiple feature maps [16]. Vgg-16 is a deep learning architecture with 16 hidden layers composed of 13 convolutional layers and 3 fully connected layers [26,27]. The key feature of this model is that it is at least one order of magnitude more efficient than the existing methods and has a default box with different positions, scales, aspect ratios, and a multiscale convolutional bounding box output using multiple feature maps [28]. The principle of SSD for target matching will first calculate the intersection over union (IOU) value between the default frame and the real target, and then will select the default frame with the largest IOU value to match the real frame. The default frame for successful matching is the positive sample. Negative samples that are successfully matched will be evaluated again. If the IOU value of the default box is not successfully matched with the real box having greater than a certain threshold (generally set to 0.5), it will be reclassified as a positive sample. For multiple feature maps, SSD uses a set of preset default bounding boxes to associate with each feature mapping unit, and the default bounding boxes are tiled on the feature map in a convolutional manner. Therefore, the position of each bounding box relative to its corresponding unit is fixed. The offset relative to the default bounding box shape in each feature map is predicted and the category of each object existing in each bounding box score from the training model loss can be calculated from Equation (1) [29].
L ( x , c , l , g ) = 1 N ( L c o n f ( x , c ) + α L l o c ( x , l , g ) )
N is the number of positive samples of the matching default bounding box. If N = 0, the loss value is defined as 0; c is the predicted value of confidence loss; l is the predicted position of the bounding box corresponding to the default box; g is the position of ground truth parameter. In this study, the spectrum information of esophageal cancer and the label information that predicts the stage and coordinates of the disease are used in the SSD detection model framework. The data batch size is 32, the global learning rate is 0.0005, and the total number of training steps is 80,000. The white-light endoscopic imaging of esophageal cancer data and the narrow-band endoscopic imaging of esophageal cancer data are separately trained. The complete flow chart of this study is shown in Figure 2.
The spectrum information obtained by image spectrum conversion has 401 characteristic dimensions in the visible light band (380–780 nm). Figure 3a,c shows the spectrum of esophageal cancer in four stages: normal, low-grade dysplasia, high-grade dysplasia, invasive cancer of white-light endoscopic images and narrow-band endoscopic images, respectively. The high-dimensional spectrum information of the white-light and narrow-band endoscopes is reduced to three-dimensions using principal component analysis (PCA) [30]. The two kinds of endoscope spectrum information after dimension can show good variability between the severity of esophageal cancer as shown in Figure 3b,d.
The color image obtained by the white-light and narrow-band endoscope needed to be subjected to a series of preprocessing. In order to increase the training data, the preprocessed images were augmented, and the label information of the data was created based on this, which contained the name and coordinate location of the disease stage. On the other hand, using 24 color blocks as the reference object, a spectrometer was used to establish an image spectrum conversion module for the endoscope. By using the multivariate regression analysis, the conversion matrix that converted the endoscopic images into the spectrum of the visible light band could be obtained. With the conversion matrix, the images could be converted into the spectrum of the visible light band to obtain the spectrum information of the esophageal cancer. Next, the spectrum information of esophageal cancer was reduced by principal component analysis. Finally, by using SSD we trained a diagnostic model to predict the number and location of esophageal cancers, based on frequency spectrum.
In this study, 155 white-light esophageal cancer spectrum data and 153 narrow-band esophageal cancer spectrum information were used to test the white-light prediction model and the narrow-band prediction model, respectively. The resulting prediction model can be presented at the same time as the corresponding original esophageal cancer image. This kind of visualization allowed us to observe and evaluate the prediction results with a very intuitive presentation effect. The results displayed a rectangular box surrounding the esophageal cancer lesion in the endoscopic image. The predicted rectangular box represented a corresponding color according to the assigned disease stage name; dysplasia corresponds to low-grade dysplasia, high-grade dysplasia corresponds to light blue-gray, invasive cancer corresponds to orange, and for normal there would not be any prediction box. The kappa statistic is frequently used to test interrater reliability [31]. The importance of reliability lies in the fact that it represents the extent to which the data collected in the study are correct representations of the variables measured. Like most correlation statistics, the kappa can range from −1 to +1. The kappa is one of the most commonly used statistics to test interrater reliability [32].

3. Results

Figure 4 shows the image results of esophageal cancer prediction presented on a white-light endoscopic imaging while Figure 5 shows the image results of esophageal cancer prediction presented on a narrow-band endoscope.
Table 1 compares the test results of the esophageal cancer spectrum data in this study with the test results using RGB esophageal cancer images. The table sorts out the spectrum data obtained by white-light endoscopes and narrow-band endoscopes. The test data results of RGB images are shown under four severity levels. The experimental data was statistically based on the confusion matrix and the results are shown in Table 2. In the white-light endoscopy spectrum data, out of the 155 images, 25, 28, 49 and 53 images represent the normal, low-grade dysplasia, high-grade dysplasia, invasive cancer case respectively. Among the prediction results of the four cancer stages, 23 in normal, 21 in low-grade dysplasia, 45 in high-grade dysplasia, and 48 in invasive cancer were predicted correctly. Therefore, the accuracy of the white-light endoscopy data test result was 88%. On the other hand, in the narrow-band endoscopy spectrum data, out of the images, 25, 27, 48 and 56 images represent the normal, low-grade dysplasia, high-grade dysplasia, invasive cancer case respectively. Among the prediction results of the four cancer stages, 24 in normal, 24 in low-grade dysplasia, 41 in high-grade dysplasia, and 51 in invasive cancer were predicted correctly. Therefore, the accuracy of the white-light endoscopy data test result was 91%. As seen from Table 2, the sensitivity rate of WLI of normal stage in RGB is 76% while for the same stage in the WLI spectrum the sensitivity was 92%. This proves that the efficiency in detecting the esophageal cancer in the normal stage has been improved by 16%. In the low-grade dysplasia stage, the sensitivity rate of WLI in RGB was 61% while for the same stage in the WLI spectrum the sensitivity was 75%. This drop in the sensitivity rate was due to fewer sample rates being available for this stage. However, in the high-grade dysplasia and invasive cancer stages the sensitivity rate in WLI RGB was 86% and 93%, respectively, while the WLI spectrum was 92 and 91%, respectively. The same pattern could be found in the NBI band also. As seen from Table 2, the sensitivity rate of NBI of normal stage in RGB was 68%, while for the same stage in the NBI spectrum the sensitivity was 96%. This proves that, similar to the WLI band, the efficiency of detecting the esophageal cancer in the early stage by using NBI spectrum was improved by 28%. In the low-grade dysplasia stage, the sensitivity rate of NBI in RGB was 83%, while for the same stage in the NBI spectrum the sensitivity was 89%. Unlike the WLI band, the sensitivity rate was increased by 6%. From this interference it is clear that for the first two stages of detecting esophageal cancer, the NBI band can be used, while for the latter two stages, the WLI band can be used. However, in the high-grade dysplasia and invasive cancer stages the sensitivity rate in NBI RGB was 88 and 98%, respectively while in the NBI spectrum, both were 91%. As seen in Table 2, the kappa WLI RBG and NBI RGB was 0.76 and 0.81 while the kappa for the WLI spectrum and NBI spectrum was 0.84. This shows that there was an increase of 0.8 and 0.3 in kappa for the WLI and NBI spectra.

4. Discussion

From this article, it is very clear that by using a hyperspectral imaging (HSI) and deep learning diagnosis model the accuracy in the early detection of esophageal cancer can be increased significantly. This study provides SSD, a fast single-shot multi-class object detector. The model is characterized by a multiscale convolutional bounding box output using multiple feature maps. The position, scale and aspect ratio of the predicted sample were greater by at least one order of magnitude when compared with the existing methods. This advantage enables the model to train the bounding box shape space more effectively. The algorithm took 19 s to predict the results of 308 test images. The accuracy of the results when using RGB image in WLI was 83% and while using the spectrum data the accuracy was increased to 88%. There was an increase of 5% in WLI. The accuracy of the results when using RGB image in NBI was 86% and while using the spectrum data the accuracy was increased to 91%. There was an increase of 5% in NBI. At present, there are not many studies on spectrum information of esophageal cancer. Even so, by increasing the amount of data for training spectrum information, the accuracy of model diagnosis can be improved significantly. This study was conducted at the Kaohsiung Medical University and a total of 45 patients took part in this study. Currently this method has not been used in any clinical diagnosis. However, in the future, if research on the band selection for spectrum information of esophageal cancer is proposed in the future, using this method to reduce the amount of spectrum information, combined with the detection and diagnosis methods of this study will be useful for the early diagnosis of esophageal cancer and will have better predictions.

5. Conclusions

This study proves that the accuracy of prediction when using spectrum data has been significantly improved and the diagnosis of narrow-band endoscopy data is more sensitive than that of white-light endoscopy.

Author Contributions

Conceptualization, C.-L.T., Y.-K.W., C.-W.H. and H.-C.W.; data curation, T.-H.C., Y.-S.T., A.M. and H.-C.W.; formal analysis, T.-H.C., I.-C.W., C.-S.C. and Y.-S.T.; funding acquisition, C.-L.T., Y.-S.T. and H.-C.W.; investigation, C.-W.H., Y.-S.T., A.M., Y.-H.C. and H.-C.W.; methodology, C.-L.T., Y.-S.T., I.-C.W., A.M., Y.-K.W. and H.-C.W.; project administration, C.-S.C., C.-W.H., Y.-S.T. and H.-C.W.; resources, H.-C.W., C.-L.T. and T.-H.C.; software, C.-L.T., T.-H.C., Y.-H.C., I.-C.W. and A.M.; supervision, H.-C.W.; validation, C.-L.T., Y.-S.T., I.-C.W., T.-H.C. and H.-C.W.; writing—original draft, A.M.; writing—review and editing, H.-C.W. and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Science and Technology, The Republic of China under the grants MOST 105-2923-E-194-003 MY3, 108-2823-8-194-002, 109-2622-8-194-001-TE1, and 109-2622-8-194-007. This work was financially/partially supported by the Advanced Institute of Manufacturing with High-tech Innovations (AIM-HI) and the Center for Innovative Research on Aging Society (CIRAS) from The Featured Areas Research Center Program within the frame-work of the Higher Education Sprout Project by the Ministry of Education (MOE), the Ditmanson Medical Foundation Chia-Yi Christian Hospital Research Program (R110-27), and Kaohsiung Armed Forces General Hospital research project 109-012 in Taiwan.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Kaohsiung Medical University Hospital (KMUH) (KMUHIRB-E(II)-20190376, KMUHIRB-E(I)-20210066).

Informed Consent Statement

Written informed consent was waived in this study because of the retrospective, anonymized nature of study design.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Holmes, R.S.; Vaughan, T.L. Epidemiology and Pathogenesis of Esophageal Cancer. Semin. Radiat. Oncol. 2007, 17, 2–9. [Google Scholar] [CrossRef]
  2. Layke, J.C.; Lopez, P.P. Esophageal cancer: A review and update. Am. Fam. Physician 2006, 73, 2187–2194. [Google Scholar]
  3. Barret, M.; Prat, F. Diagnosis and treatment of superficial esophageal cancer. Ann. Gastroenterol. 2018, 31, 256–265. [Google Scholar] [CrossRef]
  4. Wu, I.-C.; Syu, H.-Y.; Jen, C.-P.; Lu, M.-Y.; Chen, Y.-T.; Wu, M.-T.; Kuo, C.-T.; Tsai, Y.-Y.; Wang, H.-C. Early identification of esophageal squamous neoplasm by hyperspectral endoscopic imaging. Sci. Rep. 2018, 8, 13797. [Google Scholar] [CrossRef] [PubMed]
  5. Hsiao, Y.-P.; Chiu, C.-W.; Lu, C.-W.; Nguyen, H.; Tseng, Y.; Hsieh, S.-C.; Wang, H.-C. Identification of Skin Lesions by Using Single-Step Multiframe Detector. J. Clin. Med. 2021, 10, 144. [Google Scholar] [CrossRef] [PubMed]
  6. Yao, H.-Y.; Tseng, K.-W.; Nguyen, H.-T.; Kuo, C.-T.; Wang, H.-C. Hyperspectral Ophthalmoscope Images for the Diagnosis of Diabetic Retinopathy Stage. J. Clin. Med. 2020, 9, 1613. [Google Scholar] [CrossRef]
  7. Chang, S.-C.; Syu, H.-Y.; Wang, Y.-L.; Lai, C.-J.; Huang, S.-Y.; Wang, H.-C. Identifying the incidence level of periodontal disease through hyperspectral imaging. Opt. Quantum Electron. 2018, 50, 409. [Google Scholar] [CrossRef]
  8. Hsiao, Y.-P.; Wang, H.-C.; Chen, S.-H.; Tsai, C.-H.; Yang, J.-H. Identified early stage mycosis fungoides from psoriasis and atopic dermatitis using non-invasive color contrast enhancement by LEDs lighting. Opt. Quantum Electron. 2014, 47, 1599–1611. [Google Scholar] [CrossRef]
  9. Hsiao, Y.-P.; Wang, H.-C.; Chen, S.-H.; Tsai, C.-H.; Yang, J.-H. Optical perception for detection of cutaneous T-cell lymphoma by multi-spectral imaging. J. Opt. 2014, 16, 125301. [Google Scholar] [CrossRef]
  10. Jen, C.-P.; Huang, C.-T.; Chen, Y.-S.; Kuo, C.-T.; Wang, H.-C. Diagnosis of human bladder cancer cells at different stages using mul-tispectral imaging microscopy. IEEE J. Sel. Top. Quantum Electron. 2013, 20, 81–88. [Google Scholar] [CrossRef]
  11. El-Rahman, S.A. Hyperspectral image classification using unsupervised algorithms. IJACSA Int. J. Adv. Comput. Sci. Appl. 2016, 7, 198–205. [Google Scholar]
  12. Manolakis, D.; Shaw, G. Detection algorithms for hyperspectral imaging applications. IEEE Signal Process. Mag. 2002, 19, 29–43. [Google Scholar] [CrossRef]
  13. Pohl, C.; Van Genderen, J. Remote Sensing Image Fusion: A Practical Guide; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  14. Le Berre, C.; Sandborn, W.J.; Aridhi, S.; Devignes, M.; Fournier, L.; Smail-Tabbone, M.; Danese, S.; Peyrin-Biroulet, L. Application of artificial intelligence to gastroenterology and hepatology. Gastroenterology 2020, 158, 76–94.e2. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, Y.-K.; Syu, H.-Y.; Chen, Y.-H.; Chung, C.-S.; Tseng, Y.; Ho, S.-Y.; Huang, C.-W.; Wu, I.-C.; Wang, H.-C. Endoscopic Images by a Single-Shot Multibox Detector for the Identification of Early Cancerous Lesions in the Esophagus: A Pilot Study. Cancers 2021, 13, 321. [Google Scholar] [CrossRef]
  16. Cai, S.-L.; Li, B.; Tan, W.-M.; Niu, X.-J.; Yu, H.-H.; Yao, L.-Q.; Zhou, P.-H.; Yan, B.; Zhong, Y.-S. Using a deep learning system in endoscopy for screening of early esophageal squamous cell carcinoma (with video). Gastrointest. Endosc. 2019, 90, 745–753.e2. [Google Scholar] [CrossRef]
  17. Ebigbo, A.; Mendel, R.; Probst, A.; Manzeneder, J., Jr.; Papa, J.P.; Palm, C.; Messmann, H. Computer-aided diagnosis using deep learning in the evaluation of early oesophageal adenocarcinoma. Gut 2018, 68, 1143–1145. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, Y.-Y.; Xue, D.-X.; Wang, Y.-L.; Zhang, R.; Sun, B.; Cai, Y.; Feng, H.; Cai, Y.; Xu, J. Computer-assisted diagnosis of early esophageal squamous cell carcinoma using nar-row-band imaging magnifying endoscopy. Endoscopy 2019, 51, 333–341. [Google Scholar]
  19. Guo, L.; Xiao, X.; Wu, C.; Zeng, X.; Zhang, Y.; Du, J.; Bai, S.; Xie, J.; Zhang, Z.; Li, Y.; et al. Real-time automated diagnosis of precancerous lesions and early esophageal squamous cell car-cinoma using a deep learning model (with videos). Gastrointest. Endosc. 2020, 91, 41–51. [Google Scholar] [CrossRef]
  20. Kumagai, Y.; Takubo, K.; Kawada, K.; Aoyama, K.; Endo, Y.; Ozawa, T.; Hirasawa, T.; Yoshio, T.; Ishihara, S.; Fujishiro, M.; et al. Diagnosis using deep-learning artificial intelligence based on the endocytoscopic observation of the esophagus. Esophagus 2018, 16, 180–187. [Google Scholar] [CrossRef]
  21. Horie, Y.; Yoshio, T.; Aoyama, K.; Yoshimizu, S.; Horiuchi, Y.; Ishiyama, A.; Hirasawa, T.; Tsuchida, T.; Ozawa, T.; Ishihara, S.; et al. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointest. Endosc. 2018, 89, 25–32. [Google Scholar] [CrossRef]
  22. Shahidi, N.; Bourke, M.J. Can artificial intelligence accurately diagnose endoscopically curable gastrointestinal cancers? Tech. Innov. Gastrointest. Endosc. 2020, 22, 61–65. [Google Scholar] [CrossRef]
  23. Tokai, Y.; Yoshio, T.; Aoyama, K.; Horie, Y.; Yoshimizu, S.; Horiuchi, Y.; Ishiyama, A.; Tsuchida, T.; Hirasawa, T.; Sakakibara, Y.; et al. Application of artificial intelligence using convolutional neural networks in determining the invasion depth of esophageal squamous cell carcinoma. Esophagus 2020, 17, 250–256. [Google Scholar] [CrossRef]
  24. Fukuda, H.; Ishihara, R.; Kato, Y.; Matsunaga, T.; Nishida, T.; Yamada, T.; Ogiyama, H.; Horie, M.; Kinoshita, K.; Tada, T. Comparison of performances of artificial intelligence versus expert endoscopists for re-al-time assisted diagnosis of esophageal squamous cell carcinoma (with video). Gastrointest. Endosc. 2020, 92, 848–855. [Google Scholar] [CrossRef]
  25. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single Shot Multibox Detector; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  26. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:14091556. [Google Scholar]
  27. Szegedy, C.; Reed, S.; Erhan, D.; Anguelov, D.; Ioffe, S. Scalable, high-quality object detection. arXiv 2014, arXiv:14121441. [Google Scholar]
  28. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vision. 2013, 104, 154–171. [Google Scholar] [CrossRef] [Green Version]
  29. Saito, S.; Lin, Y.-C.; Nakamura, Y.; Eckner, R.; Wuputra, K.; Kuo, K.-K.; Lin, C.-S.; Yokoyama, K.K. Potential application of cell reprogramming techniques for cancer research. Cell. Mol. Life Sci. 2018, 76, 45–65. [Google Scholar] [CrossRef] [Green Version]
  30. Rodarmel, C.; Shan, J. Principal component analysis for hyperspectral image classification. Surv. Land Inf. Science. 2002, 62, 115–122. [Google Scholar]
  31. McHugh, M.L. Interrater reliability: The kappa statistic. Biochem. Med. 2012, 276–282. [Google Scholar] [CrossRef]
  32. Landis, J.R.; Koch, G.G. An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers. Biometrics 1977, 33, 363–374. [Google Scholar] [CrossRef]
Figure 1. The flow chart of spectrum conversion construction, using standard 24 color blocks (X-Rite Classic, 24 Color Checkers) as the common target object for spectrum conversion of the endoscope and spectrometer, converting the endoscopic images into 401 bands of visible light spectrum information.
Figure 1. The flow chart of spectrum conversion construction, using standard 24 color blocks (X-Rite Classic, 24 Color Checkers) as the common target object for spectrum conversion of the endoscope and spectrometer, converting the endoscopic images into 401 bands of visible light spectrum information.
Cancers 13 04593 g001
Figure 2. Flowchart of SSD esophageal cancer hyperspectral training.
Figure 2. Flowchart of SSD esophageal cancer hyperspectral training.
Cancers 13 04593 g002
Figure 3. White-light and narrow-band endoscopic imaging of esophageal cancer spectrum and dimensionality reduction data distribution map. (a) White-light endoscopic images of esophageal cancer spectrum; (b) white-light endoscopic images of esophageal cancer spectrum dimensionality reduction data in the first two principal components of the data distribution map; (c) narrow-band endoscopic imaging of esophageal cancer spectrum; (d) narrow-band endoscopic imaging of esophageal cancer spectrum dimensionality reduction data in the first two principal components of the data distribution map.
Figure 3. White-light and narrow-band endoscopic imaging of esophageal cancer spectrum and dimensionality reduction data distribution map. (a) White-light endoscopic images of esophageal cancer spectrum; (b) white-light endoscopic images of esophageal cancer spectrum dimensionality reduction data in the first two principal components of the data distribution map; (c) narrow-band endoscopic imaging of esophageal cancer spectrum; (d) narrow-band endoscopic imaging of esophageal cancer spectrum dimensionality reduction data in the first two principal components of the data distribution map.
Cancers 13 04593 g003
Figure 4. The results of the prediction of esophageal cancer on the white-light endoscopic imaging. The correct position is represented by green rectangle is shown for comparison with the predicted result. (a) The result of the esophagus test is normal. It is not displayed under the diagnosis of SSD. (b) represents the detection result of the area with low-grade dysplasia, under the diagnosis of SSD, the blue bounding box surrounds the lesion area; (c) shows the area with high-grade dysplasia. The test result shows that the gray bounding box surrounds the lesion area under the diagnosis of SSD; (d) shows the detection result of the invasive cancer area, and the orange bounding box surrounds the lesion area under the diagnosis of SSD.
Figure 4. The results of the prediction of esophageal cancer on the white-light endoscopic imaging. The correct position is represented by green rectangle is shown for comparison with the predicted result. (a) The result of the esophagus test is normal. It is not displayed under the diagnosis of SSD. (b) represents the detection result of the area with low-grade dysplasia, under the diagnosis of SSD, the blue bounding box surrounds the lesion area; (c) shows the area with high-grade dysplasia. The test result shows that the gray bounding box surrounds the lesion area under the diagnosis of SSD; (d) shows the detection result of the invasive cancer area, and the orange bounding box surrounds the lesion area under the diagnosis of SSD.
Cancers 13 04593 g004
Figure 5. The results of the prediction of esophageal cancer are displayed on the narrow-band endoscopic imaging, where the correct position (green rectangle) is compared with the predicted result; (a) is the normal esophageal test result, which is not displayed under the diagnosis of SSD; (b) shows the detection result with low-grade dysplasia. Under the diagnosis of SSD, the blue bounding box surrounds the lesion area; (c) shows with the area between high-grade dysplasia and low-grade dysplasia. The test result shows that the gray bounding box surrounds the lesion area under the diagnosis of SSD; (d) shows the detection result of the invasive cancer area, and the orange bounding box surrounds the lesion area under the diagnosis of SSD.
Figure 5. The results of the prediction of esophageal cancer are displayed on the narrow-band endoscopic imaging, where the correct position (green rectangle) is compared with the predicted result; (a) is the normal esophageal test result, which is not displayed under the diagnosis of SSD; (b) shows the detection result with low-grade dysplasia. Under the diagnosis of SSD, the blue bounding box surrounds the lesion area; (c) shows with the area between high-grade dysplasia and low-grade dysplasia. The test result shows that the gray bounding box surrounds the lesion area under the diagnosis of SSD; (d) shows the detection result of the invasive cancer area, and the orange bounding box surrounds the lesion area under the diagnosis of SSD.
Cancers 13 04593 g005
Table 1. RGB image and spectrum data test results analysis.
Table 1. RGB image and spectrum data test results analysis.
WLI RGBSensitivity (%)Precision (%)F1-Score (%)Accuracy (%)Kappa
Normal767274830.76
Low-grade dysplasia618571
High-grade dysplasia868686
Invasive cancer938589
NBI RGBSensitivity (%)Precision (%)F1-score (%)Accuracy (%)Kappa
Normal688977860.81
Low-grade dysplasia838885
High-grade dysplasia888687
Invasive cancer988490
WLI spectrumSensitivity (%)Precision (%)F1-score (%)Accuracy (%)Kappa
Normal928588880.84
Low-grade dysplasia758479
High-grade dysplasia928890
Invasive cancer919291
NBI spectrumSensitivity (%)Precision (%)F1-score (%)Accuracy(%) Kappa
Normal968992910.84
Low-grade dysplasia899291
High-grade dysplasia918990
Invasive cancer919493
Table 2. Confusion matrix statistics table of RGB image and spectrum data test results, which compares the esophageal cancer spectrum data test and the test results using RGB esophageal cancer images.
Table 2. Confusion matrix statistics table of RGB image and spectrum data test results, which compares the esophageal cancer spectrum data test and the test results using RGB esophageal cancer images.
Confusion MatrixTrue Value Total
WLI RGBNormalDysplasiaHigh-Grade
Dysplasia
Invasive Cancer
Predicted valueNormal13302112
Low-grade dysplasia11110
High-grade dysplasia31301
Invasive cancer03439
NBI RGBNormalLow-Grade DysplasiaHigh-grade dysplasiaInvasive Cancer
Predicted valueNormal25201152
Low-grade dysplasia11510
High-grade dysplasia51380
Invasive cancer60453
WLI spectrumNormalLow-Grade DysplasiaHigh-grade dysplasiaInvasive Cancer
Predicted valueNormal23211155
Low-grade dysplasia22111
High-grade dysplasia03453
Invasive cancer02248
NBI spectrumNormalLow-Grade DysplasiaHigh-grade dysplasiaInvasive Cancer
Predicted valueNormal24210153
Low-grade dysplasia12401
High-grade dysplasia01411
Invasive cancer00351
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tsai, C.-L.; Mukundan, A.; Chung, C.-S.; Chen, Y.-H.; Wang, Y.-K.; Chen, T.-H.; Tseng, Y.-S.; Huang, C.-W.; Wu, I.-C.; Wang, H.-C. Hyperspectral Imaging Combined with Artificial Intelligence in the Early Detection of Esophageal Cancer. Cancers 2021, 13, 4593. https://doi.org/10.3390/cancers13184593

AMA Style

Tsai C-L, Mukundan A, Chung C-S, Chen Y-H, Wang Y-K, Chen T-H, Tseng Y-S, Huang C-W, Wu I-C, Wang H-C. Hyperspectral Imaging Combined with Artificial Intelligence in the Early Detection of Esophageal Cancer. Cancers. 2021; 13(18):4593. https://doi.org/10.3390/cancers13184593

Chicago/Turabian Style

Tsai, Cho-Lun, Arvind Mukundan, Chen-Shuan Chung, Yi-Hsun Chen, Yao-Kuang Wang, Tsung-Hsien Chen, Yu-Sheng Tseng, Chien-Wei Huang, I-Chen Wu, and Hsiang-Chen Wang. 2021. "Hyperspectral Imaging Combined with Artificial Intelligence in the Early Detection of Esophageal Cancer" Cancers 13, no. 18: 4593. https://doi.org/10.3390/cancers13184593

APA Style

Tsai, C. -L., Mukundan, A., Chung, C. -S., Chen, Y. -H., Wang, Y. -K., Chen, T. -H., Tseng, Y. -S., Huang, C. -W., Wu, I. -C., & Wang, H. -C. (2021). Hyperspectral Imaging Combined with Artificial Intelligence in the Early Detection of Esophageal Cancer. Cancers, 13(18), 4593. https://doi.org/10.3390/cancers13184593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop