Next Article in Journal
The Value of Surveillance Imaging of Oral Squamous Cell Carcinoma
Previous Article in Journal
Pancreatic Cyst Size Measurement on Magnetic Resonance Imaging Compared to Pathology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence and Panendoscopy—Automatic Detection of Clinically Relevant Lesions in Multibrand Device-Assisted Enteroscopy

by
Francisco Mendes
1,2,†,
Miguel Mascarenhas
1,2,3,*,†,
Tiago Ribeiro
1,2,3,
João Afonso
1,2,3,
Pedro Cardoso
1,2,3,
Miguel Martins
1,2,
Hélder Cardoso
1,2,3,
Patrícia Andrade
1,2,3,
João P. S. Ferreira
4,5,
Miguel Mascarenhas Saraiva
6 and
Guilherme Macedo
1,2,3
1
Alameda Professor Hernâni Monteiro, Department of Gastroenterology, São João University Hospital, 4200-427 Porto, Portugal
2
WGO Gastroenterology and Hepatology Training Center, 4050-345 Porto, Portugal
3
Faculty of Medicine, University of Porto, Alameda Professor Hernâni Monteiro, 4200-427 Porto, Portugal
4
Department of Mechanical Engineering, Faculty of Engineering, University of Porto, Rua Dr. Roberto Frias, 4200-465 Porto, Portugal
5
DigestAID—Digestive Artificial Intelligence Development, R. Alfredo Allen n°. 455/461, 4200-135 Porto, Portugal
6
ManopH Gastroenterology Clinic, R. de Sá da Bandeira 752, 4000-432 Porto, Portugal
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2024, 16(1), 208; https://doi.org/10.3390/cancers16010208
Submission received: 5 December 2023 / Revised: 27 December 2023 / Accepted: 28 December 2023 / Published: 1 January 2024

Abstract

:

Simple Summary

Device-assisted enteroscopy is the only diagnostic and therapeutic exam capable of exploring the entire gastrointestinal tract. However, the diagnostic yield of this procedure is not sufficient enough to assure a cost-effective panendoscopy, and there is significant interobserver variability during the exam. Artificial intelligence tools have been proved to be beneficial in several areas of medicine, namely in Gastroenterology, with a strong image component. However, the development of deep learning models for application in device-assisted enteroscopy is still in an embryonic phase. The authors herein aimed to develop a multidevice convolutional neural network based on 338 exams performed in two renowned centers. The present model was able to accurately identify multiple clinically relevant lesions across the entire gastrointestinal tract, with an image processing time that favors its clinical applicability. The first worldwide panendoscopic model showed the potential of artificial intelligence in augmenting the accuracy and cost-effectiveness of device-assisted enteroscopy.

Abstract

Device-assisted enteroscopy (DAE) is capable of evaluating the entire gastrointestinal tract, identifying multiple lesions. Nevertheless, DAE’s diagnostic yield is suboptimal. Convolutional neural networks (CNN) are multi-layer architecture artificial intelligence models suitable for image analysis, but there is a lack of studies about their application in DAE. Our group aimed to develop a multidevice CNN for panendoscopic detection of clinically relevant lesions during DAE. In total, 338 exams performed in two specialized centers were retrospectively evaluated, with 152 single-balloon enteroscopies (Fujifilm®, Porto, Portugal), 172 double-balloon enteroscopies (Olympus®, Porto, Portugal) and 14 motorized spiral enteroscopies (Olympus®, Porto, Portugal); then, 40,655 images were divided in a training dataset (90% of the images, n = 36,599) and testing dataset (10% of the images, n = 4066) used to evaluate the model. The CNN’s output was compared to an expert consensus classification. The model was evaluated by its sensitivity, specificity, positive (PPV) and negative predictive values (NPV), accuracy and area under the precision recall curve (AUC-PR). The CNN had an 88.9% sensitivity, 98.9% specificity, 95.8% PPV, 97.1% NPV, 96.8% accuracy and an AUC-PR of 0.97. Our group developed the first multidevice CNN for panendoscopic detection of clinically relevant lesions during DAE. The development of accurate deep learning models is of utmost importance for increasing the diagnostic yield of DAE-based panendoscopy.

1. Introduction

Device-assisted enteroscopy (DAE) is an exam that combines diagnostic properties with tissue sampling and therapeutic endoscopy. Initially conceived for the investigation of small bowels [1], DAE’s properties make it suitable for evaluation of the entire GI tract. Typically, DAE is performed with a single- or double-balloon enteroscope, but recently, the motorized spiral enteroscope has been an alternative [2].
DAE is commonly performed in various clinical settings, more commonly after capsule endoscopy (CE) findings [3]. Ulcers and erosions are the most common pathological findings in the small bowel, with diverse etiologies (namely Crohn’s disease, refractory celiac disease, infections and neoplasms) [4]. DAE is capable of exploring a greater length of ileal mucosa than conventional ileocolonoscopy, augmenting the diagnostic accuracy for small-bowel Crohn’s disease. Additionally, in the presence of stricturing Crohn’s disease, DAE allows balloon dilation of small-bowel strictures, reducing the need for surgery.
Additionally, the identification of small-bowel tumors in CE is a common indication for DAE, with the possibility of tissue sampling being crucial to disease management [5,6]. Moreover, DAE is a useful exam in the management of polyposis syndromes, namely Peutz–Jeghers syndrome, with endoscopic polypectomy as an alternative to small-bowel resection in the course of multiple lesions found during the disease’s course [7]. Finally, DAE allows marking of the area by a small-bowel neoplasia, enhancing minimally invasive surgery.
Additionally, DAE is commonly performed in the setting of obscure gastrointestinal bleeding, particularly after a positive CE exam [8]. Angioectasias are the most common finding in the setting of small-bowel bleeding, with the possibility of argon plasma coagulation during the exam after lesion detection. However, the diagnostic yield in the presence of OGIB is modest (56%, increasing to 75% if with a positive CE exam) [8].
Additionally, there has been an increased focus on double-balloon-assisted colonoscopy. In fact, the use of a double-balloon enteroscope is useful in technically difficult colonoscopies, achieving higher cecal intubation cases and less patient discomfort in patients with difficult or previous incomplete colonoscopies [9]. However, this alternative use of the enteroscope should be accompanied by a high diagnostic accuracy. Finally, DAE is also performed in settings of altered anatomy. One of the main examples is the performance of DAE in order to access an excluded stomach in patients with Roux-en-Y gastric bypass in the setting of gastrointestinal bleeding or malignancy suspicion [10]. These alternative uses of the enteroscope guide the need to enhance the diagnostic accuracy of DAE, not only in the small bowel evaluation but also in a panendoscopic setting. Thus, despite the potential capabilities of DAE, there is a need to increase its diagnostic yield in several clinical settings.
In recent years, the application of artificial intelligence (AI) technologies in the medical field has observed exponential growth. Convolutional neural networks (CNNs) are a human cortex-inspired multi-layer architecture, with high proficiency for image pattern detection [11]. As a matter of fact, CNN models have been developed in several medical areas [12,13,14]. CE has been a focus of study in the development and application of deep learning technologies, increasing its diagnostic yield with a significant reading time reduction [15,16,17]. In fact, the development of deep learning methods has been theorized as a potential revolutionary tool to increase the diagnostic accuracy and cost-effectiveness of DAE [10]. Nevertheless, the implementation of AI models for DAE is still in the early stages. In fact, AI application in DAE has been studied for the identification of vascular lesions, protuberant lesions, ulcers and erosions [18,19,20]. Nevertheless, the clinical application of such technology is dependent on the ability to identify different types of lesions throughout a complete exam, while functioning in different devices. In this study, our group aimed to develop the first worldwide multibrand CNN for panendoscopic automatic classification of clinically relevant lesions in DAE, namely, vascular lesions, hematic residues, protuberant lesions, ulcers and erosions.

2. Materials and Methods

2.1. Study Design

A bicentric study was performed for construction of the CNN. A total of 338 exams consecutively performed at two Portuguese specialized centers (Centro Hospitalar Universitário São João and ManopH) between January 2020 and May 2023 were used for the development of the CNN. During that period, DAE was performed by three experienced gastroenterologists using three different devices: the double-balloon enteroscopy system Fujifilm EN-580T (n = 226), the single-balloon enteroscopy system Olympus EVIS EXERA II SIF-Q180 (n = 98) and the Olympus PowerSpiral Motorized Enteroscope PSF-1 (n = 14). Our study respected the Declaration of Helsinki and was developed in a non-interventional fashion. The study was approved by the ethics committee of São João University Hospital/Faculty of Medicine of the University of Porto (No. CE 407/2020). Omission of potentially identifying information of the subjects was ensured and each patient received a random number assignment to obtain effective data anonymization for researchers involved in the CNN. A legal team with Data Protection Officer (DPO) certification was responsible for the non-traceability of the data in conformity with general data protection regulation (GDPR).

2.2. Lesion Classification

The CNN comprised enteroscopy images from esophagic, gastric, enteric and colonic segments. Each segment was reviewed in order to identify several categories of lesions. The lesions selected for the model’s training and evaluation comprised a group of clinically relevant alterations in the gastrointestinal mucosa, in which the CNN could have a role with diagnostic and therapeutic implications. Vascular lesions included red spots, angioectasia and varices. Red spots were considered as punctuate flat lesions with a diameter under 1 mm, without vessel appearance. Angioectasia were defined as a reddish lesion of tortuous dilated clustered capillaries. Otherwise, varices were defined as venous dilations with a serpiginous appearance. The protruding lesions were defined as tissue elevations above the gastrointestinal epithelium, including polyps, flat lesions, subepithelial lesions and nodules. Otherwise, ulcers were considered whitish base areas of loss of epithelial covering, with surrounding swollen mucosa and a diameter of at least 5 mm. Mucosal erosions consisted of areas of minimal loss of epithelial covering and normal surrounding mucosa. The extracted images were classified by gastroenterologists with expertise in DAE (MMS, HC, PA). Non-agreeable images were discussed between DAE experts and discarded in the absence of a consensus.

2.3. CNN Development

The study design is represented through a flowchart in Figure 1. A total of 40,665 images were used for developing the model. The full dataset was divided into a training dataset (comprising around 90% of the images, n = 36,599) and a testing dataset (comprising around 10% of the images, n = 4066). The testing dataset was used for evaluating the model. In the training dataset, a 5-fold cross validation was performed, dividing the training dataset into 5 similar sized subsets. The results of each subset were used to identify the best parameters of the model that were used in the testing dataset.
The CNN was created with the Xception model pre-trained on ImageNet. The convolutional layers of the model were kept, assuring the transference of the learning to our data, while the last fully connected layers were removed. The attachment of fully connected layers was based on the number of the classes for classification of DAE images.
The model had 2 blocks with fully connected layers followed by a Dropout layer of 0.25 drop rate. A Dense layer with a size based on the number of categories to classify was added. A learning rate of 0.0001, batch size of 64 and number of epochs of 20 was set by trial and error. Our group used Keras libraries and Tensor-flow 2.3 to prepare the data and run the model. The analysis was dependent on a computer with an Intel® Xeon® Gold 6130 processor (Intel, Santa Clara, CA, USA) and a NVIDIA Quadro® RTXTM 4000 graphic processing unit (NVIDIA Corporate, Santa Clara, CA, USA).

2.4. Performance Measures and Statistical Analysis

The binary CNN calculated the probability of normal mucosa versus clinically relevant lesions for each given image (Figure 2), with higher probabilities demonstrating greater CNN prediction confidence. Heatmaps were generated based on localized features responsible for the prediction of the model (Figure 3), attempting to achieve a better comprehension of the model and guide clinical decision either during tissue sampling or therapeutic procedures. The CNN’s classification was compared to three DAE experts’, which remains the gold standard for evaluation of DAE images. Table 1 translates the confusion matrix between experts and CNN classification.
The model was evaluated through its sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV) and accuracy (Table 2). These performance measures were represented with their means and 95% confidence intervals (CI). The model’s global performance was evaluated with the precision recall (PR) curve and area under the precision recall curve (AUC-PR). Sci-kit learn version 0.22.2 was used for the statistical analysis [21].

3. Results

3.1. Construction of the Network

A CNN model was constructed with 40,655 images from 338 DAE exams. The training dataset, with 90% of the total images, was split into five similar-sized independent subsets. The remaining 10% of the total images were used for the validation dataset.
The CNN evaluated each individual image, predicting a classification with a level of certainty, later compared with the expert’s classification. The inputs of the different subsets of the training dataset allowed the trimming of the individual parameters of the CNN, which were evaluated in the validation dataset.

3.2. Global Performance of the Network

The training dataset was developed with a five-fold cross validation. Table 2 demonstrates the performance results of the folds of the training model. Thus, the training dataset had 88.7% mean sensitivity, 98.0% specificity, 92.6% PPV and 97.0% NPV, with a mean accuracy of 96.0%.
The validation dataset, with the remaining 10% of the total images, was used for evaluation of the CNN’s performance. Table 1 shows the confusion matrix between the CNN’s prediction and the experts’ classification. The deep learning model had a sensitivity of 88.9%, a specificity of 98.9%, a PPV of 95.8% and an NPV of 97.1%. The model revealed an overall accuracy of 96.8% and an AUC-PR of 0.97.

3.3. Convolutional Neural Network Computational Performance

The CNN completed the evaluation of the testing dataset within 33 s, resulting in an image processing time of 124 images per second.

4. Discussion

In this multicentric study, our group developed a CNN capable of automatic identification of clinically relevant lesions in a panendoscopic setting during DAE. Our model revealed excellent performances in all the evaluated parameters, with an overall accuracy of 96.8%. These results were accompanied by an image processing time that favors the clinical applicability of the technology. Additionally, the CNN was developed in a multidevice setting, namely, single-balloon enteroscope, double-balloon enteroscope and motorized spiral enteroscope devices, including all the types of devices used during DAE. This is, to our knowledge, the first worldwide multidevice CNN capable of detecting several types of lesions in esophagic, gastric, enteric and colonic segments during DAE.
Furthermore, it is important to discuss some methodologic points of our study. The CNN was trained with a five-fold cross validation strategy. This design assures a balanced distribution of different classes between folds, which is important in cases where class imbalance is common [22]. The choice of a five-fold cross validation strategy significantly reduced the random fluctuation typical of a single training–testing split. This methodological point aids in creating a model capable of better generalization of unseen data—which is of utmost importance in medical technology development and application [23]. Additionally, our group opted for the use of PR curves instead of the commoner receiver operating characteristic (ROC) curves to evaluate the discriminating ability of the model. The current literature regards the excessive optimism of ROC curves in evaluating a model performance in cases of data imbalance [24,25]. In cases of data imbalance, PR curves are more informative and preferred [26]. Thus, in this CNN, normal images corresponded to around 80% of the total images, favoring the use of PR curves, taking into consideration the objective of identifying all the lesions images, instead of the commoner normal images (which are part of the ROC concept).
On the other hand, beyond the increased complexity in terms of model characteristics, a CNN should be capable of being trustworthy and comprehensible. Thus, in the last few years, the concept of explainable AI has been a matter of intense discussion [27,28]. Our group tried to address this need with the development of heatmaps for each classified image, identifying the area responsible for the classification of the image (normal mucosa vs. lesion). Therefore, our group recognizes that the development of heatmaps and other explainable AI methods is of great interest to address the performance of procedures during DAE (like tissue sample or argon plasma coagulation of angioectasias), but also to ensure trustworthiness of the model and confidence among clinicians, who are responsible for its use and implementation. Thus, addressing this question is of utmost importance not only for the model’s developers but also for regulatory entities, ensuring accountability during the AI development and implementation process.
In addition to explainability, data responsibility and the ethical or legal consequences of the use of AI models are a matter of great interest [29]. Firstly, there is a need to discuss the legal responsibility in the face of an adverse outcome or incorrect diagnosis. There is still difficulty in establishing the autonomy and responsibility of an artificial intelligence model. In fact, in case of an adverse event, it is important to determine if any of the parts failed and determine responsibilities. Nevertheless, it is possible to have a misdiagnosis in the absence of a clearly determined error from any part. On the other hand, a clinician that assumes a decision based uniquely on a model’s output may be less trustworthy by the patient [30]. Therefore, it is important to have deep learning models’ outputs critically analyzed by experts in the field. Additionally, the development and implementation of deep learning models is commonly slowed by clinicians’ fear to be replaced by AI tools. Contrarily, AI models should be interpreted as a tool to increase clinicians’ diagnostic accuracy and provide more time for patient care. Nevertheless, perfect synergy is highly dependent on two factors: avoidance of the loss of skill by the clinician and a clear definition of decision protocol in cases of machine–human disagreement [31]. The authors have addressed these questions in the rewritten Discussion section of the paper.
The implementation of AI-based technologies in medicine is highly dependent on its generalization for use in multiple devices [32]. The FAIR data principles were published in 2016 as guiding tools for research data stewardship [33]. In order to fulfill the FAIR principles, data should be findable, accessible, interoperable and reusable. In the present study, the data are findable as they are assigned with unique and confidential classifiers and saved in our records. Moreover, they also respect the principle of accessibility as they can be accessed by the study investigators respecting the patient confidentiality issues and ethical principles described in the methods section. Additionally, our data are reusable, conceived to allow the continuous development of the convolutional neural network and application in another models. Taking into account the need to fulfill these four principles, our group regards the proof of methodological development in the fulfillment of system interoperability, which is not always addressed by the majority of deep learning models published in Gastroenterology. Thus, the interoperability challenge is a discussion topic in multiple science fields [34,35,36]. Our group has addressed that concern, developing a CNN that works in three different enteroscopes, comprising single-balloon, double-balloon and motorized spiral enteroscopy devices, solving a fundamental interoperability challenge for applying the technology. This is, to our knowledge, the first worldwide CNN for panendoscopic detection of clinically relevant lesions in a multidevice setting, achieved through a large image dataset.
The development of artificial intelligence models encompasses several methodological steps, translated by the technological readiness level (TRL) scale [37]. In fact, the TRL is very different between Gastroenterology fields. For instance, capsule endoscopy is one of the main areas for the development of deep learning models in Gastroenterology [38,39]. However, the majority of studies are still in an early development phase and not validated in the clinical practice. On the other hand, in colonoscopy, a computer-aided diagnostic software is already disposable in clinical practice (GI Genius, Medtronic®), being proficient in detecting colorectal polyps and predicting their histology [40]. Additionally, the software is capable of working with different endoscopy devices, achieving a solution for the interoperability challenge. However, the technology is dedicated to the evaluation of colonic mucosa, unsuitable for a panendoscopic evaluation of the gastrointestinal tract, and is centered in the detection of protuberant lesions. Our model, although being developed in a less-performed exam like device-assisted enteroscopy, focuses not only on the achievement of system interoperability but also the panendoscopic detection of multiple clinically relevant lesions. However, it is still at a low TRL and needs larger prospective real-time studies to implement the model into clinical practice.
DAE is a safe procedure, albeit with non-neglectable risks for adverse events, which are mostly minor and self-limiting. The advent of spiral motorized enteroscopy was accompanied by an increase in these minor adverse events. In fact, a work by Singh et al. revealed a minor adverse events rate of 48%, including superficial mucosal tears with middle ooze and mucosal tears [41]. Additionally, the distinction between these minor iatrogenic lesions and true lesions is a challenge in DAE. Differentiating angioectasias from iatrogenic lesions or red spots is challenging. Additionally, protuberant lesions like those verified in Peutz–Jeghers syndrome can be mistaken with irrelevant xanthelasmas. The impossibility to distinguish between these entities can prompt to unnecessary treatment, with non-neglectable risks of complications or an increase in procedure time. Deep learning models have been proposed to have a role in these clinical settings, with the possibility of not only increasing the diagnostic accuracy of an exam but also reducing the rate of unnecessary procedures. Thus, our model addressed this question, revealing 98.0% specificity with great accuracy for distinguishing between vascular lesions, ulcers, erosions and minor iatrogenic lesions developed during the exam, especially when performing spiral enteroscopy. This technical specificity is of great interest in reducing the performance of unnecessary argon plasma coagulation procedures, which are not only iatrogenic but also require higher exam completion time and are associated with increased exam-related costs.
The concept of panendoscopic evaluation of the GI tract with a single exam was introduced with the advent of minimally invasive capsule panendoscopy [42,43]. Nevertheless, capsule endoscopy (CE) is incapable of tissue sampling and therapeutic procedures. Therefore, DAE is the only exam with therapeutic purposes capable of the evaluation and management of esophageal, gastric, enteric and colonic mucosa pathologies. Thus, the development of a CNN capable of identifying lesions in a panendoscopic setting is of utmost importance to increase the clinical applicability of the model and the exam itself. Indeed, this is the first validated model to detect several clinically relevant lesions in a panendoscopic setting, achieving excellent results throughout esophagus, stomach, enteric and colonic segments of the exam. This model represents a critical milestone in implementing AI in digestive endoscopy, increasing the cost effectiveness of the exam in a panendoscopic evaluation.
In spite of the exponential growth in the development of deep learning models for CE [44,45], the application of AI technologies to DAE is still in a premature state, with scarce works applying deep learning models to augment the diagnostic performance of the exam. Additionally, the existing works were focused on detecting a specific type of lesion [18,19,20], which guarantees a diminished clinical applicability and a lower technology readiness level (TRL) of the technology. This work constitutes a landmark with the development of a CNN capable of detecting clinically relevant lesions during DAE, namely, vascular and protuberant lesions, hematic residues, ulcers and erosions. Additionally, this work was developed with images of both single-balloon, double-balloon and motorized spiral enteroscopes, solving a relevant interoperability challenge. The development of an explainable AI method like heatmap generation is of great importance to address the model’s trustworthiness and accountability.
Our group trained and validated the first panendoscopic CNN for detecting clinically relevant lesions during DAE, with high global accuracy and image processing capacities. Nevertheless, our CNN is still in an embryonic phase and not ready for clinical applicability.
This study contains some limitations. Firstly, it was developed with a retrospective design. Therefore, there is a need for larger prospective multicentric studies in the future to ensure clinical implementation of the models. Secondly, the CNN was based on still images, creating a need for real-time evaluation of panendoscopic lesions during DAE.

5. Conclusions

In conclusion, DAE is the only exam with diagnostic and therapeutic purposes capable of assuring panendoscopic evaluation of GI tract. However, the multiplicity of findings during this exam and regional differences between the mucosa of the different GI tract portions favors the implementation of AI models to increase the diagnostic ability and aid during therapeutic procedures. To our knowledge, our study reveals the first deep learning model capable of identifying clinically relevant lesions in a panendoscopic and multidevice setting (including the majority of DAE devices used in clinical practice). The development and application of these systems could amplify the indications and benefits of DAE, increasing its diagnostic yield and cost-effectiveness. In the future, larger prospective multicentric studies are needed to develop and apply these models.

Author Contributions

Conceptualization, M.M. (Miguel Mascarenhas), F.M. and G.M.; methodology, M.M. (Miguel Mascarenhas), F.M., T.R., J.A., P.C. and M.M. (Miguel Martins).; software, J.P.S.F.; validation, H.C., P.A., M.M.S. and G.M.; formal analysis, F.M., M.M. (Miguel Mascarenhas), T.R. and J.P.S.F.; investigation, F.M., M.M. (Miguel Mascarenhas) and J.P.S.F.; resources, H.C., P.A., M.M.S. and G.M.; data curation, J.P.S.F.; writing—original draft preparation, F.M. and M.M. (Miguel Mascarenhas) with equal contribution; writing—review and editing, F.M., M.M. (Miguel Mascarenhas), T.R., J.A., P.C. and M.M. (Miguel Martins); visualization, F.M., M.M. (Miguel Mascarenhas), T.R., J.A., P.C. and M.M. (Miguel Martins); supervision, H.C., P.A., M.M.S. and G.M.; project administration, M.M. (Miguel Mascarenhas); funding acquisition, M.M. (Miguel Mascarenhas). All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge NVIDIA for supporting the study by supplying graphic processing units.

Institutional Review Board Statement

This study was approved by the ethics committee of Centro Hospitalar Universitário de São João/Faculdade de Medicina da Universidade do Porto (CE 41/2021, Date of approval: 05/2021) and was conducted respecting the Declaration of Helsinki. This study is retrospective and of a non-interventional nature. Any information deemed to potentially identify the subjects was omitted.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

João P.S. Ferreira is currently an employee of DigestAID—Digestive Artificial Intelligence Development.

References

  1. Pennazio, M.; Rondonotti, E.; Despott, E.J.; Dray, X.; Keuchel, M.; Moreels, T.; Sanders, D.S.; Spada, C.; Carretero, C.; Cortegoso Valdivia, P.; et al. Small-bowel capsule endoscopy and device-assisted enteroscopy for diagnosis and treatment of small-bowel disorders: European Society of Gastrointestinal Endoscopy (ESGE) Guideline—Update 2022. Endoscopy 2023, 55, 58–95. [Google Scholar] [CrossRef]
  2. Beyna, T.; Arvanitakis, M.; Schneider, M.; Gerges, C.; Hoellerich, J.; Deviere, J.; Neuhaus, H. Total motorized spiral enteroscopy: First prospective clinical feasibility trial. Gastrointest. Endosc. 2021, 93, 1362–1370. [Google Scholar] [CrossRef]
  3. Rondonotti, E.; Spada, C.; Adler, S.; May, A.; Despott, E.J.; Koulaouzidis, A.; Panter, S.; Domagk, D.; Fernandez-Urien, I.; Rahmi, G.; et al. Small-bowel capsule endoscopy and device-assisted enteroscopy for diagnosis and treatment of small-bowel disorders: European Society of Gastrointestinal Endoscopy (ESGE) Technical Review. Endoscopy 2018, 50, 423–446. [Google Scholar] [CrossRef]
  4. Goenka, M.K.; Majumder, S.; Kumar, S.; Sethy, P.K.; Goenka, U. Single center experience of capsule endoscopy in patients with obscure gastrointestinal bleeding. World J. Gastroenterol. 2011, 17, 774–778. [Google Scholar] [CrossRef]
  5. Bettenworth, D.; Bokemeyer, A.; Kou, L.; Lopez, R.; Bena, J.F.; El Ouali, S.; Mao, R.; Kurada, S.; Bhatt, A.; Beyna, T.; et al. Systematic review with meta-analysis: Efficacy of balloon-assisted enteroscopy for dilation of small bowel Crohn’s disease strictures. Aliment. Pharmacol. Ther. 2020, 52, 1104–1116. [Google Scholar] [CrossRef]
  6. Rondonotti, E.; Koulaouzidis, A.; Yung, D.E.; Reddy, S.N.; Georgiou, J.; Pennazio, M. Neoplastic Diseases of the Small Bowel. Gastrointest. Endosc. Clin. N. Am. 2017, 27, 93–112. [Google Scholar] [CrossRef]
  7. Sakamoto, H.; Yamamoto, H.; Hayashi, Y.; Yano, T.; Miyata, T.; Nishimura, N.; Shinhata, H.; Sato, H.; Sunada, K.; Sugano, K. Nonsurgical management of small-bowel polyps in Peutz-Jeghers syndrome with extensive polypectomy by using double-balloon endoscopy. Gastrointest. Endosc. 2011, 74, 328–333. [Google Scholar] [CrossRef]
  8. Teshima, C.W.; Kuipers, E.J.; van Zanten, S.V.; Mensink, P.B. Double balloon enteroscopy and capsule endoscopy for obscure gastrointestinal bleeding: An updated meta-analysis. J. Gastroenterol. Hepatol. 2011, 26, 796–801. [Google Scholar] [CrossRef]
  9. Despott, E.J.; Murino, A.; Nakamura, M.; Bourikas, L.; Fraser, C. A prospective randomised study comparing double-balloon colonoscopy and conventional colonoscopy in pre-defined technically difficult cases. Dig. Liver Dis. 2017, 49, 507–513. [Google Scholar] [CrossRef]
  10. Nehme, F.; Goyal, H.; Perisetti, A.; Tharian, B.; Sharma, N.; Tham, T.C.; Chhabra, R. The Evolution of Device-Assisted Enteroscopy: From Sonde Enteroscopy to Motorized Spiral Enteroscopy. Front. Med. 2021, 8, 792668. [Google Scholar] [CrossRef]
  11. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef]
  12. Khurshid, S.; Friedman, S.; Reeder, C.; Di Achille, P.; Diamant, N.; Singh, P.; Harrington, L.X.; Wang, X.; Al-Alusi, M.A.; Sarma, G.; et al. ECG-Based Deep Learning and Clinical Risk Factors to Predict Atrial Fibrillation. Circulation 2022, 145, 122–133. [Google Scholar] [CrossRef]
  13. Sanyal, P.; Dey, P. Using a deep learning neural network for the identification of malignant cells in effusion cytology material. Cytopathology 2023, 34, 466–471. [Google Scholar] [CrossRef]
  14. Khan, R.A.; Fu, M.; Brent, B.; Luo, Y.; Wu, F.X. A multi-modal deep neural network for multi-class liver cancer diagnosis. Neural Netw. 2023, 165, 553–561. [Google Scholar] [CrossRef]
  15. Mascarenhas, M.; Ribeiro, T.; Afonso, J.; Ferreira, J.P.S.; Cardoso, H.; Andrade, P.; Parente, M.P.L.; Jorge, R.N.; Mascarenhas Saraiva, M.; Macedo, G. Deep learning and colon capsule endoscopy: Automatic detection of blood and colonic mucosal lesions using a convolutional neural network. Endosc. Int. Open 2022, 10, E171–E177. [Google Scholar] [CrossRef]
  16. Chu, Y.; Huang, F.; Gao, M.; Zou, D.W.; Zhong, J.; Wu, W.; Wang, Q.; Shen, X.N.; Gong, T.T.; Li, Y.Y.; et al. Convolutional neural network-based segmentation network applied to image recognition of angiodysplasias lesion under capsule endoscopy. World J. Gastroenterol. 2023, 29, 879–889. [Google Scholar] [CrossRef]
  17. Soffer, S.; Klang, E.; Shimon, O.; Nachmias, N.; Eliakim, R.; Ben-Horin, S.; Kopylov, U.; Barash, Y. Deep learning for wireless capsule endoscopy: A systematic review and meta-analysis. Gastrointest. Endosc. 2020, 92, 831–839. [Google Scholar] [CrossRef]
  18. Mascarenhas Saraiva, M.; Ribeiro, T.; Afonso, J.; Andrade, P.; Cardoso, P.; Ferreira, J.; Cardoso, H.; Macedo, G. Deep Learning and Device-Assisted Enteroscopy: Automatic Detection of Gastrointestinal Angioectasia. Medicina 2021, 57, 1378. [Google Scholar] [CrossRef]
  19. Martins, M.; Mascarenhas, M.; Afonso, J.; Ribeiro, T.; Cardoso, P.; Mendes, F.; Cardoso, H.; Andrade, P.; Ferreira, J.; Macedo, G. Deep-Learning and Device-Assisted Enteroscopy: Automatic Panendoscopic Detection of Ulcers and Erosions. Medicina 2023, 59, 172. [Google Scholar] [CrossRef]
  20. Cardoso, P.; Saraiva, M.M.; Afonso, J.; Ribeiro, T.; Andrade, P.; Ferreira, J.; Cardoso, H.; Macedo, G. Artificial Intelligence and Device-Assisted Enteroscopy: Automatic Detection of Enteric Protruding Lesions Using a Convolutional Neural Network. Clin. Transl. Gastroenterol. 2022, 13, e00514. [Google Scholar] [CrossRef]
  21. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  22. Hawkins, D.M.; Basak, S.C.; Mills, D. Assessing model fit by cross-validation. J. Chem. Inf. Comput. Sci. 2003, 43, 579–586. [Google Scholar] [CrossRef]
  23. Tougui, I.; Jilbab, A.; Mhamdi, J.E. Impact of the Choice of Cross-Validation Techniques on the Results of Machine Learning-Based Diagnostic Applications. Healthc. Inform. Res. 2021, 27, 189–199. [Google Scholar] [CrossRef]
  24. Movahedi, F.; Padman, R.; Antaki, J.F. Limitations of receiver operating characteristic curve on imbalanced data: Assist device mortality risk scores. J. Thorac. Cardiovasc. Surg. 2023, 165, 1433–1442.e1432. [Google Scholar] [CrossRef]
  25. Halligan, S.; Altman, D.G.; Mallett, S. Disadvantages of using the area under the receiver operating characteristic curve to assess imaging tests: A discussion and proposal for an alternative approach. Eur. Radiol. 2015, 25, 932–939. [Google Scholar] [CrossRef]
  26. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef]
  27. Belle, V.; Papantonis, I. Principles and Practice of Explainable Machine Learning. Front. Big Data 2021, 4, 688969. [Google Scholar] [CrossRef]
  28. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2020, 23, 18. [Google Scholar] [CrossRef]
  29. Mascarenhas, M.; Afonso, J.; Ribeiro, T.; Andrade, P.; Cardoso, H.; Macedo, G. The Promise of Artificial Intelligence in Digestive Healthcare and the Bioethics Challenges It Presents. Medicina 2023, 59, 790. [Google Scholar] [CrossRef]
  30. Watson, D.S.; Krutzinna, J.; Bruce, I.N.; Griffiths, C.E.; McInnes, I.B.; Barnes, M.R.; Floridi, L. Clinical applications of machine learning algorithms: Beyond the black box. BMJ 2019, 364, l886. [Google Scholar] [CrossRef]
  31. Taddeo, M.; Floridi, L. How AI can be a force for good. Science 2018, 361, 751–752. [Google Scholar] [CrossRef]
  32. He, J.; Baxter, S.L.; Xu, J.; Xu, J.; Zhou, X.; Zhang, K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019, 25, 30–36. [Google Scholar] [CrossRef]
  33. Wilkinson, M.D.; Dumontier, M.; Aalbersberg, I.J.; Appleton, G.; Axton, M.; Baak, A.; Blomberg, N.; Boiten, J.W.; da Silva Santos, L.B.; Bourne, P.E.; et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci. Data 2016, 3, 160018. [Google Scholar] [CrossRef]
  34. Zhang, J.; Ashrafian, H.; Delaney, B.; Darzi, A. Impact of primary to secondary care data sharing on care quality in NHS England hospitals. NPJ Digit. Med. 2023, 6, 144. [Google Scholar] [CrossRef]
  35. Rajamani, S.; Chakoian, H.; Bieringer, A.; Lintelmann, A.; Sanders, J.; Ostadkar, R.; Saupe, A.; Grilli, G.; White, K.; Solarz, S.; et al. Development and implementation of an interoperability tool across state public health agency’s disease surveillance and immunization information systems. JAMIA Open 2023, 6, ooad055. [Google Scholar] [CrossRef]
  36. Benhamou, P.Y.; Adenis, A.; Lablanche, S.; Franc, S.; Amadou, C.; Penfornis, A.; Kariyawasam, D.; Beltrand, J.; Charpentier, G. First Generation of a Modular Interoperable Closed-Loop System for Automated Insulin Delivery in Patients With Type 1 Diabetes: Lessons From Trials and Real-Life Data. J. Diabetes Sci. Technol. 2023, 17, 1433–1439. [Google Scholar] [CrossRef]
  37. Lavin, A.; Gilligan-Lee, C.M.; Visnjic, A.; Ganju, S.; Newman, D.; Ganguly, S.; Lange, D.; Baydin, A.G.; Sharma, A.; Gibson, A.; et al. Technology readiness levels for machine learning systems. Nat. Commun. 2022, 13, 6039. [Google Scholar] [CrossRef]
  38. Afonso, J.; Saraiva, M.M.; Ferreira, J.P.S.; Cardoso, H.; Ribeiro, T.; Andrade, P.; Parente, M.; Jorge, R.N.; Macedo, G. Automated detection of ulcers and erosions in capsule endoscopy images using a convolutional neural network. Med. Biol. Eng. Comput. 2022, 60, 719–725. [Google Scholar] [CrossRef]
  39. Ribeiro, T.; Mascarenhas Saraiva, M.J.; Afonso, J.; Cardoso, P.; Mendes, F.; Martins, M.; Andrade, A.P.; Cardoso, H.; Mascarenhas Saraiva, M.; Ferreira, J.; et al. Design of a Convolutional Neural Network as a Deep Learning Tool for the Automatic Classification of Small-Bowel Cleansing in Capsule Endoscopy. Medicina 2023, 59, 810. [Google Scholar] [CrossRef]
  40. Cherubini, A.; Dinh, N.N. A Review of the Technology, Training, and Assessment Methods for the First Real-Time AI-Enhanced Medical Device for Endoscopy. Bioengineering 2023, 10, 404. [Google Scholar] [CrossRef]
  41. Singh, P.; Singla, V.; Bopanna, S.; Shawl, M.R.; Garg, P.; Agrawal, J.; Arya, A.; Mittal, V.; Bhargava, R.; Madan, K. Safety and efficacy of the novel motorized power spiral enteroscopy: A single-center experience. DEN Open 2023, 3, e148. [Google Scholar] [CrossRef] [PubMed]
  42. Rondonotti, E.; Pennazio, M. Colon capsule for panendoscopy: A narrow window of opportunity. Endosc. Int. Open 2021, 9, E1860–E1862. [Google Scholar] [CrossRef] [PubMed]
  43. Vuik, F.E.R.; Moen, S.; Spaander, M.C.W. Colon capsule endoscopy as panendoscopy: Using current knowledge to enhance possibilities. Endosc. Int. Open 2022, 10, E584. [Google Scholar] [CrossRef] [PubMed]
  44. Majtner, T.; Brodersen, J.B.; Herp, J.; Kjeldsen, J.; Halling, M.L.; Jensen, M.D. A deep learning framework for autonomous detection and classification of Crohn’s disease lesions in the small bowel and colon with capsule endoscopy. Endosc. Int. Open 2021, 9, E1361–E1370. [Google Scholar] [CrossRef]
  45. Gilabert, P.; Vitria, J.; Laiz, P.; Malagelada, C.; Watson, A.; Wenzek, H.; Segui, S. Artificial intelligence to improve polyp detection and screening time in colon capsule endoscopy. Front. Med. 2022, 9, 1000726. [Google Scholar] [CrossRef]
Figure 1. Study flow chart for the training and testing phases. DAE—Device-assisted enteroscopy. The term lesion refers to the presence of clinically relevant lesions.
Figure 1. Study flow chart for the training and testing phases. DAE—Device-assisted enteroscopy. The term lesion refers to the presence of clinically relevant lesions.
Cancers 16 00208 g001
Figure 2. Output obtained from the convolutional neural network. The bars are a representation of the estimated probability by the CNN. The model output was given by the finding with the highest probability. The blue bars represent a correct prediction, whereas incorrect predictions are represented by grey bars.
Figure 2. Output obtained from the convolutional neural network. The bars are a representation of the estimated probability by the CNN. The model output was given by the finding with the highest probability. The blue bars represent a correct prediction, whereas incorrect predictions are represented by grey bars.
Cancers 16 00208 g002
Figure 3. Heatmaps generated by the convolutional neural network selecting the image location responsible for the identification of clinically relevant panendoscopic lesions. The given probability represents the level of certainty in lesion prediction.
Figure 3. Heatmaps generated by the convolutional neural network selecting the image location responsible for the identification of clinically relevant panendoscopic lesions. The given probability represents the level of certainty in lesion prediction.
Cancers 16 00208 g003
Table 1. Confusion matrix of automatic detection versus the expert’s classification in the testing dataset of the CNN model. Number of cases (relative frequency).
Table 1. Confusion matrix of automatic detection versus the expert’s classification in the testing dataset of the CNN model. Number of cases (relative frequency).
Experts Classification
Normal MucosaClinically Relevant Lesions
CNN ClassificationNormal mucosa3168 (0.97)96 (0.03)
Clinically relevant lesions34 (0.04)769 (0.96)
Table 2. Performance measures of the 5-fold cross validation of the training dataset and testing dataset for panendoscopic detection of clinically relevant lesions. N—number of patients. Sn—sensitivity. Sp—specificity. PPV—positive predictive value. NPV—negative predictive value. Acc—accuracy. ()—95% confidence interval values.
Table 2. Performance measures of the 5-fold cross validation of the training dataset and testing dataset for panendoscopic detection of clinically relevant lesions. N—number of patients. Sn—sensitivity. Sp—specificity. PPV—positive predictive value. NPV—negative predictive value. Acc—accuracy. ()—95% confidence interval values.
SnSpPPVNPVAcc
Fold 10.870.950.810.960.93
Fold 20.870.970.900.960.95
Fold 30.890.990.970.970.97
Fold 40.910.990.970.980.97
Fold 50.900.990.980.970.97
Training dataset mean
N = 38,599
0.887
(0.880–0.895)
0.980
(0.978–0.981)
0.926
(0.920–0.931)
0.970
(0.968–0.972)
0.960
(0.958–0.962)
Testing dataset
N = 4068
0.889
(0.866–0.909)
0.989
(0.985–0.993)
0.958
(0.942–0.969)
0.971
(0.965–0.976)
0.968
(0.962–0.973)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mendes, F.; Mascarenhas, M.; Ribeiro, T.; Afonso, J.; Cardoso, P.; Martins, M.; Cardoso, H.; Andrade, P.; Ferreira, J.P.S.; Mascarenhas Saraiva, M.; et al. Artificial Intelligence and Panendoscopy—Automatic Detection of Clinically Relevant Lesions in Multibrand Device-Assisted Enteroscopy. Cancers 2024, 16, 208. https://doi.org/10.3390/cancers16010208

AMA Style

Mendes F, Mascarenhas M, Ribeiro T, Afonso J, Cardoso P, Martins M, Cardoso H, Andrade P, Ferreira JPS, Mascarenhas Saraiva M, et al. Artificial Intelligence and Panendoscopy—Automatic Detection of Clinically Relevant Lesions in Multibrand Device-Assisted Enteroscopy. Cancers. 2024; 16(1):208. https://doi.org/10.3390/cancers16010208

Chicago/Turabian Style

Mendes, Francisco, Miguel Mascarenhas, Tiago Ribeiro, João Afonso, Pedro Cardoso, Miguel Martins, Hélder Cardoso, Patrícia Andrade, João P. S. Ferreira, Miguel Mascarenhas Saraiva, and et al. 2024. "Artificial Intelligence and Panendoscopy—Automatic Detection of Clinically Relevant Lesions in Multibrand Device-Assisted Enteroscopy" Cancers 16, no. 1: 208. https://doi.org/10.3390/cancers16010208

APA Style

Mendes, F., Mascarenhas, M., Ribeiro, T., Afonso, J., Cardoso, P., Martins, M., Cardoso, H., Andrade, P., Ferreira, J. P. S., Mascarenhas Saraiva, M., & Macedo, G. (2024). Artificial Intelligence and Panendoscopy—Automatic Detection of Clinically Relevant Lesions in Multibrand Device-Assisted Enteroscopy. Cancers, 16(1), 208. https://doi.org/10.3390/cancers16010208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop