Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model
Abstract
:1. Introduction
2. Materials and Methods
2.1. Dataset
2.2. Data Preprocessing
- Rotation:
- Width and height shifts:
- Shear transformation:
- Zoom:
- Horizontal flip:
2.3. DL Models
2.4. Model Design and Description
Algorithm 1: Novel U-MaskNet Segmentation Model |
1: Input: 2: Initialize: 3: for epoch = 4: for batch = 5: 6: 7: Forward Pass: 8: 9: 10: Multi-Scale Feature Integration: 11: 12: 13: Compute Loss: 14: 15: 16: 17: Advanced Loss Functions: 18: 19: 20: Backward Pass and Optimization: 21: 22: 23: end for 24: end for 25: Output: 26: Trained novel model with updated parameters |
- Convolutional layer: A convolutional layer in a neural network takes in an input with spatial dimensions and then applies the convolution operation to extract the features. Every convolutional layer employs many trainable kernels that scan the given picture and generate feature maps. This is crucial for capturing the detail in the local space, which includes edges, textures, and other spatial frequencies in the data [43,44]:where is the convolutional filter, is the input feature map, and is the bias term.
- Activation function (ReLU): The rectified linear unit (ReLU) is used in neural networks, which is an activation function that is applied to make the model non-linear [45,46]. ReLU activates only the positive channels of the input, ignoring the negative part of the input. The above benefit, in turn, helps to make the training of the network converge faster and reduces the effects of the problem of vanishing gradients:where is the rectified linear unit activation function, and is the input feature map.
- Max-pooling: Max-pooling is a down-sampling operation that decreases the size of the input feature map in the vertical and horizontal directions, conserving significant features [47]. This is carried out by choosing the maximum intensity value from a group of neighboring pixels in a particular window, ensuring a form of spatial invariance and, at the same time, decreasing the number of computations to be performed:where is the output of the max-pooling operation, and represents the pixels within the pooling window.
- Up-sampling and concatenation: Up-sampling is another operation that reconstructs the height and width dimensions of the feature map and is commonly used in the decoder section of the network to bring back the resolution of the original image. This is quite frequently performed using methods such as nearest-neighbor interpolation, bilinear interpolation, or the learned transposed convolution to generate a higher-resolution feature map. Concatenation is an operation that lays out, either in a horizontal or vertical fashion, two or more feature maps. In the architecture of the U-Net, it is utilized to connect the encoder and decoder streams of the network [48]. This combined operation is beneficial in terms of maintaining spatial information, as features from different levels of the network are merged, while retaining both high-level context and low-level spatial details:where is the concatenated feature map, is the feature map from the decoder, and is the corresponding feature map from the encoder.
- Region Proposal Network: The RPN is a neural network that generates proposals of the object or bounding boxes from the input image. It produces a set of rectangular object proposals with differences in the size and the ratio of width to height. These proposals act as the prior beliefs on where the objects might be in the image:where is the set of bounding box coordinates (x, y, width, and height), and is the Region Proposal Network.
- RoI Align: RoI Align is a function that is used for the extraction of the fixed-size feature maps from non-uniform input feature maps. It properly warps the features extracted from the input image to the proposed regions, which removes the quantization errors that are usual in RoI Pooling:where is the region of interest aligned feature map, and is the set of bounding box coordinates (x, y, width, and height).
- Mask prediction: Mask prediction is another step in Mask R-CNN, wherein a binary mask is produced regarding the RoI to predict an object’s shape within the RoI. This mask highlights the pixels that belong to the object:where is the predicted segmentation mask, and is the sigmoid activation function.
- Feature integration: In feature integration of the proposed U-MaskNet, features derived from the U-Net component and the Mask R-CNN component are integrated. This integration made sure that while one received fine-grained pixel-wise segmentation details, the other received the instance-level feature map, all of which contributed to the improvement of the segmentation:where is the multi-scale feature from U-Net.
- Multi-scale feature map integration: Multi-scale feature map integration concerns combining multi-scale feature maps from U-Net and Mask R-CNN networks in order to obtain superior feature maps. This integrated feature map maintains the details and contextual information as multi-scale, which enhances the machinery of the segmentation performance:where is the integrated feature map, and is the multi-scale feature from U-Net.
- Segmentation loss (binary cross-entropy): Segmentation loss, known as binary cross-entropy (BCE) loss, is used for estimating the difference between the predicted segmentation map and the actual segmentation map. It measures how close the pixel-wise probabilities that are predicted are to the actual labels:where is the segmentation loss, is the number of samples, is the predicted probability, and is the ground truth label.
- Bounding box regression loss: The bounding box regression loss is used for evaluating the conjunction of the regression and the border of an object from the predicted bounding box coordinates to the ground truth ones. It makes sure that the predicted bounding boxes have high and strict levels of accuracy in terms of the actual size of the objects:where is the bounding box regression loss, is the smooth L1 function, are the ground truth bounding box coordinates, and are the predicted bounding box coordinates.
- Dice loss: Dice loss is applied to estimate the dissimilarity between the segmentation map that the model predicts and the true one. It is particularly useful in handling class imbalance since it tackles the area of interest only:where is the Dice loss, is the predicted probability, and is the ground truth label.
- Total loss: Total loss in U-MaskNet is the combination of segmentation loss, Dice loss, and bounding box regression loss. In this way, this composite loss function guarantees to learn accurate segmentation maps and bounding boxes and, at the same time, to perform a good handling of class imbalance:where is the total loss.
2.5. Evaluation Metrics
3. Experimental Results
3.1. Experimentation with DeepLabv3+
3.2. Experimentation with Fully Convolutional Network (FCN)
3.3. Experimentation with DeepMask
3.4. Novel Model (U-MaskNet) Evaluation and Segmentation Results
3.5. Confusion Matrix Analysis
3.6. Model Evaluation Metrics Comparison
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Arnold, M.; Abnet, C.C.; Neale, R.E.; Vignat, J.; Giovannucci, E.L.; McGlynn, K.A.; Bray, F. Global Burden of 5 Major Types of Gastrointestinal Cancer. Gastroenterology 2020, 159, 335–349.e15. [Google Scholar] [CrossRef] [PubMed]
- Yamao, K.; Kitano, M.; Takenaka, M.; Minaga, K.; Sakurai, T.; Watanabe, T.; Kayahara, T.; Yoshikawa, T.; Yamashita, Y.; Asada, M.; et al. Outcomes of endoscopic biliary drainage in pancreatic cancer patients with an indwelling gastroduodenal stent: A multicenter cohort study in West Japan. Gastrointest. Endosc. 2018, 88, 66–75.e2. [Google Scholar] [CrossRef]
- Tajbakhsh, N.; Jeyaseelan, L.; Li, Q.; Chiang, J.N.; Wu, Z.; Ding, X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med. Image Anal. 2020, 63, 101693. [Google Scholar] [CrossRef] [PubMed]
- Ronneberger, O.F.P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
- He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
- Pogorelov, K.; Randel, K.R.; Griwodz, C.; Eskeland, S.L.; de Lange, T.; Johansen, D.; Spampinato, C.; Dang-Nguyen, D.-T.; Lux, M.; Schmidt, P.T.; et al. KVASIR. In Proceedings of the 8th ACM on Multimedia Systems Conference, New York, NY, USA, 20 June 2017; pp. 164–169. [Google Scholar]
- Lu, W.; Zhou, Y.; Wan, G.; Hou, S.; Song, S. L3-Net: Towards Learning Based LiDAR Localization for Autonomous Driving. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 6382–6391. [Google Scholar]
- Zhu, J.; Ge, M.; Chang, Z.; Dong, W. CRCNet: Global-local context and multi-modality cross attention for polyp segmentation. Biomed. Signal Process. Control 2023, 83, 104593. [Google Scholar] [CrossRef]
- Nguyen, D.C.; Nguyen, H.L. PolyPooling: An accurate polyp segmentation from colonoscopy images. Biomed. Signal Process. Control 2024, 92, 105979. [Google Scholar] [CrossRef]
- Karthikha, R.; Jamal, D.N.; Rafiammal, S.S. An approach of polyp segmentation from colonoscopy images using Dilated-U-Net-Seg—A deep learning network. Biomed. Signal Process. Control 2024, 93, 106197. [Google Scholar] [CrossRef]
- Rajasekar, D.; Theja, G.; Prusty, M.R.; Chinara, S. Efficient colorectal polyp segmentation using wavelet transformation and AdaptUNet: A hybrid U-Net. Heliyon 2024, 10, e33655. [Google Scholar] [CrossRef]
- Haider, A.; Arsalan, M.; Nam, S.H.; Hong, J.S.; Sultan, H.; Park, K.R. Multi-scale feature retention and aggregation for colorectal cancer diagnosis using gastrointestinal images. Eng. Appl. Artif. Intell. 2023, 125, 106749. [Google Scholar] [CrossRef]
- Huo, X.; Tian, S.; Yang, Y.; Yu, L.; Zhang, W.; Li, A. SPA: Self-Peripheral-Attention for central–peripheral interactions in endoscopic image classification and segmentation. Expert Syst. Appl. 2024, 245, 123053. [Google Scholar] [CrossRef]
- Içek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; pp. 424–432. [Google Scholar] [CrossRef]
- Farhadi, A.; Redmon, J. YOLOv3: An Incremental Improvement. In Computer Vision and Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Leibetseder, A.; Kletz, S.; Schoeffmann, K.; Keckstein, S.; Keckstein, J. GLENDA: Gynecologic Laparoscopy Endometriosis Dataset. In International Conference on Multimedia Modeling; Springer: Cham, Switzerland, 2020; pp. 439–450. [Google Scholar] [CrossRef]
- Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
- Bali, M.; Mahara, T. Comparison of Affine and DCGAN-based Data Augmentation Techniques for Chest X-Ray Classification. Procedia Comput. Sci. 2023, 218, 283–290. [Google Scholar] [CrossRef]
- Smirnov, E.A.; Timoshenko, D.M.; Andrianov, S.N. Comparison of Regularization Methods for ImageNet Classification with Deep Convolutional Neural Networks. AASRI Procedia 2014, 6, 89–94. [Google Scholar] [CrossRef]
- Rai, H.M. Cancer detection and segmentation using machine learning and deep learning techniques: A review. Multimed. Tools Appl. 2023, 83, 27001–27035. [Google Scholar] [CrossRef]
- Rai, H.M.; Yoo, J. Analysis of Colorectal and Gastric Cancer Classification: A Mathematical Insight Utilizing Traditional Machine Learning Classifiers. Mathematics 2023, 11, 4937. [Google Scholar] [CrossRef]
- Barillaro, L. Deep Learning Platforms: Keras. In Reference Module in Life Sciences; Elsevier: Amsterdam, The Netherlands, 2024. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. Hybrid adaptive algorithm based on wavelet transform and independent component analysis for denoising of MRI images. Meas. J. Int. Meas. Confed. 2019, 144, 72–82. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. 2D MRI image analysis and brain tumor detection using deep learning CNN model LeU-Net. Multimed. Tools Appl. 2021, 80, 36111–36141. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. Hybrid CNN-LSTM deep learning model and ensemble technique for automatic detection of myocardial infarction using big ECG data. Appl. Intell. 2021, 52, 5366–5384. [Google Scholar] [CrossRef]
- Banik, D.; Roy, K.; Krejcar, O.; Bhattacharjee, D. dHBLSN: A diligent hierarchical broad learning system network for cogent polyp segmentation. Knowl. -Based Syst. 2024, 300, 112228. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. Detection of brain abnormality by a novel Lu-Net deep neural CNN model from MR images. Mach. Learn. Appl. 2020, 2, 100004. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. A unique feature extraction using MRDWT for automatic classification of abnormal heartbeat from ECG big data with Multilayered Probabilistic Neural Network classifier. Appl. Soft Comput. 2018, 72, 596–608. [Google Scholar] [CrossRef]
- KahsayGebreslassie, A.; Gezahegn, Y.G.; Hagos, M.T.; Ibenthal, A.; Pooja. Automated Gastrointestinal Disease Recognition for Endoscopic Images. In Proceedings of the 2019 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS), Greater Noida, India, 18–19 October 2019; pp. 312–316. [Google Scholar]
- Rai, H.M.; Dashkevych, S.; Yoo, J. Next-Generation Diagnostics: The Impact of Synthetic Data Generation on the Detection of Breast Cancer from Ultrasound Imaging. Mathematics 2024, 12, 2808. [Google Scholar] [CrossRef]
- Rai, H.M.; Yoo, J.; Dashkevych, S. GAN-SkipNet: A Solution for Data Imbalance in Cardiac Arrhythmia Detection Using Electrocardiogram Signals from a Benchmark Dataset. Mathematics 2024, 12, 2693. [Google Scholar] [CrossRef]
- Pillai, R.; Sharma, N.; Gupta, R. Detection & Classification of Abnormalities in GI Tract through MobileNetV3 Transfer Learning Model. In Proceedings of the 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), Delhi, India, 6–8 July 2023; pp. 1–6. [Google Scholar]
- Dookhee, S. Gastrointestinal Endoscopic Image Classification using Transfer Learning. In Proceedings of the 2023 IEEE Third International Conference on Signal, Control and Communication (SCC), Hammamet, Tunisia, 1–3 December 2023; pp. 1–6. [Google Scholar]
- Hossain, S.; Rahman, M.; Syeed, M.M.; Uddin, M.F.; Hasan, M.; Hossain, A.; Ksibi, A.; Jamjoom, M.M.; Ullah, Z.; Samad, A. DeepPoly: Deep Learning-Based Polyps Segmentation and Classification for Autonomous Colonoscopy Examination. IEEE Access 2023, 11, 95889–95902. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar] [CrossRef]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
- Suri, J.S.; Bhagawati, M.; Agarwal, S.; Paul, S.; Pandey, A.; Gupta, S.K.; Saba, L.; Paraskevas, K.I.; Khanna, N.N.; Laird, J.R.; et al. UNet Deep Learning Architecture for Segmentation of Vascular and Non-Vascular Images: A Microscopic Look at UNet Components Buffered With Pruning, Explainable Artificial Intelligence, and Bias. IEEE Access 2022, 11, 595–645. [Google Scholar] [CrossRef]
- Ortega-Ruíz, M.A.; Karabağ, C.; Roman-Rangel, E.; Reyes-Aldasoro, C.C. DRD-UNet, a UNet-Like Architecture for Multi-Class Breast Cancer Semantic Segmentation. IEEE Access 2024, 12, 40412–40424. [Google Scholar] [CrossRef]
- Rai, H.M.; Yoo, J.; Dashkevych, S. Two-headed UNetEfficientNets for parallel execution of segmentation and classification of brain tumors: Incorporating postprocessing techniques with connected component labelling. J. Cancer Res. Clin. Oncol. 2024, 150, 220. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K.; Dashkevich, S. Automatic and accurate abnormality detection from brain MR images using a novel hybrid UnetResNext-50 deep CNN model. Biomed. Signal Process. Control 2021, 66, 102477. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K. A Novel Adaptive Feature Extraction for Detection of Cardiac Arrhythmias Using Hybrid Technique MRDWT & MPNN Classifier from ECG Big Data. Big Data Res. 2018, 12, 13–22. [Google Scholar] [CrossRef]
- Rai, H.M.; Chatterjee, K.; Dashkevych, S. The prediction of cardiac abnormality and enhancement in minority class accuracy from imbalanced ECG signals using modified deep neural network models. Comput. Biol. Med. 2022, 150, 106142. [Google Scholar] [CrossRef] [PubMed]
- Rai, H.M.; Shukla, K.K.; Goya, Y.; Amanzholova, S.; Nessibely, A. IoT-based real-time monitoring and control system for tomato cultivation. Procedia Comput. Sci. 2024, 241, 433–438. [Google Scholar] [CrossRef]
- Goyal, Y.; Rai, H.M.; Aggarwal, M.; Saxena, K.; Amanzholova, S. Revolutionizing Skin Cancer Detection: A Comprehensive Review of Deep Learning Methods. In Proceedings of the ICIMMI 2023: International Conference on Information Management & Machine Intelligence, Jaipur, India, 23–25 November 2023; pp. 1–6. [Google Scholar]
- Jha, K.; Pasbola, M.; Rai, H.M.; Amanzholova, S. Utilizing Smartwatches and Deep Learning Models for Enhanced Avalanche Victim Identification, Localization, and Efficient Recovery Strategies: An In-depth Study. In Proceedings of the ICIMMI 2023: International Conference on Information Management & Machine Intelligence, Jaipur, India, 23–25 November 2023; pp. 1–5. [Google Scholar]
- Patrick, U.; Rao, S.K.; Jagan, B.O.L.; Rai, H.M.; Agarwal, S.; Pak, W. Radar-Based Target Tracking Using Deep Learning Approaches with Unscented Kalman Filter. Appl. Sci. 2024, 14, 8332. [Google Scholar] [CrossRef]
- Jiang, A.; Yan, N.; Wang, F.; Huang, H.; Zhu, H.; Wei, B. Visible Image Recognition of Power Transformer Equipment Based on Mask R-CNN. In Proceedings of the 2019 IEEE Sustainable Power and Energy Conference (iSPEC), Beijing, China, 21–23 November 2019; pp. 657–661. [Google Scholar]
- Mewada, S.; Patel, F.; Degadwala, S.; Vyas, D. Enhancing Raga Identification in Indian Classical Music with FCN-based Models. In Proceedings of the 2023 International Conference on Sustainable Communication Networks and Application (ICSCNA), Theni, India, 15–17 November 2023; pp. 980–985. [Google Scholar]
- Liu, J. Road Crack Detection Using HDD LOSS and Dual Attention Module with DeepLabv3+. In Proceedings of the 2023 3rd International Conference on Digital Society and Intelligent Systems (DSInS), Chengdu, China, 10–12 November 2023; pp. 148–152. [Google Scholar]
- Nguyen, T.D.; Shinya, A.; Harada, T.; Thawonmas, R. Segmentation Mask Refinement Using Image Transformations. IEEE Access 2017, 5, 26409–26418. [Google Scholar] [CrossRef]
- Nakagawa, T.; Nishimoto, M.; Takaiwa, T. Disseminated Epidermolytic Acanthoma Revealed by PUVA. Dermatology 1986, 173, 150–153. [Google Scholar] [CrossRef] [PubMed]
- Duarte, E.P., Jr.; Godoi, A.F.B. Reliable Content Distribution in P2P Networks Based on Peer Groups. Int. J. Internet Distrib. Syst. 2014, 02, 5–14. [Google Scholar] [CrossRef]
- Zhang, Z.; Liu, Q.; Wang, Y. Road Extraction by Deep Residual U-Net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef]
- Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R.S., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11045. [Google Scholar] [CrossRef]
- Zhang, R.; Li, G.; Li, Z.; Cui, S.; Qian, D.; Yu, Y. Adaptive Context Selection for Polyp Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020. [Google Scholar]
- Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. PraNet: Parallel Reverse Attention Network for Polyp Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2020. [Google Scholar]
- Tomar, N.K.; Jha, D.; Bagci, U.; Ali, S. TGANet: Text-guided attention for improved polyp segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2022. [Google Scholar]
- Tong, Y.; Chen, Z.; Zhou, Z.; Hu, Y.; Li, X.; Qiao, X. An Edge-Enhanced Network for Polyp Segmentation. Bioengineering 2024, 11, 959. [Google Scholar] [CrossRef]
- Hu, Y.; Wei, J.; Jiang, Y.; Li, H.; Cui, S.; Li, Z.; Wu, S. MixPolyp: Integrating Mask, Box and Scribble Supervision for Enhanced Polyp Segmentation. arXiv 2024, arXiv:2409.16774. [Google Scholar]
- Moqurrab, S.A.; Rai, H.M.; Yoo, J. HRIDM: Hybrid Residual/Inception-Based Deeper Model for Arrhythmia Detection from Large Sets of 12-Lead ECG Recordings. Algorithms 2024, 17, 364. [Google Scholar] [CrossRef]
- Boeriu, A. Narrow-band imaging with magnifying endoscopy for the evaluation of gastrointestinal lesions. World J. Gastrointest. Endosc. 2015, 7, 110–120. [Google Scholar] [CrossRef] [PubMed]
- Li, W.; Nie, X.; Li, F.; Huang, Z.; Zeng, G. FMCA-Net: A feature secondary multiplexing and dilated convolutional attention polyp segmentation network based on pyramid vision transformer. Expert Syst. Appl. 2024, 260, 125419. [Google Scholar] [CrossRef]
- Nanni, L.; Fusaro, D.; Fantozzi, C.; Pretto, A. Improving Existing Segmentators Performance with Zero-Shot Segmentators. Entropy 2023, 25, 1502. [Google Scholar] [CrossRef] [PubMed]
- Wei, X.; Sun, J.; Su, P.; Wan, H.; Ning, Z. BCL-Former: Localized Transformer Fusion with Balanced Constraint for polyp image segmentation. Comput. Biol. Med. 2024, 182, 109182. [Google Scholar] [CrossRef]
- Nanni, L.; Lumini, A.; Fantozzi, C. Exploring the Potential of Ensembles of Deep Learning Networks for Image Segmentation. Information 2023, 14, 657. [Google Scholar] [CrossRef]
S. No. | Category | Number of Files |
0. | Dyed Lifted Polyps | 1000 |
1. | Dyed Resection Margins | 1000 |
2. | Esophagitis | 1000 |
3. | Normal Cecum | 1000 |
4. | Normal Pylorus | 1000 |
5. | Normal Z Line | 1000 |
6. | Polyps | 1000 |
7. | Ulcerative Colitis | 1000 |
Symbols | Description |
---|---|
Dataset consisting of image–label pairs | |
Learning rate | |
Total number of epochs | |
Batch size | |
Model parameters | |
Total number of samples | |
Augmentation function | |
Parameters of the U-Net model | |
Parameters of the Mask R-CNN model | |
Batch of input images | |
Batch of ground truth tables | |
Augmented input images | |
Augmented ground truth tables | |
Segmentation output of U-Net | |
Region of interest (RoI) proposals from Mask R-CNN | |
Bounding boxes from Mask R-CNN | |
Mask predictions from Mask R-CNN | |
Multi-scale features from U-Net | |
Integrated feature maps | |
Segmentation loss for U-Net | |
Segmentation loss for Mask R-CNN | |
Bounding box regression loss | |
Dice coefficient loss | |
Total loss function | |
Gradient of the total loss with respect to U-Net parameters | |
Gradient of the total loss with respect to Mask R-CNN parameters |
Evaluation Criteria | DeepLabv3+ | FCN | DeepMask | U-MaskNet |
---|---|---|---|---|
Precision | 87.46 | 98.46 | 98.45 | 98.85 |
Recall | 87.41 | 98.39 | 98.03 | 98.49 |
Dice coefficient | 87.43 | 94.12 | 94.25 | 94.35 |
IoU | 77.70 | 88.90 | 89.14 | 89.31 |
Loss | 13.26 | 5.18 | 5.11 | 4.88 |
F1 score | 91.96 | 98.44 | 98.25 | 98.68 |
AUC ROC | 98.86 | 99.94 | 99.93 | 99.96 |
Method | Precision | Recall | Dice | IoU |
---|---|---|---|---|
PolyPooling [9] | 92.9 | 94.6 | 93.7 | 88.5 |
CRCNet [8] | 92.9 | 94.6 | 93.7 | 88.5 |
U-Net [55] | 82.9 | 81.5 | 79.9 | 83.2 |
U-Net++ [56] | 89.3 | 91.0 | 88.1 | 81.7 |
ASCNet [57] | 92.2 | 90.0 | 91.3 | 90.4 |
PraNet [58] | 91.2 | 91.3 | 89.8 | 83.3 |
TGANet [59] | 91.3 | 91.2 | 88.8 | 83.4 |
Polyp [60] | NA | NA | 93.1 | 88.0 |
MixPolyp [61] | NA | NA | 85.9 | 78.5 |
Proposed (U-MaskNet) | 98.5 | 98.4 | 94.3 | 89.31 |
Method | Precision | Recall | Dice | IoU |
---|---|---|---|---|
LeNet-5 | 91.6 | 90.7 | 91.1 | 88.9 |
AlexNet | 93.4 | 93.5 | 93.4 | 87.2 |
VGG-16 | 94.7 | 93.7 | 92.8 | 86.9 |
ResNet-50 | 96.7 | 97.1 | 93.7 | 85.4 |
Inception Network | 95.4 | 96.7 | 93.4 | 86.7 |
Proposed (U-MaskNet) | 98.5 | 98.4 | 94.3 | 89.31 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pal, A.; Rai, H.M.; Frej, M.B.H.; Razaque, A. Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model. Life 2024, 14, 1488. https://doi.org/10.3390/life14111488
Pal A, Rai HM, Frej MBH, Razaque A. Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model. Life. 2024; 14(11):1488. https://doi.org/10.3390/life14111488
Chicago/Turabian StylePal, Aditya, Hari Mohan Rai, Mohamed Ben Haj Frej, and Abdul Razaque. 2024. "Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model" Life 14, no. 11: 1488. https://doi.org/10.3390/life14111488
APA StylePal, A., Rai, H. M., Frej, M. B. H., & Razaque, A. (2024). Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model. Life, 14(11), 1488. https://doi.org/10.3390/life14111488