Next Article in Journal
Comparison of Machine Learning Models and the Fatty Liver Index in Predicting Lean Fatty Liver
Next Article in Special Issue
Identifying Effective Feature Selection Methods for Alzheimer’s Disease Biomarker Gene Detection Using Machine Learning
Previous Article in Journal
Safety and Diagnostic Accuracy of the Transnasal Approach for Endobronchial Ultrasound-Guided Transbronchial Needle Aspiration (EBUS-TBNA)
Previous Article in Special Issue
Hybrid Techniques for the Diagnosis of Acute Lymphoblastic Leukemia Based on Fusion of CNN Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bi-FPN-Based Encoder–Decoder Model for Lung Nodule Image Segmentation

by
Chandra Sekhara Rao Annavarapu
1,
Samson Anosh Babu Parisapogu
2,*,†,
Nikhil Varma Keetha
1,†,
Praveen Kumar Donta
3,* and
Gurindapalli Rajita
4
1
Indian Institute of Technology (Indian School of Mines), Dhanbad 826004, India
2
Krishna Chaitanya Institute of Technology and Sciences, Markapur 523316, India
3
Distributed Systems Group, TU Wien, 1040 Vienna, Austria
4
Department of ECE, GIET University, Gunupur 765022, India
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(8), 1406; https://doi.org/10.3390/diagnostics13081406
Submission received: 12 March 2023 / Revised: 2 April 2023 / Accepted: 4 April 2023 / Published: 13 April 2023
(This article belongs to the Special Issue Deep Learning for Early Detection of Cancer)

Abstract

:
Early detection and analysis of lung cancer involve a precise and efficient lung nodule segmentation in computed tomography (CT) images. However, the anonymous shapes, visual features, and surroundings of the nodules as observed in the CT images pose a challenging and critical problem to the robust segmentation of lung nodules. This article proposes a resource-efficient model architecture: an end-to-end deep learning approach for lung nodule segmentation. It incorporates a Bi-FPN (bidirectional feature network) between an encoder and a decoder architecture. Furthermore, it uses the Mish activation function and class weights of masks with the aim of enhancing the efficiency of the segmentation. The proposed model was extensively trained and evaluated on the publicly available LUNA-16 dataset consisting of 1186 lung nodules. To increase the probability of the suitable class of each voxel in the mask, a weighted binary cross-entropy loss of each sample of training was utilized as network training parameter. Moreover, on the account of further evaluation of robustness, the proposed model was evaluated on the QIN Lung CT dataset. The results of the evaluation show that the proposed architecture outperforms existing deep learning models such as U-Net with a Dice Similarity Coefficient of 82.82% and 81.66% on both datasets.

1. Introduction

According to data released by the World Health Organization (WHO) on 3 February 2020, cancer is one of the leading causes of premature death in 134 of 183 countries. It has been observed that, especially in 2018, most of the prominent cancer deaths are due to lung cancer (1.76 million deaths). Detection and analysis of the lung nodules at an early stage facilitate efficient treatment and drastically improve a patient’s chance of survival [1]. CT scans are a widely used and highly accurate format for the purpose of screening and analyzing lung nodules, especially in differentiating the nodules from other structures. Moreover, the precise segmentation of these nodules is critical, considering the heterogeneity of the size, texture, location, and shape of the nodules, and the fact that their intensity may differ within the borders [2]. There are various types of lung nodules as observed in Figure 1 such as adhesion-type (juxtapleural and juxta-vascular), isolated, cavitary, calcified, small, and ground-glass opacity (GGO) nodules [3].
Another challenge lies in the segmentation of lung nodules, which is found in the case of nodules with small diameter and intensity comparable to that of the surrounding noise, which thereby hinders the down-sampling potential of the segmentation network, where the network cannot extract more in-depth semantic network features [4]. It significantly impacts the accuracy of the extraction of feature maps of large nodules. Based on these reasons, a robust segmentation network is necessary to accommodate the large-scale nodule (various types) problem.
Recently, convolutional neural networks (CNN) have become the mainstream architecture in the field of computer vision. One such architecture, the U-Net, is an encoder–decoder-like CNN architecture, which has shown exceptional results in segmentation of biomedical images [5]. Many modified U-Net architectures have achieved significant results in different domains of biomedical imaging. However, CNN architectures that are implemented for the task of lung nodule segmentation are still immature. Therefore, the development of advanced architectures dealing with the shortcomings of previous architectures is essential.
To deal with the challenges of efficient feature extraction and adaptation to heterogeneity of lung nodules, this paper proposes a modified U-Net architecture with a weighted bidirectional feature network (U-Det), which is appropriate for the segmentation of many forms of lung nodules. Figure 2 illustrates the pipeline of the proposed model.

Contributions

The list of the following elements provides technical contributions of this research:
  • The proposed U-Det model uses a bidirectional feature network (Bi-FPN), which functions as a feature enricher, integrating multi-scale feature fusion for efficient feature extraction.
  • Applying a data augmentation technique to deal with the small-size dataset prevents the model from over-fitting and provides better segmentation results.
  • Implementing the Mish activation function, due to its strong regularization effects, provides enhanced model training and segmentation efficiency.
  • Comparing the proposed U-Det model to the existing U-Net to the existing U-Net shows its high segmentation performance on small nodules and various categories of other pulmonary nodules.

2. Background and Related Work

This section describes the existing conventional and machine-learning-based segmentation approaches for lung nodule segmentation.

2.1. Conventional Approaches

Many conventional approaches, such as morphological methods, region-growing processes, energy-based optimization techniques, and machine learning methods, were proposed for lung nodule segmentation in the literature. In morphological methods, morphology-based operations and those based on shape hypothesis were applied to isolate lung nodules by selecting the connected region [6]. However, the isolation of lung nodules using morphological operations did not perform efficiently [7]. Following that, region-growing methods were proposed to improve lung nodule segmentation but it was observed that these methods were unable to segment juxta-vascular and juxtapleural nodules and were only well suited to isolate calcified nodules [3]. In this case, for region-growing methods, the convergence condition and irregular-shaped nodules create difficulty due to the breach of the shape hypothesis. Similarly, energy based-optimization methods were also proposed for lung nodule segmentation. These methods consist of a level set function for the characterization of the image, and an energy function to typically turn the segmentation task into an energy minimization problem [8]. However, juxtapleural nodules and low contrast nodules such as GGO nodules drastically affect the improvement of lung nodule segmentation.

2.2. Machine-Learning-Based Approaches

In the past decade, many machine-learning–based methods were proposed for lung nodule segmentation [9]. An example of one such approach is the hybrid model for classifying the lung nodules with high-level feature maps proposed by Lu et al. [10]. Further, methods based on modified support vector machines were suggested to detect small lung nodules in 3D CT scans [11,12]. Down the line, researchers also developed 3D large-scale nodule segmentation techniques based on the Hessian strategy in combination with neural networks [13]. Similarly, segmentation approaches based on multi-phase models [14], unsupervised k-means clustering, and level set algorithms [15] were also proposed.
Among the recently developed deep learning-based approaches for segmentation, CNN is a multi-layered neural network that learns to map original image files and corresponding labels hierarchically, and to change the segmentation task into the classification of voxels [16,17]. For instance, Wang et al. introduced a multi-view CNN (MV-CNN) for nodule segmentation [18]. Further, Huang et al. developed a fast and fully automated detection and segmentation approach for pulmonary nodules in thoracic CT scans using deep convolutional neural networks [19]. Moreover, the authors Subham and Raman developed a lung nodule segmentation approach Using 3-dimensional convolutional neural networks [20]. A good review is also available on pulmonary nodule detection and diagnosis using deep learning applications in computed tomography images [21]. Moreover, Havaei et al. in [22] proposed a cascaded variant of CNN (Cascaded-CNN) for brain tumor segmentation. Further extending the use of CNN to lung nodule segmentation, Shen et al. proposed a multi-crop CNN (MC-CNN) to extract salient nodule features and classify malignancy [16]. Furthermore, Wang et al. in [18] introduced a multi-view deep CNN (MV-DCNN), and Sun et al. in [23] introduced a three multi-channel region of interest (RoI)-based CNN (MCROI-CNN) for lung nodule segmentation and displayed the effectiveness of CNN over existing computer-aided diagnosis systems. Further expanding work on CNN-based architectures, Zhao et al. have advocated an enhanced pyramid deconvolution network for increasing performance on lung nodule segmentation which blends low-level fine-grained characteristics with high-level functional characteristics [24].
On the other hand, fully convolutional networks [25] are a different approach for CT image segmentation. For example, the architectures 2D U-Net and 3D U-Net, proposed by Ronnerberger et al. and Ciccek et al., respectively, are better adapted to biomedical imaging [5,26]. The fully convolutional network U-Net (FCN-UNET) architecture is a convolutional network architecture used for fast and precise segmentation of images. Similarly, the central focused CNN (CF-CNN) proposed by Wang et al. is a data-driven method without involving the shape hypothesis. It showed strong performance for the segmentation of juxtapleural nodules [27]. Moreover, Cao et al. recently proposed incorporating intensity features into the CNN architecture by implementing a dual-branch residual network (DB-ResNet) [28].
Segmentation 2D convolutional U-Network (SegU-Net), a U-Net-based architecture, has also been suggested for lung nodule segmentation [29]. For pre-processing, Wang et al. and Cao et al. implemented weighted sampling of training data to deal with the comparatively smaller size of the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) dataset [27,28]. The input slice of a CT scan is cropped to a smaller size with a random weighted sampling strategy to increase the size of the training dataset in the abovementioned method. Recently, Singadkar et al. proposed another approach toward lung nodule segmentation based on a deep deconvolutional residual network (DDRN) [30]. In DDRN, the approach was based on an encoder–decoder architecture with one-directional long-skip connections. Moreover, Wen-Fan Chen et al. developed a residual-dense-attention (RDA) U-Net network architecture for hepatocellular carcinoma segmentation [31].

3. Proposed Method

This section explains the following three phases of the proposed model: (1) data augmentation, (2) model architecture, and (3) training and post-processing. Figure 3 visualizes the proposed U-Det model architecture.

3.1. Data Augmentation

Due to the constraints of much available training data in medical image segmentation, data augmentation helps in preventing the model from over-fitting. It also improves the generalization capability of the network on data outside the training set. Along with that, it plays a vital role in building robust deep learning pipelines [32,33]. A data augmentation strategy instead of a sampling strategy was applied to input CT images of size 512 × 512 for maintaining the same input size in the proposed model. Data augmentation methods applied in the proposed network are scale, flip, shift, rotate, salt and pepper noise, and elastic deformations [34]. By applying these small transformations to images during training, variety was created in the training dataset and robustness of the proposed model was improved.

3.2. Model Architecture

The architecture of the proposed model is an end-to-end deep learning approach for lung nodule segmentation, inspired by the encoder–decoder backbone of U-Net, and the feature enricher Bi-FPN. The proposed U-Det architecture is designed to take a 512 × 512 image as an input and a 512 × 512 mask as output. It consists of three sections: the contraction, Bi-FPN, and the expansion section of depth 5, which behave similarly to an encoder, feature enricher, and decoder. Table 1 shows the corresponding layers of the model, along with their respective parameters.
Initially, the contraction section takes the augmented CT image, a slice of the CT scan. Here, the repeated application of two 3 × 3 convolutions (with the ‘same’ padding) is applied using Equation (1)
C [ m , n ] = ( I × k ) [ m , n ] = i j k [ i , j ] . I [ m i , n j ] ,
where C [ m , n ] represents the kernel convolution, I and k denote the input image and kernel, respectively, and ‘m’ and ‘n’ represent the number of ‘training samples’ and ‘features’, respectively. Further, each convolution is followed by a recently proposed non-linear Mish activation function [35], to perform strong regularization during the forward and backward pass of the model, the formula of which is shown in Equation (2)
f ( x ) = x . tanh ( ω ( x ) ) ,
where ω ( x ) is the Softplus activation function given by l n ( 1 + e x ) .
Mish implements a self-gating function, in which the input given to the gate is a scalar. The property of self-gating helps in replacing the activation functions (point-wise functions) such as rectified linear unit (ReLU). Here, the input of the gating function is a scalar with no requirement of modifying network parameters. The properties of Mish, being above-unbounded, below-bounded, non-monotonic, and smooth, play a vital part in maximizing neural network outcomes. Hence, Mish enables considerable time improvements during the forward and backward pass on Graphics Processing Unit (GPU) inference when Compute Unified Device Architecture (CUDA) is enabled and improves the efficiency of the model.
Further, a 2 × 2 max-pooling operation of stride ‘2’ is applied for down-sampling of the input image features. Here, at depth 4, a dropout layer with a dropout factor of 0.5 is used for the regularization of the model. The feature maps of the corresponding five depths of the contraction path are then fed as input to the Bi-FPN, as illustrated in Figure 3.
For efficient feature extraction, Bi-FPN implemented in the proposed method is based on conventional top-down Feature Pyramid Networks (FPN) [36]. It infuses efficient bidirectional cross-scale connections and weighted feature fusion into the model. In the Bi-FPN, multi-scale feature fusion aims to fuse features at different resolutions to obtain efficient feature extractions. In contrast, the one-way flow of information inherently limits conventional top-down FPN. Thus, Bi-FPN does not consist of nodes with only one input edge because if a node has only one input with no feature fusion, it will contribute less to the feature network to infuse different features. Moreover, the Bi-FPN has one top-down and one bottom-up path, thereby allowing the bidirectional flow of features from one depth to another in the feature network.
The Bi-FPN also incorporates additional weight for each input during feature fusion, thereby learning the importance of a particular input feature. In the Bi-FPN, dynamic learning behavior and accurate fast normalized fusion (one of the methods of incorporating weights during feature fusion) are implemented. Moreover, for improved efficiency, depth-wise separable convolution followed by batch normalization and non-linear ReLU activation function are implemented. Through the bidirectional cross-scale connections, the Bi-FPN enriches the feature maps at each depth of the network and provides an efficient fusion of features across various depths of the encoder–decoder architecture.
Incorporating a bidirectional feature network aims to improve the efficiency of feature extraction at each level of the backbone architecture and enrich the feature vectors, thus allowing a fusion of lower-level fine-grained and higher-level semantic features. As illustrated in Figure 3, the inputs of the Bi-FPN are the feature maps of the corresponding five depths of the contraction path of the backbone architecture. Then, the outputs of Bi-FPN are fed into the expansion path of the backbone network.
Inside the expansion section, the outputs of Bi-FPN are each combined with a decoder architecture to obtain a combination of lower-level fine-grained features with high-level semantic features. Each step in the expansion path consists of an up-sampling of the feature map followed by a 2 × 2 convolution (up-convolution), which halves the number of feature channels at each depth. The feature vectors obtained after up-sampling are concatenated with the corresponding feature vectors from the Bi-FPN. The concatenation operation is followed by two 3 × 3 convolutions (with the ‘same’ padding), and each is followed by the Mish activation function. In the final layer of the expansion section, the obtained 512 × 512 × 64 feature map undergoes two 3 × 3 convolutions, which are again followed by the Mish activation function. Then a 1 × 1 convolution block and a sigmoid activation function are applied as the final operation. Thus, they obtain logits corresponding to the mask of the input CT image of 512 × 512 .

3.3. Training and Post-Processing

The network training aims to increase the probability of the suitable class of each voxel in the mask. In respect to that, a weighted binary cross-entropy loss of each sample for training was utilized. The positive pixels, by the ratio of negative-to-positive voxels, in the training set were weighted to implement weighted binary cross-entropy. Since the size of the positive class in a lung nodule mask is relatively smaller compared to that of the negative class, the training set’s class weight is positive, thereby increasing the punishment for getting a positive value wrong. So the network will learn to be less biased toward outputting negative voxels due to the class imbalance in the masks. The weighted binary cross-entropy loss is formulated as follows:
L o s s = 1 N i = 1 N [ ω p × y i log y i ^ + ( 1 y i ) log ( 1 y i ^ ) ] ,
where, N represents the number of samples, ω p represents the positive prediction weights, y i represents the ground truth, and y i ^ indicates the prediction of the U-Det model.In the training approach, K-fold cross-validation [37] was utilized to obtain an accurate measure of the generalizing capability of the proposed model. Furthermore, to deal with the generation of augmented training CT images and corresponding ground truths, generators were implemented to augment input images and the generation of corresponding ground truth labels. For model optimization, the Adam model optimization algorithm [38], which updates network weights requiring little hyper-parameter fine-tuning, was utilized with the following hyper-parameters: the initial learning rate is 0.0001, Beta_1 = 0.99, Beta_2 = 0.999, and the decay rate is 1 × 10 6 . Moreover, a batch size of two samples was chosen based on the memory size of the GPU for training the model. Further, the early stopping training strategy [39] was applied to prevent over-fitting. Additionally, the strategy of reducing the learning rate of the optimizer once the model’s performance reaches a plateau was utilized. In the post-processing phase, the proposed model was designed to save the final obtained masks in a raw, metal (.mhd) format, storing volumetric data such as CT scans.

4. Data and Experiments

This section explains the details regarding the dataset, implementation, and evaluation metrics.

4.1. Data

For the experimentation and training of the proposed model, the approach utilized the publicly available dataset of the Lung Nodule Analysis 2016 (LUNA16) grand challenge [40,41,42], which is derived from the public LIDC-IDRI dataset [43,44]. In total, 888 CT scans are included in the dataset. Figure 4 illustrates the histogram of the nodule amount and diameter of nodules for the LUNA16 dataset. In addition to the LUNA16 dataset, the Quantitative Imaging Network (QIN) [45] Lung CT Segmentation dataset was used to evaluate the effectiveness of the proposed model. The QIN dataset contains 41 CT scans of non-small-cell lung cancer collected from different sources, such as LIDC, Reference Image Database to Evaluate Therapy Response (RIDER), Stanford University Medical Center, the Moffitt Cancer Center, and the Columbia University Phantom [46,47,48]. This dataset contains a total of 52 lung CT nodules across the 41 CT scans.
In the pre-processing phase, the CT images and ground truth masks were generated from CT scans, utilizing the Kaggle Data Science Bowl 2017 Competition based on annotations and a mask creation algorithm [49]. By using this method we obtained a total of 1166 CT images with corresponding ground truth masks. Further, it was partitioned into a training and a test subset containing 922 and 244 images, respectively, and we applied K-fold cross-validation of 4-folds during model training. The two subsets showed identical statistical distribution in their clinical characteristics, as depicted in Table 2. Furthermore, the same pre-processing strategy was used on the QIN dataset with the provided lung nodule locations to generate 156 CT images with corresponding ground truth masks.

4.2. Evaluation Metrics

The Dice Similarity Coefficient (DSC) was used as the key evaluation metric for assessing the segmentation performance of the U-Det model. It is a commonly used metric to calculate the difference between the outcomes of two segmentations [22,50]. In addition to the abovementioned metrics, the sensitivity (SEN) and positive predictive value (PPV) were used as additional evaluation metrics, and are formulated as follows
D S C = 2 × V ( G t S v ) V ( G t ) + V ( S v ) ,
S E N = V ( G t S v ) V ( G t ) ,
P P V = V ( G t S v ) V ( S v ) ,
where “Gt” represents the ground truth labels, and “Sv” represents the segmentation results of the proposed model. Here, the volume size measured in voxel units is represented by “V”.

4.3. Implementation Details

The experimental implementation was performed using TensorFlow (Version 2.1) deep learning framework (GPU version), Python 3.6 was the language used for coding, and CUDA 10.2 was applied for accelerated training. The experiment was carried out using Google cloud platform on a virtual instance equipped with 4 vCPUs, 15GB memory, and an SSD drive of 500 GB. During the model’s training, acceleration was performed using NVIDIA Tesla T4 GPU (14 GB video memory), and it needed about 8 h of training (20 epochs) to converge. The following link provides the source code of the implementation https://github.com/Nik-V9/U-Det, accessed on 11 March 2023.

5. Experimental Results

This section covers the details of the ablation study, overall performance of the proposed method, and experimental results.

5.1. Ablation Experiment

An ablation experiment was designed, based on the U-Net architecture with the LUNA16 test set, to verify the effectiveness of each component in the proposed architecture. The outcomes of this experiment are displayed in Table 3.

5.1.1. Effect of Mish Activation Function

From Table 3, U-Net + Mish indicates the incorporation of the Mish activation function instead of the ReLU activation function of the original U-Net architecture. The DSC score of the original U-Net was 77.84%. It can be noticed that after implementing the Mish function in U-Net, this score was increased to 78.82%. On the other hand, an encoder consisting of the contraction path of the U-Net along with the Bi-FPN was implemented. On adding the Mish activation function to the abovementioned architecture, the DSC score became 80.22%. Moreover, a version of the proposed U-Det model with ReLU was applied, which performed marginally inferior to the Mish version. It can be observed that the percentage increase in performance due to Mish was nearly 1.3. Thus it is evident that the Mish activation function is useful in the U-Det model.

5.1.2. Effect of Bi-FPN

In the description of Table 3, it can be discovered that Encoder + Bi-FPN replaced the backbone architecture with only a contraction path and the Bi-FPN functioned as the feature enricher and decoder. It can be observed that this architecture showed an improvement over the basic U-Net and achieved a DSC of 79.21%. Moreover, the ReLU version of the U-Det model is an incorporation of the Bi-FPN in the U-Net architecture; here, the DSC score showed 81.63%, which was a significant improvement over the original U-Net.
In addition to the above, even though the Bi-FPN is less computationally expensive than the U-Net architecture’s expansive path in terms of parameters, the Encoder + Bi-FPN successfully incorporates multiple feature fusions. The multi-feature fusion allows simultaneous feature map enhancement, thereby showing improvement over the U-Net architecture. In addition to that, the multiple implementations of Bi-FPN may serve as a decoder pathway, but it results in more complexity, computational expense and did not result in significant improvement in our study. Thus it can be inferred that the implementation of Bi-FPN between the expansive and contractive paths proved to be very effective in the proposed model.

5.1.3. Effect of Bi-FPN + Expansion Path

The combination of Bi-FPN and the expansion path (ReLU version of the U-Det model) was much more productive compared to the Encoder + Bi-FPN; it exhibited a DSC score of 81.63%. The addition of the expansion path of U-Net to the Encoder + Bi-FPN model helped in the proper up-sampling of low-level features and a combination of feature maps from Bi-FPN, thereby enabling the efficient fusion of high-level semantic features with low-level features.

5.1.4. Conclusion of the Ablation Study

From the observation of the DSC score of the U-Det model in Table 3 (82.82%), it is evident that the proposed U-Det model shows significant improvement over U-Net. Additionally, the effectiveness of all components and their culmination in the proposed model was verified through the ablation study.

5.2. Overall Performance

The histogram of the DSC values and the total amount of nodules, centered on every sample in the LUNA16 test set, is plotted in Figure 5 for better evaluation of the output of the U-Det model on the test set. From Figure 5, it can be quickly concluded that most of the nodules have a DSC value greater than 0.8.
For verifying the effectiveness of the Bi-FPN, the DSC results on the LUNA16 test set were compared with those of the original U-Net architecture. It can be observed that the U-Net model had a DSC of 77.84%, whereas the proposed model obtained a DSC of 82.82%, representing robust performance in the task of segmentation. Furthermore, by containing fewer parameters than the original U-Net architecture, the proposed U-Det model has shown its potential for efficient feature extraction and segmentation.
The segmentation results of complex cases, including attached (juxtapleural and juxta-vascular) and small-size nodules from both the LUNA16 and QIN Lung CT datasets, are shown in Figure 6. On observation of the results, it becomes apparent that the U-Det model outperforms the ground truth labels indicating the generalization potential of the model. The mean DSC outcomes on various lung nodule types from both datasets are shown in Table 4 and Table 5. By examining the experimental data shown in Table 4 and Table 5, it can be observed that the U-Det model’s potential for robust segmentation does not depend upon the type of nodule as it has shown exceptional performance on even small-size nodules.

6. Discussion

The following distinctions are noted when the proposed work is compared with similar other approaches.
  • To overcome the challenge of segmentation of nodules having small diameter and an intensity comparable to that of the surrounding noise, the proposed model used Bi-FPN, which functioned as a feature enricher, integrating multi-scale feature fusion for the purpose of efficient feature extraction.
  • The proposed method applied a data augmentation technique to prevent the model from over-fitting and to obtain better segmentation results.
  • The comparison of the proposed model with others showed high segmentation performance on small nodules and various other categories of pulmonary nodules.
For the experimental comparison, the results of the proposed method on LUNA16 dataset were compared with the results of methods reported in the literature such as MV-CNN [51], MCROI-CNN [23], MC-CNN [16], FCN-UNET [5], MV-DCNN [18], CF-CNN [27], and Cascaded-CNN [22]. It can be observed that the U-Det model performs comparably to human experts; the efficiency of the segmentation (in terms of DSC) of the four radiologists who worked on the LUNA16 is known to be 82.25%. Moreover, the proposed model was compared with models ranging from the original U-Net to various other recent convolution networks, including DDRN [30] and DB-ResNet [28].
In Table 6 and Table 7, and Figure 7 and Figure 8, the quantified results of the various methods on the LUNA16 test set and QIN Lung CT Segmentation dataset are represented. The outputs are shown as ’mean ± standard deviation’. As depicted in Table 6 and Table 7, the proposed method showed better performance than the existing segmentation methods on both datasets, indicating the robustness of the model.
Even though the MV-CNN [51], FCN-UNET [5], MCROI-CNN [23], MC-CNN [16], Cascaded-CNN [22], and CF-CNN [27] methods achieved good results, DDRN [30], DB-ResNet [28], and U-Det showed better performance compared to them. Although the DB-ResNet achieved good performance in various cases, its performance was hindered in the cases where the size of the nodule was less than 6 mm [28]. Similarly, for DDRN, the performance of the model was hindered when it was weakly supervised with the ground truth masks and also in the cases where the nodule size was small [30]. In contrast, the proposed method outperformed the ground truth segmentation masks and showed excellent generalization capability, as depicted in Figure 6. Further, Figure 9 illustrates the U-Det model’s performance in challenging cases such as small nodules (<6 mm), cavitary nodules, and juxta-vascular and juxtapleural nodules from the LUNA16 dataset. This observation confirmed that the proposed model showed efficient performance on various nodules, including nodules of less than 6 mm.

7. Conclusions

Lung CT image segmentation plays a vital role in the analysis of lung images, which helps in the identification of lung cancer. For lung nodule image segmentation, this paper proposed a deep-learning-based encoder–decoder model (U-Det) using Bi-FPN as a feature enricher by incorporating multi-scale feature fusion. The proposed method demonstrated encouraging precision in the segmentation of the lung nodules and obtained 82.82% and 81.66% DSC scores for the LUNA16 and QIN datasets, respectively. A comparison with similar approaches also showed the efficient performance of the proposed approach. It also achieved efficient segmentation results in daunting cases such as cavitary nodules, GGO nodules, juxtapleural nodules, and small nodules of less than 6 mm.
There are several areas of interest for further research. We employ a Bi-FPN to achieve a multi-scale fusion of features from 2D CT slices in the current work. However, it would be interesting to leverage the spatial information obtained from features of a voxel-based 3D CT scan representation. When fused across a multi-scale, these spatial features could potentially help models distinguish 3D properties and eccentricity of lung nodules for both segmentation and classification. Hence, future work will focus on developing a 3D capsule network based on the components of U-Det for fully automated malignancy classification and voxel-based segmentation of lung cancer.

Author Contributions

Conceptualization, C.S.R.A. and S.A.B.P.; supervision, C.S.R.A. and S.A.B.P.; methodology, N.V.K.; resources, P.K.D.; validation, N.V.K.; writing—review and editing, S.A.B.P., P.K.D. and G.R.; writing—original draft preparation, N.V.K.; funding acquisition, P.K.D.; visualization, G.R.; investigation, N.V.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

More details can be found at https://github.com/Nik-V9/U-Det.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. WHO Report on Cancer: Setting Priorities, Investing Wisely and Providing Care for All; WHO: Geneva, Switzerland, 2020. [Google Scholar]
  2. MacMahon, H.; Austin, J.H.; Gamsu, G.; Herold, C.J.; Jett, J.R.; Naidich, D.P.; Patz Jr, E.F.; Swensen, S.J. Guidelines for management of small pulmonary nodules detected on CT scans: A statement from the Fleischner Society. Radiology 2005, 237, 395–400. [Google Scholar] [CrossRef] [PubMed]
  3. Kubota, T.; Jerebko, A.K.; Dewan, M.; Salganicoff, M.; Krishnan, A. Segmentation of pulmonary nodules of various densities with morphological approaches and convexity models. Med. Image Anal. 2011, 15, 133–154. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, S.; Chen, X.; Zhu, Z.; Feng, B.; Chen, Y.; Long, W. Segmentation of small ground glass opacity pulmonary nodules based on Markov random field energy and Bayesian probability difference. Biomed. Eng. Online 2020, 19, 1–20. [Google Scholar] [CrossRef]
  5. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berling, Germany, 2015; pp. 234–241. [Google Scholar]
  6. Kuhnigk, J.M.; Dicken, V.; Bornemann, L.; Bakai, A.; Wormanns, D.; Krass, S.; Peitgen, H.O. Morphological segmentation and partial volume analysis for volumetry of solid pulmonary lesions in thoracic CT scans. IEEE Trans. Med. Imaging 2006, 25, 417–434. [Google Scholar] [CrossRef]
  7. Diciotti, S.; Lombardo, S.; Falchini, M.; Picozzi, G.; Mascalchi, M. Automated segmentation refinement of small lung nodules in CT scans by local shape analysis. IEEE Trans. Biomed. Eng. 2011, 58, 3418–3428. [Google Scholar] [CrossRef] [PubMed]
  8. Rebouças Filho, P.P.; da Silva Barros, A.C.; Almeida, J.S.; Rodrigues, J.; de Albuquerque, V.H.C. A new effective and powerful medical image segmentation algorithm based on optimum path snakes. Appl. Soft Comput. 2019, 76, 649–670. [Google Scholar] [CrossRef]
  9. Dharmalingham, V.; Kumar, D. A model based segmentation approach for lung segmentation from chest computer tomography images. Multimed. Tools Appl. 2019, 79, 1–26. [Google Scholar] [CrossRef]
  10. Lu, L.; Devarakota, P.; Vikal, S.; Wu, D.; Zheng, Y.; Wolf, M. Computer aided diagnosis using multilevel image features on large-scale evaluation. In Proceedings of the International MICCAI Workshop on Medical Computer Vision, Nagoya, Japan, 26 September 2013; Springer: Berling, Germany, 2013; pp. 161–174. [Google Scholar]
  11. Santos, A.M.; de Carvalho Filho, A.O.; Silva, A.C.; de Paiva, A.C.; Nunes, R.A.; Gattass, M. Automatic detection of small lung nodules in 3D CT data using Gaussian mixture models, Tsallis entropy and SVM. Eng. Appl. Artif. Intell. 2014, 36, 27–39. [Google Scholar] [CrossRef]
  12. Prabukumar, M.; Agilandeeswari, L.; Ganesan, K. An intelligent lung cancer diagnosis system using cuckoo search optimization and support vector machine classifier. J. Ambient. Intell. Humaniz. Comput. 2019, 10, 267–293. [Google Scholar] [CrossRef]
  13. Gonçalves, L.; Novo, J.; Campilho, A. Hessian based approaches for 3D lung nodule segmentation. Expert Syst. Appl. 2016, 61, 1–15. [Google Scholar] [CrossRef]
  14. Jung, J.; Hong, H.; Goo, J.M. Ground-glass nodule segmentation in chest CT images using asymmetric multi-phase deformable model and pulmonary vessel removal. Comput. Biol. Med. 2018, 92, 128–138. [Google Scholar] [CrossRef]
  15. Devi, K.Y.; Sasikala, M. Labeling and clustering-based level set method for automated segmentation of lung tumor stages in CT images. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 2299–2309. [Google Scholar] [CrossRef]
  16. Shen, W.; Zhou, M.; Yang, F.; Yu, D.; Dong, D.; Yang, C.; Zang, Y.; Tian, J. Multi-crop convolutional neural networks for lung nodule malignancy suspiciousness classification. Pattern Recognit. 2017, 61, 663–673. [Google Scholar] [CrossRef]
  17. Feng, Y.; Hao, P.; Zhang, P.; Liu, X.; Wu, F.; Wang, H. Supervoxel based weakly-supervised multi-level 3D CNNs for lung nodule detection and segmentation. J. Ambient. Intell. Humaniz. Comput. 2019. [Google Scholar] [CrossRef]
  18. Wang, S.; Zhou, M.; Gevaert, O.; Tang, Z.; Dong, D.; Liu, Z.; Tian, J. A multi-view deep convolutional neural networks for lung nodule segmentation. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju Island, Republic of Korea, 11–15 July 2017; pp. 1752–1755. [Google Scholar]
  19. Huang, X.; Sun, W.; Tseng, T.L.B.; Li, C.; Qian, W. Fast and fully-automated detection and segmentation of pulmonary nodules in thoracic CT scans using deep convolutional neural networks. Comput. Med. Imaging Graph. 2019, 74, 25–36. [Google Scholar] [CrossRef] [PubMed]
  20. Kumar, S.; Raman, S. Lung nodule segmentation using 3-dimensional convolutional neural networks. In Proceedings of the Soft Computing for Problem Solving: SocProS 2018, Vellore, India, 17–19 December 2018; Springer: Berling, Germany, 2020; Volume 1, pp. 585–596. [Google Scholar]
  21. Li, R.; Xiao, C.; Huang, Y.; Hassan, H.; Huang, B. Deep learning applications in computed tomography images for pulmonary nodule detection and diagnosis: A review. Diagnostics 2022, 12, 298. [Google Scholar] [CrossRef]
  22. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef]
  23. Sun, W.; Zheng, B.; Qian, W. Automatic feature learning using multichannel ROI based on deep structured algorithms for computerized lung cancer diagnosis. Comput. Biol. Med. 2017, 89, 530–539. [Google Scholar] [CrossRef] [PubMed]
  24. Zhao, X.; Sun, W.; Qian, W.; Qi, S.; Sun, J.; Zhang, B.; Yang, Z. Fine-grained lung nodule segmentation with pyramid deconvolutional neural network. In Proceedings of the Medical Imaging 2019: Computer-Aided Diagnosis. International Society for Optics and Photonics, San Diego, CA, USA, 16–21 February 2019; Volume 10950, p. 109503S. [Google Scholar]
  25. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  26. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berling, Germany, 2016; pp. 424–432. [Google Scholar]
  27. Wang, S.; Zhou, M.; Liu, Z.; Liu, Z.; Gu, D.; Zang, Y.; Dong, D.; Gevaert, O.; Tian, J. Central focused convolutional neural networks: Developing a data-driven model for lung nodule segmentation. Med. Image Anal. 2017, 40, 172–183. [Google Scholar] [CrossRef] [PubMed]
  28. Cao, H.; Liu, H.; Song, E.; Hung, C.C.; Ma, G.; Xu, X.; Jin, R.; Lu, J. Dual-branch residual network for lung nodule segmentation. Appl. Soft Comput. 2020, 86, 105934. [Google Scholar] [CrossRef]
  29. Rocha, J.; Cunha, A.; Mendonça, A.M. Conventional Filtering Versus U-Net Based Models for Pulmonary Nodule Segmentation in CT Images. J. Med. Syst. 2020, 44, 1–8. [Google Scholar] [CrossRef]
  30. Singadkar, G.; Mahajan, A.; Thakur, M.; Talbar, S. Deep Deconvolutional Residual Network Based Automatic Lung Nodule Segmentation. J. Digit. Imaging 2020, 33, 678–684. [Google Scholar] [CrossRef]
  31. Chen, W.F.; Ou, H.Y.; Lin, H.Y.; Wei, C.P.; Liao, C.C.; Cheng, Y.F.; Pan, C.T. Development of Novel Residual-Dense-Attention (RDA) U-Net Network Architecture for Hepatocellular Carcinoma Segmentation. Diagnostics 2022, 12, 1916. [Google Scholar] [CrossRef] [PubMed]
  32. Mikołajczyk, A.; Grochowski, M. Data augmentation for improving deep learning in image classification problem. In Proceedings of the 2018 International Interdisciplinary PhD Workshop (IIPhDW), Swinoujscie, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar]
  33. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  34. LaLonde, R.; Bagci, U. Capsules for object segmentation. arXiv 2018, arXiv:1804.04241. [Google Scholar]
  35. Misra, D. Mish: A Self Regularized Non-Monotonic Neural Activation Function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  36. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. arXiv 2019, arXiv:1911.09070. [Google Scholar]
  37. Bengio, Y.; Grandvalet, Y. No unbiased estimator of the variance of k-fold cross-validation. J. Mach. Learn. Res. 2004, 5, 1089–1105. [Google Scholar]
  38. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Caruana, R.; Lawrence, S.; Giles, C.L. Overfitting in neural nets: Backpropagation, conjugate gradient, and early stopping. In Proceedings of the Advances in Neural Information Processing Systems, Denver CO, USA, 1 January 2000; pp. 402–408. [Google Scholar]
  40. Jacobs, C.; van Rikxoort, E.M.; Twellmann, T.; Scholten, E.T.; de Jong, P.A.; Kuhnigk, J.M.; Oudkerk, M.; de Koning, H.J.; Prokop, M.; Schaefer-Prokop, C.; et al. Automatic detection of subsolid pulmonary nodules in thoracic computed tomography images. Med. Image Anal. 2014, 18, 374–384. [Google Scholar] [CrossRef]
  41. Setio, A.A.; Jacobs, C.; Gelderblom, J.; van Ginneken, B. Automatic detection of large pulmonary solid nodules in thoracic CT images. Med. Phys. 2015, 42, 5642–5653. [Google Scholar] [CrossRef] [PubMed]
  42. Van Rikxoort, E.M.; de Hoop, B.; Viergever, M.A.; Prokop, M.; van Ginneken, B. Automatic lung segmentation from thoracic computed tomography scans using a hybrid approach with error detection. Med. Phys. 2009, 36, 2934–2947. [Google Scholar] [CrossRef] [PubMed]
  43. Armato III, S.G.; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef]
  44. Setio, A.A.A.; Traverso, A.; De Bel, T.; Berens, M.S.; van den Bogaard, C.; Cerello, P.; Chen, H.; Dou, Q.; Fantacci, M.E.; Geurts, B.; et al. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Med. Image Anal. 2017, 42, 1–13. [Google Scholar] [CrossRef] [PubMed]
  45. Kalpathy-Cramer, J.; Freymann, J.B.; Kirby, J.S.; Kinahan, P.E.; Prior, F.W. Quantitative imaging network: Data sharing and competitive AlgorithmValidation leveraging the cancer imaging archive. Transl. Oncol. 2014, 7, 147–152. [Google Scholar] [CrossRef] [PubMed]
  46. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  47. Jayashree, K.; Sandy, N. Multi-site collection of lung ct data with nodule segmentations. J. Digit. Imaging 2015, 1–9. [Google Scholar]
  48. Kalpathy-Cramer, J.; Zhao, B.; Goldgof, D.; Gu, Y.; Wang, X.; Yang, H.; Tan, Y.; Gillies, R.; Napel, S. A comparison of lung nodule segmentation algorithms: Methods and results from a multi-institutional study. J. Digit. Imaging 2016, 29, 476–487. [Google Scholar] [CrossRef]
  49. Chon, A.; Balachandar, N.; Lu, P. Deep Convolutional Neural Networks for Lung Cancer Detection; Standford University: Stanford, CA, USA, 2017. [Google Scholar]
  50. Valverde, S.; Oliver, A.; Roura, E.; González-Villà, S.; Pareto, D.; Vilanova, J.C.; Ramió-Torrentà, L.; Rovira, À.; Lladó, X. Automated tissue segmentation of MR brain images in the presence of white matter lesions. Med. Image Anal. 2017, 35, 446–457. [Google Scholar] [CrossRef]
  51. Kang, G.; Liu, K.; Hou, B.; Zhang, N. 3D multi-view convolutional neural networks for lung nodule classification. PLoS ONE 2017, 12, e0188290. [Google Scholar] [CrossRef]
Figure 1. Illustrations of various types of lung nodules present in CT scans. Note: Small nodule indicates a nodule of diameter <6 mm. (a) isolated nodule; (b) juxtapleural nodule; (c) juxtavascular nodule; (d) cavitary nodule; (e) calcific nodule; (f) GGO nodule; (g) small nodule.
Figure 1. Illustrations of various types of lung nodules present in CT scans. Note: Small nodule indicates a nodule of diameter <6 mm. (a) isolated nodule; (b) juxtapleural nodule; (c) juxtavascular nodule; (d) cavitary nodule; (e) calcific nodule; (f) GGO nodule; (g) small nodule.
Diagnostics 13 01406 g001
Figure 2. Overview of the proposed model pipeline.
Figure 2. Overview of the proposed model pipeline.
Diagnostics 13 01406 g002
Figure 3. Illustration of the proposed U-Det model, where the convolutional neural network block between the down-sampling and up-sampling sections represents the Bi-FPN. The numbers at each layer of the architecture indicate the shape of feature maps at each layer, respectively. The key indicates the representation of various operations that take place in the backbone architecture.
Figure 3. Illustration of the proposed U-Det model, where the convolutional neural network block between the down-sampling and up-sampling sections represents the Bi-FPN. The numbers at each layer of the architecture indicate the shape of feature maps at each layer, respectively. The key indicates the representation of various operations that take place in the backbone architecture.
Diagnostics 13 01406 g003
Figure 4. Histogram of lung nodule sizes across the LUNA16 dataset.
Figure 4. Histogram of lung nodule sizes across the LUNA16 dataset.
Diagnostics 13 01406 g004
Figure 5. LUNA16 test set DSC distributions.
Figure 5. LUNA16 test set DSC distributions.
Diagnostics 13 01406 g005
Figure 6. Qualitative depiction representing the ground truth masks and segmentation results of the U-Det model. Here, figures (ac) are from the LUNA16 dataset, and (df) are from the QIN Lung CT Segmentation dataset. Note: The red filter represents the ground truth mask, and the blue filter represents the segmentation results.
Figure 6. Qualitative depiction representing the ground truth masks and segmentation results of the U-Det model. Here, figures (ac) are from the LUNA16 dataset, and (df) are from the QIN Lung CT Segmentation dataset. Note: The red filter represents the ground truth mask, and the blue filter represents the segmentation results.
Diagnostics 13 01406 g006
Figure 7. Quantitative segmentation results of LUNA-16 test set.
Figure 7. Quantitative segmentation results of LUNA-16 test set.
Diagnostics 13 01406 g007
Figure 8. Quantitative segmentation results of QIN Lung CT test set.
Figure 8. Quantitative segmentation results of QIN Lung CT test set.
Diagnostics 13 01406 g008
Figure 9. Visualization of segmentation results of the proposed U-Det model on heterogeneous types of lung nodules. The various types of lung nodules presented from left to right are isolated, juxtapleural, GGO and juxta-vascular, cavitary, very small size, and calcified.
Figure 9. Visualization of segmentation results of the proposed U-Det model on heterogeneous types of lung nodules. The various types of lung nodules presented from left to right are isolated, juxtapleural, GGO and juxta-vascular, cavitary, very small size, and calcified.
Diagnostics 13 01406 g009
Table 1. The layers and respective network parameters of the proposed model.
Table 1. The layers and respective network parameters of the proposed model.
Layer NameNumber of Parameters
Contraction path:
Conv2D × 10, Mish1.884 ×   10 7
MaxPool2D × 4-
Bi-FPN:
Conv2D × 51.269 ×   10 5
BatchNormalization × 123072
ReLU × 12, MaxPool2D × 3-
DepthwiseConv × 74032
Expansion path:
Conv2D × 9, Mish6.821 ×   10 6
Conv2DTrans × 4, Mish2.786 ×   10 6
Total parameters:2.858 ×   10 7
Table 2. Distribution of LUNA16 train and test sets. The values are indicated as ’mean ± standard deviation’.
Table 2. Distribution of LUNA16 train and test sets. The values are indicated as ’mean ± standard deviation’.
CharacteristicsTrain Set
(n = 922)
Test Set
(n = 244)
Diameter (mm)8.13 ± 4.609.07 ± 5.24
Margin4.03 ± 0.824.06 ± 0.76
Spiculation1.60 ± 0.791.65 ± 0.87
Lobulation1.73 ± 0.731.82 ± 0.80
Subtlety3.91 ± 0.824.06 ± 0.78
Malignancy2.95 ± 0.923.03 ± 1.00
Note: The range for all distinctive feature values except diameter was observed between 1 and 5. The ‘Margin’ characteristic shows nodule edge clarity. Both ‘Spiculation’ and ‘Lobulation’ indicate the shape characteristics of the nodule. ‘Subtlety’ explains the contrast between the nodule zone and its surrounding areas. ‘Malignancy’ reflects the possibility of this characteristic in a nodule.
Table 3. Ablation experiment on the LUNA16 test dataset. The experiment was based on the U-Net model.
Table 3. Ablation experiment on the LUNA16 test dataset. The experiment was based on the U-Net model.
MethodDSC (%)SEN (%)PPV (%)
U-Net77.84 ± 21.7477.98 ± 24.5282.52 ± 21.53
U-Net + Mish78.82 ± 22.0178.97 ± 24.8383.56 ± 21.80
Encoder + Bi-FPN79.21 ± 12.4984.40 ± 13.5176.30 ± 14.42
Encoder + Bi-FPN + Mish80.22 ± 12.3385.47 ± 13.4878.58 ± 14.34
Encoder–Decoder + Bi-FPN + ReLU81.63 ± 11.8591.06 ± 13.9677.94 ± 13.68
Proposed Method82.82 ± 11.7192.24 ± 14.1478.92 ± 17.52
Table 4. Segmentation results of the proposed model on the LUNA16 dataset for various cases such as attached and non-attached nodules, and nodules of large and small sizes.
Table 4. Segmentation results of the proposed model on the LUNA16 dataset for various cases such as attached and non-attached nodules, and nodules of large and small sizes.
(a) LUNA16 Test set
Attached
(n = 56)
Non-Attached
(n = 188)
Diameter < 6 mm
(n = 104)
Diameter ≥ 6 mm
(n = 140)
DSC (%)81.8283.1183.4082.40
Table 5. Segmentation results of the proposed model on the QIN Lung CT dataset for various cases such as attached and non-attached nodules, and nodules of large and small sizes.
Table 5. Segmentation results of the proposed model on the QIN Lung CT dataset for various cases such as attached and non-attached nodules, and nodules of large and small sizes.
(b) QIN Lung CT Segmentation dataset
Attached
(n = 34)
Non-Attached
(n = 122)
Diameter < 6 mm
(n = 54)
Diameter ≥ 6 mm
(n = 102)
DSC (%)80.0283.3083.1080.22
Table 6. Quantitative segmentation results of the proposed model compared to different types of model architectures on LUNA16 dataset.
Table 6. Quantitative segmentation results of the proposed model compared to different types of model architectures on LUNA16 dataset.
LUNA16 Test Set
Network ArchitectureDSC (%)SEN (%)PPV (%)
MV-CNN [51]75.89 ± 12.9987.16 ± 12.9170.81 ± 17.57
MCROI-CNN [23]77.01 ± 12.9385.43 ± 15.9773.52 ± 14.62
MC-CNN [16]77.51 ± 11.488.83 ± 12.3471.42 ± 14.78
FCN-UNET [5]77.84 ± 21.7477.98 ± 24.5282.52 ± 21.53
MV-DCNN [18]77.85 ± 12.9486.96 ± 15.7377.33 ± 13.26
CF-CNN [27]78.55 ± 12.4986.01 ± 15.2275.79 ± 14.73
Cascaded-CNN [22]79.83 ± 10.9186.86 ± 13.3576.14 ± 13.46
DDRN [30]81.56 ± 11.5987.35 ± 12.3977.42 ± 14.65
DB-ResNet [28]82.74 ± 10.1989.35 ± 11.7979.64 ± 13.54
Proposed Method82.82 ± 11.7192.24 ± 14.1478.92 ± 17.52
Table 7. Quantitative segmentation results of the proposed model compared to different types of model architectures on QIN Lung CT dataset.
Table 7. Quantitative segmentation results of the proposed model compared to different types of model architectures on QIN Lung CT dataset.
QIN Lung CT Segmentation Dataset
Network ArchitectureDSC (%)SEN (%)PPV (%)
FCN-UNET75.26 ± 11.8276.65 ± 16.4277.21 ± 11.57
CF-CNN77.23 ± 11.5380.12 ± 17.0776.65 ± 12.20
MV-DCNN77.89 ± 10.6481.29 ± 15.6076.95 ± 11.62
Cascaded-CNN78.89 ± 11.8987.20 ± 12.4476.74 ± 13.52
DB-ResNet80.01 ± 11.4688.13 ± 12.3479.13 ± 14.12
DDRN80.56 ± 11.0883.57 ± 11.7878.65 ± 13.86
Proposed Method81.66 ± 10.0991.11 ± 12.0180.05 ± 13.34
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Annavarapu, C.S.R.; Parisapogu, S.A.B.; Keetha, N.V.; Donta, P.K.; Rajita, G. A Bi-FPN-Based Encoder–Decoder Model for Lung Nodule Image Segmentation. Diagnostics 2023, 13, 1406. https://doi.org/10.3390/diagnostics13081406

AMA Style

Annavarapu CSR, Parisapogu SAB, Keetha NV, Donta PK, Rajita G. A Bi-FPN-Based Encoder–Decoder Model for Lung Nodule Image Segmentation. Diagnostics. 2023; 13(8):1406. https://doi.org/10.3390/diagnostics13081406

Chicago/Turabian Style

Annavarapu, Chandra Sekhara Rao, Samson Anosh Babu Parisapogu, Nikhil Varma Keetha, Praveen Kumar Donta, and Gurindapalli Rajita. 2023. "A Bi-FPN-Based Encoder–Decoder Model for Lung Nodule Image Segmentation" Diagnostics 13, no. 8: 1406. https://doi.org/10.3390/diagnostics13081406

APA Style

Annavarapu, C. S. R., Parisapogu, S. A. B., Keetha, N. V., Donta, P. K., & Rajita, G. (2023). A Bi-FPN-Based Encoder–Decoder Model for Lung Nodule Image Segmentation. Diagnostics, 13(8), 1406. https://doi.org/10.3390/diagnostics13081406

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop