Next Article in Journal
Multi-Asymmetric Irradiation Method Using a Ring Array to Obtain an Emission-Capable LED Beam Power Effect to Observe Cancer Removal Status in a Surgical Microscope
Previous Article in Journal
Editorial on the Special Issue Titled “Pathology and Diagnosis of Gynecologic Diseases”
Previous Article in Special Issue
Multilevel Threshold Segmentation of Skin Lesions in Color Images Using Coronavirus Optimization Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyper-Dense_Lung_Seg: Multimodal-Fusion-Based Modified U-Net for Lung Tumour Segmentation Using Multimodality of CT-PET Scans

1
School of Computing and Commutations, Lancaster University, Lancaster LA1 4YW, UK
2
College of Computing and Information Technology, University of Bisha, Bisha 67714, Saudi Arabia
3
Computers and Control Engineering Department, Faculty of Engineering, Tanta University, Tanta 31733, Egypt
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(22), 3481; https://doi.org/10.3390/diagnostics13223481
Submission received: 6 October 2023 / Revised: 9 November 2023 / Accepted: 15 November 2023 / Published: 20 November 2023

Abstract

:
The majority of cancer-related deaths globally are due to lung cancer, which also has the second-highest mortality rate. The segmentation of lung tumours, treatment evaluation, and tumour stage classification have become significantly more accessible with the advent of PET/CT scans. With the advent of PET/CT scans, it is possible to obtain both functioning and anatomic data during a single examination. However, integrating images from different modalities can indeed be time-consuming for medical professionals and remains a challenging task. This challenge arises from several factors, including differences in image acquisition techniques, image resolutions, and the inherent variations in the spectral and temporal data captured by different imaging modalities. Artificial Intelligence (AI) methodologies have shown potential in the automation of image integration and segmentation. To address these challenges, multimodal fusion approach-based U-Net architecture (early fusion, late fusion, dense fusion, hyper-dense fusion, and hyper-dense VGG16 U-Net) are proposed for lung tumour segmentation. Dice scores of 73% show that hyper-dense VGG16 U-Net is superior to the other four proposed models. The proposed method can potentially aid medical professionals in detecting lung cancer at an early stage.

1. Introduction

In 2020, lung cancer ranked as the second most prevalent cancer, accounting for around 11.4% of all newly identified instances of cancer, with approximately 2.2 million individuals affected. Furthermore, it was the primary cause of cancer-related mortality, responsible for approximately 18.0% of the deaths caused by cancer globally, resulting in approximately 1.8 million fatalities. Loss of appetite, exhaustion, chronic coughing, and chest pain are among the symptoms of lung cancer, which can cause unimaginable anguish for the sufferer [1].
The segmentation of lung tumours, treatment evaluation, and tumour stage classification have become significantly more accessible with the advent of PET/CT scans. Moreover, the molecular characteristics and anatomic aberrations of the target lesion can be observed with PET/CT. The PET imaging technique does not involve cutting or surgery. By detecting illness markers earlier, PET allows for earlier diagnosis than imaging modalities like MRI and CT [2]. Their metabolic processes can be analysed for their physiological function and biochemical features by studying particular organs and tissues. PET can detect molecular and cellular levels of tissue metabolism.
However, multimodality imaging technology, such as PET-CT scanners, has simultaneously made it possible to record functional and anatomical information [3]. It is a rigorous and time-consuming process for oncologists, radiologists, and pulmonologists to manually segment the lesions and tumours, leading to delays in therapy and decreased survival rates, particularly in clinics with insufficient resources. In addition, specialist knowledge and clinical experience are necessary for high-quality manual localization and segmentation. Because of this, computer-aided diagnostic (CAD) systems [4] were developed to replace radiologists’ manual viewing of lung cancer. Combining lung segmentation approaches with radiologists’ knowledge can reduce the burden on radiologists and boost their productivity and accuracy. Many recent advancements in image segmentation have allowed for more precise and effective treatment and diagnosis. Thresholding, Atlas, and Region Growing are some examples of classic automatic segmentation methods. These approaches use the shallow qualities of an image, such as grayscale, texture, gradient, and many more, to segment the object [5]. However, conventional segmentation techniques have difficulty distinguishing between tumours and surrounding healthy tissue because their intensity distributions are similar. In addition, these tasks typically involve manual processes and are characterized by a significant investment of time. Moreover, they are subject to substantial heterogeneity across operators.
Furthermore, the complexity of the background in CT images consistently provides quite different information when comparing PET and CT scans. As a result of these constraints, deep-learning-based algorithms have proven to be superior in auto-segmenting medical images [6].
Deep learning (DL) models automatically extract features and apply the learned high-dimensional abstractions for performing segmentation. The effectiveness of fully convolutional networks (FCN) for semantic segmentation is promising [7]. In an FCN, the fully connected layer is replaced by a convolutional layer. This comprehensive framework has served as the foundation for subsequent studies of semantic segmentation of medical images. Medical image segmentation commonly uses U-Net [8], built on the FCN architecture. Using skip-connection architecture, each layer’s down-sampled features are joined with their up-sampled counterparts. This mechanism is similar to an encoder-decoder, but it is more effective and does not require a great deal of disk space. FCN-based networks, such as U-Net, have surpassed manual or semi-automatic segmentation methods since the emergence of big data methods.
The U-Net architecture is a convolutional neural network (CNN) primarily used to recognize image patterns. U-Net semantic segmentation relies extensively on the categorization of image pixels. Segmenting lung tumours can be reduced to a foreground/background pixel binary classification problem. The down-sampling and up-sampling module is responsible for the U-Net architecture. The surface layer is where localization information is learned, but the down-sampling procedure, also known as the pooling procedure, may improve the volume of context data the network learns [8].
The VGG model investigates the effect of convolutional network depth on recognition accuracy in a large-scale setting. Our main contribution is a thorough evaluation of increasing-depth networks using an architecture with 3 × 3 convolution filters, which shows a significant improvement. The VGG model, a kind of CNN, was created to improve model performance when more layers are added. The VGG model takes 224 × 224 colour images as its primary input and feeds them via a sequence of convolutional layers, with filter sizes of 3 × 3 and 1 × 1 with the stride of 1 and valid padding, as well as max pooling with 2 × 2 with the stride of 2. Finally, we have a three-layer network with a SoftMax activation function and 4096 neurons in the first two layers, followed by 1000 neurons in the last layer. VGG [9] presents the two primary models, VGG16 and VGG19.
In comparison to the VGG-19 network, which has 19 layers of typical convolutional networks, the VGG-16 network [9] only has 16, with filter sizes of 3 × 3 and strides of 1. Each of the five blocks is separated from the next by a max-pooling layer. There are three interconnected layers on top of the blocks.
The significant contributions of this research are given below.
  • The inputs to the proposed architecture are PET and CT scans. Here, dense connections happen along the same pathways that process each modality individually. Finally, their features are joined together at a high layer to finish separating them.
  • Five deep models based on U-Net architecture are suggested for lung cancer segmentation in multimodal image scenarios: early fusion, late fusion, dense fusion, hyper-dense fusion, and VGG-16 U-Net.
  • The performance of the suggested models was evaluated using three types of loss functions: binary, Dice, and focal loss functions.
The remainder of the paper is structured as follows. Section 2 outlines studies on multimodal PET-CT images for segmentation based on deep learning. The proposed network architecture is described in Section 3. In Section 4, the results of the experiments and their analysis are presented. Section 5 presents the study’s findings and recommendations for moving forward.

2. Literature Review

CT and PET imaging are used in various research papers because of the unique insights they provide into the structure and function of the human body, respectively. Combining the two allows for the early detection of even the tiniest lung tumours.
Wang et al. [10] advised a DL-based dual-modality approach using CT and PET scans to develop an automated segmenting of lung tumours for radiation therapy planning. Two distinct convolution routes were built into the 3D convolutional neural network for extracting features at different resolutions from the PETs and simulated CTs, and a single deconvolution path was also built into the network. Tumour segmentation via skip connections at each granularity was achieved by aggregating the obtained characteristics from the convolution arms and feeding them into the deconvolution pathway. A panel of oncologists judged the medical effectiveness of the network-generated segmentation strategy. While this work has many promising applications, it does have some caveats. The network may struggle to produce precise segmentations when tumour edges are not precise on CT or PET.
Park et al. [11] presented a two-stage U-Net model to boost the segmentation effectiveness of lung tumours by utilizing [18F]FDG PET/CT, as precise segmentation is necessary for determining the functional size of a tumour in this imaging modality. The LifeX program was used to create the tumour volume of interest. In the first step, a 3D PET/CT volume is used to train a global U-Net, based on which a 3D binary volume is then retrieved to serve as an initial representation of the tumour’s region. In the second stage, the PET/CT slice identified in Stage 1 is sent to the U-Net, generating a 2D binary image centred on the eight adjacent slices. The major drawback of the research is the lack of a 3D volume as the final result of the suggested approach. It may cause the coronal and sagittal slices to have gaps between the binary segments.
Xiang et al. [12] recommended a modality-specific segmentation network (MoSNet) to segment lung tumours. To better understand the differences between PET and CT scans, MoSNet is taught to use modality-specific representations, while modality-fused representations are employed to convert the typical characteristics of lung tumours in both scan types. The authors suggest an adversarial approach that uses an adversarial purpose concerning a modality discriminator and a reserved modality-common illustration to reduce the modality difference’s approximation. As a result, the network’s ability to represent data for the segmentation in PET and CT scans is enhanced. By generating a map for each modality, MoSNet can explicitly quantify the weights for the attributes in each modality. However, the limitation of the research is that the proposed approach is developed for 2D thorax PET-CT slices.
Fu et al. [13] proposed a DL system for lung cancer segmentation, i.e., a multimodal spatial attention module (MSAM). It is trained to highlight tumour-related regions selectively and downplay those physiologically rising from the PET scans. Next, using the created spatial attention maps, a CNN core is trained to focus on areas of a CT image with a higher propensity for tumours. The drawback of the research is that the datasets used only had one observer defining the outlines. If numerous observers had been used to reach a consensus segmentation, things would have gone much smoother. Because of the potential vagueness of the related thresholding approach used to create the ground truth for the NSCLC dataset, the segmentation outputs require human adjustment to correct for incorrectly categorized ROIs.
Zhong et al. [14] provided an innovative method for lung tumour segmentation by bringing together a robust FCN-based 3D U-Net and a graph-cut-based co-segmentation model. Initially, high-level discriminative features for PET and CT images are learned by independently training two distinct deep U-Nets on the datasets. These features then create tumour/non-tumour masks and probability maps. The final tumour segmentation findings are obtained using the PET and CT probability mappings in a graph-cut-based co-segmentation model. Despite fusing their extracted features, the research has given different results for CT and PET.
Hwang et al. [15] recommended a new network architecture called 3C-Net, which uses numerous contexts in three distinct ways. Two decoders in the network are implemented to exploit inter-slice contextual information: a segmentation decoder and a context decoder. The context decoder receives the inter-slice difference features and uses them to predict the segmentation mask’s inter-slice difference. Having this 3D background information for each slice helps in attention direction. The prediction results from each decoder stage are used to derive a loss function for network optimization. Since two modalities are used, i.e., PET/CT data, a co-encoder block is implemented to extract mutually reinforcing features from both modalities while simultaneously acquiring contextual knowledge about them. Weights for both CT and PET are modified twice in co-encoder blocks. The co-encoder blocks take in relevant data from both modalities, allowing for interaction while maintaining spatial and structural coherence. The encoder additionally includes an asterisk spatial pyramid pooling (ASPP) block in its final step. The ASPP block aids the network in increasing the scope of its observations and avoiding the loss of spatial context, which allows the recording of visual context at various scales.
Kumar et al. [16] improved the multimodality PET-CT fusion using CNN, which learns to fuse complementary information. The proposed CNN stores modality-specific characteristics before deriving a spatially variable fusion map. It allows quantifying the relevance of each modality’s characteristics across various spots. Moreover, multiplying the fusion maps with the modality-specific feature maps yields representations of the complementary multimodality data at various positions. The recommended CNN is tested on PET-CT scans of lung tumours, where its ability to detect and separate many regions with variable fusion needs is evaluated.
Jemaa et al. [17] demonstrated a comprehensive strategy employing 2D and 3D CNN for rapid tumour classification and metabolic data retrieval from whole-body FDG-PET/CT images. This architecture is relatively economical between tumour load and healthy tissue volume, and between the intrinsic heterogeneity of the input images, which is especially important for whole-body scans due to their vast size and high asymmetry.
Zhao et al. [18] developed a novel multimodality segmentation approach that utilizes a 3D FCN, and simultaneously includes PET and CT data in tumour segmentation. Initially, the network underwent a multitask training phase, during which two parallel sub-segmentation architectures, each built with a deep CNN, were learned to generate map-like features from both modalities. The PET/CT feature maps’ characteristics were re-extracted using a weighted cross-entropy reduction technique, and a feature fusion component was then constructed using cascaded convolutional modules. The SoftMax function was also used to generate the cancer mask as the network’s final output. The research lacks an automatic setting of the weighting parameters of the loss functions, which can affect performance. Also, more effective ways for feature extraction can increase the performance of the segmentation.
Using W-net, Zhong et al. [19] evaluated 3D deep fully convolutional networks (DFCN) for tumour co-segmentation on dual-modality NSCLC and PET-CT images. CT and PET data are combined to better understand NSCLC tumours in PET-CT scans and apply DFCN co-segmentation. The recommended DFCN-based co-segmentation approach uses two connected 3D U-Nets with an encoder-decoder to exchange complementing data between PET and CT.
Bi et al. [20] developed a hyper-connected fusion model that uses a CNN-TN fusion encoder and a CNN-TN fusion decoder. With hyper connections between them, the encoder splits into three forks to independently handle PET, CT, and combined PET-CT scans. The transformer encoders process the encoded image embeddings to learn complimentary characteristics in a long-range dependency between the PET, CT, and concatenated PET-CT images. The transformer decoder combines the learnt embeddings to find characteristics important for segmentation, which are subsequently transformed into a 2D feature map. The segmentation results are then up-sampled using a convolutional neural network. The data came from the soft-tissue sarcoma databases. The data showed that the model’s Dice had a probability of 66.36%. The summaries of the literature research in lung tumour segmentation models are listed in Table 1.

3. Methodology

This section discusses the segmentation models for lung cancer in depth. The underlying functions of the proposed models are also described. The U-Net technique is the basis for the proposed models; it has demonstrated usefulness in medical image segmentation. A U-Net model is fed the CT and PET scan data. During the decoder step, further improvement is performed in the model’s recognition of the input image features by creating dense connections across layers in the same input branch of the U-Net.
The proposed architecture for lung cancer segmentation is shown in Figure 1. Figure 1 depicts the three main stages of our framework: (1) image pre-processing, (2) multimodality U-Net segmentation, and (3) medical image post-processing. The pre-processing, augmentation, and post-processing methods are discussed in the following subsections.

3.1. Image Processing

3.1.1. Image Pre-Processing

The intensity levels of the image’s pixels were normalized. In addition to resizing each image, the pixel scale value was changed from (0–255) to (0–1) to reduce the level of complexity of the images. To simplify model training, we decreased the resolution of the CT and PET scans in the dataset to 256 × 256 pixels. The dataset was divided as follows, at random: 46 examples were used for training, and another 5 were used for testing.

3.1.2. Data Augmentation

The CT-PET images were augmented throughout this phase to prevent overfitting, which helps in enhancing the performance of the model. In addition, the implementation of augmentation techniques, such as random rotations, flips, and cuts, can enhance the model’s ability to maintain invariance towards variations in feature position and orientation within the image. This feature proves to be particularly advantageous when working with real-world images that may exhibit variations in object orientation or spatial arrangement. Images are augmented in three ways, as shown in Figure 2: rotating the CT-PET by 90 degrees clockwise (2b), flipping the CT-PET upside down (2c), and left-mirroring the CT-PET (2d).

3.1.3. Image Post-Processing

The suggested framework’s ultimate stage used a morphological change and a basic thresholding technique. To lessen the impact of noise, a morphological gradient accounted for the structure of the input picture. Its effect is analogous to the difference between expanding and contracting an image.
While Equation (1) defines dilation [21] as the process of removing pixels (noises) from object boundaries, Equation (2) describes erosion [22] as the process of adding pixels (negative noises) to object boundaries.
A B = b B A b
A ! B = z E B A
where A is a set of pixels, and B is a structuring element. The thresholding technique is defined as Equation (3).
f x , y = 1             i f   x > t 0           i f   x < t
where t is the threshold value employed to determine if the pixel concerned is part of the desired ground truth or the background. These methods are then used on the expected ground truth to soften the edges and remove false positive values. After being processed, the images of tumour tissue are sharp and in sharp focus.
Figure 3 depicts the last stage in processing predicted masks, in which tiny false positive values and blobs at the borders are removed.

3.1.4. Multimodal Feature Fusion

A feature fusion strategy is deployed in medical imaging to generate a higher quality final image. Feature extraction, classification, and decision making are the three main pillars of any supervised learning-based method. To broaden the types of features collected and better understand their relationships, the early and late sequences of feature fusion are employed in the encoder portion of the core U-Net design. Features from different imaging modalities, like PET and CT, are fused serially to better characterize lung tumours.

3.1.5. Early Fusion

In early fusion, each medical image scan (CT and PET) has a single input path which contains two CNN layers with 64 units and a Relu activation function. Then, these two paths are concatenated into a single path, which is processed through a unique path in the down-sampling U-Net path. This path contains three groups of CNN architecture; each group has three CNN layers with 128, 256, and 512 units, followed by a max pooling layer. All CNNs activation functions are Relu functions. Figure 4 shows the early fusion architecture.

3.1.6. Late Fusion

In contrast to most architectures like U-Net, the encoding path is divided into N streams that serve as input for each imaging modality. Each modality learns a unique feature set using images from the other. The two modalities’ feature maps are combined at each network’s high-level feature layer. This process solves the problem of the early fusion strategy. These feature sets are combined into one feature set and then subjected to the last phase of a multimodal classifier’s training. The U-Net down-sampling path contains four groups of CNN layers. Each group contains three sequential CNN layers with 64, 128, 256, and 512 units, respectively, followed by a max pooling layer. All CNNs have a Relu activation function. At this point, the two paths are concatenated to generate the input of the U-Net up-sampling path. Figure 5 shows the late fusion architecture.

3.1.7. Dense Fusion

For lung cancer segmentation, the dense-fusion-based U-Net provides two down-sampling routes, one for CT and one for PET images. Eight CNN deep learning building blocks are used along each possible route. All layers preceding the current layer are inputs to the current CNN layer. A max-pooling layer follows each pair of consecutive CNN layers. The dimensions of the CNN layer are (in order) 64, 128, 256, and 512. The Relu activation function is standard in all CNNs. The input to the U-Net up-sampling path is generated by concatenating the outputs of the paths following the design described in each path. The dense fusion architecture is shown in Figure 6.

3.1.8. Hyper-Dense Fusion

Deep learning is essential when an application requires a deep layer to function effectively and efficiently. Reducing the overfitting impact is one of several benefits of using dense architecture for multimodality U-Net medical image segmentation. The layers in the same input path provide inputs to all net layers for dense design, which is necessary for U-Nets with multiple input paths. Each layer feeds its immediate successor and those in adjacent input channels in hyper-dense fusion. As the network learns the intricate connections between the modalities at each level of abstraction, the hyper-dense connectivity produces a more robust feature representation than early/late fusion in a multimodal situation. The hyper-dense fusion layout is depicted in Figure 7.

3.2. Loss Functions

In the proposed method, we thoroughly investigated and compared the models using a variety of loss functions. Segmenting an image is essentially a pixel-level classification problem. Each pixel in an image contributes to the overall image, and specific clusters of pixels define particular aspects. Semantic image segmentation is a technique that divides these pixels into their respective components. While designing intricate, deep learning architectures for image segmentation, choosing the loss/objective function is crucial. Loss functions can be broken down into several types based on distribution, region, boundary, and compound. The proposed analysis uses three distinct loss functions, i.e., binary cross-entropy, Dice, and focal. The representation as the network discovers the many interconnections between modalities at every level of abstraction, rather than the binary early/late fusion approach.

3.2.1. Binary Cross-Entropy

Cross-entropy [23] is used to quantify the dissimilarity between two probability distributions for a specific random variable or sequence of events. Segmentation, being a categorization at the pixel level, is extensively employed for this purpose, and it achieves good results. The formula for cross-entropy is given in Equation (4).
L B C E y , y ^ = y log y ^ + 1 y log 1 y ^
where y ^ is the predicted value by the trained model.

3.2.2. Focal Loss

Focal loss (FL) [24] can be understood as a form of binary cross-entropy. It reduces the importance of simple instances, letting the model devote more time and energy to mastering complex ones. It is effective when there is a significant imbalance between the classes. The formula for focal loss is given in Equation (5)
F L p t = α t 1 p t γ log p t
where ( 1 p t ) γ is the modulating factor. Here, γ > 0, and when γ = 1, the focal loss will be the same as the cross-entropy loss. Similarly, γ can be set by the inverse class frequency or used as a hyperparameter with a typical range of [0, 1].

3.2.3. Dice Loss

In computer vision, the Dice coefficient is a metric utilized extensively to calculate the degree of similarity between two images. In the latter half of 2016, it was also adapted as a Dice loss function [25]. The formula for Dice loss is given in Equation (6).
F D L y ,   p ^ = 1 2 y p ^ + 1 y + p ^ + 1
In this case, 1 is added to both the numerator and denominator to prevent the function from being undefined in exceptional cases like y = p ^ = 0 .

4. Hyper-Dense VGG16 U-Net Segmentation Proposed Model

In various computer vision problems, shortcut connections between layers have become increasingly popular since the emergence of residual learning [26]. Unlike in conventional networks, these links back-propagate gradients immediately, which helps prevent gradient-vanishing issues and allows for more complex architectures. The idea of shortcut connections was expanded upon by DenseNet [27], which specified that each layer’s inputs should correspond to the outputs of all layers that came before them. Densely connected convolutional neural networks (CNNs) are built using the feed-forward principle, which entails adding direct connections from any layer to all succeeding layers. Deep networks are more accessible and more accurate to train because of this connectivity. This section proposes independently expanding U-Net to support DenseNet connections within the same multiple N streams of PET and CT modalities. Higher level layers of the proposed extension will also use the late fusion strategy.
The inspiration for this comes from three separate observations. First, all architectural feature maps are connected by short paths, enabling implicit deep supervision. Second, the network’s information and gradients are better able to flow because of the direct connections between all layers. Finally, the regularizing effect of dense connections makes it less likely that training data will be too small for a given task.
Using dense and hyper-dense connections has been demonstrated to have many benefits when segmenting medical images. More information can be gleaned from medical images when the VGG architecture is used for feature extractions. We propose a multimodality U-Net medical image segmentation model using hyper-dense connections and the VGG16 model.
Our primary objective was to refine an existing deep-learning model for lung cancer segmentation. To do this, we modified the U-Net design and used it as the starting point. The encoder and the decoder are both CNNs, making up the basic U-Net architecture. The encoder extracts features by first performing convolutional operations and then down-sampling. The usual convolutional processes follow the up-sampling and concatenation layer of the decoder branch. Connecting feature maps from the encoder network is made possible via a skip link that connects the same-level layers of the decoder and encoder, with the up-sampled feature map conveying coarse global context information. This helps with recovering local characteristics after down-sampling.
According to this model, U-Net takes data via two distinct input paths, one for each image type. The architecture of both paths is VGG16, with dense and hyper-dense connections between them. This architecture was proposed so that image classification and segmentation tasks may take advantage of VGG, dense, and ultra-dense networks. The suggested VGG16 U-Net model’s components are given in Figure 8.
Figure 8 depicts the proposed hyper-dense VGG16 U-Net model, built upon the U-Net. Both CT and PET images can be fed into the model. The segmented image of lung cancer is the product of the model. In the suggested approach, input images for both CT and PET were 128 × 128. Each image input type has its dedicated input path, each with 16 CNNs (the number of CNNs in VGG16). Each dataset was processed through CNNs of varying sizes (64, 128, 256, and 512). Figure 6 shows that both input paths are incredibly well connected, and there are also many connections between the two. All convolutional neural networks used ReLU activation. The decoder’s structure comprises four groups of convolutional neural networks (CNNs) of varying sizes (1024, 512, 256, and 128).

5. Experiments

The efficiency of the proposed U-Net models for segmenting lung tumours was measured across various performance criteria. The STS dataset was used for both training and testing the models. Experiments compared the newly developed models to benchmarked models widely utilized on the same dataset and other datasets.

5.1. Experimental Setup

All experiments were run on servers in the Google Colaboratory environment, and the recommended models for segmenting lung tumours were built using a TensorFlow and Keras backend with an NVIDIA Tesla P100 -PCIE GPU and 32.0 GB RAM. For the training phase, we employed the Adam optimizer with the following settings: learning rate = 0.0001, 1 = 0.9, 2 = 0.999, and epsilon = 1 × 10−8. One hundred epochs of training were used. The intensity levels of the image’s pixels were normalized to remove any potential for ambiguity. The dataset was arbitrarily divided into 70% for training, 20% for validation, and 10% for testing.

5.2. Dataset Description

The proposed models are trained on data from a study of soft-tissue sarcomas (STSs) [28]. STS includes many types of scans, such as CT, PET, and MRI, but in our research only the CT and PET scans were used. In this dataset, a cohort of 51 patients with histologically proven soft-tissue sarcomas (STSs) of the extremities was retrospectively evaluated. It included 27 females and 24 males, with ages ranging from 16 to 83 years old, and with various cancer degrees: low, intermediate, and high. The PET slice volumes had a thickness of 3.27 mm and a median in-plane resolution of 5.47 mm × 5.47 mm (range: 3.91–5.47 mm). All images used in our tests were downsized to 128 pixels on the longest dimension. For our experiments, we excluded slices without tumour pixels in the ground truth. In total, over 3000 pairs of PET-CT image slices (3063 STS) were included.

5.3. Performance Metrics

The efficiency of the suggested approach was assessed using the most commonly employed metric for evaluating segmentation tasks [29]: the Dice score ( D i c e ), the most crucial segmentation performance measure. It is defined by Equation (7).
D i c e = 2     T p 2     T p + F p + F n
In addition, measures of accuracy, sensitivity, and specificity were used. Equations (8) to (10) also provide definitions for them.
A c c u r a c y = T p + T n T p + T n + F n + F p
S e n s i t i v i t y = T p T p + F n
S p e c i f i c i t y = T n T n + F p
where the four primary blocks for computing these metrics were defined as true positive ( T p ), true negative ( T n ), false positive ( F p ), and false negative ( F n ) values.

6. Results and Discussion

The effectiveness of the proposed models is discussed in this section. In this section, we report the results of our model assessments, broken down into four categories: loss function comparisons, same-dataset comparisons, cross-dataset comparisons, and cross-model comparisons. Dice, IoU, accuracy, spectral sensitivity, and area under the curve (AUC) were utilized as performance measures.

6.1. Loss-Function-Based Comparison

Adjustments to the loss functions form the basis for a new comparative evaluation of the models. Focal loss functions, Dice, and binary cross-entropy are employed in our research. These operations are among the most well-known and often used in deep learning for image segmentation. The outcomes are displayed in Table 2, Table 3 and Table 4 for binary, Dice, and focal loss functions.
Figure 9, Figure 10 and Figure 11 depict the findings using various loss functions, like cross-entropy, focal loss, and Dice loss. In contrast, the performance measures for the proposed models using the metrics Dice, IoU, accuracy, sensitivity, and specificity are given in Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16, respectively.
The suggested hyper-dense VGG16 model outperforms the other models in Dice for all types of loss functions, as seen in Table 2, Table 3 and Table 4.The focal loss function is the only option if one wants the best Dice performance possible. Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 offer graphical representations of the evaluation outcomes. The five presented models are compared in Figure 9, Figure 10, Figure 11 and Figure 12 regarding the binary cross-entropy, Dice, and focal loss functions used as performance indicators. Figure 9 shows that the suggested hyper VGG16 model outperforms the others in terms of Dice accuracy (improved by 7%), IOU accuracy (improved by 9%), and specificity accuracy (improved by 0.4%). However, the late fusion model’s accuracy and sensitivity are unparalleled. The results of the Dice loss function are shown in Figure 10, and it is evident that the suggested model outperforms the previously introduced models in terms of Dice, specificity, and sensitivity. Finally, the proposed model outperforms the other established models regarding the focal loss function performance, achieving 73% for Dice. Figure 10, Figure 11, Figure 12 and Figure 13 present visual representations of the performance above metrics about the loss function employed. Figure 13 demonstrates that the most outstanding value for the focal loss function is found with the Dice metric. The segmentation results of the proposed model for various loss functions are displayed in Figure 17. In Figure 17, the lung tumour segmentation results generated by hyper-dense VGG16 are compared to the ground truth, employing various loss functions such as binary, Dice, and focal. The observations from Figure 17 indicate that the focal loss function yields the most accurate predictions, capturing even the segmentation of small tumour portions and producing a predicted segmentation mask that closely aligns with the ground truth segmentation. Conversely, when utilizing the binary cross-entropy loss function, the segmentation results tend to be slightly larger. The Dice loss function, however, provides the least accurate predictions, as it fails to segment small tumour portions and produces a larger overall segmentation compared to the ground truth.

6.2. Same Dataset Models in the State of the Art

Table 5 compares the Dice, IoU, accuracy, specify, and sensitivity scores of the proposed model to those of the benchmarked research conducted on the same STS dataset. As shown in Table 5, the dense fusion model outperforms the other models using the same dataset. Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 shown visual representations of the performance metrics which were used in performance evaluation and the comparison of the proposed models.
According to the data in the table and the graphs above, the suggested hyper-dense VGG16 model outperforms state-of-the-art models trained on the same dataset by a margin of Dice improvement equal to 17%. The proposed architecture for feature extraction between the two types of inputs and applying the VGG16 architecture of image feature extraction led to this enhancement.

6.3. Different Datasets in the State of the Art

Table 6 compares the Dice, IOU, accuracy, specificity, and sensitivity metrics of the proposed model to those of benchmarked studies that employed different datasets, demonstrating the model’s superior performance. Table 6 show that the dense fusion model has a higher Dice value than the other models. Figure 23, Figure 24, Figure 25, Figure 26 and Figure 27 show visual representations of the performance metrics used in performance evaluation and the comparison of the proposed models.
Based on the data in the table, we infer that the suggested hyper-dense VGG16 model outperforms the other state-of-the-art lung cancer segmentation models trained on a separate dataset by a margin of Dice improvement equal to 14%. The proposed architecture for feature extraction between the two types of inputs and applying the VGG16 architecture of image feature extraction led to this enhancement.

7. Conclusions and Future Scope

Despite significant tumour detection and treatment developments, lung cancer remains a leading cause of mortality globally. This study offers a deep learning model based on hyper-dense VGG16 deep technologies to segment lung cancer using U-Net as a backbone structure with multimodal input images. A comprehensive evaluation of the proposed model was included: introducing four models to compare with the proposed one, comparing the proposed model with others presented in the state of the art, and comparing the proposed model with the others using the same and different datasets. The hyper-dense VGG16 model performed the best out of all the analyses performed on this dataset, receiving a Dice score of 73%. The hyper-dense VGG16 model achieved better performance because the extraction of features in each layer is not only a function of the previous layer input, but a function of all previous inputs in the same input branch, and it has inputs from the other input branches which make the feature extraction deeper and more accurate. However, further research based on the outcomes of this study can go in a few different directions, such as employing a combination of CT, PET, and MRI as input to the segmentation model to verify the suggested model’s generalizability and introducing the performance study of the proposed model with respect to a combination of two types of loss functions such as binary and Dice loss functions. Furthermore, the model’s performance can be tested by experimenting with various hyperparameter values of the offered methodologies.

Author Contributions

G.M.A. contributed in terms of conceptualization, formal analysis, methodology, software, validation, visualization, writing—original draft, and writing—review and editing. N.M.’s involvement consisted of dataset’s resources. Q.N. led the conceptualization, project administration, and supervision. R.J. contributed through conceptualization, supervision, and review—editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by The Engineering and Physical Sciences Research Council (EPSRC) under Grant EP/P009727/1, and in part by the Leverhulme Trust under Grant RF-2019-492.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used for this study is publicly available and contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, J.; Li, J.; Wu, J.; Song, Y.; Hu, S.; Hong, J.; Wang, W. Change in Symptom Clusters Perioperatively in Patients with Lung Cancer. Eur. J. Oncol. Nurs. 2021, 55, 102046. [Google Scholar] [CrossRef] [PubMed]
  2. Ju, W.; Xiang, D.; Zhang, B.; Wang, L.; Kopriva, I.; Chen, X. Random Walk and Graph Cut for Co-Segmentation of Lung Tumor on PET-CT Images. IEEE Trans. Image Process. 2015, 24, 5854–5867. [Google Scholar] [CrossRef] [PubMed]
  3. Greco, C.; Rosenzweig, K.; Cascini, G.L.; Tamburrini, O. Current status of PET/CT for tumour volume definition in radiotherapy treatment planning for non-small cell lung cancer (NSCLC). Lung Cancer 2007, 57, 125–134. [Google Scholar] [CrossRef] [PubMed]
  4. Cellina, M.; Cè, M.; Irmici, G.; Ascenti, V.; Khenkina, N.; Toto-Brocchi, M.; Martinenghi, C.; Papa, S.; Carrafiello, G. Artificial Intelligence in Lung Cancer Imaging: Unfolding the Future. Diagnostics 2022, 12, 2644. [Google Scholar] [CrossRef] [PubMed]
  5. Cellina, M.; Cacioppa, L.M.; Cè, M.; Chiarpenello, V.; Costa, M.; Vincenzo, Z.; Pais, D.; Bausano, M.V.; Rossini, N.; Bruno, A. Artificial Intelligence in Lung Cancer Screening: The Future is Now. Cancers 2023, 15, 4344. [Google Scholar] [CrossRef] [PubMed]
  6. Hosny, K.M.; Elshoura, D.; Mohamed, E.R.; Vrochidou, E.; Papakostas, G.A. Deep Learning and Optimization-Based Methods for Skin Lesions Segmentation: A Review. IEEE Access 2023, 11, 85467–85488. [Google Scholar] [CrossRef]
  7. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  8. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar]
  9. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  10. Wang, S.; Mahon, R.; Weiss, E.; Jan, N.; Taylor, R.J.; McDonagh, P.R.; Quinn, B.; Yuan, L. Automated Lung Cancer Segmentation Using a PET and CT Dual-Modality Deep Learning Neural Network. Int. J. Radiat. Oncol. Biol. Phys. 2023, 115, 529–539. [Google Scholar] [CrossRef] [PubMed]
  11. Park, J.; Kang, S.K.; Hwang, D.; Choi, H.; Ha, S.; Seo, J.M.; Eo, J.S.; Lee, J.S. Automatic Lung Cancer Segmentation in [18F]FDG PET/CT Using a Two-Stage Deep Learning Approach. Nucl. Med. Mol. Imaging 2023, 57, 86–93. [Google Scholar] [CrossRef] [PubMed]
  12. Xiang, D.; Zhang, B.; Lu, Y.; Deng, S. Modality-Specific Segmentation Network for Lung Tumor Segmentation in PET-CT Images. IEEE J. Biomed. Health Inform. 2023, 27, 1237–1248. [Google Scholar] [CrossRef]
  13. Fu, X.; Bi, L.; Kumar, A.; Fulham, M.; Kim, J. Multimodal Spatial Attention Module for Targeting Multimodal PET-CT Lung Tumor Segmentation. IEEE J. Biomed. Health Inform. 2021, 25, 3507–3516. [Google Scholar] [CrossRef]
  14. Zhong, Z.; Kim, Y.; Zhou, L.; Plichta, K.; Allen, B.; Buatti, J.; Wu, X. 3D fully convolutional networks for co-segmentation of tumors on PET-CT images. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 228–231. [Google Scholar] [CrossRef]
  15. Hwang, S.E.; Hwang, D.; Kang, S.K.; Choi, H.; Ha, S.; Eo, J.S.; Lee, J.S. 3C-Net: Deep Learning-based Lung Cancer Segmentation Using Multi-Context Information on FDG PET/CT Images. J. Nucl. Med. 2022, 63, 3349. [Google Scholar]
  16. Kumar, A.; Fulham, M.; Feng, D.; Kim, J. Co-Learning Feature Fusion Maps from PET-CT Images of Lung Cancer. IEEE Trans. Med. Imaging 2020, 39, 204–217. [Google Scholar] [CrossRef] [PubMed]
  17. Jemaa, S.; Fredrickson, J.; Carano, R.A.D.; Nielsen, T.; de Crespigny, A.; Bengtsson, T. Tumor Segmentation and Feature Extraction from Whole-Body FDG-PET/CT Using Cascaded 2D and 3D Convolutional Neural Networks. J. Digit. Imaging 2020, 33, 888–894. [Google Scholar] [CrossRef] [PubMed]
  18. Zhao, X.; Li, L.; Lu, W.; Tan, S. Tumor co-segmentation in PET/CT using multi-modality fully convolutional neural network. Phys. Med. Biol. 2019, 64, 015011. [Google Scholar] [CrossRef] [PubMed]
  19. Zhong, Z.; Kim, Y.; Plichta, K.; Allen, B.G.; Zhou, L.; Buatti, J.; Wu, X. Simultaneous cosegmentation of tumors in PET-CT images using deep fully convolutional networks. Med. Phys. 2019, 46, 619–633. [Google Scholar] [CrossRef] [PubMed]
  20. Bi, L.; Fu, X.; Liu, Q.; Song, S.; Feng, D.D.; Fulham, M.; Kim, J. Hyper-Connected Transformer Network for Co-Learning Multi-Modality PET-CT Features. arXiv 2022, arXiv:2210.15808. [Google Scholar]
  21. Gil, J.Y.; Kimmel, R. Efficient Dilation, Erosion, Opening and Closing Algorithms. In Mathematical Morphology and Its Applications to Image and Signal Processing; Springer: Boston, MA, USA, 2002; pp. 301–310. [Google Scholar] [CrossRef]
  22. Sreedhar, K. Enhancement of Images Using Morphological Transformations. Int. J. Comput. Sci. Inf. Technol. 2012, 4, 33–50. [Google Scholar] [CrossRef]
  23. Ma, Y.D.; Liu, Q.; Qian, Z.B. Automated image segmentation using improved PCNN model based on cross-entropy. In Proceedings of the 2004 International Symposium on Intelligent Multimedia, Video and Speech Processing, Hong Kong, China, 20–22 October 2004; pp. 743–746. [Google Scholar] [CrossRef]
  24. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  25. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, 14 September 2017; Springer: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 39, pp. 4700–4708. [Google Scholar]
  28. Vallières, M.; Freeman, C.R.; Skamene, S.R.; El Naqa, I. A radiomics model from joint FDG-PET and MRI texture features for the prediction of lung metastases in soft-tissue sarcomas of the extremities. Phys. Med. Biol. 2015, 60, 5471–5496. [Google Scholar] [CrossRef] [PubMed]
  29. Müller, D.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes 2022, 15, 210. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Block diagram for the lung cancer segmentation framework.
Figure 1. Block diagram for the lung cancer segmentation framework.
Diagnostics 13 03481 g001
Figure 2. Some examples of the augmentation process of CT and PET images for STS: (a) the main CT-PET, (b) rotating the CT-PET by 90 degrees clockwise, (c) flipping the CT-PET upside down, and (d) left-mirroring the CT-PET. Red arrows indicate the tumour region.
Figure 2. Some examples of the augmentation process of CT and PET images for STS: (a) the main CT-PET, (b) rotating the CT-PET by 90 degrees clockwise, (c) flipping the CT-PET upside down, and (d) left-mirroring the CT-PET. Red arrows indicate the tumour region.
Diagnostics 13 03481 g002
Figure 3. An example of the post-processing effect: (a) predicted mask and (b) image after mask post-processing.
Figure 3. An example of the post-processing effect: (a) predicted mask and (b) image after mask post-processing.
Diagnostics 13 03481 g003
Figure 4. Early fusion architecture.
Figure 4. Early fusion architecture.
Diagnostics 13 03481 g004
Figure 5. Late fusion architecture.
Figure 5. Late fusion architecture.
Diagnostics 13 03481 g005
Figure 6. Dence fusion architecture.
Figure 6. Dence fusion architecture.
Diagnostics 13 03481 g006
Figure 7. Hyper-dense fusion architecture.
Figure 7. Hyper-dense fusion architecture.
Diagnostics 13 03481 g007
Figure 8. The proposed hyper-dense VGG16 U-Net model architecture.
Figure 8. The proposed hyper-dense VGG16 U-Net model architecture.
Diagnostics 13 03481 g008
Figure 9. Binary cross-entropy function.
Figure 9. Binary cross-entropy function.
Diagnostics 13 03481 g009
Figure 10. Dice loss function.
Figure 10. Dice loss function.
Diagnostics 13 03481 g010
Figure 11. Focal loss function.
Figure 11. Focal loss function.
Diagnostics 13 03481 g011
Figure 12. Dice metric.
Figure 12. Dice metric.
Diagnostics 13 03481 g012
Figure 13. IOU metric.
Figure 13. IOU metric.
Diagnostics 13 03481 g013
Figure 14. Accuracy metric.
Figure 14. Accuracy metric.
Diagnostics 13 03481 g014
Figure 15. Specificity metric.
Figure 15. Specificity metric.
Diagnostics 13 03481 g015
Figure 16. Sensitivity metric.
Figure 16. Sensitivity metric.
Diagnostics 13 03481 g016
Figure 17. The comparison of lung tumour segmentation results, along with the segmentation outcomes for corresponding enlarged tumour regions, using our proposed hyper-dense VGG16 model with various loss functions (“Binary”, “Dice”, and “Focal”). The green contours outline the “Ground Truth” segmentation, and the blue contours outline results from the proposed model.
Figure 17. The comparison of lung tumour segmentation results, along with the segmentation outcomes for corresponding enlarged tumour regions, using our proposed hyper-dense VGG16 model with various loss functions (“Binary”, “Dice”, and “Focal”). The green contours outline the “Ground Truth” segmentation, and the blue contours outline results from the proposed model.
Diagnostics 13 03481 g017
Figure 18. Dice [13,20].
Figure 18. Dice [13,20].
Diagnostics 13 03481 g018
Figure 19. Accuracy [13,20].
Figure 19. Accuracy [13,20].
Diagnostics 13 03481 g019
Figure 20. Specificity [13,20].
Figure 20. Specificity [13,20].
Diagnostics 13 03481 g020
Figure 21. IOU [13,20].
Figure 21. IOU [13,20].
Diagnostics 13 03481 g021
Figure 22. Sensitivity [13,20].
Figure 22. Sensitivity [13,20].
Diagnostics 13 03481 g022
Figure 23. Dice [13,16].
Figure 23. Dice [13,16].
Diagnostics 13 03481 g023
Figure 24. Accuracy [13,16].
Figure 24. Accuracy [13,16].
Diagnostics 13 03481 g024
Figure 25. Specificity [13,16].
Figure 25. Specificity [13,16].
Diagnostics 13 03481 g025
Figure 26. IOU [13,16].
Figure 26. IOU [13,16].
Diagnostics 13 03481 g026
Figure 27. Sensitivity [13,16].
Figure 27. Sensitivity [13,16].
Diagnostics 13 03481 g027
Table 1. Literature summary.
Table 1. Literature summary.
Author CT-Only ExtractorPET-Only ExtractorFeature FusionDataset Description
Wang et al. [10]3D CNN3D CNN3D CNNPrivate clinic dataset comprising 290 pairs of CT and PET.
Park et al. [11]Global U-NetGlobal U-NetRegional U-NetPrivate data of 887 individuals with lung cancer.
Xiang et al. [12]Dual-stream encoderDual-stream encoderDecoder branch126 PET-CT scans containing NSCLC
Fu et al. [13]Encoder-decoder backbone CNNMultimodal spatial attention module (MSAM).CNN architecture containing skip connections.Two clinical PET-CT datasets of NSCLC and STS.
Zhong et al. [14]3D U-Net3D U-NetGraph-cut-based co-segmentation modelPET-CT scans from lung cancer patients.
Hwang et al. [15]Shared co- encoderShared co-encoderShared co-encoderF-18-FDG PET/CT scans from a private hospital.
Kumar et al. [16]Encoder using multiscale outputAn encoder using multiscale outputDecoder using multiscale multimodal inputBiopsy-proven NSCLC FDG PET-CT scans.
Jemaa et al. [17]--2D U-Net and selected VNetPatients with non-Hodgkin’s lymphoma and NSCLC, which includes 3664 FDG-PET/CT images from head to toe.
Zhao et al. [18]VNetVNetVoxel-wise addition, along with VNetPrivate clinical dataset having 3D PET/CT images.
Zhong et al. [19]Encoder using multiscale outputAn encoder using multiscale outputDecoder using multiscale multimodal inputNSCLC patients who received stereotactic radiation treatment.
Bi et al. [20]CNN-TN EncoderCNN-TN EncoderTN-CNN decoderNon-small-cell lung cancer (NSCLC) and one soft-tissue sarcoma (STS) dataset.
Table 2. Binary cross-entropy.
Table 2. Binary cross-entropy.
DiceIOUACCSenSpec
Late0.678820.536510.985160.738850.99068
Early0.680660.540750.983970.738160.99083
Dense0.695690.540160.980950.684010.99225
Hyper0.718510.578180.983810.723020.99284
Hyper + VGG160.725320.586870.982780.692090.99423
Table 3. Dice loss function.
Table 3. Dice loss function.
DiceIOUACCSenSpec
Late0.514790.697340.978060.646040.99048
Early0.514650.696710.979930.654520.99135
Dense0.514850.511910.981120.672530.99112
Hyper0.640810.697250.982950.679580.99046
Hyper + VGG160.668280.522220.980480.695060.99102
Table 4. Focal loss function.
Table 4. Focal loss function.
DiceIOUACCSenSpec
Late0.712170.57040.983470.710460.99327
Early0.661120.510110.979430.63510.99232
Dense0.715540.574030.981980.701310.99347
Hyper0.727130.587170.984360.717860.99339
Hyper + VGG160.730110.556640.981030.674720.99362
Table 5. Comparison of the proposed and benchmarked models on the STS dataset.
Table 5. Comparison of the proposed and benchmarked models on the STS dataset.
DiceIOUACCSenSpec
Fu et al. [13]62.26--64.7499.7
Bi et al. [20]66.36--69.9399.69
Late0.7121710.57040.983470.710460.99327
Early0.6611160.510110.979430.63510.99232
Dense0.7155390.574030.981980.701310.99347
Hyper0.727130.587170.984360.717860.99339
Hyper + VGG160.73.01090.556640.981030.674720.99362
Table 6. Comparison of the proposed and benchmarked models on different datasets.
Table 6. Comparison of the proposed and benchmarked models on different datasets.
DiceIOUACCSenSpec
Fu et al. [13]67.83--99.976.16
Kumar et al. [16]63.85----
Late0.7121710.5703970.9834710.7104620.993269
Early0.6611160.5101140.9794280.6350950.992316
Dense0.7155390.574030.9819760.7013140.993468
Hyper0.727130.5871710.9843620.7178610.993387
Hyper + VGG160.7301090.5566350.9810340.6747170.99362
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alshmrani, G.M.; Ni, Q.; Jiang, R.; Muhammed, N. Hyper-Dense_Lung_Seg: Multimodal-Fusion-Based Modified U-Net for Lung Tumour Segmentation Using Multimodality of CT-PET Scans. Diagnostics 2023, 13, 3481. https://doi.org/10.3390/diagnostics13223481

AMA Style

Alshmrani GM, Ni Q, Jiang R, Muhammed N. Hyper-Dense_Lung_Seg: Multimodal-Fusion-Based Modified U-Net for Lung Tumour Segmentation Using Multimodality of CT-PET Scans. Diagnostics. 2023; 13(22):3481. https://doi.org/10.3390/diagnostics13223481

Chicago/Turabian Style

Alshmrani, Goram Mufarah, Qiang Ni, Richard Jiang, and Nada Muhammed. 2023. "Hyper-Dense_Lung_Seg: Multimodal-Fusion-Based Modified U-Net for Lung Tumour Segmentation Using Multimodality of CT-PET Scans" Diagnostics 13, no. 22: 3481. https://doi.org/10.3390/diagnostics13223481

APA Style

Alshmrani, G. M., Ni, Q., Jiang, R., & Muhammed, N. (2023). Hyper-Dense_Lung_Seg: Multimodal-Fusion-Based Modified U-Net for Lung Tumour Segmentation Using Multimodality of CT-PET Scans. Diagnostics, 13(22), 3481. https://doi.org/10.3390/diagnostics13223481

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop