Next Article in Journal
Secure Encapsulation Schemes Using Key Recovery System in IoMT Environments
Next Article in Special Issue
Fully Parallel Implementation of Otsu Automatic Image Thresholding Algorithm on FPGA
Previous Article in Journal
THz-TDS Reflection Measurement of Coating Thicknesses at Non-Perpendicular Incidence: Experiment and Simulation
Previous Article in Special Issue
Computational Methods for Liver Vessel Segmentation in Medical Imaging: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Guided Network with Densely Connected Convolution for Skin Lesion Segmentation

College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2021, 21(10), 3462; https://doi.org/10.3390/s21103462
Submission received: 13 April 2021 / Revised: 9 May 2021 / Accepted: 11 May 2021 / Published: 16 May 2021
(This article belongs to the Special Issue Advances in Image Segmentation: Theory and Applications)

Abstract

:
The automatic segmentation of skin lesions is considered to be a key step in the diagnosis and treatment of skin lesions, which is essential to improve the survival rate of patients. However, due to the low contrast, the texture and boundary are difficult to distinguish, which makes the accurate segmentation of skin lesions challenging. To cope with these challenges, this paper proposes an attention-guided network with densely connected convolution for skin lesion segmentation, called CSAG and DCCNet. In the last step of the encoding path, the model uses densely connected convolution to replace the ordinary convolutional layer. A novel attention-oriented filter module called Channel Spatial Fast Attention-guided Filter (CSFAG for short) was designed and embedded in the skip connection of the CSAG and DCCNet. On the ISIC-2017 data set, a large number of ablation experiments have verified the superiority and robustness of the CSFAG module and Densely Connected Convolution. The segmentation performance of CSAG and DCCNet is compared with other latest algorithms, and very competitive results have been achieved in all indicators. The robustness and cross-data set performance of our method was tested on another publicly available data set PH2, further verifying the effectiveness of the model.

1. Introduction

One third of cancers worldwide are skin cancers [1]. Currently, 2 to 3 million cases of non-melanoma skin cancer and 132,000 cases of melanoma skin cancer occur globally each year. It is estimated that in the United States, there were 96,480 new cases and 7230 deaths from melanoma in 2019. Melanoma accounts for less than 5% of all skin cancers, but 75% of skin cancer deaths are related to melanoma. Studies have shown that the 5-year survival rate of patients with advanced malignant melanoma is only 15%, while the final cure rate of early patients is as high as 95% [2]. Patients with benign melanoma only need to be found early and removed to prevent the disease from being life-threatening [3]. Therefore, the diagnosis of benign and malignant melanoma as well as early and late stages can play an extremely important role in the survival of melanoma patients.
Dermoscopy is an image formed based on the imaging principle of removing the reflection of the skin surface and visually enhancing the deeper skin [4]. Compared with visual inspection, dermoscopy can improve the diagnostic accuracy rate by 20% [5]. Skin lesion segmentation is one of the important steps in the computer-aided diagnosis of various skin diseases. As shown in Figure 1, due to the huge differences in the size, location, shape and color of the lesions in different patients, and a large number of artifacts, including inherent skin features (such as hair, blood vessels) and artificial artifacts (such as bubbles, ruler marks, uneven lighting, incomplete lesion areas, etc.), the automatic segmentation of lesions in dermoscopic images is very challenging. In addition, the low contrast between the lesion and the surrounding texture also hinders the automatic segmentation of the lesion area.
With the help of computer-aided diagnosis, segmentation of skin lesion images to obtain lesion areas helps doctors quickly identify the location of the lesion area, which can improve the diagnosis rate and enable patients to be treated early. In the early stage, the skin lesion segmentation image use edge detection, threshold segmentation, active contour (expectation maximization, level set, clustering, etc.) or region-based (region growth, iterative stochastic region merging, etc.) hybrid technology [6,7]. Although the above segmentation algorithm has a certain segmentation effect, it relies too much on the quality of manual feature selection and introduces prior information. Moreover, it is difficult for the recognition model based on artificial features to obtain good generalization ability for skin lesions images with highly changed clinical manifestations.
Recently, supervised methods have achieved promising results in the field of computer vision, but they rely on annotated training data sets, which require the proficiency of humans and related background knowledge. In contrast, unsupervised learning makes data-driven decisions by obtaining insights directly from the data itself. Unsupervised learning is applied to all aspects of image processing. Ahmed et al. [8] proposed a low-rank tensor with a sparse mixture of Gaussian (LRTSMoG) decomposition algorithm for natural crack detection. He proposed algorithm models jointly the LRST pattern by using a tensor decomposition framework. In particular, the weak natural crack information can be extracted from strong noise. Gupta et al. [9] investigate the utility of unsupervised machine learning and data visualisation for tracking changes in user activity over time. The goal of semi-supervised and unsupervised image segmentation is to greatly reduce or even eliminate the need for training data, thereby minimizing the burden on clinicians when training the segmentation model. Li et al. [10] present a novelsemi-supervised method for skin lesion segmentation, where the network is optimizedby the weighted combination of a common supervised loss for labeled inputs only anda regularization loss for both labeled and unlabeled data. Pathan et al. [11] proposed a deep clustering architecture and formal image analysis for image segmentation. The main idea is based on an unsupervised learning method, clustering the images of the severity of the disease in the sample of the subject, and then segmenting the image to highlight and outline the area of interest. Feyjie et al. [12] proposed a novel small-scale learning framework for semantic segmentation, in which unlabeled images can be used in each plot.
Recently, the Deep Convolutional Neural Network (DCNN) model segment skin lesions into pixel-level classification problems and achieve remarkable success [13,14,15,16]. No matter from the deep full convolutional neural network (FCN), which earliest uses in image segmentation [17], to UNet [18], which extends the architecture of FCN, and various extension models based on UNet architecture, they all have the contradiction between semantic information and spatial location information. The feature semantic information in the deep layers of the network is richer, but through step-by-step pooling, the spatial resolution of the feature is continuously reduced, and the spatial position information is continuously lost. If the output is directly upsampled to the original input resolution, the segmentation result will be rougher [19]. For dermoscopy images, due to its small grayscale changes and relatively blurred boundaries, the underlying features with rich spatial location information are particularly important for restoring feature spatial resolution.
In recent years, the attention mechanism has been widely used in deep neural networks and has been widely used in many tasks [20,21,22]. Most applications in computer vision and computer graphics involve the concept of image filtering to reduce noise or extract useful image structure. The guided filter [23] is an edge-preserving image filter, which serves as a special extension path that transfers the structural information extracted from the low-level feature map to the high-level feature map. It is effective and efficient in a variety of computer vision and computer graphics applications, including noise reduction, detail smoothing/enhancement, HDR compression, image extinction/feathering, haze removal, and joint upsampling.
In this article, we propose a Channel Spatial Attention-guided network with Densely Connected Convolutional (CSAG and DCCNet) based on deep CNN. The model is equipped with a novel and effective Channel Spatial Fast Attention-guided Filter module (CSFAG) for dermoscopic image segmentation. Compared with the existing attention methods widely used in CNN, the CSFAG module has three advantages. First, we directly impose some constraints on the unknown output by considering the guide image and use spatial attention and channel attention to collect information around the feature map. The calculated attention weight is applied to the guide feature map, so that the model can further capture the correlation of dimensional features, focus on the lesion area, and reduce the influence of noise on the segmentation performance of the model. Secondly, the CSFAG module supports the fusion of multi-resolution features. This module can recover spatial information by filtering low-resolution feature maps and high-resolution feature maps, and merge structural information of various resolution levels to better retain spatial location information. Third, adding the CSFAG module we proposed to the network can greatly improve the performance of lesion segmentation, and the module can be seamlessly integrated into multiple basic segmentation network architectures, with good robustness.
Our specific work content is as follows:
(1)
We have designed a novel Channel Spatial Fast Attention-guided Filter (referred to as CSFAG). Through a large number of ablation experiments, it is verified that the CSFAG module is superior to other mainstream attention modules and can be combined with multiple basic segmentation networks, which can effectively improve the segmentation performance of the model.
(2)
Embed the proposed CSFAG module into the jump connection of the M-Net segmentation network to form CSAG and DCCNet. In the last step of the CSAG and DCCNet encoding path, densely connected convolutions are used to replace ordinary convolutional layers, and through the idea of “collective knowledge”, the gap between low-level features is bridged and features are effectively aggregated.
(3)
On the ISIC-2017 data set, the segmentation performance of CSAG and DCCNet was compared with other latest algorithms. The six indicators of accuracy, sensitivity, specificity, Dice coefficient, Jaccard coefficient, and Matthew correlation coefficient were all achieved very competitive results. CSAG and DCCNet trained in ISIC-2017 was tested on another publicly available data set, PH2, to verify the robustness and cross-dataset performance of our method.
The rest of this paper is arranged as follows: related work is introduced in Section 2. Section 3 introduces the proposed network architecture in detail; experiments and comparison results are explained in Section 4; finally, Section 5 gives our conclusion.

2. Work Organization

2.1. Semantic Segmentation Model

The method based on full convolutional network (FCN) has made great progress in semantic segmentation. The initial popular depth learning method for semantic segmentation tasks is patch classification, which uses image blocks around the pixel to classify each pixel independently. J. Long et al. [17] first applied the FCN to the end-to-end training of image segmentation, which makes the convolutional neural network can perform dense pixel prediction without a fully connected layer. This method to produce image segmentation maps of any size. Since then, in the field of semantic segmentation, this model is adopted by almost all advanced methods. Except for the fully connected layer, another problem with convolutional neural networks for semantic segmentation is the use of pooling layers. Although the pooling layer expands the receptive field and aggregates the context, it causes the loss of location information.
There are two different structures to solve this problem. The first is the encoder-decoder structure. The encoder gradually reduces the spatial dimension through the pooling layer, and the decoder gradually repairs the details and spatial dimensions of the object through methods such as bilinear interpolation. The encoder and the decoder are usually embedded with a skip connection, which can better help the decoder to repair the details of the target. U-Net [18] is the most commonly used structure in this method. Since then, the UNet-based structure has derived multiple segmentation networks, such as V-Net [24], UNet++ [25], MultiResUNet [26], UNET 3+ [27]. The second method is to use the atrous convolution structure to remove the pooling layer. On the premise of not reducing the spatial dimension, the atrous convolutional layer improves the receptive field index, captures multi-scale context information and maintains the spatial position relative of the feature map, while the pooing will introduce translation invariance. Classical semantic segmentation networks from Deeplab series networks [28,29,30] to PSPNet [31] all use atrous convolution.

2.2. Attention Mechanism

As we all know, attention plays an important role in human perception [32,33]. The attention mechanism has been widely used in many tasks. Earlier, the Google Deep Mind team first used the attention mechanism on the RNN model for image description problems [34]. Subsequently, they proposed a model based on the attention mechanism for the recognition of multiple objects in the image [35]. Typical attention models on a single image are residual attention network [36] and squeeze-and-excitation Networks (SENet) [37]. The residual attention network includes two attention components, a stacked network structure composed of multiple attention components, and residual attention learning that combines the residual structure with the attention mechanism. SENet includes a squeeze and excitation block to retain the channel attention introduced for each residual block. PSANet [38] aggregates the contextual information of each location through the predicted attention map. A 2 Net [39] propose a dual attention block, which can collect informative global features from the entire time and space distribution of the image. DANet [40] is based on the self-attention mechanism to capture rich contextual relevance to solve the scene segmentation task, while applying spatial and channel attention to collect information around the feature map. Sanghyun Woo [41] proposed an attention module called CBAM, which can be embedded into classic deep networks to improve model performance. Huisi Wu et al. [42] proposed a deep learning model equipped with a new and efficient adaptive dual attention module (ADAM) to automatically segment skin lesions from dermoscopic images. Most of the above methods only focus on feature information from the two parts of spatial and channel. However, in some recent studies, some scholars use both spatial and temporal domains for object segmentation with self-attention. Hamad et al. [43] proposed dilated causal convolution with multi-head self attention for sensor human activity recognition. The multi-headed self-attention is used to enable the model to focus on important and relevant time steps more than the insignificant time steps from the sequential feature maps during recognition. Hu et al. [44] proposes a hybrid multi-dimensional features fusion structure of spatial and temporal segmentation model for automated thermography defects detection. They design a new attention block to provide spatiotempo-ral attention to focus on semantically meaningful regionsof the volumetric data and recalibrate the feature mapsadaptively, based on the weighted channels.
However, all of the above methods only emphasize the local focus of the pixel relationship or the global focus of the entire image. Due to the different shapes, sizes and colors of lesions; skin types and characteristics; inherent skin characteristics, a single attention mechanism or the existing two-dimensional attention module with dimensional integration still cannot cope with the challenges in skin lesion segmentation. In this work, we combined image filtering with attention, and designed a Channel and Spatial Fast Attention-guided Filter (CSFAG) module to preserve the smooth characteristics of the edges of skin lesions without being subject to gradient inversion The effect of artifacts. The CSFAG module performs well in terms of quality and efficiency.

2.3. Dermoscopic Image Segmentation

Earlier, Kawahara, etc. proposed a fully convolutional network based on Alex Net to extract the surface features of melanoma [45]. Subsequently, Long [17] proposed a fully convolutional neural network FCN, which uses a fully convolutional layer instead of a fully connected layer to convert the classification convolutional neural network into a segmentation network. Lequan uses ResNet based on the FCN network structure to segment the lesion area on the ISIC dermoscopic image data set [46]. UNet, proposed by Ronneberger et al. [18], is one of the most popular FCN structures for medical image segmentation. Tschandl P is based on migration learning and uses the LinkNet structure to use the ResNet classification network pre-trained on the ImageNet data set for the coding part of the segmentation. It has achieved a significant performance improvement on the task of segmentation of skin lesions [47]. The DeepLab series of work introduced delitated convolution to reduce the loss of the resolution of the coding part and increase the receptive field. Goyal applied DeepLabV3+ to the task of skin lesions segmentatio [48]. In order to reduce the number of parameters that make the network lightweight, Md. Hasan et al. uses depthwise separable convolution instead of standard convolution, and projects the learned distinguishing features onto the pixel space at different stages of the encoder [49]. However, in clinical practice, due to the complexity of the lesion and the significant increase in the number of dermoscopic images, the lightweight, robustness and ability of the segmentation model to be combined with multiple basic networks are becoming more and more important. In our work, our goal is to develop a segmentation model that is easy to transplant and has good robustness.

3. Proposed CSAG and DCCNet Model

The CSAG and DCCNet proposed in this paper is an end-to-end multi-label deep network, which consists of multiple novel channel and space fast attention-oriented filter modules (CSFAG), densely connected convolution, multi-scale input layer, UNet and side output layers. In the CSFAG module, the guided filter is combined with the attention module, and a novel attention-oriented filter is designed and embedded in the skip connection of CSAG and DCCNet. In the last step of the CSAG and DCCNet coding path, densely connected convolution is used to replace the ordinary convolutional layer. Using dense connection convolution in the last step of the encoding path can recover spatial information with the help of rich context information, further enriching context information and reducing the difficulty of training. CSAG and DCCNet uses U-Net as the basic network structure, and builds an image pyramid input on the left side of the coding layer to achieve a multi-scale layer of multi-level receiving field fusion; on the right side of the decoder, it introduces a side output layer, average all side output maps as the final prediction map. Simultaneously use the multi-label loss function to update the parameters to train the model The architecture of this model is shown in Figure 2.

3.1. CSFAG Module

The concept of image filtering is widely used in computer vision and computer graphics. Simple linear translation invariant (LTI) filters, such as Gaussian filters, Laplacian filters and Sobel filters, are widely used for image blur/sharpening, edge detection and feature extraction. The kernel of the LTI filter is spatially invariant and has no relation to the image content. Considering the above reasons, we hope to merge other information from the given guide image during the image filtering process so that the image filter can establish a connection with the structural information in the dermoscopic image, in order to better improve the feature extraction ability of the model. We consider using guided images to directly impose some constraints on the output, and use spatial attention and channel attention to collect information around the feature map. Therefore, this article is inspired by the above-mentioned methods and designs a channel and space fast attention-oriented filter (CSFAG) module. The module is mainly composed of two parts, the fast guide filter and the attention module. The model architecture of this module is shown in Figure 3.
The input of the CSFAG module includes the guided feature map G and the filtered feature map F, and the output is a high-resolution feature map O. First, we use bilinear interpolation to scale the guiding feature map G to generate a low-resolution feature map G l . G l and filtered feature map F through attention module to generate attention feature map A. We minimize the reconstruction error [50] between G l and F to obtain the coefficients W l and B l of the CSFAG module. By bilinearly up-sampling W l and B l , we obtain the coefficients W h and B h which match the guiding feature map G. Finally, the output O of the CSFAG module is obtained through linear transformation.
Specifically, the channel and space fast attention-oriented filter constructs a local square window ω k with a radius of r and a pixel k as the center to realize a local linear model. Assuming that G l i is the pixel of G l , the output O k i of ω k is obtained by linear transformation:
O k i = W k G l i + B k , i ω k
In order to determine the coefficients W k and B k , it is necessary to minimize the reconstruction error between the outputs O k i and the filtered feature F i of all pixels in the window ω k , as shown in Equation (2), where λ is a regularization parameter which controls the smoothness. A i is the attention weight at position i, obtained by the attention module:
min W , B k E W k , B k = i θ k A i 2 W k G l + B k 2 + λ W k 2
W k = 1 | ω | i ω k G l i F i μ k F ¯ k σ k 2 + λ
B k = F ¯ k W k μ k
In Equations (3) and (4), μ k and σ k are the mean and variance of G in the window ω k , | ω | is the number of pixels in ω k , and F ¯ k = 1 | ω | i ω is the average of F in ω k .
After calculating the coefficients W k and B k in the window ω k , we get the output O k i corresponding to each window. Average O k i obtained from different windows to generate O i which is equal to the average coefficient of all windows overlapping with i. As shown in Equation (5),
O i = 1 N k k Ω i W k G i + 1 N K k Ω i B k = W l G l + B l
where Ω i is the set of all windows including position i, and * is element-wise multiplication. After up-sampling W l and B l to obtain W h and B l respectively, the final output is calculated as O = W h G + B h .
The calculation process of the CSFAG module is described in detail in Algorithm 1.
In order to further demonstrate that the CSFAG module can help the model generate clearer boundaries, highlight the lesion area and reduce background influence, We randomly selected a melanoma dermoscopy image in the test set of the ISIC2017 data set for visual analysis. We select the CSFAG module on the jump connection in the fourth layer of the CSAG and DCCNet model as the test module. For showing the role of the CSFAG module more clearly, we perform three convolution operations on the input of the CSFAG module, the output of the intermediate attention module and the final output of the CSFAG module to reduce the dimensionality. Then use bilinear interpolation to restore the feature size to 256 × 256 . We have drawn a CAM heat map to more clearly show the changes produced by the CSFAG feature map. As shown in Figure 4, we can clearly see from the CAM heat map, after the attention module in the CSFAG module, the features are more focused on the lesion area. In particular, the final output results generated after the fast guide filter, the model further aggregates the characteristics of the space and the channel, enhances the discriminative ability of the network, and makes the network good use of the given characteristics.
Algorithm 1 Channel Spatial Fast Attention-guided Filter
Input: 
the guided feature map G and the filtered feature map F, parameters are r and λ
Output: 
high-resolution feature map O
1:
G l = f s u b s a m p l e ( G )
2:
F C A = f C h a n n e l A t t e n t i o n ( F ) , F S A = f S p a t i a l A t t e n t i o n ( F C A )
G C A = f C h a n n e l A t t e n t i o n ( G l ) , G S A = f S p a t i a l A t t e n t i o n ( G C A )
3:
A 0 = F S A + G S A , A 1 = F R e l u ( A 0 ) , A 2 = F c o n v o l u t i o n ( A 1 ) , A = F s i g m o d ( A 2 )
4:
ω k = a local square window with a radius of r and a pixel k as the ccenter
5:
μ k = 1 N i = 0 N G l , i ω k
6:
σ k 2 = 1 N i = 0 N ( i μ ) 2 G l , i ω k
7:
F ¯ k = 1 | ω | i = 0 N F , i ω k
8:
W k = 1 | ω | i ω k G l i F i μ k F ¯ k σ k 2 + λ
9:
min W k , B k E W k , B k = i ω k A i 2 W k G l + B i 2 + λ W k 2
10:
O k i = W k G l i + B k , i ω k
11:
W l = 1 N k k Ω i W k , B l = 1 N k k Ω i B k
12:
O i = W l G l + B l , i G
13:
W h = f u p s a m p l e ( W l ) , B h = f u p s a m p l e ( B l )
14:
O = W h G + B h
15:
 return O

Channel and Spatial Attention Learning

The attention mechanism on the channel has been proposed in Hu et al.’s SENet [37], and it has been verified that it can improve network performance. The channel attention module and the spatial attention module [41] are shown in Figure 5. The Channel Attention Module compresses the feature map in the spatial dimension to obtain a one-dimensional vector before performing operations. As shown in Figure 5a, the input feature map F C R C × H × W is passed through global max pooling and global average pooling based on width and height respectively, and then passed through a multilayer perceptron (MLP). The MLP output feature maps F A v g C R C × 1 × 1 and F M a x C R C × 1 × 1 are summed element-by-element, so that channel attention maps X c can be generated, as shown in Equation (6). After the sigmoid activation operation, the final channel attention feature map d is generated, as shown in Equation (7). The channel attention featuremap and input featuremap are subjected to elementwise multiplication operations to generate the input features required by the spatial attention module.
X c = F A v g C F M a x C , X c R C × 1 × 1
where ⊕ represents addition element by element.
A c = i = 0 c 1 1 + exp X i , 1 , 1 , A c R C × 1 × 1
where C is the number of channels, X i , 1 , 1 is the element with coordinate ( i , 1 , 1 ) , and ⋃ is element-by-element contacting.
The Spatial Attention Module compresses the channel, and performs average pooling and maximum pooling in the channel dimensions. As shown in Figure 5b, the feature map A c R C × H × W output by the channel attention module is used as the input feature map of this module. First, the average pooling and maximum pooling operations along the channel axis are used to obtain two two-dimensional feature maps F A v g S and F M a x S (the specific process is shown in Equations (9) and (10)), and then concat the two results based on the channel. The feature map of the merged 2 channels is subjected to a convolution operation to reduce the dimension to 1 channel to generate a spatial attention map X s , as shown in Equation (8), where W and b represent MLP weight and MLP biase respectively. C a t (.) means concatenate. C o n v (.) means convolution operation.
X s = C o n v ( C a t ( F A v g S ; F Max ) W + b , X s R 1 × H × W )
F A v g S ( x , y ) = 1 C i = 1 c X i , x , y , F A v g S R 1 × H × W
F M a x S ( x , y ) = Max i = 0 X i , x , y , F M a x S R 1 × H × W
Then generate spatial attention feature A S through sigmoid, as shown in Equation (11), and finally multiply the feature and the input feature of the module to obtain the final generated feature.
A S = m H n w 1 1 + exp X s m , n , A s R 1 × H × W
where m, n represent mth position and nth position respectively. ⋃ represents the contact element by element.

3.2. Densely Connected Convolution Module

The densely connected convolution module [51] is composed of multiple dense blocks, and each dense block performs two convolution operations. The structure of densely connected convolution is shown in Figure 6. The distinctive feature of the densely connected convolution module is that the input of each block is the concatenation of all feature maps generated by all previous blocks, and each layer performs a series of continuous transformations. The idea of densely connected convolution has some advantages over conventional convolution. First, it helps the network learn diversified functional characteristics, rather than redundant functions. In addition, this idea allows information to flow through the network and reuse functions to increase the representativeness of the network. Dense connectivity ensures the maximum information path between layers by connecting all layers. The output features of all convolutional layers in the dense block are connected in series along the channel axis. Let X i 1 and X i be the input and output of the i t h tightly connected dense block, respectively. The output X i of the i t h dense block F i can be obtained in the following manner.
X i = F i X i 1 , X i 2 , , X 0
where [.] represents the element series along the channel axis, and the input X 0 is the initial element f 1 . Because the input of the convolutional layer is repeatedly cascaded, the number of channels used for the input of the next convolutional layer increases with the growth rate n.

4. Results and Discussion

4.1. Data Set and Processing

4.1.1. Data Set

The data used in this article were from the ISIC2017 [52] challenge data set released by the International Skin Imaging Collaboration (ISIC) and the PH2 [53] data set provided by the Pedro Hispano Dermatology Department of the Hospital and the Tecnico Lisboa University of Porto Research Group in Matosinhos, Portugal. The ISIC2017 data set provided 2750 dermoscopic images in RBG format, of which 2000 images were used in the training phase, 600 images were used in the Test Phase, and 150 images were used in the Validation Phase. The PH2 data set contained 200 8-bit RGB dermatoscope images with a fixed size of 768 × 560 pixels. All images provided lesion boundaries given by professional clinicians. Table 1 summarizes the distribution of the two data sets.
Ne, Me and SK stand for Benign Nevus, Melanoma or Seborrheic Keratosis, respectively.

4.1.2. Processing

We combined 150 test data and 2000 training data as training data for the network. In order to train our proposed network structure more conveniently, we scaled the width of images of different sizes to 256 px according to the aspect ratio, and then filled the black borders on the top and bottom of the image to increase the height to 256 px. In order to better allow the network to learn the brightness, tone, and vividness of the dermatoscope image, we converted the image in RGB format to HSV format. As shown in Figure 7, the color components of R, G, and B in the RGB image were all related to the amount of light irradiated to the object. Therefore, the image description based on these components made it difficult to distinguish the object. Unlike RGB, HSV separates brightness or image intensity from chromaticity or color information and is more stable to changes in external lighting. HSV images can detect objects with specific colors and reduce the influence of light intensity from the outside. Therefore, we used HSV images as an effective supplement to training data to train the segmentation network. At the same time, the two formats of images were rotated horizontally, vertically, horizontally and vertically to expand the number of training images, and finally 17,200 training images were generated.

4.2. Experimental Setup

4.2.1. Evaluation Metrics

We used Sensitivity (SEN), Specificity (SPE), Accuracy (ACC), Jaccard (JAC), Dice Coefficient (DIC) and Matthew Correlation Coefficient (MCC) to more accurately judge the segmentation performance of CSAG and DCCNet proposed in this paper. Among them, SEN, SPE, and ACC are common statistical measures used to judge the performance of binary classification. JAC and DIC are used to evaluate the similarity between the segmentation results and ground truth. MCC is the correlation coefficient between the prediction result and ground truth. The above evaluation indicators are directly calculated from the confusion matrix. The calculation method of these six evaluation indicators refers to Equations (13)–(18), where T P represents the correct segmentation of skin lesion pixels, and F N is the wrong segmentation of skin lesion pixels. If the segmentation of non-lesion pixels is correctly classified as non-lesion, it is regarded as T N . Otherwise, they are F P .
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N F P + T N
A c c u r a c y = T P + T N T P + F N + T N + F P
J A C = T P T P + F N + F P
D I C = 2 × T P 2 × T P + F P + F N
M C C = T P · T N F P · F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )

4.2.2. Implementation Details

All experiments in this article were run on Ubuntu 16.04 system, performed on a workstation equipped with Intel(R) Xeon(R) Gold 5218 CPU 2.30 GHz, NVIDIA Quadro RTX 6000 (24 G). Use Python 3.6 and Pytorch 1.0.0 deep learning framework for programming. CSAG and DCCNet used stochastic gradient descent with momentum, the momentum parameter was 0.9, the weight decay coefficient was 5 × 10 4 , the initial learning rate was 0.1, the exponential decay was 10% every 50 epochs, and the mini-batch size was 16. We used the Softmax function for final classification. The radius r of the local square window constructed in the CSFAG module was 4, and the regularization parameter λ was set to 0.001.

4.3. Ablation Analysis

4.3.1. Discussion on the Number of Densely Connected Blocks at the Bottom

In the last step of the CSAG and DCCNet encoding path proposed in this article, densely connected convolution was used to replace the ordinary convolutional layer. However, setting several densely connected blocks could achieve the best segmentation results, which is a problem worth discussing. Therefore, this article used 600 dermoscopy images on the ISIC-2017 data set for testing, and the distribution of the 600 test images is shown in Table 2. We used the control variable method. In the skip connection part of CSAG and DCCNet, the CSFAG module was still embedded, and only the number of dense connection blocks D used in the last step of the encoding path changed. We set D = 1, 2, 3, and 4 respectively. When D = 1, there was no densely connected convolution block in this layer, which was similar to the last coding layer of the standard U-Net.
The specific experimental results are shown in Table 2. When the number of densely connected blocks was set to 3, the best results were achieved on ACC, SPE, DIC, JAC, and MCC. When D = 4, all indicators fell. In subsequent experiments, three densely connected blocks were set.

4.3.2. Structure Ablation

The CSAG and DCCNet proposed in this paper uses U-Net as the basic network structure, and builds an image pyramid input on the left side of the coding layer to achieve a multi-scale layer of multi-level receiving field fusion; on the right side of the decoder, introduces a side output layer, average all side output maps as the final prediction map. This structure also constitutes M-Net [54]. In addition, the use of densely connected convolution and CSFAG module are also key factors for the CSAG and DCCNet model to show excellent performance in skin lesion segmentation. In this regard, we designed a set of ablation experiments, taking the UNet network structure as the baseline model, and adding image pyramid input, side output layers and multi-label loss functions to this basic structure to verify that these methods can improve the segmentation of dermoscopic images. In the last step of the MNet encoding path, densely connected convolutions are used instead of ordinary convolutional layers. We named them M-Net+Dense Convolutions. The CSFAG module is embedded on the MNet skip connection, and we named it M-Net+CSFAG. Experiments have further proved that these two methods are effective. The specific segmentation performance indicators are shown in Table 3 and Table 4. The data in all tables were tested using different models on the 600 test set images in the ISIC2017 data set, calculating the various indicators of each test image, and averaging the various indicators of the 600 test images. The results obtained for Nevus cases, Melanoma cases, SK cases, and overall were all strictly trained for 200 epochs.
From the results in Table 3 and Table 4, we can see that whether it was in Benign Nevus, Melanoma and Seborrheic Keratosis lesions, MNet achieved greater improvement in various indicators compared with UNet, which verified Pyramid input, side output and multi-label loss function methods were helpful to improve the performance of dermoscopy image segmentation. From the results of M-Net+Dense Convolutions and M-Net+CSFAG, compared with MNet, the four key indicators of ACC, DIC, MCC, and JAC all achieved certain improvements. This further proves that these two methods are effective. We applied Dense Convolutions and CSFAG to MNet together to design the CSAG and DCCNet proposed in this article. From the specific segmentation results, especially in the segmentation of melanoma lesions and seborrheic keratosis, very competitive results could be obtained.
As presented in Figure 8, we visually display the segmentation results of UNet, MNet, M-Net+Dense Convolutions, M-Net+CSFAG, and CSAG and DCCNet proposed in this section. For Figure 8f, we zoomed in on the lesion area on the image to visually show the comparison between the five models verified in the structure ablation experiment and the ground truth segmentation results. It can be clearly seen in the figure that, compared with UNet, the segmentation result of MNet had clearer edges and more concentrated recognition of the lesion area. For M-Net+Dense Convolutions, M-Net+CSFAG and ours, compared with MNet, the segmentation results on moles and melanoma lesions had little difference. We believe that for dermoscopic images of moles and melanoma lesions, due to the large difference in the front background and the relatively concentrated lesion area, a simpler model could obtain better segmentation results. However, for the small difference between the front background and the blurred boundary of the lesion area, similar to the dermoscopic image of the last two rows in Figure 8, it was difficult to obtain a more ideal segmentation result using a more basic segmentation network structure such as UNet or MNet. Therefore, CSAG and DCCNet was designed in this paper. From the segmentation results of the last two rows in Figure 8, CSAG and DCCNet could identify larger lesion areas for images with small differences in the front background and blurred boundaries of the lesion area, the segmentation result obtained was closer to the ground truth.

4.3.3. Attention Module Ablation

The Channel Spatial Fast Attention-guided Filter (CSFAG) module designed in this paper is the main reason for the improvement of segmentation performance. In order to verify the effectiveness of the CSFAG module and verify that it had better feature aggregation performance than other attention modules, we designed a set of ablation experiments. We use M-Net+Dense Convolutions in the structural ablation part as the Baseline, and only changed the attention module embedded in the jump connection to verify the advanced nature of the CSFAG module proposed in this article. Three classic, lightweight, and general attention modules were selected for comparison with the CSFAG module. The first was the Squeeze-and-Excitation module in SE-Net [37]. The SE module performs attention or gating operations on the channel dimension to drive the model to pay more attention to the channel features with the most information, while suppressing those unimportant channel features. The second attention module is the Feature Pyramid Attention (FPA) module proposed by Li and Xiong et al. [55]. This module performs attention operations on pixels, adopts the idea of global pooling of PSPnet [31], and adds the result of pooling to the result of convolution with attention. The last attention module is the CBAM module proposed by Sanghyun Woo [41]. This module combines the attention mechanism of spatial and channel, and learned how to effectively emphasize or compress and extract intermediate features, so that the model pays more attention to the target object itself.
Table 5 and Table 6 show the segmentation performance of the model formed by the combination of SE Block, FPA, CBAM and Baseline and CSAG and DCCNet (ours) on three skin lesions. From the perspective of various indicators, the segmentation performance of the network structure with the added attention module was greatly improved compared with the baseline. For CSAG and DCCNet, especially in the three key indicators of DIC, JAC, and MCC, CSAG and DCCNet has achieved good results. We think the reasons why CSAG and DCCNet could obtain higher results are considered as follows. SE Block only studied the channel attention mechanism of feature maps, and ignored the information in the spatial dimension. Although the FPA module used the pyramid structure to expand the range of the receptive field, it lost the pixel-level positioning information, which had a greater impact on the dermoscopic image segmentation. CBAM could learn what to pay attention to and where to pay attention in the channel and spatial dimensions, but it is difficult to integrate the structural information contained in low-resolution feature maps and high-resolution feature maps, and retains less edge information. In addition, for the parameters of each model, SE Block only paid attention to the characteristics at the channel level, making the model parameters smaller. The FPA module uses 5 × 5 and 7 × 7 large convolution kernels, resulting in a large amount of model parameters. CBAM applies attention to both channel and spatial dimensions at the same time. Compared with SE Block, the amount of CBAM module parameters also increased. However, the CSFAG module proposed in this paper used the CBAM module as the attention module and combined the guide filter, so the parameter amount was slightly higher than the CBAM module parameter amount.
For the problems of the above several attention modules, in order to further improve the segmentation performance of the model, we designed the CSFAG module. In Figure 9, we show the segmentation results of the four network structures for ablation experiments in this section. From Figure 9, we found that compared to the other three attention modules, the segmentation results obtained by CSAG and DCCNet (ours) were closer to ground truth. Especially for dermoscopic images with relatively blurry lesions, CSAG and DCCNet (ours) could segment larger lesions and get better segmentation results. In order to compare the segmentation results intuitively, we marked the segmented edge contours with lines of different colors and superimposed them on the original image, and enlarged the lesion area, as shown in Figure 9g. Obviously, the blue line representing CSAG and DCCNet (ours) was closer to the red line representing ground truth than the other three modules, which proves that the CSFAG module could achieve better segmentation performance on the baseline model.
In order to further illustrate the superiority of the CSFAG module, we drew the ROC curve and PR curve on the baseline-based SE block, FPA, CBAM model and CSAG and DCCNet (ours) respectively on overall, Nevus cases, Melanoma cases, and SK cases. The ROC curve gave information between false positive pixels and true positive pixels in the form of scores based on the threshold change on the probability map. When the ratio between the positive sample and the negative sample was large, the PR curve could better reflect the true performance of the classification. The difference from the upper left convex of the ROC curve was that the PR curve had an upper right convex effect. As shown in Figure 10, the areas of our ROC curve and PR curve were larger than other attention modules. This further showed that the CSFAG module had better performance than other attention modules, and better segmentation of background information and skin lesion area information.

4.3.4. Robustness Test of CSFAG Module

In order to verify that the CSFAG module proposed in this paper had good robustness and prove that it can be combined with multiple basic segmentation networks, we designed this set of experiments. We used U-Net [18], SegNet [56] and M-Net [54] three basic segmentation networks as the backbone network, embed the CSFAG module into the skip connection, named U-Net+CSFAG, SegNet+CSFAG and M-Net+CSFAG. Table 7 and Table 8 summarize the comparison of the segmentation performance of the three basic architectures and the new model combined with the CSFAG module and the proposed CSAG and DCCNet (ours). From Table 7 and Table 8, we can see that although the new network formed after adding the CSFAG module had a large increase in the amount of model parameters, the new network was higher than the basic network in six indicators, which showed that the CSFAG module could significantly improve the segmentation performance of the network and is easy to transplant, and has good robustness.
The results of UNet and SegNet in Table 7 and Table 8 were slightly lower than those of Al-Masni et al. [57] in Table 9 and Table 10. We considered the differences caused by different parameter settings and training methods.Al-Masni et al. used the weight parameters of the model trained on VGG-16 on the ImageNet dataset as the initial weight of SegNet, and only fine-tuned the weight of the segmentation network. They used Theano and Keras for programming, select the AdamOptimizer optimization algorithm, set the batch to 20, and used NVIDIA GeForce GTX 1080 (16 G) GPU for training. However, when we reproduced U-Net and SegNet, we set random initial weights, performed on a workstation equipped with Intel(R) Xeon(R) Gold 5218 CPU 2.30 GHz, NVIDIA Quadro RTX 6000 (24 G). We used Python 3.6 and Pytorch 1.0.0 deep learning framework for programming. Using the stochastic gradient descent method with momentum, the momentum parameter was 0.9, and the weight attenuation coefficient was 5 × 10−4.
As shown in Figure 11, in order to visually display the improvement of segmentation performance brought by the CSFAG module, we visually display the segmentation results of U-Net, SegNet, M-Net and the new attention-guided network they generate and CSAG and DCCNet. For (c3), (d3), (e3) in Figure 11, we use the red line to represent the ground truth, and the blue line to represent the segmentation result of ours. We zoomed in on the lesion area on the picture to visualize the comparison between the three models of the CSFAG module and the ground truth segmentation results. (c3) shows the comparison of U-Net and U-Net+CSFAG segmentation results. (d3) shows the comparison of SegNet and SegNet+CSFAG segmentation results. (e3) shows the comparison of M-Net and M-Net+CSFAG segmentation results Contrast. It can be clearly seen in the figure that the three backbone networks were compared with the new attention-guided filter network. The new attention-guided filter network could significantly improve the segmentation performance and make the segmentation boundary more accurate. Especially in the dermoscopic image segmentation with small difference between the front and background, the network after adding the CSFAG module could distinguish larger lesion areas. It can be seen that adding the CSFAG module that we proposed to the network could greatly improve the performance of lesion segmentation. The module could be seamlessly integrated into multiple basic segmentation network architectures, and had good robustness.

4.3.5. Comparing with Existing Technology by Lesion Type

Table 9 and Table 10 compare the CSAG and DCCNet proposed in this paper with the latest methods proposed by Al-Masni et al. [57] and Goyal et al. [58]. From the results shown in Table 9 and Table 10, compared with other models, CSAG and DCCNet achieved higher results in all indicators in the segmentation of three different types of skin lesions, especially in ACC, The four key indicators of DIC, JAC and MCC achieved the very competitive results. This proves that our proposed CSAG and DCCNet could produce more accurate segmentation boundaries, especially in the segmentation of melanoma lesions and seborrheic keratosis images. Therefore, we believe that the CSAG and DCCNet proposed in this paper was an effective supplement to the dermoscopic image segmentation method.
Since there are few papers on calculating segmentation performance according to the types of skin lesions, there is a lack of comparable data. Most studies do not divide the test set according to the type of lesion, but evaluate the segmentation performance of all dermoscopic images in the test set together. In order to further compare with existing dermoscopic image segmentation methods, we have compiled the results in Table 11. From the results shown in Table 11, CSAG and DCCNet achieved high results in the two indicators of ACC and SPE, and also achieved more competitive results in the two key indicators of DIC and JAC. This further proved that the CSAG and DCCNet proposed in this paper was an effective supplement to the existing dermoscopic image segmentation methods.

4.3.6. Results on the PH2 Data Set

We used the independent dermoscopic image data set PH2 to verify the cross-data performance of our proposed network. We used the model trained on the ISIC2017 dataset to test the segmentation performance on the PH2 dataset. As shown in Table 12 and Table 13, in the segmentation of Benign Nevus and melanoma, our proposed model obtained competitive results. This showed that our proposed model had good robustness and cross-data performance. Figure 12 shows the CSAG and DCCNet segmentation results. The first two columns are moles and the last two columns are melanomas.

5. Conclusions

In this article, we propose and implement a novel and robust skin lesion segmentation depth model called CSAG and DCCNet. In the last step of the encoding path, the model uses densely connected convolution instead of ordinary convolutional layers. In order to achieve better information fusion, highlight the foreground and reduce the impact of the background, we designed a novel attention-guided filter module, Channel Spatial Fast Attention-guided Filter (CSFAG for short), and embedded it in the CSAG and DCCNet segmentation network Jumping connection. Secondly, the model uses U-Net as the basic network structure, and builds an image pyramid input layer on the left side of the coding layer; on the right side of the decoder, introduces a side output layer, and averages all side output images as the final prediction image. In order to verify its effect, we evaluated the model using two publicly available datasets (ISIC-2017 Challenge and PH2 dataset). The results show that both densely connected convolution and CSFAG modules can improve the segmentation performance of the network, and the combination of them to form CSAG and DCCNet is better than some of the latest algorithms for skin lesion segmentation. Through a large number of ablation experiments, we have verified that the CSFAG module is superior to other mainstream attention modules and can be combined with multiple basic segmentation networks, effectively improving the segmentation performance of the model. CSAG and DCCNet trained on ISIC-2017 data set was tested on another publicly available data set PH2 data set to verify the robustness and cross-data set performance of our method. In the future, we believe that more interference in dermoscopic images is a key factor affecting segmentation performance, data purification on it is a very effective work. We look forward to the combination of the new dermoscopic image preprocessing method and the proposed model to get a better segmentation model, and apply the model to other medical images to prove the robustness of the method.

Author Contributions

Conceptualization, S.T. and Y.J.; methodology, S.T.; software, S.T.; validation, S.T., S.C. and C.W.; formal analysis, S.T.; resources, Y.J.; data curation, Z.M.; writing—original draft preparation, S.T.; writing—review and editing, S.T. and Y.J.; visualization, S.C. and C.W.; supervision, Z.M.; funding acquisition, Y.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (61962054), in part by the the National Natural Science Foundation of China (61163036), in part by the 2016 Gansu Provincial Science and Technology Plan Funded by the Natural Science Foundation of China (1606RJZA047), in part by the 2012 Gansu Provincial University Fundamental Research Fund for Special Research Funds; Gansu Province Postgraduate Supervisor Program in Colleges and Universities (1201-16), in part by the Northwest Normal University’s Third Phase of Knowledge and Innovation Engineering Research Backbone Project (nwnu-kjcxgc-03-67).

Institutional Review Board Statement

Ethical review and approval are not applicable for thisreview paper.

Informed Consent Statement

An informed consent statement is not applicable.

Data Availability Statement

We used two public dermoscopy datasets to evaluate the proposed segmentation network, namely the ISIC-2017 challenge dataset [52] and the PH2 dataset [53]. The ISIC-2017 challenge data set is provided by the International Skin Imaging Collaboration (ISIC). The PH2 data set was collected by the Pedro Hispano Department of Dermatology in the hospital and the Técnico Lisboa University of Porto research team in Matosinhos, Portugal. The url of ISIC2017 is https://challenge.isic-archive.com/landing/2017, accessed on 16 May 2021.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CSAG and DCCNetChannel Spatial Attention-guided Network with Densely Connected Convolutional
CSFAGChannel Spatial Fast Attention-guided Filter
FCNFull Convolutional Network
DCNNDeep Convolutional Neural Network
SENSensitivity
SPESpecificity
ACCAccuracy
JACJaccard
DICDice Coefficient
MCCMatthew Correlation Coefficient

References

  1. Ge, Z.; Demyanov, S.; Chakravorty, R.; Bowling, A.; Garnavi, R. Skin disease recognition using deep saliency features and multimodal learning of dermoscopy and clinical images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017; pp. 250–258. [Google Scholar]
  2. Barker, C.A.; Postow, M.A. Combinations of radiation therapy and immunotherapy for melanoma: A review of clinical outcomes. Int. J. Radiat. Oncol. Biol. Phys. 2014, 88, 986–997. [Google Scholar] [CrossRef] [Green Version]
  3. Celebi, M.E.; Kingravi, H.A.; Uddin, B.; Iyatomi, H.; Aslandogan, Y.A.; Stoecker, W.V.; Moss, R.H. A methodological approach to the classification of dermoscopy images. Comput. Med. Imaging Graph. 2007, 31, 362–373. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Woltsche, N.; Piana, S.; Ferrara, G.; Hofmann-Wellenhof, R.; Zalaudek, I. Three dermoscopic signs of growth of pigmented lesions. J. Am. Acad. Dermatol. 2015. [Google Scholar] [CrossRef]
  5. Vestergaard, M.E.; Macaskill, P.H.P.M.; Holt, P.E.; Menzies, S.W. Dermoscopy compared with naked eye examination for the diagnosis of primary melanoma: A meta-analysis of studies performed in a clinical setting. Br. J. Dermatol. 2008, 159, 669–676. [Google Scholar] [CrossRef] [PubMed]
  6. Celebi, M.E.; Iyatomi, H.; Schaefer, G.; Stoecker, W.V. Lesion border detection in dermoscopy images. Comput. Med. Imaging Graph. 2009, 33, 148–153. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Oliveira, R.B.; Mercedes Filho, E.; Ma, Z.; Papa, J.P.; Pereira, A.S.; Tavares, J.M.R. Computational methods for the image segmentation of pigmented skin lesions: A review. Comput. Methods Progr. Biomed. 2016, 131, 127–141. [Google Scholar] [CrossRef] [Green Version]
  8. Ahmed, J.; Gao, B.; Lok Woo, W. Sparse low-rank tensor decomposition for metal defect detection using thermographic imaging diagnostics. IEEE Trans. Ind. Inform. 2020, 17, 1810–1820. [Google Scholar] [CrossRef]
  9. Gupta, P.; McClatchey, R.; Caleb-Solly, P. Tracking changes in user activity from unlabelled smart home sensor data using unsupervised learning methods. Neural Comput. Appl. 2020, 32, 12351–12362. [Google Scholar] [CrossRef] [Green Version]
  10. Li, X.; Yu, L.; Chen, H.; Fu, C.W.; Heng, P.A. Semi-supervised skin lesion segmentation via transformation consistent self-ensembling model. arXiv 2018, arXiv:1808.03887. [Google Scholar]
  11. Pathan, S.; Tripathi, A. Y-net: Biomedical Image Segmentation and Clustering. arXiv 2020, arXiv:2004.05698. [Google Scholar]
  12. Feyjie, A.R.; Azad, R.; Pedersoli, M.; Kauffman, C.; Ayed, I.B.; Dolz, J. Semi-supervised few-shot learning for medical image segmentation. arXiv 2020, arXiv:2003.08462. [Google Scholar]
  13. Yuan, Y.; Chao, M.; Lo, Y.C. Automatic skin lesion segmentation using deep fully convolutional networks with jaccard distance. IEEE Trans. Med. Imaging 2017, 36, 1876–1886. [Google Scholar] [CrossRef] [PubMed]
  14. Li, H.; He, X.; Zhou, F.; Yu, Z.; Ni, D.; Chen, S.; Wang, T.; Lei, B. Dense deconvolutional network for skin lesion segmentation. IEEE J. Biomed. Health Inform. 2018, 23, 527–537. [Google Scholar] [CrossRef] [PubMed]
  15. Yuan, Y.; Lo, Y.C. Improving dermoscopic image segmentation with enhanced convolutional-deconvolutional networks. IEEE J. Biomed. Health Inform. 2017, 23, 519–526. [Google Scholar] [CrossRef] [Green Version]
  16. Bi, L.; Kim, J.; Ahn, E.; Kumar, A.; Fulham, M.; Feng, D. Dermoscopic image segmentation via multistage fully convolutional networks. IEEE Trans. Biomed. Eng. 2017, 64, 2065–2074. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  19. Lin, G.; Milan, A.; Shen, C.; Reid, I. Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1925–1934. [Google Scholar]
  20. Lin, G.; Shen, C.; Van Den Hengel, A.; Reid, I. Efficient piecewise training of deep structured models for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3194–3203. [Google Scholar]
  21. Lin, Z.; Feng, M.; Santos, C.N.D.; Yu, M.; Xiang, B.; Zhou, B.; Bengio, Y. A structured self-attentive sentence embedding. arXiv 2017, arXiv:1703.03130. [Google Scholar]
  22. Shen, T.; Zhou, T.; Long, G.; Jiang, J.; Pan, S.; Zhang, C. Disan: Directional self-attention network for rnn/cnn-free language understanding. arXiv 2017, arXiv:1709.04696. [Google Scholar]
  23. He, K.; Sun, J. Fast guided filter. arXiv 2015, arXiv:1505.00996. [Google Scholar]
  24. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  25. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  26. Ibtehaz, N.; Rahman, M.S. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 2020, 121, 74–87. [Google Scholar] [CrossRef]
  27. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Wu, J. UNet 3+: A Full-Scale Connected UNet for Medical Image Segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  28. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  29. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  30. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  31. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  32. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  33. Corbetta, M.; Shulman, G.L. Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 2002, 3, 201. [Google Scholar] [CrossRef]
  34. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. arXiv 2014, arXiv:1406.6247. [Google Scholar]
  35. Ba, J.; Mnih, V.; Kavukcuoglu, K. Multiple object recognition with visual attention. arXiv 2014, arXiv:1412.7755. [Google Scholar]
  36. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  38. Zhao, H.; Zhang, Y.; Liu, S.; Shi, J.; Loy, C.C.; Lin, D.; Jia, J. Psanet: Point-wise spatial attention network for scene parsing. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 267–283. [Google Scholar]
  39. Chen, Y.; Kalantidis, Y.; Li, J.; Yan, S.; Feng, J. A2-Nets: Double Attention Networks. arXiv 2018, arXiv:1810.11579. [Google Scholar]
  40. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  41. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  42. Wu, H.; Pan, J.; Li, Z.; Wen, Z.; Qin, J. Automated Skin Lesion Segmentation via an Adaptive Dual Attention Module. IEEE Trans. Med. Imaging 2020, 40, 357–370. [Google Scholar] [CrossRef] [PubMed]
  43. Hamad, R.A.; Kimura, M.; Yang, L.; Woo, W.L.; Wei, B. Dilated Causal Convolution with Multi-Head Self Attention for Sensor Human Activity Recognition. Neural Comput. Appl. 2021, 1–18. Available online: https://link.springer.com/article/10.1007/s00521-021-06007-5 (accessed on 16 May 2021).
  44. Hu, B.; Gao, B.; Woo, W.L.; Ruan, L.; Jin, J.; Yang, Y.; Yu, Y. A Lightweight Spatial and Temporal Multi-Feature Fusion Network for Defect Detection. IEEE Trans. Image Process. 2020, 30, 472–486. [Google Scholar] [CrossRef]
  45. Kawahara, J.; Hamarneh, G. Multi-resolution-tract CNN with hybrid pretrained and skin-lesion trained layers. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Athens, Greece, 17 October 2016; Springer: Cham, Switzerland, 2016; pp. 164–171. [Google Scholar]
  46. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  47. Tschandl, P.; Sinz, C.; Kittler, H. Domain-specific classification-pretrained fully convolutional network encoders for skin lesion segmentation. Comput. Biol. Med. 2019, 104, 111–116. [Google Scholar] [CrossRef]
  48. Goyal, M.; Ng, J.; Yap, M.H. Multi-class lesion diagnosis with pixel-wise classification network. arXiv 2018, arXiv:1807.09227. [Google Scholar]
  49. Hasan, M.K.; Dahal, L.; Samarakoon, P.N.; Tushar, F.I.; Martí, R. DSNet: Automatic dermoscopic skin lesion segmentation. Comput. Biol. Med. 2020, 103738. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  51. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  52. Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Halpern, A. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
  53. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH 2-A dermoscopic image database for research and benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar]
  54. Fu, H.; Cheng, J.; Xu, Y.; Wong, D.W.K.; Liu, J.; Cao, X. Joint optic disc and cup segmentation based on multi-label deep network and polar transformation. IEEE Trans. Med. Imaging 2018, 37, 1597–1605. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180. [Google Scholar]
  56. Badrinarayanan, V.; Handa, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for robust semantic pixel-wise labelling. arXiv 2015, arXiv:1505.07293. [Google Scholar]
  57. Al-Masni, M.A.; Al-Antari, M.A.; Choi, M.T.; Han, S.M.; Kim, T.S. Skin lesion segmentation in dermoscopy images via deep full resolution convolutional networks. Comput. Methods Programs Biomed. 2018, 162, 221–231. [Google Scholar] [CrossRef] [PubMed]
  58. Goyal, M.; Oakley, A.; Bansal, P.; Dancey, D.; Yap, M.H. Skin Lesion Segmentation in Dermoscopic Images with Ensemble Deep Learning Methods. IEEE Access 2019. [Google Scholar] [CrossRef]
  59. Wen, H. II-FCN for skin lesion analysis towards melanoma detection. arXiv 2017, arXiv:1702.08699. [Google Scholar]
  60. Sarker, M.M.K.; Rashwan, H.A.; Akram, F.; Banu, S.F.; Saleh, A.; Singh, V.K. SLSDeep: Skin lesion segmentation based on dilated residual and pyramid pooling networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Cham, Switzerland, 2018; pp. 21–29. [Google Scholar]
  61. Jahanifar, M.; Tajeddin, N.Z.; Asl, B.M.; Gooya, A. Supervised saliency map driven segmentation of lesions in dermoscopic images. IEEE J. Biomed. Health Inform. 2018, 23, 509–518. [Google Scholar] [CrossRef] [Green Version]
  62. Bi, L.; Kim, J.; Ahn, E.; Kumar, A.; Feng, D.; Fulham, M. Step-wise integration of deep class-specific learning for dermoscopic image segmentation. Pattern Recognition 2019, 85, 78–89. [Google Scholar] [CrossRef] [Green Version]
  63. Wei, Z.; Song, H.; Chen, L.; Li, Q.; Han, G. Attention-based DenseUnet network with adversarial training for skin lesion segmentation. IEEE Access 2019, 7, 136616–136629. [Google Scholar] [CrossRef]
  64. Lei, B.; Xia, Z.; Jiang, F.; Jiang, X.; Ge, Z.; Xu, Y.; Wang, S. Skin lesion segmentation via generative adversarial networks with dual discriminators. Med. Image Anal. 2020, 64, 101716. [Google Scholar] [CrossRef] [PubMed]
  65. Jiang, Y.; Cao, S.; Tao, S.; Zhang, H. Skin Lesion Segmentation Based on Multi-Scale Attention Convolutional Neural Network. IEEE Access 2020, 8, 122811–122825. [Google Scholar] [CrossRef]
  66. Xie, F.; Yang, J.; Liu, J.; Jiang, Z.; Zheng, Y.; Wang, Y. Skin lesion segmentation using high-resolution convolutional neural network. Comput. Methods Programs Biomed. 2020, 186, 105241. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A typical pictorial presentation of the skin lesion images with different challenging for segmentation.
Figure 1. A typical pictorial presentation of the skin lesion images with different challenging for segmentation.
Sensors 21 03462 g001
Figure 2. CSAG and DCCNet model architecture.
Figure 2. CSAG and DCCNet model architecture.
Sensors 21 03462 g002
Figure 3. CSFAG model architecture.
Figure 3. CSFAG model architecture.
Sensors 21 03462 g003
Figure 4. Visual display of CSFAG module input and output (A dermoscopy image of melanoma is a case, the red line is ground truth).
Figure 4. Visual display of CSFAG module input and output (A dermoscopy image of melanoma is a case, the red line is ground truth).
Sensors 21 03462 g004
Figure 5. Attention module; (a) channel attention submodule; (b) spatial attention submodule.
Figure 5. Attention module; (a) channel attention submodule; (b) spatial attention submodule.
Sensors 21 03462 g005
Figure 6. Densely connected convolution module.
Figure 6. Densely connected convolution module.
Sensors 21 03462 g006
Figure 7. Visual description of the input dermoscopy image in different color spaces.
Figure 7. Visual description of the input dermoscopy image in different color spaces.
Sensors 21 03462 g007
Figure 8. Visualized results of structural ablation experiments. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original image; (b) the ground truth; (c) U-Net; (d) 1. M-Net; (d) 2. M-Net+Dense Convolutions; (d) 3. M-Net+ CSFAG; (e) CSAG and DCCNet model; (f) ground truth (red) and U-Net+ (purple), MNet (green), M-Net+Dense Convolutions (yellow), M-Net+CSFAG (black) and ours (blue) segmentation results comparison chart. All pictures are preprocessed.
Figure 8. Visualized results of structural ablation experiments. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original image; (b) the ground truth; (c) U-Net; (d) 1. M-Net; (d) 2. M-Net+Dense Convolutions; (d) 3. M-Net+ CSFAG; (e) CSAG and DCCNet model; (f) ground truth (red) and U-Net+ (purple), MNet (green), M-Net+Dense Convolutions (yellow), M-Net+CSFAG (black) and ours (blue) segmentation results comparison chart. All pictures are preprocessed.
Sensors 21 03462 g008
Figure 9. Visualization results of the attention module ablation. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original skin lesion image; (b) Ground truth corresponding to the lesion; (c) SE Block+Baseline model lesion segmentation results; (d) FPA+Baseline model lesion segmentation results; (e) CBAM+Baseline model lesion segmentation Results; (f) CSAG and DCCNet (ours) model lesion segmentation results; (g) ground truth (red) and SEnet+Baseline (purple), FPA+Baseline (green), CBAM+Baseline (yellow) and ours (blue) segmentation Results comparison chart. All pictures are preprocessed.
Figure 9. Visualization results of the attention module ablation. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original skin lesion image; (b) Ground truth corresponding to the lesion; (c) SE Block+Baseline model lesion segmentation results; (d) FPA+Baseline model lesion segmentation results; (e) CBAM+Baseline model lesion segmentation Results; (f) CSAG and DCCNet (ours) model lesion segmentation results; (g) ground truth (red) and SEnet+Baseline (purple), FPA+Baseline (green), CBAM+Baseline (yellow) and ours (blue) segmentation Results comparison chart. All pictures are preprocessed.
Sensors 21 03462 g009
Figure 10. Visualized results of ROC curve and PR curve.
Figure 10. Visualized results of ROC curve and PR curve.
Sensors 21 03462 g010
Figure 11. Visualization results of the robustness test of the CSAG module on the ISIC 2017 data set. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original image; (b) the ground truth; (c) 1. U-Net; (c) 2. U-Net+CSFAG; (c) 3. Comparison of segmentation results ground truth (red), U-Net (purple), U-Net+CSFAG (green) and ours (blue); (d) 1. SegNet; (d) 2. SegNet+CSFAG; (d) 3. Comparison of segmentation results of ground truth (red) and SegNet (white), SegNet+CSFAG (black) and ours (blue) (e) 1. M-Net; (e) 2. M-Net+CSFAG; (e) 3. Comparison of segmentation results of ground truth (red) and M-Net (yellow), M-Net+CSFAG (pink) and ours (blue) (f) CSAG and DCCNet (ours) model. All pictures are preprocessed.
Figure 11. Visualization results of the robustness test of the CSAG module on the ISIC 2017 data set. The first two lines are Benign Nevus, the middle two lines are Melanoma lesions, and the last two lines are Seborrheic Keratosis lesions. (a) Original image; (b) the ground truth; (c) 1. U-Net; (c) 2. U-Net+CSFAG; (c) 3. Comparison of segmentation results ground truth (red), U-Net (purple), U-Net+CSFAG (green) and ours (blue); (d) 1. SegNet; (d) 2. SegNet+CSFAG; (d) 3. Comparison of segmentation results of ground truth (red) and SegNet (white), SegNet+CSFAG (black) and ours (blue) (e) 1. M-Net; (e) 2. M-Net+CSFAG; (e) 3. Comparison of segmentation results of ground truth (red) and M-Net (yellow), M-Net+CSFAG (pink) and ours (blue) (f) CSAG and DCCNet (ours) model. All pictures are preprocessed.
Sensors 21 03462 g011
Figure 12. Visualization results for the PH2 data set. (a) Original skin lesion image; (b) ground truth corresponding to the lesion. (c) CSARM-CNN model lesion segmentation results; (d) ground truth (red) and segmentation results (blue) comparison chart.
Figure 12. Visualization results for the PH2 data set. (a) Original skin lesion image; (b) ground truth corresponding to the lesion. (c) CSARM-CNN model lesion segmentation results; (d) ground truth (red) and segmentation results (blue) comparison chart.
Sensors 21 03462 g012
Table 1. The distribution of The ISIC Challenge 2017 data set and PH2 datasets.
Table 1. The distribution of The ISIC Challenge 2017 data set and PH2 datasets.
DatasetISIC 2017PH2
MeSKNeTotalMeSKNeTotal
Training data40429614502150----
Test data1179039360040-160200
Table 2. Determination of the numberD of densely connected blocks on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
Table 2. Determination of the numberD of densely connected blocks on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
MethodOverall
ACCSENSPEDICJACMCC
D = 195.0487.5896.4084.9675.3081.49
D = 295.2187.5196.9985.9077.9182.51
D = 395.9487.0399.3586.9778.8583.93
D = 495.6186.5097.4686.3577.8983.50
Table 3. ISIC-2017 data set, performance comparison of U-Net, M-Net, M-Net+DC (Dense Convolutions), M-Net+CSFAG and CSAG and DCCNet (ours) in sensitivity, specificity and accuracy. Bold data indicates that the value is the maximum value in this indicator.
Table 3. ISIC-2017 data set, performance comparison of U-Net, M-Net, M-Net+DC (Dense Convolutions), M-Net+CSFAG and CSAG and DCCNet (ours) in sensitivity, specificity and accuracy. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
SENSPEACCSENSPEACCSENSPEACCSENSPEACC
U-Net8.6M81.8097.5393.8370.3198.0889.6257.8498.3489.0774.5397.5992.32
M-Net10.92M85.9697.3196.0578.6097.5892.1772.6098.3291.6983.2298.3294.58
M-Net+DC13.79M87.6397.9896.1379.6297.9793.3468.8099.6092.8984.4699.1595.10
M-Net+CSFAG24.89M89.9297.6396.8279.4097.2093.2469.9498.0293.3986.5497.4195.85
Ours28.74M89.8599.4397.4680.2399.5194.7080.2499.7094.9987.0399.3595.94
Table 4. On the ISIC-2017 data set, performance comparison of U-Net, M-Net, M-Net+DC (Dense Convolutions), M-Net+CSFAG and CSAG and DCCNet (ours) in Jaccard coefficient, Dice coefficient and Matthew correlation coefficient. Bold data indicates that the value is the maximum value in this indicator.
Table 4. On the ISIC-2017 data set, performance comparison of U-Net, M-Net, M-Net+DC (Dense Convolutions), M-Net+CSFAG and CSAG and DCCNet (ours) in Jaccard coefficient, Dice coefficient and Matthew correlation coefficient. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
DICJACMCCDICJACMCCDICJACMCCDICJACMCC
U-Net8.6M78.7164.8975.1273.1657.6866.7364.3947.4958.5574.3259.1369.64
M-Net10.92M86.8977.3484.8081.7269.2078.3674.3259.1371.5683.6874.0881.30
M-Net+DC13.79M87.1478.1685.5683.3172.1377.1178.1962.8073.8584.7476.6581.78
M-Net+CSFAG24.89M88.4679.3186.6883.7072.7179.8379.1665.5175.2085.6376.8683.20
Ours28.74M90.6582.9089.1984.2473.6878.1082.1472.7877.5686.9778.8583.93
Table 5. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
Table 5. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
SENSPEACCSENSPEACCSENSPEACCSENSPEACC
SE Block+Baseline26.62M90.2597.0596.8978.7798.6893.4178.2398.0494.0185.9197.9795.44
FPA+Baseline32.08M88.2698.6796.2176.1298.2293.5777.2098.1293.4582.2998.1595.35
CBAM+Baseline27.40M89.0198.2597.0676.0597.7593.5981.6996.6894.1682.7697.7095.99
Baseline13.79M87.6397.9896.1379.6297.9793.3468.8099.6092.8984.4699.1595.10
Ours28.74M89.8599.4397.4680.2399.5194.7080.2499.7094.9987.0399.3595.94
Table 6. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
Table 6. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
DICJACMCCDICJACMCCDICJACMCCDICJACMCC
SE Block+Baseline26.62M87.6178.4786.1384.1573.5278.7981.7672.1080.9185.8978.4882.63
FPA+Baseline32.08M87.3677.8985.9083.9973.6477.2081.3670.3476.8385.0977.0582.89
CBAM+Baseline27.40M88.5579.4586.7484.9272.9177.9082.4771.1778.9885.1677.1582.38
Baseline13.79M87.1478.1685.5683.3172.1377.1178.1962.8073.8584.7476.6581.78
Ours28.74M90.6582.9089.1984.2473.6878.1082.1472.7877.5686.9778.8583.93
Table 7. The performance comparison of sensitivity, specificity and accuracy of the three basic architecture networks and their new network structure with the CSAG and DCCNet (ours) model on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
Table 7. The performance comparison of sensitivity, specificity and accuracy of the three basic architecture networks and their new network structure with the CSAG and DCCNet (ours) model on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
SENSPEACCSENSPEACCSENSPEACCSENSPEACC
UNet8.60M81.8097.5393.8370.3198.0889.6257.8498.3489.0774.5397.5992.32
UNet+CSFAG23.15M87.3599.3196.0073.8997.6890.1682.9898.8791.4985.6798.4594.34
SegNet28.08M82.8398.6595.6073.2499.0291.6965.6298.9892.6480.1999.0095.46
SegNet+CSFAG42.78M85.5198.9996.3174.4799.1493.2982.6998.1091.8282.4198.9995.34
M-Net10.92M85.9697.3196.0578.6097.5892.1772.6098.3291.6983.2298.3294.58
M-Net+CSFAG24.89M88.8297.6396.8279.4097.2093.2469.9498.0293.3986.5497.4195.85
Ours28.74M89.8599.4397.4680.2399.5194.7080.2499.7094.9987.0399.3595.94
Table 8. The performance comparison of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of the three basic architecture networks and their new network structure with the CSAG and DCCNet (ours) model on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
Table 8. The performance comparison of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of the three basic architecture networks and their new network structure with the CSAG and DCCNet (ours) model on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
MethodParamsNevus CasesMelanoma CasesSK CasesOverall
DICJACMCCDICJACMCCDICJACMCCDICJACMCC
UNet8.60M78.7164.8975.1273.1657.6866.7364.3947.4958.5574.3259.1369.64
UNet+CSFAG23.15M86.2875.8783.8876.7562.2771.3072.4656.8267.1481.0868.1877.77
SegNet28.08M83.6671.9181.2877.8463.7373.2873.6758.3170.1583.0771.0580.70
SegNet+CSFAG42.78M86.4376.1084.4182.0469.5578.5875.3560.4570.8085.0674.0082.36
M-Net10.92M86.8977.3484.8081.7269.2078.3674.3259.1371.5683.6874.0881.30
M-Net+CSFAG24.89M88.4679.3186.6883.7072.7179.8379.1665.5175.2085.6376.8683.20
Ours28.74M90.6582.9089.1984.2473.6878.1082.1472.7877.5686.9778.8583.93
Table 9. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
Table 9. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
MethodNevus CasesMelanoma CasesSK CasesOverall
SENSPEACCSENSPEACCSENSPEACCSENSPEACC
U-Net [57]76.7697.2692.8958.7196.8184.9843.8197.6484.8367.1597.2490.14
FCN-AlexNet [58]82.4497.5894.8472.3596.2387.8271.7097.9289.3578.8697.3792.65
FCN-32s [58]83.6796.6994.5974.3696.3288.9475.8096.4189.4580.6796.7292.72
FCN-16s [58]84.2396.9194.6775.1496.2789.9475.4896.2588.8381.1496.6892.74
FCN-8s [58]83.9197.2294.5578.3795.9689.6369.8596.5787.4080.7296.8792.52
DeepLabV3+ [58]88.5497.2195.6777.7196.3789.6574.5998.5590.0684.3497.2593.66
Mask-RCNN [58]87.2596.3895.3278.6395.6389.3182.4194.8890.8584.8496.0193.48
SegNet [57]85.1996.3093.9373.7894.2687.9070.5892.5087.2980.0595.3791.76
FrCN [57]88.9597.4495.6278.9196.0490.7882.3794.0891.2985.4096.6994.03
Ensemble-A [58]92.0895.3795.5984.6294.2090.8587.4994.4191.7289.9395.0094.04
Ours89.8599.4397.4680.2399.5194.7080.2499.7094.9987.0399.3595.94
Table 10. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different modules and CSFAG modules on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
Table 10. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different modules and CSFAG modules on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
MethodNevus CasesMelanoma CasesSK CasesOverall
DICJACMCCDICJACMCCDICJACMCCDICJACMCC
U-Net [57]82.1669.7178.0570.8254.8363.7157.8840.7363.8976.2761.6471.23
FCN-AlexNet [58]85.6177.0182.9175.9464.3270.3575.0963.7671.5182.1572.5578.75
FCN-32s [58]85.0876.3982.2978.3967.2372.7076.1864.7872.1082.4472.8678.89
FCN-16s [58]85.6077.3982.9279.2268.4173.2675.2364.1171.4282.8073.6579.31
FCN-8s [58]84.3376.0781.7380.0869.5874.3968.0156.5465.1481.0671.8777.81
DeepLabV3+ [58]88.2981.0985.9080.8671.3076.0177.0567.5574.6285.1677.1582.28
Mask-RCNN [58]88.8380.9185.3880.2870.6974.9580.4870.7476.3185.5877.3981.99
SegNet [57]85.6974.9781.8479.1165.4571.0372.5456.9164.3282.0969.6376.79
FrCN [57]89.6881.2886.9084.0272.4477.9081.8369.2576.1187.0877.1183.22
Ensemble-A [58]89.2882.1186.3383.5474.5378.0882.5373.4578.6187.1479.3483.57
Ours90.6582.9089.1984.2473.6878.1082.1472.7877.5686.9778.8583.93
Table 11. Performance evaluation of accuracy, Dice coefficient, Jaccard coefficient, sensitivity, specificity, and performance of different modules and CSFAG modules on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
Table 11. Performance evaluation of accuracy, Dice coefficient, Jaccard coefficient, sensitivity, specificity, and performance of different modules and CSFAG modules on the ISIC-2017 data set. Bold data indicates that the value is the maximum value in this indicator.
ModelYearACCDICJACSENSPE
CDNN [59]20170.9340.8490.7650.8250.975
SLSDeep [60]20180.9360.8780.7820.8160.983
Jahanifar M et al. [61]20180.9300.8390.7490.8100.981
Bi L et al. [62]20190.8620.8570.7770.9670.941
Att-DenseUnet [63]20190.93290.87860.80350.87340.9314
DAGAN [64]20200.9350.8590.7710.8350.976
CSARM-CNN [65]20200.9580.8460.7330.8020.994
DSC [66]20200.9380.8620.7830.8700.964
Ours 0.9590.8690.7880.8700.993
Table 12. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
Table 12. Performance evaluation of sensitivity, specificity, and accuracy performance of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
MethodNevus CasesMelanoma CasesOverall
SENSPEACCSENSPEACCSENSPEACC
FCN [47]95.3594.0994.4490.3094.0292.8290.3094.0292.82
SegNet [47]91.5796.5795.1975.5096.8386.0486.5396.6193.36
U-Net [47]86.6897.6394.6070.5898.4784.3681.6397.7692.55
FrCn [47]94.4895.4695.2091.5796.5594.6493.7295.6595.08
Ours93.3098.1997.3585.3495.7291.0089.7597.7395.90
Table 13. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
Table 13. Performance evaluation of Jaccard coefficient, Dice coefficient and Matthew correlation coefficient of different attention modules and CSFAG modules on the ISIC-2017 data set using baseline as the basic structure. Bold data indicates that the value is the maximum value in this indicator.
MethodNevus CasesMelanoma CasesOverall
DICJACMCCDICJACMCCDICJACMCC
FCN [47]90.4682.5986.7889.0380.2283.7189.0380.2283.71
SegNet [47]91.3284.0387.9984.5573.2373.8989.3680.7784.64
U-Net [47]89.8881.6386.3282.0469.5571.7387.6177.9582.78
FrCn [47]91.3884.1388.1592.9286.7788.6291.7784.7988.30
Ours92.3785.8390.7889.6183.8785.0290.9783.4488.34
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tao, S.; Jiang, Y.; Cao, S.; Wu, C.; Ma, Z. Attention-Guided Network with Densely Connected Convolution for Skin Lesion Segmentation. Sensors 2021, 21, 3462. https://doi.org/10.3390/s21103462

AMA Style

Tao S, Jiang Y, Cao S, Wu C, Ma Z. Attention-Guided Network with Densely Connected Convolution for Skin Lesion Segmentation. Sensors. 2021; 21(10):3462. https://doi.org/10.3390/s21103462

Chicago/Turabian Style

Tao, Shengxin, Yun Jiang, Simin Cao, Chao Wu, and Zeqi Ma. 2021. "Attention-Guided Network with Densely Connected Convolution for Skin Lesion Segmentation" Sensors 21, no. 10: 3462. https://doi.org/10.3390/s21103462

APA Style

Tao, S., Jiang, Y., Cao, S., Wu, C., & Ma, Z. (2021). Attention-Guided Network with Densely Connected Convolution for Skin Lesion Segmentation. Sensors, 21(10), 3462. https://doi.org/10.3390/s21103462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop