Next Article in Journal
Oncofetal Chondroitin Sulfate Is a Highly Expressed Therapeutic Target in Non-Small Cell Lung Cancer
Next Article in Special Issue
Reinforcement Learning for Precision Oncology
Previous Article in Journal
Targeting Bone Metastasis in Cancers
Previous Article in Special Issue
Automated PD-L1 Scoring Using Artificial Intelligence in Head and Neck Squamous Cell Carcinoma
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis

1
Department of Electrical Engineering and Electronics, University of Liverpool, Liverpool L69 3GJ, UK
2
Department of Computer Engineering, University of Diyala, Baqubah 32010, Iraq
*
Author to whom correspondence should be addressed.
Cancers 2021, 13(17), 4494; https://doi.org/10.3390/cancers13174494
Submission received: 3 August 2021 / Revised: 2 September 2021 / Accepted: 4 September 2021 / Published: 6 September 2021
(This article belongs to the Collection Artificial Intelligence in Oncology)

Abstract

:

Simple Summary

Melanoma is the most dangerous type of skin cancer. It grows quickly and has the ability to spread to any organ. This study aims to evaluate and benchmark deep learning models for automatic melanoma diagnosis considering nineteen convolutional neural networks and ten criteria. Multi-Criteria Decision Making methods (MCDMs) are exploited to conduct the benchmarking and subsequently selecting the optimal model considering the predefined criteria. The study findings would help in the model selection, designing quick and reliable diagnostic tools based on image data, and contributing to the development of more accurate and efficient point-of-care diagnostic and detection systems.

Abstract

Melanoma is the most invasive skin cancer with the highest risk of death. While it is a serious skin cancer, it is highly curable if detected early. Melanoma diagnosis is difficult, even for experienced dermatologists, due to the wide range of morphologies in skin lesions. Given the rapid development of deep learning algorithms for melanoma diagnosis, it is crucial to validate and benchmark these models, which is the main challenge of this work. This research presents a new benchmarking and selection approach based on the multi-criteria analysis method (MCDM), which integrates entropy and the preference ranking organization method for enrichment of evaluations (PROMETHEE) methods. The experimental study is carried out in four phases. Firstly, 19 convolution neural networks (CNNs) are trained and evaluated on a public dataset of 991 dermoscopic images. Secondly, to obtain the decision matrix, 10 criteria, including accuracy, classification error, precision, sensitivity, specificity, F1-score, false-positive rate, false-negative rate, Matthews correlation coefficient (MCC), and the number of parameters are established. Third, entropy and PROMETHEE methods are integrated to determine the weights of criteria and rank the models. Fourth, the proposed benchmarking framework is validated using the VIKOR method. The obtained results reveal that the ResNet101 model is selected as the optimal diagnosis model for melanoma in our case study data. Thus, the presented benchmarking framework is proven to be useful at exposing the optimal melanoma diagnosis model targeting to ease the selection process of the proper convolutional neural network architecture.

1. Introduction

Skin cancer is the most frequent type of cancer and can be highly truculent [1]. In the UK, more than 100,000 new cases of skin cancer are reported each year [2]. In 2016, 1319 death cases from non-melanoma skin cancer and 2285 death cases from melanoma skin cancer were reported [3,4]. The initial stage in melanoma diagnosing is usually a visual assessment of the skin lesions. In comparison to inspection with the naked eye, dermatoscopy is one of the dermatologists’ most popular imaging procedures, and a frequently used diagnostic tool that enhances and improves the diagnosis of malignant and benign pigmented skin lesions [5].
A dermoscopy magnifies the surface of the skin lesion, allowing better visualization of deeper skin structures. It provides improved diagnostic accuracy of skin lesions, enabling the dermatologist to examine them more thoroughly. There are two main dermoscopy modes: non-polarized dermoscopy (NPD) and polarized dermoscopy (PD). Non-polarized dermoscopy (NPD) is integrated with a magnification lens and light-emitting diodes to provide illumination, enabling the visualization of subsurface structures in the epidermis. Non-polarized dermoscopy (NPD) requires direct contact of the glass plate with the skin surface and the presence of a liquid interface, such as alcohol, liquid paraffin, water, or ultrasound gel. The interface fluid dramatically increases the penetration of light, reduces scattered radiation, and produces a clear, low-reflection image, which allows excellent visualization of the superficial layers of the skin from the epidermis to the dermal–epidermal junction (DEJ). Like NPD, polarized dermoscopy (PD) contains light-emitting diodes to provide illumination and are equipped with a magnification lens. However, PDs use two polarized filters to achieve cross-polarization. NPD does not require direct contact with the skin and does not require the use of immersion liquids. PD allows visualization of subsurface structures located at the dermal–epidermal junction (DEJ) or superficial dermis. PD nearly blinds to the skin’s surface and structures in the superficial epidermis. Hence, non-polarized dermoscopy reveals superficial features, while polarized dermoscopy shows deeper structures, inferring that the use of both methods can provide complementary information [6,7]. Melanoma is diagnosed in two ways: visual inspection and biopsy. ABCDE (asymmetric, shape, border, color, diameter, and evolution) [8] are the main criteria used for visual screening of melanoma lesions based on a geometric description. Because the ABCDE approach is entirely dependent on the practitioner’s visual acuity and experience, this approach can be performed efficiently only by trained dermatologists [9].
However, manual review by dermatologists is a time-consuming, controversial, and error-prone task. The number of required dermatologists comparing the size of the population in the United States, Australia, and the UK is considerably low [10,11,12]. In the USA, the required number of dermatologists should be more than 4 per 100,000 individuals, which is the number that is suggested to adequately care for a population. However, it is currently estimated at 3.4 per 100,000 individuals. Similarly, there are just 550 practicing dermatologists in Australia, which is almost 15 % less than what is required to meet the needs of the population [11]. In the UK, the Royal College of Physicians (RCP) [13] recommends one full-time equivalent (FTE) consultant per 62,500 of the population. The RCP recommends 989 FTE consultant dermatologists. The British Association of Dermatologists (BAD) [14] found that there are 813 dermatology specialists in the UK Compared to the RCP’s recommendations, the BAD show a shortfall in the region of 250 consultants [12]. Hence, melanoma patients may not be aware of the severity of their disease if they do not undergo inspection by skilled specialists during an early stage of the disease and, thus, miss the ideal time to treat their conditions.
These obstacles encourage and inspire researchers to create automated melanoma diagnosis methods, using computer-aided diagnosis (CAD) systems. For non-experienced dermatologists, the CAD tool could provide a user-friendly environment, used as a second opinion in melanoma cancer diagnosis [15,16]. A large volume of skin images were collected in recent years, and sophisticated deep learning-based models [17] were successfully trained to perform automatic analysis of these skin images due to the industrial advancement of both computer hardware represented by graphics card capabilities and software technologies. These breakthroughs prompted expectations that automated diagnostic tools which will be available in the near future to examine and diagnose all types of skin lesions without the requirement of human experience [18].
Many automated melanoma diagnosis systems based on deep learning techniques, especially deep convolutional neural networks (CNNs), were recently developed. The new methods have significantly advanced the state of the art in skin lesion analysis. The CNN can automatically extract and learn high-level features, increasing the robustness of melanoma images’ inter- and intra-class variability [19,20]. With the rapid increase in the number of automatic recognition of melanoma from dermoscopy images using CNNs, comparing results among pieces of works and evaluation has become an awkward task. This limitation is due to methodological constraints and the absence of some of the standard metrics used to evaluate the performance of the models in terms of sensitivity, specificity, specificity, etc. To overcome these limitations, we assess and benchmark the deep learning models applied for melanoma diagnosis by considering similar methodological constraints, similar experimental settings and parameter setups, and similar evaluation criteria for all the deep learning models used in this study. Due to the existence of trade-offs and conflict among performance evaluation criteria during the evaluation process, the benchmarking of DL models is dealt with as a multiple criteria problem [21]. Accordingly, multi-criteria decision-making schemes (MCDM) can be exploited to benchmark the convolutional neural network models used for melanoma diagnosis.
Multi-criteria decision-making methods (MCDM) are an application of decision theory that handles multi-objective choice. It is a strategy for assessing and comparing multiple solutions (alternatives) considering competing criteria. It is a widely used decision-making approach in the field of operational research that deals with several criteria to find an optimal solution for decision makers. MCDM techniques find the optimal selection by ranking the performance of the alternatives, where the highest rank is assigned the best feasible alternative (solution) [21,22,23]. Two key problems could arise during the evaluation and benchmarking of deep convolutional neural network models for melanoma detection. First, what are the suitable criteria for the evaluation? Second, what is the proper benchmarking approach for selecting the optimal model considering the provided criteria?. Thus, the motivation of this work is to present a framework for evaluating and benchmarking multiple deep learning models for melanoma detection, using various evaluation criteria.
In light of the concerns mentioned above and given the rapid development of deep learning algorithms for melanoma diagnosis, it is crucial to validate and benchmark these models, which is the main challenge of this work. This research direction aims to conduct a comprehensive evaluation and benchmark of convolutional neural networks for melanoma diagnosis. The benchmarking is accomplished by prioritizing convolutional network architectures and then selecting the optimal architecture, given specific criteria. The contribution of our work can be represented in four-fold as follows:
  • The proposed study provides an appropriate and powerful linkage between the multi-criteria decision-making techniques and the objective performance evaluation criteria, which are typically used to evaluate the deep learning models. This integration with decision-making schemes helps to rank the learning models based on multiple conflicting criteria and select the optimal model in our case study.
  • This is the first study that introduces the application of a multi-criteria decision-making approach based on merging entropy and PROMETHEE methods to help prioritize the deep convolutional neural networks used for melanoma diagnosis and select the optimal model considering various criteria.
  • This study presents a comprehensive evaluation of 19 convolutional neural network models with a two-class classifier. The models are trained and evaluated on a dataset of 991 dermoscopic images considering 10 performance evaluation metrics.
  • The findings of our investigations would aid and expedite the timely deployment of artificial intelligence (AI)–assisted CAD systems to clinics and hospitals with regard to easing model selection under different criteria.
The remainder of this paper is presented as follows: in Section 2, the materials and proposed methods are described and explained; the experiments and results of the proposed system are designed, reported and discussed in Section 3; and finally, the work is concluded in Section 4.

2. Materials and Methods

2.1. Materials

To carry out our experiments, dermoscopic images were collected from the openly available International Skin Imaging Collaboration (ISIC 2017) dataset [18]. Melanoma, seborrheic keratosis, and nevus, shown in Figure 1, are the three types of lesions represented in the dataset. Melanoma is a cancerous skin tumor with a high mortality rate. Seborrheic keratosis and nevus, the other two types of lesions, are benign skin tumors formed from different cells. Although the ISIC Challenge 2017 included three subtasks with annotations for three classes (nevus, seborrheic keratosis, and melanoma), we only consider the melanoma subtask versus the remaining classes, producing a two-class classification task. The ISIC (2017) dataset comprises 2000 training images and 600 test images. In the training set, there are 374 melanoma images and 1626 non-melanoma images. The test set contains 117 melanoma images and 483 non-melanoma images. In total, both training and test data comprise 491 melanoma images and 2109 non-melanoma images. The percentage of melanoma images in the dataset is 19%. This ratio shows a highly imbalanced data distribution between the two classes. Our study does not target to develop a new method for melanoma diagnosis competing with other methods in which particular strategies are designed to remedy and alleviate the effect of imbalanced data. Instead, this study aims to evaluate and benchmark the existing CNNs architectures considering multiple conflicting criteria. The condition of benchmarking in this study is set for balanced data. Thus, to maintain the balance of classes distribution, all the melanoma images (491) in the dataset are collected, whereas only the first 500 non-melanoma images are gathered, producing 991 dermoscopic images in total. The data are split into five folds for training and testing. In each of the five training cycles, four folds are used for training, and the hold-out set is used for testing the network performance. Thus, in each training process, this generates 393 images (melanoma) and 400 images (non-melanoma) for training, and 98 images (melanoma) and 100 images (non-melanoma) for testing.

2.2. Methods

Our developed evaluation and the benchmarking system illustrated in Figure 2 comprises five main stages, including data preparation, designing of CNN models, training of CNN models, evaluation criteria establishment, and benchmarking of CNN models using MCDM. In the first and second phases of the proposed framework, depicted as red and orange blocks in Figure 2, the data are prepared, and deep convolutional neural networks are implemented (different versions of a specific CNN architecture are considered; for instance; VGG16, VGg19). In the third phase, depicted as a grey block, the CNN models are trained. The key evaluation criteria are identified and measured by evaluating the trained models on test data. In the final phases, shown as blue and green blocks, MCDM methods are employed to prioritize the alternatives (i.e., CNN models). The blue block shows the construction of the decision matrix (models as rows and criteria as columns); then, the entropy method is applied to calculate and generate the weights of criteria. Finally, the MCDM methods (PROMETHEE and VIKOR) are exploited to rank CNN models and report the optimal CNN architecture considering the provided decision matrix and the weights of criteria. Although PROMETHEE and VIKOR are different statistical methods, the input data of these methods are the same, which are the weights of criteria and the decision matrix. These methods are independent; therefore, they are applied to the given input data separately. In this section, each phase of the proposed framework is described as follows:

2.2.1. Pre-Trained Convolutional Neural Network Models (CNNs)

The key CNN baseline architectures that have been applied in this study are summarized below:
  • AlexNet: In 2012, AlexNet [24] substantially surpassed all previous classification methods, winning the ImageNet Large Scale Visual Recognition Competition (ILSVRC) by reducing top-5 error from 26% to 15.33%. The network’s design was similar to the LeNet network developed by Yann LeCun et al. [25], but it was deeper, with more filters per layer and layered convolutional layers. 11 × 11 , 5 × 5 , 3 × 3 convolutions filters, max pooling, dropout, data augmentation, ReLU activations, and SGD with momentum were all included. After each convolutional layer, added ReLU activations were added. AlexNet was trained using two Nvidia Geforce GTX 580 GPUs for six days, which is why their network is divided into two pipelines.
  • VGG16,19: Simonyan and Zisserman presented the VGG architecture in 2014 [26]. It is a straightforward design, with only blocks made up of an incremental number of convolution layers and 3 × 3 filters. Furthermore, max-pooling blocks follow convolution blocks to reduce the size of the activation maps obtained. Finally, a classification block is employed, consisting of two dense layers and a final output layer. The numbers 16 and 19 refer to how many weighted layers each network includes. On the other hand, this network has a couple of drawbacks: it takes too long to learn and has a lot of parameters.
  • InceptionV1,V3: Google implemented inception building blocks in GoogLeNet (Inceptionv1) [27]. These blocks function well together and result in a model that is easy to generalize. GoogLeNet is made up of nine Inception modules that are stacked one on top of the other. There are a total of 27 layers, 5 of which are pooling layers. The total number of layers used in the network design is about 100. New revisions of the model appeared as the model was updated regularly. Inception-v2 and Inception-v3 [28] were released within a short time gap in 2015. Except for a few features, Inception-v2 integrates all of GoogLeNet’s features. Filter banks were increased in width in Inception-v2 to eliminate the “representational bottleneck”. All of the changes from Inception-v2 were included in Inception-v3. Furthermore, Inception-v3 underwent additional changes, such as the use of a higher resolution input and the use of the RMSProp optimiser, which significantly reduced the cost function.
  • InceptionResNetV2: Inception V4 was launched in 2016 by Google researchers in conjunction with Inception-ResNet. By implementing Inception-V4, the main goal of this network architecture was to reduce the complexity of the Inception V3 model, which provided state-of-the-art accuracy on the ILSVRC2015 challenge. This architecture also investigates the use of residual networks on the Inception model [29].
  • ResNet18,50,101: The ResNet architecture, founded by He et al. in 2015 [30], was a major turning point in the introduction of an extraordinary form of architecture focused on “modules” or “networks within networks”. The principle of residual connections was first implemented in these networks. ResNet comes in various sizes and numbers of layers—such as ResNet18, RerNet50, and RerNet101—but the most common is ResNet50, which has 50 layers with weights. Despite having many more layers than the VGG, ResNet50 needs nearly five times less memory. This is because, instead of dense layers, this network uses a layer called GlobalAveragePooling in the classification stage, which transforms the 2D feature maps of the last layer in the feature extraction stage into an n-classes vector that is used to measure the likelihood of belonging to each class.
  • DenseNet201: DenseNet [31] is very similar to ResNet, but there are a few key differences. DenseNet concatenates the output of the previous layer with the output of the next layer. At the same time, ResNet follows an additive approach that combines the previous layer (identity) with the next layer. DenseNet model was founded mainly to address the vanishing gradient’s impact on high-level neural networks’ layers. Using the composite function operation, the previous layer’s output becomes the second layer’s input. Convolution, pooling, batch normalization, and non-linear activation layers form this composite process. DenseNet comes in a variety of types, including DenseNet-121, DenseNet-169, and DenseNet-201. The numbers represent the number of the neural network’s layer.
  • Xception: Xception [32] is an extension of the Inception architecture that uses depthwise separable convolutions to replace the regular Inception modules. The mapping of cross-channel and spatial correlations in the feature maps of convolutional neural networks can be fully decoupled in this network. The authors called their proposed architecture Xception, which stands for “Extreme Inception,” since this hypothesis is a stronger version of the hypothesis that underlies the Inception architecture. In a nutshell, the Xception architecture is a depthwise separable convolution layers stack with residual connections. This makes it very simple to establish and change the architecture.
  • MobileNet: MobileNet [33] is a convolutional neural network designed for mobile and embedded vision uses. They are based on a streamlined architecture that builds lightweight deep neural networks with low latency for mobile and embedded devices, using depthwise separable convolutions. The width multiplier and resolution multiplier parameters are added to make it easier to tune MobileNet. The depthwise convolution in MobileNets applies a single filter to each input channel. After that, the pointwise convolution applies a 1 × 1 convolution to combine the depthwise convolution’s outputs. A separate layer for filtering and a separate layer for combining are used in depthwise separable convolution. This factorization has the effect of reducing the computation and model size drastically.
  • NASNetMobile and NASNetLarge: Google Brain built Neural Architecture Search (NASNet) [34]. The authors suggested that an architectural building block be detected on a small dataset and then transferred to a larger dataset. They generally look for the best convolutional layer or cell on a small dataset first, then stack together more copies of this cell to extend to the larger dataset. A new regularization technique called ScheduledDropPath was proposed, which significantly enhances the generalization of the NASNet models. With a smaller model size and lower complexity, the NASNet method achieves state-of-the-art results. While the overall architecture of NASNet is predefined, the blocks or cells are not. Alternatively, a reinforcement learning search technique is used to find them. The authors developed different versions of NASNets with different computational requirements. The larger model, NASNetlarge, is a convolutional neural network trained on over onen million images from the ImageNet database, while the smaller model, NASNetMobile, is optimized for mobile devices.
  • ShuffleNet: ShuffleNet [35] is a convolutional neural network optimized for mobile devices with minimal processing capacity developed by Megvii Inc. (Face++). The network architecture design considers two new operations to lower computation costs while retaining accuracy: pointwise group convolution and channel shuffle. It specializes in common mobile platforms, such as drones, robots, and smartphones, and aims for the best accuracy in minimal computational resources.
  • DarkNet19,53: The backbone of YOLOv2 is a convolutional neural network called Darknet-19 [36]. It generally employs 3 × 3 filters and twice the number of channels after each pooling phase, similar to VGG models. It leverages global average pooling to produce predictions and 1 × 1 filters to compress the feature representation among 3 × 3 convolutions, identical to the work on Network in Network (NIN). Batch normalization is a technique for stabilizing training and accelerating convergence. Darknet-53 [37], on the other hand, is a convolutional neural network that serves as the backbone for the YOLOv3 object detection method. The utilization of residual connections and more layers are an enhancement over its predecessor, Darknet-19.
  • EfficientNetB0: EfficientNetB0 [38] is a convolutional neural network that scales depth, width, and resolution dimensions, using a compound coefficient. Unlike the traditional methodology, which arbitrarily scales network dimensions, the EfficientNetB0 scaling strategy scales network dimensions with a set of predetermined scaling coefficients. According to the compound scaling approach, if the input image is larger, the network needs more layers and channels to widen the receptive field and catch more fine-grained patterns on the larger image. In addition to squeeze-and-excitation blocks [39], the base of EfficientNet is built on MobileNetV2’s inverted bottleneck residual blocks [33].
  • SqueezeNet: DeepScale, UC Berkeley, and Stanford University collaborated to develop SqueezeNet [40]. With 50× fewer parameters, SqueezeNet reaches AlexNet-level accuracy on ImageNet. Additionally, the authors were able to compress SqueezeNet to less than 0.5 MB, using model compression approaches (510× smaller than AlexNet). Smaller convolutional neural networks (CNNs) require less communication across servers during distributed training and less bandwidth. They are also more feasible to be deployed on FPGAs and hardware with restricted computational resources and limited memory.

2.2.2. Benchmarking Criteria

This section presents elaboration for the criteria taken into consideration in this study. The choice of criteria in MCDM methods is highly dependent on the decision-making context, and the problem handled. As we deal with a classification problem, our study has established the most popular measurements typically used for classifiers’ evaluation as criteria. The performance of each CNN model was evaluated in this stage, using 10 evaluation metrics. We utilized the test accuracy, F1-score, sensitivity, specificity, precision, false-positive rate and false-negative rate, Matthews correlation coefficient (MCC), classification error, network complexity to evaluate each of the model targeted for study in this research.
  • Accuracy: this metric measures how close the predicted value is to the actual data values. It can be defined using the following formula:
    A c c u r a c y ( A c c ) = t p + t n t p + t n + f p + f n
    t p : True Positive, t n : True Negative, f p : False Positive, f n : False Negative
  • Classification error: This refers to the number of samples incorrectly classified (false positives and false negatives). It can be defined as follows:
    C l a s s i f i c a t i o n E r r o r ( E r r ) = 1 A c c
  • Precision: The precision metric tests the ability of the classifier to reject irrelevant samples. The formula of this metric can be defined as follows:
    P r e c i s i o n ( P r e ) = t p t p + f p
  • Sensitivity: The sensitivity metric measures the proportion of the correctly detected relevant samples. It can be represented as follows:
    S e n s i t i v i t y ( S n ) = t p t p + f n
  • F1-Score: The F1-score can be obtained by the weighted average of sensitivity (recall) and precision, where the relative contribution of both recall and precision to the F1-score are equal. The F1-score can be defined as follows:
    F 1 S c o r e = 2 ( P r e c i s i o n × R e c a l l ) P r e c i s i o n + R e c a l l
    where R e c a l l = S e n s i t i v i t y
  • Specificity: It describes the ability of the classifier to detect the true negative rate. The formula of specificity can be defined using the following equation:
    S p e c i f i c i t y ( S p ) = t n t n + f p
  • False-Positive Rate (FPR): This is the proportion of negative examples wrongly categorized as positive. This metric is also known as the miss rate and is represented as follows:
    F a l s e P o s i t i v e R a t e ( F P R ) = f p f p + t n
  • False-Negative rate (FNR): This is the proportion of negative examples wrongly categorized as positive. This metric is also known as the fall-out rate. This evaluation criterion is introduced as follows:
    F a l s e N e g a t i v e R a t e ( F N R ) = f n f n + t p
  • Matthews Correlation Coefficient (MCC): The MCC is a correlation coefficient that yields a value between 1 and + 1 for actual and estimated binary classifications. A coefficient of + 1 shows ideal prediction, 0 shows random prediction, and 1 indicates complete disagreement between predictions and the ground truth. The MCC can be defined as follows:
    M C C = t p × t n f p × f n t p + f p t p + f n t n + f p t n + f n
  • CNN Complexity: This refers to the number of parameters existing in the pre-trained CNN.

2.2.3. Multi-Criteria Decision Making (MCDM)

Multi-criteria decision making typically involves six phases: (i) problem formulation, (ii) identification of requirements, (iii) goal setting, (iv) identification of alternatives, (v) development of criteria, and (vi) the identification and application of decision-making techniques. This process can be carried out using various mathematical procedures chosen based on the problem at hand, and the level of complexity ascribed to the decision-making process [41,42]. This study has formulated the CNN models benchmarking as the research goal, considering 19 CNNs as alternatives and 10 criteria. For decision making, preference ranking organization method for enrichment evaluation (PROMETHEE) [43], an MCDM method, is adopted to generate the ranking list and to produce the optimal model selection, using the criteria’s weights computed by the entropy method. For validating the optimal model selection, another MCDM method called VlseKriterijumska Optimizacija I Kompromisno Resenje (VIKOR) in Serbian [44], which means multi-criteria optimization and compromise solution, is also applied. This section describes the MCDM methods exploited to rank the CNN models and selects the optimal model, given the criteria mentioned earlier, using the data in our case study.
  • Entropy: This method computes relative weights by objectively interpreting the relative intensities of the criteria significance based on data discrimination [45]. MDCM’s generated decision matrix D M is defined by m alternatives (19 CNN models) and k criteria (10 criteria), which are represented as follows:
    D M = x i j m × k
    From the constructed decision matrix D M , the procedure of entropy weighting method described in [45] is followed to measure the weights w j . x i j refers to each entry in the D M , where i = 1 , , m , j = 1 , , k . The steps of the entropy weighting method [45] are described as follows:
    Step1: Normalizing the decision matrix using the following equation:
    p i j = x i j i = 1 m x i j , 1 i m , 1 j k
    Step2: Measuring the entropy value for each criterion as follows:
    e j = g j = 1 k p i j ln p i j , g = 1 / ln m , 1 i m , 1 j k .
    Step3: Determining the inherent contrast intensity of each criterion as follows:
    d i = 1 e j , 1 j k
    Step4: The entropy weights of criteria are then defined as follows:
    w j = d j / j = 1 k d j , 1 j k
  • PROMETHEE: The PROMETHEE is an outranking approach for ranking and selecting a finite collection of alternatives based on often competing criteria. Compared to other multi-criteria analysis methods, PROMETHEE II is an uncomplicated complete (not partial) ranking method in terms of conception and application. The stepwise procedure of PROMETHEE II can be defined as follows, giving the provided decision matrix and the weights of criteria:
    Step 1: Determining of deviations based on pairwise comparisons as follows:
    d j ( a , b ) = g j ( a ) g j ( b )
    where d j ( a , b ) refers to the difference between the evaluations of a and b on each criterion.
    Step 2: Preference function application:
    P j ( a , b ) = F j d j ( a , b ) j = 1 , , k
    where P j ( a , b ) denotes the preference of alternative a with regard to alternative b on each criterion, as a function of d j ( a , b ) .
    Step 3: Calculating an overall or global preference index using the following formula:
    π ( a , b ) = j = 1 k P j ( a , b ) w j
    where π ( a , b ) of a over b represents the weighted sum p ( a , b ) for each criterion, and w j is the weight w j related to the j th criterion.
    Step 4: Calculating the partial ranking PROMETHEE I (outranking flows) using the following equations:
    ϕ + ( a ) = 1 m 1 b = 1 m π ( a , b )
    ϕ ( a ) = 1 m 1 b = 1 m π ( b , a )
    where ϕ + ( a ) and ϕ ( a ) represent the positive outranking flow and negative outranking flow for each alternative, respectively.
    Step 5: Calculating the complete ranking PROMETHEE II (outranking flows) using the following equations:
    ϕ ( a ) = ϕ + ( a ) ϕ ( a )
    where ϕ ( a ) represents the outranking flow for each alternative.
  • VIKOR: The VIKOR approach [44] was initially developed to optimize complex systems that involve various parameters. Using the predefined weights, the VIKOR provides a compromise ranking list and suggests a compromise solution. VIKOR creates a multi-criteria rating index based on a specific “closeness” metric to the “ideal” solutions [44]. The VIKOR methodology’s compromise ranking algorithm can be described as follows, giving the provided decision matrix and the weights of criteria.
    Step1: Determining the best value as x j * and the worst value as x j of the criteria as j = 1 , 2 , , k . This also leads to configure the criteria as beneficial and non-beneficial values. The beneficial attributes require being maximized, while the non-beneficial ones need to be minimized, which are identified as follows:
    Rule1: Best value for beneficial criteria is x j * = m a x x i j , and for non-beneficial is x j * = m i n x i j ,
    Rule2: Worst value for beneficial criteria is x j = m i n x i j , and for non-beneficial is x j = m a x x i j .
    Step2: Determining the values of S i and R i , where i = 1 , 2 , , m using the following equations:
    S i = j = 1 k w j x j * x i j / x j * x j , R i = max j w j x j * x i j / x j * x j ,
    where w j are the weights of criteria computed using the entropy method.
    Step3: Determining the values of S * and R * as follows:
    S * = min i S i , R * = min i R i , S = max i S i , R = max i R i
    Step4: Determining the values of Q i ; where i = 1 , 2 , , m and v is defined as the weight of the scheme of “the majority of criteria” using the following equation:
    Q i = v S i S * / S S * + 1 v R i R * / R R *
    Step5: Ranking the alternatives by sorting the values of Q i in ascending order.

3. Experimental Results and Discussion

3.1. Experimental Setup and Training

During the experimental process, 19 CNN models pre-trained on ImageNet dataset [46] were modified and re-trained using transfer learning and fine-tuning strategies to classify the skin lesion into two classes: cancerous (melanoma) or non-cancerous (non-melanoma). The characteristics of the CNN architectures in terms of number of total layers, number of learnable layers, size of CNN, size of the input image, and number of parameters in each network architecture are described in Table 1. In the training of models, binary cross-entropy was preferred as a cost function, and the stochastic gradient descent with momentum (SGDM) optimizer to minimize the cost function. The softmax activation function was used in the output layer of the models. Each model was trained through six epochs, and the training was repeated for a total of five times. The batch size is set to 10, providing 79 iterations per epoch and 474 iterations for six epochs. The learning rate value was set to 0.0003 and momentum of 0.9. The learning curves of 19 CNN models are presented in Figure A1.
To provide fair performance evaluation and benchmarking among the nineteen models, we opted to use a fixed number of epochs for all models. Figure A1 shows that all the models stopped training at the same endpoint, and the trained models were deployed from this endpoint to conduct the testing phase. We aimed to compare the performance of the networks under the same constraints and conditions. So, choosing the optimal number of epochs to train a particular model was not considered. Considering learning the models under the same conditions, if one model encounters overfitting and subsequently fails to achieve good accuracy on the unseen test set, whereas another model has not undergone overfitting, the later model is preferred over the former model. However, in Figure A1, it can be noticed that the training and validation curves show a steady learning behavior, and there is no indication of overfitting. In order to prevent potential overfitting during the training, the online data augmentation is applied by using various image transformation methods, such as vertical and horizontal flipping, random translation in the range of [ 30 , 30 ] , and random scaling in the range of [ 0.9 , 1.1 ] .
In most of the CNN models, the last layer is the learnable weights of fully connected layers. Thus, to apply the transfer learning and fine-tune the network, using our data, these completely connected layers are replaced with a new, fully-connected layer, comprising two neurons adhering to the two classes in our study. Instead of fully connected layers, the last learnable layer in some networks, such as SqueezeNet, is a 1 × 1 convolutional layer. In this scenario, the old convolutional layer is replaced by a new convolutional layer with the same number of filters as classes.

3.2. Results of the Experiments and Discussion

To examine the classification performance of the models, nine evaluation metrics widely used in classification tasks are used, including accuracy, classification error, precision, sensitivity, specificity, F1-score, false-positive rate, false-negative rate, and Matthews correlation coefficient. Table 2 depicts the evaluation performance of the 19 CNN models describing the average value and the standard deviation of a specific criterion over the five folds. This study reveals the high evaluation performance of the CNN models for melanoma diagnosis, employing a balanced number of dermoscopic images through a thorough analysis of 19 pre-trained CNNs using a specific parameter configuration and learning technique for the networks.
As shown in Table 2, the ResNet101 model reported the best average test accuracy and MCC with 94.34% and 88.96%, respectively, compared to other CNN models. The highest F1-score with a value of 93.96% was attained by Densenet201, followed by ResNet101 with a value of 93.89%. Furthermore, Inceptionv3 achieved the highest specificity and precision values with 96.8% and 96.11%, followed by 96% specificity achieved by MobileNetv2 and 95.36% precision achieved by ResNet101. DenseNet201 produced the highest sensitivity of 93.47%, followed by 92.86% reported in ResNet101. It can also be noticed that Inceptionv3 attained the lowest FPR of 3.2%, while DenseNet201 revealed the lowest FNR of 6.53%, and the smallest error, 5.66%, was reported by ResNet101. According to the minimum number of parameters, SqueezeNet has 1.24 million parameters, which is the optimal number, compared to other CNN models. Table 2 also explores the deviation among the accuracies reported from the five folds and exposes the difficulty in recognizing the best model based on the variation of the accuracies in the five folds. Likely, Table 3 and Figure 3 show that there is no superior CNN model over others, due to the lack of a CNN model that achieves the best accuracies through the five folds. This would lead to difficulty in selecting the best model, while considering other criteria.
Figure 4 exhibits the trade-off and conflict among the evaluation criteria of the 19 CNN models. For instance, a trade-off between sensitivity (true positive rate) and specificity (true negative rate) should be considered, where DenseNet201 reports the highest sensitivity, whereas Inceptionv3 attains the highest specificity. Precision is also independent and has a trade-off with accuracy. Accuracy is the degree of veracity, while precision is the degree of reproducibility. That means that it is possible to be very precise but not very accurate, and it is also possible to be accurate without being precise. The best quality detection is both accurate and precise. Inceptionv3 achieves the highest precision, whereas Resnet101 reveals the best accuracy. It should also produce a trade-off between FNR and FPR, where Inceptionv3 reports the lowest FPR, while DenseNet201 reports the lowest FNR. Thus, it is crucial to make a trade-off between the models that could achieve the optimal diagnosis by reducing the number of negative cases falsely diagnosed as positive and the models that could reach the optimal diagnosis by reducing the number of positive instances falsely diagnosed as negative. The F1-score is also needed to achieve a balance between precision and sensitivity, where Densenet201 provides the best F1-Score followed by Resnet101. For the number of parameters required to determine the network complexity, SqueezeNet has the lighter network architecture, compared to VGG19, which has the largest network architecture. Although SqueezeNet is optimal in terms of network complexity, it still shows moderate-low accuracy performance through the five folds shown in Figure 3. Additionally, there is a conflict between the criteria that are required to be minimized (such as FNR, FPR, Err, and the number of parameters) and the criteria targeted to be maximized (such as Acc, Sen, Spe, Pre, F1-score, and Mathew).
From Figure 4, it can also be noticed that there is no superior CNN model, due to the conflict among evaluation criteria and the difficulty to optimize all criteria simultaneously. Hence, selecting the best deep learning model for automated melanoma diagnosis considering multiple conflicted criteria is a difficult task, due to the variance of the criteria significance, the conflict among these criteria, and the trade-off among them. Therefore, benchmarking CNN architectures for melanoma detection is crucial for selecting the optimal model, achieving a trade-off among the 10 pre-defined evaluation criteria. The multiple criteria decision-making method (MCDM) [43,44] is targeted to apply and rank the 19 models according to their performance, considering the trade-off among the criteria. Thus, the best-selected networks could be easily adopted to construct an ensemble learning system for melanoma diagnosis or even use the optimal network to construct a system using a single model.
To achieve the goal of our study by generating a ranking list for CNN models and selecting the optimal solution, the PROMETHEE method [43] is applied considering the 19 alternatives (CNN models) and 10 criteria. To further validate the decision made by PROMETHEE, we also applied the VIKOR approach [44] using the same data setting and configuration. First, the decision matrix D M is constructed using m alternatives, in our case 19, and the k criteria, in our case 10, producing D M of size 19 × 10 . The criteria are then classified into two categories according to the required optimization strategy. The first category includes the criteria that require minimization, including classification error, false-positive rate, false-negative rate and number of parameters, known as non-beneficial criteria. Unlikely, the second category includes the criteria that require maximization, including accuracy, sensitivity, specificity, precision, F1-score and MCC, known as beneficial criteria. The Equations (24) and (25) defined below are used for normalizing the non-beneficial and beneficial criteria, respectively. The normalized criteria are shown in Table 4.
x ¯ i j = x j min x i j
x ¯ i j = x i j x j max
x i j refers to the entries of the decision matrix D M , where i = 1 , , m , j = 1 , , k , k represents the number of alternatives (19 CNN models), and m defines the number of criteria (10 criteria).
To measure the weights of criteria, the entropy method [45] is exploited and applied on the normalized D M producing the weight values of 0.964825438, 0.804398756, 0.985470611, 0.951881312, −1.420375792, −1.473036988, 1.02152041, 0.49110277, −1.294287661, −0.031498856 for accuracy, sensitivity, specificity, F1-score, FNR, FPR, precision, MCC, classification error and number of parameters, respectively. The obtained weights, along with the normalized D M , are used to make the optimal selection, using the PROMETHEE method [43]. The equations used to measure the ranking list are described earlier in Section 2.2.3. We have used the threshold function as the preference function (0 if d ≤ 0 and 1 if d ≥ 0) required in Step 2 in the stepwise procedure of PROMETHEE. To calculate the complete ranking list, ϕ ( a ) represents the outranking flow for each alternative as shown in Table 5. The highest ϕ ( a ) value indicates the compromised solution, which could be chosen as the optimal model. PROMETHEE reports a value of 150.84, the highest ϕ ( a ) for the ResNet101 CNN model and 133.24 as the second-best value for the DenseNet201 model.
To validate the model selection made by PROMETHEE, the VIKOR [44] method is also applied, considering the same weights and the same D M . Unlike PROMETHEE, the lowest Q value in VIKOR indicates the compromised solution, which could be chosen as the optimal model, shown in Table 5. VIKOR reports a value of 0, the lowest Q for the ResNet101 CNN model, and 0.079 as the second-lowest value for the DenseNet201 model. Thus, the mathematical consistency of the judgements coming out of PROMETHEE II was tested and proven. Hence, the effectiveness of the model ranking produced by PROMETHEE II was validated by demonstrating the agreement between two different statistical methods, considering the same conflicting criteria.
To provide a direct and explicit comparison between the two decision-making methods, PROMETHEE and VIKOR, Table 6 elaborates the optimal CNN model selection in both approaches. It can be noticed that until the seventh rank, the two methods have a similar decision for the optimal CNN model selection. Likewise, the ranks 10, 11, 12, 13, 15, 18 and 19 provide the exact model recommendation by both approaches. On the other hand, the decision made by methods has slightly different priorities for the 8, 9, 14, 16 and 17 levels. The suggested framework’s findings show that the best model selection decision based on numerous conflict factors is robust and reliable.
This work developed a new multi-criteria decision-making methodology that aids in assessing the criteria that influence the decision to choose a specific CNN model, prioritizing the models and selecting the best model. When software developers need to find an effective CNN model that meets specified requirements for constructing a robust CAD system, the proposed approach of revealing the CNN models’ priorities would be beneficial and valuable. Finally, our study may provide and draw a new line in the evaluation and benchmark of the deep learning models for various diseases. Although the proposed benchmarking framework has made progress in benchmarking the models used for melanoma diagnosis from dermoscopy images, there is still space for improvement in research work. In future work, we aim to study the effect of the model selection, considering different criteria. The criteria that are to be considered include (i) training the models under several transfer learning scenarios and data augmentation strategies, (ii) exploring the impact of several optimization schemes, and (iii) testing various class balancing and weighting techniques. We also consider training the models on several datasets, targeting the effect of variation among datasets. These reported limitations and suggested improvements are currently part of the authors’ ongoing research.

4. Conclusions

Medical diagnostics tools based on deep learning of medical images are becoming more widely recognized as clinically relevant AI-based solutions. However, developing appropriate deep neural network models and training strategies for clinical uses is a research area that needs to be investigated. The inaccurate selection of melanoma diagnosis model could be costly to medical organizations, especially when more accurate and efficient diagnosis models are urgently needed. This study investigated the performance of some of these networks for melanoma diagnosis, utilizing dermoscopic images after a thorough evaluation of 19 pre-trained CNNs, using particular evaluation criteria, parameter settings and training strategies. An MCDM-based methodology is presented for evaluating, benchmarking, and ranking melanoma diagnostic models and selecting the most optimal model. The study findings would help in the model selection, designing quick and reliable diagnostic tools based on image data, and contributing to the development of more accurate and efficient point-of-care diagnostic and detection systems. Other image modalities, such as non-dermoscopic (clinical) images, can also be used to train and test the network architecture of the pre-trained models. Therefore, we aim to adapt our proposed network designs in the future to include not only dermoscopic but also clinical images. We would also like to expand the number of training samples and investigate other deep learning training methodologies.

Author Contributions

S.A. conceived and designed the experiments; S.A. and B.A.-B. performed the experiments; S.A., B.A.-B. and W.A.-N. analyzed the data and wrote the paper; W.A.-N. supervised and ran the project. All authors have read and agreed to the published version of the manuscript.

Funding

Saeed Alzahrani was funded by the Kingdom of Saudi Arabia government.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

A publicly available dataset was analyzed in this study. This data can be found in https://challenge.isic-archive.com/data, accessed on 18 May 2021. Both the data analyzed during the current study and code are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DLDeep Learning
CNNConvolutional Neural Networks
MMalignant
BBenign
CADComputer-Aided Diagnosis
FPRFalse Positive Rate
FNRFalse Negative Rate
MCDMMulti-Criteria Decision Making
PROMETHEEPreference Ranking Organization Method for Enrichment of Evaluations
VIKORVisekriterijumska Optimizacija I Kompromisno Resenje

Appendix A

Figure A1. The performance of the CNN models visualizing training, validation and loss curves.
Figure A1. The performance of the CNN models visualizing training, validation and loss curves.
Cancers 13 04494 g0a1aCancers 13 04494 g0a1bCancers 13 04494 g0a1c

References

  1. Ries, L.A.; Harkins, D.; Krapcho, M.; Mariotto, A.; Miller, B.; Feuer, E.J.; Clegg, L.X.; Eisner, M.; Horner, M.J.; Howlader, N.; et al. SEER Cancer Statistics Review 1975–2003. National Cancer Institute . 2006. Available online: https://seer.cancer.gov/archive/csr/1975_2003/ (accessed on 5 September 2021).
  2. Skin Cancer Statistics. Available online: https://www.nhs.uk/conditions/non-melanoma-skin-cancer/ (accessed on 1 May 2021).
  3. Non-Melanoma Skin Cancer Statistics. Available online: https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/non-melanoma-skin-cancer (accessed on 1 May 2021).
  4. Melanoma Skin Cancer Statistics. Available online: https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/melanoma-skin-cancer (accessed on 1 May 2021).
  5. Rosendahl, C.; Tschandl, P.; Cameron, A.; Kittler, H. Diagnostic accuracy of dermatoscopy for melanocytic and nonmelanocytic pigmented lesions. J. Am. Acad. Dermatol. 2011, 64, 1068–1073. [Google Scholar] [CrossRef] [PubMed]
  6. Nirmal, B. Dermatoscopy: Physics and principles. Indian J. Dermatopathol. Diagn. Dermatol. 2017, 4, 27. [Google Scholar] [CrossRef]
  7. Pan, Y.; Gareau, D.S.; Scope, A.; Rajadhyaksha, M.; Mullani, N.A.; Marghoob, A.A. Polarized and nonpolarized dermoscopy: The explanation for the observed differences. Arch. Dermatol. 2008, 144, 828–829. [Google Scholar] [CrossRef] [PubMed]
  8. Smith, D.; Bowden, T. Using the ABCDE approach to assess the deteriorating patient. Nurs. Stand. 2017, 32, 51. [Google Scholar] [CrossRef]
  9. Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer Statistics. CA Cancer J. Clin. 2021, 71, 7–33. [Google Scholar] [CrossRef]
  10. Glazer, A.M.; Rigel, D.S. Analysis of trends in geographic distribution of US dermatology workforce density. JAMA Dermatol. 2017, 153, 472–473. [Google Scholar] [CrossRef]
  11. Skin Cancer Check-Ups a Long Time Coming as Australia Faces Huge Shortage of Dermatologists. Available online: https://www.abc.net.au/news/2021-06-14/gps-to-help-ease-growing-skin-specialist-waiting-times/100211834 (accessed on 27 August 2021).
  12. Eedy, D. Dermatology: A specialty in crisis. Clin. Med. 2015, 15, 509. [Google Scholar] [CrossRef] [Green Version]
  13. Royal College of Physicians Dermatology. London: RCP. Available online: www.rcplondon.ac.uk/sites/default/files/dermatology.pdf (accessed on 27 August 2021).
  14. British Association of Dermatologists Clinical Services. London: BAD. Available online: www.bad.org.uk/healthcare-professionals/clinical-services/ (accessed on 27 August 2021).
  15. Alzahrani, S.; Al-Nuaimy, W.; Al-Bander, B. Seven-point checklist with convolutional neural networks for melanoma diagnosis. In Proceedings of the 2019 8th European Workshop on Visual Information Processing (EUVIP), Rome, Italy, 28–31 October 2019; pp. 211–216. [Google Scholar]
  16. Nami, N.; Giannini, E.; Burroni, M.; Fimiani, M.; Rubegni, P. Teledermatology: State-of-the-art and future perspectives. Expert Rev. Dermatol. 2012, 7, 1–3. [Google Scholar] [CrossRef] [Green Version]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  18. Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 168–172. [Google Scholar]
  19. Naeem, A.; Farooq, M.S.; Khelifi, A.; Abid, A. Malignant melanoma classification using deep learning: Datasets, performance measurements, challenges and opportunities. IEEE Access 2020, 8, 110575–110597. [Google Scholar] [CrossRef]
  20. Pérez, E.; Reyes, O.; Ventura, S. Convolutional neural networks for the automatic diagnosis of melanoma: An extensive experimental study. Med. Image Anal. 2021, 67, 101858. [Google Scholar] [CrossRef] [PubMed]
  21. Keeney, R.L.; Raiffa, H.; Meyer, R.F. Decisions with Multiple Objectives: Preferences and Value Trade-Offs; Cambridge University Press: Cambridge, UK, 1993. [Google Scholar]
  22. Baltussen, R.; Niessen, L. Priority setting of health interventions: The need for multi-criteria decision analysis. Cost Eff. Resour. Alloc. 2006, 4, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Thokala, P.; Devlin, N.; Marsh, K.; Baltussen, R.; Boysen, M.; Kalo, Z.; Longrenn, T.; Mussen, F.; Peacock, S.; Watkins, J.; et al. Multiple criteria decision analysis for health care decision making—An introduction: Report 1 of the ISPOR MCDA Emerging Good Practices Task Force. Value Health 2016, 19, 1–13. [Google Scholar] [CrossRef] [Green Version]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  25. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  26. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  27. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  28. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  29. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2016; pp. 770–778. [Google Scholar]
  31. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  32. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  33. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  34. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8697–8710. [Google Scholar]
  35. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  36. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  37. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  38. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  39. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  40. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50× fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  41. Jahan, A.; Edwards, K.L.; Bahraminasab, M. Multi-Criteria Decision Analysis for Supporting the Selection of Engineering Materials in Product Design; Butterworth-Heinemann: Oxford, UK, 2016. [Google Scholar]
  42. Ivlev, I.; Vacek, J.; Kneppo, P. Multi-criteria decision analysis for supporting the selection of medical devices under uncertainty. Eur. J. Oper. Res. 2015, 247, 216–228. [Google Scholar] [CrossRef]
  43. Behzadian, M.; Kazemzadeh, R.B.; Albadvi, A.; Aghdasi, M. PROMETHEE: A comprehensive literature review on methodologies and applications. Eur. J. Oper. Res. 2010, 200, 198–215. [Google Scholar] [CrossRef]
  44. Liou, J.J.; Tsai, C.Y.; Lin, R.H.; Tzeng, G.H. A modified VIKOR multiple-criteria decision method for improving domestic airlines service quality. J. Air Transp. Manag. 2011, 17, 57–61. [Google Scholar] [CrossRef]
  45. Hainmueller, J. Entropy balancing for causal effects: A multivariate reweighting method to produce balanced samples in observational studies. Political Anal. 2012, 20, 25–46. [Google Scholar] [CrossRef] [Green Version]
  46. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
Figure 1. Example of images used to conduct this study. Both nevus and seborrhoeic keratosis are classified as non-melanoma in our experiments.
Figure 1. Example of images used to conduct this study. Both nevus and seborrhoeic keratosis are classified as non-melanoma in our experiments.
Cancers 13 04494 g001
Figure 2. The block diagram of the proposed framework used to benchmark CNN models for melanoma diagnosis. M refers to malignant (melanoma) and B refers to benign (non-melanoma).
Figure 2. The block diagram of the proposed framework used to benchmark CNN models for melanoma diagnosis. M refers to malignant (melanoma) and B refers to benign (non-melanoma).
Cancers 13 04494 g002
Figure 3. The obtained accuracies over five folds in the nineteen CNN models. It shows that there is no superior CNN model over others due to the lack of a CNN model that achieves the best accuracies through the five folds. This would lead to difficulty selecting the best model while considering another conflicting criterion, such as the network complexity.
Figure 3. The obtained accuracies over five folds in the nineteen CNN models. It shows that there is no superior CNN model over others due to the lack of a CNN model that achieves the best accuracies through the five folds. This would lead to difficulty selecting the best model while considering another conflicting criterion, such as the network complexity.
Cancers 13 04494 g003
Figure 4. The mean value over the five folds for specific evaluation criteria, along with the number of parameters (the network complexity). No single model achieves the best performance in all evaluation criteria. If a CNN model achieves the best evaluation performance in some evaluation criteria, it may fail to gain superior performance in the remaining criteria.
Figure 4. The mean value over the five folds for specific evaluation criteria, along with the number of parameters (the network complexity). No single model achieves the best performance in all evaluation criteria. If a CNN model achieves the best evaluation performance in some evaluation criteria, it may fail to gain superior performance in the remaining criteria.
Cancers 13 04494 g004
Table 1. Characteristics of the pre-trained CNN architectures adopted in our study.
Table 1. Characteristics of the pre-trained CNN architectures adopted in our study.
Network#Layers#Learnable LayersNetwork Size (MB)Input Image Size#Para (Millions)
AlexNet [24]258227 227 × 227 61
Vgg16 [26]4116515 224 × 224 138
Vgg19 [26]4719535 224 × 224 144
GoogleNet (Inceptionv1) [27]1442227 224 × 224 7
Inceptionv3 [28]3154889 299 × 299 23.9
ResNet18 [30]711844 224 × 224 11.7
ResNet50 [30]1775096 224 × 224 25.6
ResNet101 [30]347101167 224 × 224 44.6
InceptionResv2 [29]824164209 299 × 299 55.9
Xception [32]1707185 299 × 299 22.9
DenseNet201 [31]70820177 224 × 224 20
MobileNetv2 [33]1545313 224 × 224 3.5
ShuffleNet [35]172505.4 224 × 224 1.4
NasnetMobile [34]913*20 224 × 224 5.3
NasnetLarge [34]1243*332 331 × 331 88.9
DarkNet19 [36]641978 256 × 256 20.8
DarkNet53 [37]18453155 256 × 256 41.6
EfficientNetB0 [38]2908220 224 × 224 5.3
SqueezeNet [40]68185.2 227 × 227 1.24
Table 2. The evaluation performance of the 19 CNN models describing the mean value (m) ± standard deviation (s) of a specific criterion over the five folds.
Table 2. The evaluation performance of the 19 CNN models describing the mean value (m) ± standard deviation (s) of a specific criterion over the five folds.
NetworkmACC ±  sACCmSen ±  sSenmSpe ±  sSpemF1 ±  sF1mFNR ±  sFNRmFPR ±  sFPRmPre ±  sPremMathew ±  sMathewmErr ±  sErr
AlexNet [24]87.07 ± 5.1184.9 ± 10.9589.2 ± 3.786.39 ± 6.2815.1 ± 10.9510.8 ± 3.788.57 ± 3.4974.6 ± 9.8712.93 ± 5.11
Vgg16 [26]89.7 ± 6.2386.94 ± 9.3492.4 ± 6.589.18 ± 6.913.06 ± 9.347.6 ± 6.591.98 ± 6.479.76 ± 12.110.3 ± 6.23
Vgg19 [26]87.37 ± 7.0183.27 ± 1191.4 ± 10.3386.58 ± 7.7616.73 ± 118.6 ± 10.3391.29 ± 9.0275.64 ± 13.3812.63 ± 7.01
GoogleNet (Inceptionv1) [27]87.78 ± 5.8787.55 ± 8.8888 ± 1187.65 ± 5.9212.45 ± 8.8812 ± 1188.71 ± 9.1176.3 ± 11.2212.22 ± 5.87
Inceptionv3 [28]92.93 ± 8.0188.98 ± 11.8296.8 ± 4.3292.29 ± 9.0511.02 ± 11.823.2 ± 4.3296.11 ± 5.4986.18 ± 15.557.07 ± 8.01
ResNet18 [30]90 ± 5.6889.18 ± 4.7190.8 ± 10.1389.97 ± 5.3810.82 ± 4.719.2 ± 10.1391.23 ± 9.3280.41 ± 11.3410 ± 5.68
ResNet50 [30]92.42 ± 7.0790.2 ± 11.2494.6 ± 5.2291.95 ± 7.819.8 ± 11.245.4 ± 5.2294.2 ± 5.6985.21 ± 13.857.58 ± 7.07
ResNet101 [30]94.34 ± 7.2892.86 ± 12.1495.8 ± 3.1993.89 ± 8.267.14 ± 12.144.2 ± 3.1995.36 ± 3.9488.96 ± 14.025.66 ± 7.28
InceptionResv2 [29]90.3 ± 7.9688.57 ± 10.5492 ± 5.7989.87 ± 8.6311.43 ± 10.548 ± 5.7991.34 ± 6.7780.71 ± 15.829.7 ± 7.96
Xception [32]88.99 ± 6.7990 ± 7.8588 ± 8.889.02 ± 6.6810 ± 7.8512 ± 8.888.39 ± 8.0678.3 ± 13.5911.01 ± 6.79
DenseNet201 [31]93.94 ± 4.9793.47 ± 3.8694.4 ± 8.7393.96 ± 4.76.53 ± 3.865.6 ± 8.7394.75 ± 7.6488.15 ± 9.66.06 ± 4.97
MobileNetv2 [33]90.81 ± 7.2485.51 ± 11.9596 ± 3.3989.9 ± 8.1414.49 ± 11.954 ± 3.3995.23 ± 4.3282.25 ± 13.989.19 ± 7.24
ShuffleNet [35]86.06 ± 6.8480.61 ± 9.1691.4 ± 14.3385.24 ± 6.4619.39 ± 9.168.6 ± 14.3391.99 ± 11.5273.6 ± 13.1913.94 ± 6.84
NasnetMobile [34]86.57 ± 6.4780.82 ± 12.7192.2 ± 5.9785.25 ± 7.8719.18 ± 12.717.8 ± 5.9791.28 ± 5.3674.09 ± 12.1213.43 ± 6.47
NasnetLarge [34]91.31 ± 7.0888.16 ± 7.2494.4 ± 7.790.96 ± 7.2211.84 ± 7.245.6 ± 7.794.04 ± 7.982.84 ± 14.178.69 ± 7.08
DarkNet19 [36]86.77 ± 4.1481.02 ± 5.4392.4 ± 3.3685.79 ± 4.6518.98 ± 5.437.6 ± 3.3691.22 ± 3.9573.98 ± 8.1513.23 ± 4.14
DarkNet53 [37]89.19 ± 6.1583.88 ± 9.8894.4 ± 2.9788.26 ± 7.2216.12 ± 9.885.6 ± 2.9793.42 ± 478.87 ± 11.7910.81 ± 6.15
EfficientNetB0 [38]86.87 ± 3.4485.31 ± 3.8688.4 ± 4.8886.56 ± 3.5114.69 ± 3.8611.6 ± 4.8887.96 ± 4.9173.86 ± 7.0313.13 ± 3.44
SqueezeNet [40]84.65 ± 2.3886.73 ± 4.9582.6 ± 6.1984.83 ± 2.3113.27 ± 4.9517.4 ± 6.1983.34 ± 4.869.66 ± 4.7515.35 ± 2.38
Table 3. The obtained accuracies over five folds in the 19 CNN models.
Table 3. The obtained accuracies over five folds in the 19 CNN models.
ModelFold1Fold2Fold3Fold4Fold5
AlexNet78.2889.986.8790.489.9
Vgg1682.3286.8786.8795.9696.46
Vgg1979.880.8190.489.3996.46
Inceptionv182.3284.8583.8491.9295.96
Inceptionv379.890.9196.9799.4997.47
ResNet1882.8385.3592.9392.4296.46
ResNet5081.3190.9192.9397.9898.99
ResNet10181.8294.4496.9798.9999.49
InceptionResv277.2788.8993.4393.9497.98
Xception78.2886.8790.9193.4395.45
DenseNet20186.3691.9496.4697.9897.47
MobileNetv281.3186.8789.997.4798.48
ShuffleNet77.2783.8484.8588.3895.96
NasnetMobile78.2886.3685.8685.8696.46
NasnetLarge80.388.8992.9396.4697.98
DarkNet1981.3185.8684.8590.9190.91
DarkNet5379.2987.3791.4193.9493.94
EfficientNetB084.3483.8485.3588.8991.92
SqueezeNet82.3284.8582.3285.8687.88
Table 4. Normalized decision matrix. Alter.—alternative; Cr.—criterion.
Table 4. Normalized decision matrix. Alter.—alternative; Cr.—criterion.
Alter./ Cr.ACCSenSpeF1-ScoreFNRFPRPreMCCErrPara
AlexNet0.92290.90830.92150.91940.43250.29630.92150.83860.43770.0203
Vgg160.95080.93010.95450.94910.50000.42110.95700.89660.54950.0090
Vgg190.92610.89090.94420.92150.39030.37210.94980.85030.44810.0086
Inceptionv10.93050.93670.90910.93280.52450.26670.92300.85770.46320.1771
Inceptionv30.98510.95201.00000.98220.59261.00001.00000.96880.80060.0519
ResNet180.95400.95410.93800.95750.60350.34780.94920.90390.56600.1060
ResNet500.97960.96500.97730.97860.66630.59260.98010.95780.74670.0484
ResNet1011.00000.99350.98970.99930.91460.76190.99221.00001.00000.0278
InceptionResv20.95720.94760.95040.95650.57130.40000.95040.90730.58350.0222
Xception0.94330.96290.90910.94740.65300.26670.91970.88020.51410.0541
DenseNet2010.99581.00000.97521.00001.00000.57140.98580.99090.93400.0620
MobileNetv20.96260.91480.99170.95680.45070.80000.99080.92460.61590.3543
ShuffleNet0.91220.86240.94420.90720.33680.37210.95710.82730.40600.8857
NasnetMobile0.91760.86470.95250.90730.34050.41030.94970.83280.42140.2340
NasnetLarge0.96790.94320.97520.96810.55150.57140.97850.93120.65130.0139
DarkNet190.91980.86680.95450.91300.34400.42110.94910.83160.42780.0596
DarkNet530.94540.89740.97520.93930.40510.57140.97200.88660.52360.0298
EfficientNetB00.92080.91270.91320.92120.44450.27590.91520.83030.43110.2340
SqueezeNet0.89730.92790.85330.90280.49210.18390.86710.78300.36871.0000
Table 5. Ranking for decision making represented by the values of ϕ in PROMETHEE and Q in VIKOR. The highest ϕ value is the best, whereas the lowest Q is the best.
Table 5. Ranking for decision making represented by the values of ϕ in PROMETHEE and Q in VIKOR. The highest ϕ value is the best, whereas the lowest Q is the best.
Model ϕ : PROMETHEEQ: VIKORPROMETHEEVIKOR
AlexNet−86.540043650.784232851515
Vgg1616.318776280.5104848889
Vgg19−63.81243590.747666591313
Inceptionv1−57.199666870.680966911212
Inceptionv3132.20506340.1846634633
ResNet1815.255469340.461465498
ResNet50115.16330970.2125113244
ResNet101150.8418215011
InceptionResv228.4254640.449610977
Xception−29.982036890.607874251111
DenseNet201133.23556050.0799838922
MobileNetv272.892307950.4216718166
ShuffleNet−106.98197140.85949251818
NasnetMobile−89.200936460.848541617
NasnetLarge73.31931010.3346168555
DarkNet19−76.305652630.810737721416
DarkNet531.4566820090.569573371010
EfficientNetB0−95.93019790.782394291714
SqueezeNet−133.160823111919
Table 6. Optimal CNN model selection in PROMETHEE versus VIKOR approach.
Table 6. Optimal CNN model selection in PROMETHEE versus VIKOR approach.
Model RankPROPMETHEEVIKOR
1ResNet101ResNet101
2DenseNet201DenseNet201
3Inceptionv3Inceptionv3
4ResNet50ResNet50
5NasnetLargeNasnetLarge
6MobileNetv2MobileNetv2
7InceptionResv2InceptionResv2
8Vgg16ResNet18
9ResNet18Vgg16
10DarkNet53DarkNet53
11XceptionXception
12Inceptionv1Inceptionv1
13Vgg19Vgg19
14DarkNet19EfficientNetB0
15AlexNetAlexNet
16NasnetMobileDarkNet19
17EfficientNetB0NasnetMobile
18ShuffleNetShuffleNet
19SqueezeNetSqueezeNet
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alzahrani, S.; Al-Bander, B.; Al-Nuaimy, W. A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis. Cancers 2021, 13, 4494. https://doi.org/10.3390/cancers13174494

AMA Style

Alzahrani S, Al-Bander B, Al-Nuaimy W. A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis. Cancers. 2021; 13(17):4494. https://doi.org/10.3390/cancers13174494

Chicago/Turabian Style

Alzahrani, Saeed, Baidaa Al-Bander, and Waleed Al-Nuaimy. 2021. "A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis" Cancers 13, no. 17: 4494. https://doi.org/10.3390/cancers13174494

APA Style

Alzahrani, S., Al-Bander, B., & Al-Nuaimy, W. (2021). A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis. Cancers, 13(17), 4494. https://doi.org/10.3390/cancers13174494

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop