Next Article in Journal
Ultrasound Imaging versus Radiographs in Differentiating Periapical Lesions: A Systematic Review
Previous Article in Journal
MRI of Finger Pulleys at 7T—Direct Characterization of Pulley Ruptures in an Ex Vivo Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Skin Lesion Classification on Ultrasound Images

by
Péter Marosán-Vilimszky 
1,2,*,
Klára Szalai 
3,
András Horváth 
1,
Domonkos Csabai 
2,
Krisztián Füzesi 
2,
Gergely Csány 
2 and
Miklós Gyöngy 
1,2
1
Faculty of Information Technology and Bionics, Pázmány Péter Catholic University, Práter u. 50/A, 1083 Budapest, Hungary
2
Dermus Kft., Sopron út 64, 1116 Budapest, Hungary
3
Department of Dermatology, Venereology and Dermatooncology, Semmelweis University, Mária u. 41, 1085 Budapest, Hungary
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(7), 1207; https://doi.org/10.3390/diagnostics11071207
Submission received: 20 June 2021 / Accepted: 30 June 2021 / Published: 3 July 2021
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The growing incidence of skin cancer makes computer-aided diagnosis tools for this group of diseases increasingly important. The use of ultrasound has the potential to complement information from optical dermoscopy. The current work presents a fully automatic classification framework utilizing fully-automated (FA) segmentation and compares it with classification using two semi-automated (SA) segmentation methods. Ultrasound recordings were taken from a total of 310 lesions (70 melanoma, 130 basal cell carcinoma and 110 benign nevi). A support vector machine (SVM) model was trained on 62 features, with ten-fold cross-validation. Six classification tasks were considered, namely all the possible permutations of one class versus one or two remaining classes. The receiver operating characteristic (ROC) area under the curve (AUC) as well as the accuracy (ACC) were measured. The best classification was obtained for the classification of nevi from cancerous lesions (melanoma, basal cell carcinoma), with AUCs of over 90% and ACCs of over 85% obtained with all segmentation methods. Previous works have either not implemented FA ultrasound-based skin cancer classification (making diagnosis more lengthy and operator-dependent), or are unclear in their classification results. Furthermore, the current work is the first to assess the effect of implementing FA instead of SA classification, with FA classification never degrading performance (in terms of AUC or ACC) by more than 5%.

1. Introduction

1.1. Motivation

Skin cancer is a disease that is causing a growing problem in the developed world. For instance, one in five Americans are expected to get skin cancer during their lifetime, with an estimated 5.8% rise in melanoma cases for 2021 and a 77% rise in the incidence of non-melanoma skin cancer between 1994 and 2014 [1]. While malignant melanoma (MM) is the most deadly form of skin cancer, thankfully it is about 20 times less common than other forms of skin cancer, with basal cell carcinoma (BCC) being the most common non-melanoma skin cancer [2]. Due to the relative shortage of dermatologists in the midst of increases in skin cancer incidence, the role of computer aided diagnostic approaches is gaining increasing prominence.
Deep neural network-based optical approaches using tens of thousands of clinical records, including dermoscopy images, have achieved an accuracy of around 94% on automated skin lesion classification [3,4]. Despite the high accuracy of optics-based melanoma detection, the addition of subsurface information from ultrasound imaging can further improve classification accuracy [5].
In the last few decades, there has been increased interest in the use of dermatologic ultrasound for skin lesion diagnosis. The appearance of different cancerous and noncancerous skin lesions on ultrasound images, as well as their quantitative acoustical parameters, have been extensively reported [6,7,8,9,10,11,12,13,14,15,16,17,18]. In the following two subsections, ultrasound-based lesion diagnostic methods in general are briefly presented, followed by a brief review of the use of the aforementioned skin-specific features for skin cancer diagnosis.

1.2. Overview of Ultrasound-Based Lesion Diagnostic Methods

There are several lesion types, the automated classification of which is commonly studied in ultrasound imaging. Some of these are: benign versus malignant breast lesion differentiation [19,20,21,22,23,24]; thyroid cancer detection [25,26,27,28] or liver disease classification [29,30]. Lesion diagnosis techniques belong either to the traditional Computer-Aided Diagnostic (CAD) class of methods, or to the relatively recent class of deep learning (DL) methods. Both groups are briefly considered in turn, with the reader referred to the following review articles for a more extensive overview [31,32,33,34,35].
Typically, the CAD system pipeline consists of four main steps, namely pre-processing, segmentation, feature extraction, and classification. Unlike DL methods, CAD methods reduce the large size of the data input to a relatively small number of explicitly defined features. In the context of ultrasound-imaged lesions, previous works [31,34] have shown that promising features are: texture features, such as Laws’ texture energy, local binary patterns, wavelet features, contrast of gray level values or gray level co-occurrence matrices [21,22,25,26,29,30]; morphological features, such as spiculation, depth-to-width ratio, elliptic-normalized circumference and skeleton, long axis-to-short axis ratio [20,21,22]; statistical-model-based features such as those based on the Nakagami or K-Distributions [19]; and finally, bioinspired or domain-knowledge-based descriptors, such as those describing shape, calcifications, posterior shadow and echo, or echo characteristics [23].
In contrast to traditional CAD methods, DL methods, which do not require predefined features, have experienced an upsurge in popularity in virtually all areas of image processing. Due to their considerable ability in approximating arbitrary functions, DL methods can potentially replace any number of the above steps in the traditional CAD pipeline, provided that a sufficient number of training data are made available. This requirement is usually met in the case of “grand challenges” using popular lesion types, where anyone can test their algorithm on publicly available datasets. Recent examples of the use of DL in ultrasound-based lesion diagnosis are to be found in the classification of thyroid nodules [27,28], breast lesion differentiation [24] and lung ultrasound for the detection of COVID-19-associated lesions [36].
Despite their understandable popularity, since DL methods are based on large interconnected neural networks, their explainability (and thus the predictability of failure cases) poses a challenge to their verification. Moreover, as hinted earlier, DL techniques require large datasets (typically on the order of 1000 recordings). Due to the current relative lack of large datasets for skin ultrasound, ultrasound-based skin lesion classification techniques mostly rely on traditional CAD systems. This is the topic of the next subsection.

1.3. Ultrasound-Based Differential Diagnosis of Benign and Malignant Skin Lesions

Regarding the topic of skin cancer, the reader is first directed to two reviews of skin cancer detection methods in general [37,38], followed by two reviews of ultrasound-based skin cancer diagnosis [39,40]. These reviews of ultrasound methods highlight a number of studies where a number of quantitative and semi-quantitative parameters—such as echogenicity, homogeneity, shape, margins and location of the lesions, as well as the posterior acoustic shadow and dermal echogenicity ratio—are shown to be promising features in differential diagnosis [41,42,43,44,45,46,47]. These works, however, do not aim to provide automated classification, as they require fully manual segmentation and examine the diagnostic potential of features rather than combining them into a classification framework. Therefore, lesion classification accuracy values are either missing, not detailed properly, or do not reach the desired level (60%+) [40].
A number of more recent studies have moved towards providing a skin cancer classification framework. Csabai et al. [48] and Andrékuté et al. [49] combined an semi-automated (SA) segmentation method with a fully-automatic feature extraction and classification method based on acoustical, textural and shape features. Csabai et al. [48] examined three kinds of lesion types, namely MMs, BCCs and benign nevi. Using a manually selected seeding region, an active contour model (ACM) was used to segment the lesion. Five shape features and seven first-order texture parameters were defined and their mean and standard deviation were input as features into a support vector machine (SVM) model. In terms of the area under the receiver operating characteristic (ROC) curve (AUC) metric, they reported a classification performance of 86% for the differentiation between nevi from cancerous lesions and 90% for BCC vs. nevi. Andrékuté et al. [49] used a somewhat different approach: following a manual selection of those A-lines that contained the lesion, the lesion boundaries were automatically calculated for each A-line independently (A-lines are one-dimensional sections of the B-mode image in the depth direction, the direction of propagation of the ultrasound pulse). From these A-lines, 29 features were extracted for binary classification between MMs and benign melanocytic skin tumors (MST). An AUC performance of 89.0 ± 0.6 % was obtained.
In another strand of research, Kia et al. [50] presented an automatic classification method for differentiating between healthy tissues, benign lesions, BCCs and melanomas. Although 98% sensitivity was attained, this was achieved at the cost of the specificity being only 5%, making the diagnostic value of the algorithm extremely limited. (Although, judging from the context in which the performance values were reported, it is possible the authors may have meant to write a specificity of 95%.) In addition, healthy skin without lesions was included in the testing set, making a comparison with other classification articles difficult, if they do not consider the differentiation of lesion-free skin necessary. A more recent work from the group is based on tissue frequency analysis [51]. It uses a 384-element-long feature vector from frequency space to train the above-mentioned neural network. The algorithm calculated the features using the whole sonograms without applying any kind of segmentation before it. Their work reached an accuracy of 95.9% using 220 malignant and 180 benign lesions for training, testing and evaluation; a very promising result. However, since the report did not specify the four classes of skin type for which differential diagnosis was performed, more reports from their work are needed before the method can be verified.
A final work worthy of mention is that of Tiwari et al. [5], where skin lesion classification is performed based on parameters collected and combined from a number of different imaging modalities, namely ultrasonograpy, dermatoscopy and spectrophotometry. Although the results are outstanding (with an AUC of 99.9 % ), purely ultrasound-based classification is unfortunately not proposed or evaluated in the work.
In the table below (Table 1), the performance of the most relevant ultrasound-based classification methods is presented. They have been selected from the literature included above on the basis of having clearly defined and documented performance measures differentiating between skin lesions only.

1.4. Aims of Current Work

The current work aims to present a framework for ultrasound-based skin cancer diagnosis that differentiates between three common skin lesion types: benign nevi, BCC and MM. In contrast to techniques that require some form of manual segmentation, the use of an automated segmentation method [52] makes skin cancer detection fully automatic, which, considering the time limits imposed on dermatological visits [53], would significantly improve the utility of the skin cancer detection method. Although many methods exist for optical-based automated skin cancer diagnosis [3,4], as mentioned earlier, ultrasound has the potential to improve on the accuracy of fully optical-based methods [5]. The aim of the current work, therefore, is to assess how an fully-automated (FA) skin lesion detection method compares with reference SA methods and to compare with relevant results in the literature.

2. Materials and Methods

In the current section, the data processing pipeline is described from the point of data collection to data processing using feature extraction, classification and, finally, measures to evaluate classification performance. All the code used in the work is available to download on GitHub (Available online: https://github.com/marosanp/skin-lesion-us, accessed on 20 June 2021).

2.1. Ultrasound Data Collection

Data were collected at the Department of Dermatology, Venereology and Dermatooncology, Semmelweis University, Budapest, Hungary, as part of an ethically approved study. Informed consent was obtained from the participating patients for the anonymised use of the data for research and publication [52].
The source of the examined dataset was a commercial high-frequency ultrasound imager (HI VISION Preirus with 5–18 MHz EUP-L75 transducer connected to Hitachi Preirus, Hitachi, Tokyo, Japan). The current study involved N = 310 B-mode ultrasound images, containing skin lesions with a thickness of 1–2 mm. Three different types of skin lesions were distinguished, including 110 benign nevi, 130 BCCs and 70 recordings of MMs. Figure 1 illustrates a representative ultrasound image of each examined lesion type.

2.2. Segmentation

Three different segmentation techniques were implemented and compared in the current study: one was an FA algorithm to study the accuracy of ultrasound-based FA skin cancer detection, while the other two were SA algorithms used as a reference. Note that the same segmentation method was used for training and testing rather than selecting one as the ground truth during training.
The first technique performs FA lesion segmentation based on an initial seeding step and a growing step, described in detail in [52], and described briefly below. The seeding step begins with a pre-processing substep to make ultrasound images from different machines similar to each other. This is followed by layer extraction substeps (above skin, epidermis, and dermis), and a lesion detection substep within the dermis. Each layer extraction substep first performs an intensity-based clustering or multilevel thresholding method combined with prior geometric information to return an initial estimate of the layer region; then performs a refinement of the region estimate based on an ACM and morphological operations. The seeding step is concluded by a lesion detection substep that incorporates information from the layer extraction substeps with prior geometrical assumptions about the arrangement of the layers and lesions. The automated seeding is followed by a growing step, using ACMs to extract the final lesion mask. To the best of our knowledge, this is the only fully automatic segmentation algorithm for ultrasound images of skin cancer-suspicious lesions that works on images from multiple imaging systems [52].
Two SA segmentation algorithms were also used for comparison purposes. Both of them require manual seeding for lesion localization and execute an ACM-based growing step on the initial seed masks for final boundary delimitation. The first, freehand seeding method used a freehand drawing around the lesion borders (using the MATLAB command freehand). This freehand seeding method simulated a careful manual segmentation since it allows any errors to be corrected using an ACM method. The second seeding method generated the largest area rectangle (LAR) from the freehand drawing and used that as the seed. This is similar to someone choosing a rectangular seed, as found in other works [54,55], and preferred in practice to freehand seeding due to the higher selection speed involved. We also have to mention that freehand seeding adds a significant variance and impairs the reproducibility of the algorithm as was demonstrated in [52]. In the current implementation, the difference was that the LAR was derived from the freehand seeding itself to allow meaningful comparison between the two.
The above-defined three segmentation techniques were chosen based on the following considerations. Our primary motivation of the work is to compare FA classification with SA classification. To the best of our knowledge, our previous work [52] is the only FA seeding-based method that was applied on more than one type of skin ultrasound image, demonstrating its robustness; therefore, this was chosen as the FA segmentation algorithm. Considering reference SA segmentation methods, freehand-seeding-based SA segmentation was considered a good approximation to ground truth, combining human knowledge with the filling in of spatially fine details. Since the freehand method is highly time-consuming, the LAR-based SA segmentation is considered a good simulation of a rapid human input into the segmentation workflow [52]. Since the freehand method was considered to be the most reliable segmentation, this was treated as the ground truth segmentation. In particular, the success rate of the FA segmentation was defined as the proportion of lesions where the Dice coefficient between the FA and freehand SA exceeded 10%.

2.3. Feature Extraction

Using the segmentation described in the previous subsection, 93 features were extracted from the lesion, lesion boundary, and the area of the dermis under the lesion. Then, feature selection was made by examining the SVM-based weights on the training set. In this way, 62 features were selected from 93 examined (Table 2).
The 62 features were calculated using the following image regions:
  • Lesion region: all the pixels inside the lesion mask;
  • Dermis region: pixels of the region of the dermis being right under the lesion mask;
  • Lesion boundary: a lane of pixels being located within a fixed distance from the lesion mask boundary.
The features can be grouped into first-order textural, second-order textural, and shape features. First-order textural features express information about the distribution of individual pixel intensity values, while second-order textural features express spatial correlation between pixel intensities [56], and, in the current work, are based on the gray level co-occurrence matrix (GLCM) [57,58,59]. Lastly, some shape-based features were also extracted. All three groups of features are presented below.

2.3.1. First-Order Textural Features

First-order textural features can be broadly categorized according to the properties or regions concerned, hence the subgroups; attenuation, lesion contrast, lesion boundary, and statistical.
Attenuation-based features, such as attenuation, contrast of attenuation and heterogeneity of attenuation, examine the lesion region and its shadowing in the dermis region right under the lesion mask. Contrast parameters, such as lesion- contrast-based heterogeneity and the mean lesion contrast, examine the contrast of each boundary line emanating radially from the inner edge of the lesion [48]. The above two subgroups, as well as the mean boundary belonging to the boundary subgroup, were adopted from Csabai et al. [48].
Lesion boundary region-based features, such as mean boundary, boundary heterogeneity, boundary contrast, boundary heterogeneity contrast, boundary-lesion contrast, dermis-lesion heterogeneity contrast and boundary-lesion heterogeneity contrast, are calculated based on the expressions presented in Table 2. Statistical features, such as skewness, kurtosis and entropy, were also selected.

2.3.2. Second-Order Textural Features

For most of the second-order (GLCM) textural descriptors (contrast, correlation I. correlation II. dissimilarity, maximum probability, difference variance, difference entropy and information measure of correlation I.), the descriptors were calculated in both the vertical and horizontal directions, for both the lesion region and the dermis region of the images. For some of the second-order descriptors (energy, entropy, homogeneity I & II, and information measurement of correlation II) only the vertical GLCMs were calculated for both regions.
Further details and calculations of the above listed GLCM-based co-occurrence texture statistics are to be found in the work of Uppulari [59].

2.3.3. Shape Features

Shape features are derived from parameters describing the shape of the lesion boundary. Some shape parameters, such as standard deviation of curvature and circularity, were extracted by Csabai et al. [48]. Further shape features, such as axis ratio, perimeter-area ratio and compactness (ratio of perimeter and the length of the major axis of the lesion mask), are also extracted.
The feature names with corresponding indices in the feature set (idx) are presented in Table 2, with references (to those taken from the literature) or with a short description (in the cases of newly introduced descriptors).

2.4. Classification

To aid comparison, the approach for the classification methodology (training and testing) closely resembled that of Andrékuté et al. [49]. Namely, an SVM-based classifier [60] was used, with ten-fold classification implemented in the following manner. First, ten separate groups were selected randomly, with the same ratios of lesion types. Then, each group in turn was selected as the test group, with the remaining nine groups merged into a training set. The classification performance was then averaged over the ten training instances, using accuracy (ACC) and AUC as performance metrics for the binary classification. In addition to binary classification (as implemented by Andrékuté et al. [49]), multiclass classification was also carried out in the current work and evaluated using ACC.
The following classes were distinguished for both binary and multiclass classification: ‘Nevus vs. others’, ‘MM vs. others’, ‘BCC vs. others’, ‘Nevus vs. BCC’, ‘Nevus vs. MM’, ‘BCC vs. MM’. In the cases of the binary classifications, ‘Nevus vs. BCC’, ‘Nevus vs. MM’ and ‘BCC vs. MM’ classifications were performed on datasets containing only lesions from the relevant two groups, while the corresponding multiclass classification results were obtained using datasets containing all three lesion types. ‘Nevus vs. others’, ‘MM vs. others’ and ‘BCC vs. others’ classifications were performed using datasets containing all three types of lesions in both binary and multiclass cases. Here, the binary classification training was conducted using two classes (a certain type of lesion versus all other lesions), while the multiclass classifications used three lesion type classes for training.
The parameter set (Algorithm 1) and the entire workflow (Algorithm 2) of the current work is presented as an algorithm in pseudo-code form at the end of the section, focusing on the details of the classification method. Details of the FA segmentation algorithm are presented in Marosán et al. [52], where figures of the procedure are also provided. Feature extraction is detailed in Table 2 and Section 2.3, Feature extraction.
Algorithm 1: Skin lesion classification algorithm—Part I. Variables and Methods.
Inputs
N//where N is the number of ultrasound images collected
I 1 , I 2 , I N // I j is the jth image collected (images are indexed by j)
L 1 , L 2 , L N // L j is the lesion class of the jth image

Outputs
C 1 , C 2 , C N , where C j { N , B C C , M M , O } // C j is the predicted class label of the jth image (nevus, BCC, MM, other)

Parameters
S e g //segmentation types: {FA, SA LAR, SA Freehand}
C l a s s //classification types: {‘Nevus vs. BCC’, ‘Nevus vs. MM’, ‘BCC vs. MM’, ‘Nevus vs. others’, ‘MM vs. others’, ‘BCC vs. others’}
m a s k L 1 , m a s k L 2 , m a s k L N // m a s k L j is the lesion mask of the jth image
m a s k D 1 , m a s k D 2 , m a s k D N // m a s k D j is the dermis mask of the jth image
F 1 , F 2 , F N // F j is the feature vector of the jth image
F 1 G , F 2 G , F 10 G // F k G is the kth group of feature vectors
C 1 G , C 2 G , C 10 G // C k G is the kth group of label predictions of feature vectors

Methods
s e g m e n t i ( ) //segmentation of lesion and dermis masks
f e a t u r e _ e x t r a c t i o n ( ) //first and second order textural & shape feature extraction, based on dermis and lesion crops
g r o u p i n g ( ) //selects randomly 10 groups, with the same ratios of lesion types
l a b e l i n g ( ) //image labeling based on the following classification types: ‘Nevus vs. others’, ‘MM vs. others’, ‘BCC vs. others’, ‘Nevus vs. BCC’, ‘Nevus vs. MM’, ‘BCC vs. MM’
s v m _ t r a i n ( ) //train SVM model, based on feature vector & label pair set
s v m _ p r e d i c t ( ) //predict label for feature vector, using pre-defined SVM model
e v a l u a t e ( ) //compute ACC & AUC values based on pre-defined and predicted labels

2.5. Performance Metrics

The evaluation of classification methods was carried out using the performance metrics listed below. Numerous works have applied [48,49] the following metrics previously.
The sensitivity S e n s calculates the proportion of positive cases that are correctly detected:
S e n s = T P T P + F N ,
with T P , F N denoting the number of true positives and false negatives in the classification, respectively.
Algorithm 2: Skin lesion classification algorithm—Part II. Procedure.
Diagnostics 11 01207 i001
The specificity S p e c calculates the proportion of negative cases that are correctly detected:
S p e c = T N T N + F P ,
with T N , F P denoting the number of true negatives and false positives in the classification, respectively.
The accuracy A C C describes the classification accuracy, namely, the ratio of the number of correctly detected cases to the number of all examined cases
A C C = T N + T P T N + T P + F N + F P .
The receiver operating characteristic R O C curve is a graphical plot that displays how S e n s varies with 1 S p e c . Considering R O C as a function, its definition can be stated as
S e n s = R O C 1 S p e c .
Finally, using the above definition for the R O C , the area under the ROC curve A U C can be defined as:
A U C = 0 1 R O C ( x ) d x ,
where x = 1 S p e c .

3. Results and Discussion

3.1. Overview of Classification Performance

The ROC curves of the various binary classifications for all three segmentation methods are presented in Figure 2. From these curves, the AUC can be calculated, the values of which are summarized in Table 3.
In general, nevi are well distinguished from cancerous lesions: with one exception, the relevant classifications (‘Nevus vs. others’, ‘Nevus vs. BCC’, ‘Nevus vs. MM’) yield AUC values over 90% (highlighted in bold in the table).
Although the FA method generally fares worse than the SA methods, the performance is always within 5% of the best performing method.
The ACC values for both binary and multiclass classifications are summarized in Table 4. The table depicts a similar trend to the AUC results in Table 4: the nevi are generally well distinguished from cancerous lesions, with most such classifications reaching accuracies above 80% (highlighted in bold in the table). The notable exception is the multiclass classification of ’Nevus vs. MM’, which could be, in part, due to the relatively low number (N = 70) of melanoma recordings.
It should also be noted that multiclass classifications generally show a worse performance than binary classifications. This is arguably because multiclass classification forces the training classes to be smaller, outweighing the advantage that arises from being able to train on adequately distinct classes separately.
As before, the FA method generally shows a poorer performance; however, with one exception (‘BCC vs. others’, multiclass), the difference from the best method is never worse than 5%.
The current results also compare favourably to the results of Andrékuté et al. [49], where nevi were distinguished from melanoma with an accuracy of 82.4%: with the SA methods, the binary classification achieved an accuracy of 85.0%, while the FA method also achieved a comparable 81.1%.
For a comparison with other relevant works (all using SA segmentation), please see Table 1. Although the comparison is more difficult with other works, since they mostly set the sensitivity to 100% and then observe specificity, the current work nevertheless seems to fare well. For example, for those works where AUC values are provided, the current work is always superior in its corresponding AUC values. The specificity at 100% sensitivity shows more variable results in the current work.
Overall, the classification of nevi from cancerous lesions, and in particular from both cancerous lesion types (‘Nevus vs. others’), provides the best classification performance. Using multiclass classification, the FA method can distinguish nevi from cancerous lesions with an accuracy of over 85%.

3.2. Comparison of FA and SA Classification Performance with Representative Images

The reason behind a difference in classification performance can be a direct result of differences in segmentation during the classification phase; or it can be an indirect result of training on differently segmented images. For the FA method, 83.5% of the lesions were detected correctly, making its classification performance being close to that of the other methods somewhat surprising. To try and offer putative explanations for differences between classification performances, a presentation of images (with overlaid segmentation contours) that led to different classifications could be informative. In order to be consistent, the classification discussed will be the best performing one, namely ‘Nevus vs. others’. Before proceeding to discuss the images, it should be noted for context that, for this classification task, all three methods were jointly successful in 73.6% of the images, all three failed together in 5.2% of the cases, while there were differing classifications in 21.3% of the cases. In the following subsections, those cases where differing classifications are given are considered.

3.2.1. Cases When FA Fails While SA Methods Perform Correctly

In most cases of differing classifications, the SA methods succeeded while the FA method failed. Figure 3 presents examples for the three trends observed when analyzing this subset. In some cases, the FA segmentation method detected an image region fully outside the real lesion. As can be seen in Figure 3a, there are cases in which the image structure can be misleading when searching for the lesion location. In other cases, the FA segmentation did not detect parts of the lesion (Figure 3b) or included additional areas that were not part of the lesion (Figure 3c). In the latter case, the example shows a shadowing effect in the dermis region next to the lesion, which potentially misled the FA segmentation method.

3.2.2. Cases When the Two SA Methods Return Different Classifications

The SA segmentation methods, based on freehand and LAR segmentation, produced similar results generally; however, in certain cases, they led to different lesion classification results, such as in the cases presented in Figure 4. In the minority (29%) of such cases (2.5% of all images), the LAR-based method was the one failing the classifications. These were due to the shape of the lesion being such that the LAR segmentation could cover only a small portion of the lesion, which could not be expanded to cover the entire lesion even with the subsequent ACM method (Figure 4a).
In the majority (71%) of the cases discussed here (6.3% of all images), the freehand-based method led to failing classifications while the LAR produced correct results. These were presumably due to human error in the freehand segmentation, since the pixel boundaries of the freehand lesion segmentation are relatively arbitrary in contrast to the LAR-based method, where the ACM model finds the boundaries of the lesion with a higher precision (Figure 4b).

3.2.3. Cases When the SA Methods Both Fail While the FA Method Performs Correctly

In some cases, the ‘Nevus vs. others’ type classification was correct for the FA segmentation while failing for both SA segmentations. Figure 5 presents notable cases for such segmentations. Figure 5a shows a case in which the FA segmentation detected a similarly shaped and sized but slightly shifted region from that detected by the two SA segmentations. In some cases, such as the one presented in Figure 5b, the FA segmentation detected additional image regions as part of the lesion in comparison to the results of the SA segmentations, leading to a correct classification result.
In the case shown in Figure 5c, all three segmentation results matched closely; however, slight differences in their borders led to different classification results. This example emphasizes the significance of small details in automated ultrasound-image-based lesion classification performance.

3.3. Sensitivity of Classification to Changes in Lesion Segmentation

The last example in the previous subsection (Figure 5c) presented an interesting case, since it showed that slight changes in the border of the segmented lesion could lead to different classification results. Similar issues of classification sensitivity have been addressed elsewhere in the computer vision literature [61]; however, it is a challenging topic to address, partly due to the large search space involved in the sensitivity analysis. In the current work, the topic has been partially addressed in the following manner.
The ultrasound image on Figure 5 depicting a nevus was chosen as the target of the investigation, since this is where the phenomenon of changing classification due to a small change in lesion border was observed. Taking the freehand-based SA segmentation as the reference, the region was progressively grown/shrunk at various regions of the border (right edge, bottom edge, right and bottom, entire border), with the classification noted at each step.
Figure 6 shows the results. Figure 6b shows that if the segmented region is shrunk only at the right edge, then a modest shrinkage of only 2 pixels’ width is able to change the classification from wrong to correct. This is consistent with the behaviour previously noted in Figure 5c, where the correct FA classification has a segmentation that is slightly shrunk at the right edge compared to the other two (Figure 6a). Interestingly, where other sides of the segmented regions are also modified, it becomes more difficult for the classification to be corrected. This could be because a hyperechoic patch on the right edge of the lesion perturbs the classification, and the segmentation is otherwise correct.
Although a more systematic analysis is beyond the scope of the current paper, the above preliminary analysis does show some insight into the relatively small yet significant perturbations in classification that can occur due to variabilities in segmentation.

3.4. Feature Performance

The aim of this section is to evaluate the features used in the classification framework by identifying those with the largest contribution to the classification performance. Since in the SVM classifier model the features do not act in isolation but are part of a non-linear system where features support each other, the performance of a feature was deemed to be better evaluated by measuring how much performance degraded when it was left out, rather than the performance it achieves on its own. Thus, the contribution of each of the 62 selected features was examined one-by-one as follows: one feature was removed and binary classification was performed using the remaining 61 features and using LAR-based segmentation. The AUC and ACC performance metrics were computed for all 62 cases for the six types of classification (similarly to Table 3 and Table 4 earlier when all features had been used). The top features—the absence of which caused the largest deterioration in classification performance—are shown in Table 5, with the worst feature also shown for reference. By removing one of the top features, the accuracy performance dropped by around 2 to 5%.
Table 5 shows several features that performed well on different classification cases, regarding both AUC and ACC performance measures. Those features that appear at least three times in the top four are in bold and are as follows (with feature indices in parentheses as listed in Section 2.3, Feature extraction): boundary heterogeneity (7); boundary contrast (8); skewness (13); entropy (15); axis ratio (18); compactness (20); and difference variance (26). Of these features, four are first-order textural features (7,8,13,15), two are shape-based features (18,20), and one is a second-order textural feature (26). When considering the possibility of applying the trained classifier to skin ultrasound images from other devices, the two shape-based features are arguably the most transposable; with the other features, some form of domain adaptation may be required [62]. In both the ‘Nevus vs. others’ and ‘Nevus vs. MM’ classification tasks, shape-based features were prominent, in agreement with the results of [49], where shape-based features provided the highest classification performance for distinguishing between nevi and melanoma.

3.5. Runtime Measurements

While the segmentation seeding algorithm was implemented in Python, the growing step, feature extraction, classification and evaluation were implemented in MATLAB R2018b (MathWorks, Inc., Natick, MA, USA). Table 6 details the computational cost of the proposed algorithm. Each ultrasound image had a size of 900 × 400 pixels and the computer used had an Intel Core i7-7500U CPU (2.70 GHz) processor and 16 GB RAM.

4. Conclusions

An automated framework for skin lesion classification was presented and an FA and two SA segmentation methods were compared. Ultrasound images of three types of lesions were used for the classifications: MMs, BCCs and benign nevi. Both binary and multiclass classifications were performed and evaluated, using six types of class differentiations, for all three segmentation methods, separately.
The best results were obtained generally for ‘Cancerous vs. Non-cancerous’ (‘Nevus vs. others’, ‘Nevus vs. BCC’, ‘Nevus vs. MM’) type binary classifications (>90% AUC). The two SA methods produced generally better results than the FA method, but with relatively slight differences (between 0.7–3.9%), such that the FA also provided an AUC at around 91% for the binary classification between nevi and the two cancerous lesion types. The achieved accuracies were similar to those obtained by Andrékuté et al. [49] when they differentiated between nevi and melanoma: 85.0% with the SA methods and 81.1% with the FA method, compared with 82.4%. The classification of nevi from cancerous lesions had even higher accuracies of above 85% even with the FA method. The result demonstrates the viability of FA skin cancer classification from ultrasound images.
Since features can be highly dependent on the type of ultrasound image they are applied to, it is worth noting that, in the case of the best performing classification task of distinguishing nevi from cancerous lesions, the top two performing features in terms of accuracy were shape-based features, since such features are more adaptable for different types of ultrasound images. Nevertheless, future work should focus on applying domain adaptation techniques to ensure the classification framework here presented can also be applied to skin ultrasound images produced by other devices.

Author Contributions

Conceptualization, P.M.-V., K.S., A.H., D.C. and M.G.; methodology, P.M.-V., A.H., D.C., K.F. and M.G.; software, P.M.-V., D.C. and K.F.; validation, P.M.-V., K.S., A.H., G.C. and M.G.; formal analysis, P.M.-V., A.H., G.C. and M.G.; investigation, P.M.-V., A.H., G.C. and M.G.; resources, P.M.-V., K.S., D.C. and G.C.; data curation, P.M.-V., K.S., D.C. and G.C.; writing—original draft preparation, P.M.-V., G.C. and M.G.; writing—review and editing, P.M.-V., G.C. and M.G.; visualization, P.M.-V.; supervision, M.G. and A.H.; project administration, M.G.; funding acquisition, M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was in part supported by Pázmány Péter Catholic University KAP programmes. The APC was funded by Pázmány Péter Catholic University. The collaboration with Jedlik Innováció Kft under the GINOP-2.1.7-15-2016-02201 programme is gratefully acknowledged.

Institutional Review Board Statement

The acquisition of the ultrasound images was conducted according to the guidelines of the Declaration of Helsinki, and approved by the competent authority National Institute of Pharmacy and Nutrition, Budapest, Hungary (protocol code OGYÉI/16798/2017).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the data acquisition.

Acknowledgments

The authors thank the generosity of Miklós Sárdy (director) and Sarolta Kárpáti (former director) of the Department of Dermatology, Venereology and Dermatooncology of Semmelweis University) for their scientific support and encouragement of the current research. The authors also thank Jan D’hooge, from Department of Cardiovascular Sciences, KU Leuven, for advising on the second-order textural features.

Conflicts of Interest

Some of the co-authors are employed by and hold financial interest in Dermus Kft. The company may plan to use the scientific results in its own research and development efforts. Other than that, the authors have no potential conflicts of interest to disclose.

Abbreviations

The following abbreviations are used in this manuscript:
FAfully-automated
SAsemi-automated
SVMSupport Vector Machine
ROCreceiver operating characteristic
AUCarea under the curve
ACCclassification accuracy
MMmalignant melanoma
BCCbasal cell carcinoma
CADComputer-Aided Diagnostic
DLdeep learning
MSTmelanocytic skin tumor
NANot Available
LARlargest area rectangle
ACMactive contour model
GLCMgray level co-occurence matrix

References

  1. American Cancer Society, Cancer Facts & Figures 2021. Available online: https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/annual-cancer-facts-and-figures/2021/cancer-facts-and-figures-2021.pdf (accessed on 21 May 2021).
  2. Apalla, Z.; Lallas, A.; Sotiriou, E.; Lazaridou, E.; Ioannides, D. Epidemiological trends in skin cancer. Dermatol. Pract. Concept. 2017, 7. [Google Scholar] [CrossRef] [Green Version]
  3. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  4. Ha, Q.; Liu, B.; Liu, F. Identifying Melanoma Images using EfficientNet Ensemble: Winning Solution to the SIIM-ISIC Melanoma Classification Challenge. arXiv 2020, arXiv:2010.05351. [Google Scholar]
  5. Tiwari, K.A.; Raišutis, R.; Liutkus, J.; Valiukevičienė, S. Diagnostics of Melanocytic Skin Tumours by a Combination of Ultrasonic, Dermatoscopic and Spectrophotometric Image Parameters. Diagnostics 2020, 10, 632. [Google Scholar] [CrossRef] [PubMed]
  6. Edwards, C.; Al-Aboosi, M.; Marks, R. The use of A-scan ultrasound in the assessment of small skin tumours. Br. J. Dermatol. 1989, 121, 297–304. [Google Scholar] [CrossRef] [PubMed]
  7. Cammarota, T.; Pinto, F.; Magliaro, A.; Sarno, A. Current uses of diagnostic high-frequency US in dermatology. Eur. J. Radiol. 1998, 27, S215–S223. [Google Scholar] [CrossRef]
  8. Raju, B.I.; Swindells, K.J.; Gonzalez, S.; Srinivasan, M.A. Quantitative ultrasonic methods for characterization of skin lesions in vivo. Ultrasound Med. Biol. 2003, 29, 825–838. [Google Scholar] [CrossRef]
  9. Uhara, H.; Hayashi, K.; Koga, H.; Saida, T. Multiple Hypersonographic Spots in Basal Cell Carcinoma. Dermatol. Surg. 2007, 33, 1215–1219. [Google Scholar] [CrossRef]
  10. Samimi, M.; Perrinaud, A.; Naouri, M.; Maruani, A.; Perrodeau, E.; Vaillant, L.; Machet, L. High-resolution ultrasonography assists the differential diagnosis of blue naevi and cutaneous metastases of melanoma: Ultrasonography for differential diagnosis of naevi and melanoma metastases. Br. J. Dermatol. 2010, 163, 550–556. [Google Scholar] [CrossRef]
  11. Machet, L.; Samimi, M.; Georgesco, G.; Mourtada, Y.; Naouri, M.; Marc, J.; Ossant, F.; Patat, F.; Vaillant, L. High Resolution Ultrasound Imaging of Melanocytic and Other Pigmented Lesions of the Skin. Ultrasound Imaging 2011. [Google Scholar] [CrossRef] [Green Version]
  12. Wortsman, X. Sonography of the Primary Cutaneous Melanoma: A Review. Radiol. Res. Pract. 2012, 2012, 814396. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Mandava, A.; Ravuri, P.; Konathan, R. High-resolution ultrasound imaging of cutaneous lesions. Indian J. Radiol. Imaging 2013, 23, 269. [Google Scholar] [CrossRef] [PubMed]
  14. Piotrzkowska-Wroblewska, H.; Litniewski, J.; Szymanska, E.; Nowicki, A. Quantitative Sonography of Basal Cell Carcinoma. Ultrasound Med. Biol. 2015, 41, 748–759. [Google Scholar] [CrossRef]
  15. Dybiec, E.; Pietrzak, A.; Adamczyk, M.; Michalska-Jakubus, M.; Wawrzycki, B.; Lotti, T.; Rutkowski, P.; Krasowska, D. High frequency ultrasonography of the skin and its role as an auxillary tool in diagnosis of benign and malignant cutaneous tumors—A comparison of two clinical cases. Acta Dermatovenerol. Croat. ADC 2015, 23, 43–47. [Google Scholar] [PubMed]
  16. Bard, R.L. High-Frequency Ultrasound Examination in the Diagnosis of Skin Cancer. Dermatol. Clin. 2017, 35, 505–511. [Google Scholar] [CrossRef]
  17. Wortsman, X. Atlas of Dermatologic Ultrasound; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  18. Catalano, O.; Roldán, F.A.; Varelli, C.; Bard, R.; Corvino, A.; Wortsman, X. Skin cancer: Findings and role of high-resolution ultrasound. J. Ultrasound 2019, 22, 423–431. [Google Scholar] [CrossRef]
  19. Shankar, P.M.; Dumane, V.; Reid, J.M.; Genis, V.; Forsberg, F.; Piccoli, C.W.; Goldberg, B.B. Classification of ultrasonic B-mode images of breast masses using Nakagami distribution. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2001, 48, 569–580. [Google Scholar] [CrossRef]
  20. Chen, C.M.; Chou, Y.H.; Han, K.C.; Hung, G.S.; Tiu, C.M.; Chiou, H.J.; Chiou, S.Y. Breast lesions on sonograms: Computer-aided diagnosis with nearly setting-independent features and artificial neural networks. Radiology 2003, 226, 504–514. [Google Scholar] [CrossRef]
  21. Singh, B.K.; Verma, K.; Thoke, A. Fuzzy cluster based neural network classifier for classifying breast tumors in ultrasound images. Expert Syst. Appl. 2016, 66, 114–123. [Google Scholar] [CrossRef]
  22. Prabusankarlal, K.M.; Thirumoorthy, P.; Manavalan, R. Assessment of combined textural and morphological features for diagnosis of breast masses in ultrasound. Hum. Centric Comput. Inf. Sci. 2015, 5, 1–17. [Google Scholar] [CrossRef] [Green Version]
  23. Sahiner, B.; Chan, H.P.; Roubidoux, M.A.; Hadjiiski, L.M.; Helvie, M.A.; Paramagul, C.; Bailey, J.; Nees, A.V.; Blane, C. Malignant and benign breast masses on 3D US volumetric images: Effect of computer-aided diagnosis on radiologist accuracy. Radiology 2007, 242, 716–724. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Han, S.; Kang, H.K.; Jeong, J.Y.; Park, M.H.; Kim, W.; Bang, W.C.; Seong, Y.K. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys. Med. Biol. 2017, 62, 7714. [Google Scholar] [CrossRef] [PubMed]
  25. Acharya, U.R.; Faust, O.; Sree, S.V.; Molinari, F.; Suri, J.S. ThyroScreen system: High resolution ultrasound thyroid image characterization into benign and malignant classes using novel combination of texture and discrete wavelet transform. Comput. Methods Programs Biomed. 2012, 107, 233–241. [Google Scholar] [CrossRef] [PubMed]
  26. Acharya, U.R.; Chowriappa, P.; Fujita, H.; Bhat, S.; Dua, S.; Koh, J.E.; Eugene, L.; Kongmebhol, P.; Ng, K.H. Thyroid lesion classification in 242 patient population using Gabor transform features from high resolution ultrasound images. Knowl. Based Syst. 2016, 107, 235–245. [Google Scholar] [CrossRef]
  27. Chi, J.; Walia, E.; Babyn, P.; Wang, J.; Groot, G.; Eramian, M. Thyroid nodule classification in ultrasound images by fine-tuning deep convolutional neural network. J. Digit. Imaging 2017, 30, 477–486. [Google Scholar] [CrossRef] [PubMed]
  28. Sharifi, Y.; Bakhshali, M.A.; Dehghani, T.; Danai Ashgzari, M.; Sargolzaei, M.; Eslami, S. Deep learning on ultrasound images of thyroid nodules. Biocybern. Biomed. Eng. 2021, 41, 636–655. [Google Scholar] [CrossRef]
  29. Virmani, J.; Kumar, V.; Kalra, N.; Khandelwal, N. SVM-based characterization of liver ultrasound images using wavelet packet texture descriptors. J. Digit. Imaging 2013, 26, 530–543. [Google Scholar] [CrossRef] [Green Version]
  30. Acharya, U.R.; Fujita, H.; Sudarshan, V.K.; Mookiah, M.R.K.; Koh, J.E.; Tan, J.H.; Hagiwara, Y.; Chua, C.K.; Junnarkar, S.P.; Vijayananthan, A.; et al. An integrated index for identification of fatty liver disease using radon transform and discrete cosine transform features in ultrasound images. Inf. Fusion 2016, 31, 43–53. [Google Scholar] [CrossRef]
  31. Huang, Q.; Zhang, F.; Li, X. Machine learning in ultrasound computer-aided diagnostic systems: A survey. BioMed Res. Int. 2018, 2018, 5137904. [Google Scholar] [CrossRef]
  32. Wang, B.; Liu, M.; Zhu, M.; Eisenbrey, J. Artificial intelligence in ultrasound imaging: Current research and applications. Dep. Radiol. Fac. Pap. 2019, 75, 53–61. [Google Scholar]
  33. Zhou, B.; Yang, X.; Liu, T. Artificial Intelligence in Quantitative Ultrasound Imaging: A Review. arXiv 2020, arXiv:2003.11658. [Google Scholar]
  34. Junaid, M.J.A.; Kumar, R. Data Science is all Set to Revolutionize the Ultrasound Diagnosis in Medical Health Care. In Proceedings of the 2020 International Conference on Computation, Automation and Knowledge Management (ICCAKM), Dubai, United Arab Emirates, 9–10 January 2020; pp. 73–78. [Google Scholar]
  35. Wang, S. Applications of Automated Machine Learning Diagnosis in Medical Ultrasound. Master’s Thesis, Drexel University, Philadelphia, PA, USA, 2020. [Google Scholar]
  36. McDermott, C.; Lacki, M.B.; Sainsbury, B.; Henry, J.; Filippov, M.; Rossa, C. Sonographic diagnosis of COVID-19: A review of image processing for lung ultrasound. Front. Big Data 2021, 4, 2. [Google Scholar] [CrossRef]
  37. Narayanamurthy, V.; Padmapriya, P.; Noorasafrin, A.; Pooja, B.; Hema, K.; Firus Khan, A.Y.; Nithyakalyani, K.; Samsuri, F. Skin cancer detection using non-invasive techniques. RSC Adv. 2018, 8, 28095–28130. [Google Scholar] [CrossRef] [Green Version]
  38. Heibel, H.D.; Hooey, L.; Cockerell, C.J. A Review of Noninvasive Techniques for Skin Cancer Detection in Dermatology. Am. J. Clin. Dermatol. 2020, 21, 513–524. [Google Scholar] [CrossRef] [PubMed]
  39. Bhatta, A.K.; Keyal, U.; Liu, Y. Application of high frequency ultrasound in dermatology. Discov. Med. 2018, 26, 237–242. [Google Scholar] [PubMed]
  40. Dinnes, J.; Bamber, J.; Chuchu, N.; Bayliss, S.E.; Takwoingi, Y.; Davenport, C.; Godfrey, K.; O’Sullivan, C.; Matin, R.N.; Deeks, J.J.; et al. High-frequency ultrasound for diagnosing skin cancer in adults. Cochrane Database Syst. Rev. 2018. [Google Scholar] [CrossRef]
  41. Harland, C.; Bamber, J.; Gusterson, B.; Mortimer, P. High frequency, high resolution B-scan ultrasound in the assessment of skin tumours. Br. J. Dermatol. 1993, 128, 525–532. [Google Scholar] [CrossRef]
  42. Dummer, W. Preoperative Characterization of Pigmented Skin Lesions by Epiluminescence Microscopy and High-Frequency Ultrasound. Arch. Dermatol. 1995, 131, 279. [Google Scholar] [CrossRef]
  43. Lassau, N.; Spatz, A.; Avril, M.F.; Tardivon, A.; Margulis, A.; Mamelle, G.; Vanel, D.; Leclere, J. Value of high-frequency US for preoperative assessment of skin tumors. Radio Graph. 1997, 17, 1559–1565. [Google Scholar] [CrossRef] [Green Version]
  44. Harland, C.; Kale, S.; Jackson, P.; Mortimer, P.; Bamber, J. Differentiation of common benign pigmented skin lesions from melanoma by high-resolution ultrasound. Br. J. Dermatol. 2000, 143, 281–289. [Google Scholar] [CrossRef]
  45. Clément, A.; Hoeffel, C.; Fayet, P.; Benkanoun, S.; Sahut D’izarn, J.; Oudjit, A.; Legmann, P.; Gorin, I.; Escande, J.; Bonnin, A. Value of high frequency (20 mhZ) and doppler ultrasound in the diagnosis of pigmented cutaneous tumors. J. Radiol. 2001, 82, 563–571. [Google Scholar] [PubMed]
  46. Bessoud, B.; Lassau, N.; Koscielny, S.; Longvert, C.; Avril, M.F.; Duvillard, P.; Rouffiac, V.; Leclère, J.; Roche, A. High-frequency sonography and color Doppler in the management of pigmented skin lesions. Ultrasound Med. Biol. 2003, 29, 875–879. [Google Scholar] [CrossRef]
  47. Rallan, D.; Bush, N.L.; Bamber, J.C.; Harland, C.C. Quantitative Discrimination of Pigmented Lesions Using Three-Dimensional High-Resolution Ultrasound Reflex Transmission Imaging. J. Investig. Dermatol. 2007, 127, 189–195. [Google Scholar] [CrossRef] [PubMed]
  48. Csabai, D.; Szalai, K.; Gyongy, M. Automated classification of common skin lesions using bioinspired features. In Proceedings of the 2016 IEEE International Ultrasonics Symposium (IUS), Tours, France, 18–21 September 2016; pp. 1–4. [Google Scholar] [CrossRef]
  49. Andrėkutė, K.; Linkevičiūtė, G.; Raišutis, R.; Valiukevičienė, S.; Makštienė, J. Automatic Differential Diagnosis of Melanocytic Skin Tumors Using Ultrasound Data. Ultrasound Med. Biol. 2016, 42, 2834–2843. [Google Scholar] [CrossRef]
  50. Kia, S.; Setayeshi, S.; Shamsaei, M.; Kia, M. Computer-aided diagnosis (CAD) of the skin disease based on an intelligent classification of sonogram using neural network. Neural Comput. Appl. 2013, 22, 1049–1062. [Google Scholar] [CrossRef]
  51. Kia, S.; Setayeshi, S.; Pouladian, M.; Ardehali, S.H. Early diagnosis of skin cancer by ultrasound frequency analysis. J. Appl. Clin. Med. Phys. 2019, 20, 153–168. [Google Scholar] [CrossRef]
  52. Marosán, P.; Szalai, K.; Csabai, D.; Csány, G.; Horváth, A.; Gyöngy, M. Automated seeding for ultrasound skin lesion segmentation. Ultrasonics 2021, 110, 106268. [Google Scholar] [CrossRef]
  53. Levell, N.J.; Jones, S.K.; Bunker, C.B. Dermatology. Royal College of Physicians. 2013. Available online: https://www.bad.org.uk/library-media/documents/consultant%20physicians%20working%20with%20patients%202013.pdf (accessed on 21 May 2021).
  54. Sciolla, B.; Cowell, L.; Dambry, T.; Guibert, B.; Delachartre, P. Segmentation of Skin Tumors in High-Frequency 3-D Ultrasound Images. Ultrasound Med. Biol. 2017, 43, 227–238. [Google Scholar] [CrossRef]
  55. Nguyen, K.L.; Delachartre, P.; Berthier, M. Multi-Grid Phase Field Skin Tumor Segmentation in 3D Ultrasound Images. IEEE Trans. Image Process. 2019, 28, 3678–3687. [Google Scholar] [CrossRef]
  56. Marosán, P. Detection of Myocardial Infarction in Echocardiograms. Master’s Thesis, Pázmány Péter Catholic University, Budapest, Hungary, 2016. [Google Scholar]
  57. Haralick, R.M.; Shanmugam, K.; Dinstein, I. Textural Features for Image Classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  58. Soh, L.K.; Tsatsoulis, C. Texture analysis of SAR sea ice imagery using gray level co-occurrence matrices. IEEE Trans. Geosci. Remote Sens. 1999, 37, 780–795. [Google Scholar] [CrossRef] [Green Version]
  59. Uppuluri, A. Avinash Uppuluri (2016). GLCM Texture Features, MATLAB Central File Exchange. Retrieved July. 2016. Available online: https://www.mathworks.com/matlabcentral/fileexchange/22187-glcm-texture-features (accessed on 20 June 2021).
  60. Manthiri, A.S. Avinash Uppuluri (2017). Multi Class SVM, MATLAB Central File Exchange. Retrieved July. 2017. Available online: https://www.mathworks.com/matlabcentral/fileexchange/62061-multi-class-svm (accessed on 20 June 2021).
  61. Su, J.; Vargas, D.V.; Sakurai, K. One pixel attack for fooling deep neural networks. IEEE Trans. Evol. Comput. 2019, 23, 828–841. [Google Scholar] [CrossRef] [Green Version]
  62. Wang, M.; Deng, W. Deep visual domain adaptation: A survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Representative ultrasound images of the three examined lesion types. (a) Nevus; (b) BCC; (c) MM. The primary layers of the skin are marked by the red (epidermis) and green (dermis) marks on the sides of the images. The lesions are marked by a white letter ‘L’ placed above them on the images.
Figure 1. Representative ultrasound images of the three examined lesion types. (a) Nevus; (b) BCC; (c) MM. The primary layers of the skin are marked by the red (epidermis) and green (dermis) marks on the sides of the images. The lesions are marked by a white letter ‘L’ placed above them on the images.
Diagnostics 11 01207 g001
Figure 2. Receiver operating characteristic (ROC) curves using (a) freehand-seeding-based SA segmentation; (b) LAR-seeding-based SA segmentation; (c) FA segmentation and binary classification.
Figure 2. Receiver operating characteristic (ROC) curves using (a) freehand-seeding-based SA segmentation; (b) LAR-seeding-based SA segmentation; (c) FA segmentation and binary classification.
Diagnostics 11 01207 g002aDiagnostics 11 01207 g002b
Figure 3. Examples of images with segmentation results leading to a correct ‘Nevus vs. others’ type binary classification in the cases of both SA methods (freehand: blue mark; LAR: green mark) while failing the classification results for the FA method (red mark). Three typical cases are presented: (a) an image region being totally separated from that of the lesion misleading the FA segmentation; (b) FA segmentation detecting the lesion only partially; (c) FA segmentation including misleading image areas nearby the lesion as part of the lesion segmentation.
Figure 3. Examples of images with segmentation results leading to a correct ‘Nevus vs. others’ type binary classification in the cases of both SA methods (freehand: blue mark; LAR: green mark) while failing the classification results for the FA method (red mark). Three typical cases are presented: (a) an image region being totally separated from that of the lesion misleading the FA segmentation; (b) FA segmentation detecting the lesion only partially; (c) FA segmentation including misleading image areas nearby the lesion as part of the lesion segmentation.
Diagnostics 11 01207 g003aDiagnostics 11 01207 g003b
Figure 4. Examples of images with segmentation results leading to a different ‘Nevus vs. others’ type binary classification in the case of the SA methods (freehand: blue mark; LAR: green mark). Two typical cases are presented: (a) freehand segmentation leads to correct classification, but LAR-based segmentation detects the lesion only partially so its classification fails; (b) SA freehand segmentation fails because of arbitrary boundary selection, but the ACM model, used by SA LAR, corrects the mistake.
Figure 4. Examples of images with segmentation results leading to a different ‘Nevus vs. others’ type binary classification in the case of the SA methods (freehand: blue mark; LAR: green mark). Two typical cases are presented: (a) freehand segmentation leads to correct classification, but LAR-based segmentation detects the lesion only partially so its classification fails; (b) SA freehand segmentation fails because of arbitrary boundary selection, but the ACM model, used by SA LAR, corrects the mistake.
Diagnostics 11 01207 g004aDiagnostics 11 01207 g004b
Figure 5. Examples of images with segmentation results leading to a correct ‘Nevus vs. others’ type binary classification in the case of the FA method (red mark) while failing for both SA methods (freehand: blue mark; LAR: green mark). Three notable cases are presented: (a) FA segmentation detected a similar but slightly shifted region from that detected by SA segmentation; (b) FA segmentation detected additional image regions as part of the lesion in comparison to the results of SA segmentations; (c) All three segmentation results matched closely, however slight differences in their borders led to different classification results.
Figure 5. Examples of images with segmentation results leading to a correct ‘Nevus vs. others’ type binary classification in the case of the FA method (red mark) while failing for both SA methods (freehand: blue mark; LAR: green mark). Three notable cases are presented: (a) FA segmentation detected a similar but slightly shifted region from that detected by SA segmentation; (b) FA segmentation detected additional image regions as part of the lesion in comparison to the results of SA segmentations; (c) All three segmentation results matched closely, however slight differences in their borders led to different classification results.
Diagnostics 11 01207 g005aDiagnostics 11 01207 g005b
Figure 6. Sensitivity analysis of skin lesion classification procedure. (a) Zoomed ultrasound image from Figure 5c, with segmentations denoted using the previously used color scheme (freehand SA, LAR SA, FA). Taking the freehand-based SA lesion border as a reference, the border is shrunk/grown along various regions of the border ((b): right; (c): bottom; (d): right and bottom; (e): entire region). The classification result is drawn on the corresponding border change, with gray signifying incorrect classification, and white correct classification.
Figure 6. Sensitivity analysis of skin lesion classification procedure. (a) Zoomed ultrasound image from Figure 5c, with segmentations denoted using the previously used color scheme (freehand SA, LAR SA, FA). Taking the freehand-based SA lesion border as a reference, the border is shrunk/grown along various regions of the border ((b): right; (c): bottom; (d): right and bottom; (e): entire region). The classification result is drawn on the corresponding border change, with gray signifying incorrect classification, and white correct classification.
Diagnostics 11 01207 g006aDiagnostics 11 01207 g006b
Table 1. Comparison of ultrasound-based skin lesion differential diagnosis methods focusing on benign and malignant skin lesion. In contrast to the current proposed method, none of the methods are fully automated since they do not employ fully-automated (FA) segmentation. NA stands for Not Available.
Table 1. Comparison of ultrasound-based skin lesion differential diagnosis methods focusing on benign and malignant skin lesion. In contrast to the current proposed method, none of the methods are fully automated since they do not employ fully-automated (FA) segmentation. NA stands for Not Available.
Lesions ComparedFeatures UsedPerformance
AUC/Sens./Spec.
Ref.
nevi vs. MMacoustic shadowing, dermal echo ratio
mean & std echogenicity, entry echo line
NA/100%/<30% [44]
BCC vs. othershigh frequency and Doppler ultrasoundNA/91%/14%[45]
MM vs. othersechostructure, homogeneity, lesion margin
Color Doppler, intralesional vessels
NA/100%/32%
NA/100%/34%
[46]
[46]
nevi vs. MMsurface & intra-lesional reflectance
attenuation, param. relative uniformity
NA/100%/55%[47]
nevi vs. cancerousshape & texture features86%/100%/19% [48]
BCC vs. nevishape & texture features90%/100%/45% [48]
MM vs. MSTacoustical, textural & shape features89%/85%/79% [49]
Table 2. Feature set used for ultrasound-based skin lesion classification.
Table 2. Feature set used for ultrasound-based skin lesion classification.
FeatureIdxReference/Description
First-order textural features (Attenuation, Contrast, Boundary, Statistical features)
Attenuation  [48]
      Attenuation1
      Contrast of attenuation2
      Heterogeneity of attenuation3
Contrast  [48]
      Lesion contrast-based heterogeneity4
      Mean lesion contrast5
Boundary
      Mean boundary6 [48]
      Boundary heterogeneity7 s t d ( L B )
      Boundary contrast8 a v g ( L B ) / a v g ( L )
      Boundary heterogeneity contrast9 s t d ( L B ) / s t d ( L )
      Boundary-lesion contrast10 [ a v g ( L B ) a v g ( L ) ] / a v g ( L B )
      Dermis-lesion heterogeneity contrast11 [ s t d ( D ) s t d ( L ) ] / s t d ( D )
      Boundary-lesion heterogeneity contrast12 [ s t d ( L B ) s t d ( L ) ] / s t d ( L B )
Statistical features
      Skewness13
      Kurtosis14
      Entropy15
Shape features
      Standard deviation of curvature16 [48]
      Circularity17 [48]
      Axis ratio18 M a / m a
      PA ratio19 P / A
      Compactness20 P / M a
Second-order (GLCM) textural features [57,58,59]
      Contrast21, 29, 42, 50
      Correlation I.22, 30, 43, 51
      Correlation II.23, 31, 44, 52
      Dissimilarity24, 32, 45, 53
      Energy33, 54
      Entropy34, 55
      Homogeneity I.35, 56
      Homogeneity II.36, 57
      Maximum probability25, 37, 46, 58
      Difference variance26, 38, 47, 59
      Difference entropy27, 39, 48, 60
      Information measure of correlation I.28, 40, 49, 61
      Information measure of correlation II.41, 62
List of symbols
      avg—average
      std—standard deviation
      L—Lesion region (all the pixels inside the lesion mask)
      D—Dermis region (pixels of the region of the dermis being right under the lesion mask)
      LB—Lesion boundary (a lane of pixels being located within a fixed distance from the
      lesion mask edges)
      Ma—length of major axis of the lesion mask
      ma—length of minor axis of the lesion mask
      P—perimeter of lesion mask
      A—area of lesion mask
Table 3. Area under the curve (AUC) values of binary classification. Values over 90% are highlighted.
Table 3. Area under the curve (AUC) values of binary classification. Values over 90% are highlighted.
AUCSA SegmentationSA SegmentationFA Segmentation
(Binary)(Freehand)(LAR)
      Nevus vs. others0.9210.9530.914
      MM vs. others0.7860.7580.750
      BCC vs. others0.8570.8580.840
      Nevus vs. BCC0.9300.9570.921
      Nevus vs. MM0.9250.9330.896
      BCC vs. MM0.7640.7350.783
Table 4. Mean accuracy (ACC) values of binary/multiclass classification. Values over 80% are highlighted in bold.
Table 4. Mean accuracy (ACC) values of binary/multiclass classification. Values over 80% are highlighted in bold.
Mean ACCSA SegmentationSA SegmentationFA Segmentation
(Binary/Multiclass)(Freehand)(LAR)
      Nevus vs. others0.842/0.8810.881/0.8390.848/0.855
      MM vs. others0.777/0.7520.784/0.7680.761/0.777
      BCC vs. others0.784/0.7810.784/0.7810.765/0.713
      Nevus vs. BCC0.879/0.8080.892/0.7920.850/0.750
      Nevus vs. MM0.850/0.6610.850/0.6670.811/0.689
      BCC vs. MM0.705/0.6250.670/0.6350.745/0.620
Table 5. Features performance. AUC (a) and ACC (b) measures of binary classifications using LAR-based segmentation are presented for the cases of leaving one feature out from the feature set of the classification framework. The results (AUC and ACC) are presented in ascending order, showing the omitted features in descending order regarding their contribution to the performance of the full framework. The features are represented by their index in the feature set of the framework. The top four rows for each of the six binary classification tasks are presented, together with the last row as a reference. The full list of the features (with feature names and corresponding indices) can be found in Section 2.3, Feature extraction. Features appearing at least three times in the top four are highlighted in bold.
Table 5. Features performance. AUC (a) and ACC (b) measures of binary classifications using LAR-based segmentation are presented for the cases of leaving one feature out from the feature set of the classification framework. The results (AUC and ACC) are presented in ascending order, showing the omitted features in descending order regarding their contribution to the performance of the full framework. The features are represented by their index in the feature set of the framework. The top four rows for each of the six binary classification tasks are presented, together with the last row as a reference. The full list of the features (with feature names and corresponding indices) can be found in Section 2.3, Feature extraction. Features appearing at least three times in the top four are highlighted in bold.
a.
Nevus vs. OthersMM vs. OthersBCC vs. OthersNevus vs. BCCNevus vs. MMBCC vs. MM
idxAUCidxAUCidxAUCidxAUCidxAUCidxAUC
180.910180.738150.850200.927180.866110.738
150.91370.77740.850490.92770.90540.747
130.918340.78090.85120.928200.908260.749
70.918320.780320.854100.92890.909170.752
... ... ... ... ... ...
550.929110.794140.864550.93740.930180.771
b.
Nevus vs. OthersMM vs. OthersBCC vs. OthersNevus vs. BCCNevus vs. MMBCC vs. MM
idxACCidxACCidxACCidxACCidxACCidxACC
200.82610.771150.76860.875180.789130.680
180.83280.771340.774150.87580.82870.690
410.832110.77160.777210.879130.833210.690
430.832200.77180.777260.879150.839260.690
... ... ... ... ... ...
90.868260.784390.79080.896190.86150.725
Table 6. Runtime measurements of the proposed algorithm. For segmentation and feature extraction, the mean runtime over all images; for classification tasks, the mean runtime over all folds was calculated.
Table 6. Runtime measurements of the proposed algorithm. For segmentation and feature extraction, the mean runtime over all images; for classification tasks, the mean runtime over all folds was calculated.
MethodRuntime EnvironmentMean Runtime [s]
segmentation
      lesion detectionPython 3.72.209
      border segmentationMATLAB2.414
feature extractionMATLAB0.377
binary classification
      training (lesion class vs. others)MATLAB3.129
      training (class 1 vs. class 2)MATLAB2.043
      predictionMATLAB0.002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Marosán-Vilimszky , P.; Szalai , K.; Horváth , A.; Csabai , D.; Füzesi , K.; Csány , G.; Gyöngy , M. Automated Skin Lesion Classification on Ultrasound Images. Diagnostics 2021, 11, 1207. https://doi.org/10.3390/diagnostics11071207

AMA Style

Marosán-Vilimszky  P, Szalai  K, Horváth  A, Csabai  D, Füzesi  K, Csány  G, Gyöngy  M. Automated Skin Lesion Classification on Ultrasound Images. Diagnostics. 2021; 11(7):1207. https://doi.org/10.3390/diagnostics11071207

Chicago/Turabian Style

Marosán-Vilimszky , Péter, Klára Szalai , András Horváth , Domonkos Csabai , Krisztián Füzesi , Gergely Csány , and Miklós Gyöngy . 2021. "Automated Skin Lesion Classification on Ultrasound Images" Diagnostics 11, no. 7: 1207. https://doi.org/10.3390/diagnostics11071207

APA Style

Marosán-Vilimszky , P., Szalai , K., Horváth , A., Csabai , D., Füzesi , K., Csány , G., & Gyöngy , M. (2021). Automated Skin Lesion Classification on Ultrasound Images. Diagnostics, 11(7), 1207. https://doi.org/10.3390/diagnostics11071207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop