Next Article in Journal
New N-Adducts of Thiadiazole and Thiazoline with Levoglucosenone and Evaluation of Their Significant Cytotoxic (Anti-Cancer) Activity
Next Article in Special Issue
Evaluation of Extra-Prostatic Extension on Deep Learning-Reconstructed High-Resolution Thin-Slice T2-Weighted Images in Patients with Prostate Cancer
Previous Article in Journal
Prognostic Value of Metastatic Lymph Node Ratio and Identification of Factors Influencing the Lymph Node Yield in Patients Undergoing Curative Colon Cancer Resection
Previous Article in Special Issue
AI-Powered Segmentation of Invasive Carcinoma Regions in Breast Cancer Immunohistochemical Whole-Slide Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identification of Skin Lesions by Snapshot Hyperspectral Imaging

1
Department of Dermatology, Ditmanson Medical Foundation Chiayi Christian Hospital, Chia Yi City 60002, Taiwan
2
Department of Mechanical Engineering, National Chung Cheng University, 168, University Rd., Min Hsiung, Chia Yi City 62102, Taiwan
3
Department of Dermatology, Dalin Tzu Chi General Hospital, No. 2, Min-Sheng Rd., Dalin Town, Chia Yi City 62247, Taiwan
4
School of Nursing, Shinawatra University, 99 Moo 10, Bangtoey, Samkhok, Pathum Thani 12160, Thailand
5
Division of General Surgery, Department of Surgery, Kaohsiung Armed Forces General Hospital, 2, Zhongzheng 1st.Rd., Lingya District, Kaohsiung City 80284, Taiwan
6
Director of Technology Development, Hitspectra Intelligent Technology Co., Ltd., Kaohsiung City 80661, Taiwan
*
Authors to whom correspondence should be addressed.
Cancers 2024, 16(1), 217; https://doi.org/10.3390/cancers16010217
Submission received: 4 December 2023 / Revised: 28 December 2023 / Accepted: 1 January 2024 / Published: 2 January 2024
(This article belongs to the Special Issue Advances in Oncological Imaging)

Abstract

:

Simple Summary

This research revolutionizes dermatological diagnostics by integrating artificial intelligence (AI) and hyperspectral imaging (HSI) to identify skin cancer lesions, particularly Mycosis fungoides (MF). It differentiates MF from conditions such as psoriasis and atopic dermatitis using a dataset of 1659 skin images. This study employs a novel AI algorithm alongside advanced techniques for precise lesion segmentation and classification, moving diagnosis from color to spectral analysis. This non-invasive and efficient method marks a significant advancement in the early and accurate detection of skin malignancies. The model’s high performance is validated by its sensitivity, specificity, and accuracy, making it a vital tool in dermatology for identifying skin cancers and inflammatory conditions.

Abstract

This study pioneers the application of artificial intelligence (AI) and hyperspectral imaging (HSI) in the diagnosis of skin cancer lesions, particularly focusing on Mycosis fungoides (MF) and its differentiation from psoriasis (PsO) and atopic dermatitis (AD). By utilizing a comprehensive dataset of 1659 skin images, including cases of MF, PsO, AD, and normal skin, a novel multi-frame AI algorithm was used for computer-aided diagnosis. The automatic segmentation and classification of skin lesions were further explored using advanced techniques, such as U-Net Attention models and XGBoost algorithms, transforming images from the color space to the spectral domain. The potential of AI and HSI in dermatological diagnostics was underscored, offering a noninvasive, efficient, and accurate alternative to traditional methods. The findings are particularly crucial for early-stage invasive lesion detection in MF, showcasing the model’s robust performance in segmenting and classifying lesions and its superior predictive accuracy validated through k-fold cross-validation. The model attained its optimal performance with a k-fold cross-validation value of 7, achieving a sensitivity of 90.72%, a specificity of 96.76%, an F1-score of 90.08%, and an ROC-AUC of 0.9351. This study marks a substantial advancement in dermatological diagnostics, thereby contributing significantly to the early and precise identification of skin malignancies and inflammatory conditions.

1. Introduction

Cutaneous T-cell lymphomas (CTCLs) are a group of rare dermatological malignancies of idiopathic origin, characterized by the neoplastic proliferation of T lymphocytes that play a vital role in the immune defense against pathogenic microorganisms. In CTCLs, aberrant T-cell growth predisposes the skin’s stratified epithelium to bacterial infiltration. The incipient phase of the disease is known as Mycosis fungoides (MF), which typically manifests as erythematous and scaling plaques, predominantly in sun-protected areas, or as waxing and waning nodules; these lesions disseminate systemically as a result of the migration of malignant T cells, including Sézary cells [1,2,3,4]. Early intervention is paramount to mitigate the symptomatology of the disease and decelerate its evolution. However, the premycotic presentation of MF often mimics the clinical features of psoriasis (PsO) and atopic dermatitis (AD), thereby complicating the differential diagnosis [5,6,7]. Contemporary diagnostic modalities for MF encompass the histopathological examination of cutaneous biopsies [8,9,10,11], peripheral blood analyses [12,13,14,15], and flow cytometric evaluation of T-cell clonality [13,16,17,18], complemented by lymph node biopsies and cross-sectional imaging such as computed tomography (CT) of the thoracoabdominal cavity [19,20] or fluorodeoxyglucose positron emission tomography (FDG-PET) [21,22,23,24]. In current clinical practice, the evaluation of MF predominantly relies on the histopathological analysis of tissue biopsies, lymphatic imaging modalities, hematological assessments, and various other invasive diagnostic techniques. This multifaceted approach to diagnosis, while thorough, is inherently complex and protracted, thereby imposing a considerable psychological burden on the patient.
In contemporary research, the image interpretation capabilities of artificial intelligence (AI) models have been demonstrated to attain a level of diagnostic accuracy that is comparable to that of traditional, invasive testing methodologies, thus, indicating their potential utility as adjunct diagnostic tools. This study explored the utilization of AI for computer-aided diagnosis, employing a singular multi-frame algorithm to discern three distinct dermatological disorders—MF, PsO, and AD—by analyzing cutaneous lesion imagery [25]. Validation against histopathological specimens and optical coherence tomography (OCT) corroborated the AI algorithm’s proficiency, showing a 93% accuracy in the detection of these conditions. The model’s diagnostic sensitivity notably varied across diseases, exhibiting superior performance for MF due to the typically larger lesion size. Meanwhile, the identification of AD was comparatively less precise due to the diminutive and less conspicuous nature of the lesions. The implementation of deep learning and object detection algorithms in dermatological diagnosis has garnered considerable interest within the scientific community. A notable enhancement has been reported in the Single Shot MultiBox Detector (SSD) algorithm by Wang et al. [26], published in 2023. This study introduced the SSD-KD, an innovative technique for classifying skin diseases via knowledge distillation, which capitalizes on intra-instance relational features and self-supervised dual relational knowledge distillation to augment the efficiency of lightweight models operating within computational constraints. Empirical evaluations on the extensive ISIC 2019 skin disease dermoscopic image dataset revealed that the SSD-KD achieved a commendable 85% accuracy in categorizing eight varied skin conditions while optimizing the model’s parameter efficiency and computational demands.
The academic research community has shown a growing interest in the development of methodologies for locating skin lesions and segmenting affected areas, a task of paramount importance in dermatological diagnostics. Sumithra et al. [27] introduced a novel approach for the automatic segmentation and classification of skin lesions. Their method incorporates filtering to eliminate extraneous elements, region-based segmentation, and feature extraction for classification. The technique demonstrated promising results, achieving F-measures of 46.71% using a Support Vector Machine (SVM), 34% with k-Nearest Neighbors (k-NN), and 61% for a fusion of the two, tested on a dataset comprising 726 samples from 141 images across five different disease categories. Pandiyan et al. [28] addressed the crucial challenge of classifying skin conditions, a task with significant health implications if not promptly and accurately addressed. Their methodology employed a dynamic graph cut algorithm for skin lesion segmentation, paired with a Naïve Bayes classifier for disease classification. The proposed approach outperformed existing state-of-the-art methods, achieving accuracy rates of 94.3% for benign cases, 91.2% for melanoma, and 92.9% for keratosis, as evaluated on the ISIC 2017 dataset. This advancement holds considerable potential for aiding medical professionals in the early diagnosis and treatment of various skin conditions, thereby enhancing patient care outcomes. Wen et al. [29] proposed the challenging task of skin lesion segmentation in dermoscopic images, a critical step in the precise analysis of melanoma. This network demonstrated its superiority over existing methods across multiple skin lesion segmentation datasets, marking a substantial advancement in the field. However, traditional convolutional neural networks (CNNs) have shown limitations in capturing global context information, often resulting in suboptimal segmentation outcomes.
However, the reliance on spatial information in image processing methodologies presents several limitations, particularly in the context of dermatological imaging. The diversity in photographic data, stemming from varying angles of capture and the multifarious presentation of skin lesions, poses a significant challenge. As illustrated in Figure 1, lesions associated with MF, PsO, and AD exhibit considerable variability in their anatomical distribution. They can manifest in areas ranging from the upper body with small lesion sizes (approximately 5 cm2) to extensive involvement of the back, chest, flanks, or lower body. Moreover, these lesions present a wide array of morphologies, from superficial erythematous streaks to deeper epidermal disruptions, some of which may be complicated by infection and bleeding. Given these factors, the automatic localization of lesions by using spatial information alone is not deemed essential. The static nature of the photographic data further diminishes the necessity for employing object-detection algorithms such as YOLOv5 or CNN-based approaches. Clinicians are typically able to identify the location of lesions with relative ease on the basis of surface examination. Most deep-learning CNN algorithms leverage spatial information for feature recognition. However, this approach may not yield definitive results due to the diverse anatomical distribution and morphological heterogeneity of the lesions. Despite these challenges, distinguishing between these three disease types based solely on skin color during clinical examination can be arduous. Consequently, a spectral discriminant method that capitalizes on color space conversion techniques and utilizes the spectral domain is urgently needed. Color space conversion involves translating a color’s representation from one basis to another to ensure that an image, when converted from one color space to another, retains its original appearance as closely as possible. Such an approach could differentiate the extent of invasive lesions on the basis of skin color variations, thus offering a more discernible advantage. The adoption of advanced methods, such as converting from the color space to the spectral domain, represents a noninvasive diagnostic solution that has the potential to achieve high classification accuracy.
In this study, a transformation algorithm from color space to the spectral domain was employed by utilizing hyperspectral imaging (HSI) algorithms to address the challenging task of detecting early-stage invasive lesions in T-cell skin cancer, namely MF, as opposed to milder skin conditions such as AD and PsO. As depicted in Figure 2, input images were initially subjected to coarse segmentation via a Unet-Attention mechanism. Precise accuracy in segmentation is not paramount; the primary objective is to delineate the affected skin from the healthy skin for subsequent analysis within the HSI model. This approach conserves model capacity and enhances accuracy by eliminating the noise caused by moles and other skin pigments. The output from the HSI model underwent dimensional reduction, and it was then classified using the XGBoost algorithm. This framework offers a sophisticated method for distinguishing three commonly confused skin lesions by converting images to the spectral domain within a compact model while maintaining a high level of classification accuracy.

2. Materials and Methods

2.1. Sample Preparation

A total of 1659 skin images were collected for training purposes. This dataset included 333 images of MF, 421 images of PsO, 351 images of AD, and 554 images of normal skin. These images were categorized into four distinct groups on the basis of the dermatologists’ tissue analyses, forming separate sets for training. Expert data annotation was carried out in accordance with the evaluations of pathology slices and OCT images. A digital camera (E-M10 Mark III by Olympus) was employed for capturing the skin photographs. The imaging process was directly conducted by a physician, focusing on the patient’s lesion sites. Disease classification was based on the observations by the dermatologists who reviewed the pathological tissue sections and confirmed the presence of symptomatic manifestations in the patients.
To ensure the uniform stability of illumination, it is imperative that the patient’s position, camera, and arrangement of light sources remain constant. This measure helps prevent aberration and chromatic aberration. The specific scanning position was clinically determined, based on the patient’s reports of their condition and the treating physician’s direct visual observation. Once the clinical lesions were located, multispectral analysis was conducted at these sites. Patients were instructed to alter their position for imaging, allowing a comprehensive coverage of the upper body from both front and back, akin to bilateral positions. For smaller affected areas, such as private areas, armpits, or arms, patients were asked to position themselves for close-up imaging. The data distribution is depicted in Figure 3, which illustrates the variety in the distribution of skin lesions and highlights the most frequent locations of these lesions on the body. A commonality among the three disease types was the prevalence of lesions on larger skin areas, such as the chest and back. Following these were lesions commonly found on the lateral aspects of the chest, shoulders, and arms. Some close-up shots were included to enhance the color accuracy of the lesions.

2.2. Segmentation Task

U-Nets are a widely utilized model for image segmentation. They are noted for their effectiveness across various medical imaging applications and their efficiency upon deployment. The structure of a U-Net enables high segmentation accuracy even with a limited number of training samples, thus distinguishing it from other models. Its architecture is founded on multi-stage cascaded convolutional neural networks, and these layers are tasked with extracting regions of interest and providing dense predictions. This makes a U-Net particularly adept at delivering superior performance in medical image segmentation. The attention mechanism, commonly employed in natural language processing, utilizes class-specific pooling to enhance the accuracy and speed of classification tasks. These attention maps emphasize pertinent regions, thus making them more significant for specific health-related datasets and improving the model’s applicability and performance in medical contexts.
The U-Net segmentation model was developed using PyTorch and trained using Binary Cross-Entropy (BCE) Loss with Logits. The model underwent training for 200 epochs by utilizing the Adam optimizer. The training parameters included a learning rate of 0.001, a weight decay of 0.0005, and a batch size of 32.

2.3. Model HSI-Feature Extraction Task

In this study, HSI was employed using a digital camera (E-M10 Mark III/Olympus) for skin photography and a spectrometer (Ocean Optics, QE65000) to establish the correlation between these two devices by capturing a shared target. The standard 24-color card (X-Rite Classic, 24 Color Checkers) served as the common target. This card encompasses a range of natural colors, representative of various natural objects such as human skin, leaves, and the blue sky. The digital camera captures these color blocks and translates them into digital images, while the spectrometer concurrently measures light from the 24 color blocks to delineate spectral information. The resultant images were then converted into a set of sRGB channel values. A transformation matrix was derived from these datasets. The entire modeling process is depicted in Figure 4.

2.3.1. Spectrometer to XYZ Color Space Conversion

In spectrometry, the transformation of spectral data R(λ), ranging from 380 nm to 780 nm, into the XYZ color space involves the light source spectrum S(λ) used in the camera and the XYZ color matching functions x ¯ ( λ ) , y ¯ ( λ ) , and z ¯ ( λ ) (Figure S1). The Y component of the XYZ space significantly correlates with brightness. The maximum Y value from the light source spectrum is scaled to 100 to calibrate for luminance, setting a brightness ratio k as detailed in Equation (4). This scaling ensures consistent luminance representation within a standard range. The spectral data are converted into the XYZ value (XYZspectrometer) and regulated in the XYZ color gamut space by using Equations (1)–(3) as follows:
X = k 380   n m 780   n m S λ R ( λ ) x ¯ ( λ )   d λ ,
Y = k 380   n m 780   n m S λ R λ y ¯ λ   d λ ,
Z = k 380   n m 780   n m S λ R ( λ ) z ¯ ( λ )   d λ ,
where S(λ) is the spectral density of the sample; R(λ) is the spectral density of the standard illuminant; and x ¯ ( λ ) , y ¯ ( λ ) , and z ¯ ( λ ) are the components of color matching functions. The value k, which is the brightness ratio, is given by
k = 100 / 380   n m 780   n m S λ y ¯ λ     d λ

2.3.2. Non-Linear XYZ Correction

The analysis of the camera’s nonlinear response showed that the spectrum analyzer exhibited a linear response. This investigation focused on the brightness values (denoted as Y values) for the 19th–24th color patches on a 24-color checker chart representing the gradient of grayscale changes. A linear regression analysis of the Y values corresponding to the XYZ color space of these patches revealed the camera’s nonlinearity. Figure 5 illustrates that a third-order polynomial best describes the camera’s nonlinear response, as evidenced by the coefficient of determination (R2) value. The R2 value for the third-order polynomial regression was 0.99997, which was higher than that of linear and quadratic polynomial regressions. Consequently, the camera’s nonlinear response can be effectively corrected through a third-order polynomial model. The correction factor for the nonlinear response is denoted as Vnon-linear, as shown in Equation (5).
V n o n l i n e a r = X 3   Y 3   Z 3   X 2   Y 2   Z 2   X   Y   Z   1 T
Noise in a camera image refers to the combined spatial and temporal fluctuations in the recorded signal, under the assumption of consistent, even lighting. Dark current, which is influenced by both time and temperature, can lead to the accumulation of noise on the camera sensor. Consequently, it was treated as a constant in the analytical model. This constant is denoted as Vdark, as defined in Equation (6) as follows:
V d a r k = α
Color noise, also known as chrominance noise, represents a random deviation in color relative to the image’s original colors. In contrast to luminance noise, color noise is typically linked to sensor warming or heating. In the analysis of color noise, an issue arises with respect to color matching. Given that the camera image was transformed into the XYZ color space, accounting for the interdependencies among the X, Y, and Z values became imperative. These relationships were encapsulated within the XYZ color-matching functions, which are graphically represented in Figure S1. The functions x ¯ , y ¯ , and z ¯ were integrally linked to the spectral power distribution. Consequently, a comprehensive enumeration of the potential interactions among X, Y, and Z was conducted, and the collective outcomes were encapsulated within the variable Vcolor, as delineated in Equation (7).
V c o l o r = X Y Z   X Y   Y Z   X Z   X   Y   Z T
After a thorough assessment of all identified error sources was conducted, the variable Vcolor served as the foundational element for the application of non-linear response correction, through multiplication with Vnon-linear. The resultant product was then normalized to the third order to preclude the possibility of over-correction. Subsequently, the constant Vdark was incorporated, leading to the formation of the variable matrix V. This process is systematically delineated in Equation (8). The variable matrix V was then reintroduced to derive the correction matrix C, which constituted the final corrective framework for the system.
V = [ X 3   Y 3   Z 3   X 2 Y   X 2 Z   Y 2 Z   X Y 2   X Z 2   Y Z 2 X Y Z   X 2   Y 2   Z 2   X Y   X Z   Y Z   X   Y   Z   α ] T

2.3.3. Correction Matrix C and the Calibration Camera and Spectrometer

Error factors, such as nonlinear response, dark current, and color noise, may occur when the camera is shooting. Therefore, the correction matrix C, which can be used to correct the camera, was finally obtained, as shown in Equation (9):
C = X Y Z s p e c t r o m e t e r × p i n v V ,
where V is defined as Equation (8) with X, Y, and Z components derived from XYZcamera, and [XYZspectrometer] is the matrix created by X, Y, and Z components obtained from the spectrometer.
Equation (8) was employed to extend the XYZcamera matrix to the V matrix, resulting in the corrected values of X, Y, and Z (denoted as XYZcorrection), as delineated in Equation (10). Given that the wavelength band utilized in this study falls within the visible light spectrum (380–780 nm), the outcome of this correction can be characterized in terms of chromatic aberration.
X Y Z c o r r e c t i o n = C × V ,
where V is defined as Equation (8) with X, Y, and Z components derived from XYZcamera, and [C] is the correction matrix obtained from Equation (9).

2.3.4. Principal Component Analysis of Reflectance Spectrum

In the process of converting the XYZ values (denoted as XYZcorrection), which are acquired subsequent to camera calibration, into spectral data, principal component analysis (PCA) was employed on the reflectance spectrum data (Rspectrometer) of the standard 24-color chart. This analysis yielded the principal components and their associated scores (including eigenvalues) of the reflectance spectrum for the 24-color checker chart.
The results of PCA on the reflectance spectrum (Rspectrometer) demonstrated that the initial 12 principal components (EVs) accounted for 99.99% of the total variance within the data (Figure 6) and their density details in the spectrum dataset (Figure S2). The cumulative explained variance plot displays the total variance captured by successive principal components. When the plot reaches its 12th point at 99.99%, it signifies that the first 12 components collectively explain 99.99% of the dataset’s variance. Dimensionality reduction was achieved by leveraging these 12 principal components, thereby extracting the principal component scores. The subsequent multivariate regression analysis utilized these scores, with Vcolor serving as a predictor variable. Vcolor was chosen due to its comprehensive delineation of the correlations among X, Y, and Z values. This analytical approach facilitated the derivation of the transformation matrix M, which bridges the measurement gap between the camera and the spectrometer, as explicated in Equation (11).
M = s c o r e × p i n v V c o l o r
where Vcolor is defined as Equation (7) with X, Y, and Z components derived from XYZspectrometer; [score] is the principal component scores obtained from 12 sets of principal components for dimensionality reduction of Rspectrometer; and Rspectrometer is the spectral signal obtained from the spectrometer of 24-color patches ranging from 380 nm to 780 nm with a 1 nm resolution.

2.3.5. The Hyperspectrum

The hyperspectrum is denoted by Sspectrum, as indicated in Equation (12):
[ S s p e c t r u m ] = E V M V c o l o r
where Vcolor is defined as Equation (7) with X, Y, and Z components derived from XYZcorrection; [EV] is the 12 principal components obtained from the dimensionality reduction of Rspectrometer; Rspectrometer is the spectral signal obtained from the spectrometer of 24-color patches ranging from 380 nm to 780 nm with a resolution of 1 nm; and [M] is the transformation matrix obtained from Equation (11).

2.4. Classification Task

By utilizing dimensionality reduction based on the t-Distributed Stochastic Neighbor Embedding (t-SNE) model [30], the feature extraction capabilities of the HSI model were assessed, and an appropriate classification method was determined. As depicted in Figure 7, the t-SNE analysis was conducted using 5000 data point samples from three distinct diseased skin groups. The model exhibited a distinct clustering of features specific to the PsO skin group in comparison to the other groups. However, minimal overlap was observed between the feature clusters of the AD and MF groups. Overall, the t-SNE model demonstrated relatively clear segregation among the three skin disease groups. These results showed that the HSI model possesses the capacity for effective high-level representation classification among various skin disease groups. The visualization outputs of the t-SNE model informed the selection of the XGBoost method as a feature extraction and classification technique.
Extreme Gradient Boosting (XGBoost) is known for its high accuracy in supervised learning problems, often comparable to that of other deep learning models. One of the key strengths of XGBoost lies in its rapid training capabilities, enabled by parallel computation. This efficiency is particularly advantageous when utilizing GPU resources, making the algorithm well-suited for handling high-dimensional data such as hyperspectral data. At its core, XGBoost is built upon the decision tree ensembles algorithm, allowing it to process training data without the need for normalization. This approach is rooted in a strategic focus. Instead of uniformly training on all samples, XGBoost prioritizes subsets that demonstrate poor performance. This selection process is skewed towards choosing samples that were misclassified in previous iterations rather than randomly selecting at an equal probability. While XGBoost is primarily tailored for discrete data types, such as tabular data, and may not be the optimal choice for unstructured data, its speed of training is a notable advantage over Artificial Neural Networks (ANNs). Moreover, ANNs often rely on a large volume of training samples and benefit significantly from transfer learning, which is a strategy that is less feasible with specialized data, such as that pertaining to skin diseases. This limitation makes XGBoost a more practical choice in scenarios where data specificity and training efficiency are paramount.

2.5. Training Strategy and Performance Evaluation

Given that the study involved only 34 patients, employing a fixed division of training, validation, and testing in the proportions of 70%, 20%, and 10% proved to be unsuitable for out-of-sample testing. This finding was particularly evident when considering that only three patients were involved in the testing phase, each representing a distinct disease group. Thus, k-fold cross-validation was utilized to address this challenge and prevent data leakage. In k-fold cross-validation, the original dataset is evenly divided into k subsets. Within these subsets, one is designated as the test set, and the remaining k−1 subsets are used as the training set. This process is repeated for each fold, allowing the model’s performance to be evaluated across all folds. Three different values of k were tested to assess the effect of the k-fold division on the training model: 3, 5, and 7. K-fold cross-validation is particularly beneficial for comprehensive and accurate model evaluation, especially when dealing with limited datasets. The training data are randomly divided into k parts (where k is an integer, typically 5 or 10). The model is then trained k times, with each iteration using one part as validation data and the remaining parts as training data. The final evaluation of the model is the average of the results from these k training iterations. This methodology offers a more objective and accurate evaluation, which is crucial for determining the model’s suitability for the current data and the specific problem at hand.
The commonly used metrics for performance evaluation are accuracy, sensitivity, specificity, F1-score, and area under the receiver operator curve (ROC-AUC). The metrics are shown in Section S2 of the Supplementary Materials.

3. Results and Discussions

3.1. Evaluate the Performance of the HSI Model

The simulated spectrum was juxtaposed with the reflection spectrum (denoted as Rspectrometer) of the 24-color card. The root-mean-square error (RMSE) was computed to quantify the discrepancy between the two spectra. The calculated average RMSE was 0.0525 (Table S1). Color blocks 13–18 exemplify the variance between the simulated spectrum and the measured spectrum of the 24-color card, as illustrated in Figure 8. These six color blocks were selected for investigation due to their representation of six filters commonly used in chromatic correction. The comparison between the six typical colors demonstrates the correlation between the measured and simulated spectra, indicating that the spectral reproduction algorithm has successfully produced simulation results approximating the measured spectra. Consequently, it can be inferred that the accuracy of the hyperspectral model in converting from real data is notably effective. The simulated spectrum was transformed into the L*a*b* color space for a comparative analysis using the CIEDE 2000 color difference metric (Section S1.5 in Supplementary Materials). This comparison yielded an average color difference of 0.28, as depicted in Table 1. The color difference of the 24-color blocks was illustrated on the CIE 1931 chromaticity diagram for a 2° standard observer, as shown in Figure 9. In this diagram, a black line connects two central points, with “red” indicating simulated colors and “green” representing measured colors. The minimal discrepancy between the simulated and actual colors demonstrated the effectiveness of the HSI algorithm in calibrating the correlation between the camera and the spectrometer.

3.2. Segmentation Model Results

Figure 10 presents the performance of the U-Net Attention models by using example images from various sites affected by the three skin disease groups. The model effectively distinguished between damaged and normal skin in the segmented areas, which is crucial because the objective of segmentation tasks is to isolate damaged skin from other elements such as normal skin, melanin spots, and areas affected by different injuries (such as bleeding or scratches). Given the limited number of samples, which influences the training efficiency of a deep learning model, excessive focus on model regularization through hyperparameter tuning was not a primary concern. However, metrics, such as the cross-entropy, dice coefficient, and intersection over union (IoU), were utilized to ensure the efficacy of the segmentation model. The training outcomes, as indicated by a cross-entropy of 0.3220, a dice coefficient of 0.8447, and an IoU of 0.8521, demonstrated the robust performance of the model. The metric-based evaluations and visual assessments of the test images confirmed the high effectiveness of the U-Net Attention model in this context.

3.3. Classification Task Results

In this study, the performance of the model was rigorously evaluated using k-fold cross-validation with distinct k values of 3, 5, and 7. As shown in Table 2, for k = 3, the model achieved a sensitivity of 85.61%, a specificity of 95.25%, an F1-score of 86.26%, and an ROC-AUC of 0.9051. When the k value was increased to 5, a notable improvement was found in all metrics, that is, the sensitivity, specificity, F-1 score, and ROC-AUC increased to 89.20%, 96.35%, 89.19%, and 0.9270, respectively. Further enhancement was observed with k = 7, where the model reached its peak performance, exhibiting a sensitivity of 90.72%, a specificity of 96.76%, an F1-score of 90.08%, and an ROC-AUC of 0.9351. This progressive improvement in the model performance with increasing k values underscored the effectiveness of k-fold cross-validation in refining the model’s predictive accuracy.
Figure 11 presents the confusion matrices for the k-fold cross-validation at k = 3, 5, and 7, providing insights into the model’s classification accuracy across four categories: MF, PsO, AD, and normal skin. For k = 3, the confusion matrix illustrated a balanced performance across all categories, with notable precision in classifying PsO (37 out of 43) and normal (49 out of 54) cases. However, a few instances of MF and AD were misclassified as other categories. When k = 5, the confusion matrix displayed improved accuracy, especially in correctly identifying all PsO cases (35 out of 39), and a significant reduction in misclassification of normal cases. Finally, at k = 7, the confusion matrix further refined the model’s performance, particularly in the precise identification of AD cases (33 out of 33); though, a slight increase in misclassified cases was observed in the normal category. These matrices collectively demonstrated a gradual enhancement of the model’s ability to distinguish between these dermatological conditions, thus reflecting the efficacy of higher k-fold values in achieving more accurate and reliable classification results.
Figure 12 illustrates the ROC curves for the three datasets, (a) k = 3, (b) k = 5, and (c) k = 7, across four classes (0, normal; 1, MF; 2, PsO; and 3, AD). For k = 3, the AUC values for the classes were impressively high, ranging from 0.94 to 0.97, with the highest for class 2 (PsO) at 0.97, indicating a robust classification capability. In the k = 5 dataset, a noticeable uptick in the AUC values was observed, affirming the model’s consistency and enhanced accuracy, particularly for class 2 (PsO) and class 3 (AD), with AUCs of 0.97 and 0.95, respectively. When k = 7, the model maintained a similar level of performance, with AUC values for class 0 (normal) and class 1 (MF) remaining steady at 0.96, whereas class 2 (PsO) reached a peak AUC of 0.98. The slight decrease in class 3’s AUC to 0.94 in k = 7 indicated a nuanced differentiation in the model’s predictive ability across different folds. Overall, these results highlighted the efficacy of the applied model in distinguishing among the four classes, with subtle variations observed across different k-fold settings.

3.4. Comparison with Other Existing Studies

In this section, we critically examine the performance of existing models relative to our proposed ensemble model, which integrates three distinct components: U-Net Attention for segmentation, a Hyperspectral model for feature extraction, and XGBoost for classification. The evaluation of the ensemble model was carried out according to the following criteria: segmentation efficacy, as measured by the IoU index; and classification efficiency, gauged by the ROC-AUC index in comparison to existing models with segmentation capabilities. Additionally, given that our model is designed for rapid inference, the impact on prediction speed was also a crucial factor under consideration.
To assess deployment efficiency, we measured the prediction time for each image and computed the average. Table 3 presents the model’s performance evaluation based on model size, IoU segmentation accuracy, AUC-ROC classification accuracy, and average prediction time per image, aiming to compare the practical deployment efficiency against existing models. The model’s prediction speed, denoted by the number of images processed per second, is a critical efficiency metric. Table 3 indicates that the count of non-trainable parameters is largely dictated by the architectural structure of the segmentation module, specifically the U-Net Attention model. The inclusion of the hyperspectral imaging module has resulted in a slight increase in the model size. Our ensemble model’s segmentation efficiency primarily hinges on the U-Net Attention module. While segmentation is crucial, the priority is not heavily placed on the localization of damaged skin areas but rather on differentiating them accurately from other skin lesions. Consequently, our model excelled in classification efficacy, evidenced by a classification accuracy of 0.9351, achieved using a 7-fold training approach.
In terms of prediction speed, our model demonstrated competent performance at an acceptable rate. It is noted that a trade-off exists between accuracy and inference time; a higher accuracy might lead to an increased inference time. These results highlight that our hybrid models, despite longer deployment times, prioritize diagnostic accuracy, a critical aspect in medical systems.

4. Conclusions

In this study, a novel approach leveraging HSI algorithms for the early detection of invasive lesions in T-cell skin cancer, particularly MF, was introduced. This method distinctively contrasts with traditional diagnostic processes for milder skin conditions such as AD and PsO. By utilizing a transformation algorithm from the color space to the spectral domain, coarse segmentation was applied via a Unet-Attention mechanism to the input images. This initial segmentation, while not necessitating high precision, effectively distinguished between affected and healthy skin, thereby optimizing the HSI model’s capacity and accuracy by filtering out irrelevant noise such as moles and other skin pigments. Subsequently, the HSI model’s output was subjected to dimensional reduction and classified using the XGBoost algorithm. This framework presents an advanced and compact solution for accurately differentiating among three frequently misdiagnosed skin lesions by transforming images into the spectral domain.
As described in the Introduction, this study was confined to the identification of early-stage skin cancer, specifically MF, which is frequently misdiagnosed as either of two inflammatory diseases, PsO or AD. Consequently, the primary limitation is that the study’s focus is narrow and susceptible to interference from noise when other skin diseases are present which manifest with characteristics dissimilar to the aforementioned three or presenting with lesions and discolorations on the skin that appear nearly identical to those of the three diseases. Furthermore, another significant constraint is the generalizability of the clinical examination applications due to the limited participant pool of 34 individuals, rendering the sample size insufficiently varied to supplant direct physician diagnosis. The diagnostic system developed herein serves only to augment physician diagnosis, offering a reference to aid in formulating subsequent treatment plans. In addition, limited skin phototypes also impact research outcomes. This study predominantly involved Taiwanese patients with skin phototypes ranging from III (light brown) to IV (moderate brown), according to Fitzpatrick’s classification [31]. Consequently, observations on skin areas with phototypes exceeding IV (dark brown), such as tanned skin, may result in chromatic aberration. Therefore, it is anticipated that future studies will incorporate a more diverse array of skin phototypes to broaden the research dataset. Skin phototype variations are a noted limitation in hyperspectral imaging, a spectral domain signal processing technology where uneven lighting significantly impacts spectral results. To mitigate aberrations, patient positioning, camera setup, and lighting must be consistent, with scanning positions clinically determined and adjusted for comprehensive coverage. White balance and chromatic adaptation algorithms, including the Gray World and Bradford methods, are employed to ensure color uniformity across different skin phototypes, effectively handling chromatic aberration and diverse skin conditions.
In an age where advancements in medical technology are paramount, this study echoes the necessity of continual innovation in the realm of medical diagnostics. Similar to the evolving approach to managing diseases such as dementia and retinal disorders, the application of HSI in dermatology indicates a transformative step in medical imaging. By combining the precision of spectral analysis with the prowess of AI algorithms, this research paves the way for more accurate and early detection of skin cancers, thus setting a new standard in dermatological diagnostics and patient care. The success of this study not only reinforces the potential of hyperspectral conversion technology in skin lesion analysis but also inspires further exploration into its applicability across various medical imaging domains.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers16010217/s1, Figure S1: XYZ color matching function; Figure S2: 12 principal components of Rspectrometer; Table S1: Root Mean Square Error of Sspectrum and Rspectrometer.

Author Contributions

Conceptualization, H.-Y.H., H.-T.N., P.S., P.-H.L. and H.-C.W.; Data curation, H.-T.N. and T.-L.L.; Formal analysis, H.-T.N., T.-L.L. and P.-H.L.; Funding acquisition, H.-Y.H., P.S. and P.-H.L.; Investigation, T.-L.L. and P.S.; Methodology, H.-Y.H., H.-T.N. and H.-C.W.; Project administration, H.-C.W.; Resources, T.-L.L., P.-H.L. and H.-C.W.; Software, H.-T.N.; Supervision, H.-C.W.; Validation, H.-Y.H., H.-T.N., T.-L.L. and P.S.; Writing—original draft, H.-Y.H., H.-T.N. and P.S.; Writing—review & editing, P.-H.L. and H.-C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Science and Technology Council of The Republic of China under the grants NSTC 112-2221-E-194-036. This work was financially/partially supported by the Ditmanson Medical Foundation Chia-Yi Christian Hospital, and the National Chung Cheng University Joint Research Program (CYCH-CCU-2023-02), and Kaohsiung Armed Forces General Hospital Research Program KAFGH_D_112030 in Taiwan.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of the Ditmanson Medical Foundation Chia-Yi Christian Hospital (CYCH) (IRB2022084).

Informed Consent Statement

Written informed consent was waived in this study because of the retrospective, anonymized nature of the study design.

Data Availability Statement

The data presented in this study are available in this article upon request to the corresponding author.

Conflicts of Interest

Author Hsiang-Chen Wang has been involved as a Director of Technology Development in Company, Hitspectra Intelligent Technology Co., Ltd.

References

  1. Querfeld, C.; Zain, J.; Rosen, S.T. Primary cutaneous T-Cell lymphomas: Mycosis fungoides and Sezary syndrome. T-Cell NK-Cell Lymphomas Biol. Nov. Ther. 2019, 176, 225–248. [Google Scholar]
  2. Stoll, J.R.; Willner, J.; Oh, Y.; Pulitzer, M.; Moskowitz, A.; Horwitz, S.; Myskowski, P.; Noor, S.J. Primary cutaneous T-cell lymphomas other than mycosis fungoides and Sézary syndrome. Part I: Clinical and histologic features and diagnosis. J. Am. Acad. Dermatol. 2021, 85, 1073–1090. [Google Scholar] [CrossRef] [PubMed]
  3. García-Díaz, N.; Piris, M.Á.; Ortiz-Romero, P.L.; Vaqué, J.P. Mycosis fungoides and sézary syndrome: An integrative review of the pathophysiology, molecular drivers, and targeted therapy. Cancers 2021, 13, 1931. [Google Scholar] [CrossRef] [PubMed]
  4. Dobos, G.; Lazaridou, I.; de Masson, A. Mycosis Fungoides and Sézary Syndrome: Microenvironment and Cancer Progression. Cancers 2023, 15, 746. [Google Scholar] [CrossRef] [PubMed]
  5. Hsiao, Y.-P.; Wang, H.-C.; Chen, S.-H.; Tsai, C.-H.; Yang, J.-H. Identified early stage mycosis fungoides from psoriasis and atopic dermatitis using non-invasive color contrast enhancement by LEDs lighting. Opt. Quantum Electron. 2015, 47, 1599–1611. [Google Scholar] [CrossRef]
  6. Fujimura, T.; Okuyama, R.; Ito, Y.; Aiba, S. Profiles of Foxp3+ regulatory T cells in eczematous dermatitis, psoriasis vulgaris and mycosis fungoides. Br. J. Dermatol. 2008, 158, 1256–1263. [Google Scholar] [CrossRef] [PubMed]
  7. Nashan, D.; Faulhaber, D.; Ständer, S.; Luger, T.; Stadler, R. Mycosis fungoides: A dermatological masquerader. Br. J. Dermatol. 2007, 156, 1–10. [Google Scholar] [CrossRef] [PubMed]
  8. Cerroni, L. Mycosis fungoides-clinical and histopathologic features, differential diagnosis, and treatment. Semin. Cutan. Med. Surg. 2018, 37, 2–10. [Google Scholar] [CrossRef]
  9. Shamim, H.; Johnson, E.F.; Gibson, L.E.; Comfere, N. Mycosis fungoides with spongiosis: A potential diagnostic pitfall. J. Cutan. Pathol. 2019, 46, 645–652. [Google Scholar] [CrossRef]
  10. van Santen, S.; Jansen, P.M.; Quint, K.D.; Vermeer, M.H.; Willemze, R. Plaque stage folliculotropic mycosis fungoides: Histopathologic features and prognostic factors in a series of 40 patients. J. Cutan. Pathol. 2020, 47, 241–250. [Google Scholar] [CrossRef]
  11. Hodak, E.; Geskin, L.; Guenova, E.; Ortiz-Romero, P.L.; Willemze, R.; Zheng, J.; Cowan, R.; Foss, F.; Mangas, C.; Querfeld, C. Real-Life barriers to diagnosis of early mycosis fungoides: An international expert panel discussion. Am. J. Clin. Dermatol. 2023, 24, 5–14. [Google Scholar] [CrossRef] [PubMed]
  12. Roelens, M.; de Masson, A.; Ram-Wolff, C.; Maki, G.; Cayuela, J.M.; Marie-Cardine, A.; Bensussan, A.; Toubert, A.; Bagot, M.; Moins-Teisserenc, H. Revisiting the initial diagnosis and blood staging of mycosis fungoides and Sézary syndrome with the KIR3DL2 marker. Br. J. Dermatol. 2020, 182, 1415–1422. [Google Scholar] [CrossRef] [PubMed]
  13. Horna, P.; Wang, S.A.; Wolniak, K.L.; Psarra, K.; Almeida, J.; Illingworth, A.J.; Johansson, U.; Craig, F.E.; Torres, R. Flow cytometric evaluation of peripheral blood for suspected Sézary syndrome or mycosis fungoides: International guidelines for assay characteristics. Cytom. Part B Clin. Cytom. 2021, 100, 142–155. [Google Scholar] [CrossRef] [PubMed]
  14. Lyapichev, K.A.; Bah, I.; Huen, A.; Duvic, M.; Routbort, M.J.; Wang, W.; Jorgensen, J.L.; Medeiros, L.J.; Vega, F.; Craig, F.E. Determination of immunophenotypic aberrancies provides better assessment of peripheral blood involvement by mycosis fungoides/Sézary syndrome than quantification of CD26− or CD7− CD4+ T-cells. Cytom. Part B Clin. Cytom. 2021, 100, 183–191. [Google Scholar] [CrossRef] [PubMed]
  15. Allen, P.B.; McCook-Veal, A.A.; Switchenko, J.M.; Paulino, D.M.; Niyogusaba, T.; Baird, K.M.; Tarabadkar, E.S.; Lechowicz, M.J. Staging lymph nodes and blood at diagnosis in mycosis fungoides identifies patients at increased risk of progression to advanced stage: A retrospective cohort study. Cancer 2023, 129, 541–550. [Google Scholar] [CrossRef] [PubMed]
  16. Guitart, J. Sézary syndrome and mycosis fungoides flow cytometric evaluation: The clinicians’ perspective. Cytom. Part B Clin. Cytom. 2021, 100, 129–131. [Google Scholar] [CrossRef] [PubMed]
  17. Vermeer, M.H.; Moins-Teisserenc, H.; Bagot, M.; Quaglino, P.; Whittaker, S. Flow cytometry for the assessment of blood tumour burden in cutaneous T-cell lymphoma: Towards a standardized approach. Br. J. Dermatol. 2022, 187, 21–28. [Google Scholar] [CrossRef] [PubMed]
  18. Illingworth, A.; Johansson, U.; Huang, S.; Horna, P.; Wang, S.A.; Almeida, J.; Wolniak, K.L.; Psarra, K.; Torres, R.; Craig, F.E. International guidelines for the flow cytometric evaluation of peripheral blood for suspected Sézary syndrome or mycosis fungoides: Assay development/optimization, validation, and ongoing quality monitors. Cytom. Part B Clin. Cytom. 2021, 100, 156–182. [Google Scholar] [CrossRef]
  19. Hodak, E.; Sherman, S.; Papadavid, E.; Bagot, M.; Querfeld, C.; Quaglino, P.; Prince, H.; Ortiz-Romero, P.; Stadler, R.; Knobler, R. Should we be imaging lymph nodes at initial diagnosis of early-stage mycosis fungoides? Results from the PROspective Cutaneous Lymphoma International Prognostic Index (PROCLIPI) international study. Br. J. Dermatol. 2021, 184, 524–531. [Google Scholar] [CrossRef]
  20. Calvani, J.; De Masson, A.; de Margerie-Mellon, C.; de Kerviler, É.; Ram-Wolff, C.; Gruber, A.; Meignin, V.; Brice, P.; Sadoux, A.; Mourah, S. Image-guided lymph node core-needle biopsy predicts survival in mycosis fungoides and Sézary syndrome. Br. J. Dermatol. 2021, 185, 419–427. [Google Scholar] [CrossRef]
  21. Jonak, C.; Tittes, J.; Brunner, P.M.; Guenova, E. Mycosis fungoides and Sézary syndrome. JDDG J. Der Dtsch. Dermatol. Ges. 2021, 19, 1307–1334. [Google Scholar] [CrossRef] [PubMed]
  22. McDonald, E.O.; Amanullah, A.A.; Park, P.S.U.; Song, W.; Werner, T.J.; Alavi, A.; Revheim, M.-E. The role of 18F-FDG PET/CT in primary cutaneous lymphoma: An educational review. Ann. Nucl. Med. 2023, 37, 328–348. [Google Scholar] [CrossRef] [PubMed]
  23. Gallamini, A.; Zwarthoed, C.; Borra, A. Positron emission tomography (PET) in oncology. Cancers 2014, 6, 1821–1889. [Google Scholar] [CrossRef] [PubMed]
  24. Salem, A.E.; Shah, H.R.; Covington, M.F.; Koppula, B.R.; Fine, G.C.; Wiggins, R.H.; Hoffman, J.M.; Morton, K.A. PET-CT in Clinical Adult Oncology: I. Hematologic Malignancies. Cancers 2022, 14, 5941. [Google Scholar] [CrossRef] [PubMed]
  25. Hsiao, Y.-P.; Chiu, C.-W.; Lu, C.-W.; Nguyen, H.T.; Tseng, Y.S.; Hsieh, S.-C.; Wang, H.-C. Identification of skin lesions by using single-step multiframe detector. J. Clin. Med. 2021, 10, 144. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, Y.; Wang, Y.; Cai, J.; Lee, T.K.; Miao, C.; Wang, Z.J. Ssd-kd: A self-supervised diverse knowledge distillation method for lightweight skin lesion classification using dermoscopic images. Med. Image Anal. 2023, 84, 102693. [Google Scholar] [CrossRef] [PubMed]
  27. Sumithra, R.; Suhil, M.; Guru, D. Segmentation and classification of skin lesions for disease diagnosis. Procedia Comput. Sci. 2015, 45, 76–85. [Google Scholar] [CrossRef]
  28. Balaji, V.; Suganthi, S.; Rajadevi, R.; Kumar, V.K.; Balaji, B.S.; Pandiyan, S. Skin disease detection and segmentation using dynamic graph cut algorithm and classification through Naive Bayes classifier. Measurement 2020, 163, 107922. [Google Scholar] [CrossRef]
  29. Wu, H.; Chen, S.; Chen, G.; Wang, W.; Lei, B.; Wen, Z. FAT-Net: Feature adaptive transformers for automated skin lesion segmentation. Med. Image Anal. 2022, 76, 102327. [Google Scholar] [CrossRef]
  30. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3149–3157. [Google Scholar]
  31. Chan, I.L.; Cohen, S.; da Cunha, M.G.; Maluf, L.C. Characteristics and management of Asian skin. Int. J. Dermatol. 2019, 58, 131–143. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The distribution of skin lesions in MF, PsO, and AD is remarkably varied. These lesions can manifest on the upper body with small areas starting from 5 cm2, extending to larger lesions that spread over extensive regions of the (a) chest, (b) back, (c) arms, flanks, or lower body. The morphological presentation of these lesions ranges from superficial red streaks on the skin to deeper lesions in the epidermal layers, which may be associated with infection and bleeding. These lesions, which appear as red clusters, are relatively challenging to distinguish with the naked eye. As depicted in the figure, the lesions are identified as [(ac)] MF, [(d,e)] PsO, and (f) AD.
Figure 1. The distribution of skin lesions in MF, PsO, and AD is remarkably varied. These lesions can manifest on the upper body with small areas starting from 5 cm2, extending to larger lesions that spread over extensive regions of the (a) chest, (b) back, (c) arms, flanks, or lower body. The morphological presentation of these lesions ranges from superficial red streaks on the skin to deeper lesions in the epidermal layers, which may be associated with infection and bleeding. These lesions, which appear as red clusters, are relatively challenging to distinguish with the naked eye. As depicted in the figure, the lesions are identified as [(ac)] MF, [(d,e)] PsO, and (f) AD.
Cancers 16 00217 g001
Figure 2. The experimental workflow is as follows: Input images were initially segmented in a coarse manner by using Unet-Attention. The exact precision in this segmentation phase is not critical. The primary aim is to segregate the affected skin from the normal skin for processing in the hyperspectral imaging (HSI) model. This step aids in conserving model capacity and enhances accuracy by filtering out noise, such as that from moles and other skin pigments, which are removed during this stage. Subsequently, the output from the HSI model underwent dimensional reduction, and it was classified using the XGBoost algorithm.
Figure 2. The experimental workflow is as follows: Input images were initially segmented in a coarse manner by using Unet-Attention. The exact precision in this segmentation phase is not critical. The primary aim is to segregate the affected skin from the normal skin for processing in the hyperspectral imaging (HSI) model. This step aids in conserving model capacity and enhances accuracy by filtering out noise, such as that from moles and other skin pigments, which are removed during this stage. Subsequently, the output from the HSI model underwent dimensional reduction, and it was classified using the XGBoost algorithm.
Cancers 16 00217 g002
Figure 3. The data distribution displays the number of images in accordance with their shooting angles, encompassing back-view shots, side-view shots, and close-up shots.
Figure 3. The data distribution displays the number of images in accordance with their shooting angles, encompassing back-view shots, side-view shots, and close-up shots.
Cancers 16 00217 g003
Figure 4. A schematic of the proposed method is displayed. Hyperspectral imaging facilitates the conversion between the camera and the spectrometer, employing standard 24-color blocks (X-Rite Classic, 24 Color Checkers) as calibration references. The digital camera captures these color blocks, translating the information into digital images. Concurrently, the spectrometer delineates spectral information by measuring the light from the 24 color blocks. Subsequently, the camera’s images are algorithmically transformed into a spectrum equivalent to that derived from the spectrometer, utilizing hyperspectral imaging techniques.
Figure 4. A schematic of the proposed method is displayed. Hyperspectral imaging facilitates the conversion between the camera and the spectrometer, employing standard 24-color blocks (X-Rite Classic, 24 Color Checkers) as calibration references. The digital camera captures these color blocks, translating the information into digital images. Concurrently, the spectrometer delineates spectral information by measuring the light from the 24 color blocks. Subsequently, the camera’s images are algorithmically transformed into a spectrum equivalent to that derived from the spectrometer, utilizing hyperspectral imaging techniques.
Cancers 16 00217 g004
Figure 5. Polynomial regression analysis of the brightness values of these color checker blocks was conducted to represent the gradient of grayscale changes. This analysis comprises (a) first-order (linear) regression, (b) second-order (quadratic) regression, and (c) third-order (cubic) polynomial regression.
Figure 5. Polynomial regression analysis of the brightness values of these color checker blocks was conducted to represent the gradient of grayscale changes. This analysis comprises (a) first-order (linear) regression, (b) second-order (quadratic) regression, and (c) third-order (cubic) polynomial regression.
Cancers 16 00217 g005
Figure 6. Cumulative explained variance of the first 12 principal components in the spectrum obtained from the principal component analysis.
Figure 6. Cumulative explained variance of the first 12 principal components in the spectrum obtained from the principal component analysis.
Cancers 16 00217 g006
Figure 7. The t-SNE analysis utilized 5000 data points, sampled from three different skin disease groups. This analysis revealed a clear clustering of features unique to the PsO skin group, setting it apart from the other groups.
Figure 7. The t-SNE analysis utilized 5000 data points, sampled from three different skin disease groups. This analysis revealed a clear clustering of features unique to the PsO skin group, setting it apart from the other groups.
Cancers 16 00217 g007
Figure 8. The correlation between the measured spectrum and the spectrum simulated by the algorithm, representing six basic color filters commonly used in chromatic correction, was assessed. This comparison revealed that the algorithm’s simulated reproduction closely matched the measured spectrum. The largest deviation occurred at longer wavelengths, specifically those exceeding 600 nm.
Figure 8. The correlation between the measured spectrum and the spectrum simulated by the algorithm, representing six basic color filters commonly used in chromatic correction, was assessed. This comparison revealed that the algorithm’s simulated reproduction closely matched the measured spectrum. The largest deviation occurred at longer wavelengths, specifically those exceeding 600 nm.
Cancers 16 00217 g008
Figure 9. The color difference of 24-color blocks is shown on the CIE 1931 chromaticity diagram 2° standard observer. The black line connecting the center of the two points is “red” to indicate simulated colors and “green” to indicate measured colors.
Figure 9. The color difference of 24-color blocks is shown on the CIE 1931 chromaticity diagram 2° standard observer. The black line connecting the center of the two points is “red” to indicate simulated colors and “green” to indicate measured colors.
Cancers 16 00217 g009
Figure 10. Visualization of the segmentation by the U-Net Attention model illustrates that most lesion areas are effectively isolated from the surrounding normal skin areas. The lesions manifest in various forms: either as close-up images on the arm (a,b), as clusters comprising numerous small, scattered spots (c,e,f), or even as a large area (d) distinctly segmented areas, in contrast to other skin regions.
Figure 10. Visualization of the segmentation by the U-Net Attention model illustrates that most lesion areas are effectively isolated from the surrounding normal skin areas. The lesions manifest in various forms: either as close-up images on the arm (a,b), as clusters comprising numerous small, scattered spots (c,e,f), or even as a large area (d) distinctly segmented areas, in contrast to other skin regions.
Cancers 16 00217 g010
Figure 11. Confusion matrix for k-folds (a) 3, (b) 5, and (c) 7.
Figure 11. Confusion matrix for k-folds (a) 3, (b) 5, and (c) 7.
Cancers 16 00217 g011
Figure 12. Receiver operating characteristics for (a) k = 3; (b) k = 5, and (c) k = 7 folds.
Figure 12. Receiver operating characteristics for (a) k = 3; (b) k = 5, and (c) k = 7 folds.
Cancers 16 00217 g012
Table 1. Chromatic difference diagram of measured and simulated spectra of 24-color blocks.
Table 1. Chromatic difference diagram of measured and simulated spectra of 24-color blocks.
Color No.MeasuredSimulatedL* (Measured)a* (Measured)b* (Measured)L* (Simulated)a* (Simulated)b* (Simulated)CIEDE 2000
1 38.9424.3536.8938.8124.9238.230.52
2 67.9332.0949.0867.9532.0348.940.05
3 49.450.079.8749.34−0.669.681.01
4 42.33−4.0539.4542.59−3.9738.030.57
5 55.8916.1711.3255.8816.1211.330.04
6 69.10−19.3931.4869.10−19.4631.450.05
7 65.3246.3382.1165.3746.4381.810.14
8 39.098.89−10.4939.239.36−10.400.49
9 55.1158.9646.8755.0659.0347.190.14
10 31.2224.816.3431.7825.366.050.57
11 71.87−5.8376.5671.87−5.8576.510.02
12 74.8734.2690.1374.8434.1690.250.08
13 27.8711.77−23.1027.6011.34−23.260.46
14 53.97−27.3150.5053.95−27.2250.740.11
15 45.7666.2051.5245.7665.9651.010.17
16 83.7623.22100.1583.7623.26100.180.02
17 54.1058.2824.0154.1158.3024.010.01
18 48.03−23.70−0.0848.08−23.35−0.020.19
19 95.4715.5246.2895.4815.5246.280.00
20 80.9813.5840.1380.9513.6240.110.04
21 66.3911.3534.0066.5211.4734.180.14
22 52.199.4028.1451.909.2928.260.31
23 36.456.8421.3336.466.7921.440.09
24 21.364.8714.7621.053.6815.291.55
average0.28
Table 2. Model training performance was evaluated across different k-fold values (3, 5, and 7) with respect to four performance metrics.
Table 2. Model training performance was evaluated across different k-fold values (3, 5, and 7) with respect to four performance metrics.
k-FoldSensitivitySpecificityF1-ScoreROC-AUC
k385.61%95.25%86.26%0.9051
k589.20%96.35%89.19%0.9270
k790.72%96.76%90.08%0.9351
Table 3. Comparison of model parameters and performances.
Table 3. Comparison of model parameters and performances.
ModelNumber of ParamsSize
(MB)
IoUAUC-ROCTime Prediction per Image
(second)
U-Net31M1240.80030.84900.0741
U-Net++55M2200.89770.93340.1182
U-Net Attention35M1400.85210.90970.0792
DeepLabv341M1640.87300.95970.0836
Ours (3-folds)35M1480.85210.90510.0810
Ours (5-folds)35M1480.85210.92700.0810
Ours (7-folds)35M1480.85210.93510.0810
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, H.-Y.; Nguyen, H.-T.; Lin, T.-L.; Saenprasarn, P.; Liu, P.-H.; Wang, H.-C. Identification of Skin Lesions by Snapshot Hyperspectral Imaging. Cancers 2024, 16, 217. https://doi.org/10.3390/cancers16010217

AMA Style

Huang H-Y, Nguyen H-T, Lin T-L, Saenprasarn P, Liu P-H, Wang H-C. Identification of Skin Lesions by Snapshot Hyperspectral Imaging. Cancers. 2024; 16(1):217. https://doi.org/10.3390/cancers16010217

Chicago/Turabian Style

Huang, Hung-Yi, Hong-Thai Nguyen, Teng-Li Lin, Penchun Saenprasarn, Ping-Hung Liu, and Hsiang-Chen Wang. 2024. "Identification of Skin Lesions by Snapshot Hyperspectral Imaging" Cancers 16, no. 1: 217. https://doi.org/10.3390/cancers16010217

APA Style

Huang, H. -Y., Nguyen, H. -T., Lin, T. -L., Saenprasarn, P., Liu, P. -H., & Wang, H. -C. (2024). Identification of Skin Lesions by Snapshot Hyperspectral Imaging. Cancers, 16(1), 217. https://doi.org/10.3390/cancers16010217

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop