Next Article in Journal
Artificial Intelligence-Based Cervical Cancer Screening on Images Taken during Visual Inspection with Acetic Acid: A Systematic Review
Previous Article in Journal
Polyclonal Antibody Generation against PvTRAg for the Development of a Diagnostic Assay for Plasmodium vivax
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Medical Image Classifications for 6G IoT-Enabled Smart Health Systems

by
Mohamed Abd Elaziz
1,2,3,4,*,
Abdelghani Dahou
5,
Alhassan Mabrouk
6,
Rehab Ali Ibrahim
1 and
Ahmad O. Aseeri
7,*
1
Department of Mathematics, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
2
Artificial Intelligence Research Center (AIRC), Ajman University, Ajman 346, United Arab Emirates
3
Faculty of Computer Science & Engineering, Galala University, Suze 435611, Egypt
4
Department of Electrical and Computer Engineering, Lebanese American University, Byblos P.O. Box 36, Lebanon
5
Mathematics and Computer Science Department, University of Ahmed DRAIA, Adrar 01000, Algeria
6
Mathematics and Computer Science Department, Faculty of Science, Beni-Suef University, Beni-Suef 62521, Egypt
7
Department of Computer Science, College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(5), 834; https://doi.org/10.3390/diagnostics13050834
Submission received: 11 January 2023 / Revised: 3 February 2023 / Accepted: 19 February 2023 / Published: 22 February 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
As day-to-day-generated data become massive in the 6G-enabled Internet of medical things (IoMT), the process of medical diagnosis becomes critical in the healthcare system. This paper presents a framework incorporated into the 6G-enabled IoMT to improve prediction accuracy and provide a real-time medical diagnosis. The proposed framework integrates deep learning and optimization techniques to render accurate and precise results. The medical computed tomography images are preprocessed and fed into an efficient neural network designed for learning image representations and converting each image to a feature vector. The extracted features from each image are then learned using a MobileNetV3 architecture. Furthermore, we enhanced the performance of the arithmetic optimization algorithm (AOA) based on the hunger games search (HGS). In the developed method, named AOAHG, the operators of the HGS are applied to enhance the AOA’s exploitation ability while allocating the feasible region. The developed AOAG selects the most relevant features and ensures the overall model classification improvement. To assess the validity of our framework, we conducted evaluation experiments on four datasets, including ISIC-2016 and PH2 for skin cancer detection, white blood cell (WBC) detection, and optical coherence tomography (OCT) classification, using different evaluation metrics. The framework showed remarkable performance compared to currently existing methods in the literature. In addition, the developed AOAHG provided results better than other FS approaches according to the obtained accuracy, precision, recall, and F1-score as performance measures. For example, AOAHG had 87.30%, 96.40%, 88.60%, and 99.69% for the ISIC, PH2, WBC, and OCT datasets, respectively.

1. Introduction

The emergence of the Internet of medical things (IoMT) and 6G technologies has provided the medical field with a new opportunity and methodologies to improve the diagnosis and prediction of diseases [1,2,3,4]. A large quantity of data, including computed tomographic (CT) images, are generated quickly at a short timescale, raising the problem of efficiently processing such images in real-time to help the medical field detect cancerous diseases in their early stages [5,6,7]. However, the availability and accessibility of medical images have always been limited for researchers due to privacy concerns, holding back the desired rapid advancement in the healthcare domain. Meanwhile, CT images are of low resolution, noisy, and difficult to process, which challenges routine diagnosis in terms of accuracy and precision [8,9,10,11].
In the era of advanced communication technologies such as 6G, providing a real-time medical diagnosis is a critical issue [12]. An early detection of diseases affecting sensitive human body areas, such as blood, breast, lung, and skin, can help limit the spread of the disease and protect the affected body parts. Without providing an accurate and quick diagnosis, the spread of diseases and tumors can be enormously infectious, which can cause a high rate of mortality [13]. For instance, skin cancer detection and prediction is a significant challenge in medical imaging, which is still under development. The health sector can benefit from the rapid development of medical equipment, communication technologies, and the IoMT to provide quality service at an efficient scale.
The IoMT is a collection of internet-related devices that assist healthcare procedures and activities [14]. The IoMT relates to employing the intelligent Internet of things (IoT) and modern communications to service medical personnel, medicines, and medical equipment and facilities to enable the gathering, monitoring, controlling, and a faster access to personal health data. IoT techniques’ software in medicine covers nearly all field areas, including physician ID, remote hospital emergency services, home healthcare of medical products and replacement parts, hospital instruments and the clinical waste surveillance of medical devices, blood management staff, infectious disease control, and others. In addition, the 6G-enabled IoMT provides ultrafast and accurate responses while reducing the workload and the cost of the research and development of the medical field. It is anticipated that the 6G communications system will play an essential role in providing the necessary transmission rate, stability, accessibility, and architecture [15]. Compared to the traditional diagnosis methods used for cancerous disease detection at an early stage, the 6G-enabled IoMT provides a necessary platform for processing enormous healthcare data, including hundreds of slices of CT scans [16,17].
Although the 6G-enabled IoMT has proven helpful for developing embedded systems that can detect illness with the same precision as an expert, it relies heavily on the developed algorithms based on deep learning (DL) and optimization algorithms [18]. The embedded systems in the 6G-enabled IoT can benefit from the capability of DL methods in medical image processing and cancerous disease identification. The adoption of DL methods may assist in avoiding recurrent problems that require a considerable time to solve and the need for a large number of well-labelled training data. Therefore, transfer learning (TL) can assist in overcoming some of these issues by integrating pretrained DL models [19]. DL models incorporate several techniques, such as structure design, model training, model size, feature representation, and hyperparameter optimization. On the other hand, metaheuristic (MH) optimization techniques have proven effective in addressing various complicated optimization issues for computer-aided diagnosis. For instance, Silva et al. [20] improved the hyperparameters of a CNN using particle swarm optimization (PSO) to reduce the false positives (FPs), while detecting lung nodules in lung scans owing to their similar patterns and low population density, which can produce misleading data. Moreover, Surbhi et al. [21] utilized adaptable PSO to automatically diagnose brain tumors to reduce noise and enhance image quality.
This paper introduces a framework to improve the diagnostic imaging identification efficiency, designed to be integrated into the 6G-enabled IoMT. In addition, it aims to overcome several problems, including (1) the curse of dimensionality, (2) the slow inference, and (3) the low performance. The framework comprises two phases: (1) a feature extraction using a deep learning model with transfer learning and (2) a feature selection using a developed optimization algorithm. In the first phase, a developed deep learning architecture is constructed based on MobileNetV3, which acts as a core component for feature extraction. The pretrained MobileNetV3 is employed to learn and extract medical image representations during the deep learning model training using CT images. The pretrained MobileNetV3 has been selected as the deep-learning-based model of choice due to its lightweight design that can be operated on resource-constrained devices with limited energy and resource consumption. In the second phase, a newly developed feature selection method is introduced to enhance the behaviour of the arithmetic optimization algorithm using hunger games algorithms, named AOAHG. The AOAHG method selects only the most relevant features and ensures the overall model classification improvement and efficiency. A thorough assessment of the suggested framework is presented and compared to various state-of-the-art methods utilizing four real-world datasets. In general, the main motivation for using this combination between the AOA, HGS, and MobileNetV3 was based on the shown performance of each performance on different applications. HGS has been applied to handle engineering problems [22], crisis events detection [23], node clustering and multihops [24], feature selection [25], and others [26]. AOA has been applied to solve structural design optimization [27], functionally graded material [28], robot path planning [29], human activity recognition [30], and others [31]. Therefore, this combination can improve the performance of medical image classifications in a 6G IoT-enabled smart health environment. To the best our knowledge, this is the first time the HGS, AOA, and MobileNetV3 have been integrated under a single framework for IoMT.
The main contributions of this work can be summarized as follows:
  • Proposing a 6G-enabled IoMT method that reduces human involvement in medical facilities while providing rapid diagnostic results. The new method is designed to be integrated into resource-constrained systems.
  • Using the transfer learning approach to extract the features from the medical images.
  • Enhancing the ability of the arithmetic optimization algorithm as a feature selection technique using operators of the hunger games search.
  • Evaluating the developed 6G-IoMT model using four datasets and comparing its performance with other state-of-the-art techniques.
The rest of this paper provides a background on transfer learning for extracting features. Section 4 provides the developed 6G-enabled IoMT framework. Section 5 discusses the image diagnosis framework’s outcomes. Lastly, in Section 6, the conclusion and future scope are discussed.

2. Related Works

The power of classification to aid in medical diagnosis makes it an important field of research. As a result of classification optimization, researchers have improved classification performance by applying deep learning and transfer learning in the IoMT. In addition, using metaheuristic optimization algorithms in conjunction with convolution neural networks (CNN) for medical image classification is presented in this section. Table 1 summarizes the literature review on the used datasets in our study.

2.1. IoMT-Based Deep Learning

Due to the spread of contagious diseases which can cause a pandemic, a reliable infrastructure offering conventional diagnostic tools and systems has emerged in the IoMT. The IoMT relies on the IoT infrastructure, which lowers the information transmission latency and the complexity of centralized diagnosis processes. The IoMT offers several solutions for the medical field, including monitoring systems, medical information-sharing mediums, remote consulting, and automatic report generation. Thus, technology facilitates the life of patients by offering to monitor and consult systems while helping the medical staff by reducing human intervention and human faults. The IoMT collects information from the patient using different types of sensors, devices, and clinical records, which can be stored and shared on a cloud-based centre [32]. For instance, computed aided diagnosis (CAD) technologies rely on IoT technologies to offer medical image classification, which can be built using several IoT and deep learning techniques [33]. Furthermore, self-monitoring systems can be seen as valuable components in the IoMT system such as weight and activity monitoring in diet, cardiovascular fitness, heartbeat, and nutrition planning programs [34,35,36].
Recently, the IoMT technologies have been evolving with the development of the artificial intelligence field, especially with the breakthrough of DL algorithms [18]. As a result, DL has enhanced both the specialist and the patient experience in the IoMT ecosystem providing accurate and fast diagnosis reports and helping the early prevention of disease spread. For instance, Rodrigues et al. [37] developed a vital healthcare system based on DL techniques, such as transfer learning, to classify skin lesions. Han et al. [38] investigated using DL techniques to process CT scan images and perform lung and stroke region segmentation. The authors established a communication channel with the patient relying on IoT technologies to provide diagnosis reports and consultations. Bianchetti et al. [39] proposed an automated ML system using tumour histotypes of dPET (dynamic positron emission tomography) data for adenocarcinoma lung cancer classification. Hossen et al. [40] developed a framework based on a federated learning approach and a convolution neural network (CNN) to classify human skin diseases and preserve data privacy.
Unlike fully automated DL systems, the IoMT still relies on the medical expert’s intervention to validate the results generated by a DL model or assess the accuracy of the DL model. However, DL models have shown a remarkable and accurate performance in many medical applications where they can help in decision-making and the early detection of infectious diseases from big data. Thus, developing a robust DL model to perform a specific task is vital to provide the patients with the best medicament and control the disease in its early stages [41].

2.2. Transfer Learning on Medical Images

In recent years, pretrained models for different applications have outperformed the regular learning process and training models from scratch. Thus, the performance on various applications has increased, and the learning time has been reduced [42]. The transfer learning process aims to transfer the knowledge learned while solving specific tasks to a new related task. For instance, Cheplygina et al. [43] addressed the use of transfer learning and different learning approaches in the medical field to perform medical image analysis. Transfer learning can be applied while fine-tuning all or specific layers to adapt the previously learned knowledge to the new related task. For instance, Ayan and Ünver [44] fine-tuned two pretrained models, including Xception and VGG16, trained on a large set of images from the ImageNet dataset. The fine-tuned models were used to detect pneumonia in chest X-ray images where the VGG16 exceeded the Xception model in terms of detection accuracy.
The ability to extract features from the VGG and ResNet models using bilinear classification techniques combined with SVM classifiers yielded the best results on several test sets [45]. A combination of data-driven approaches and InceptionV3 was used to train roughly 13W dermatology images, with findings on the testing set comparable to those of physicians [46]. Skin lesion segmentation was utilized to categorize melanoma in the ISBI-2016 skin lesion analysis towards cancer diagnosis [47]. As a result of this, the final classification had to be performed step by step. Multiple CNNs employing dynamic pattern training were used to simulate cancer intraclass conflict and associated noise interference in [48]. Kawahara et al. [49] decided to employ a pretrained CNN to identify skin images throughout their entire dataset rather than starting from scratch with randomly initialized parameters. After that pretraining, the CNN’s number of training rounds was considerably decreased, and the accuracy percentage for five classes was 84.8%. Lopez et al. [50] applied a deep learning method for early detection. It was developed using an adapted VGGNet design and a transfer learning technique. A sensitivity value of 78.56% was achieved using the ISIC archive dataset using the developed model. The performance of a CNN model for detecting lesions was tested using a dataset that was both extended and unaugmented in [51]. The researchers noted that deep learning approaches could be practical, and more data had to be collected. In addition, the network performed better on the additional dataset than other models. Yu et al. [47] implemented a very deep residual network-based multistage model for automatically detecting melanomas in dermoscopy images. They merged VGG and ResNet networks with the SVM classifier to improve the model’s detection performance. Zhang et al. [52] developed a deep synergic learning (SDL) model based on multiple deep CNNs in parallel with a sharing strategy for mutual learning. The authors validated the model’s performance on the ImageCLEF and ISIC datasets for medical image classification tasks.
Most of the well-known pretrained models in computer vision are based on convolution blocks, such as Inception, MobileNet, ResNet, DenseNet, and EfficientNet [53]. Furthermore, Transformer-based pretrained models were first established for language modelling and have been widely adopted for computer vision tasks. Transformer-based pretrained models benefit from the attention mechanism to learn contextual feature representation. For instance, ResViT [54] is a residual Vision-Transformer-based model for medical image tasks. ResViT synthesizes multimodal MRI and CT images in an adversarial learning process that relies on residual convolutional and transformer building blocks.

2.3. Medical Images Classification Using FS Optimizers

Currently, metaheuristic (MH) optimization techniques are applied to find solutions for different optimization problems. Those MH techniques provide a set of solutions rather than a single answer, supporting them in efficiently exploring the search space. Thus, they provide better results than traditional optimization approaches [55].
In the same context, Ravi K Samala et al. [56] presented an approach to the multilayered pathway used to predict breast cancer. They developed a two-stage approach consisting of transfer learning and determining features, respectively. To train pretrained CNNs, ROIs from large lesions were used. A random forest classification was developed based on the learned CNN. A genetic algorithm (GA) was used to select the relevant features. Silva et al. [20] optimized the hyperparameters of a CNN using PSO for the false-positive reduction in CT lung images.
Shankar K. et al. [57] developed the grey wolf optimization (GWO) technique for Alzheimer’s disease using brain imaging analysis. Then, a CNN was used to extract the features from the retrieved images. Goel et al. [58] developed an OptCoNet as an optimized CNN architecture for recognizing COVID-19 patients as having pneumonia or not. The GWO was used to determine the parameters of the convolution layer. To improve architectures for denoising images, Mohamed et al. [59] developed an enhanced version of firefly algorithms (FFA) to categorize the images as abnormal and normal. There was a significant enhancement in performance as a result of this adjustment. The diagnosis of melanoma was improved using the whale optimization algorithm (WOA) and levy flight [60]. These methods have some limitations, such as premature convergence, primarily when worked in a large search space [61]. These limitations have a negative impact on the prediction performance, especially in the IoMT environment. Therefore, the main objective of this paper was to determine the best solutions to improve the convergence rate by reducing the number of selected features.
As part of our developed study, to overcome these problems, transfer learning is integrated with metaheuristic optimization to build the IoMT framework. The qualities of this framework enable excellent performance and affordable computing expenses and address the financial concerns discussed earlier. Treating and detecting infections in or out of the clinic is essential. In order to use the IoMT system, all we need is an internet-connected device and a digital copy of the examination. The service’s quick reply allows for meaningful data throughout a session.
Table 1. The literature review on selected datasets.
Table 1. The literature review on selected datasets.
DSModel / SourceMethodology
ISIC-2016CUMED [47]Integrating a fully convolutional residual network (FCRN) and other very deep residual networks for classification.
BL-CNN [45]Combining two different types of deep CNN (DCNN) features as local and global features, using deep ResNet for the global features and a bilinear (BL) pooling technique to extract local features.
DCNN-FV [62]Integrating a ResNet method and a local descriptor encoding strategy. The local descriptors were based on a Fisher vector (FV) encoding to build a global image representation.
MC-CNN [52]Using multiple DCNNs simultaneously and enabling them to mutually learn from each other.
MFA [63]Cross-net-based combination of several fully convolutional were suggested. Used multiple CNNs for selecting semantic regions, local color and patterns in skin images. The FV was used to encode the selected features.
FUSION [64]MobileNet and DenseNet were coupled to boost feature selectivity, computation complexity, and parameter settings.
PH2ANN [65]A decision support system mad a doctor’s decision easier utilizing four distinct ML algorithms, where the artificial neural network (ANN) achieved the best performance.
DenseNet201-SVM [66]U-Net was used with spatial dropout to solve the problem of overfitting, and different augmentation effects were applied on the training images to increase the data samples.
DenseNet201-KNN [37]Combined twelve CNN models as resource extractors with seven different classifier configurations, which the greatest results obtained using the DenseNet201 model with a KNN classifier.
ResNet50-NB [67]A ResNet model was applied to map images and learn features through TL. The extracted features were optimized using a grasshopper optimization algorithm with a naïve Bayes classification.
Blood-CellCNN-SVM [68]A CNN with SVM-based classifiers with features derived by a kernel principal component analysis of the intensity and histogram data was able to classify images.
CNN [69]An SVM and a granularity feature were used to detect and classify blood cells independently. CNNs were utilized to automatically extract high-level features from blood cells, and these features were then used to identify the other 3 types of blood cells using a random forest.
CNN-Augmentation [70]The extraction and selection of features, as well as the classification of white blood cells, were all automated. A DL approach was used to automate the entire procedure with CNNs for binary and multiclass classification.

3. Background

Improved deep learning for extracting features and two feature selection algorithms, the arithmetic optimization algorithm and the hunger games search are all presented in the following.

3.1. Enhanced Deep Learning

It has been shown that DL methods are effective in various tasks, such as the categorization [71] and segmentation of images and object identification [72]. There is still much to know about the difficulties of these activities, particularly regarding the quality and effect of the acquired representations. Many DL architectures and learning methods have been developed during the last decade. With its many topologies, layouts, settings, and training procedures, the CNN is among the most studied DL models. Instead of conventional convolution operation, depthwise separable convolutions may be used on embedded devices or edge apps since they replace the existing convolutions. In order to overcome the drawbacks of conventional convolution operation, numerous DL models have adopted the idea of depthwise separable convolutions, such as EfficientNet [73]. The depthwise separable convolutions differ from conventional convolution operations in that they are applied individually to each input port. As a result, the models are operationally affordable and can be learned using lower parameters and little training time. MobileNetV3 [74] is available in two structures based on the model size: MobileNetV3-large and MobileNetV3-small. Compared to MobileNetV2, the MobileNetV3 structure is intended to reduce delay and improve accuracy. MobileNetV3-large, for example, increased accuracy by 3.2% over MobileNetV2 while decreasing latency by 20%. The NetAdapt method was used to find the best network topology and kernel dimensions for the convolution layer on MobileNetV3. The MobileNetV3 structure comprises the following fundamental components: a depth-separable convolution operation with a specific convolution kernel, a batch normalization, and an activation function. Next, the depthwise-separable, fully connected layer’s mutual information calculations and the retrieval of hidden units use 1 × 1 convolutions. Third, a global average pooling makes feature maps more manageable regarding their spatial dimension. Furthermore, by using an inverted residual block [75], we may avoid the bottlenecks caused by the residual skipped-connection method. These blocks make up the inverted residual one: (a) using the 1 × 1 extension and convolutions, as well as the depthwise convolution kernels of size 1 × 1 , for more complicated representations and to reduce model computations; (b) a convolutional layer with depth separation; (c) a method for retaining a skip connection. In addition, the squeeze-and-excite (SE) block [74] can be used to choose the appropriate features channel by channel. Finally, a rectified linear unit (ReLU) and the h-swish activation function are interchangeable terms for the same thing, the activation function.

3.2. Arithmetic Optimization Algorithm

The arithmetic optimization algorithm (AOA) [76] is an MH technique that depends on essential functions to find the optimal solution. Like other MH techniques, it begins with a randomized number of candidate alternatives (X) and the best-obtained or nearly optimal solution. For the AOA to begin functioning, the search stage should be selected first (i.e., exploration or exploitation). In the following search stages, the maths optimizer accelerated ( M O A ) is used and defined as in Equation (1).
M O A ( t ) = M i n + t × M a x M i n T
The variable t denotes the current repetition and ranges from one to the maximum allowable number of epochs (T). The terms indicate the accelerating function’s lowest and greatest values, M i n and M a x .
To discover an ideal option, AOA’s exploration agents examine the research scope at random locations across multiple areas, using two primary search methods (the divide technique and the multiply technique described in Equation (2).
x i , j ( t + 1 ) = X b j ÷ ( M O P ) × ( U L j × μ + L B j ) , r 2 > 0.5 X b j × M O P × ( U L j × μ + L B j ) , O t h e r w i s e
where U L j = U B j L B j . In this scenario, x i ( t + 1 ) represents the ith solution during the next repetition, x i , j ( t ) represents the jth location of the ith solution in the latest iteration, and X b j represents the jth place in the optimal method thus far. ϵ is a tiny integer number. The jth location’s minimum and maximum bounds are denoted by U B j and L B j , respectively. The μ = 0.5 process parameters regulate the search behaviour.
M O P ( t ) = 1 + t 1 / α T 1 / α
where M O P ( t ) in Equation (3) represents the probability of the maths optimizer ( M O P ). The current iteration is represented by t, while the total number of iterations is represented by (T). The exploitation accuracy across iterations is defined by the sensitivity parameter α = 5 .
It is necessary to do this stage of exploitation by only researching if r 1 is less than the existing M O A ( t ) quantity (see Equation (1)). In AOA, the exploitation operators (subtraction and addition) discover the research scope intensely across several populated areas and methods to produce a solution based on two primary search techniques (i.e., subtraction and addition) that are modelled in Equation (4).
x i , j ( t + 1 ) = X b j M O P × ( U L j × μ + L B j ) , r 3 > 0.5 X b j + M O P × ( U L j × μ + L B j ) , O t h e r w i s e

3.3. Hunger Games Search

The hunger games search (HGS) algorithm was developed by [77] as an optimization technique that resembles organismal biology. As a result of the HGS, a creature’s capacity to use hunger as a physiological incentive for all of these things is one of its most distinguishing features. HGS mathematical modelling begins with a population of N alternatives X before obtaining F i t i estimates for each alternative’s fitness function. The modernization step is instead carried out using the given formula in Equation (5).
X = X t × 1 + r a n d , r 1 < l W 1 × X b + R × W 2 × X b i , r 1 > l , r 2 > E W 1 × X b R × W 2 × X b i , r 1 > l , r 2 < E
where X b i = X b X t . The two variables r 1 and r 2 represent random numbers, and the parameter r a n d produces random numbers from a normally distributed set. The parameter R determines the search area and may be dependent on the number of rounds as defined in Equation (6).
R = 2 × s × r a n d s , s = 2 × 1 t T
where E indicates the parameter, which is specified as in Equation (7).
E = sech F i t i F i t b
F i t b indicates the fitness function’s highest value, whereas S e c h denotes the hyperbolic value, defined as in Equation (8).
sech x = 2 e x e x .
Additionally, W 1 and W 2 are the hunger weights from Equations (9) and (10).
W 1 = H i × N S H × r 4 , r 3 < l 1 , r 3 > l
W 2 = 2 1 e ( | H i S H | ) × r 5
S H represents the solution’s hunger-experiencing accumulation, and the parameters S H correlate to r 3 , r 4 and r 5 being random integers with ranges in the interval [0, 1], as follows:
S H = i H i
H i = 0 , F i t i = F i t b H i + H n , o t h e r w i s e
where H n represents the new hunger, and it is formulated as:
H n = L H × 1 + r , T H < L H T H , o t h e r w i s e
T H = 2 F i t i F i t b F i t w F i t b × r 6 × U B L B
Moreover, there is a lower value provided by F i t w for the fitness function; in addition, r 6 [ 0 , 1 ] is a randomised number that indicates if hunger has positive or negative effects based on various variables.

4. Developed Approach

To accomplish our approach, we created a 6G-enabled IoMT framework. It is capable of transmitting data quicker than a 5G-enabled system. In bandwidth, 6G may reach microseconds, significantly improving its speed over 5G’s milliseconds [78]. Furthermore, 6G enables real-time broadcast and processing better quality images and assists artificial intelligence in achieving real-time broadcast and execution. Nevertheless, only low-latency and high-bandwidth wireless communication technologies can satisfy the developing requirements of DL and IoMT. Therefore, based on the 6G network and DL model concepts, we suggest incorporating the combined DL and FS optimizer algorithms presented in the following subsections into our 6G-enabled IoMT framework.

4.1. Feature-Extraction-Based Deep Learning

To identify and extract feature information, we utilized a transfer learning approach. Pretrained models for image recognition tasks are helpful because they speed up training and implication. Instead of building models from scratch, it is possible to fine-tune a few layers while the model’s weights are fine-tuned. We replaced the model’s top part with new layers for classification and feature extraction. MobileNetV3 was used as a core block for extracting features after fine-tuning its weights on different task-specific datasets.
MobileNetV3 was adjusted and trained to retrieve feature representations from input with a size equal to 224 × 224 . The ImageNet data [75] were used to train the MobileNetV3 model and produce pretrained versions based on the model size (large or small). We used the dataset representing images of skin cancer, blood cells, and optical tomography to fine-tune the MobileNetV3-Large pretrained model. In our experiments, we replaced the MobileNetV3 model’s classification layer with two layers represented as 1 × 1 pointwise convolutions to extract the image representations and fine-tune the model for the classification task.
The 1 × 1 pointwise convolution is often used to categorize and extract features that have similar applications to those of multilayer perceptrons (MLPs). After fine-tuning the MobileNetV3 layers, we fed the extracted features to a 1 × 1 pointwise convolution which learned task-specific features. The MobileNet3 core layers are a combination of inverted residual blocks stacked sequentially. Each inverted residual block consists of several components derived from the MobileNetV2 structure, including a 1 × 1 expansion convolution, a depthwise-separable convolution, a squeeze-and-excite block, a 1 × 1 projection convolution, and a skip-connection mechanism. Furthermore, a kernel of size 3 × 3 is used in the depthwise-separable convolution with an activation function which can be placed in the following order ( 3 × 3 C o n v ) ( B N ) ( R e L U / h s w i s h ) ( 1 × 1 C o n v ) ( B N ) ( R e L U / h s w i s h ) . A depth-separable fully connected layer with various nonlinearity variables, including hard swish (h-swish) or ReLU, may be included in each construction block. These functions are described in Equations (15) and (16).
R e L U ( x ) = max ( 0 , x )
h s w i s h ( x ) = x × σ ( x )
where σ ( x ) specifies the piecewise linear difficult analogue functional, where σ ( x ) = R e L U 6 ( x + 3 ) 6 . The output of the 1 × 1 pointwise convolution placed before the classification layer ( 1 × 1 pointwise convolution) is the feature extraction block that generates the learned image embeddings during the network training and fine-tuning. Each extracted image embedding is represented with a 128-feature vector. The developed model was trained on each dataset for 100 epochs with a batch size of 32 with an early stopping strategy (20 epochs). The RMSprop algorithm, with the learning rate of 1 × 10 4 was applied to modify the model’s weight and bias values. For this reason, we employed a dropout layer and data augmentation with randomized horizontal flips, randomized crops, colour jitters, and periodic vertical flips to counteract the model’s overfitting problem. The Pytorch framework was used to implement the model, and the training was conducted on an Nvidia RTX1080 GPU.

4.2. The Developed FS Algorithm

This article aims to provide a novel technique for enhancing the efficiency of the arithmetic optimization algorithm (AOA). This was achieved by using the operators of the hunger games search (HGS) algorithm. Whenever the AOA could not discover the optimal solution within a specified iteration, a much more effective searching focused on the HGS was applied to enhance the exploration ability. The HGS enhanced the capacity to do global and regional searches concurrently.
The basic steps of the FS technique, called AOAHG, are shown in Figure 1. The initial stage in the developed AOAHG was to create the set of N agents X, reflecting the FS challenge solutions. This procedure was obtained by using the following formula:
X i = r a n d ( U L ) + L , i = 1 , 2 , , N , j = 1 , 2 , , D i m
The term D i m denotes the number of features. As a result, the available dimensionality was restricted to values between U and L. We used the following equation to obtain the binary form of each X i :
B X i j = 1 i f X i j > 0.5 0 o t h e r w i s e
As a further step, we calculated the fitness value for X i as in Equation (19), based on its binary form B X i .
F i t i = λ × γ i + ( 1 λ ) × | B X i | D i m ,
In this case, the proportion of features associated is denoted as ( | B X i | D i m ) . γ i is the validation loss of the SVM. In general, the SVM is often applied because it is more reliable and has fewer parameters than other classifiers. The value of the parameter λ balances the ratio between the accuracy of a classifier’s predictions and the selection of features.
The following procedure was used to adjust a solution X i by using either the HGS or AOA operators. This was accomplished through the use of the probability P i associated with each X i . While the HGS may take longer, it was used if the probability of P i was less than the M O A , as defined by the given equations:
X i j = X i j A i f P i > M O A X i j H G o t h e r w i s e
where M O A is specified in Equation (1). The value of X i j A is updated using the operators of the AOA as described in Equation (2).
X i j A = T h e f i r s t r u l e o f E q u a t i o n ( 2 ) i f P A > 0.5 T h e s e c o n d r u l e o f E q u a t i o n ( 2 ) o t h e r w i s e
where P A [ 0 , 1 ] is a random sample variable that is utilised to keep the AOA operators comparable throughout solution updates.
The HGS operators were then applied on the upgraded population of X. The following formula yielded X i j H G :
X i j H G = W 1 × X i j R × W 2 | X i j X ( t ) | i f P H > 0.5 W 1 × X i j + R × W 2 | X i j X ( t ) | o t h e r w i s e
where W 1 and W 2 are defined in Equations (9) and (10), respectively. If P H was greater than 0.5, the first HGS rule was applied; otherwise, the second HGS rule was applied.
Additionally, the search space [ U , L ] was dynamically changed throughout the finding process as follows:
L j = m i n ( X i j )
U j = m a x ( X i j )
The next stage was to determine if the closure conditions were met, and if so, the optimum solution was given. If this happened, the upgrade procedure was repeated from the beginning. The suggested AOAHG’s pseudocode is given in Algorithm 1.
Algorithm 1 Pseudocode of the developed AOAHG algorithm
1:
Initialize the parameters.
2:
Split the dataset into training and testing sets after extracting the features.
3:
Initialize the number of solutions (N).
4:
repeat
5:
      Determine the value of the fitness function.
6:
      Find the best solution.
7:
      Update the M O A value using Equation (1).
8:
      Update the M O P value using Equation (3).
9:
      Calculate the hunger weight of each position using Equations (9) and (10).
10:
    Enhance H i using Equation (12).
11:
    for  i = 1 to N do
12:
        for  j = 1 to P o s i t i o n s  do
13:
              Generate a random values in [0, 1] ( P i , P A , and P H ).
14:
              if  P i > M O A  then
15:
                     Position limitations can be adjusted for new seeds.
16:
                     if  P A > 0.5 then
17:
                            Update ith solutions’ positions by the first rule in Equation (2).
18:
                     else
19:
                             Update ith solutions’ positions by the second rule in Equation (2).
20:
              else
21:
                     if  P H > 0.5 then
22:
                             Update ith solutions’ positions by the first rule in Equation (22).
23:
                     else
24:
                             Update ith solutions’ positions by the second rule in Equation (22).
25:
until The iteration (t) criterion has been met.
26:
Return the best solution.

4.3. Sixth-Generation-Enabled IoMT Framework

The suggested 6G-enabled IoMT architecture is shown in Figure 2. The terminal intelligence of the IoT first collected diagnostic images, and if the expert’s aim was to learn the framework, the input images could be transmitted through a 6G network. Then, the data collected from the multiaccess edge-computing servers could be uploaded to a cloud computing service.
The three primary processes were still in place in cloud computing. In the first stage, the DL design’s features were retrieved, as discussed in Section 4.1. As a second stage, we used the modified AOA depending on an HGS (AOAHG) to select the significant features, as illustrated in Section 4.2. Finally, once the classifier had been learned, it could be distributed across several API forecasting/prediction nodes, saving on transmission fees.
On the other hand, if the user’s goal was to test the case/disease of the collected image, the test pattern in the API prediction/forecasting tools was employed. API forecasting enabled the system’s authorized training product to forecast anything without retraining, saving time and reducing internet traffic. Finally, the sender/specialist was given the last diagnostic and several evaluation metrics such as accuracy, F1-score, and others to back up the system’s forecasts.
The time complexity of the developed method depended on the AOAHG and MobileNetV3. The complexity of the developed AOAHG method was represented as O ( N × ( T × D + N + 1 ) where N ,   T , and D are the number of solutions, iterations, and dimensions, respectively. In addition, MobileNetV3 had around 3 million trainable parameters.

5. Experimental Studies and Results

5.1. Dataset

Four medical datasets were employed for our experimental assessment, including a white blood cells (WBC) dataset, retinal optical coherence tomography (OCT) images, and skin images to identify malignant ones. To perform skin cancer classification, two datasets of dermatoscopic images were used: PH2 [79] and ISIC-2016 [80]. Figure 3 depicts a sample of images from the tested datasets.

5.1.1. WBC Dataset

The accessible data utilized in this research were classified into four types, as described in [81], and included the following: eosinophil, lymphocyte, monocyte, and neutrophil. The WBC dataset contains microscopic images of 3120 eosinophils, 3103 lymphocytes, 3098 monocytes, and 3123 neutrophils. Each picture has a resolution of 320 × 240 pixels and a depth of 24 bits. Furthermore, the dataset was divided into two parts: 80% for training and 20% for testing. To be more specific, the training set had 2496 eosinophils, 2484 lymphocytes, 2477 monocytes, and 2498 neutrophils, while the testing set contained 620 from monocytes, 624 from neutrophils, also, 623 from eosinophils and lymphocytes.

5.1.2. OCT Dataset

In this section, we introduce the description of the OCT dataset, which consists of 84,484 OCT B-scans obtained using 4686 patients (collected at the Shiley Eye Institute of the University of California, San Diego (UCSD)). These images are categorized into four classes, DME, CNV, drusen, and normal, which contain 8866, 37,455, 11,598, and 26,565 images, respectively. Additionally, this dataset includes 968 test images and 83,516 training images. For the training, 37,213, 11,356, 8624, and 26,323 images were used from each class, respectively, and for the testing set, we used all images from each class.

5.1.3. PH2 Dataset

A sample size of 200 dermoscopy images was included in the PH2 dataset, comprising 80 common nevus, 80 atypical nevus, and 40 melanoma. The data were split into 85% training and 15% testing sets.

5.1.4. ISIC Dataset

In all, 1179 samples were included from the ISIC-2016 dataset, which was divided into two categories including benign and cancerous. The ISIC-2016 dataset contains 248 images of malignant tumours and 1031 images of benign tumours. Furthermore, the data were divided into 70% and 30% training and testing sets, respectively. For the training, we used 173 malignant and 727 benign images, whereas for the testing, we used 75 malignant and 304 benign images.
To assess the efficiency of the developed method for classifying medical images, the recall, precision, balanced accuracy, accuracy, and F1-score were used.

5.2. Experimental Results and Discussion

This section summarises the results of the experiments conducted to evaluate the efficiency of the developed 6G-IoMT approach. We assessed our developed FS method against other FS based on MH approaches, including the Aquila optimizer (AO) [82], PSO, GWO, moth-flame optimization (MFO) [83], bat algorithm (BAT) [84], Archimedes optimization algorithm (ArchOA) [85], chaos game optimization (CGO) [86], hunger games search (HGS) [77], and arithmetic optimization algorithm (AOA) [76]. After that, extreme gradient boosting (XGB), the K-nearest neighbours (KNN), random forest (RF), and support vector machine (SVM) classifiers were assessed against each other. All tests used a population size of 50 and 20 iterations. The other parameters were set according to the original implementation.

5.2.1. Results of FS Methods

Results from the ISIC-2016 dataset and PH2 dataset can be found in Table 2. Table 3 contains the findings from the WBC dataset and OCT dataset.
From Table 2, the SVM-based AOAHG provided better results than other approaches on the ISIC-2016 dataset. The accuracy of the AOAHG algorithm using the SVM was 87.34%, representing the best efficiency, followed by MFO, which achieved the second rank with 86.54%. ArchOA, AOA, and HGS followed the previous two algorithms (AOAHG and MFO). The BAT and CGO algorithms, with 86.02%, followed the preceding methods. The algorithms that followed were the PSO (85.75%), AO (85.49%), HGS (84.96%), and GWO (84.43%) algorithms. For the value of precision, the developed AOAHG method achieved 86.53%, followed by the MFO with an 85.60% accuracy.
The results of the recall of the AOAHG method were better than others. The developed algorithm was followed by MFO, which had a success rate of 85.54%. The ArchOA and AOA all had the same value of recall, that is, 86.28%. For the vote, the PSO obtained 86.02%, and both CGO and BAT achieved 85.75%. Finally, the GWO algorithm had the worse outcome at 84.43%. The presented AOAHG method outscored the other methods with an F1-score of 86.47%. The AOA and ArchOA obtained the exact value of 85.73%. Next, MFO, CGO, BAT, and AO had an F1-score of 85.57%, 85.50%, 85.00%, and 84.86%, respectively.
For the PH2 dataset, Table 2 illustrates that the AOAHG method significantly improved the determination of features when using an SVM classification method; this was evident across all metrics. According to the accuracy measure, AOAHG correctly classified 96.43% of the testing samples when an SVM was used. In addition, these results were significantly different from the accuracy of other FS approaches. Moreover, the AOAHG had the better precision value of any SVM methods at 96.44%, the highest of any other optimizer algorithm. AOA, CGO, BAT, ArchOA, MFO, GWO, and PSO placed second. Following these optimizers were the AO and the HGS methods, which achieved an 96.70% accuracy. As a further analysis, the recall metric for the SVM classification model was 96.43% for AOAHG, which indicated that the developed method had the maximum effectiveness. The developed AOAHG method was the best optimizer based on the F1-score with 96.43%. The PSO, GWO, MFO, ArchOA, AO, BAT, HGS, CGO, and AOA approaches had an F1-score of nearly 96.07%. Furthermore, the presented AOAHG method had the highest balanced accuracy value, nearly 97.02%. These algorithms came in second place for all the other optimization algorithms with 96.73%. Nevertheless, when these ten optimizers were combined with the KNN, XGB, and RF classifiers, the outcomes had the poorest overall performance measures compared to those of the SVM classification algorithm.
Table 3 shows the comparison between the AOAHG approach and other optimizers using the white blood cell dataset. Based on the results, the AOAHG algorithm based on an SVM provided a better accuracy (nearly 88.62%) than the other algorithms. The AOA was second with an accuracy of 88.58%. Then, the MFO and HGS techniques had the same outcome (i.e., 88.54%), and both the CGO and GWO approaches obtained the same accuracy value, 88.50%. In addition, from those results, it can be noticed that the ArchOA and AOA had the worst score at 88.26%.
Moreover, the recall values of AOAHG were better than those of the other methods. The AOA, HGS, and MFO methods all had similar recall values (i.e., around 88.54%). Finally, the AO and ArchOA had a worse outcome of 88.26%. The developed AOAHG technique outperformed other methods according to the F1-score, which was 88.80%. The AOA was second, with 88.76%, and the ArchOA obtained the worst F1-score value of 88.44%. Based on the results of the balanced accuracy, the AOAHG algorithm provided better results with an accuracy of 88.62%. Similarly, according to the balanced accuracy, the AOA was ranked second (88.58%).
Table 3 shows the results of the algorithms applied on the OCT dataset. From those results, it can be noticed that the AOAHG method was better than the other optimization methods. The best performance based on the accuracy measure was the AOAHG approach using the SVM with a 99.69% accuracy. At the same time, the CGO and HGS methods were ranked second with an accuracy 99.59%. Based on the precision value, the developed AOAHG approach achieved a score of 99.69%. The CGO and HGS algorithms followed with a precision of 99.59%. The precision of the AOA, BAT, ArchOA, MFO, and GWO algorithms was 99.40%. On the other side, the AOAHG approach had the highest recall measure performance of any SVM classifiers at 99.69%. Coming in second were HGS and CGO, which both had a recall of 99.59%. Five optimizers had a common recall value of almost 99.38%: GWO, MFO, ArchOA, BAT, and AOA. On the other hand, PSO and AO performed the worst at 99.28%. Our new algorithm (i.e., AOAHG) outperformed earlier ones with 99.69% on the F1-score metric. Following that, CGO and HGS both obtained 99.59% for the F1-score. Nevertheless, that was not the end of it. With a result of 99.28%, AO and PSO came in dead last in the competition. The AOAHG algorithm had the most excellent actual quality, with a balanced accuracy of 99.69%. HGS and CGO obtained 99.59% and came in second. Following that, AOA, BAT, ArchOA, MFO, and GWO had a balanced accuracy of 99.38%. However, PSO and AO had the worst results, with a balanced accuracy of just 99.28%.
From a different viewpoint, as shown in Figure 4, the average outcomes of the ten feature selection optimizers investigated on the four classifiers (i.e., SVM, KNN, RF, and XGB) on the four chosen datasets, PH2, ISIC, WBC, and OCT, are given in Figure 4. From Figure 4a it can be noticed that the overall average accuracy on the PH2 dataset was nearly 96.11% and 95.68% for the SVM and KNN, respectively. In addition, the overall balanced accuracy of the SVM classifier was the best (96.76%). It was followed by the KNN (96.10%), the XGB (93.84%), and the RF (92.14%) classifiers. Moreover, the best F1-score from the ten optimization techniques was obtained by the SVM at about 96.11%; the KNN was second with 95.69%. Furthermore, the XGB outperformed the RF algorithm, with a success rate of 93.08% for XGB and 92.27% for RF, whereas the SVM was better than the other classifiers based on the recall value. Furthermore, the SVM achieved 96.11%, while the KNN classifier achieved 95.68%. Finally, the XGB and RF algorithms obtained 93.07% and 92.22%, respectively. In terms of precision measure, the SVM classification algorithm delivered superior results compared to the KNN, XGB, and RF classifiers, with 95.73%, 93.56%, and 92.93%, respectively.
As shown in Figure 4b, the average accuracy of the ten optimization techniques on the ISIC dataset using the SVM was 85.91%; the RF algorithm took second place with 85.49%. Furthermore, the XGB achieved 84.75%, outperforming the KNN, which achieved 84.28%. Moreover, the RF was better than the other classifiers in terms of balanced accuracy. To be more specific, the RF achieved 74.38%, while the XGB achieved 73.82%. Finally, the SVM and KNN algorithms obtained 73.39% and 72.22%, respectively. Regarding the F1-score measure, the SVM classification algorithm delivered superior results compared to those of the RF, XGB, and KNN classifiers, with 85.04%, 84.39%, and 83.74%, respectively. Additionally, the overall average recall was approximately 85.91% for the SVM classifier, whereas the RF classifier came in second with 85.49%. XGB obtained a higher percentage (with 84.75%) compared to the KNN classification algorithm. In addition, the SVM classifier precision was the highest at 85.01%. It was followed by the RF (84.08%), the XGB (84.18%), and the KNN (83.46%).
As shown in Figure 4c, the SVM classifier achieved the highest accuracy on the WBC dataset with 88.46% followed by the KNN (88.44%), RF (88.39%), and XGB (88.30%) classifiers, respectively. Meanwhile, the BA metric of all optimization approaches outperformed the SVM classifier by 88.46% and the KNN classifier by 88.44%, respectively. The RF classifier surpassed the XGB classifier, with an accuracy equal to 88.39% for RF and 88.29% for XGB. The SVM classifier scored an average F1-score equal to 88.65%, whereas the KNN scored 88.63%. The RF obtained a higher F1-score (with 88.60%) than the XGB classification algorithm. The SVM classification algorithm delivered better results than the KNN, RF, and XGB classifiers, with a recall equal to 88.44%, 88.39%, and 88.30%, respectively. Additionally, the RF was better than the other classifiers based on the precision score. For instance, the RF achieved 90.51%, while the KNN and SVM achieved 90.49%. Finally, the XGB algorithm obtained 90.47%.
As shown in Figure 4d, the SVM classification algorithm delivered a superior average accuracy score on the OCT dataset compared to XGB, KNN, and RF classifiers, with 99.30%, 99.28%, and 99.28%, respectively. The average balanced accuracy was 99.43% for the SVM classifier, whereas the XGB classifier came in second with 99.30%. KNN obtained a lower average balanced accuracy (with 99.28%) than the RF classification algorithm. From a different perspective, the overall F1-score of the SVM classifier was the best (99.43%). It was preceded by the XGB (99.30%), the KNN (99.28%), and the RF (99.28%) classifiers. Meanwhile, the SVM scored a better recall than the other classifiers, equal to 99.43%, while the recall for XGB was 99.30%. The KNN and RF algorithms obtained 99.28%. Moreover, the SVM classifier achieved an average precision equal to 99.45%, followed by the XGB classifier with 99.32%. In addition, the KNN and RF classifiers achieved the same precision at 99.30%.
Figure 5 presents the average accuracy of the experimented classifiers on the four datasets using different optimization strategies. As shown in Figure 5, the SVM showed a significantly better performance compared to the other classifiers in terms of accuracy. To be more precise, the SVM had an accuracy of 92.48%, while the KNN method had an accuracy of 91.92%. Finally, the XGB and RF algorithms achieved an accuracy of 91.35% and 91.34%, respectively.
The whole operation took less time to finish than it did for a consumer, as shown in Figure 6, which displays the average execution time for the optimizers on the selected datasets. According to the findings, the KNN classification model was the fastest (i.e., less time), followed by the SVM classification algorithm, which required 2.7829 s to finish. In all, 22.5073 s were needed to complete the RF classification. The XGB required the longest duration at 41.4983 s.
For the four datasets, using the SVM classifier, the developed AOAHG and ArchOA took an average of 1.7031 and 1.9405 s to execute, as shown in Figure 7. Overall, these times were faster than those of other similar methods. The AO optimizer completed in 2.3397 s, while the HGS, MFO, CGO, GWO, BAT, and AOA optimizers completed in 2.3704 s, 2.4415 s, 2.7837 s, 3.1439 s, 3.3205 s, and 3.3388 s, respectively. For instance, the PSO algorithm achieved the longest execution time (4.4469 s).
From a different viewpoint, Figure 8 illustrates each feature selection strategy on four datasets and the corresponding average accuracy. Using various optimizers, the SVM classifier outperformed the AOAHG technique on average with a 93.02% accuracy. The MFO method ranked second with an accuracy of 92.63%. The AOA outperformed the CGO, ArchOA, and BAT algorithm, averaging accuracies of 92.55%, 92.50%, and 92.45%, respectively. The PSO obtained an accuracy of 92.39%. These three optimizers produced the worst results, with an average balanced accuracy of 92.29% (HGS), 92.28% (AO), and 92.1% (GWO).
To summarise, for the ISIC-2016, PH2, WBC, and OCT datasets, the AOAHG algorithm alongside the SVM classifier obtained the best accuracy. In addition, our suggested method also produced the quickest results (i.e., the least execution time).

5.2.2. Compared Methods

Other medical image categorization approaches are examined in this section to compare our developed method. Table 4 summarises the findings of many critical methods. Medical image categorization requires the development of highly accurate technologies. Comparing our approach to other models evaluated on the same datasets is critical. Table 4 compares the accuracy of various illness detection approaches using the ISIC, PH2, WBC, and OCT datasets.
The ISIC dataset was used to evaluate various skin cancer detection techniques, including integrated feature fusion [45], corroborated by Fisher coding and extensive deep networks [62], interactive model of multi-CNN learning [52], and fusing Fisher vectors with CNN data [63].
In [65], the authors built a decision framework using a convolutional neural network to evaluate the PH2 dataset for skin cancer detection. A U-Net could automatically identify malignant tumours, according to [66]. Rodrigues et al. [37] used transfer learning and a CNN as components of their IoT architecture. A hierarchical structure based on two-dimensional elements in the picture and ResNet were presented in [67] for improved deep learning. The following identification techniques were utilized to recognize and estimate essential blood cells in the WBC dataset. A CNN was used to perform classification, as described in [68]. Additionally, Ref. [69] took advantage of a selectivity feature and an SVM. CNNs were offered as a deep learning strategy in [70] for automating the whole operation.
Six well-known classification algorithms were implemented and tested in [87] to validate their performance on the OCT dataset. The algorithms included transfer learning [87] and IFCNN [88]. IFCNN [88] used numerous convolutional features inside of a CNN and a recurrent fusion method to identify OCT images. Huang et al. [89] devised a special layer guided convolutional neural network (LGCNN) to discriminate between the typical retina and three common macular disorders. Kermany et al. [87] introduced an image-based deep learning (IBDL) technique that adjusted the channel’s parameters and was utilized as a feature representation. Sun et al. [90] used sparse coding and dictionary learning based on the scale-invariant feature transform (SIFT) metaphor to identify AMD, DME, and routine images. Ji et al. [91] used Inception V3 via transfer learning as a feature extractor where a CNN was added on top of the pretrained Inception V3 network after eliminating the top layers to detect feature-space alterations.
Overall, our technique allows us to eliminate unnecessary features from high-dimensional representations of the input medical image extracted from a CNN network. Nevertheless, our framework’s primary flaw is that it is time- and memory-intensive. The next stage is to simplify the framework and make it more efficient. Other techniques of augmentation may be investigated in the future to further enhance our current system.

6. Conclusions and Future Works

The attractive characteristics of 6G compared to earlier generations of wireless networks have lately generated a considerable attention in business and academic fields. In our study, the developed framework depended on the cloud centre’s classification models being trained before they could be put to work. They were then transmitted to the cloud centre after being analysed at the cloud centre using learned representations from medical images obtained from edge devices on IoT/fog computing nodes. MobileNetV3 was modified and fine-tuned using medical pictures to determine more complex and informative representations and extract image embedding. Furthermore, a novel metaheuristic algorithm that relied on the arithmetic optimization algorithm (AOA) and hunger games search was developed as a feature selection method to filter only relevant features from image embedding. Convergence was accelerated, and feature vectors were improved as a result. In order to determine how well the developed framework model performed, it was sent to a simulated medical imaging cloud centre or assessed using fog computing and a copy of the developed algorithm. The developed framework was tested on the ISIC-2016, PH2, WBC, and OCT datasets. The results showed that the presented technique outperformed other methods already used for feature selection. In addition, the results of the assessments with other new medical image categorization technologies showed that the IoMT technique developed could enhance the overall performance and services. An increased amount of medical information, as well as its use in medical treatment, will be assessed as part of future research. Combining multiple classification techniques is also an intriguing research topic since it may enable practitioners to improve the performance of current approaches. In addition, the hyperparameters optimization of deep learning models can be investigated as using the wrong hyperparameters can limit the performance of the model.

Author Contributions

Conceptualization, M.A.E., A.D., A.M., R.A.I. and A.O.A.; methodology, M.A.E., A.D., A.M., R.A.I. and A.O.A.; software, M.A.E., A.D., A.M. and R.A.I.; validation, M.A.E., A.D., A.M. and R.A.I.; formal analysis, M.A.E., A.D., A.M., R.A.I. and A.O.A.; investigation, M.A.E., A.D., A.M., R.A.I. and A.O.A.; writing—original draft preparation, M.A.E., A.D., A.M. and R.A.I.; writing—review and editing, M.A.E., A.D., A.M., R.A.I. and A.O.A.; visualization, M.A.E., A.D., A.M., R.A.I. and A.O.A.; supervision, M.A.E., A.D., A.M. and R.A.I.; project administration, M.A.E., A.D., A.M., R.A.I. and A.O.A.; and funding acquisition, A.O.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number (IF-PSAU-2022/01/19574).

Data Availability Statement

The data are available from the authors upon request.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

References

  1. Dao, N.N. Internet of wearable things: Advancements and benefits from 6G technologies. Future Gener. Comput. Syst. 2022, 138, 172184. [Google Scholar] [CrossRef]
  2. Koundal, D.; Sharma, B.; Guo, Y. Intuitionistic based segmentation of thyroid nodules in ultrasound images. Comput. Biol. Med. 2020, 121, 103776. [Google Scholar] [CrossRef]
  3. Singh, K.; Sharma, B.; Singh, J.; Srivastava, G.; Sharma, S.; Aggarwal, A.; Cheng, X. Local statistics-based speckle reducing bilateral filter for medical ultrasound images. Mob. Netw. Appl. 2020, 25, 2367–2389. [Google Scholar] [CrossRef]
  4. Zhang, G.; Navimipour, N.J. A comprehensive and systematic review of the IoT-based medical management systems: Applications, techniques, trends and open issues. Sustain. Cities Soc. 2022, 2022, 103914. [Google Scholar] [CrossRef]
  5. Krishnadas, P.; Chadaga, K.; Sampathila, N.; Rao, S.; Prabhu, S. Classification of Malaria Using Object Detection Models. Informatics 2022, 9, 76. [Google Scholar] [CrossRef]
  6. Sampathila, N.; Chadaga, K.; Goswami, N.; Chadaga, R.P.; Pandya, M.; Prabhu, S.; Bairy, M.G.; Katta, S.S.; Bhat, D.; Upadya, S.P. Customized Deep Learning Classifier for Detection of Acute Lymphoblastic Leukemia Using Blood Smear Images. Healthcare 2022, 10, 1812. [Google Scholar] [CrossRef]
  7. Acharya, V.; Dhiman, G.; Prakasha, K.; Bahadur, P.; Choraria, A.; Prabhu, S.; Chadaga, K.; Viriyasitavat, W.; Kautish, S.; Sushobhitha , M.; et al. AI-assisted tuberculosis detection and classification from chest X-rays using a deep learning normalization-free network model. Comput. Intell. Neurosci. 2022, 2022, 2399428. [Google Scholar] [CrossRef]
  8. Faruqui, N.; Yousuf, M.A.; Whaiduzzaman, M.; Azad, A.; Barros, A.; Moni, M.A. LungNet: A hybrid deep-CNN model for lung cancer diagnosis using CT and wearable sensor-based medical IoT data. Comput. Biol. Med. 2021, 139, 104961. [Google Scholar] [CrossRef]
  9. Hu, M.; Zhong, Y.; Xie, S.; Lv, H.; Lv, Z. Fuzzy system based medical image processing for brain disease prediction. Front. Neurosci. 2021, 15, 714318. [Google Scholar] [CrossRef]
  10. Aurna, N.F.; Yousuf, M.A.; Taher, K.A.; Azad, A.; Moni, M.A. A classification of MRI brain tumor based on two stage feature level ensemble of deep CNN models. Comput. Biol. Med. 2022, 146, 105539. [Google Scholar] [CrossRef]
  11. Yang, J.; Shi, R.; Wei, D.; Liu, Z.; Zhao, L.; Ke, B.; Pfister, H.; Ni, B. MedMNIST v2-A large-scale lightweight benchmark for 2D and 3D biomedical image classification. Sci. Data 2023, 10, 41. [Google Scholar] [CrossRef]
  12. Nayak, S.; Patgiri, R. 6G communication technology: A vision on intelligent healthcare. In Health Informatics: A Computational Perspective in Healthcare; Springer: Berlin/Heidelberg, Germany, 2021; pp. 1–18. [Google Scholar]
  13. Eid, M.M.; Rashed, A.N.Z.; Bulbul, A.A.M.; Podder, E. Mono-rectangular core photonic crystal fiber (MRC-PCF) for skin and blood cancer detection. Plasmonics 2021, 16, 717–727. [Google Scholar] [CrossRef]
  14. Jin, B.; Zhao, Y.; Liang, Y. Internet of things medical image detection and pediatric renal failure dialysis complicated with respiratory tract infection. Microprocess. Microsyst. 2021, 83, 104016. [Google Scholar] [CrossRef]
  15. Wang, W.; Liu, F.; Zhi, X.; Zhang, T.; Huang, C. An Integrated deep learning algorithm for detecting lung nodules with low-dose CT and its application in 6G-enabled internet of medical things. IEEE Internet Things J. 2020, 8, 5274–5284. [Google Scholar] [CrossRef]
  16. Abd Elaziz, M.; Mabrouk, A.; Dahou, A.; Chelloug, S.A. Medical Image Classification Utilizing Ensemble Learning and Levy Flight-Based Honey Badger Algorithm on 6G-Enabled Internet of Things. Comput. Intell. Neurosci. 2022, 2022, 5830766. [Google Scholar] [CrossRef]
  17. Mabrouk, A.; Dahou, A.; Elaziz, M.A.; Díaz Redondo, R.P.; Kayed, M. Medical Image Classification Using Transfer Learning and Chaos Game Optimization on the Internet of Medical Things. Comput. Intell. Neurosci. 2022, 2022, 9112634. [Google Scholar] [CrossRef]
  18. Liu, Y.; Wang, J.; Li, J.; Niu, S.; Song, H. Machine learning for the detection and identification of internet of things (iot) devices: A survey. arXiv 2021, arXiv:2101.10181. [Google Scholar]
  19. Karimi, D.; Warfield, S.K.; Gholipour, A. Transfer learning in medical image segmentation: New insights from analysis of the dynamics of model parameters and learned representations. Artif. Intell. Med. 2021, 116, 102078. [Google Scholar] [CrossRef]
  20. da Silva, G.L.F.; Valente, T.L.A.; Silva, A.C.; de Paiva, A.C.; Gattass, M. Convolutional neural network-based PSO for lung nodule false positive reduction on CT images. Comput. Methods Progr. Biomed. 2018, 162, 109–118. [Google Scholar] [CrossRef]
  21. Vijh, S.; Sharma, S.; Gaurav, P. Brain tumor segmentation using OTSU embedded adaptive particle swarm optimization method and convolutional neural network. In Data Visualization and Knowledge Engineering; Springer: Berlin/Heidelberg, Germany, 2020; pp. 171–194. [Google Scholar]
  22. Onay, F.K.; Aydemir, S.B. Chaotic hunger games search optimization algorithm for global optimization and engineering problems. Math. Comput. Simul. 2022, 192, 514–536. [Google Scholar] [CrossRef]
  23. Adel, H.; Dahou, A.; Mabrouk, A.; Abd Elaziz, M.; Kayed, M.; El-Henawy, I.M.; Alshathri, S.; Amin Ali, A. Improving crisis events detection using distilbert with hunger games search algorithm. Mathematics 2022, 10, 447. [Google Scholar] [CrossRef]
  24. Yang, Y.; Wu, Y.; Yuan, H.; Khishe, M.; Mohammadi, M. Nodes clustering and multi-hop routing protocol optimization using hybrid chimp optimization and hunger games search algorithms for sustainable energy efficient underwater wireless sensor networks. Sustain. Comput. Inform. Syst. 2022, 35, 100731. [Google Scholar] [CrossRef]
  25. Devi, R.M.; Premkumar, M.; Jangir, P.; Kumar, B.S.; Alrowaili, D.; Nisar, K.S. BHGSO: Binary hunger games search optimization algorithm for feature selection problem. CMC-Comput. Mater. Contin. 2022, 70, 557–579. [Google Scholar]
  26. Fahim, S.R.; Hasanien, H.M.; Turky, R.A.; Alkuhayli, A.; Al-Shamma’a, A.A.; Noman, A.M.; Tostado-Véliz, M.; Jurado, F. Parameter identification of proton exchange membrane fuel cell based on hunger games search algorithm. Energies 2021, 14, 5022. [Google Scholar] [CrossRef]
  27. Kaveh, A.; Hamedani, K.B. Improved arithmetic optimization algorithm and its application to discrete structural optimization. Structures 2022, 35, 748–764. [Google Scholar] [CrossRef]
  28. Khatir, S.; Tiachacht, S.; Le Thanh, C.; Ghandourah, E.; Mirjalili, S.; Wahab, M.A. An improved Artificial Neural Network using Arithmetic Optimization Algorithm for damage assessment in FGM composite plates. Compos. Struct. 2021, 273, 114287. [Google Scholar] [CrossRef]
  29. Wang, R.B.; Wang, W.F.; Xu, L.; Pan, J.S.; Chu, S.C. An adaptive parallel arithmetic optimization algorithm for robot path planning. J. Adv. Transp. 2021, 2021, 3606895. [Google Scholar] [CrossRef]
  30. Dahou, A.; Al-qaness, M.A.; Abd Elaziz, M.; Helmi, A. Human activity recognition in IoHT applications using arithmetic optimization algorithm and deep learning. Measurement 2022, 199, 111445. [Google Scholar] [CrossRef]
  31. Khodadadi, N.; Snasel, V.; Mirjalili, S. Dynamic arithmetic optimization algorithm for truss optimization under natural frequency constraints. IEEE Access 2022, 10, 16188–16208. [Google Scholar] [CrossRef]
  32. Gupta, K.D.; Sharma, D.K.; Ahmed, S.; Gupta, H.; Gupta, D.; Hsu, C.H. A Novel Lightweight Deep Learning-Based Histopathological Image Classification Model for IoMT. Neural Process. Lett. 2021, 2021, 1–24. [Google Scholar]
  33. Sekhar, A.; Biswas, S.; Hazra, R.; Sunaniya, A.K.; Mukherjee, A.; Yang, L. Brain tumor classification using fine-tuned GoogLeNet features and machine learning algorithms: IoMT enabled CAD system. IEEE J. Biomed. Health Inform. 2021, 26, 983–991. [Google Scholar] [CrossRef]
  34. Abeltino, A.; Bianchetti, G.; Serantoni, C.; Ardito, C.F.; Malta, D.; De Spirito, M.; Maulucci, G. Personalized Metabolic Avatar: A Data Driven Model of Metabolism for Weight Variation Forecasting and Diet Plan Evaluation. Nutrients 2022, 14, 3520. [Google Scholar] [CrossRef]
  35. Bianchetti, G.; Abeltino, A.; Serantoni, C.; Ardito, F.; Malta, D.; De Spirito, M.; Maulucci, G. Personalized self-monitoring of energy balance through integration in a web-application of dietary, anthropometric, and physical activity data. J. Pers. Med. 2022, 12, 568. [Google Scholar] [CrossRef]
  36. Serantoni, C.; Zimatore, G.; Bianchetti, G.; Abeltino, A.; De Spirito, M.; Maulucci, G. Unsupervised clustering of heartbeat dynamics allows for real time and personalized improvement in cardiovascular fitness. Sensors 2022, 22, 3974. [Google Scholar] [CrossRef]
  37. Rodrigues, D.D.A.; Ivo, R.F.; Satapathy, S.C.; Wang, S.; Hemanth, J.; Reboucas Filho, P.P. A new approach for classification skin lesion based on transfer learning, deep learning, and IoT system. Pattern Recognit. Lett. 2020, 136, 8–15. [Google Scholar] [CrossRef]
  38. Han, T.; Nunes, V.X.; Souza, L.F.D.F.; Marques, A.G.; Silva, I.C.L.; Junior, M.A.A.F.; Sun, J.; Reboucas Filho, P.P. Internet of Medical Things—Based on Deep Learning Techniques for Segmentation of Lung and Stroke Regions in CT Scans. IEEE Access 2020, 8, 71117–71135. [Google Scholar] [CrossRef]
  39. Bianchetti, G.; Taralli, S.; Vaccaro, M.; Indovina, L.; Mattoli, M.; Capotosti, A.; Scolozzi, V.; Calcagni, M.L.; Giordano, A.; De Spirito, M.; et al. Automated detection and classification of tumor histotypes on dynamic PET imaging data through machine-learning driven voxel classification. Comput. Biol. Med. 2022, 145, 105423. [Google Scholar] [CrossRef]
  40. Hossen, M.N.; Panneerselvam, V.; Koundal, D.; Ahmed, K.; Bui, F.M.; Ibrahim, S.M. Federated machine learning for detection of skin diseases and enhancement of internet of medical things (IoMT) security. IEEE J. Biomed. Health Inform. 2022, 27, 835–841. [Google Scholar] [CrossRef]
  41. Jain, S.; Nehra, M.; Kumar, R.; Dilbaghi, N.; Hu, T.; Kumar, S.; Kaushik, A.; Li, C.Z. Internet of medical things (IoMT)-integrated biosensors for point-of-care testing of infectious diseases. Biosens. Bioelectron. 2021, 179, 113074. [Google Scholar] [CrossRef]
  42. Han, X.; Zhang, Z.; Ding, N.; Gu, Y.; Liu, X.; Huo, Y.; Qiu, J.; Yao, Y.; Zhang, A.; Zhang, L.; et al. Pre-trained models: Past, present and future. AI Open 2021, 2, 225–250. [Google Scholar] [CrossRef]
  43. Cheplygina, V.; de Bruijne, M.; Pluim, J.P. Not-so-supervised: A survey of semi-supervised, multi-instance, and transfer learning in medical image analysis. Med. Image Anal. 2019, 54, 280–296. [Google Scholar] [CrossRef] [Green Version]
  44. Ayan, E.; Ünver, H.M. Diagnosis of pneumonia from chest X-ray images using deep learning. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019; pp. 1–5. [Google Scholar]
  45. Ge, Z.; Demyanov, S.; Bozorgtabar, B.; Abedini, M.; Chakravorty, R.; Bowling, A.; Garnavi, R. Exploiting local and generic features for accurate skin lesions classification using clinical and dermoscopy imaging. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 986–990. [Google Scholar]
  46. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  47. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef]
  48. Guo, Y.; Ashour, A.S.; Si, L.; Mandalaywala, D.P. Multiple convolutional neural network for skin dermoscopic image classification. In Proceedings of the 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 6–8 December 2018; pp. 365–369. [Google Scholar]
  49. Kawahara, J.; BenTaieb, A.; Hamarneh, G. Deep features to classify skin lesions. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 1397–1400. [Google Scholar]
  50. Lopez, A.R.; Giro-i Nieto, X.; Burdick, J.; Marques, O. Skin lesion classification from dermoscopic images using deep learning techniques. In Proceedings of the 2017 13th IASTED International Conference on Biomedical Engineering (BioMed), Innsbruck, Austria, 20–21 February 2017; pp. 49–54. [Google Scholar]
  51. Ayan, E.; Ünver, H.M. Data augmentation importance for classification of skin lesions via deep learning. In Proceedings of the 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; pp. 1–4. [Google Scholar]
  52. Zhang, J.; Xie, Y.; Wu, Q.; Xia, Y. Medical image classification using synergic deep learning. Med. Image Anal. 2019, 54, 10–19. [Google Scholar] [CrossRef]
  53. Hosseinzadeh Taher, M.R.; Haghighi, F.; Feng, R.; Gotway, M.B.; Liang, J. A systematic benchmarking analysis of transfer learning for medical image analysis. In Domain Adaptation and Representation Transfer, and Affordable Healthcare and AI for Resource Diverse Global Health; Springer: Berlin/Heidelberg, Germany, 2021; pp. 3–13. [Google Scholar]
  54. Dalmaz, O.; Yurt, M.; Çukur, T. ResViT: Residual vision transformers for multimodal medical image synthesis. IEEE Trans. Med. Imaging 2022, 41, 2598–2614. [Google Scholar] [CrossRef]
  55. Anwar, S.M.; Majid, M.; Qayyum, A.; Awais, M.; Alnowami, M.; Khan, M.K. Medical image analysis using convolutional neural networks: A review. J. Med. Syst. 2018, 42, 226. [Google Scholar] [CrossRef] [Green Version]
  56. Samala, R.K.; Chan, H.P.; Hadjiiski, L.M.; Helvie, M.A.; Richter, C.; Cha, K. Evolutionary pruning of transfer learned deep convolutional neural network for breast cancer diagnosis in digital breast tomosynthesis. Phys. Med. Biol. 2018, 63, 095005. [Google Scholar] [CrossRef]
  57. Shankar, K.; Lakshmanaprabu, S.; Khanna, A.; Tanwar, S.; Rodrigues, J.J.; Roy, N.R. Alzheimer detection using Group Grey Wolf Optimization based features with convolutional classifier. Comput. Electr. Eng. 2019, 77, 230–243. [Google Scholar]
  58. Goel, T.; Murugan, R.; Mirjalili, S.; Chakrabartty, D.K. OptCoNet: An optimized convolutional neural network for an automatic diagnosis of COVID-19. Appl. Intell. 2021, 51, 1351–1366. [Google Scholar] [CrossRef]
  59. Elhoseny, M.; Shankar, K. Optimal bilateral filter and convolutional neural network based denoising method of medical image measurements. Measurement 2019, 143, 125–135. [Google Scholar] [CrossRef]
  60. Zhang, N.; Cai, Y.X.; Wang, Y.Y.; Tian, Y.T.; Wang, X.L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 101756. [Google Scholar] [CrossRef]
  61. El-Shafeiy, E.; Sallam, K.M.; Chakrabortty, R.K.; Abohany, A.A. A clustering based Swarm Intelligence optimization technique for the Internet of Medical Things. Expert Syst. Appl. 2021, 173, 114648. [Google Scholar] [CrossRef]
  62. Yu, Z.; Jiang, X.; Zhou, F.; Qin, J.; Ni, D.; Chen, S.; Lei, B.; Wang, T. Melanoma recognition in dermoscopy images via aggregated deep convolutional features. IEEE Trans. Biomed. Eng. 2018, 66, 1006–1016. [Google Scholar] [CrossRef]
  63. Yu, Z.; Jiang, F.; Zhou, F.; He, X.; Ni, D.; Chen, S.; Wang, T.; Lei, B. Convolutional descriptors aggregation via cross-net for skin lesion recognition. Appl. Soft Comput. 2020, 92, 106281. [Google Scholar] [CrossRef]
  64. Wei, L.; Ding, K.; Hu, H. Automatic skin cancer detection in dermoscopy images based on ensemble lightweight deep learning network. IEEE Access 2020, 8, 99633–99647. [Google Scholar] [CrossRef]
  65. Ozkan, I.A.; Koklu, M. Skin lesion classification using machine learning algorithms. Int. J. Intell. Syst. Appl. Eng. 2017, 5, 285–289. [Google Scholar] [CrossRef] [Green Version]
  66. Al Nazi, Z.; Abir, T.A. Automatic skin lesion segmentation and melanoma detection: Transfer learning approach with u-net and dcnn-svm. In Proceedings of the International Joint Conference on Computational Intelligence, Budapest, Hungary, 2–4 November 2020; pp. 371–381. [Google Scholar]
  67. Afza, F.; Sharif, M.; Mittal, M.; Khan, M.A.; Hemanth, D.J. A hierarchical three-step superpixels and deep learning framework for skin lesion classification. Methods 2021, 202, 88–102. [Google Scholar] [CrossRef]
  68. Habibzadeh, M.; Krzyżak, A.; Fevens, T. White blood cell differential counts using convolutional neural networks for low resolution images. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, 9–13 June 2013; pp. 263–274. [Google Scholar]
  69. Zhao, J.; Zhang, M.; Zhou, Z.; Chu, J.; Cao, F. Automatic detection and classification of leukocytes using convolutional neural networks. Med. Biol. Eng. Comput. 2017, 55, 1287–1301. [Google Scholar] [CrossRef]
  70. Sharma, M.; Bhave, A.; Janghel, R.R. White blood cell classification using convolutional neural network. In Soft Computing and Signal Processing; Springer: Berlin/Heidelberg, Germany, 2019; pp. 135–143. [Google Scholar]
  71. Mabrouk, A.; Díaz Redondo, R.P.; Dahou, A.; Abd Elaziz, M.; Kayed, M. Pneumonia Detection on Chest X-ray Images Using Ensemble of Deep Convolutional Neural Networks. Appl. Sci. 2022, 12, 6448. [Google Scholar] [CrossRef]
  72. Ignatov, A.; Romero, A.; Kim, H.; Timofte, R. Real-time video super-resolution on smartphones with deep learning, mobile ai 2021 challenge: Report. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2535–2544. [Google Scholar]
  73. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
  74. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  75. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 1–26 July 2016; pp. 770–778. [Google Scholar]
  76. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  77. Yang, Y.; Chen, H.; Heidari, A.A.; Gandomi, A.H. Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst. Appl. 2021, 177, 114864. [Google Scholar] [CrossRef]
  78. Giordani, M.; Polese, M.; Mezzavilla, M.; Rangan, S.; Zorzi, M. Toward 6G networks: Use cases and technologies. IEEE Commun. Mag. 2020, 58, 55–61. [Google Scholar] [CrossRef]
  79. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH 2-A dermoscopic image database for research and benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar]
  80. Gutman, D.; Codella, N.C.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin lesion analysis toward melanoma detection: A challenge at the international symposium on biomedical imaging (ISBI) 2016, hosted by the international skin imaging collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar]
  81. Liang, G.; Hong, H.; Xie, W.; Zheng, L. Combining convolutional neural network with recursive neural network for blood cell image classification. IEEE Access 2018, 6, 36188–36197. [Google Scholar] [CrossRef]
  82. Abualigah, L.; Yousri, D.; Abd Elaziz, M.; Ewees, A.A.; Al-qaness, M.A.; Gandomi, A.H. Aquila Optimizer: A novel meta-heuristic optimization Algorithm. Comput. Ind. Eng. 2021, 157, 107250. [Google Scholar] [CrossRef]
  83. Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  84. Yang, X.S. A new metaheuristic bat-inspired algorithm. In Nature inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. [Google Scholar]
  85. Hashim, F.A.; Hussain, K.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W. Archimedes optimization algorithm: A new metaheuristic algorithm for solving optimization problems. Appl. Intell. 2021, 51, 1531–1551. [Google Scholar] [CrossRef]
  86. Talatahari, S.; Azizi, M. Chaos Game Optimization: A novel metaheuristic algorithm. Artif. Intell. Rev. 2021, 54, 917–1004. [Google Scholar] [CrossRef]
  87. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef]
  88. Fang, L.; Jin, Y.; Huang, L.; Guo, S.; Zhao, G.; Chen, X. Iterative fusion convolutional neural networks for classification of optical coherence tomography images. J. Vis. Commun. Image Represent. 2019, 59, 327–333. [Google Scholar] [CrossRef]
  89. Huang, L.; He, X.; Fang, L.; Rabbani, H.; Chen, X. Automatic classification of retinal optical coherence tomography images with layer guided convolutional neural network. IEEE Signal Process. Lett. 2019, 26, 1026–1030. [Google Scholar] [CrossRef]
  90. Sun, Y.; Li, S.; Sun, Z. Fully automated macular pathology detection in retina optical coherence tomography images using sparse coding and dictionary learning. J. Biomed. Opt. 2017, 22, 016012. [Google Scholar] [CrossRef] [Green Version]
  91. Ji, Q.; He, W.; Huang, J.; Sun, Y. Efficient deep learning-based automated pathology identification in retinal optical coherence tomography images. Algorithms 2018, 11, 88. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flowchart showing the developed FS algorithm.
Figure 1. Flowchart showing the developed FS algorithm.
Diagnostics 13 00834 g001
Figure 2. The suggested 6G-enabled IoMT framework diagram.
Figure 2. The suggested 6G-enabled IoMT framework diagram.
Diagnostics 13 00834 g002
Figure 3. Sample of images from: ISIC, PH2, WBC, and OCT datasets.
Figure 3. Sample of images from: ISIC, PH2, WBC, and OCT datasets.
Diagnostics 13 00834 g003
Figure 4. Average results from the four classifiers on the selected datasets.
Figure 4. Average results from the four classifiers on the selected datasets.
Diagnostics 13 00834 g004
Figure 5. Average accuracy on the four classifiers.
Figure 5. Average accuracy on the four classifiers.
Diagnostics 13 00834 g005
Figure 6. Average execution time of the classifiers across the datasets.
Figure 6. Average execution time of the classifiers across the datasets.
Diagnostics 13 00834 g006
Figure 7. Average execution times of the SVM across the datasets.
Figure 7. Average execution times of the SVM across the datasets.
Diagnostics 13 00834 g007
Figure 8. Average accuracy of the SVM across the datasets.
Figure 8. Average accuracy of the SVM across the datasets.
Diagnostics 13 00834 g008
Table 2. Classification results (%) of each FS algorithm using two skin datasets (ISIC and PH2). ET is the execution time.
Table 2. Classification results (%) of each FS algorithm using two skin datasets (ISIC and PH2). ET is the execution time.
Alg.Cls.ISICPH2
ACBAFRPETACBAFRPET
PSOSVM85.7572.0484.7885.7584.680.1396.0796.7396.0796.0796.100.12
XGB84.4373.7284.1484.4383.920.2692.8693.4592.8892.8693.403.61
KNN84.1771.5583.5384.1783.210.0895.7196.1395.7295.7195.770.24
RF85.2274.7284.9085.2284.680.3391.7991.6791.8591.7992.600.56
GWOSVM84.4371.2183.6584.4383.340.1696.0796.7396.0796.0796.100.09
XGB82.8571.7382.6282.8582.430.2292.8693.7592.8592.8693.432.57
KNN84.1772.5583.7384.1783.450.0695.7196.1395.7295.7195.770.17
RF85.2273.7184.7285.2284.460.3091.7991.6791.8591.7992.600.48
MFOSVM86.5473.0385.5786.5485.600.1596.0796.7396.0796.0796.100.09
XGB85.4976.3985.3885.4985.280.2393.2194.0593.2293.2193.582.63
KNN82.5971.0782.3182.5982.080.0695.7196.1395.7295.7195.770.17
RF85.2273.7184.7285.2284.460.3192.1491.9692.2192.1492.870.50
ArchOASVM86.2874.8885.7386.2885.510.0996.0796.7396.0796.0796.100.06
XGB83.6473.2383.4783.6483.320.1991.7992.5691.8091.7992.601.69
KNN84.9674.0584.5984.9684.350.0595.3695.8395.3795.3695.400.11
RF85.7574.5585.2785.7585.030.2993.5793.7593.6193.5794.000.45
OASVM85.4973.3884.8685.4984.600.1496.0796.7396.0796.0796.070.06
XGB84.7073.3984.2784.7084.010.2793.5794.3593.5893.5793.971.81
KNN85.4974.3885.0485.4984.790.0795.7196.1395.7295.7195.770.12
RF86.2875.8885.9086.2885.690.3392.1491.9692.2192.1492.870.47
BATSVM86.0272.2085.0086.0284.970.1096.0796.7396.0796.0796.100.08
XGB85.7575.0585.3685.7585.130.1992.8693.7592.8692.8693.282.48
KNN82.3268.3981.5582.3281.120.0595.7196.1395.7295.7195.770.16
RF85.7575.0585.3685.7585.130.2892.5092.2692.5692.5093.150.46
HGSSVM84.9673.0584.4084.9684.120.1296.0796.7396.0796.0796.070.07
XGB85.4975.3985.2285.4985.020.2392.5093.1592.5292.5093.152.07
KNN84.9673.0584.4084.9684.120.0795.7196.1395.7295.7195.770.15
RF84.9674.0584.5984.9684.350.3092.1492.5692.1892.1492.850.47
CGOSVM86.0274.7185.5086.0285.270.1496.0796.7396.0796.0796.100.07
XGB84.9673.0584.4084.9684.120.2293.5794.3593.5893.5793.972.15
KNN84.9673.5584.5084.9684.230.0695.7196.1395.7295.7195.770.14
RF85.2273.2184.6385.2284.360.3092.8692.5692.9192.8693.440.46
AOASVM86.2874.8885.7386.2885.510.1696.0796.7396.0796.0796.100.11
XGB85.7573.5485.0885.7584.850.2593.9394.6493.9493.9394.272.86
KNN84.9671.0483.9984.9683.780.0795.7196.1395.7295.7195.770.18
RF85.4973.8884.9585.4984.690.3191.4391.3791.5091.4392.330.50
AOAHGSVM87.3474.5386.4787.3486.530.0696.4397.0296.4396.4396.440.10
XGB84.4372.7283.9684.4383.670.0993.5794.3593.5893.5793.973.01
KNN84.1772.5583.7384.1783.450.0495.7196.1395.7295.7195.770.19
RF85.7575.0585.3685.7585.130.2691.7991.6791.8591.7992.600.53
Table 3. Classification Results (%) of each FS algorithm using WBC dataset and OCT dataset. ET is the execution time.
Table 3. Classification Results (%) of each FS algorithm using WBC dataset and OCT dataset. ET is the execution time.
Alg.Cls.WBCOCT
ACBAFRPETACBAFRPET
PSOSVM88.4688.4688.6588.4690.491.199.2899.2899.2899.2899.3016
XGB88.4288.4188.6488.4290.6058.099.1799.1799.1899.1799.20178
KNN88.4288.4288.6188.4290.448.499.2899.2899.2899.2899.303
RF88.4688.4688.6688.4690.536.499.2899.2899.2899.2899.30104
GWOSVM88.5088.5088.6988.5090.510.999.3899.3899.3899.3899.4011
XGB88.4288.4288.6588.4290.6344.399.3899.3899.3899.3899.40137
KNN88.5088.5088.6988.5090.556.599.1799.1799.1899.1799.202
RF88.4288.4188.6188.4290.435.599.3899.3899.3899.3899.4090
MFOSVM88.5488.5488.7488.5490.590.899.3899.3899.3899.3899.409
XGB88.5088.5088.7188.5090.5842.599.3899.3899.3899.3899.40119
KNN88.5088.5088.7088.5090.606.199.1799.1799.1899.1799.202
RF88.4688.4688.6688.4690.555.499.1799.1799.1899.1799.2089
ArchOASVM88.2688.2588.4488.2690.320.499.3899.3899.3899.3899.407
XGB88.3088.3088.5288.3090.4515.799.2899.2899.2899.2899.3090
KNN88.5888.5888.7588.5890.551.999.1799.1799.1899.1799.201
RF88.4688.4688.6788.4690.583.399.1799.1799.1899.1799.2074
AOSVM88.2688.2588.4788.2690.440.699.2899.2899.2899.2899.309
XGB88.3488.3388.5688.3490.4829.999.3899.3899.3899.3899.40125
KNN88.4688.4688.6588.4690.514.099.2899.2899.2899.2899.302
RF88.3488.3388.5588.3490.484.499.4899.4899.4899.4899.4976
BATSVM88.3488.3488.5188.3490.230.899.3899.3899.3899.3899.4012
XGB88.0688.0588.3188.0690.4243.099.2899.2899.2899.2899.30141
KNN88.4688.4688.6388.4690.416.599.4899.4899.4899.4899.492
RF88.3088.2988.5288.3090.445.599.2899.2899.2899.2899.3088
HGSSVM88.5488.5488.7488.5490.620.499.5999.5999.5999.5999.599
XGB88.2688.2588.4788.2690.3920.599.3899.3899.3899.3899.40122
KNN88.3888.3888.5888.3890.462.799.1799.1799.1899.1799.202
RF88.4688.4688.6788.4690.624.099.3899.3899.3899.3899.4086
CGOSVM88.5088.5088.6888.5090.491.099.5999.5999.5999.5999.5910
XGB87.9487.9388.2187.9490.4136.699.1799.1799.1899.1799.20125
KNN88.2288.2188.4388.2290.325.499.1799.1799.1899.1799.202
RF88.2288.2188.4488.2290.415.299.1799.1799.1899.1799.2086
AOASVM88.5888.5888.7688.5890.570.899.3899.3899.3899.3899.4012
XGB88.4288.4188.6288.4290.5447.699.1799.1799.1899.1799.20142
KNN88.4288.4288.6188.4290.447.499.2899.2899.2899.2899.302
RF88.4288.4188.6288.4290.515.999.0799.0799.0799.0799.1086
AOAHGSVM88.6288.6288.8088.6290.591.099.6999.6999.6999.6999.696
XGB88.3088.3088.5188.3090.4048.799.3899.3899.3899.3899.4068
KNN88.4688.4688.6388.4690.426.799.5999.5999.5999.5999.591
RF88.3888.3788.5988.3890.525.899.3899.3899.3899.3899.4062
Table 4. Accuracy (AC) results of the developed method and other existing methods.
Table 4. Accuracy (AC) results of the developed method and other existing methods.
DSModelAC (%)YearRef.
ISICBL-CNN85.002017[45]
DCNN-FV86.812018[62]
MC-CNN86.302019[52]
MFA86.812020[63]
AOAHG + SVM87.30presentOurs
PH2ANN92.502017[65]
DenseNet + SVM92.002020[66]
DenseNet + KNN93.162020[37]
ResNet + NB95.402021[67]
AOAHG + SVM96.40presentOurs
WBCCNN + SVM85.002013[68]
CNN87.082017[69]
CNN + Augm87.002019[70]
AOAHG + SVM88.60presentOurs
OCTTransfer Learning80.302018[87]
IFCNN87.302019[88]
LGCNN89.902019[89]
IBDL94.572018[87]
ScSPM97.752017[90]
InceptionV398.862018[91]
AOAHG + SVM99.69presentOurs
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elaziz, M.A.; Dahou, A.; Mabrouk, A.; Ibrahim, R.A.; Aseeri, A.O. Medical Image Classifications for 6G IoT-Enabled Smart Health Systems. Diagnostics 2023, 13, 834. https://doi.org/10.3390/diagnostics13050834

AMA Style

Elaziz MA, Dahou A, Mabrouk A, Ibrahim RA, Aseeri AO. Medical Image Classifications for 6G IoT-Enabled Smart Health Systems. Diagnostics. 2023; 13(5):834. https://doi.org/10.3390/diagnostics13050834

Chicago/Turabian Style

Elaziz, Mohamed Abd, Abdelghani Dahou, Alhassan Mabrouk, Rehab Ali Ibrahim, and Ahmad O. Aseeri. 2023. "Medical Image Classifications for 6G IoT-Enabled Smart Health Systems" Diagnostics 13, no. 5: 834. https://doi.org/10.3390/diagnostics13050834

APA Style

Elaziz, M. A., Dahou, A., Mabrouk, A., Ibrahim, R. A., & Aseeri, A. O. (2023). Medical Image Classifications for 6G IoT-Enabled Smart Health Systems. Diagnostics, 13(5), 834. https://doi.org/10.3390/diagnostics13050834

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop