Next Article in Journal
Biomechanics Assist Measurement, Modeling, Engineering Applications, and Clinical Decision Making in Medicine
Previous Article in Journal
Deep Learning for Detecting COVID-19 Using Medical Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization

by
Hanaa ZainEldin
1,
Samah A. Gamel
1,
El-Sayed M. El-Kenawy
2,*,
Amal H. Alharbi
3,
Doaa Sami Khafaga
3,*,
Abdelhameed Ibrahim
1,* and
Fatma M. Talaat
4
1
Computer Engineering and Control Systems Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
2
Department of Communications and Electronics, Delta Higher Institute of Engineering and Technology, Mansoura 35111, Egypt
3
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
Machine Learning & Information Retrieval Department, Faculty of Artificial Intelligence, Kafrelsheikh University, Kafrelsheikh 33511, Egypt
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(1), 18; https://doi.org/10.3390/bioengineering10010018
Submission received: 17 October 2022 / Revised: 12 November 2022 / Accepted: 17 December 2022 / Published: 22 December 2022
(This article belongs to the Section Biochemical Engineering)

Abstract

:
Diagnosing a brain tumor takes a long time and relies heavily on the radiologist’s abilities and experience. The amount of data that must be handled has increased dramatically as the number of patients has increased, making old procedures both costly and ineffective. Many researchers investigated a variety of algorithms for detecting and classifying brain tumors that were both accurate and fast. Deep Learning (DL) approaches have recently been popular in developing automated systems capable of accurately diagnosing or segmenting brain tumors in less time. DL enables a pre-trained Convolutional Neural Network (CNN) model for medical images, specifically for classifying brain cancers. The proposed Brain Tumor Classification Model based on CNN (BCM-CNN) is a CNN hyperparameters optimization using an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. There is an optimization of hyperparameters followed by a training model built with Inception-ResnetV2. The model employs commonly used pre-trained models (Inception-ResnetV2) to improve brain tumor diagnosis, and its output is a binary 0 or 1 (0: Normal, 1: Tumor). There are primarily two types of hyperparameters: (i) hyperparameters that determine the underlying network structure; (ii) a hyperparameter that is responsible for training the network. The ADSCFGWO algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that uses both algorithms’ strengths. The experimental results show that the BCM-CNN as a classifier achieved the best results due to the enhancement of the CNN’s performance by the CNN optimization’s hyperparameters. The BCM-CNN has achieved 99.98% accuracy with the BRaTS 2021 Task 1 dataset.

1. Introduction

Recently, digital medical images have been essential for detecting numerous illnesses. It is additionally used for training and research. The need for electronic medical images is growing dramatically; for example, in 2002, the Department of Radiology at the University Hospital of Geneva produced between 12,000 and 15,000 images daily [1]. An efficient and exact computer-aided diagnostic system is required for medical report creation and medical image research. The old method of manually evaluating medical imaging is time-consuming, inaccurate, and prone to human error. Over the medical diseases, the brain tumor has become a serious issue, ranking 10th among the major causes of death in the US. It is reported that 700,000 persons have brain tumors, of which 80 percent are benign and 20 percent are malignant [2]. According to estimates by the American Cancer Society from 2021, 78,980 adults have been diagnosed with a brain tumor, with 55,150 noncancerous and 24,530 malignant tumors (13,840 men and 10,690 females) [3]. According to studies, brain tumor is the top cause of cancer deaths in children and adults worldwide [4].
The most typical kind of brain disease is a brain tumor. It is an unregulated development of brain cells. Brain tumors are always classified into brain tumors, both primary and secondary. The first starts in the brain and usually stays there, whereas the latter starts as cancer somewhere else in the body and spreads to the brain [5]. There are two different forms of tumors: malignant and benign. A benign tumor is a slow-growing tumor that does not infiltrate nearby tissues; on the other hand, a malignant which is a very aggressive tumor that spreads from one location to another. The World Health Organization (WHO) grades a brain tumor as I-IV. Tumors in categories I and II are regarded as slow-growing, while tumors in categories III and IV are always malignant and have a worse prognosis [6].
In recent decades, many imaging techniques such as X-ray, Magneto Encephalo Graphy (MEG), Computed Tomography (CT), Ultrasonography, Electro Encephalo Graphy (EEG), Single-Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), and Magnetic Resonance Imaging (MRI) have emerged that not only exhibit the detailed and complete facets of brain tumors but also help doctors to accurately diagnose the tumor and determine the correct treatment mechanism [4]. MRI is considered the most popular imaging technique for detecting brain tumors [7]. Without subjecting the patients to excessive ionization radiation, MRI is a non and excellent soft tissue contrast imaging technique that gives essential information about brain tumor shape, location, and size.
The brain tumor diagnosis is highly time intensive and largely depends on the radiologist’s skills and knowledge. Because there are more patients, the amount of data that must be processed has grown significantly, making traditional techniques cost and incorrect [8]. The difficulties are associated with significant brain tumor size, shape, and intensity variations for the same tumor type and similar manifestations of other disease types. A misclassification of a brain tumor can result in major consequences and reduce the patient’s survivability. There is a rise in interest in building automated technologies for processing images to overcome the limitations of manual diagnosis [4,9] and other related applications [10,11,12]. Several systems for computer-aided diagnosis (CAD) have been created recently to diagnose brain tumors automatically.
In recent years, among many other applications, artificial intelligence (AI) has demonstrated promising results as a decision support system to assist in the detection of diseases and the establishment of precise medical diagnoses. In order to address practical problems researchers and governments focus on machine learning, a branch of artificial intelligence [13,14]. Machine learning, for instance, may predict the COVID-19 outbreak in the COVID-19 pandemic challenge by determining how risky the virus is and then scaling up the level of the methods performed. In the realm of medical analysis, machine learning algorithms are frequently used for things such as COVID-19 prediction [15], Alzheimer’s disease progression [16], brain tumor development [17], breast cancer progression [18], and other disorders [19,20,21]. Deep learning and machine learning are essential for identifying diseases and resolving medical problems.
Many researchers investigated numerous algorithms for detecting and classifying brain tumors with high performance and less error. Deep Learning (DL) techniques have recently been widely employed to build automatic systems that can accurately classify or segment brain tumors in less time. DL enables the use of a pre-trained Convolutional Neural Network (CNN) model [22] for medical imagery, specifically for the classification of brain tumors, which has been created for various applications, including GoogLeNet [23], AlexNet, and ResNet-34 [24]. DL is made up of a multi-layered deep neural network [25]. The backpropagation algorithm is used by a neural network (NN) to reduce the error between the target and actual value. Nevertheless, even as the number of layers increases, developing artificial neural network models gets more difficult.
The main contributions of the current work are:
  • Introduce an enhanced model to improve brain tumor diagnosis.
  • It proposes a Brain Tumor Classification Model (BCM-CNN) based on an advanced 3D model using Enhanced Convolutional Neural Network (BCM-CNN).
  • The proposed Brain Tumor Classification Model (BCM-CNN) is based on two submodules; (i) CNN hyperparameters optimization using an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm followed by trained Model, and (ii) segmentation model.
  • The ADSCFGWO algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that uses both algorithms’ strengths.
  • The experimental results show that the BCM-CNN as a classifier achieved the best results due to the enhancement of the CNN’s performance by the CNN optimization’s hyperparameters.
The remainder of the paper is structured as follows: A brief review of state-of-the-art deep learning methods for finding brain tumors is discussed in Section 2. The proposed technique is described in detail in Section 3. Section 4 depicts simulation and experimental results. The conclusion and future works are presented in Section 5.

2. Related Work

This section introduces a collection of cutting-edge DL-based brain tumor classification techniques. Based on DL and transfer learning algorithms, there are numerous methods for classifying brain tumors. State-of-the-art techniques can be classified into deep learning-based, machine learning-based, and hybrid-based techniques. Table 1 summarizes different classification techniques for a brain tumor.

2.1. Deep Learning-Based Techniques

B. Srikanth et al. presented [26] a 16-layer VGG-16 deep NN, which accepts improved images from a prior pre-processing phase as input and moves them through the convolution layer for extracting the features and downsampling (Convolution, ReLU, Max-Pooling). Their proposed approach increased the precision of brain tumor MR image multi-classification. To avoid the overfitting problem, completely linked and SoftMax layers are employed. Lastly, after 20 training iterations, their proposed model achieves the best outcomes, which have a 98 percent accuracy.
GS Tandel et al. [27] The researcher developed five clinical multiclass datasets. They used a transfer learning-based Convolutional Neural Network (CCN) to improve performance in brain tumor classification by employing MRI images. The proposed CNN model was compared to six alternative ML classification approaches, including Decision Tree (DT), Naive Bayes (NB), Linear Discrimination (LD), K-nearest Neighbor, and Support Vector Machine (SVM). The five types of multiclass classification brain tumor datasets are considered, and the proposed CNN-based (DL) model technique beats the six types of machine learning model techniques. For the five classes, the CNN-based AlexNet achieved a mean accuracy rate of 87.14, 93.74, 95.97, 96.65, and 100 percent using three different cross-validation procedures, K2, K5, and K10, respectively.
The authors in [28] proposed a CNN technique for a three-class classification to distinguish between three kinds of brain tumors, including glioma, meningioma, and pituitary tumors. They used a pre-trained GoogleNet for feature extraction from brain MRI scans. To identify the extracted features, proven-based classifications are used. The suggested approach outperforms existing approaches with an average classification accuracy of 98%. Precision, F-score, recall, specificity, and the Area Under the Curve (AUC) are performance metrics employed in the study. According to the result of the research, transfer learning techniques is a highly effective strategy when medical pictures are scarce.
Regarding a three-class brain tumor classification, ref. [29] suggested a deep inception residual network. They have adjusted the output layer of the ResNet V2 network with a dense network and a softmax layer. The suggested model maximizes brain tumor classification accuracy. The proposed model was tested on a publicly accessible brain tumor imaging dataset with 3064 pictures. The accuracy of the proposed model exceeds state-of-the-art techniques by 99.69%.
Using the concept of transfer learning, ref. [30] presented a brain tumor classification technique with MRI images. VGG16, ResNet50, DenseNet, and VGG19 networks use transfer learning to detect the most frequent brain cancers. Deep transfer learning algorithms are trained and evaluated on the publicly available Figshare dataset, which contains 3064 T1-weighted MRI scans from 233 patients with three common brain tumor types: glioma (1426 pictures), pituitary tumor (930 images), and meningioma (708 photos). The suggested model enhances the classification performance by 99.02% compared to ResNet50 and Adadelta.
The RCNN approach was used to design a new architecture for brain tumor classification that was tested using two openly accessible datasets from Figshare and Kaggle [31]. The authors presented a method for brain tumor detection that uses a low-complexity architecture to reduce the processing time of a traditional RCNN structure. Firstly, to identify glioma and healthy tumor MRI images, they used a Two-Channel CNN, a low-complex framework, which improves accuracy by 98.21%. Afterward, this framework is employed as a feature extractor in an RCNN to identify tumor areas in a Glioma MRI dataset that is categorized from a preceding phase. Lastly, the tumor region is bounded by boxes. This approach has been used for two more tumor types: meningioma and pituitary tumors. With an overall confidence level of 98.8%, their approach could achieve a low execution time in comparison to state-of-the-art techniques.
ImageNet-based Vision Transformer (ViT) models (B/16, B/32, L/16, and L/32) that have been trained and fine-tuned were proposed by [32] for brain tumor classification purposes. Validation and testing were performed on a three-classes brain tumor dataset from figshare that included 3064 T1w contrast-enhanced (CE) MRI slices with gliomas, meningiomas, and pituitary tumors. L/32 was the highest model, gaining 98.2% in the total test accuracy at a resolution of 384 × 384 . The ensembles of all four ViT algorithms showed an average testing accuracy of 98.7% at the same resolution, surpassing the performance of each algorithm at both resolutions and their ensembling at resolution 224 × 224 .

2.2. Machine Learning-Based Techniques

Pareek et al. [33] presented a method that detects if there is a tumor or not and then classifies the tumor type. The proposed method was tested on 150 T1-weighted MRI brain imaging for identifying brain tumors. The supervised approach was utilized for the classification process, and the principal component analysis was employed for feature extraction. They also assessed the tumor’s area and volume to determine the tumor’s levels. The findings of the experiments demonstrate that KSVM is 97 percent accurate in classifying brain tumors.
A novel method proposed in [34] produces excellent results and outperforms earlier techniques. To improve MRI quality and to build an exclusionary feature set, the suggested method employs normalization, densely speeded-up powerful features, and histogram of gradient approaches. In the classifying stage, they use a support vector machine. The proposed system has been tested on a significant dataset. The accuracy obtained with this method is 90.27 percent compared to state-of-the-art techniques. Regarding experimental findings, this strategy outperformed the most recent techniques. These findings were obtained by a rigorous statistical study (k-fold cross-validation), demonstrating the recommended method’s accuracy and robustness.
A quantum Fully Self neural network (QFS-Net) network using qubits/three states of quantum for segmentation of the brain lesions has been proposed [35] as a method to benefit from the capabilities of quantum correlations. The advanced quantum back-propagation approach used in supervised QINN networks is replaced with a ground-breaking supervised qutrit-based counter-propagating technique in the QFS-Net. This method enables the propagation of iterative quantum states throughout the network’s layers.

2.3. Hybrid-Based Techniques

Khairandish et al. [36] Presented a hybrid approach that combines CNN and SVM with threshold-based segmentation in terms of classification. The hybrid proposed CNN-SVM demonstrates enhanced overall accuracy with 98.4959%. To extract features from tumor regions and adjacent tissues, pre-trained AlexNet, GoogLeNet, ShuffleNet, and ResNet-18 networks are employed [37]. Although deep features are crucial in the identification process, some low-level data about tumors may be lost. As a result, a shallow network is made to learn low-level data. Deep and shallow features are blended to compensate for the loss of data. With the fused feature sets, SVM and k-NN classifiers are trained. Data augmentation and ROI expansion simultaneously enhance the average sensitivity by roughly 11.72 percent, according to experimental results. These findings support the theory that the tissues around the tumor contain significant data. Not only that, but feature fusion may substitute for missing low-level information. Furthermore, the deep feature extractor process is conducted with the ResNet-18. their experimental results are competitive compared to state-of-the-art techniques.
Authors in [38] proposed a deep learning-based automatic multimodal classification technique for categorizing different brain tumors. The suggested approach comprises five essential phases. In the first phase, an edge-based histogram and the Discrete Cosine Transform (DCT) are used to implement the linear contrasting stretching. Phase two involves the deep learning component Extractions are made. Deep learning feature extraction is performed in the second phase. Two pre-CNN networks, namely VGG16 and VGG19, were employed for feature extraction. The Extreme Learning Machine (ELM) and a correntropy-based strategy were both employed in the third stage to choose the best features. The resistant covariant features based on Partial Least Squares (PLS) were combined into one matrix. ELM received the merged matrix to conduct the classification model. The suggested technique was tested using the three datasets (BraTs2015, BraTs2017, and BraTs2018) with an accuracy of 97.8%, 96.9%, and 92.5%, respectively.
A hybrid deep learning-based technique to classify brain tumors with ISLES2015 and BRATS2015 datasets was proposed in [39]. DLS techniques such as VGG16, VGG19, and ResNet50 are used for experimental results. Then, the classifiers SoftMax, SVM-RBF, and SVM-Cubic are used to construct the multi-class classification, and performance is calculated according to the total accuracy reached by each method. The results of this study proved that VGG19 with SVM-Cubic has significantly greater accuracy (96%) than other methods.
Irmak et al. [40] proposed three-distinct Convolutional networks are suggested for three distinct classification architectures. Detection of brain tumor accuracy reaches 99.33 percent in the first CNN architecture. The accuracy of the second CNN model architecture reaches approximately 92.66 percent. The second CNN architecture can categorize brain tumors into five types: normal, meningioma, glioma, metastatic, and pituitary. With an accuracy of 98.14 percent, the third CNN architecture successfully categorizes brain tumors into Grade II, Grade III, and Grade IV. The state-of-the-art CNN algorithms such as Inceptionv3, AlexNet, ResNet-50, GoogleNet, and VGG-16 compared to the suggested CNN models. Utilizing the grid search optimization technique, all the essential model parameters of Convolutional networks are automatically identified. Publicly released clinical datasets are used to generate acceptable detection results.
Table 1. Different brain tumor classification techniques.
Table 1. Different brain tumor classification techniques.
ReferenceUsed TechniqueDatasetAccuracyAdvantagesDisadvantages
[26]16-layer VGG-16 deep NNHospitals’ dataset from 2010–2015, China98%Improve multi-class brain tumor classification accuracy.Small dataset
[27]CNN-based DL modelREMBRANDT100% for two-class classificationAI-based transfer learning surpasses machine learning for brain tumor classification.
[28]CNN technique for a three-class classificationFigshare98%Transfer learning techniques is a highly effective strategy when medical pictures are scarce.Samples from the category meningioma were misclassified. overfitting with smaller training data.
[31]RCNN-based modelOpenly accessible datasets from Figshare and Kaggle98.21%Low execution time that is optimal for real time processing. Operate on limited-resources systemsLimited to object detection and need to implement brain segmentation.
[36]Hybrid CNN-SVMBRATS 201598.49%Provide effective classification technique for brain tumorNeed to consider the size and location of brain tumor.
[37]SVM and k-NN classifiersFigshare, 201797.25%Extended ROI’s shallow and deep properties improve classifier performanceAccuracy need to be improved
[29]Deep inception residual networkPublicly accessible brain tumor imaging dataset with 3064 pictures99.69%Achieves high classification performance.Large number of parameters. Maximum computational time.
[40]CNN model for multi-classificationPublicly released clinical datasets99.33%CNN models can help doctors and radiologists validate their first brain tumor assessment
[33]Kernel-based SVMFigshare97%Can detect whether brain tumor is benign and malignant.Small dataset. Classification accuracy need to be increased.
[30]Transfer learning-based classificationFigshare99.02%Accurately detect brain tumors. Transfer learning in healthcare can help doctors make quick, accurate decisions.Classification accuracy need to be increased.
[34]Multi-classification modelThree different publicly datasets90.27%Low computational time. Help doctors in making better classification decisions for brain cancers.Classification accuracy need to be increased.
[32] ImageNet-based ViT Figshare 98.7% Accurately detect brain tumors. Helps radiologists make the right patient-based decision Needs to consider the size and location of brain tumor
[38] Deep learning-based automatic multimodal classification BraTs 2015, BraTs 2017, BraTs 2018 97.8% Feature extraction improved classification accuracy and reduced processing time Classification accuracy need to be increased
[39] Hybrid deep learning-based ISLES2015 and BRATS2015 96% Perform multi-class classification for brain tumor Classification accuracy need to be increased

3. Brain Tumor Classification Model Based CNN (BCM-CNN)

This section proposes a Brain Tumor Classification Model (BCM-CNN) based on an advanced model using a Convolutional Neural Network. The overall architecture of the proposed model is shown in Figure 1. The BCM-CNN is used to diagnose a brain tumor. It consists of a hyperparameters optimization, followed by an Inception-ResnetV2 training model. The model’s output is a binary 0 or 1 (0: Normal, 1: Tumor) and uses common pre-trained models (Inception-ResnetV2) to enhance the brain tumor diagnosis process.

3.1. CNN Hyperparameters Optimization

This subsection discusses the selected hyperparameters. The configuration’s hyperparameters are variables that are not part of the model and whose values cannot be inferred from the data. Two main categories of hyperparameters exist: (i) a network structure-determining hyperparameter; (ii) the network is trained by the hyperparameter. Table 2 contains a list of the hyperparameters that were examined in this study.

3.2. ADSCFGWO for CNN Hyperparameters

To select the most important characteristics from the metamaterial dataset in order to achieve the best possible performance, the adaptive dynamic sine cosine fitness grey wolf optimizer, abbreviated as ADSCFGWO, was initially introduced in [41]. This algorithm draws from both the sine cosine and grey wolf algorithms in an adaptable framework that makes use of both algorithms’ strengths. To estimate the double T-shape monopole antenna properties, the ADSCFGWO algorithm additionally optimizes a bidirectional recurrent neural network (BRNN). In this work, the optimization of the CNN hyperparameters is based on the adaptive dynamic sine cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. The ADSCFGWO algorithm is shown in Algorithm 1.
The population in the potential solution of the ADSCFGWO algorithm, X i ( i = 1 , 2 , , n ) with size n, is split into two groups: the exploration group, n 1 , and exploitation group, n 2 . The exploration group’s job is to use the search space to discover new locations where the greatest possible solution might be located. The exploitation group’s job is to use an objective function to enhance the best solution’s quality. These two groups cooperate in the suggested optimization process to trade responsibilities and required data that can hasten the retrieval of the optimum solution. The effective avoidance of the local optima and the precise exploration of the search space are advantages of this collaboration. The ADSCFGWO optimization technique has two key characteristics: first, it maintains correct control over the equilibrium between the exploitation and exploration groups; and second, it uses a dynamic mechanism to avoid steady regions in the search space.
The fittest solutions are denoted by ( S α ) , ( S β ) , and ( S δ ) . The position update in the direction of the prey position is estimated during the search process as
X ( t + 1 ) = T 1 + T 2 + T 3 3
where T 1 , T 2 , a n d   T 3 are calculated as
T 1 = S α A 1 . D ,   T 2 = S β A 2 . D ,   T 3 = S δ A 3 . D
where D is calculated as | C 1 . ( F α S α + F β S β + F δ S δ ) X ( t ) | . The A and C vectors are defined as A = 2 a . r 1 a and C = 2 r 2 , where the vectors values r 1 and r 2 are randomly selected from the range [ 0 , 1 ] . The values of a are determined in the range [ 0 , 2 ] and is calculated as a = 2 t . 2 T M a x for T M a x iterations. The fitness functions are calculated as
F α = F α F α + F β + F δ ,   F β = F β F α + F β + F δ ,   F δ = F δ F α + F β + F δ
The ADSCFGWO method automatically balances the subgroups of the population’s exploitation and exploration. The algorithm uses a 70/30 system in which two groups—exploration and exploitation groups—represent 70% of the population. A large number of participants in the exploration group early in the optimization process helps with the discovery of novel and intriguing search regions. The overall fitness of agents increases when more exploitative agents can increase their fitness values, but the proportion of agents engaged in exploration falls quickly from 70% to 30%. If a better solution cannot be identified, using an elitism approach ensures convergence by keeping the process leader in consecutive populations. ADSCFGWO may at any point increase the size of the exploration group, provided that the leader’s fitness has not dramatically increased over the course of three consecutive iterations.
The suggested ADSCFGWO algorithm’s computational complexity can be stated as in Table 3 for population n and iterations t m a x . From this analysis, the complexity of computations is O ( t m a x × n ) and O ( t m a x × n × d ) with d dimension.
Algorithm 1 ADSCFGWO algorithm
1:
Initialize ADSCFGWO population X i ( i = 1 , 2 , , n ) with size n, iterations t M a x , fitness function F n , parameters (a, A 1 , A 2 , A 3 , C 1 , C 2 , r 1 , r 2 , r 3 , r 4 )
2:
Calculate fitness function F n for each X i
3:
Find best solutions as S α , S β , S δ
4:
Set t = 1
5:
while   t t M a x do
6:
  Update  r 1 by r 1 = a 1 t M a x i t e r
7:
  for ( i = 1 : i n 1 ) do
8:
      DynamicSearch( F n )
9:
      Update Fitness by Equation (3)
10:
    Update positions from GWO as X ( t + 1 ) = T 1 + T 2 + T 3 3
11:
    if ( r 4 < 0.5 ) then
12:
      Update positions from SCA as X ( t + 1 ) = X ( t ) + r 1 × sin ( r 2 ) × | r 3 S α X ( t ) |
13:
    end if
14:
  end for
15:
  for ( i = 1 : i n 2 ) do
16:
    DynamicSearch( F n )
17:
    Update Fitness by Equation (3)
18:
    Update positions from GWO as X ( t + 1 ) = T 1 + T 2 + T 3 3
19:
    if ( r 4 0.5 ) then
20:
      Update positions from SCA as X ( t + 1 ) = X ( t ) + r 1 × cos ( r 2 ) × | r 3 S α X ( t ) |
21:
    end if
22:
  end for
23:
  Update fitness function F n for each X i
24:
  Update parameters
25:
  Find best solutions as S α , S β , S δ
26:
end while
27:
Return best solution X *
28:
DynamicSearch( F n )
29:
if (Best F n is same for three iterations) then
30:
  Increase exploration group solutions ( n 1 )
31:
  Decrease exploitation group solutions ( n 2 )
33:
end if

3.3. 3D U-Net Architecture Segmentation Model

U-Net [42] is a network that is used for fast and accurate image segmentation. It comprises an expanded pathway and a contracting pathway. The contracting pathway adheres to the standard convolutional network design. Two 3 × 3 unpadded convolution layers are applied repeatedly, and after them, a ReLU activation function and a 2 × 2 max-pooling with stride 2 are applied for down-sampling. The number of features at every stage in the down-sampling process is doubled. The expanding pathway consists of an up-sampling process, a 2 × 2 convolution layer that reduces the size of the feature map, a combination with the proportionally clipped feature map from the contracting pathway, and two 3 × 3 convolution layers, each accompanied by a ReLU activation function. All the 64-component extracted features are mapped to the required number of categories in the last layer using a 1 × 1 convolution layer. The model includes 23 convolutional layers overall. The reason for using a U-net network is that it is fast compared to other networks. On a modern GPU, segmentation of a 512 × 512 picture consumes less than a second.
Numerous U-Net-based variant networks have been proposed since U-Net [42]’s extensive research and application in medical image segmentation in 2015; 3D U-Net [43] is the most representative of these. Figure 2 depicts the 3D U-Net’s structure. The encoder-decoder architecture of this model expands on the prior U-Net (2D). The encoder component performs feature extraction from an analysis of the input image. The associated decoder produces a segmented mask. The mask extraction is supervised by this model by minimizing a cost function. 3D U-Net differs from 2D U-Net in that its features are extracted and restored using 3D convolution, 3D max-pooling, and 3D deconvolution blocks in turn after the volume data are input. In addition, batch normalization is added by 3D U-Net to prevent bottlenecks and hasten convergence. For the segmentation process, the dataset is partitioned into a train, validation, and test datasets.

4. Experimental Results

This section describes the used dataset, the Performance metrics used in CNN, the implementation of the proposed strategy, and the experiments conducted. The parameters for the ADSCFGWO algorithm’s configuration are shown in Table 4.

4.1. Dataset Description

The used dataset is BRaTS 2021 Task 1 Dataset [44]. As training, validation, and testing data for this year’s BraTS challenge, a sizable number of multi-institutional regular clinically acquired multi-parametric MRI (mpMRI) images of glioma with pathologically confirmed diagnosis and accessible MGMT promoter methylation status are used. For Task 1, the datasets utilized in this year’s competition have been updated with many additional routines clinically collected mpMRI scans since BraTS’20. To quantitatively assess the projected tumor segmentations, expert neuroradiologists create and approve ground truth annotations of tumor sub-regions for each patient in the training, validation, and testing datasets. As shown in Figure 3, the dataset is partitioned into a train, validation, and test datasets. Figure 4 illustrates an example of the dataset.
The data augmentation technique is used in this study to artificially generate fresh training data from the current data. As a sort of data augmentation, picture augmentation produces altered representations of the training dataset’s images. The input dataset is subjected to several image transformations, such as horizontal and vertical shift, horizontal and vertical flip, random rotation, and random zoom. The shift augmentation maintains the same image dimensions while shifting all of the MRI image’s pixels in either a horizontal or vertical direction. When flipping horizontally or vertically, all pixels’ rows and columns are reversed. The MRI image is randomly rotated between 0 and 360 degrees clockwise using the rotation augmentation. The zoom augmentation’s final step involves randomly zooming the MRI image by a factor between [0.9, 1.1].

4.2. Performance Metrics Used in CNN

The conventional computer-aided diagnostic approach may be tested using a variety of key performance metrics, including accuracy, precision, F1-score, recall, specificity, and sensitivity. The number of cases that were accurately identified as defective is shown by the letter T P , which stands for True Positive. False Positive, abbreviated as F P , refers to the number of cases that were incorrectly identified as defective. Additionally, F N stands for False Negative and reflects the number of occurrences that were incorrectly classified as non-defective. T N is for True Negative, which represents the number of cases that were correctly identified as non-defective. The metrics are defined as in Table 5.

4.3. The BCM-CNN Evaluation

As shown in Table 6, the effectiveness of the suggested approach (BCM-CNN) is evaluated in comparison to the previously widely used classifiers CNN [22], Decision Tree (DT) [45], Linear Discriminant (LD) [46], Support Vector Machine (SVM) [47], and K-Nearest Neighbor (K-NN) [48]. The default parameters are used for these methods.
Samples from the texted dataset are employed in the classification experiment. As a result of the BCM-CNN based on the ADSCFGWO algorithm boosting the performance of the CNN after altering its hyperparameters, the BCM-CNN delivered the best results when employed as a classifier, with an accuracy of (0.99980004), Sensitivity (TRP) of (0.99980004), Specificity (TNP) of ( 0.99980004), Pvalue (PPV) of (0.99980004), Nvalue (NPV) of (0.99980004), and F1-score of (0.9998). After the SVM-Linear model, which has an accuracy score of (0.968992248), the K-NN model, which has an accuracy score of (0.965250965), and finally, the LD model, which has an accuracy score of (0.961538462), the simple CNN model gets the second-best accuracy with a score of (0.9765625). The DT model was only able to achieve the lowest level of accuracy with (0.956022945). This came about as a consequence of the fact that the method that was proposed resulted in an improvement in CNN’s overall performance.
Table 7 shows the proposed BCM-CNN-based classifier’s statistical description and a comparison of classifiers based on 11 runs (run the algorithm 11 times) and 80 iterations ( t M a x in Algorithm 1) for 10 agents (Population size n in Algorithm 1) of the ADSCFGWO algorithm. This is to confirm the stability of the proposed method compared to other methods. Table 8 presents the compared and the proposed classifier’s test results using a one-way ANOVA (analysis of variance) test. In contrast, Table 9 discusses the comparison and the proposed classifiers test results using the Wilcoxon Signed-Rank test. With a p-value of less than 0.05, this statistical test demonstrates the significant difference between the suggested BCM-CNN classifier’s results and those of other classifiers.
The accuracy of the proposed BCM-CNN and comparative methods is shown by the box plot in Figure 5. This graph demonstrates the maximum accuracy results that the BCM-CNN-based optimization algorithm was able to produce. Based on the number of values with the Bin Center range (0.946–1.0), the accuracy histogram for the algorithms that have been presented and compared is shown in Figure 6, which attests to the stability of the suggested algorithm.
Figure 7 displays the residual, QQ (quantile-quantile), homoscedasticity plots, and heat map for the proposed and compared techniques. The possible problems can be observed in the residual values and plots as opposed to the plot of the original dataset. The independent variable is plotted on the horizontal axis, while the residual values are plotted on the vertical axis. The ideal situation is achieved if the residual values are scattered randomly and uniformly along the horizontal axis. The residual value is calculated as follows when the mean and the sum of the residuals are both equal to zero: (Actual-Predicted values). Figure 7 displays the residual plot. To determine if a model is linear or nonlinear and which one is best, plot patterns in a residual plot can be used. The projected scores for the dependent variable are examined visually together with the homogeneity of variance or heteroscedasticity. When the error term, also known as noise or random disturbance in the relationship between the dependent and independent variables, is constant across all values of the independent variables, this situation is referred to as homoscedasticity. The heteroscedasticity plot, shown in Figure 7, improves the precision of the research results. Any infraction can be quickly and easily detected.
The QQ plot is also shown in Figure 7. A probability plot is one illustration. By plotting the quantiles against one another, two probability distributions are primarily compared. It is possible to see that the point distributions in the QQ plot fit on the line in the illustration. Since the relationship between the actual and projected residuals is linear, the suggested technique is effective. Figure 7 serves as a tool for data visualization and displays heat maps for the offered and contrasted algorithms. The intensity of a two-dimensional color scale indicates the complexity of an algorithm. The color fluctuation provides obvious visual cues as to how the proposed solution is superior to the comparable algorithms. The ADSCFGWO algorithm’s performance in feature selection, as seen in Figure 7, is supported by these figures.

4.4. 3D U-Net Segmentation Model

There are four classes in the segmentation process. Segmentation classes are NOT tumor, non-enhancing tumor (RED color), EDEMA (Green color), and ENHANCING (yellow color). These classes were converted into three classes later. Figure 8 illustrates samples of images and masks with a positive brain tumor. For more precious and fast detection of brain tumor, the 3D U-net segmentation model has been implemented on the BRaTS 2021 dataset. The dataset is divided into 70% training, 20% validation, and 10% testing. Implementation is constructed online on Kaggle. U-net model enhances segmentation validation accuracy up to 99.33%, and validation loss up to 0.01 as shown in Figure 9. We can conclude that our proposed model can detect brain tumor with high accuracy compared to state-of-the-art techniques in terms of classification and segmentation.

5. Conclusions and Future Work

A large number of researchers looked at a wide variety of algorithms with the goal of accurately detecting and classifying brain cancers in a quick and efficient manner. Deep learning (DL) makes it possible to use a Convolutional Neural Network (CNN) model that has already been pre-trained for the analysis of medical pictures, in particular for the categorization of brain tumors. The fundamental objective of this research is to develop an improved model with the intention of making brain tumor diagnosis more accurate. A Convolutional Neural Network (CNN) that is based on a Brain Tumor Classification Model (BCM-CNN) was proposed in this paper. The optimization of the CNN’s hyperparameters was based on an adaptive dynamic sine-cosine fitness grey wolf optimizer (ADSCFGWO) algorithm. The BCM-CNN was used as a classifier in the experiments, and the results reveal that it produced the best results due to the enhancement of the performance of the CNN after the optimization was performed. The BCM-CNN was given the BRaTS 2021 Task 1 dataset, and it performed with an accuracy of 99.99%. The main limitation of the proposed algorithm is that it takes a long time to process due to the extra optimization steps. It may not be applicable as the size of the trained data is limited, so we intend to solve this issue in future work by generalizing more data. We also intend to make a prediction in future work, not just classification.

Author Contributions

Conceptualization, H.Z.; Methodology, S.A.G. and D.S.K.; Software, E.-S.M.E.-K.; Validation, A.H.A. and A.I.; Formal analysis, H.Z.; Investigation, S.A.G., A.I. and F.M.T.; Resources, A.H.A. and D.S.K.; Data curation, F.M.T.; Writing—original draft, H.Z., S.A.G. and D.S.K.; Writing—review & editing, E.-S.M.E.-K. and A.I.; Visualization, A.H.A., D.S.K. and F.M.T.; Supervision, E.-S.M.E.-K.; Project administration, E.-S.M.E.-K.; Funding acquisition, H.Z., S.A.G. and A.I. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R120), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
DLDeep Learning
CNNConvolutional Neural Network
BCM-CNNBrain Tumor Classification Model based on CNN
ADSCFGWOAdaptive Dynamic Sine-Cosine Fitness Grey Wolf Optimizer
WHOWorld Health Organization
MEGMagneto Encephalon Graph
CTComputed Tomography
EEGUltrasonography, Electro Encephalon Graph
SPECTSingle-Photon Emission Computed Tomography
PETPositron Emission Tomography
MRIMagnetic Resonance Imaging
CADComputer-Aided Diagnosis
NNNeural Network
FCFully Connected
DTDecision Tree
NBNaive Bayes
LDLinear Discrimination
SVMSupport Vector Machine
AUCArea Under the Curve
LDLinear Discriminant
QQQuantile-Quantile

References

  1. Al-Galal, S.A.Y.; Alshaikhli, I.F.T.; Abdulrazzaq, M.M. MRI brain tumor medical images analysis using deep learning techniques: A systematic review. Health Technol. 2021, 11, 267–282. [Google Scholar] [CrossRef]
  2. Rahman, M.L.; Reza, A.W.; Shabuj, S.I. An internet of things-based automatic brain tumor detection system. Indones. J. Electr. Eng. Comput. Sci. 2022, 25, 214–222. [Google Scholar] [CrossRef]
  3. Key Statistics for Brain and Spinal Cord Tumors. Available online: https://www.cancer.org/cancer/brain-spinal-cord-tumors-adults/about/key-statistics.html (accessed on 20 September 2022).
  4. Ayadi, W.; Elhamzi, W.; Charfi, I.; Atri, M. Deep CNN for Brain Tumor Classification. Neural Process. Lett. 2021, 53, 671–700. [Google Scholar] [CrossRef]
  5. Liu, J.; Li, M.; Wang, J.; Wu, F.; Liu, T.; Pan, Y. A survey of MRI-based brain tumor segmentation methods. Tsinghua Sci. Technol. 2014, 19, 578–595. [Google Scholar] [CrossRef]
  6. Amin, J.; Sharif, M.; Haldorai, A.; Yasmin, M.; Nayak, R.S. Brain tumor detection and classification using machine learning: A comprehensive survey. Complex Intell. Syst. 2021, 8, 3161–3183. [Google Scholar] [CrossRef]
  7. Jayade, S.; Ingole, D.T.; Ingole, M.D. Review of Brain Tumor Detection Concept using MRI Images. In Proceedings of the 2019 International Conference on Innovative Trends and Advances in Engineering and Technology (ICITAET), Shegoaon, India, 27–28 December 2019. [Google Scholar] [CrossRef]
  8. Yang, Y.; Yan, L.F.; Zhang, X.; Han, Y.; Nan, H.Y.; Hu, Y.C.; Hu, B.; Yan, S.L.; Zhang, J.; Cheng, D.L.; et al. Glioma Grading on Conventional MR Images: A Deep Learning Study With Transfer Learning. Front. Neurosci. 2018, 12, 804. [Google Scholar] [CrossRef] [Green Version]
  9. Nazir, M.; Shakil, S.; Khurshid, K. Role of deep learning in brain tumor detection and classification (2015 to 2020): A review. Comput. Med. Imaging Graph. 2021, 91, 101940. [Google Scholar] [CrossRef]
  10. El-Kenawy, E.S.M.; Mirjalili, S.; Abdelhamid, A.A.; Ibrahim, A.; Khodadadi, N.; Eid, M.M. Meta-Heuristic Optimization and Keystroke Dynamics for Authentication of Smartphone Users. Mathematics 2022, 10, 2912. [Google Scholar] [CrossRef]
  11. El-kenawy, E.S.M.; Albalawi, F.; Ward, S.A.; Ghoneim, S.S.M.; Eid, M.M.; Abdelhamid, A.A.; Bailek, N.; Ibrahim, A. Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm. Mathematics 2022, 10, 3144. [Google Scholar] [CrossRef]
  12. El-Kenawy, E.S.M.; Mirjalili, S.; Alassery, F.; Zhang, Y.D.; Eid, M.M.; El-Mashad, S.Y.; Aloyaydi, B.A.; Ibrahim, A.; Abdelhamid, A.A. Novel Meta-Heuristic Algorithm for Feature Selection, Unconstrained Functions and Engineering Problems. IEEE Access 2022, 10, 40536–40555. [Google Scholar] [CrossRef]
  13. Ibrahim, A.; Mirjalili, S.; El-Said, M.; Ghoneim, S.S.M.; Al-Harthi, M.M.; Ibrahim, T.F.; El-Kenawy, E.S.M. Wind Speed Ensemble Forecasting Based on Deep Learning Using Adaptive Dynamic Optimization Algorithm. IEEE Access 2021, 9, 125787–125804. [Google Scholar] [CrossRef]
  14. El-kenawy, E.S.M.; Abutarboush, H.F.; Mohamed, A.W.; Ibrahim, A. Advance Artificial Intelligence Technique for Designing Double T-shaped Monopole Antenna. Comput. Mater. Contin. 2021, 69, 2983–2995. [Google Scholar] [CrossRef]
  15. Samee, N.A.; El-Kenawy, E.S.M.; Atteia, G.; Jamjoom, M.M.; Ibrahim, A.; Abdelhamid, A.A.; El-Attar, N.E.; Gaber, T.; Slowik, A.; Shams, M.Y. Metaheuristic Optimization Through Deep Learning Classification of COVID-19 in Chest X-Ray Images. Comput. Mater. Contin. 2022, 73, 4193–4210. [Google Scholar] [CrossRef]
  16. Lee, G.; Nho, K.; Kang, B.; Sohn, K.A.; Kim, D. Predicting Alzheimer’s disease progression using multi-modal deep learning approach. Sci. Rep. 2019, 9, 1952. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Sharma, K.; Kaur, A.; Gujral, S. Brain Tumor Detection based on Machine Learning Algorithms. Int. J. Comput. Appl. 2014, 103, 7–11. [Google Scholar] [CrossRef]
  18. Agrawal, M.; Jain, V. Prediction of Breast Cancer based on Various Medical Symptoms Using Machine Learning Algorithms. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 28–30 April 2022; pp. 1242–1245. [Google Scholar] [CrossRef]
  19. Rabbi, M.F.; Mahedy Hasan, S.M.; Champa, A.I.; AsifZaman, M.; Hasan, M.K. Prediction of Liver Disorders using Machine Learning Algorithms: A Comparative Study. In Proceedings of the 2020 2nd International Conference on Advanced Information and Communication Technology (ICAICT), Dhaka, Bangladesh, 28–29 November 2020; pp. 111–116. [Google Scholar] [CrossRef]
  20. Swain, D.; Pani, S.K.; Swain, D. A Metaphoric Investigation on Prediction of Heart Disease using Machine Learning. In Proceedings of the 2018 International Conference on Advanced Computation and Telecommunication (ICACAT), Bhopal, India, 28–29 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  21. Zhu, W.; Sun, L.; Huang, J.; Han, L.; Zhang, D. Dual Attention Multi-Instance Deep Learning for Alzheimer’s Disease Diagnosis With Structural MRI. IEEE Trans. Med Imaging 2021, 40, 2354–2366. [Google Scholar] [CrossRef]
  22. Abdelhamid, A.A.; El-Kenawy, E.S.M.; Alotaibi, B.; Amer, G.M.; Abdelkader, M.Y.; Ibrahim, A.; Eid, M.M. Robust Speech Emotion Recognition Using CNN+LSTM Based on Stochastic Fractal Search Optimization Algorithm. IEEE Access 2022, 10, 49265–49284. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar] [CrossRef] [Green Version]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef] [Green Version]
  25. Alhussan, A.A.; Khafaga, D.S.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Abdelhamid, A.A. Pothole and Plain Road Classification Using Adaptive Mutation Dipper Throated Optimization and Transfer Learning for Self Driving Cars. IEEE Access 2022, 10, 84188–84211. [Google Scholar] [CrossRef]
  26. Srikanth, B.; Suryanarayana, S.V. Multi-Class classification of brain tumor images using data augmentation with deep neural network. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  27. Tandel, G.S.; Balestrieri, A.; Jujaray, T.; Khanna, N.N.; Saba, L.; Suri, J.S. Multiclass magnetic resonance imaging brain tumor classification using artificial intelligence paradigm. Comput. Biol. Med. 2020, 122, 103804. [Google Scholar] [CrossRef]
  28. Deepak, S.; Ameer, P. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef] [PubMed]
  29. Kokkalla, S.; Kakarla, J.; Venkateswarlu, I.B.; Singh, M. Three-class brain tumor classification using deep dense inception residual network. Soft Comput. 2021, 25, 8721–8729. [Google Scholar] [CrossRef] [PubMed]
  30. Özlem, P.; Güngen, C. Classification of brain tumors from MR images using deep transfer learning. J. Supercomput. 2021, 77, 7236–7252. [Google Scholar] [CrossRef]
  31. Kesav, N.; Jibukumar, M. Efficient and low complex architecture for detection and classification of Brain Tumor using RCNN with Two Channel CNN. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6229–6242. [Google Scholar] [CrossRef]
  32. Tummala, S.; Kadry, S.; Bukhari, S.A.C.; Rauf, H.T. Classification of Brain Tumor from Magnetic Resonance Imaging Using Vision Transformers Ensembling. Curr. Oncol. 2022, 29, 7498–7511. [Google Scholar] [CrossRef]
  33. Pareek, M.; Jha, C.K.; Mukherjee, S. Brain Tumor Classification from MRI Images and Calculation of Tumor Area. In Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; pp. 73–83. [Google Scholar] [CrossRef]
  34. Ayadi, W.; Charfi, I.; Elhamzi, W.; Atri, M. Brain tumor classification based on hybrid approach. Vis. Comput. 2020, 38, 107–117. [Google Scholar] [CrossRef]
  35. Konar, D.; Bhattacharyya, S.; Panigrahi, B.K.; Behrman, E.C. Qutrit-Inspired Fully Self-Supervised Shallow Quantum Learning Network for Brain Tumor Segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6331–6345. [Google Scholar] [CrossRef]
  36. Khairandish, M.; Sharma, M.; Jain, V.; Chatterjee, J.; Jhanjhi, N. A Hybrid CNN-SVM Threshold Segmentation Approach for Tumor Detection and Classification of MRI Brain Images. IRBM 2022, 43, 290–299. [Google Scholar] [CrossRef]
  37. Öksüz, C.; Urhan, O.; Güllü, M.K. Brain tumor classification using the fused features extracted from expanded tumor region. Biomed. Signal Process. Control 2022, 72, 103356. [Google Scholar] [CrossRef]
  38. Khan, M.A.; Ashraf, I.; Alhaisoni, M.; Damaševičius, R.; Scherer, R.; Rehman, A.; Bukhari, S.A.C. Multimodal Brain Tumor Classification Using Deep Learning and Robust Feature Selection: A Machine Learning Application for Radiologists. Diagnostics 2020, 10, 565. [Google Scholar] [CrossRef]
  39. Kadry, S.; Nam, Y.; Rauf, H.T.; Rajinikanth, V.; Lawal, I.A. Automated Detection of Brain Abnormality using Deep-Learning-Scheme: A Study. In Proceedings of the 2021 Seventh International conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25–27 March 2021. [Google Scholar] [CrossRef]
  40. Irmak, E. Multi-Classification of Brain Tumor MRI Images Using Deep Convolutional Neural Network with Fully Optimized Framework. Iran. J. Sci. Technol. Trans. Electr. Eng. 2021, 45, 1015–1036. [Google Scholar] [CrossRef]
  41. Khafaga, D.S.; Alhussan, A.A.; El-Kenawy, E.S.M.; Ibrahim, A.; Eid, M.M.; Abdelhamid, A.A. Solving Optimization Problems of Metamaterial and Double T-Shape Antennas Using Advanced Meta-Heuristics Algorithms. IEEE Access 2022, 10, 74449–74471. [Google Scholar] [CrossRef]
  42. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science; Springer International Publishing: Berlin, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  43. Özgün, Ç.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016; Springer International Publishing: Berlin, Germany, 2016; pp. 424–432. [Google Scholar] [CrossRef] [Green Version]
  44. BRaTS 2021 Task 1 Dataset, RSNA-ASNR-MICCAI Brain Tumor Segmentation (BraTS) Challenge 2021. Available online: https://www.kaggle.com/datasets/dschettler8845/brats-2021-task1?select=BraTS2021_Training_Data.tar (accessed on 20 September 2022).
  45. Nour, M.; Cömert, Z.; Polat, K. A Novel Medical Diagnosis model for COVID-19 infection detection based on Deep Features and Bayesian Optimization. Appl. Soft Comput. 2020, 97, 106580. [Google Scholar] [CrossRef] [PubMed]
  46. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef] [Green Version]
  47. Benmahamed, Y.; Teguar, M.; Boubakeur, A. Application of SVM and KNN to Duval Pentagon 1 for transformer oil diagnosis. IEEE Trans. Dielectr. Electr. Insul. 2017, 24, 3443–3451. [Google Scholar] [CrossRef]
  48. Benmahamed, Y.; Kemari, Y.; Teguar, M.; Boubakeur, A. Diagnosis of Power Transformer Oil Using KNN and Nave Bayes Classifiers. In Proceedings of the 2018 IEEE 2nd International Conference on Dielectrics (ICD), Budapest, Hungary, 1–5 July 2018. [Google Scholar] [CrossRef]
Figure 1. The proposed BCM-CNN model steps.
Figure 1. The proposed BCM-CNN model steps.
Bioengineering 10 00018 g001
Figure 2. Structure of the 3D U-Net.
Figure 2. Structure of the 3D U-Net.
Bioengineering 10 00018 g002
Figure 3. Dataset distribution.
Figure 3. Dataset distribution.
Bioengineering 10 00018 g003
Figure 4. Example of dataset images.
Figure 4. Example of dataset images.
Bioengineering 10 00018 g004
Figure 5. Box plot of accuracy for the BCM-CNN model under consideration and the contrasting models.
Figure 5. Box plot of accuracy for the BCM-CNN model under consideration and the contrasting models.
Bioengineering 10 00018 g005
Figure 6. Histogram of accuracy for the BCM-CNN model under consideration and the contrasting models.
Figure 6. Histogram of accuracy for the BCM-CNN model under consideration and the contrasting models.
Bioengineering 10 00018 g006
Figure 7. The heat map, residual, QQ, and homoscedasticity plots of the ADSCFGWO and comparable algorithms.
Figure 7. The heat map, residual, QQ, and homoscedasticity plots of the ADSCFGWO and comparable algorithms.
Bioengineering 10 00018 g007
Figure 8. Visualize image with mask of a positive brain tumor.
Figure 8. Visualize image with mask of a positive brain tumor.
Bioengineering 10 00018 g008
Figure 9. Training and validation performance parameters; (a) Accuracy, (b) Loss, (c) Dice coef.
Figure 9. Training and validation performance parameters; (a) Accuracy, (b) Loss, (c) Dice coef.
Bioengineering 10 00018 g009
Table 2. CNN hyperparameters setup.
Table 2. CNN hyperparameters setup.
ParameterValue
CNN training options (Default)
RateDropFactor
Momentum Learn
L2Regularization
LearnRateDropPeriod
GradientThreshold
GradientThresholdMethod
ValidationData
VerboseFrequency
ValidationPatience
ValidationFrequency
ResetInputNormalization
CNN training options (Custom)
InitialLearnRate
ExecutionEnvironment
MiniBatchSize
MaxEpochs
Verbose
Shuffle
LearnRateSchedule
Optimizer

0.1000
0.9000
1.0000 × 10 4
10
Inf
l2norm
imds
50
Inf
50
1

1.0000 × 10 4
gpu
8
20
0
every-epoch
piecwise
ADSCFGWO
Table 3. ADSCFGWO algorithm’s computational complexity.
Table 3. ADSCFGWO algorithm’s computational complexity.
No.OperationComplexity
1Initialization O ( 1 )
2Calculate objective function O ( n )
3Finding best solutions O ( n )
4Updating position of current grey wolf by fitness O ( t m a x × n )
5Updating position of current individual by Sine Cosine O ( t m a x × n )
6Updating objective function O ( t m a x × n )
7Finding best fitness O ( t m a x )
8Updating parameters O ( t m a x )
9Producing the best fitness O ( 1 )
Table 4. Parameters for the ADSCFGWO algorithm’s configuration.
Table 4. Parameters for the ADSCFGWO algorithm’s configuration.
ParameterValue
a[0, 2]
r 1 , r 2 , r 3 , r 4 [0, 1]
# Runs (Repeat the whole algorithm)11
# Iterations ( t M a x )80
# Agents (Population size n)10
Table 5. Performance metrics used in CNN.
Table 5. Performance metrics used in CNN.
No.MetricsCalculation
1Accuracy T P + T N T P + T N + F P + F N
2Sensitivity T P T P + F N
3Specificity T N T N + F P
4Precision (PPV) T P T P + F P
5Negative Predictive Value (NPV) T N T N + F N
6F1-score 2 × P P V × T P R P P V + T P R
Table 6. The performance of the proposed method (BCM-CNN) versus basic classifiers.
Table 6. The performance of the proposed method (BCM-CNN) versus basic classifiers.
Accuracy Sensitivity (TRP)Specificity (TNP)p Value (PPV)N Value (NPV)F1-Score
BCM-CNN0.999800040.999800040.999800040.999800040.999800040.9998
CNN0.97656250.9756097560.9771986970.9661835750.9836065570.970874
SVM-Linear0.9689922480.9569377990.9771986970.9661835750.9708737860.961538
K-NN0.9652509650.9569377990.9708737860.9569377990.9708737860.956938
LD0.9615384620.9478672990.9708737860.9569377990.9646302250.952381
DT0.9560229450.9478672990.9615384620.9433962260.9646302250.945626
Table 7. Proposed BCM-CNN classifier’s statistical description and a comparison of classifiers.
Table 7. Proposed BCM-CNN classifier’s statistical description and a comparison of classifiers.
BCM-CNNCNNSVM-LinearK-NNLDDT
Number of values111111111111
Minimum0.99980.97660.9690.96530.95150.946
25% Percentile0.99980.97660.9690.96530.96150.956
Median0.99980.97660.9690.96530.96150.956
75% Percentile0.99980.97660.9690.96530.96150.956
Maximum0.99980.98660.9720.96530.97150.966
Range00.010.00300.020.02
10% Percentile0.99980.97660.9690.96530.95350.948
90% Percentile0.99980.98520.9720.96530.97030.9649
95% CI of median
Actual confidence level98.83%98.83%98.83%98.83%98.83%98.83%
Lower confidence limit0.99980.97660.9690.96530.96150.956
Upper confidence limit0.99980.97960.97190.96530.96540.9602
Mean0.99980.97770.96950.96530.96190.9564
Std. Deviation00.003060.00119500.004620.004649
Std. Error of Mean00.00092260.000360300.0013930.001402
Lower 95% CI of mean0.99980.97570.96870.96530.95880.9533
Upper 95% CI of mean0.99980.97980.97030.96530.9650.9595
Coefficient of variation0.000%0.3130%0.1232%0.000%0.4803%0.4860%
Geometric mean0.99980.97770.96950.96530.96190.9564
Geometric SD factor11.0031.00111.0051.005
Lower 95% CI of geo. mean0.99980.97570.96870.96530.95880.9533
Upper 95% CI of geo. mean0.99980.97980.97030.96530.9650.9595
Harmonic mean0.99980.97770.96950.96530.96190.9564
Lower 95% CI of harm. mean0.99980.97570.96870.96530.95880.9533
Upper 95% CI of harm. mean0.99980.97980.97030.96530.9650.9595
Quadratic mean0.99980.97770.96950.96530.96190.9564
Lower 95% CI of quad. mean0.99980.97570.96870.96530.95880.9533
Upper 95% CI of quad. mean0.99980.97980.97030.96530.9650.9595
Skewness 2.8871.924 −0.2076−0.2118
Kurtosis 8.5362.047 4.0013.839
Sum1110.7610.6610.6210.5810.52
Table 8. Compared and proposed classifiers test results for ANOVA.
Table 8. Compared and proposed classifiers test results for ANOVA.
SSDFMSF (DFn, DFd)p Value
Treatment (between columns)0.0132350.002646F (5, 60) = 295.4p < 0.0001
Residual (within columns)0.0005374600.000008957--
Total0.0137765---
Table 9. Compared and proposed classifiers test results for Wilcoxon Signed-Rank.
Table 9. Compared and proposed classifiers test results for Wilcoxon Signed-Rank.
BCM-CNNCNNSVM-LinearK-NNLDDT
Theoretical median000000
Actual median0.99980.97660.9690.96530.96150.956
Number of values111111111111
Wilcoxon Signed Rank Test
Sum of signed ranks (W)666666666666
Sum of positive ranks666666666666
Sum of negative ranks000000
P value (two tailed)0.0010.0010.0010.0010.0010.001
Exact or estimate?ExactExactExactExactExactExact
Significant (alpha = 0.05)?YesYesYesYesYesYes
How big is the discrepancy?
Discrepancy0.99980.97660.9690.96530.96150.956
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

ZainEldin, H.; Gamel, S.A.; El-Kenawy, E.-S.M.; Alharbi, A.H.; Khafaga, D.S.; Ibrahim, A.; Talaat, F.M. Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization. Bioengineering 2023, 10, 18. https://doi.org/10.3390/bioengineering10010018

AMA Style

ZainEldin H, Gamel SA, El-Kenawy E-SM, Alharbi AH, Khafaga DS, Ibrahim A, Talaat FM. Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization. Bioengineering. 2023; 10(1):18. https://doi.org/10.3390/bioengineering10010018

Chicago/Turabian Style

ZainEldin, Hanaa, Samah A. Gamel, El-Sayed M. El-Kenawy, Amal H. Alharbi, Doaa Sami Khafaga, Abdelhameed Ibrahim, and Fatma M. Talaat. 2023. "Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization" Bioengineering 10, no. 1: 18. https://doi.org/10.3390/bioengineering10010018

APA Style

ZainEldin, H., Gamel, S. A., El-Kenawy, E. -S. M., Alharbi, A. H., Khafaga, D. S., Ibrahim, A., & Talaat, F. M. (2023). Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization. Bioengineering, 10(1), 18. https://doi.org/10.3390/bioengineering10010018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop