Next Article in Journal
Methods to Improve Dynamic System Response of Power Compensators Using Supercapacitors in Low-Voltage Ride-Through (LVRT) Conditions
Next Article in Special Issue
Multi-Objective Quantum-Inspired Seagull Optimization Algorithm
Previous Article in Journal
ASMAC: An Adaptive Slot Access MAC Protocol in Distributed VANET
Previous Article in Special Issue
Forensic Analysis on Internet of Things (IoT) Device Using Machine-to-Machine (M2M) Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Deep Learning-Based Approach for Brain Tumor Classification

1
Department of Software Engineering, University of Engineering and Technology, Taxila 44000, Pakistan
2
Department of Software Engineering, University of Science and Technology Bannu, Bannu 28100, Pakistan
3
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences (UCAS), Shenzhen 518055, China
4
Electrical Engineering Department, Faculty of Engineering & Technology, Future University in Egypt, New Cairo 11845, Egypt
5
Department of Computer Engineering and Networks, College of Engineering, Prince Sattam Bin Abdulaziz University, Wadi Addawasir 11991, Saudi Arabia
6
Software College, Northeastern University, Shenyang 110169, China
7
Department of Electrical Engineering, Government College University, Lahore 54000, Pakistan
8
Faculty of Engineering, Uni de Moncton, Moncton, NB E1A3E9, Canada
9
International Institute of Technology and Management, Libreville BP1989, Gabon
10
Spectrum of Knowledge Production & Skills Development, Sfax 3027, Tunisia
11
Department of Electrical and Electronic Engineering Science, School of Electrical Engineering, University of Johannesburg, Johannesburg 2006, South Africa
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(7), 1146; https://doi.org/10.3390/electronics11071146
Submission received: 13 March 2022 / Revised: 29 March 2022 / Accepted: 29 March 2022 / Published: 5 April 2022
(This article belongs to the Special Issue Intelligent Data Sensing, Processing, Mining, and Communication)

Abstract

:
Brain tumors (BTs) are spreading very rapidly across the world. Every year, thousands of people die due to deadly brain tumors. Therefore, accurate detection and classification are essential in the treatment of brain tumors. Numerous research techniques have been introduced for BT detection as well as classification based on traditional machine learning (ML) and deep learning (DL). The traditional ML classifiers require hand-crafted features, which is very time-consuming. On the contrary, DL is very robust in feature extraction and has recently been widely used for classification and detection purposes. Therefore, in this work, we propose a hybrid deep learning model called DeepTumorNet for three types of brain tumors (BTs)—glioma, meningioma, and pituitary tumor classification—by adopting a basic convolutional neural network (CNN) architecture. The GoogLeNet architecture of the CNN model was used as a base. While developing the hybrid DeepTumorNet approach, the last 5 layers of GoogLeNet were removed, and 15 new layers were added instead of these 5 layers. Furthermore, we also utilized a leaky ReLU activation function in the feature map to increase the expressiveness of the model. The proposed model was tested on a publicly available research dataset for evaluation purposes, and it obtained 99.67% accuracy, 99.6% precision, 100% recall, and a 99.66% F1-score. The proposed methodology obtained the highest accuracy compared with the state-of-the-art classification results obtained with Alex net, Resnet50, darknet53, Shufflenet, GoogLeNet, SqueezeNet, ResNet101, Exception Net, and MobileNetv2. The proposed model showed its superiority over the existing models for BT classification from the MRI images.

1. Introduction

The human brain is a command center and an essential organ of the human nervous system responsible for accomplishing daily life activities. The brain collects stimuli or signals from the body’s sensory organs, handles processing, and directs the ultimate decisions and output information to the muscles. BTs is one of the most severe situations related to the human brain, where a group of abnormal brain cells grows in an undisciplined manner [1]. BTs can be divided into two main types: primary and secondary metastatic. The primary brain tumors (BTs) are generally non-cancerous and originate from human brain cells. In contrast, secondary metastatic tumors spread to the brain with blood flow from other body parts.
Furthermore, the World Health Organization (WHO) classified BTs into four categories (Grade I–IV) depending on their malignancy or benignity. The standard approaches for detecting and analyzing BTs are magnetic resonance imaging (MRI) and computer tomography (CT) [2]. Grade III and Grade IV malignant BTs are fast-growing; they spread to other body parts and affect healthy cells. Thus, early BT detection and classification helps doctors plan proper treatment based on MRI and other images [3]. Glioma, pituitary and meningioma are the three main types of primary brain tumors. Pituitary BTs are generally benign and grow in the pituitary glands, the base layer of the brain that produces some essential hormones in the body [4]. Gliomas develop from the glial cells of the brain [5]. Meningioma tumors generally grow on the protective membrane of the brain and spinal cord [6]. The separation of normal brain tissue from abnormal tissues is critical in BT detection. Due to size, shape, and location variations, BT detection becomes more energizing and is still an open problem. The concepts of medical image processing are used in BT analysis (i.e., classification, segmentation, and detection) [7]. BT classification is a necessary procedure to identify the tumor type at an early stage, if there are any. Many modernistic, computer-aided diagnosis systems are presented in biomedical image processing to help radiologists guide patience and better classify a BT [8]. A BT is a hazardous disease, and it causes shorter life when there are high-grade tumors. To be precise, the diagnosis of a BT plays a vital role in treatment and is helpful for the patient’s life [9]. Due to high variance, low contrast in nasopharyngeal carcinoma (NPC), and disrupted edges in magnetic resonance images (MRIs). Accurate tumor segmentation is critical in the guidance of a radiologist [10] to identify tumors better. There are numerous deep learning architectures proposed in the literature for BT segmentation, such as DensNet [11], ResNet [12], and InceptionNet [13].
In the literature, ML and DL are the two main techniques implemented for BT detection [14,15]. Various studies have been proposed that employed machine learning methods, such as support vector machines (SVM) [16,17], k-nearest neighbor (KNN) [18], principal component analysis (PCA) [19], decision trees, and artificial neural networks (ANNs) [20,21]. However, these methods work on hand-crafted features, while the mean features need to be extracted for the training process. Therefore, the detection and classification accuracy depend on the quality of the features. Machine learning classifiers are time-consuming and require large memory for large datasets [22]. Additionally, CNN layers are widely used for image and speech feature extraction [23]. Artificial neural networks are also used for the extraction of different features, as each neuron is connected to another neuron [24]. However, in deep learning, the last layers are fully connected and perform well in medical imaging. For example, the CNN is the most common DL model mostly used for image classification [25].
This inspired us to propose a DL-based approach to enhance existing algorithms’ accuracy and performance in detecting various types of BTs and evaluate the approach on a publicly available dataset. For this purpose, we propose a hybrid DeepTumorNet model that identifies and classifies BTs under the following three main types: meningiomas, pituitary, and gliomas. The proposed method adopts the mechanism of deep learning for feature extraction and a Softmax classification layer for variety. The proposed model recorded the highest ever classification accuracy on the (CE-MRI) dataset, which is publicly available on figshare, compared with the traditional method (i.e., Google net [26], Alex net [27,28,29], Resnet50 [12], Squeezenet [28], DensNet [11], darknet53 [29], Mobilenetv2 [30], Resnet101 [31], and Shufflenet [32]). Furthermore, with the proposed research study, we aim to answer the following research question: How efficiently and correctly does the deep learning algorithm identify and classify BTs into different BT diseases?
Our main contribution to this research includes the following.
We propose a hybrid deep learning model for three types of BT classification (pituitary, meningiomas, and gliomas), where the proposed DL classification approach outperforms the existing state-of-the-art approaches by showing the highest accuracy on the CE-MRI dataset, and extensive experiments are performed with nine other pretrained deep learning models using transfer learning techniques. We also compare the results of the proposed model with the previously proposed methods.
This research article is organized as follows. Section 2 presents the related works. Section 3 details the proposed methodology. Section 4 describes the results and discussion. Section 5 gives the conclusion and future work as a result of this study.

2. Related Works

Recent human works on computer-aided medical diagnosis provide improved performance. For instance, Gupta et al. [33] integrated canny detection with an ANN for BTs identification. They adjusted the sizes of previously treated images by RGB2grey conversion and implemented a clever edge detection method to remove the qualities from BTs pictures. In the end, an ANN was implemented for BTs identification. Mehrotra et al. [34] proposed a DL network for two classes (malignant and benign) of dataset classification. They utilized five different pre-trained deep learning models (AlexNet, GoogLeNet, ResNet101, ResNet50, and SqueezeNet). AlexNet accomplished the most acceptable accuracy of 99.04%. Moreover, it was observed that the model depended on the chosen optimizer.
Ramzan et al. [35] proposed a DL model for BTs detection using a fusion of hand-crafted and deep features. The authors segmented BTs by applying a grab cut algorithm and some morphological operations. Moreover, deep learning (VVG19) and handmade components extracted for the tumor segments were fused through a serial-based method. These concatenated features were supplied to multiple classifiers. Raja et al. [36] enhanced the MR images using the nonsubsampled contourlet transform (NSCT) and NSCT fused separately. Low- and high-frequency sub-bands were connected independently.
Furthermore, the inverse NSCT was applied to obtain enhanced images and then extract the textured features from the enhanced images. The adaptive neuro-fuzzy inference system (ANFIS) method was used to classify these features into normal and glioma BT images. The classified glioma images were segmented using a morphological operation.
Sultan et al. [37] proposed a DL model based on a CNN using two publicly available datasets containing 3064 (meningioma, glioma, and pituitary tumors) and 516 (Grade II, Grade III, and Grade IV) images. The proposed model achieved a highest accuracy of 96.13% and 98.7%, respectively. Kumar et al. [38] introduced a hybrid approach for BT detection and classification. The proposed strategy extracted the features from MR images by applying a discrete wavelet transform and using PCA for feature reduction. Using a kernel SVM trained with the reduce feature, the proposed hybrid performed better than traditional DL methods by increasing the accuracy and decreeing the root mean square error. Pitchai et al. [39] integrated an artificial neural network (ANN) along with a fuzzy k-means algorithm for BT segregation and detection. The GLCM-extracted features are given to the ANN for a variety of standard and abnormal MRI images.
Furthermore, the fuzzy k-means algorithm segregates tumors from abnormal MRIs. Amin et al. [40] proposed an automatic detection and classification method for three publicly available BT datasets. Their proposed strategy comprises three steps. They segment the region of interest using different pre-processing techniques and morphological operations. Moreover, each candidate’s lesion intensity, texture, and shape features are extracted and classified with an SVM. The proposed technique is more robust than other existing methods and returned 97.1% accuracy. Özyurt et al. [41] proposed a BT detection method. First, they segmented the MRI tumor images using the NS-EMFSE method. They extracted features from the segmented images using AlexNet and then the ML model, and KNN and a SVM were utilized to detect and classify the BT images as benign or malignant using the SVM, achieving 95.62% accuracy. Kaplan et al. [42] modified the local binary patterns (LBP) feature for BT detection and classification. They extracted the nLBP and αLBP features and performed classification with different ML classifiers such as KNN, random forest (RF), ANN, A1DE, and linear discriminant analysis (LDA). Their proposed method achieved the highest accuracy of 95.56%, and nLBP d = 1 feature extraction with the KNN model, where d is the distance between two neighbors’ pixels.
Çinar et al. [43] proposed a hybrid CNN architecture for BT detection by modifying a deep learning model. They removed the last five layers of ResnNet50 and added eight new layers. Their proposed hybrid ResNet50 model obtained 97.2% accuracy, while the single ResNet50 model obtained 92.53% accuracy. Mohsen et al. [44] introduced a DNN for BT detection and classification, where the MR images are segmented into normal and abnormal by using the fuzzy c-means clustering technique. The features are extracted from segmented images using DWT.
Moreover, these features are reduced by using PCA. Finally, classification is performed by a deep neural network. Like a CNN, the proposed methodology achieved the highest accuracy of 98.4%. Lather et al. [45] investigated different BT segmentation and detection techniques. Their study surveyed six segmentation techniques and presented a detailed review of previous researchers’ work on segmentation and detection. Abd-Ellah et al. [46] suggested two-phase models for BT detection and localization. To detect brain tumors, the proposed system in the first phase classifies MR images as normal or abnormal by using a CNN for features extraction and the ECOC-SVM approach for classification. In the second phase, the tumor is localized in strange images by designing a new five-layer region based on R-CNN. The proposed two-phase multi-model achieved 99.5% detection accuracy. Marghalani et al. [47] proposed an automatic classification method for BT and Alzheimer’s disease. They extracted a bag of features (SURF and SIFT) and used a SVM as a classifier.
The proposed automatic classification method classified the MRI data set into three classes—MRIs of brain tumors, MRIs of Alzheimer’s disease, and MRIs of normal brains—with an average accuracy of 97%, which was greater than that of the SURF-based features. Swati et al. [48] used block-wise fine-tuned CNN models for BT detection and classification. The fine-tuned VGG-19 for the block-wise fine-tuning technique achieved 94.84% classification accuracy in less training time than the hand-crafted features. Kumar et al. [49] introduced a new optimized DL mechanism for BT detection and classification named Dolphin-SCA based on a deep CNN. For segmentation, the researcher used a fuzzy deformable fusion model with a dolphin echolocation-based sine cosine algorithm (Dolphin-SCA). The extracted features were used in a deep neural network with Dolphin-SCA based on the power LDP and statistical extracted features. The proposed technique achieved 96.3% classification accuracy.
Deepak et al. [50] used pre-trained GoogLeNet for feature extraction and proven classifier models for BT detection and classification. The proposed approach recorded 98% accuracy compared with the state-of-the-art methods. Raja et al. [51] presented a hybrid model for BT detection and classification that involves different backgrounds (i.e., pre-processing using a non-local express filter and segmentation using the Bayesian fuzzy method). Afterward, various image features were captured and extracted using theoretic measures, scattering transform, and the wavelet packet Tsallis entropy method. Finally, classification was carried out using a hybrid approach based on deep autoencoder with a Softmax regression and obtained 98.5% accuracy. Rammurthy et al. [52] proposed a new BT detection technique based on DL, namely Whale Harris Hawks optimization, by combining the whale optimization algorithm (WOA) with the Harris hawks optimization (HHO) algorithm. At first, the tumors in the images are segmented using cellular automata, and different features like the size, variance, mean, and kurtosis are extracted, while the elements are classified for better BTs detection with the proposed Whale Harris Hawks optimization (WHHO). The proposed method reached its highest accuracy at 81.6%. Bahadure et al. [53] used a skull skipping algorithm to eliminate the non-brain parts from MR images to detect BTs based on Berkeley wavelet transformation (BWT) and segmented features (shape, texture, color, and contrast) with an SVM classifier. The experimental results obtained 96.51% accuracy. Waghmare et al. [54] implemented different CNN architectures for BTs detection and classification. Fine-tuned VGG-16 increased the classification accuracy of the augmented data set and reached the most acceptable accuracy at 95.71%.

3. Materials and Methods

This section highlights and elaborates the proposed research methodology for fine-grained BTs classification. The proposed method is mainly described in two steps. First, we elaborated on the research dataset utilized for fine-grained BTs classification. Secondly, we thoroughly elaborated on the proposed deep learning-based approach and its architecture to detect and classify brain MRI images into meningiomas, gliomas, and pituitary tumors.

3.1. Dataset and Preprocessing

This research study utilized a publicly available CE-MRI dataset [55]. The MRI images in the dataset were collected over 5 years (2005–2010) from 233 different patients having BTs at Nanfang Hospital Guangzhou China and General Hospital Tianjin Medical University in China. The dataset comprises 3062 MRI images of 3 distinct types of BT in 233 patients, including gliomas (1426), meningiomas (708), and pituitary tumors (930) in 3 different views. The details of the research dataset are explained in Figure 1 and Table 1. The dataset images are in 2D volumes and have a 512 × 512 resolution with a 0.49 × 0.49 mm2 pixel size. Additionally, the tumor region in the MRI was bordered manually by three experienced radiologists [56]. The dataset images available on figshare are in .mat format and have a 512 × 512 resolution. The proposed model was designed with an input layer size of 224 ×224. Therefore, the dataset was pre-processed to make it pursuable for the proposed approach, as shown in Figure 2. Initially, the MRI images were normalized (i.e., mat to .jpg format) and then resized by using the resize function in Matlab [57] according to the image input sizes of our proposed deep learning model and other pretrained models. Therefore, the MRI images were resized to 224 × 224 pixels, and the DarkNet19 images were resized to 256 × 256 pixels. Furthermore, the dataset images were split into 70% for training and 30% for testing. We used all the 3064 brain tumor images for the experiments, where around 2146 images (495 meningiomas, 652 pituitary tumors, and 999 gliomas) were used for training. The remaining 918 images (213 meningiomas, 278 pituitary tumors, and 428 gliomas) for testing.

3.2. DeepTumorNet

The hybrid DeepTumorNet model follows the basic structure of the CNNs. As training a convolution network from scratch is a challenging job, sometimes, it may take months [30,48]. Therefore, instead of training a new deep learning classifier from scratch, a pretrained classifier would be better to train the proposed deep learning approach. For this purpose, we used GoogLeNet as a base model, as GoogLeNet [26] remained the winner of the ILSVRC (2014) ImageNet competition. GoogLeNet has 22 learnable layers and 144 layers, including 2 convolution layers, 4 max_pooling layers, 1 average pooling layer, 2 normalization layers, 1 fully connected layer, and 9 inception layers modules [43]. Additionally, each inception module was comprised of one max-pooling and six convolutional layers. The input layer of GoogLeNet was renewed to 224 × 224 × 1. The ReLU activation function was used in the pretrained GoogLeNet approach. At the same time, the ReLU activation function neglected all negative values and used zero instead. Alternatively, Leaky ReLU is an improved version of ReLU, and it replaces all negative values with positive ones [58].
Meanwhile, in the proposed DeepTumorNet classifier, the last 5 layers of GoogLeNet were removed, and 15 new layers were added instead of 5 layers. Furthermore, the ReLU activation function in the feature map layer was changed to the Leaky ReLU activation function to increase the proposed model’s expressiveness and overcome the dying ReLU problem without disturbing the primary convolution neural network architecture. The total count of layers after these changes increased from 144 to 154.
The first convolution layer used a filter (patch) size of 7 × 7, which immediately reduced the image size. The second convolution layer had a depth of two and leveraged the 1 × 1 convolution block, which is the effect of dimensionality reduction. Furthermore, the inception module of GoogLeNet has different convolution kernels, such as convolution kernels of 1 × 1, 3 × 3, and 5 × 5 which extract the features at different scales, starting from the most delicate features to the core features. The larger convolution kernel covers a larger area to compute the features. Similarly, the 1 × 1 convolution kernel gives further details and reduces the amount of computation. The new additions include four convolutional layers with a tiny filter size of 1 × 1. In addition, increasing the number of convolution layers in the CNN gave us more detailed, accurate, and robust features. These four later convolutional layers extracted high-level features compared with the initial layer, which extracted low-level features.
Moreover, the global average pooling layer increased the accuracy at the network’s end. Furthermore, the ReLU activation function in the feature map layer was changed to the Leaky ReLU activation function to increase the proposed model’s expressiveness and overcome the dying ReLU problem. Therefore, the proposed hybrid model extracted more detailed, discriminative, and deep features because of the additional layer, resulting in better classification performance than the mentioned state-of-the-art pretrained deep learning models.
The structure of the proposed hybrid DeepTumorNet approach is shown in Figure 2. The name, type, number of the filter, filter size, and epsilon of the added layer are elaborated in Table 2. Below, we discuss each layer of the proposed hybrid DeepTumorNet model in detail.

3.2.1. Image Input Layer

The proposed DeepTumorNet model starting from the image layer included the model’s input, which specified the image input size, which in our case was 224 × 224 × 1. Such a number corresponds to the input image’s width, height, and channel size (1 in the case of grayscale images and 3 in the case of color images). The images were first read from in the input layer for processing.

3.2.2. Convolutional Layer

The convolutional layer was utilized to recover the deep learning features from an input image (producing a feature map). The mathematical operation presents two arguments: an image matrix and a filter size (size represents the height and width of the filters) [59]. Our hybrid model uses different filter sizes of 7 × 7, 5 × 5, and 1 × 1 in the convolutional layers and 3 × 3 in the max-pooling layers. Convolutional layers add padding with the input of the feature map by using the “Padding” name–value pair. The discrete time convolution method is elaborated in Equation (1):
s t = x w t = a = x a w t a
where W is the kernel filter, x is the input to the method, t is the time taken, and s is the results. In the case of two-dimensional input data being taken, Equation (2) is considered:
S i , j = I K i , j = m n I i , j K i m , j n  
The terms i and j show the areas of the desired matrix required after the deep learning convolution method. The preferred technique in this procedure is set so that the filter’s center is in the first position.
If cross-entropy is to be accomplished in the proposed approach, Equation (3) is utilized:
S i , j = I K i , j = m n I i + m , j + n K m , n

3.2.3. Activation Function

The activation functions are often used in DL-based models for nonlinear transformation processes. Sigmoid, Tanh, and ReLU activation functions were the most widely used and preferred activation functions developed in the past. However, ReLU provides an output of zero for all negative inputs (i.e., deactivating negative inputs), resulting in dead neurons (the dying ReLU problem). A neuron is “dead” if it always outputs 0 and is stuck on the opposing side. We used the leaky ReLU activation function instead of ReLU in the feature map, an value-added form of the ReLU activation function, to address the dying ReLU problem [60]. In the case of leaky ReLU, the negative number (x) output is a tiny linear component of x instead of 0. Moreover, in the last 15 added layers, we used a clipped ReLU activation function which performed a thresholding operation, where any desired input value that was less than 0 was set to 0, and any desired input value that was above the ceiling was set to the specified ceiling. The formulas of the activation functions are shown in Equations (4)–(7):
ReLU:
f x = 0 , x < 0 x , x 0 , f x = 0 , x < 0   1 , x 0  
Sigmoid:
f x = 1 1 + e x , f x = f x 1 f x
Tanh:
tan h ( x ) = 2 1 + e 2 x 1 , f x = 1 f ( x ) 2
Clipped ReLU:
f x = 0 , x < 0 x 0 x < c e i l i n g c e i l i n g , x c e i l i n g
Leaky ReLU:
f x = x , x 0 s c a l e x , x < 0
In the leaky ReLU function, output x is for the positive inputs, and it outputs a small value that is 0.01 times x in the case of negative values. Hence, no neuron is deactivated in this case, and we would no longer encounter dead neurons.

3.2.4. Batch Normalization Layer

The batch normalization layer was utilized to normalize the outputs generated by the proposed convolution layers. Normalization shortens the training time of the proposed DeepTumorNet model to achieve the learning process more efficiently and quickly. The batch normalization process is given in Equations (8)–(10):
Y i = X i μ β σ 2 β + ε
σ β = I M ( X i μ β ) 2
μ β = 1 M i = 1 M X i
where M is the total number of input data, X_i = 1, …, M, μ_β is the stack’s average value, σ_β is the stack’s standard deviation, and Yi is the new values obtained as a result of the normalization procedure.

3.2.5. Pooling Layer

After the convolution layer, the pooling layer was used to simplify the information from the convolution layer (a downsampling procedure to decrease the size of the feature map and remove unnecessary data). Average and maximal pooling are the two most common pooling strategies. In the last 15 layers, we used global average pooling. In pooling, the network does not carry out any learning. For the pooling process, 3 × 3-sized filters were used. The pooling process is given in Equation (12):
S = w 2 × h 2 × d 2
w 2 = w 1 f   A + 1  
h 2 = h 1 f A + 1  
d 2 = d 1
where w1 represents the width of the images of the MRI, h1 denotes the height of the input MRI image, d1 denotes the value of the depth of the input MRI image size, f represents the filter size, A represents the number of steps utilized, and S represents the size of the manufactured image.

3.2.6. Fully Connected Layer

In the proposed model, the convolutional layers are followed by a fully connected layer. This is accomplished by merging all of the features learned by the preceding layers over a number of images. This layer determines the most significant patterns in order to categorize the images. The output size value in the final completely linked layer is 3, as in the proposed research study, the number of classes (meningioma, glioma, and pituitary) is 3. For this, the obtained output value of the proposed FC layer is 3. Equations (16) and (17) are used for this purpose:
U i l = j w j i l 1 y j l 1
y i l = f u i l + b l
where l is the total number of the layers, i and j are the total number of neurons, yli is the value created in the proposed output layer, wl-1ji is the hidden layer’s weight value, yl-1i is the value of the input neurons, uli is the output layer’s value, and b(l) is the value of deviation.

3.2.7. Softmax Layer

The activation function makes the output of the fully linked layer more normalized. Softmax executes the probabilistic calculation coming from the network and creates the work in positive numbers for each class. The Softmax method is given in Equation (18):
P y = j | x i , W , b = e x p X T W j j = 1 n e x p X T W j  
where A, s, W, and b are weight vectors.

3.2.8. Classification Layer

The last layer of the proposed model is the classification layer, which is used to produce the output by utilizing each input. The Softmax activation function returned a probability distribution [61].

3.2.9. Training Parameters

We used a trial-and-error-based approach to perform experiments with the parameters shown in Table 3. We constantly monitored the improvement of the training validation accuracy and error to find the optimal convergence of each CNN. If there was no validation accuracy or error increase, the training was terminated automatically. We used stochastic gradient descent (SGD) to train the proposed DeepTumorNet model with an initial and final learning rate of 0.01 and a minibatch size of 10 images. The proposed DeepTumorNet model was trained on 120 epochs for brain tumor classification to obtain the optimum results.

4. Results and Discussion

This research paper was intended to classify three different types of BTMRI images correctly. The CEMRI dataset was classified with pre-trained deep learning models and the proposed hybrid DeepTumorNet model, where 70% of the dataset was used for training purposes and 30% for testing purposes. The results were achieved in the Matlab environment with computer resources, namely an i5 processor and 8 GB of RAM. There are various approaches for the measurement of deep learning network classification performance. A confusion matrix is widely used for classification tasks in the CNN process. The most desired and preferred measures included precision, recall, accuracy, and F1 score. These calculations are computed through a confusion matrix [62]. The general form of the confusion matrix utilized in this research is shown in Table 4.
Here, PGG is the images from the dataset which predicted glioma; PMG is the images from the dataset which were actually meningioma and predicted wrongly as glioma; PPG is the images from the dataset which were actually pituitary and predicted wrongly as glioma; PGM is the images from the dataset which were actually glioma and predicted wrongly as meningioma; PMM is the images from the dataset which were actually meningioma and predicted correctly as meningioma; PPM is the images from the dataset which were actually pituitary and predicted wrongly as meningioma; PGP is the images from the dataset which were actually glioma and predicted wrongly as pituitary; PMP is the images from the dataset which were actually meningioma and predicted wrongly as pituitary; and PPP is the images from the dataset which were actually pituitary and predicted correctly as pituitary. Furthermore, the accuracy, precision, recall, and F1-score are defined below.

4.1. Performance Metrics (Accuracy, Precision, Recall, and F1 Score)

Classification accuracy is the proportion between the correct predictions and the total data elements.
The equation that calculates the accuracy is shown in Equation (19) and Figure 3:
A c c = T P + T N T P + F P + T N + F N
where TP is true positive, or the estimated accurate amount of data, FP is false positive, or the fact that it is harmful and predicated as positive, TN is true negative, which is genuinely harmful and is predicated negatively, and FN is false negative, or the information that it is positive and predicated as unfavorable.
Precision (positive predicted values) is the proportion of optimistic forecasts that belong to the all positive class. The equation of the precision value is shown in Equation (20):
p r e c = T P T P + F P
Recall is defined as the number of true positive (TP) outcomes divided by the total number of elements corresponding to the positive class. The calculation for the recall value is presented in Equation (21):
  R e c a l l = T P T P + F N  
The F1 score is the average of the precision and recall. The calculation for the F1 score is presented in Equation (22):
F 1 m e a s u r e = 2 p r e c r e c a l l p r e c r e c a l l  
Additionally, after the proposed hybrid DeepTumorNet model was trained, the performance values of the accuracy, precion, recall, and F1 score were measured as presented in Table 5. The proposed hybrid deeptumorNet model obtained the highest accuracy, precision, recall, and F1 score values of 99.67%, 99.6%, 100%, and 99.6%, respectively. The loss and accuracy curves of the proposed DeepTumorNet approach are shown in Figure 3. The loss function shows how well the proposed hybrid model classified the tumor images into fine-grained tumors (gliomas, meningioma, and pituitary). The loss and accuracy of the proposed model after epoch 71 almost remained the same, demonstrating that the proposed hybrid approach classified the brain tumors with higher accuracy even at lower epochs than 120. The training and validation process of the proposed DeepTumorNet model is presented in Figure 3. A confusion matrix is shown in Table 6, which summarizes the correct and incorrect classifications of our proposed method.

4.2. Comparison of DeepTumorNet with the Pretrained Transfer Learning Approaches

This section compares some pre-trained transfer learning models with the proposed hybrid DeepTumorNet model. This experimentation aimed to evaluate the usefulness of the hybrid proposed model for BTs classification. At the same time, we compared the evaluation performance of our DeepTumorNet model with nine classical DL models (i.e., ExceptionNet, MobileNetv2, SqueezeNet, ShuffleNet, DenseNet, ShuffleNet, ResNet50, MobileNetv2, DarkNet-53, ResNet101, and AlexNet). All of these comparable DL models were employed using a transfer learning configuration trained on the ImageNet database. We changed the last three layers of each pre-trained model to adapt them to the target number of classes. The fully connected (FC) layer in every model was removed, and a new FC layer having an output size of three was introduced using the fine-tuning strategy, as we had three classes as output. Every model had a different input size, and like the size of the input images, GoogLeNet’s was 224 × 224, while 227 × 227 was the for SqueezeNet and other models. All models were fine-tuned with the same experimental parameters shown in Table 6. The fine-tuning classification method is illustrated in Figure 4. To perform the deep learning training, the dataset’s images were split at a ratio of 30% and 70% for testing and training, respectively, to obtain reliable results and efficient performance for the deep neural networks as shown in Figure 3. The results are presented in Table 5.
We witnessed that our method expressively outperformed the nine comparative DL models from the resulting observations. However, the Darknet53 and MobileNetV2 pretrained classifiers performed relatively better and close to the proposed hybrid DeepTumorNet model by achieving the second and third highest performance in classifying brain MRI images to meningioma, glioma, and pituitary types.

4.3. Comparative Results of DeepTumorNet with the State-of-the-Art Classification Approaches

Previously, researchers conducted relevant studies on BTs using different machine learning and DL techniques with a similar dataset. However, in some studies, the authors used binary classification classes (malignant and benign) of data and multiple classes (fine-grained classification) with fewer images. It is noticeable in Table 7 that the proposed deep learning-based hybrid model’s performance values were enhanced compared with the state-of-the-art approaches in the literature.
We compared the obtained results of our proposed DeepTumorNet approach with the state-of-the-art approaches in the literature [27,30,31,35,56,57]. The comparative results indicated the greater efficiency of the hybrid DeepTumorNet method over these techniques. Additionally, it was essential to use hand-crafted engineering for classification, which is computationally more complex. Moreover, the proposed model performed well than the mentioned methods in terms of accuracy, as depicted in Table 7. The proposed hybrid model comprised 154 layers, including newly added layers followed by the leaky ReLU activation function, decreasing the dying ReLU problem. Therefore, the proposed hybrid model extracted more descriptive and discriminative details and accurate features for classification.
In the proposed hybrid model, GoogLeNet was modified by adding and replacing the last 5 layers with 15 new layers and replacing the ReLU activation function with leaky ReLU. The newly proposed method performed better than the base model GoogLeNet in classification accuracy. The DeepTumorNet model is a convolutional neural network architecture with fewer hardware resources. It also requires a very convenient time for training. The increasing number of epoch and dataset sizes increased the time complexity of the network. However, the proposed deep learning-based hybrid model extracted more descriptive, detailed, and discriminative deep features. The proposed model can be embedded in an MRI machine for real-time BT classification. Moreover, it will help neurologists and surgeons in the treatment of BT patients.

5. Conclusions and Future Work

This research intended to classify BTs using different convolution neural networks and a new hybrid model. The GoogLeNet architecture was utilized as a base for the proposed DeepTumorNet framework. The last 5 layers of GoogLeNet were eliminated and 15 new deep layers were added in place of the 5 layers. Furthermore, the ReLU activation function was changed into the leaky ReLu activation function without disturbing the primary convolution neural network architecture. The total count of layers after the changes increased from 144 to 154. The proposed hybrid model reached a highest ever classification accuracy of 99.67%. In addition, we deployed nine deep pretrained CNN models using the transfer learning technique on the CE-MRI dataset to identify the BT types and compared their results with the proposed hybrid model. The experimental results demonstrated that the proposed hybrid model more accurately discriminated the brain tumors. Moreover, the proposed method computed more descriptive and discriminative details and accurate features for brain classification, resulting in high accuracy compared with the other state-of-art approaches.
Furthermore, it is evident from experimentation that the pretrained CNN model using transfer learning techniques produced the utmost performance. However, the hybrid framework accuracy reached the maximal compared with the rest of the pretrained models. Furthermore, in future work, experimenting with the dataset with a small number of malignant brain MRI images and a significant number of normal brain MRIs should be performed, as the proposed model extracted more detailed, discriminative, and accurate features. Therefore, before classifying the brain MRI images into two classes (i.e., malignant and benign), an efficient segmentation technique should be applied to brain MRI data. After that, the proposed model can accurately detect and classify benign and malignant images using segmented images.
Additionally, there is a possibility that other CNN networks can be converted to hybrid approaches to show better classification results with less time complexity. We aim to classify BTCT images and other large BT datasets [65] with the proposed DeepTumorNet model. Moreover, we plan to check the efficiency of the proposed hybrid method for other forms of medical image analysis, such as lung cancer, COVID-19, and pneumonia detection.

Author Contributions

Conceptualization, A.R.; methodology, A.R.; software, A.R., validation, A.R., H.A. and I.A.; formal analysis, A.R., D.J., J.A.K., I.A., Y.I.D. and A.S.S.; investigation, I.A. and A.U.R.; resources, H.A., J.A.K., Y.I.D., A.U.R. and H.H.; data curation, A.S.S., I.A., D.J., Y.I.D. and A.U.R.; writing—original draft preparation, A.R. and I.A.; writing—review and editing, I.A., Y.I.D., A.U.R. and H.H.; visualization, J.A.K. and A.U.R.; supervision, I.A. and J.A.K.; project administration, I.A., Y.I.D. and H.H.; funding acquisition, I.A., A.S.S. and H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Future University Researchers Supporting Project Number FUESP-2020/48 at Future University in Egypt, New Cairo 11845, Egypt.

Data Availability Statement

The datasets used in this investigation are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this paper:
AI.Artificial intelligence
MLMachine learning
DLDeep learning
TLTransfer Learning
BTBrain tumors
MRI.Magnetic resonance imaging
K-NNK-nearest neighbor
SVMSupport vector machine
CNNConvolutional neural network
PCA.Principal component analysis
LDALinear discriminant analysis

References

  1. Zhao, L.; Jia, K. Multiscale CNNs for brain tumors segmentation and diagnosis. Comput. Math. Methods Med. 2016, 2016, 8356294. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Saddique, M.; Kazmi, J.H.; Qureshi, K. A hybrid approach of using symmetry technique for brain tumors. Comput. Math. Methods Med. 2014, 2014, 712783. [Google Scholar] [CrossRef] [PubMed]
  3. Shree, N.V.; Kumar, T.N. Identification and classification of BTMRI images with feature extraction using DWT and probabilistic neural network. Brain Inform. 2018, 5, 23–30. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Komninos, J.; Vlassopoulou, V.; Protopapa, D.; Korfias, S.; Kontogeorgos, G.; Sakas, D.E.; Thalassinos, N.C. Tumors metastatic to the pituitary gland: Case report and literature review. J. Clin. Endocrinol. Metab. 2004, 2, 574–580. [Google Scholar] [CrossRef] [Green Version]
  5. DeAngelis, L.M. Brain tumors. N. Engl. J. Med. 2001, 344, 114–123. [Google Scholar] [CrossRef] [Green Version]
  6. Louis, D.N.; Perry, A.; Reifenberger, G.; Von Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization classification of tumors of the central nervous system: A summary. Acta Neuropathol. 2016, 132, 803–820. [Google Scholar] [CrossRef] [Green Version]
  7. Chahal, P.K.; Pandey, S.; Goel, S. A survey on brain tumors detection techniques for MR images. Multimed. Tools Appl. 2020, 79, 21771–21814. [Google Scholar] [CrossRef]
  8. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumors classification using deep CNN with extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [Google Scholar] [CrossRef]
  9. Rehman, A.; Naz, S.; Razzak, M.I.; Akram, F.; Imran, M. A deep learning-based framework for automatic brain tumors classification using transfer learning. Circuits Syst. Signal Process. 2020, 39, 757–775. [Google Scholar] [CrossRef]
  10. Wang, Y.; Zu, C.; Hu, G.; Luo, Y.; Ma, Z.; He, K.; Wu, X.; Zhou, J. Automatic tumor segmentation with deep convolutional neural networks for radiotherapy applications. Neural Process. Lett. 2018, 48, 1323–1334. [Google Scholar] [CrossRef]
  11. Jégou, S.; Drozdzal, M.; Vazquez, D.; Romero, A.; Bengio, Y. The one hundred layers tiramisu: Fully convolutional denseness for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 11–19. [Google Scholar]
  12. Zhang, Q.; Cui, Z.; Niu, X.; Geng, S.; Qiao, Y. Image segmentation with pyramid dilated convolution based on ResNet and U-Net. In Proceedings of the International Conference on Neural Information Processing, Guangzhou, China, 14 November 2017; pp. 364–372. [Google Scholar]
  13. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar]
  14. Ding, Y.; Zhang, C.; Lan, T.; Qin, Z.; Zhang, X.; Wang, W. Classification of Alzheimer’s disease based on the combination of morphometric feature and texture feature. In Proceedings of the 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Washington, DC, USA, 9–12 November 2015; pp. 409–412. [Google Scholar]
  15. Ijaz, A.; Ullah, I.; Khan, W.U.; Ur Rehman, A.; Adrees, M.S.; Saleem, M.Q.; Cheikhrouhou, O.; Hamam, H.; Shafiq, M. Efficient algorithms for E-healthcare to solve multiobject fuse detection problem. J. Healthc. Eng. 2021, 2021, 9500304. [Google Scholar]
  16. Ahmad, I.; Liu, Y.; Javeed, D.; Ahmad, S. A decision-making technique for solving order allocation problem using a genetic algorithm. IOP Conf. Ser. Mater. Sci. Eng. 2020, 853, 012054. [Google Scholar] [CrossRef]
  17. Binaghi, E.; Omodei, M.; Pedoia, V.; Balbi, S.; Lattanzi, D.; Monti, E. Automatic segmentation of MR brain tumors images using support vector machine in combination with graph cut. In Proceedings of the 6th International Joint Conference on Computational Intelligence (IJCCI), Rome, Italy, 22–24 October 2014; pp. 152–157. [Google Scholar]
  18. Bahadure, N.B.; Ray, A.K.; Thethi, H.P. The comparative approach of MRI-based brain tumors segmentation and classification using genetic algorithm. J. Digit. Imaging 2018, 31, 477–489. [Google Scholar] [CrossRef]
  19. Kaya, I.E.; Pehlivanlı, A.Ç.; Sekizkardeş, E.G.; Ibrikci, T. PCA based clustering for brain tumors segmentation of T1w MRI images. Comput. Methods Programs Biomed. 2017, 140, 19–28. [Google Scholar] [CrossRef]
  20. Zikic, D.; Glocker, B.; Konukoglu, E.; Criminisi, A.; Demiralp, C.; Shotton, J.; Thomas, O.M.; Das, T.; Jena, R.; Price, S.J. Forests for tissue-specific segmentation of high-grade gliomas in multi-channel MR. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Berlin, Heidelberg, 1 October 2012; pp. 369–376. [Google Scholar]
  21. Nikam, R.D.; Lee, J.; Choi, W.; Banerjee, W.; Kwak, M.; Yadav, M.; Hwang, H. Ionic sieving through one-atom-thick 2D material enables analog nonvolatile memory for neuromorphic computing. Adv. Electron. Mater. 2021, 17, 2100142. [Google Scholar] [CrossRef]
  22. Zhang, J.P.; Li, Z.W.; Yang, J. A parallel SVM training algorithm on large-scale classification problems. In Proceedings of the 2005 International Conference on Machine Learning and Cybernetics, Guangzhou, China, 18–21 August 2005; pp. 1637–1641. [Google Scholar]
  23. Ye, F.; Yang, J. A deep neural network model for speaker identification. Appl. Sci. 2021, 11, 3603. [Google Scholar] [CrossRef]
  24. Ijaz, A.; Liu, Y.; Javeed, D.; Shamshad, N.; Sarwr, D.; Ahmad, S. A review of artificial intelligence techniques for selection & evaluation. IOP Conf. Ser. Mater. Sci. Eng. 2020, 853, 012055. [Google Scholar]
  25. Akbari, H.; Khalighinejad, B.; Herrero, J.L.; Mehta, A.D.; Mesgarani, N. Towards reconstructing intelligible speech from the human auditory cortex. Sci. Rep. 2019, 9, 874. [Google Scholar] [CrossRef] [Green Version]
  26. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  27. Harish, P.; Baskar, S. MRI based detection and classification of brain tumor using enhanced faster R-CNN and Alex Net model. Mater. Today Proc. 2020. [Google Scholar] [CrossRef]
  28. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. AlexNet-level accuracy with 50× fewer parameters and <0.5 MB model size. In Proceedings of the 5th International Conference Learning Representation, Toulon, France, 24–26 April 2016; pp. 1–8. [Google Scholar]
  29. Saleem, S.; Amin, J.; Sharif, M.; Anjum, M.A.; Iqbal, M.; Wang, S.H. A deep network designed for segmentation and classification of leukemia using fusion of the transfer learning models. Complex Intell. Syst. 2021, 7, 1–16. [Google Scholar] [CrossRef]
  30. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  33. Gupta, A.; Verma, A.; Kaushik, D.; Garg, M. Applying deep learning approach for brain tumors detection. Mater. Today Proc. 2020, 50, 1–13. [Google Scholar] [CrossRef]
  34. Mehrotra, R.; Ansari, M.A.; Agrawal, R.; Anand, R.S. A transfer learning approach for AI-based classification of brain tumors. Mach. Learn. Appl. 2020, 2, 100003. [Google Scholar] [CrossRef]
  35. Ramzan, F.; Khan, M.U.G.; Rehmat, A.; Iqbal, S.; Saba, T.; Rehman, A.; Mehmood, Z. A deep learning approach for automated diagnosis and multi-class classification of Alzheimer’s disease stages using resting-state fMRI and residual neural networks. J. Med. Syst. 2020, 44, 37. [Google Scholar] [CrossRef]
  36. Selvapandian, A.; Manivannan, K. Fusion-based glioma brain tumors detection and segmentation using ANFIS classification. Comput. Methods Programs Biomed. 2018, 166, 33–38. [Google Scholar] [CrossRef]
  37. Sultan, H.H.; Salem, N.M.; Al-Atabany, W. Multi-classification of brain tumors images using deep neural Network. IEEE Access 2019, 7, 69215–69225. [Google Scholar] [CrossRef]
  38. Tufail, A.B.; Ijaz Ahmad, A. Diagnosis of diabetic retinopathy through retinal fundus images and 3D convolutional neural networks with limited number of samples. Wirel. Commun. Mob. Comput. 2021, 2021, 6013448. [Google Scholar] [CrossRef]
  39. Pitchai, R.; Supraja, P.; Victoria, A.H.; Madhavi, M. Brain tumors egmentation using deep learning and fuzzy K-means clustering for magnetic resonance images. Neural Processing Lett. 2021, 53, 2519–2532. [Google Scholar] [CrossRef]
  40. Amin, J.; Sharif, M.; Yasmin, M.; Fernandes, S.L. A distinctive approach in brain tumors detection and classification using MRI. Pattern Recognit. Lett. 2020, 139, 118–127. [Google Scholar] [CrossRef]
  41. Özyurt, F.; Sert, E.; Avci, E.; Dogantekin, E. Brain tumors detection based on Convolutional Neural Network with neutrosophic expert maximum fuzzy sure entropy. Measurement 2019, 147, 106830. [Google Scholar] [CrossRef]
  42. Kaplan, K.; Kaya, Y.; Kuncan, M.; Ertunç, H.M. Brain tumors classification using modified local binary patterns (LBP) feature extraction methods. Med. Hypotheses 2020, 139, 109696. [Google Scholar] [CrossRef] [PubMed]
  43. Çinar, A.; Yildirim, M. Detection of tumors on brain MRI images using the hybrid convolutional neural network architecture. Med. Hypotheses 2020, 139, 109684. [Google Scholar] [CrossRef] [PubMed]
  44. Mohsen, H.; El-Dahshan, E.S.; El-Horbaty, E.S.; Salem, A.B. Classification using deep learning neural networks for brain tumors. Future Comput. Inform. J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  45. Lather, M.; Singh, P. Investigating brain tumors segmentation and detection techniques. Procedia Comput. Sci. 2020, 167, 121–130. [Google Scholar] [CrossRef]
  46. Javeed, D.; Gao, T.; Khan, M.T.; Ahmad, I. A Hybrid Deep Learning-Driven SDN Enabled Mechanism for Secure Communication in Internet of Things (IoT). Sensors 2021, 21, 4884. [Google Scholar] [CrossRef]
  47. Marghalani, B.F.; Arif, M. Automatic classification of brain tumors and Alzheimer’s disease in MRI. Procedia Comput. Sci. 2019, 163, 78–84. [Google Scholar] [CrossRef]
  48. Swati, Z.N.K.; Zhao, Q.; Kabir, M.; Ali, F.; Ali, Z.; Ahmed, S.; Lu, J. Brain tumors classification for MR images using transfer learning and fine-tuning. Comput. Med. Imaging Graph. 2019, 75, 34–46. [Google Scholar] [CrossRef]
  49. Kumar, S.; Mankame, D.P. Optimization drove deep convolution neural network for brain tumors classification. Biocybern. Biomed. Eng. 2020, 40, 1190–1204. [Google Scholar] [CrossRef]
  50. Deepak, S.; Ameer, P.M. Brain tumors classification using in-depth CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef]
  51. Raja, P.S. Brain tumors classification using a hybrid deep autoencoder with Bayesian fuzzy clustering-based segmentation approach. Biocybern. Biomed. Eng. 2020, 40, 440–453. [Google Scholar] [CrossRef]
  52. Ramamurthy, D.; Mahesh, P.K. Whale Harris Hawks optimization-based deep learning classifier for brain tumors detection using MRI images. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 1–14. [Google Scholar] [CrossRef]
  53. Bahadure, N.B.; Ray, A.K.; Thethi, H.P. Image analysis for MRI-based brain tumors detection and feature extraction using biologically inspired BWT and SVM. Int. J. Biomed. Imaging 2017, 2017, 9749108. [Google Scholar] [CrossRef] [Green Version]
  54. Waghmare, V.K.; Kolekar, M.H. Brain tumors classification using deep learning. In The Internet of Things for Healthcare Technologies; Springer: Singapore, 2021; Volume 73, pp. 155–175. [Google Scholar]
  55. BT. Data Set. Available online: https://figshare.com/articles/dataset/brain_tumor_dataset/1512427 (accessed on 3 January 2022).
  56. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced performance of brain tumors classification via tumor region augmentation and partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  57. Resize Function. Available online: https://www.mathworks.com/products/matlab.html (accessed on 4 January 2022).
  58. Agarwal, M.; Gupta, S.; Biswas, K.K. A new conv2d model with modified relu activation function for identifying disease type and severity in cucumber plants. Sustain. Comput. Inform. Syst. 2021, 30, 100473. [Google Scholar]
  59. Botta, B.; Gattam, S.S.; Datta, A.K. Eggshell crack detection using deep convolutional neural networks. J. Food Eng. 2022, 315, 110798. [Google Scholar] [CrossRef]
  60. Thakur, S.; Kumar, A. X-ray and CT-scan-based automated detection and classification of COVID-19 using convolutional neural networks (CNN). Biomed. Signal Process. Control 2021, 69, 102920. [Google Scholar] [CrossRef]
  61. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-margin Softmax Loss for Convolutional Neural Networks. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19 June 2016; pp. 1–7. [Google Scholar]
  62. Ruuska, S.; Hämäläinen, W.; Kajava, S.; Mughal, M.; Matilainen, P.; Mononen, J. Evaluation of the confusion matrix method invalidating an automated system for measuring feeding behavior of cattle. Behav. Process. 2018, 148, 56–62. [Google Scholar] [CrossRef] [Green Version]
  63. Anaraki, A.K.; Ayati, M.; Kazemi, F. Magnetic resonance imaging-based brain tumors grades classification and grading via convolutional neural networks and genetic algorithms. Biocybern. Biomed. Eng. 2019, 39, 63–74. [Google Scholar] [CrossRef]
  64. Abiwinanda, N.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain tumors classification using convolutional neural network. In World Congress on Medical Physics and Biomedical Engineering; Springer: Singapore, 2018; pp. 183–189. [Google Scholar]
  65. Ullah, Z.; Farooq, M.U.; Lee, S.H.; An, D. A hybrid image enhancement-based brain MRI images classification technique. Med. Hypotheses 2020, 143, 109922. [Google Scholar] [CrossRef]
Figure 1. Three different tumors (meningioma, glioma, and pituitary tumor) in three different views.
Figure 1. Three different tumors (meningioma, glioma, and pituitary tumor) in three different views.
Electronics 11 01146 g001
Figure 2. Architecture of proposed hybrid model.
Figure 2. Architecture of proposed hybrid model.
Electronics 11 01146 g002
Figure 3. Classification performance (accuracy (%)) of the proposed hybrid model.
Figure 3. Classification performance (accuracy (%)) of the proposed hybrid model.
Electronics 11 01146 g003
Figure 4. Diagram of transfer learning-based classification.
Figure 4. Diagram of transfer learning-based classification.
Electronics 11 01146 g004
Table 1. Explanation of CE-MRI dataset.
Table 1. Explanation of CE-MRI dataset.
Tumor ClassPatientsImagesView of MRI.No. of MRI. Images
Meningioma82708A *209
C *268
S *231
Pituitary62930A *291
C *319
S *320
Glioma891426A *494
C *437
S *495
Total2333064 3064
* Axial = A, * Coronal = C, and * Sagittal = S.
Table 2. Characteristics of added layers in the proposed hybrid model.
Table 2. Characteristics of added layers in the proposed hybrid model.
S.NoLayer NameTypeNo of FilterFilter SizeEpsilon
1block_16_expandConv9601 × 1
2block_16_expand_BNBatch Norm 0.001
3block_16_expand_reluClipped ReLU Layer
4block_16_depthwiseGrouped Conv9603 × 3
5block_16_depthwise_BNBatch Norm 0.001
6block_16_depthwise_reluClipped ReLU Layer 0.001
7block_16_projectConv3201 × 1
8block_16_project_BNBatch Norm 0.001
9Conv_1Conv12801 × 1
10Conv_1_bnBatch Norm 0.001
11out_reluClipped ReLU Layer
12global_average_pooling2d_1Global Average Pooling
13LogitsFully Connected
14Logits_softmaxSoftmax
15ClassificationLayer_LogitsClassification Layer
Table 3. Parameter values used in training networks.
Table 3. Parameter values used in training networks.
NameSGDM
MiniBtachSize10
Number of Epochs120
Initial Learning Rate0.01
Shuffleevery epoch
Validation Frequency50
Table 4. General confusion matrix.
Table 4. General confusion matrix.
Predicated Classes
GliomasMeningiomasPituitary
Actual ClassGliomasPGGPMGPPG
MeningiomasPGMPMMPPM
PituitaryPGPPMPPPP
Table 5. Experimentation results of pretrained models.
Table 5. Experimentation results of pretrained models.
ModelAccuracyPrecisionRecallF1-Score
Proposed Model99.67%99.6%100%99.66%
AlexNet97.8%97.6%97.66%97.66%
GoogLeNet98.26%98%98.66%98.33%
Shufflenet98.37%98.33%98.66%98.33%
ResNet5098.60%98.33%98.66%98.33%
MobileNet V299%99%99%99%
SqueezeNet97.91%97.66%98%97.66%
Darknet5399.13%99%99.33%99%
Resnet10198.91%98.66%99%98.66%
ExceptionNet98.69%98.33%98.33%98%
Table 6. Confusion matrix of proposed hybrid model.
Table 6. Confusion matrix of proposed hybrid model.
Predicated Classes
GliomasMeningiomasPituitary
Actual ClassGliomas4260101
Meningiomas0121100
Pituitary0000278
Table 7. Comparative study of the proposed model with recent ML and DL models.
Table 7. Comparative study of the proposed model with recent ML and DL models.
AuthorTechniqueClassification TypeDatasetAccuracy (%)
Mehrotra et al., (2020) [27]PT-CNN: AlexNetBinary ClassT1-weighed MRI
(Benign = 224; malignant = 472)
99.04%
Kaplan et al., (2020) [35]LBP SVM KNNMulti-ClassT1-weighed CE-MRI
(meningiomas = 708; gliomas = 1426; pituitary = 930)
95.56%
Sultan et al., (2019) [30]CNNMulti-ClassT1-weighed CE-MRI
(meningiomas = 208; gliomas = 492; pituitary = 289)
96%
Anaraki et al., (2019) [63]GA-CNNMulti-ClassT1-weighed CE-MRI
(meningiomas = 708; gliomas = 1426; pituitary = 930)
94.20%
Kumar et al., (2017) [31]GWO+M-SVMMulti-ClassT1-weighed CE-MRI
(meningiomas = 248; gliomas = 12; pituitary = 55)
95.23%
Bahadur et al., (2017) [64]BWT+SVMBinaryT2-weighted images
(normal = 67; abnormal = 134)
95%
Abiwinanda et al., (2019)CNNMulti-ClassT1-weighed CE-MRI
(meningiomas = 708, gliomas = 1426; pituitary = 930)
84%
Proposed methodDeep CNNMulti-ClassT1-weighed CE-MRI
(meningiomas = 708; gliomas = 1426; pituitary = 930)
99.67%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Raza, A.; Ayub, H.; Khan, J.A.; Ahmad, I.; S. Salama, A.; Daradkeh, Y.I.; Javeed, D.; Ur Rehman, A.; Hamam, H. A Hybrid Deep Learning-Based Approach for Brain Tumor Classification. Electronics 2022, 11, 1146. https://doi.org/10.3390/electronics11071146

AMA Style

Raza A, Ayub H, Khan JA, Ahmad I, S. Salama A, Daradkeh YI, Javeed D, Ur Rehman A, Hamam H. A Hybrid Deep Learning-Based Approach for Brain Tumor Classification. Electronics. 2022; 11(7):1146. https://doi.org/10.3390/electronics11071146

Chicago/Turabian Style

Raza, Asaf, Huma Ayub, Javed Ali Khan, Ijaz Ahmad, Ahmed S. Salama, Yousef Ibrahim Daradkeh, Danish Javeed, Ateeq Ur Rehman, and Habib Hamam. 2022. "A Hybrid Deep Learning-Based Approach for Brain Tumor Classification" Electronics 11, no. 7: 1146. https://doi.org/10.3390/electronics11071146

APA Style

Raza, A., Ayub, H., Khan, J. A., Ahmad, I., S. Salama, A., Daradkeh, Y. I., Javeed, D., Ur Rehman, A., & Hamam, H. (2022). A Hybrid Deep Learning-Based Approach for Brain Tumor Classification. Electronics, 11(7), 1146. https://doi.org/10.3390/electronics11071146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop