Next Article in Journal
Castration-Resistant Prostate Cancer Outcome Prediction Using Phased Long Short-Term Memory with Irregularly Sampled Serial Data
Next Article in Special Issue
Sleep State Classification Using Power Spectral Density and Residual Neural Network with Multichannel EEG Signals
Previous Article in Journal
Mechanical Behavior of Circular Steel Tubular Beam-Columns Corroded Uniformly in Atmospheric Environment
Previous Article in Special Issue
Compressed-Sensing Magnetic Resonance Image Reconstruction Using an Iterative Convolutional Neural Network Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network

by
Milica M. Badža
1,2,* and
Marko Č. Barjaktarović
2
1
Innovation center, School of Electrical Engineering, University of Belgrade, Bulevar kralja Aleksandra 73, 11000 Belgrade, Serbia
2
School of Electrical Engineering, University of Belgrade, Bulevar kralja Aleksandra 73, 11000 Belgrade, Serbia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(6), 1999; https://doi.org/10.3390/app10061999
Submission received: 14 January 2020 / Revised: 11 March 2020 / Accepted: 12 March 2020 / Published: 15 March 2020
(This article belongs to the Special Issue Medical Signal and Image Processing)

Abstract

:
The classification of brain tumors is performed by biopsy, which is not usually conducted before definitive brain surgery. The improvement of technology and machine learning can help radiologists in tumor diagnostics without invasive measures. A machine-learning algorithm that has achieved substantial results in image segmentation and classification is the convolutional neural network (CNN). We present a new CNN architecture for brain tumor classification of three tumor types. The developed network is simpler than already-existing pre-trained networks, and it was tested on T1-weighted contrast-enhanced magnetic resonance images. The performance of the network was evaluated using four approaches: combinations of two 10-fold cross-validation methods and two databases. The generalization capability of the network was tested with one of the 10-fold methods, subject-wise cross-validation, and the improvement was tested by using an augmented image database. The best result for the 10-fold cross-validation method was obtained for the record-wise cross-validation for the augmented data set, and, in that case, the accuracy was 96.56%. With good generalization capability and good execution speed, the new developed CNN architecture could be used as an effective decision-support tool for radiologists in medical diagnostics.

1. Introduction

Cancer is the second leading cause of death globally, according to the World Health Organization (WHO) [1]. Early detection of cancer can prevent death, but this is not always possible. Unlike cancer, a tumor could be benign, pre-carcinoma, or malign. Benign tumors differ from malign in that benign generally do not spread to other organs and tissues and can be surgically removed [2].
Some of the primary brain tumors are gliomas, meningiomas, and pituitary tumors. Gliomas are a general term for tumors that arise from brain tissues other than nerve cells and blood vessels. On the other hand, meningiomas arise from the membranes that cover the brain and surround the central nervous system, whereas pituitary tumors are lumps that sit inside the skull [3,4,5,6]. The most important difference between these three types of tumors is that meningiomas are typically benign, and gliomas are most commonly malignant. Pituitary tumors, even if benign, can cause other medical damage, unlike meningiomas, which are slow-growing tumors [5,6]. Because of the information mentioned above, the precise differentiation between these three types of tumors represents a very important step of the clinical diagnostic process and later effective assessment of patients.
The most common method for differential diagnostics of tumor type is magnetic resonance imaging (MRI). However, it is susceptible to human subjectivity, and a large amount of data is difficult for human observation. Early brain–tumor detection mostly depends on the experience of the radiologist [7]. The diagnostics of the tumor could not be complete before establishing whether it is benign or malignant. In order to examine whether the tissue is benign or malignant, a biopsy is usually performed. Unlike tumors elsewhere in the body, the biopsy of the brain tumor is not usually obtained before definitive brain surgery [8]. In order to obtain precise diagnostics, and to avoid surgery and subjectivity, it is important to develop an effective diagnostics tool for tumor segmentation and classification from MRI images [7].
The development of new technologies, especially artificial intelligence and machine learning, has had a significant impact on medical field, providing an important support tool for many medical branches, including imaging. Different machine-learning methods for image segmentation and classification are applied in MRI image processing to provide radiologists with a second opinion.
Since 2012, the Perelman School of Medicine at the University of Pennsylvania, Center for Biomedical Image Computing & Analytics (CBICA) has been running an online competition, the Multimodal Brain Tumor Segmentation Challenge (BRATS) [9]. The image databases used in BRATS are made publicly available after the competition is finished. Different classification algorithms designed using these image databases can be found in many papers [10,11,12,13,14]. However, the databases are usually small, on average about 285 images, and they often contain images showing two tumor levels, low and high level of glioma tumor, acquired in the axial plane [10].
In addition, classification has been carried out on other image databases, which are also quite small [15,16,17,18]. Mohsen et al. used 66 images to classify four types of images showing brain tumors: tumor-free, glioblastoma, sarcoma, and metastasis. By using a deep neural network (DNN), they obtained an accuracy of 96.97% [17].
In the literature, there are other algorithms and different modifications of the pre-trained networks that are used for image analysis, classification, and segmentation. Different approaches have been tested on other medical databases, both on MRI images of brain tumors and on tumors from different parts of the human body [19,20]. These papers were not considered further, as the focus was on the papers using the same MRI image database that we used.
Cheng et al., who were the first to present the image database used in this paper, classified the tumor types using augmented tumor region of interest, image dilatation, and ring-form partition. They extracted features using intensity histogram, gray level co-occurrence matrix, and bag-of-words models, and achieved an accuracy of 91.28% [21]. More papers that used the same database are discussed in Section 3. There, we discuss different types of networks, pre-trained ones, capsule net networks, other architectures of convolutional networks, and combinations with neural networks for feature extraction and classifiers for the output result. The discussion also concerns approaches using different modifications of the database, as well as the original one. The papers that used the original or augmented database are listed in the tables for better comparison.
The biggest problem with classifying and segmenting the MRI images with some neural networks lies in the number of images in the database. In addition, MRI images are acquired in different planes, so the option of using all the available planes could enlarge the database. As this could generally affect the classification output by overfitting, pre-processing is required before feeding the images into the neural network. However, one of the known advantages of convolutional neural networks (CNN) is that the pre-processing and the feature engineering do not have to be performed.
The aim of this research is firstly to examine the classification of three tumor types from an imbalanced database with a CNN. Although considered large compared to other available MRI image databases, this database is still far smaller than databases generally used in the field of artificial intelligence. We wanted to show that the performance of the small architecture could compare favorably with the performance of the more complex ones. Using a simpler network requires fewer resources for training and implementation. This is a crucial problem to address because limited available resources make it difficult to use the system in clinical diagnostics and on mobile platforms. If the system is needed to be used in everyday clinical diagnostics, it should be generally applicable. We wanted to examine the network’s generalization capability for clinical studies and to show how the subject-wise cross-validation approach gives more realistic results for further implementation.
In this paper, we present a new CNN architecture for brain tumor classification of three tumor types: meningioma, glioma, and pituitary tumor from T1-weighted contrast-enhanced magnetic resonance images. The network performance was tested using four approaches: combinations of two 10-fold cross-validation methods (record-wise and subject-wise) and two databases (original and augmented). The results are presented using the confusion matrices and accuracy metric. A comparison with the comparable state-of-the-art methods is also presented.

2. Methodology

2.1. Image Database

The image database, provided as a set of slices, used in this paper contains 3064 T1-weighted contrast-enhanced MRI images acquired from Nanfang Hospital and General Hospital, Tianjin Medical University, China from 2005 to 2010. It was first published online in 2015, and the last modified version was realized in 2017 [22]. There are three types of tumors: meningioma (708 images), glioma (1426 images), and pituitary tumor (930 images). All images were acquired from 233 patients in three planes: sagittal (1025 images), axial (994 images), and coronal (1045 images) plane. The examples of different types of tumors, as well as different planes, are shown in Figure 1. The tumors are marked with a red outline. The number of images is different for each patient.

2.2. Image Pre-Processing and Data Augmentation

Magnetic resonance images from the database were of different sizes and were provided in int16 format. These images represent the input layer of the network, so they were normalized and resized to 256 × 256 pixels.
In order to augment the dataset, we transformed each image in two ways. The first transformation was image rotation by 90 degrees. The second transformation was flipping images vertically [23]. In this way, we augmented our dataset three times, resulting in 9192 images.

2.3. Network Architecture

Tumor classification was performed using a CNN developed in Matlab R2018a (The MathWorks, Natick, MA, USA). The network architecture consists of input, two main blocks, classification block, and output, as shown in Figure 2. The first main block, Block A, consists of a convolutional layer which as an output gives an image two times smaller than the provided input. The convolutional layer is followed by the rectified linear unit (ReLU) activation layer and the dropout layer. In this block, there is also the max pooling layer which gives an output two times smaller than the input. The second block, Block B, is different from the first only in the convolution layer, which retains the same output size as the input size of that layer. The classification block consists of two fully connected (FC) layers, of which the first one represents the flattened output of the last max pooling layer, whereas, in the second FC layer, the number of hidden units is equal to the number of the classes of tumor. The whole network architecture consists of the input layer, two Blocks A, two Blocks B, classification block, and output layer; altogether, there are 22 layers, as shown in Table 1.

2.4. Training Network

We used a k-fold cross-validation method to test the network performance [24]. Two different approaches were implemented, and both consisted of 10-fold cross-validation. The first approach was to randomly divide the data into 10 approximately equal portions so that each tumor category was equally present in each portion, referred to as record-wise cross-validation. The second approach was to randomly divide the data into 10 approximately equal portions where the data from a single subject could only be found in one of the sets. Each set, therefore, contained data from a couple of subjects regardless of the tumor class, referred to as subject-wise cross-validation. The second approach was implemented to test the generalization capability of the network in medical diagnostics [25]. The generalization capability in clinical practice represents the ability to predict the diagnosis based on the data obtained from subjects from which there are no observations in the training process. Therefore, observations from individuals in the training set must not appear in the test set. If this is not the case, complex predictors can pick up a confounding relationship between identity and diagnostic status and so produce unrealistically high prediction accuracy [26]. In order to compare the performance of our network with other state-of-the-art methods, we also tested our network without k-fold cross-validation (one test). In all the above-mentioned methods, two data portions were used for the test, two for validation, and six for training. Both datasets, normal and augmented, were tested using all the methods.
The network was trained using an Adam optimizer, with a mini-batch size equal to 16 and data shuffling in every iteration. The early-stop condition that affects when the process of network training will stop corresponds to one epoch. More specifically, it was tuned to finish the training process after the one epoch, when the loss starts to increase. The regularization factor was set to 0.004, and the initial learning rate to 0.0004. The weights of the convolutional layers were initialized using a Glorot initializer, also known as Xavier initializer [27].
The training process was stopped when the loss on the validation set got larger than or was equal to the previous lowest loss for 11 times. The network was trained and tested on a single graphical processing unit (GPU), CUDA device, GeForce GTX 1050 Ti.

3. Results and Discussion

Results of the developed CNN are shown in Table 2 and visualized using the confusion matrices, as shown in Figures 3 and 5–7. In confusion matrices, non-white rows represent network output classes, and non-white columns correspond to real classes in Figures 3 and 5–7. The numbers/percentages of correctly classified images are shown on the diagonal. The last row represents the sensitivity, whereas the last column corresponds to the specificity. Overall accuracy is shown in the bottom-right field. The upper number in the non-white boxes corresponds to the number of images, and the lower number represents the percentage of the whole class database in the training or test set. In order to neglect the imbalance of classes of tumors in the database, we have also shown mean average precision, recall, and F1-score in Table 2.
Figure 3 shows confusion matrices for the record-wise 10-fold cross-validation approach for testing data obtained from the original dataset. The classification error for the testing set after cross-validation is equal to 4.6%.
Examples of classified images from the original dataset with record-wise 10-fold cross-validation are shown in Figure 4, with the tumors outlined in red. Figure 4 comprises confusion matrices, where the rows show examples of images for the outputs, and columns emphasize what the intended target was, illustrating correctly and wrongly classified images.
Confusion matrices for the record-wise 10-fold cross-validation method for testing data from the augmented dataset are shown in Figure 5. The classification error for the testing data is 3.4%.
Figure 6 presents confusion matrices for the subject-wise 10-fold cross-validation testing method for testing data obtained from the original dataset. The classification error for the testing set is 15.7%. This result shows that the network has a smaller accuracy than for record-wise cross-validation, which was expected because the predictions were made based on previously unseen data.
Confusion matrices for the subject-wise 10-fold cross-validation approach for testing data from the augmented dataset are shown in Figure 7. The classification error for the testing data is 11.5%.
The proposed architecture of the CNN had only 4.3 million weights, and it obtained better results with augmented data, which was expected because the data set is not especially extensive. Even with the augmented data set, the subject-wise accuracy is lower than the accuracy obtained with the record-wise cross-validation because, with the augmentation, we only increased the number of images for individual patients, not the number of patients. As a consequence of splitting the data with the subject-wise method, increasing the number of patients was more important. The first class of tumors, meningioma, had the lowest sensitivity and specificity for all four testing methods. This is easily explained given that meningioma is the hardest to discern from the other two types of tumors based on the place of origin and overall features. The execution speed was quite good with an average of less than 15 ms per image.

Comparison with State-of-the-Art-Methods

There are several papers that use the same database for brain tumor classification. In order to compare our results with those of previous studies, we selected only those papers which had designed neural networks, used whole images as input for classification, and tested their networks with a k-fold cross-validation method, as shown in Table 3. We also compared our results with those of researchers who had not tested the network with k-fold cross-validation, as shown in Table 4. A comparison with the studies that used designed neural networks and an augmented dataset, but did not test it by k-fold cross-validation, is presented in Table 5.
In the literature, there are also studies that used the same database for classification with pre-trained networks [23,35,36,37,38,39,40] or, as input, they use only tumor region or some features that are extracted from the tumor region [7,21,23,41,42]. Similarly, in several papers, researchers have modified this database prior to classification [36,43,44,45,46,47]. The designed networks are usually simpler than already-existing pre-trained networks and have faster execution speed. To our knowledge, the best results using the pre-trained network are 98.69% [40] and 98.66% accuracy [36]. Rehman et al. [40] preprocessed images with contrast enhancement and augmented the dataset. The augmentation was fivefold, with rotations of 90, 180, and 270 degrees and horizontal and vertical flipping. The best result was obtained with a fine-tuned VGG16 trained using stochastic gradient descent with momentum. Although our approach has a 1.41% higher classification error, it has 4.3 million weights as opposed to the VGG16, which is a very deep network with 138 million weights. Very deep networks such as VGG16 and AlexNet require dedicated hardware for real-time performance. Kutlu and Avcı [36] also modified the database, using only those images that were taken in the axial plane, and used only 100 images of each tumor type. For feature extraction, they used the pre-trained AlexNet, and, for testing, they performed a 5-fold cross-validation method. It is unclear how the algorithm will perform on the whole dataset and what its generalization capabilities are.
Developing the network which uses only the region of the tumor or some other segmented part as input is better in terms of speed of execution, but also requires methods for segmentation or a dedicated expert who would mark those parts.
To our knowledge, the best result in the literature using the segmented image parts as inputs are presented by Tripathi and Bag [41], with 94.64% accuracy. For input to the classifiers, they use features that are extracted from the segmented brain from the image. They tested their approach using a 5-fold cross-validation method.

4. Conclusions

A new CNN architecture for brain tumor classification was presented in this study. The classification was performed using a T1-weighted contrast-enhanced MRI image database which contains three tumor types. As input, we used whole images, so it was not necessary to perform any preprocessing or segmentation of the tumors. Our designed neural network is simpler than pre-trained networks, and it is possible to run it on conventional modern personal computers. This is possible because the algorithm requires many less resources for both training and implementation. The importance of developing smaller networks is also linked to the possibility of deploying the algorithm on mobile platforms, which is significant for diagnostics in developing countries [48]. In addition, the network has a very good execution speed of 15 ms per image. In order to test the network, we used record-wise and subject-wise 10-fold cross-validation on both the original and augmented image database. In clinical diagnostics, the generalization capability implies predictions for subjects from whom we have no observations. With this in mind, the observations from individuals in the training set must not appear in the test set. If this condition is not met, complex predictors can have unrealistically high prediction accuracy due to the confounding dependency between the identity and the diagnosis of a patient [26]. In relation to that knowledge, we have committed subject-wise cross-validation.
A comparison with the comparable state-of-the-art methods shows that our network obtained better results. The best result for 10-fold cross-validation was achieved for the record-wise method and, for the augmented dataset, and the accuracy was 96.56%. To our knowledge, in the literature, there is no paper that shows tested generalization, through subject-wise k-fold method, for this image database. For the subject-wise approach, we obtained an accuracy of 88.48% for the augmented dataset. The average test execution was less than 15 ms per image. These results show that our network has a good generalization capability and good execution speed, so it could be used as an effective decision-support tool for radiologists in medical diagnostics.
Regarding further work, we will consider other approaches to database augmentation (e.g., increasing number of subjects) in order to improve the generalization capability of the network. One of the main improvements will be adjusting the architecture so that it could be used during brain surgery, classifying and accurately locating the tumor [49]. Detecting the tumors in the operating room should be performed in real-time and real-world conditions; thus, in that case, the improvement would also involve adapting the network to a 3D system [50]. By keeping the network architecture simple, detection in real time could be possible. In future, we will examine the performance of our designed neural network, as well as improved ones, on other medical images.

Author Contributions

Conceptualization, M.M.B.; Methodology, M.M.B. and M.Č.B.; Software, M.M.B.; Validation, M.M.B. and M.Č.B.; Writing—original draft, M.M.B.; Writing—review and editing, M.Č.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

We thank Vladislava Bobić and Ivan Vajs, the researchers at the Innovation Center School of Electrical Engineering, University of Belgrade, and Petar Atanasijević, the teaching assistant at the School of Electrical Engineering, for their valuable comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization—Cancer. Available online: https://www.who.int/news-room/fact-sheets/detail/cancer. (accessed on 5 November 2019).
  2. Priya, V.V. An Efficient Segmentation Approach for Brain Tumor Detection in MRI. Indian J. Sci. Technol. 2016, 9, 1–6. [Google Scholar]
  3. Cancer Treatments Centers of America—Brain Cancer Types. Available online: https://www.cancercenter.com/cancer-types/brain-cancer/types (accessed on 30 November 2019).
  4. American Association of Neurological Surgeons—Classification of Brain Tumours. Available online: https://www.aans.org/en/Media/Classifications-of-Brain-Tumors (accessed on 30 November 2019).
  5. DeAngelis, L.M. Brain Tumors. New Engl. J. Med. 2001, 344, 114–123. [Google Scholar] [CrossRef] [Green Version]
  6. Louis, D.N.; Perry, A.; Reifenberger, G.; Von Deimling, A.; Figarella-Branger, M.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization Classification of Tumors of the Central Nervous System: A summary. Acta Neuropathol. 2016, 131, 803–820. [Google Scholar] [CrossRef] [Green Version]
  7. Afshar, P.; Plataniotis, K.N.; Mohammadi, A. Capsule Networks for Brain Tumor Classification Based on MRI Images and Coarse Tumor Boundaries. In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1368–1372. [Google Scholar]
  8. Byrne, J.; Dwivedi, R.; Minks, D. Tumours of the brain. In Nicholson T (ed) Recommendations Cross Sectional Imaging Cancer Management, 2nd ed.; Royal College of Radiologists: London, UK, 2014; pp. 1–20. Available online: https://www.rcr.ac.uk/publication/recommendations-cross-sectional-imaging-cancer-management-second-edition (accessed on 5 November 2019).
  9. Center for Biomedical Image Computing & Analytics (CBICA). Available online: http://braintumorsegmentation.org/ (accessed on 5 November 2019).
  10. Mlynarski, P.; Delingette, H.; Criminisi, A.; Ayache, N. Deep learning with mixed supervision for brain tumor segmentation. J. Med Imaging 2019, 6, 034002. [Google Scholar] [CrossRef]
  11. Amin, J.; Sharif, M.; Yasmin, M.; Fernandes, S.L. Big data analysis for brain tumor detection: Deep convolutional neural networks. Futur. Gener. Comput. Syst. 2018, 87, 290–297. [Google Scholar] [CrossRef]
  12. Amin, J.; Sharif, M.; Raza, M.; Yasmin, M. Detection of Brain Tumor based on Features Fusion and Machine Learning. J. Ambient. Intell. Humaniz. Comput. 2018, 1–17. [Google Scholar] [CrossRef]
  13. Usman, K.; Rajpoot, K. Brain tumor classification from multi-modality MRI using wavelets and machine learning. Pattern Anal. Appl. 2017, 20, 871–881. [Google Scholar] [CrossRef] [Green Version]
  14. Pereira, S.; Meier, R.; Alves, V.; Reyes, M.; Silva, C. Automatic Brain Tumor Grading from MRI Data Using Convolutional Neural Networks and Quality Assessment. In Understanding and Interpreting Machine Learning in Medical Image Computing Applications; Springer: Berlin/Heidelberg, Germany, 2018; pp. 106–114. [Google Scholar]
  15. Farhi, L.; Zia, R.; Ali, Z.A. 5 Performance Analysis of Machine Learning Classifiers for Brain Tumor MR Images. Sir Syed Res. J. Eng. Technol. 2018, 1, 6. [Google Scholar] [CrossRef]
  16. Vijh, S.; Sharma, S.; Gaurav, P. Brain Tumor Segmentation Using OTSU Embedded Adaptive Particle Swarm Optimization Method and Convolutional Neural Network. Emerg. Trends Comput. Expert Technol. 2019, 13, 171–194. [Google Scholar]
  17. Mohsen, H.; El-Dahshan, E.-S.A.; El-Horbaty, E.-S.M.; Salem, A.-B.M. Classification using deep learning neural networks for brain tumors. Futur. Comput. Inform. J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  18. Veeraraghavan, A.K.; Roy-Chowdhury, A.; Chellappa, R. Matching shape sequences in video with applications in human movement analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1896–1909. [Google Scholar] [CrossRef] [PubMed]
  19. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Akkus, Z.; Galimzianova, A.; Hoogi, A.; Rubin, D.L.; Erickson, B.J. Deep Learning for Brain MRI Segmentation: State of the Art and Future Directions. J. Digit. Imaging 2017, 30, 449–459. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  22. Cheng, J. Brain Tumor Dataset. 2017. Available online: https://doi.org/10.6084/m9.figshare.1512427.v5 (accessed on 10 September 2019).
  23. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [Google Scholar] [CrossRef]
  24. Wong, T.-T. Performance evaluation of classification algorithms by k-fold and leave-one-out cross validation. Pattern Recognit. 2015, 48, 2839–2846. [Google Scholar] [CrossRef]
  25. Saeb, S.; Lonini, L.; Jayaraman, A.; Mohr, D.C.; Kording, K.P. The need to approximate the use-case in clinical machine learning. GigaScience 2017, 6, 1–9. [Google Scholar] [CrossRef] [Green Version]
  26. Little, M.A.; Varoquaux, G.; Saeb, S.; Lonini, L.; Jayaraman, A.; Mohr, D.C.; Kording, K.P. Using and understanding cross-validation strategies. Perspectives on Saeb et al. GigaScience 2017, 6, 1–6. [Google Scholar] [CrossRef]
  27. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, Chia Laguna Resort, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  28. Phaye, S.S.R.; Sikka, A.; Dhall, A.; Bathula, D. Dense and Diverse Capsule Networks: Making the Capsules Learn Better. arXiv 2018, arXiv:1805.04001. [Google Scholar]
  29. Pashaei, A.; Ghatee, M.; Sajedi, H. Convolution neural network joint with mixture of extreme learning machines for feature extraction and classification of accident images. J. Real Time Image Process. 2019, 1–16. [Google Scholar] [CrossRef]
  30. Gumaei, A.; Hassan, M.M.; Hassan, R.; Alelaiwi, A.; Fortino, G. A Hybrid Feature Extraction Method With Regularized Extreme Learning Machine for Brain Tumor Classification. IEEE Access 2019, 7, 36266–36273. [Google Scholar] [CrossRef]
  31. Pashaei, A.; Sajedi, H.; Jazayeri, N. Brain Tumor Classification via Convolutional Neural Network and Extreme Learning Machines. In Proceedings of the 2018 8th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 25–26 October 2018; pp. 314–319. [Google Scholar]
  32. Afshar, P.; Mohammadi, A.; Plataniotis, K.N. Brain Tumor Type Classification via Capsule Networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3129–3133. [Google Scholar]
  33. Kurup, R.V.; Sowmya, V.; Soman, K.P. Effect of Data Pre-processing on Brain Tumor Classification Using Capsulenet. In International Conference on Intelligent Computing and Communication Technologies; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2019; pp. 110–119. [Google Scholar]
  34. Srinivasan, K.; Nandhitha, N.M. Development of Deep Learning algorithms for Brain Tumor classification using GLCM and Wavelet Packets. Caribb. J. Sci. 2019, 53, 1222–1228. [Google Scholar]
  35. Sultan, H.H.; Salem, N.; Al-Atabany, W. Multi-Classification of Brain Tumor Images Using Deep Neural Network. IEEE Access 2019, 7, 69215–69225. [Google Scholar] [CrossRef]
  36. Kutlu, H.; Avcı; Avci, E. A Novel Method for Classifying Liver and Brain Tumors Using Convolutional Neural Networks, Discrete Wavelet Transform and Long Short-Term Memory Networks. Sensors 2019, 19, 1992. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Swati, Z.N.K.; Zhao, Q.; Kabir, M.; Ali, F.; Ali, Z.; Ahmed, S.; Lu, J. Brain tumor classification for MR images using transfer learning and fine-tuning. Comput. Med. Imaging Graph. 2019, 75, 34–46. [Google Scholar] [CrossRef]
  38. Shaikh, M.; Kollerathu, V.A.; Krishnamurthi, G. Recurrent Attention Mechanism Networks for Enhanced Classification of Biomedical Images. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1260–1264. [Google Scholar]
  39. Fernando, T.; Denman, S.; Aristizabal, D.E.A.; Sridharan, S.; Laurens, K.R.; Johnston, P.; Fookes, C. Neural Memory Plasticity for Anomaly Detection. arXiv 2019, arXiv:1910.05448. [Google Scholar]
  40. Rehman, A.; Naz, S.; Razzak, M.I.; Akram, F.; Razzak, I. A Deep Learning-Based Framework for Automatic Brain Tumors Classification Using Transfer Learning. Circuits Syst. Signal Process. 2019, 39, 757–775. [Google Scholar] [CrossRef]
  41. Tripathi, P.C.; Bag, S. Non-invasively Grading of Brain Tumor Through Noise Robust Textural and Intensity Based Features. In Computational Intelligence in Pattern Recognition; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2019; pp. 531–539. [Google Scholar]
  42. Ismael, M.; Abdel-Qader, I. Brain Tumor Classification via Statistical Features and Back-Propagation Neural Network. In Proceedings of the 2018 IEEE International Conference on Electro/Information Technology (EIT), Rochester, MI, USA, 3–5 May 2018; pp. 0252–0257. [Google Scholar]
  43. Kotia, J.; Kotwal, A.; Bharti, R. Risk Susceptibility of Brain Tumor Classification to Adversarial Attacks. In Proceedings of the Advances in Intelligent Systems and Computing, Krakow, Poland, 2–4 October 2019; pp. 181–187. [Google Scholar]
  44. Anaraki, A.K.; Ayati, M.; Kazemi, F. Magnetic resonance imaging-based brain tumor grades classification and grading via convolutional neural networks and genetic algorithms. Biocybern. Biomed. Eng. 2019, 39, 63–74. [Google Scholar] [CrossRef]
  45. Abiwinanda, N.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain Tumor Classification Using Convolutional Neural Network. In World Congress on Medical Physics and Biomedical Engineering 2018 Prague; Springer: Singapore, 2018; pp. 183–189. [Google Scholar]
  46. Kaldera, H.N.T.K.; Gunasekara, S.R.; Dissanayake, M.B. Brain tumor Classification and Segmentation using Faster R-CNN. In Proceedings of the 2019 Advances in Science and Engineering Technology International Conferences (ASET), Dubai, UAE, 26 March–10 April 2019; pp. 1–6. [Google Scholar]
  47. Kharrat, A.; Neji, M. Feature selection based on hybrid optimization for magnetic resonance imaging brain tumor classification and segmentation. Appl. Med. Inform. 2019, 41, 9–23. [Google Scholar]
  48. Portable Ultrasound Enables Anytime, Anywhere Imaging. Available online: https://healthtechmagazine.net/article/2018/07/portable-ultrasound-enables-anytime-anywhere-imaging (accessed on 3 February 2020).
  49. Moccia, S.; Foti, S.; Routray, A.; Prudente, F.; Perin, A.; Sekula, R.F.; Mattos, L.S.; Balzer, J.R.; Fellows-Mayle, W.; De Momi, E.; et al. Toward Improving Safety in Neurosurgery with an Active Handheld Instrument. Ann. Biomed. Eng. 2018, 46, 1450–1464. [Google Scholar] [CrossRef]
  50. Zaffino, P.; Pernelle, G.; Mastmeyer, A.; Mehrtash, A.; Zhang, H.; Kikinis, R.; Kapur, T.; Spadea, M.F. Fully automatic catheter segmentation in MRI with 3D convolutional neural networks: Application to MRI-guided gynecologic brachytherapy. Phys. Med. Boil. 2019, 64, 165008. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Representation of normalized magnetic resonance imaging (MRI) images showing different types of tumors in different planes. In the images, the tumor is marked with a red outline. The example is given for each tumor type in each of the planes.
Figure 1. Representation of normalized magnetic resonance imaging (MRI) images showing different types of tumors in different planes. In the images, the tumor is marked with a red outline. The example is given for each tumor type in each of the planes.
Applsci 10 01999 g001
Figure 2. Schematic representation of convolutional neural network (CNN) architecture containing the input layer, two Blocks A, two Blocks B, classification block and output. Block A and Block B differ only in the convolution layer. Convolution layer in Block A gives an output two times smaller than the input, whereas the convolutional layer in Block B gives the same size output as input.
Figure 2. Schematic representation of convolutional neural network (CNN) architecture containing the input layer, two Blocks A, two Blocks B, classification block and output. Block A and Block B differ only in the convolution layer. Convolution layer in Block A gives an output two times smaller than the input, whereas the convolutional layer in Block B gives the same size output as input.
Applsci 10 01999 g002
Figure 3. Confusion matrices for the original dataset with record-wise 10-fold cross-validation for testing data.
Figure 3. Confusion matrices for the original dataset with record-wise 10-fold cross-validation for testing data.
Applsci 10 01999 g003
Figure 4. Example of classified images from the original dataset with record-wise 10-fold cross-validation. The tumors are marked with a red outline.
Figure 4. Example of classified images from the original dataset with record-wise 10-fold cross-validation. The tumors are marked with a red outline.
Applsci 10 01999 g004
Figure 5. Confusion matrices for the augmented dataset with record-wise 10-fold cross-validation for testing data.
Figure 5. Confusion matrices for the augmented dataset with record-wise 10-fold cross-validation for testing data.
Applsci 10 01999 g005
Figure 6. Confusion matrices for the original dataset with subject-wise 10-fold cross-validation for testing data.
Figure 6. Confusion matrices for the original dataset with subject-wise 10-fold cross-validation for testing data.
Applsci 10 01999 g006
Figure 7. Confusion matrices for the augmented dataset with subject-wise 10-fold cross-validation during testing.
Figure 7. Confusion matrices for the augmented dataset with subject-wise 10-fold cross-validation during testing.
Applsci 10 01999 g007
Table 1. New CNN architecture. All network layers are listed with their properties.
Table 1. New CNN architecture. All network layers are listed with their properties.
Layer No.Layer NameLayer Properties
1Image Input256 × 256 × 1 images
2Convolutional16 5 × 5 × 1 convolutions with stride [2 2] and padding ‘same’
3Rectified Linear Unit Rectified Linear Unit
4Dropout50% dropout
5Max Pooling2 × 2 max pooling with stride [2 2] and padding [0 0 0 0]
6Convolutional32 3 × 3 × 16 convolutions with stride [2 2] and padding ‘same’
7Rectified Linear UnitRectified Linear Unit
8Dropout50% dropout
9Max Pooling2 × 2 max pooling with stride [2 2] and padding [0 0 0 0]
10Convolutional64 3 × 3 × 32 convolutions with stride [1 1] and padding ‘same’
11Rectified Linear UnitRectified Linear Unit
12Dropout50% dropout
13Max Pooling2 × 2 max pooling with stride [2 2] and padding [0 0 0 0]
14Convolutional128 3 × 3 × 64 convolutions with stride [1 1] and padding ‘same’
15Rectified Linear UnitRectified Linear Unit
16Dropout50% dropout
17Max Pooling2 × 2 max pooling with stride [2 2] and padding [0 0 0 0]
18Fully Connected1024 hidden neurons in fully connected (FC) layer
19Rectified Linear UnitRectified Linear Unit
20Fully Connected3 hidden neurons in fully connected layer
21Softmaxsoftmax
22Classification Output3 output classes, “1” for meningioma, “2” for glioma, and “3” for a pituitary tumor
Table 2. Results from testing the network with 10-fold cross-validation and one test, with two different cross-validation methods and different datasets.
Table 2. Results from testing the network with 10-fold cross-validation and one test, with two different cross-validation methods and different datasets.
Division/DatasetTesting ApproachTest Accuracy [%]Average Precision [%]Average Recall [%]Average F1-Score [%]
record-wise/the original dataset10-fold95.4094.8195.0794.93
One test97.3995.4496.9496.11
record-wise/the augmented dataset10-fold96.5695.7996.5196.11
One test97.2897.1597.8297.47
subject-wise/the original dataset10-fold84.4581.4082.7281.86
One test90.3985.9985.8485.91
subject-wise/the augmented dataset10-fold88.4886.4887.8286.97
One test91.8483.9482.1881.78
Table 3. Comparison of results of different network architectures, trained and tested on the original dataset, which use whole images as input and are tested using the k-fold cross-validation method.
Table 3. Comparison of results of different network architectures, trained and tested on the original dataset, which use whole images as input and are tested using the k-fold cross-validation method.
Referencek-Fold Cross-Validation Method/Data DivisionAccuracy [%]Average Precision [%]Average Recall [%]Average F1-Score [%]
Phaye et al. [28]8-fold; data division not stated95.03XXX
Pashaei et al. [29]5-fold; 80% data in training set, 20% in test93.68XXX
Gumaei et al. [30]5-fold; 80% data in training set, 20% in test92.61XXX
Pashaei et al. [31]10-fold; 70% data in training set, 30% in test.93.6894.6091.4393.00
Proposed10-fold; 60% data in training set, 20% in validation, 20% in test.95.40 194.81 195.07 194.94 1
1 The best obtained result.
Table 4. Comparison of results of different network architectures, trained, and tested on the original dataset, which use whole images as input and are not tested using the k-fold cross-validation method.
Table 4. Comparison of results of different network architectures, trained, and tested on the original dataset, which use whole images as input and are not tested using the k-fold cross-validation method.
ReferenceData DivisionAccuracy [%]Average Precision [%]Average Recall [%]Average F1-Score [%]
Afshar et al. [32]data division not stated86.56XXX
Vimal Kurup et al. [33]80% data in training set, 20% in test92.6092.6794.6793.33
Srinivasan et al. [34]75% data in training set, 25% in test93.30X91.0072.00
Proposed60% data in training set, 20% in validation, 20% in test97.39 195.44 196.94 196.11
1 The best obtained result.
Table 5. Comparison of results of different network architectures, trained and tested on the augmented dataset, which use whole images as input and are not tested using the k-fold cross-validation method.
Table 5. Comparison of results of different network architectures, trained and tested on the augmented dataset, which use whole images as input and are not tested using the k-fold cross-validation method.
ReferenceData DivisionAccuracy [%]Average Precision [%]Average Recall [%]Average F1-Score [%]
Sultan et al. [35]68% data in training set, 32% in validation and test96.1396.0694.43X
Proposed60% data in training set, 20% in validation, 20% in test97.28 197.15 197.82 197.47 1
1 The best obtained result.

Share and Cite

MDPI and ACS Style

Badža, M.M.; Barjaktarović, M.Č. Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network. Appl. Sci. 2020, 10, 1999. https://doi.org/10.3390/app10061999

AMA Style

Badža MM, Barjaktarović MČ. Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network. Applied Sciences. 2020; 10(6):1999. https://doi.org/10.3390/app10061999

Chicago/Turabian Style

Badža, Milica M., and Marko Č. Barjaktarović. 2020. "Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network" Applied Sciences 10, no. 6: 1999. https://doi.org/10.3390/app10061999

APA Style

Badža, M. M., & Barjaktarović, M. Č. (2020). Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network. Applied Sciences, 10(6), 1999. https://doi.org/10.3390/app10061999

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop