Next Article in Journal
Applications of Converged Various Forces for Detection of Biomolecules and Novelty of Dielectrophoretic Force in the Applications
Previous Article in Journal
Correcting Decalibration of Stereo Cameras in Self-Driving Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

MRI Segmentation and Classification of Human Brain Using Deep Learning for Diagnosis of Alzheimer’s Disease: A Survey

1
Department of Information and Communications Engineering, Chosun University, Gwangju 61452, Korea
2
Division of Computer & Electronic Systems Engineering, Hankuk University of Foreign Studies, Yongin 17035, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(11), 3243; https://doi.org/10.3390/s20113243
Submission received: 16 April 2020 / Revised: 25 May 2020 / Accepted: 3 June 2020 / Published: 7 June 2020
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Many neurological diseases and delineating pathological regions have been analyzed, and the anatomical structure of the brain researched with the aid of magnetic resonance imaging (MRI). It is important to identify patients with Alzheimer’s disease (AD) early so that preventative measures can be taken. A detailed analysis of the tissue structures from segmented MRI leads to a more accurate classification of specific brain disorders. Several segmentation methods to diagnose AD have been proposed with varying complexity. Segmentation of the brain structure and classification of AD using deep learning approaches has gained attention as it can provide effective results over a large set of data. Hence, deep learning methods are now preferred over state-of-the-art machine learning methods. We aim to provide an outline of current deep learning-based segmentation approaches for the quantitative analysis of brain MRI for the diagnosis of AD. Here, we report how convolutional neural network architectures are used to analyze the anatomical brain structure and diagnose AD, discuss how brain MRI segmentation improves AD classification, describe the state-of-the-art approaches, and summarize their results using publicly available datasets. Finally, we provide insight into current issues and discuss possible future research directions in building a computer-aided diagnostic system for AD.

1. Introduction

Magnetic resonance imaging (MRI) is used to analyze the anatomical structures of the brain due to its high spatial resolution and ability to contrast soft tissue. It is known that MRI is generally associated with fewer health risks compared to other modalities like computed tomography (CT) and positron emission tomography (PET) [1]. Over the past few decades, tremendous progress has been made in assessing brain injuries and exploring brain anatomy with MRI [2]. Disorders such as Alzheimer’s disease (AD) and multiple sclerosis [3] associated with the brain can be identified using MRI. Tissue atrophy is a popular indicator that is used in diagnosing AD. The segmentation of brain MRI taken at different times is also used to measure structural changes in the brain. Accurate detection and classification of unhealthy tissue and its surrounding healthy structures are also important in the diagnosis of conditions such as AD. A large amount of data is required for more accurate diagnoses. However, it can be challenging for clinicians to analyze large and complex MRI datasets and to extract important information manually. Moreover, due to various inter- or intra-operator variability issues [4], manual analysis of brain MRI is time-consuming and vulnerable to errors. Hence, it is necessary to develop an automated segmentation method to provide accurate results with high confidence. Computerized techniques for MRI segmentation, visualization, and registration have recently been used on large scale datasets to assist clinicians in making qualitative diagnoses.
A number of clinical applications include brain MRI segmentation because it impacts the results of the entire analysis process. A number of classical machine learning-based approaches have been developed for the segmentation of brain tissue types such as gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF). Abnormal tissues of the brain in patients with AD can be segmented in MRI [5]. However, the extraction of the imaging features for such segmentation requires elaborate engineering techniques and specialized expertise.
The main objective of brain MRI segmentation is to divide the image into well-defined regions, where each region consists of a set of pixels that share the same range of intensities, texture, or neighborhood. Our review discusses the segmentation of these regions from brain images using deep learning techniques. The segmentation of GM, WM, and CSF from brain MRI is challenging due to their tissue intensities, non-uniformity (bias), noise artifacts, and partial volume effect. To overcome these difficulties, several deep learning techniques for brain MRI segmentation have been developed and will be reviewed. We also review various deep learning techniques for the early diagnosis of AD, which is a type of dementia that can cause thinking, memory, and behavioral issues. For the MRI classification of AD, significant patterns from raw data are considered, and these patterns are grouped into different categories based on their characteristics. Significant advances in imaging technology have led to the development of different applications of image segmentation and classification. Brain segmentation plays an important role and can be a building block for AD diagnosis. Semantic segmentation techniques with Freesurfer [6] can predict the volume of the brain from MRI scans of patients with AD [7]. Unsupervised hierarchical segmentation methods for AD diagnosis can detect homogeneous regions and separate them from coarse to finer levels, providing more flexibility for multi-level analysis than single-level semantic segmentation [8,9]. Furthermore, precuneus atrophy and the hippocampus are the most sensitive biological indicators of AD, particularly at an early stage [10]. The precuneus, an area of the posteromedial cortex, has recently received significant attention in functional neuroimaging studies. Precuneus atrophy observed in the AD group suggests that MRI volumetric assessment of precuneus volume, in addition to the hippocampal volume, might be a useful radiological index for the diagnosis of AD.
Figure 1 shows the boundaries of the left precuneus in the sagittal plane, and the entire anterior-posterior extent of the hippocampus (including its head, body, and tail) outlined on coronal MR images. These anatomical boundaries are used for volumetric measurements in the precuneus and hippocampus.
Traditional machine learning approaches have relatively lower performance with larger amounts of input data. It can be challenging to detect brain abnormalities correctly and to find a solution for the automatic segmentation of brain structures. Such challenges mainly arise from the changes in settings for the acquisition of MRI scans, fluctuations in the appearance of pathology, normal anatomical variations in brain morphology, and imperfections in image acquisition. The limitations of traditional machine learning methods can be overcome by deep learning-based approaches. Moreover, deep learning can also be used to perform quantitative analysis of brain MRI through the self-learning of features, by which new features can be recognized. Deep learning has been acquiring substantial attention in various medical image analyses [11], such as computer-aided diagnosis of breast lesions [12], pulmonary nodules [13], and histopathological diagnosis [14].
This paper aims to provide an outline of progressive deep learning methods in the area of MRI segmentation of normal and abnormal (AD) brain tissue. Furthermore, we analyze existing problems in the segmentation of brain MRI that can be overcome with deep learning. In summary, the main objectives of this review are to:
  • Provide an overview of the current deep learning approaches for brain MRI segmentation and classification of AD.
  • Identify the application challenges in the segmentation of brain structure MRI and classification of AD.
  • Show that MRI segmentation of the brain structure can improve the accuracy of diagnosing AD.
The rest of the paper is organized as follows. A brief overview of publicly available brain MRI datasets, followed by a brain MRI analysis, is presented in Section 2. An overview of convolutional neural networks (CNN) architecture, segmentation of brain structure MRI using deep learning, and how segmentation improves the classification of AD are described in Section 3. The evaluation measures for brain MRI segmentation is presented in Section 4. Finally, we conclude with a general discussion and explore future directions in the field of brain MRI segmentation.

2. MRI Dataset for Brain Analysis

The data evaluation framework of three-dimension (3D) cross-sectional brain MRI is used to classify patients with AD and to segment brain tissue types (CSF, GM, and WM). Publicly available datasets such as open access series of imaging studies (OASIS) [15], Alzheimer’s disease neuroimaging initiative (ADNI) [16], medical image computing and computer-assisted intervention (MICCAI) [17], and internet brain segmentation repository (IBSR) [18] are popularly used for segmentation of brain MRI and AD diagnosis. Table 1 shows the details of the OASIS, ADNI, IBSR, and MICCAI datasets. The details of these datasets are described below and then followed by an analysis of brain MRI at various stages.

2.1. Public Dataset for Brain MRI

2.1.1. OASIS

The OASIS dataset [15] was created by Washington University, where the Alzheimer’s Disease Research Centre manages a large amount of longitudinal and cross-sectional brain MRI data from non-demented and demented subjects. The longitudinal dataset contains multiple scans of each subject over a period of time, and the cross-sectional category includes the details of 416 subjects aged between 18 and 96 years. The risk factor of AD can be measured with the clinical dementia rating (CDR) [19] and mini-mental state examination (MMSE) [20]. The subjects are examined with risk factors as non-demented for CDR 0, very mild dementia for CDR 0.5, mild dementia for CDR 1, and moderate dementia for CDT 2, as indicated in [15].

2.1.2. ADNI

AD usually observed in elderly individuals can be diagnosed using the ADNI dataset [16], which contains details of MRI scans for 843 subjects with scanner intensity fields ranging from 1.5 T to 3 T. It is observed that patients with lower cognitive abilities including thinking and memory loss are associated with mild cognitive impairment (MCI). They also have a high risk of transforming to AD or any other types of dementia and are grouped separately from AD.

2.1.3. IBSR

The IBSR dataset [18] is used to evaluate and develop segmentation techniques for brain images. This dataset provides manually guided expert segmentation results along with the MRI data. It consists of 20 real T1-Weighted (T1-W) MRI with manually guided expert segmentation results, referred to as the ground truth. In addition, each MRI volume contains around 60 coronal T1-W slices with a 3.1 mm resolution (slice gap between successive slices) and 18 cortical T1-W slices with a resolution of 1.5 mm. The subject volumes of this dataset have dimensions of 256 × 256 × 128 pixels and different voxel spacings: 0.84 × 0.84 × 1.5 mm3 0.94 × 0.94 × 1.5 mm3, and 1.0 × 1.0 × 1.5 mm3. In addition, Massachusetts General Hospital has provided manual segmentation of 32 noncortical structures.

2.1.4. MICCAI

The MICCAI-2012 dataset [17] consists of 35 T1-w MRI volumes and manual segmentation of 134 structures obtained from Neuromorphometrics, Inc., Scotts Valley, CA, USA. It is mainly used for the segmentation of tissue, tumor, and structure. This dataset started with 80 real and synthetic cases in 2012. The size of the training and testing data has increased over the years. The MICCAI 2012 challenge in multi-atlas labeling is used for sub-cortical structure segmentation.

2.2. Pre-Processing for Brain MRI Analysis

Figure 2 illustrates a typical pipeline for the segmentation stages of brain MRI analysis, conventionally proposed in the literature from a top-level perspective. The overall pipeline consists of four main stages, including pre-processing, data-preparation, segmentation, and post-processing. Different pre-processing tasks are required after acquiring MRI so that the images can be used for the segmentation of various tissue types of the brain. Figure 3 shows examples of brain MRI in pre-processing. The pre-processing for MRI includes brain extraction, bias field correction, and image registration.
Brain extraction: Brain MRI shows tissue as well as parts of the head, eye, fat, spinal cord, and skull [21]. Skull stripping to extract brain tissue from the non-brain tissue needs to be performed to identify the voxels as brain or non-brain. The output of skull stripping can be a new image only with brain voxels or a binary value assigning value 1 for brain voxels and value 0 for the rest of the tissue. In general, brain voxels include the brain stem, CSF, GM, and WM of the cerebral cortex, subcortical structures, and cerebellum, whereas non-brain voxels include the scalp, matter, eyes, bones, dura, skin, muscles, and fat [22]. An example of an original T1-W brain MRI and corresponding skull stripped output is shown in Figure 3a,b, respectively.
Bias field correction: Image contrast due to magnetic field inhomogeneity [23] can be fine-tuned with the help of bias field correction. The bias field depends on the strength of the magnetic field and is almost negligible when the MRI is performed at 0.5 T. However, when the magnetic field strength is 1.5 T, 3 T, or larger, it is considered as strong and can impact the analysis of the MRI. Figure 3c shows the bias field, and Figure 3b,d shows the MRI before and after bias field correction, respectively.
Noise reduction: Noise reduction is the process of lowering locally variant Rician noise noticed in MRI [23]. This is considered to be less critical for classification applications using deep learning [24].
Image registration: Image registration is mainly used to convert the alignment of the images from spatial to common anatomical spaces [25] and has two types, inter- and intra-patient image registration. The inter-patient image registration is used to standardize MRI onto standard stereotaxic spaces, whereas intra-patient registration aids in aligning MRI of different sequences (T1- and T2-W images) to obtain multi-channel characterization for each position within the brain. After pre-processing, data preparation is performed with data augmentation or patch-based strategies from the input volumes. Then, segmentation or classification according to the objective of the analysis is performed based on input modalities and patch dimensions. Finally, the results obtained could be refined by choosing the largest groups only or smoothing regions.

3. Review of Brain MRI Segmentation and Diagnosis

In this section, we provide a comprehensive literature review on the segmentation of the brain structure and classification of brain MRI for diagnosing AD. Moreover, we briefly discuss CNN architecture, followed by the segmentation of brain structures using deep learning techniques. Then, the classification of AD using deep learning is presented. Finally, we discuss how the segmentation of brain MRI improves the classification accuracy of MRI for AD.

3.1. Overview of CNN Architecture

Deep learning refers to neural networks with a deep number of layers (usually more than five) that extract a hierarchy of features from raw input images. Traditional machine learning algorithms [26,27,28,29,30] extract features manually, whereas deep learning extracts complex, high-level features from the images and trains a large amount of data, thus resulting in greater accuracy. Owing to significantly increased GPU processing power, deep learning methods allow us to train a vast amount of imaging data and increase accuracy despite variations in images.
Various applications such as image segmentation, genotype/phenotype detection, classification of diseases, object detection, and speech recognition utilize different deep learning approaches. Some of the popular deep learning algorithms include deep Boltzmann machines, CNNs, stacked auto-encoders, and deep neural networks. CNNs are prominently used for image segmentation and classification. Even though CNNs were first introduced in 1989 [31], they received more attention after observing their excellent performance in the ImageNet [32] Competition in 2012 [33]. It is reported that, by applying CNN on a dataset with millions of images with 1000 various classes, the error rate can be reduced to half, compared to earlier best computing approaches [34]. CNN architecture is increasingly complex, with a large number of layers, including neurons with millions of weights and a large number of connections between different neurons, indicating that the computational complexity of CNN architecture is high.
Figure 4 shows the basic block diagram of CNN, which consists of layers of convolution, pooling, activation function, and fully connected layers with each layer performing specific functions. Input images are convolved across the kernel by the convolutional layer to produce feature maps. In the pooling layer, as the value transferred to the successive layer, the results obtained from preceding convolutional layers are downsampled using the maximum or average of the specified neighborhood. The most popular activation functions are the rectified linear unit (ReLU) [14] and the leaky ReLU [14], which is a modification of ReLU. The ReLU transforms data nonlinearly by clipping off negative input values to zero and passing positive input values as output. The results of the last CNN layer are coupled to loss function (e.g., scores are normalized into a multinomial distribution over labels by cross-entropy loss) to provide a forecast of the input data. Finally, network parameters are obtained by decreasing the loss function between prediction and ground truth labels along with regularization constraints. In addition, weights of the network are updated at each iteration (e.g., using stochastic gradient descent) using backpropagation until the convergence.
Table 2 shows various segmentation strategies using single-modality, multi-modality, semantic-wise, patch-wise and cascaded using the CNN architecture. Single-modality refers to a single source of information and is adaptable to different scenarios. In contrast, multi-modality utilizes multi-source information and provides exact localization, highlighting any pathognomonic changes and metabolic activity of the target tissue in the case of positron emission tomography (PET). Semantic-wise approaches link each pixel of an image with its class label. Segmentation labels are mapped with the input image so that it minimizes loss function. This allows segmentation maps to be generated for any image size. The computational complexity of this method is much lower than other approaches [34]. This principle in segmentation is used for most of the state-of-the-art approaches [35,36,37], as will be described in Section 3.2. The patch-wise approach takes small patches from high-resolution images. That is, the input images are split as a number of local patches and trained. This can result in better local information by predicting the information for an individual patch. Moreover, the model can be trained with local details in patch-wise approaches but does require higher computational complexity. Cascaded CNN is characterized by the first network, which provides the first classification, and the second network receiving the outputs of the first network (as input) to refine the classification. Cascaded CNN achieves competitive results in comparison with other CNN approaches. Despite the success in segmentation performance using deep learning, there are also several problems and limitations.
Challenges in brain MRI analysis for segmentation and classification using deep learning:
  • Deep learning used in big data analytics: The major challenge lies in the difficulty of obtaining a large enough dataset to train and improve the accuracy of the model properly. Deep learning faces difficulties in dealing with the volume (high-dimensional decision space, and a large number of objectives), variety (modeling using heterogeneous data and knowledge transfer between problems), variability (robustness over time and online knowledge acquisition) and veracity (noisy fitness evaluations and surrogate-assisted optimization) of big data [50]. To overcome this problem, the author in [51] suggested various optimization techniques such as global optimization, which reuse the knowledge extracted from the vast amount of high dimensional, heterogeneous, and noisy data. On the other hand, complex optimization techniques provide efficient solutions by formulating new insights and methodologies for optimization problems that take advantage of using deep learning approaches when dealing with big data problems. The traditional machine learning approaches show better performance with less input data. As the amount of data increases beyond a certain critical point, the performance of traditional machine learning approaches becomes steady, whereas deep learning approaches tend to increase [51]. Deep learning architectures, such as deep neural networks, deep belief networks, and recurrent neural networks, have been applied to research fields including medical image analysis, bioinformatics, and computer vision, where they often produce impressive results, that are comparable to and superior to human experts in some cases.
  • Scalability of deep learning approaches: The scalability of deep learning needs to consider not only accuracy but several other measures regarding computational resources. Scalability plays a vital role in deep learning. As data expands in terms of variability, variety, veracity, and volume, it becomes increasingly difficult to scale computing performance using enterprise-class servers and storage in line with the increase. Scalability can be achieved by implementing deep learning techniques on a high-performance computing (HPC) system (super-computing, cluster, sometimes considered cloud computing), which offers immense potential for data-intensive business computing [50]. The ability to generate data, which is important where data is not available for learning the system (especially for computer vision tasks, such as inverse graphics).
  • Multi-task, transfer learning, or multi-module learning: Learning simultaneously from several domains or with various models is one of the significant challenges in deep learning. Currently, one of the most significant limitations to transfer learning is the problem of negative transfer. Transfer learning only works if the initial and target problems are similar enough for the first round of training to be relevant. If the first round of training is too different, the model may perform worse than if it had never been trained at all. There are no clear standards on what types of training are sufficiently related, or how this should be measured.

3.2. Segmentation of Brain MRI Using Deep Learning

To perform a quantitative analysis of the brain tissue, and large-scale study of intracranial volume, accurate automated segmentation of brain structures such as GM, WM, and CSF in MRI is crucial. The traditional approaches used for the segmentation of brain tissues include the Atlas-based approach and pattern recognition approach: Atlas-based approaches [52,53,54,55] match intensity information between an atlas and target images. Atlas-based and registration are among the methods which are widely used for human brain segmentation [56,57,58] but do not provide robust results for small and highly variable structures like the hippocampus, due to limitations in registration and variability in reliable ground truth data. In pattern recognition approaches [37,59,60], brain tissues are classified based on the set of local intensity features. Recently, hippocampal atrophy has been proposed as a biomarker of AD [61,62]. The hippocampus is a part of the brain’s limbic system surrounded by different kinds of tissue. A number of studies have shown that a lower hippocampal volume is observed in patients with AD [63,64]. Hence, MRI segmentation of the hippocampus could have practical value in clinical applications [65]. However, segmentation for the hippocampus in MRI is challenging due to its small size, partial volume effects, anatomical variability, low contrast, low signal-to-noise ratio, indistinct boundary, and proximity to the Amygdaloid body. Furthermore, manual segmentation requires time-consuming expert analysis. A recent study shows that segmenting the hippocampus, thresholding or region growing using conventional methods do not achieve acceptable results [66]. Wang et al. [67] proposed a region growing algorithm based on seed, which is simple and effective, yet fails to obtain promising results because of the unclear edges of the hippocampus [68].
Table 3 shows the list of studies on the segmentation of brain tissues based on CNN and also the type of public dataset used in the analysis. In addition, the summary of segmentation strategies with their image dimension followed by classifiers used in the CNN architecture are described. Furthermore, an overview of existing methods based on deep learning for the segmentation of brain tissues and anatomical segmentation (e.g., the hippocampus) is summarized in Table 4.
Challenges in brain MRI segmentation:
  • Large variations in brain anatomical structures due to phenotypes, age, gender, and disease. It is difficult to apply one specific segmentation method to all phenotypic categories for reliable performance [77].
  • It is challenging to process cytoarchitectural variations, such as gyral folds, sulci depths, thin tissue structures, and smooth boundaries between different tissues. This can result in confusing categorical labeling for distinct tissue classes. This is difficult even for human experts.
  • The low contrast of anatomical structure in T1, T2, and FLAIR modalities results in low segmentation performances.
  • Manual segmentation for brain MRI is laborious, subjective, and time-consuming, and requires sophisticated knowledge of brain anatomy. Thus, it is difficult to obtain enough amount of ground truth data for building a segmentation model.
  • The noisy background in the ordinary image for segmentation is challenging because it is hard to assign an accurate label to each pixel/voxel with learned features.
  • The segmentation of the hippocampus, which is one of the most important biomarkers for AD, is difficult due to its small size and volume [65], as well as its anatomical variability, partial volume effects, low contrast, low signal-to-noise ratio, indistinct boundary and proximity to the Amygdaloid body.

3.3. Brain MRI Classification of AD Diagnosis Using Deep Learning

The segmentation of brain MRI is carried out to eliminate unnecessary details and to locate relevant objects from the processed images. The detailed analysis of the tissue structures from the segmented MRI leads to a more precise classification of specific brain disorders such as AD. AD is more common in elderly individuals, and it is considered to be a common form of dementia. Patients with AD suffer from the degradation of cognitive abilities over time. In advanced cases, patients struggle with activities of daily life, ultimately resulting in an inability to self-care. In this disease, nerve cells and tissues of the human brain are affected. Initially, the frontal lobe of the cerebral cortex, which helps in planning, thinking, and remembering, and the hippocampus, which plays a crucial role in the development of new memories, can be affected. Although vulnerability to AD increases in those over the age of 65 years, AD is not solely associated with old age [78]. A recent study [79] estimates that more than 90 million people will have AD by 2050. Despite considerable research to discover treatments for AD and halt or delay its progression, so far, there have not been promising results [80].
Table 5 summarizes the classification methods for the diagnosis of AD using CNN architectures on the public datasets (OASIS and ADNI). Furthermore, considerable research efforts have been made for the classification of AD. The applications and key features of the methods are described in Table 6.
Challenges in the diagnosis of AD:
  • The automatic classification of AD is quite challenging due to the low contrast of the anatomical structure in MRI. The presence of noisy or outlier pixels in MRI scans due to various scanning conditions may also result in a reduction of the classification accuracy.
  • The major challenge in AD is that it is difficult to make long-term tracking and investigation of the patient’s pathology. Thus, it is not easy to track the transition of AD status. In the ADNI dataset [16], there are only 152 transitions in total out of the entire dataset of 2731 MRIs. Due to the lack of the MRIs in terms of tracking the transition of AD status, it is likely for the model to overfit without generalizing distinctions between different stages of AD.
  • It is well known that AD is not only diagnosed from clinical stages of brain MRI, but also occurs through abnormal amyloid β peptide (Aβ) and tau (τ) protein activity around neurons and their temporal relationship with the different phases of AD in different stages. The factors mentioned above should be considered as multi-modal biomarkers as well as brain MRI. Thus, complexity during the process of treating AD is due to diverse factors regulating its pathology.
  • Data multimodality in the diagnosis of AD
    Since neuroimaging data (i.e., MRI or PET) and genetic data (single nucleotide polymorphism (SNP)) have different data distributions, different numbers of features and different levels of discriminative ability to AD diagnosis (e.g., SNP data in their raw form are less effective in AD diagnosis), simple concatenation of the features from multimodality data will result in an inaccurate prediction model [105,106] due to heterogeneity.
    High dimensionality issue: One neuroimage scan normally contains millions of voxels, while the genetic data of a subject has thousands of AD-related SNPs.

3.4. The Segmentation of Brain MRI Improves the Classification of AD

It is known that AD is the major cause of dementia for most Caucasians [107]. An important pathological characteristic of AD is diffuse brain atrophy, which includes atrophy of the cerebral cortex, enlargement of the ventricles, and atrophy of the medial temporal lobe (MTL) structures such as the hippocampus [108,109]. It is reported that AD pathology is associated with GM and WM tissues, and it was discovered that abnormalities of these tissues are highly correlated with cognitive decline [110,111,112]. To more clearly capture how AD evolves, neuropsychological and anatomical information from the patient needs to be examined at different transitional phases of the disease. From these aspects, the populations suffering from MCI have a high chance of converting disease status from MCI to AD. The results in [113,114] show that the risk among individuals converting from MCI to AD is significantly increased as compared to normal individuals. Among various types of MCIs, the amnesic form (aMCI) is commonly found in most individuals. As the highest annual incidence of conversion from aMCI to AD is observed [113,114], aMCI is considered to be the prodromal phase of AD. In previous studies [108,115,116,117], the atrophy of MTL structures, such as the hippocampus, amygdala, and entorhinal and parahippocampal cortices, increases with progression of the disease. Considerable research effort has focused on the segmentation of the hippocampus to evaluate the volume or shape of the brain [109,118,119,120,121,122,123,124,125,126,127,128,129,130]. Cortical thickness and GM tissue maps [131] have also been shown to have high predictive value in the diagnosis of AD [131]. In addition, studies on group differences based on voxel-based morphometry (VBM) [132], deformation-based morphometry (DBM) [133], or tensor-based morphometry (TBM) [134] were investigated. The VBM is a spatially specific and unbiased method of analysis of MRI, showing the regional GM and WM volume at a voxel scale [135]. This method has been applied to AD and MCI, reflecting patterns of GM irregularities that are suitable for the clinical phase of the disease, and also predict the risk of conversion from MCI to AD [136,137,138]. The presence and extent of WM atrophy are more problematic, and often other MRI methods are used to analyze the presence of microscopic tissue impair in AD [139]. Furthermore, the CSF and structural imaging markers are considered to be primary indicators amended to the current diagnostic criteria of MCI and AD [116]. The CSF markers (Ab42, t-tau, and p-tau) with the volumes of the hippocampus and lateral ventricles, can be combined for distinguishing between HC and MCI, while the CSF Ab42 marker with the shape of the hippocampus and lateral ventricles is a good combination for identifying MCI and AD [140]. Recently, deep learning approaches showed better performance for the automatic segmentation of the hippocampus and classification of AD. The deep CNN model is constructed to learn features of the 3D patches extracted based on the hippocampal segmentation results for the classification task [141]. A multi-task deep learning (MDL) method was proposed for joint hippocampal segmentation and clinical score regression using MRI scans [142]. Choi et al. [143] shown that deep CNN based biomarkers are strongly correlated with future cognitive decline using PET.
The overall block diagram of the diagnosis of AD usually adopted in conventional methods is depicted in Figure 5. The first stage is to acquire the slices from the MRI. This is followed by data preparation with pre-processing, where non-relevant information is removed, and the data are reoriented for more straightforward interpretation. Segmentation using deep learning is performed on the pre-processed data to extract the attributes from brain MRI. Based on attributes such as surface area, the center of gravity, average intensity, and standard deviation, the classifier predicts the output based on prior knowledge using deep learning architecture and decides whether the patient is developing AD or not.

4. Evaluation Metrics for Brain MRI Segmentation

In medical image analysis, validation and quantitative comparison are common problems of different segmentation methods. For validation, ground truth data is required for the comparison of segmented output. In a real scenario, there is not enough amount of ground truth data available for the assessment of acquired data in humans. Thus, the ground truth data of patients are manually generated by experts after the MRI acquisition. Although this is the only way to validate the real patient’s MRI data, this validation must be carefully considered because manual segmentation is prone to errors, highly subjective, and difficult to reproduce (even by the same expert) [144]. To overcome these limitations, several alternate validation methods with software simulation and phantoms have been proposed. Software simulation generates artificial images that simulate the process of real acquisition. Likewise, ground truth is known, and the various acquisition parameters, imaging artifacts can be managed and analyzed independently. This type of validation requires less effort and is more flexible to implement. However, the software simulation method does not consider all factors which might impact real image acquisition, and the images obtained for software simulation are only an approximate estimation of the real images. MRI scanners are used to produce phantom images that appear to be more realistic compared to the ones obtained from software simulations. However, phantom images are not flexible. Additionally, software simulations and phantom imaging results are expensive and time-consuming. To evaluate the overlap between the predicted brain MRI and the given ground truth image, various similarity measures were used [145].
The most well-known evaluation measure is the Dice coefficient [146]. The quality of two binary label masks can be compared with this volume. Let us define that D P is the mask by the human evaluator and D Q is the mask generated by a segmented algorithm, the Dice overlap is then evaluated as Equation (1).
D i c e   ( D P , D Q ) = 2 | D P D Q | | D P | + | D Q |
where | . | represents the number of voxels. The overlap measure has values in the range of [0, 1], with 0 indicating no match between the two masks and 1 indicating that the two masks are identical. The Jaccard index [147] is also used as a similarity measure for the comparison of two binary label masks, and it is expressed as Equation (2).
J a c c a r d   I n d e x   ( D P , D Q ) = | D P D Q | | D P | + | D Q | | D P D Q |
The ratio of true positives to the sum of true and false positives is known as the positive predicted value (PPV). It is also called as precision and expressed as Equation (3).
P P V ( D P , D Q ) = | D P D Q | | D P D Q | + | D P c D Q |
The ratio of true positives to the sum of true and false positives is known as the positive predicted value (PPV). It is also called as precision and expressed as Equation (3). The ratio of true positives to the sum of true positives and false negatives is known as the true positives rate and is calculated as Equation (4).
T P R ( D P , D Q ) = | D P D Q | | D P D Q | + | D P D Q c |
The ratio of true positives to the sum of true positives and false negatives is known as the lesion true positive rate (LTPR). Considering the list of lesions,   L P , as the 18-connected components of D P and similarly,   L Q , as the 18-connected components of D Q . It is expressed as Equation (5).
L T P R ( D P , D Q ) = |   L P L Q | |   L P L Q | + |   L P L Q c |
where |   L P L Q | refers to the overlap between a connected component of D P and D Q . It is shown that the human evaluator and segmentation algorithm have identical lesions. Furthermore, the lesion wise ratio of false positives to the sum of false positives and true negatives is known as the lesion false positive rate (LFPR), which can be expressed as Equation (6).
L F T R ( D P , D Q ) = | L P c L Q | | L P c L Q | + | L P c L Q c |
where L R c is the 18-connected components of D P c . The absolute difference in volume divided by the true volume is called the absolute volume difference (AVD) and is expressed as Equation (7).
A V D ( D P , D Q ) = Max   ( | D P | , | D Q | ) Min   ( | D P | , | D Q | ) | D P |
The average symmetric surface distance (ASSD) is the average of the distance (in millimeters) from the lesions in D P to the nearest lesion identified in D Q plus the distance from the lesions in D P to the nearest lesion identified in D Q . It is described as Equation (8).
A S S D   ( D P , D Q ) = p   L P d ( r ,   L Q ) + q   L Q d ( r ,   L P ) 2
where d ( r ,   L Q ) is the distance from the lesion r in   L P to the nearest lesion in L Q . The value of 0 refers to D P and D Q being identical. The detail about similarity measures for validation and evaluation in medical image analysis is found in [132]. Table 7 shows the summary of the validation measures of brain segmentation and their mathematical formulation concerning the number of true positives (TP), false positives (FP), and false-negative (FN) at voxel and lesion levels (TPL, FPL, and FNL).

5. Discussion and Future Directions

Two-dimensional (2D) networks are used in most of the works, which are summarized in Table 3. Table 8 displays information about the various architectures in relation to their performance. The best generalization can be achieved with the help of optimized layers of architecture and choosing the better hyper-parameters with advanced training procedures.
Table 8 shows the summary of the results in the reviewed paper using deep learning approaches for brain structure segmentation results reported in the recent investigations in DSC and JI values as evaluation metrics. Per the results in Table 8, Wachinger et al. [71] show better results in terms of DSC compared to the other methods using the MICCAI dataset because the author introduced explicit within-brain location information through Cartesian and spectral coordinates to help the classifier to discriminate one class from another. Furthermore, Dolz et al. [70] implemented a 3D fully convolution network architecture which was used to segment sub-cortical structures in the brain MRI and the result outperforms other methods in terms of DSC in IBSR datasets. However, it is noted that these results are not able to be accurately compared because these studies used different datasets and experimental conditions.
It is necessary to have a universal architecture to provide a good representation of the underlying input image without suffering from significant over-fitting to enable accurate comparison. In previous studies on AD-related brain disease diagnosis, MRI-based computer-aided diagnosis methods [148] usually contain three fundamental components, i.e., pre-determination of ROIs, extraction of imaging features, and construction of classification models. Depending on the scales of the pre-defined ROIs in MRI for subsequent feature extraction and classifier construction, these methods can be further divided into three categories, i.e., voxel-level, region-level, and patch-level morphological pattern analysis methods. Specifically, voxel-based methods [149,150] attempt to identify voxel-wise disease-associated microstructures for AD classification. In contrast, region-based methods [87,142] extract quantitative features from pre-segmented brain regions to construct classifiers for identifying patients from HCs. To capture brain changes in local regions for the early diagnosis of AD, patch-based methods [128,129] adopt an intermediate scale (between the voxel-level and region-level) of feature representations for MRI to construct classifiers. Table 9 shows the summary of the various state-of-the-art methods for the classification of AD vs. HC and pMCI (progressive MCI) vs. sMCI (stable MCI) on MRI datasets. In contrast to the studies that only used MRI datasets of ADNI-1, Lian et al. [95] evaluated the method on larger cohorts composed of 1,457 subjects from both ADNI-1 and ADNI-2. The authors used a more challenging evaluation protocol (i.e., independent training and testing sets) and obtained competitive classification performance, especially for pMCI vs. sMCI conversion prediction. The results in [85] outperform other methods in AD vs. HC classification because transfer learning can be used to enhance generality of the features capturing the AD biomarkers with a three stacked 3D convolutional autoencoder network, which is pre-trained on the ADNI datasets. The extracted features are used as AD biomarkers in lower layers of a 3D CNN network. Similar performances were observed in [109], where authors propose a combined use of AlexNet and the data permutation scheme, and it outperforms other methods in AD classification owing to spatial correlation features generated by using localized convolution kernels in CNN and more informative slice selection that trains AlexNet more effectively.
Evaluating the model on different datasets is one of the suggested practice methods. Some works [37,43,151] have validated the models on three or more different public datasets and reported the results. These kinds of practices can be used to make the model robust and can also be applied to a dataset that consists of data from different imaging modalities and MRI scanners [152]. It is exceedingly challenging to train deep CNNs with low-resolution MRI, and the need for a short prediction time makes it impossible to train deep CNNs [115]. Therefore, training such networks can be performed by designing a method with faster convolution. The computation speed of CNN was improved by using FFT algorithms and fast matrix multiplication methods [153,154], yet the training algorithms of deep CNNs can be improved using variants of SGD [155] and their parallelized implementations. It is expected to improve the performance of deep CNNs with the high optimization method using emerging new algorithms with less or no hyper-parameters, which constitute one of the major bottlenecks for most users.
One of the main contributors to the steep rise of deep learning has been the widespread availability of GPU. GPU has parallel computing engines that perform with execution threads larger than central processing units (CPUs). It is known that deep learning on GPUs is typically 10–30 times faster than on CPUs [156]. Another driving force behind the popularity of deep learning methods is the wide availability of open-source software packages. Deep learning libraries such as Caffe [157], PyTorch [158], Tensorflow [159], and Theano [160] are most often used. Summaries of the hardware and software requirements for brain MRI segmentation and classification of AD using deep learning approaches are described in Table 10.
According to recent research, deep learning is promising for the analysis of brain MRI and can overcome the issues associated with the earlier state-of-the-art machine learning algorithms. Brain MRI analysis using computer-aided techniques has been challenging because of its complex structure, irregular appearance, imperfect image acquisition, non-standardized MR scales, imaging protocol variations, and presence of pathology. Hence, more generic methods using deep learning are preferable to manage these vulnerabilities.
Even though deep learning techniques in brain MRI analysis have great potential, there are still some limitations. It does not display competitive results on relatively small datasets, whereas it outperforms other methods on large datasets like ImageNet [161]. Several studies have shown that most of the methods consistently achieve better results when the size of the training datasets is increased [162,163]. There has been an increasing demand for large-scale datasets so that deep learning techniques can be used more efficaciously. It is challenging to acquire a large amount of brain MRI data due to legal and privacy issues. Therefore, it is necessary to develop a deep learning method using many different brain MRI datasets. One solution is to augment the data from the existing dataset. For this, random transformations such as translation, flipping, deformation, and rotation called data augmentation can be applied to the original data to increase the size of the dataset and are applied in deep learning methods. Many studies have shown that data augmentation leads to benefits that introduce random variations to the original data and reduce overfitting [33,150]. Furthermore, supervised learning techniques construct predictive models by learning from a large number of training examples, where each training example has a label indicating its ground-truth output. Although current techniques have achieved great success, it is difficult to get strong supervision information like fully ground truth labels due to the high cost of data labeling in many applications. In circumstances in which there are difficulties in obtaining large brain MRI datasets with ground truth annotations, deep learning techniques should work with weak supervision, which can be trained on small datasets. Deep weakly-supervised learning models can be a solution to identify diseases in brain MRI without the need for a large number of ground truth annotations. These models allow us to simultaneously classify brain MRI, yielding pixel-wise localization scores, thereby identifying the corresponding regions of interests (ROIs). Transfer learning could be used to share well-performing deep learning models, which are trained on normal and pathological brain MRI data, among the brain imaging research community. The generalization ability of these models could then be improved across datasets with less effort required.

6. Conclusions

In medical imaging, the advancement of computational intelligence, deep learning, and computer-aided detection has attracted much attention in brain segmentation and AD diagnosis. Although deep learning methods have a high impact on the quantitative analysis of brain MRI, there is still difficulty in finding a robust, generic method. Pre-processing initialization and post-processing can affect the performance of deep learning techniques. Here, we reviewed the-state-of-the-art studies of brain structure and classification of brain MRI for the diagnosis of AD. Moreover, we discussed how brain structure segmentation improves the classification performance of AD. The segmentation for brain MRI helps to facilitate the interpretation and classification of AD. Brain MRI segmentation can be challenging work due to the images having a noisy background, partial volume effect, and low contrast. Furthermore, the automatic classification of AD is quite challenging due to the low contrast of the anatomical structure in MRI. To overcome these difficulties, various methods for segmentation have been proposed, with varying complexity. These methods have resulted in more accurate results in the past few decades. The segmentation of the brain structure and classification of AD using deep learning approaches has gained attention due to the ability to provide efficacious results over a large-scale data set as well as to learn and make decisions on its own.

Author Contributions

Conceptualization, N.Y., J.Y.C. and B.L.; methodology, N.Y., J.Y.C., and B.L.; formal analysis, N.Y., J.Y.C., and B.L.; investigation, N.Y., J.Y.C., and B.L.; writing—original draft preparation, N.Y. and J.Y.C.; writing—review and editing, B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (No. NRF-2019R1A4A1029769 and NRF-2019R1I1A3A01058959).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Miriam, R.; Malin, B.; Jan, A.; Patrik, B.; Fredrik, B.; Jorgen, R.; Ingrid, L.; Lennart, B.; Richard, P.; Martin, R.; et al. PET/MRI and PET/CT hybrid imaging of rectal cancer–description and initial observations from the RECTOPET (Rectal Cancer trial on PET/MRI/CT) study. Cancer Imaging 2019, 19, 52. [Google Scholar]
  2. Rebecca, S.; Diana, M.; Eric, J.; Choonsik, L.; Heather, F.; Michael, F.; Robert, G.; Randell, K.; Mark, H.; Douglas, M.; et al. Use of Diagnostic Imaging Studies and Associated Radiation Exposure for Patients Enrolled in Large Integrated Health Care Systems, 1996–2010. JAMA 2012, 307, 2400–2409. [Google Scholar]
  3. Hsiao, C.-J.; Hing, E.; Ashman, J. Trends in electronic health record system use among office-based physicians: United States, 2007–2012. Natl. Health Stat. Rep. 2014, 1, 1–18. [Google Scholar]
  4. Despotović, I.; Goossens, B.; Philips, W. MRI Segmentation of the Human Brain: Challenges, Methods, and Applications. Comput. Math. Methods Med. 2015, 2015, 1–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Akkus, Z.; Galimzianova, A.; Hoogi, A.; Rubin, D.L.; Erickson, B.J. Deep Learning for Brain MRI Segmentation: State of the Art and Future Directions. J. Digit. Imaging 2017, 30, 449–459. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Mulder, E.R.; De Jong, R.A.; Knol, D.L.; Van Schijndel, R.A.; Cover, K.S.; Visser, P.J.; Barkhof, F.; Vrenken, H.; Alzheimer’s Disease Neuroimaging Initiative. Hippocampal volume change measurement: Quantitative assessment of the reproducibility of expert manual outlining and the automated methods FreeSurfer and FIRST. NeuroImage 2014, 92, 169–181. [Google Scholar] [CrossRef]
  7. Corso, J.J.; Sharon, E.; Dube, S.; El-Saden, S.; Sinha, U.; Yuille, A. Efficient Multilevel Brain Tumor Segmentation with Integrated Bayesian Model Classification. IEEE Trans. Med. Imaging 2008, 27, 629–640. [Google Scholar] [CrossRef] [Green Version]
  8. Yang, C.; Sethi, M.; Rangarajan, A.; Ranka, S. Supervoxel-Based Segmentation of 3D Volumetric Images. Appl. E Comput. 2017, 10111, 37–53. [Google Scholar]
  9. Killiany, R.J.; Hyman, B.T.; Gómez-Isla, T.; Moss, M.B.; Kikinis, R.; Jolesz, F.; Tanzi, R.; Jones, K.; Albert, M.S. MRI measures of entorhinal cortex vs hippocampus in preclinical AD. Neurology 2002, 58, 1188–1196. [Google Scholar] [CrossRef]
  10. Ryu, S.Y.; Kwon, M.J.; Lee, S.B.; Yang, D.W.; Kim, T.W.; Song, I.U.; Yang, P.S.; Kim, H.J.; Lee, A.Y. Measurement of Precuneal and Hippocampal Volumes Using Magnetic Resonance Volumetry in Alzheimer’s Disease. J. Clin. Neurol. 2010, 6, 196–203. [Google Scholar] [CrossRef] [Green Version]
  11. Lin, D.; Vasilakos, A.V.; Tang, Y.; Yao, Y. Neural networks for computer-aided diagnosis in medicine: A review. Neurocomputing 2016, 216, 700–708. [Google Scholar] [CrossRef]
  12. Kooi, T.; Litjens, G.; Van Ginneken, B.; Gubern-Mérida, A.; Sánchez, C.I.; Mann, R.M.; Heeten, A.D.; Karssemeijer, N. Large scale deep learning for computer aided detection of mammographic lesions. Med. Image Anal. 2017, 35, 303–312. [Google Scholar] [CrossRef] [PubMed]
  13. Cheng, J.-Z.; Ni, N.; Chou, Y.-H.; Qin, J.; Tiu, C.-M.; Chang, Y.-C.; Huang, C.-S.; Shen, D.; Chen, C.-M. Computer-Aided Diagnosis with Deep Learning Architecture: Applications to Breast Lesions in US Images and Pulmonary Nodules in CT Scans. Sci. Rep. 2016, 6, 2445. [Google Scholar] [CrossRef] [Green Version]
  14. Geert, L.; Clara, S.; Nadya, T.; Meyke, H.; Iris, N.; Iringo, K.; Christina, H.; Peter, B.; Bram, G.; Jeroen, L. Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci. Rep. 2016, 6, 26286. [Google Scholar]
  15. Marcus, D.; Wang, T.H.; Parker, J.; Csernansky, J.G.; Morris, J.C.; Buckner, R.L. Open Access Series of Imaging Studies (OASIS): Cross-sectional MRI Data in Young, Middle Aged, Nondemented, and Demented Older Adults. J. Cogn. Neurosci. 2007, 19, 1498–1507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Jack, C.R.; Bernstein, M.A.; Fox, N.; Thompson, P.; Alexander, G.; Harvey, D.; Borowski, B.; Britson, P.J.; Whitwell, J.L.; Ward, C.; et al. The Alzheimer’s disease neuroimaging initiative (ADNI): MRI methods. J. Magn. Reson. Imaging 2008, 27, 685–691. [Google Scholar] [CrossRef] [Green Version]
  17. Warfield, S.; Landman, B. MICCAI 2012 Workshop on Multi-Atlas Labeling, in: MICCAI Grand Challenge and Workshop on Multi-Atlas Labeling; CreateSpace Independent Publishing Platform: Scotts Valley, CA, USA, 2012. [Google Scholar]
  18. IBSR Dataset. Available online: https://www.nitrc.org/projects/ibsr (accessed on 4 June 2020).
  19. Morris, C. The clinical dementia rating (CDR). Neurology 1993, 43, 2412. [Google Scholar] [CrossRef]
  20. Rubin, E.H.; Storandt, M.; Miller, J.P.; Kinscherf, D.A.; Grant, E.A.; Morris, J.C.; Berg, L. A prospective study of cognitive function and onset of dementia in cognitively healthy elders. Arch. Neurol. 1998, 55, 395–401. [Google Scholar] [CrossRef]
  21. Duta, N.; Sonka, M. Segmentation and interpretation of MR brain images. An improved active shape model. IEEE Trans. Med. Imaging 1998, 17, 1049–1062. [Google Scholar] [CrossRef]
  22. Rogowska, J. Overview and fundamentals of medical image segmentation. In Handbook of Medical Image Processing and Analysis; Elsevier: Amsterdam, The Netherlands, 2000. [Google Scholar]
  23. Vovk, U.; Pernus, F.; Likar, B. A Review of Methods for Correction of Intensity Inhomogeneity in MRI. IEEE Trans. Med. Imaging 2007, 26, 405–421. [Google Scholar] [CrossRef]
  24. Coupe, P.; Yger, P.; Prima, S.; Hellier, P.; Kervrann, C.; Barillot, C. An optimized blockwise nonlocal means denoising filter for 3-D magnetic resonance images. IEEE Trans. Med. Imaging 2008, 27, 425–441. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Arno, K.; Jesper, A.; Babak, A.; John, A.; Brian, A.; Ming-Chang, C.; Gary, C.; Louis, C.; James, G.; Pierre, H.; et al. Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. NeuroImage 2009, 46, 786–802. [Google Scholar]
  26. Zhang, Y.-D.; Dong, Z.; Wu, L.; Wang, S. A hybrid method for MRI brain image classification. Expert Syst. Appl. 2011, 38, 10049–10053. [Google Scholar] [CrossRef]
  27. Neffati, S.; Ben Abdellafou, K.; Jaffel, I.; Taouali, O.; Bouzrara, K. An improved machine learning technique based on downsized KPCA for Alzheimer’s disease classification. Int. J. Imaging Syst. Technol. 2018, 29, 121–131. [Google Scholar] [CrossRef]
  28. Saraswathi, S.; Mahanand, B.; Kloczkowski, A.; Suresh, S.; Sundararajan, N. Detection of onset of Alzheimer’s disease from MRI images using a GA-ELM-PSO classifier. In Proceedings of the 2013 Fourth International Workshop on Computational Intelligence in Medical Imaging (CIMI), Singapore, 16–19 April 2013; pp. 42–48. [Google Scholar]
  29. Ding, Y.; Zhang, C.; Lan, T.; Qin, Z.; Zhang, X.; Wang, W. Classification of Alzheimer’s disease based on the combination of morphometric feature and texture feature. In Proceedings of the 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Washington, DC, USA, 9–12 November 2015; pp. 409–412. [Google Scholar]
  30. Kwon, G.-R.; Kim, J.-I.; Kwon, G.-R. Diagnosis of Alzheimer’s Disease Using Dual-Tree Complex Wavelet Transform, PCA, and Feed-Forward Neural Network. J. Health Eng. 2017, 2017, 1–13. [Google Scholar] [CrossRef]
  31. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  32. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Li, F. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  33. Alex, K.; Ilya, S.; Hg, E. Imagenet classification with deep convolutional neural networks. In Proceedings of the NIPS, IEEE Neural Information Processing System Foundation, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  34. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  35. Zhang, W.; Li, R.; Deng, H.; Wang, L.; Lin, W.; Ji, S.; Shen, D. Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage 2015, 108, 214–224. [Google Scholar] [CrossRef]
  36. Brebisson, A.; Montana, G. Deep neural networks for anatomical brain segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7 June 2015. [Google Scholar]
  37. Moeskops, P.; Benders, M.; Chiţǎ, S.M.; Kersbergen, K.J.; Groenendaal, F.; De Vries, L.S.; Viergever, M.A.; Išgum, I. Automatic segmentation of MR brain images of preterm infants using supervised classification. NeuroImage 2015, 118, 628–641. [Google Scholar] [CrossRef]
  38. Dong, N.; Wang, L.; Gao, Y.; Shen, D. Fully convolutional networks for multi-modality isointense infant brain image segmentation. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 1342–1345. [Google Scholar]
  39. Brosch, T.; Tang, L.Y.W.; Yoo, Y.; Li, D.K.B.; Traboulsee, A.; Tam, R. Deep 3D Convolutional Encoder Networks with Shortcuts for Multiscale Feature Integration Applied to Multiple Sclerosis Lesion Segmentation. IEEE Trans. Med. Imaging 2016, 35, 1229–1239. [Google Scholar] [CrossRef]
  40. Shakeri, M.; Tsogkas, S.; Ferrante, E.; Lippe, S.; Kadoury, S.; Paragios, N.; Kokkinos, I. Sub-cortical brain structure segmentation using F-CNNs. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016. [Google Scholar]
  41. Kong, Z.; Luo, J.; Xu, S.; Li, T. Automatical and accurate segmentation of cerebral tissues in fMRI dataset with combination of image processing and deep learning. Opt. Biophotonics Low-Resour. Settings IV 2018, 10485. [Google Scholar] [CrossRef]
  42. Milletari, F.; Ahmadi, S.-A.; Kroll, C.; Plate, A.; Rozanski, V.; Maiostre, J.; Levin, J.; Dietrich, O.; Ertl-Wagner, B.; Bötzel, K.; et al. Hough-CNN: Deep learning for segmentation of deep brain regions in MRI and ultrasound. Comput. Vis. Image Underst. 2017, 164, 92–102. [Google Scholar] [CrossRef] [Green Version]
  43. Mehta, R.; Majumdar, A.; Sivaswamy, J. Brainsegnet: A convolutional neural network architecture for automated segmentation of human brain structures. J. Med. Imaging 2017, 4, 024003. [Google Scholar] [CrossRef] [PubMed]
  44. Kamnitsas, K.; Ledig, C.; Newcombe, V.; Simpson, J.P.; Kane, A.; Menon, D.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef]
  45. Pereira, S.; Pinto, A.; Alves, V.; Silva, C. Brain Tumor Segmentation Using Convolutional Neural Networks in MRI Images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef]
  46. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; LaRochelle, H. Brain tumor segmentation with Deep Neural Networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [Green Version]
  47. Dou, Q.; Chen, H.; Yu, L.; Zhao, L.; Qin, J.; Wang, D.; Mok, V.C.-T.; Shi, L.; Heng, P.A. Automatic Detection of Cerebral Microbleeds from MR Images via 3D Convolutional Neural Networks. IEEE Trans. Med. Imaging 2016, 35, 1182–1195. [Google Scholar] [CrossRef]
  48. Chen, H.; Dou, Q.; Yu, L.; Qin, J.; Heng, P.A. VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images. NeuroImage 2018, 170, 446–455. [Google Scholar] [CrossRef]
  49. Lyksborg, M.; Puonti, O.; Agn, M.; Larsen, R. An Ensemble of 2D Convolutional Neural Networks for Tumor Segmentation. Appl. E Comput. 2015, 9127, 201–211. [Google Scholar]
  50. Chen, X.-W.; Lin, X. Big Data Deep Learning: Challenges and Perspectives. IEEE Access 2014, 2, 514–525. [Google Scholar] [CrossRef]
  51. Alom, Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sagan, V.; Nasrin, M.S.; Hasan, M.; Van Essen, B.C.; Awwal, A.A.S.; Asari, V.K. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics 2019, 8, 292. [Google Scholar] [CrossRef] [Green Version]
  52. Srhoj-Egekher, V.; Benders, M.; Viergever, M.A.; Išgum, I. Automatic neonatal brain tissue segmentation with MRI. SPIE Med. Imaging 2013, 8669, 86691K. [Google Scholar] [CrossRef]
  53. Anbeek, P.; Išgum, I.; Van Kooij, B.J.M.; Mol, C.P.; Kersbergen, K.J.; Groenendaal, F.; Viergever, M.A.; De Vries, L.S.; Benders, M. Automatic Segmentation of Eight Tissue Classes in Neonatal Brain MRI. PLoS ONE 2013, 8, e81895. [Google Scholar] [CrossRef]
  54. Vrooman, H.; Cocosco, C.; Lijn, F.; Stokking, R.; Ikram, M.; Vernooij, M.; Breteler, M.; Niessen, W.J. Multi-spectral brain tissue segmentation using automatically trained k-nearestneighbor classification. NeuroImage 2007, 37, 71–81. [Google Scholar] [CrossRef] [PubMed]
  55. Makropoulos, A.; Gousias, I.S.; Ledig, C.; Aljabar, P.; Serag, A.; Hajnal, J.V.; Edwards, A.D.; Counsell, S.; Rueckert, D. Automatic Whole Brain MRI Segmentation of the Developing Neonatal Brain. IEEE Trans. Med. Imaging 2014, 33, 1818–1831. [Google Scholar] [CrossRef] [PubMed]
  56. Christensen, G.E.; Rabbitt, R.D.; Miller, M.I. Deformable templates using large deformation kinematics. IEEE Trans. Image Process. 1996, 5, 1435–1447. [Google Scholar] [CrossRef]
  57. Davatzikos, C.; Prince, J. Brain image registration based on curve mapping. In Proceedings of the IEEE Workshop on Biomedical Image Analysis, Seattle, WA, USA, 24–25 June 1994; pp. 245–254. [Google Scholar]
  58. Carmichael, O.T.; Aizenstein, H.J.; Davis, S.W.; Becker, J.T.; Thompson, P.M.; Meltzer, C.C.; Liu, Y. Atlas-based hippocampus segmentation in Alzheimer’s disease and mild cognitive impairment. NeuroImage 2005, 27, 979–990. [Google Scholar] [CrossRef] [Green Version]
  59. Wang, L.; Gao, Y.; Shi, F.; Li, G.; Gilmore, J.; Lin, W.; Shen, D. Links: Learning-based multi-source integration framework for segmentation of infant brain images. NeuroImage 2014, 108, 734–746. [Google Scholar]
  60. Chiţǎ, S.M.; Benders, M.; Moeskops, P.; Kersbergen, K.J.; Viergever, M.A.; Išgum, I. Automatic segmentation of the preterm neonatal brain with MRI using supervised classification. SPIE Med Imaging 2013, 8669, 86693. [Google Scholar]
  61. Cuingnet, R.; Gerardin, E.; Tessieras, J.; Auzias, G.; Lehéricy, S.; Habert, M.-O.; Chupin, M.; Benali, H.; Colliot, O. Automatic classification of patients with Alzheimer’s disease from structural MRI: A comparison of ten methods using the ADNI database. NeuroImage 2011, 56, 766–781. [Google Scholar] [CrossRef] [Green Version]
  62. Wolz, R.; Julkunen, V.; Koikkalainen, J.; Niskanen, E.; Zhang, D.P.; Rueckert, D.; Soininen, H.; Lötjönen, J.M.P.; The Alzheimer’s Disease Neuroimaging Initiative. Multi-Method Analysis of MRI Images in Early Diagnostics of Alzheimer’s Disease. PLoS ONE 2011, 6, e25446. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Braak, H.; Braak, E. Staging of alzheimer’s disease-related neurofibrillary changes. Neurobiol. Aging 1995, 16, 271–278. [Google Scholar] [CrossRef]
  64. Dubois, B.; Feldman, H.H.; Jacova, C.; DeKosky, S.T.; Barberger-Gateau, P.; Cummings, J.; Delacourte, A.; Galasko, U.; Gauthier, S.; Jicha, G.; et al. Research criteria for the diagnosis of Alzheimer’s disease: Revising the NINCDS–ADRDA criteria. Lancet Neurol. 2007, 6, 734–746. [Google Scholar] [CrossRef]
  65. Maryam, H.; Bashir, B.; Jamshid, D.; Tim, E. Segmentation of the Hippocampus for Detection of Alzheimer’s Disease, Advances in Visual Computing; Springer: Berlin/Heidelberg, Germany, 2012; pp. 42–50. [Google Scholar]
  66. Claude, I.; Daire, J.-L.; Sebag, G. Fetal Brain MRI: Segmentation and Biometric Analysis of the Posterior Fossa. IEEE Trans. Biomed. Eng. 2004, 51, 617–626. [Google Scholar] [CrossRef] [PubMed]
  67. Wang, Y.; Zhu, Y.; Gao, W.; Xu, H.; Lin, Z. The measurement of the volume of stereotaxic MRI hippocampal formation applying the region growth algorithm based on seeds. In Proceedings of the IEEE/ICME International Conference on Complex Medical Engineering, Harbin, China, 22–25 May 2011. [Google Scholar]
  68. Tan, P. Hippocampus Region Segmentation for Alzheimer’s Disease Detection; Technische Universiteit Eindhoven: Eindhoven, The Netherlands, 2009. [Google Scholar]
  69. Bao, S.; Chung, A. Multi-scale structured cnn with label consistency for brain MR image segmentation. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2016, 6, 1–5. [Google Scholar] [CrossRef]
  70. Dolz, J.; Desrosiers, C.; Ben Ayed, I. 3D fully convolutional networks for subcortical segmentation in MRI: A large-scale study. NeuroImage 2018, 170, 456–470. [Google Scholar] [CrossRef] [Green Version]
  71. Wachinger, C.; Reuter, M.; Klein, T. DeepNAT: Deep convolutional neural network for segmenting neuroanatomy. NeuroImage 2018, 170, 434–445. [Google Scholar] [CrossRef]
  72. Khagi, B.; Kwon, G.; Lama, R. Comparative analysis of Alzheimer’s disease classification by CDR level using CNN, feature selection, and machine-learning techniques. Int. J. Imaging Syst. Technol. 2019, 29, 297–310. [Google Scholar] [CrossRef]
  73. Bernal, J.; Kushibar, K.; Cabezas, M.; Valverde, S.; Oliver, A.; Llado, X. Quantitative Analysis of Patch-Based Fully Convolutional Neural Networks for Tissue Segmentation on Brain Magnetic Resonance Imaging. IEEE Access 2019, 7, 89986–90002. [Google Scholar] [CrossRef]
  74. Jiong, W.; Zhang, Y.; Wang, K.; Tang, X. Skip Connection U-Net for White Matter Hyperintensities Segmentation from MRI. IEEE Access 2019, 7, 155194–155202. [Google Scholar] [CrossRef]
  75. Chen, C.-C.C.; Chai, J.-W.; Chen, H.-C.; Wang, H.C.; Chang, Y.-C.; Wu, Y.-Y.; Chen, W.-H.; Chen, H.-M.; Lee, S.-K.; Chang, C.-I. An Iterative Mixed Pixel Classification for Brain Tissues and White Matter Hyperintensity in Magnetic Resonance Imaging. IEEE Access 2019, 7, 124674–124687. [Google Scholar] [CrossRef]
  76. Pengcheng, L.; Zhao, Y.; Liu, Y.; Chen, Q.; Liu, F.; Gao, C. Temporally Consistent Segmentation of Brain Tissue from Longitudinal MR Data. IEEE Access 2020, 8, 3285–3293. [Google Scholar] [CrossRef]
  77. Pulkit, K.; Pravin, N.; Chetan, A. U-SegNet: Fully Convolutional Neural Network based Automated Brain tissue segmentation Tool. In Proceedings of the 2018 25th IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018. [Google Scholar]
  78. Wortmann, M. World Alzheimer’s report 2014: Dementia and risk reduction. Alzheimer’s Dement. 2015, 11, 837. [Google Scholar] [CrossRef]
  79. Morra, J.H.; Tu, Z.; Apostolova, L.G.; Green, A.E.; Toga, A.W.; Thompson, P.M. Comparison of AdaBoost and Support Vector Machines for Detecting Alzheimer’s Disease Through Automated Hippocampal Segmentation. IEEE Trans. Med. Imaging 2009, 29, 30–43. [Google Scholar] [CrossRef] [Green Version]
  80. Brookmeyer, R.; Johnson, E.; Ziegler-Graham, K.; Arrighi, H.M. Forecasting the global burden of Alzheimer’s disease. Alzheimer’s Dement. 2007, 3, 186–191. [Google Scholar] [CrossRef] [Green Version]
  81. Siqi, L.; Liu, S.; Cai, W.; Pujol, S.; Kikinis, R.; Feng, D.D. Early diagnosis of Alzheimer’s disease with deep learning. In Proceedings of the 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April–2 May 2014; pp. 1015–1018. [Google Scholar]
  82. Suk, H.; Lee, S.; Shen, D. Hierarchical feature representation and multimodal fusion with deep learning for AD-MCI diagnosis. NeuroImage 2014, 101, 569–582. [Google Scholar] [CrossRef] [Green Version]
  83. Payan, A.; Montana, G. Predicting Alzheimer’s disease: A neuroimaging study with 3D convolutional neural networks. In Proceedings of the ICPRAM 2015-4th International Conference on Pattern Recognition Applications and Methods, Lisbon, Portugal, 10–12 January 2015. [Google Scholar]
  84. Andres, O.; Munilla, J.; Gorriz, J.; Ramirez, J. Ensembles of Deep Learning Architectures for the Early Diagnosis of the Alzheimer’s Disease. Int. J. Neural Syst. 2016, 26, 1650025. [Google Scholar] [CrossRef]
  85. Hosseini, E.; Gimel’farb, G.; El-Baz, A. Alzheimer’s disease diagnostics by a deeply supervised adapTable 3D convolutional network, Computational and Mathematical Methods in Medicine. arXiv 2016, arXiv:1607.00556. [Google Scholar]
  86. Sarraf, S.; Tofighi, G. Classification of Alzheimer’s disease using fMRI data and deep learning convolutional neural networks. arXiv 2016, arXiv:1603.08631. [Google Scholar]
  87. Liu, M.; Zhang, J.; Adeli, E.; Shen, D. Landmark-based deep multi-instance learning for brain disease diagnosis. Med. Image Anal. 2017, 43, 157–168. [Google Scholar] [CrossRef]
  88. Aderghal, K.; Benois-Pineau, J.; Karim, A.; Gwenaelle, C. Fuseme: Classification of sMRI images by fusion of deep CNNs in 2D+E projections. In Proceedings of the 15th International Workshop on Content-Based Multimedia Indexing, Florence, Italy, 19–21 June 2017. [Google Scholar]
  89. Shi, J.; Zheng, X.; Li, Y.; Zhang, Q.; Ying, S. Multimodal Neuroimaging Feature Learning with Multimodal Stacked Deep Polynomial Networks for Diagnosis of Alzheimer’s Disease. IEEE J. Biomed. Health Inform. 2018, 22, 173–183. [Google Scholar] [CrossRef] [PubMed]
  90. Korolev, S.; Safiullin, A.; Belyaev, M.; Dodonova, Y. Residual and plain convolutional neural networks for 3D brain MRI classification. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017. [Google Scholar]
  91. Jyoti, I.; Zhang, Y. Brain MRI analysis for Alzheimer’s disease diagnosis using an ensemble system of deep convolutional neural networks. Brain Inform. 2018, 5, 2. [Google Scholar] [CrossRef]
  92. Lu, D.; Popuri, K.; Ding, W.; Balachandar, R.; Faisal, M. Multimodal and multiscale deep neural networks for the early diagnosis of Alzheimer’s disease using structural MRI and FDG-pPET images. Sci. Rep. 2017, 8, 1–13. [Google Scholar]
  93. Khvostikov, A.; Aderghal, K.; Benois-Pineau, J.; Krylov, A.; Catheline, G. 3D CNN-based classification using sMRI and md-DTI images for Alzheimer disease studies. arXiv 2018, arXiv:1801.05968. [Google Scholar]
  94. Aderghal, K.; Khvostikov, A.; Krylov, A.; Benois-Pineau, J.; Afdel, K.; Catheline, G. Classification of Alzheimer Disease on Imaging Modalities with Deep CNNs Using Cross-Modal Transfer Learning. In Proceedings of the 2018 IEEE 31st International Symposium on Computer-Based Medical Systems (CBMS), Karlstad, Sweden, 18–21 June 2018. [Google Scholar]
  95. Lian, C.; Liu, M.; Zhang, J.; Shen, D. Hierarchical Fully Convolutional Network for Joint Atrophy Localization and Alzheimer’s Disease Diagnosis Using Structural MRI. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 880–893. [Google Scholar] [CrossRef]
  96. Liu, M.; Cheng, D.; Wang, K.; Wang, Y.; The Alzheimer’s Disease Neuroimaging Initiative. Multi-Modality Cascaded Convolutional Neural Networks for Alzheimer’s Disease Diagnosis. Neuroinformatics 2018, 16, 295–308. [Google Scholar] [CrossRef]
  97. Lee, B.; Ellahi, W.; Choi, J.Y. Using Deep CNN with Data Permutation Scheme for Classification of Alzheimer’s Disease in Structural Magnetic Resonance Imaging (sMRI). IEICE Trans. Inf. Syst. 2019, E102, 1384–1395. [Google Scholar] [CrossRef] [Green Version]
  98. Feng, C.; ElAzab, A.; Yang, P.; Wang, T.; Zhou, F.; Hu, H.; Xiao, X.; Lei, B. Deep Learning Framework for Alzheimer’s Disease Diagnosis via 3D-CNN and FSBi-LSTM. IEEE Access 2019, 7, 63605–63618. [Google Scholar] [CrossRef]
  99. Mefraz, K.; Abraham, N.; Hon, M. Transfer Learning with Intelligent Training Data Selection for Prediction of Alzheimer’s Disease. IEEE Access 2019, 7, 72726–72735. [Google Scholar] [CrossRef]
  100. Ruoxuan, C.; Liu, M. Hippocampus Analysis by Combination of 3-D DenseNet and Shapes for Alzheimer’s Disease Diagnosis. IEEE J. Biomed. Health Inform. 2019, 23, 2099–2107. [Google Scholar] [CrossRef]
  101. Ahmed, S.; Choi, K.Y.; Lee, J.J.; Kim, B.C.; Kwon, G.-R.; Lee, K.H.; Jung, H.Y.; Goo-Rak, K. Ensembles of Patch-Based Classifiers for Diagnosis of Alzheimer Diseases. IEEE Access 2019, 7, 73373–73383. [Google Scholar] [CrossRef]
  102. Fang, X.; Liu, Z.; Xu, M. Ensemble of deep convolutional neural networks based multi-modality images for Alzheimer’s disease diagnosis. IET Image Process. 2020, 14, 318–326. [Google Scholar] [CrossRef]
  103. Kam, T.-E.; Zhang, H.; Jiao, Z.; Shen, D. Deep Learning of Static and Dynamic Brain Functional Networks for Early MCI Detection. IEEE Trans. Med. Imaging 2020, 39, 478–487. [Google Scholar] [CrossRef] [PubMed]
  104. Shi, Y.; Suk, H.-I.; Gao, Y.; Lee, S.-W.; Shen, D. Leveraging Coupled Interaction for Multimodal Alzheimer’s Disease Diagnosis. IEEE Trans. Neural Networks Learn. Syst. 2020, 31, 186–200. [Google Scholar] [CrossRef]
  105. Di Paola, M.; Di Iulio, F.; Cherubini, A.; Blundo, C.; Casini, A.R.; Sancesario, G.; Passafiume, D.; Caltagirone, C.; Spalletta, G. When, where, and how the corpus callosum changes in MCI and AD: A multimodal MRI study. Neurology 2010, 74, 1136–1142. [Google Scholar] [CrossRef] [PubMed]
  106. Liu, S.; Liu, S.; Cai, W.; Che, H.; Pujol, S.; Kikinis, R.; Feng, D.; Fulham, M. Adni Multimodal neuroimaging feature learning for multiclass diagnosis of Alzheimer’s disease. IEEE Trans. Biomed. Eng. 2014, 62, 1132–1140. [Google Scholar] [CrossRef] [Green Version]
  107. Loveman, E.; Green, C.; Fleming, J.; Takeda, A.; Picot, J.; Payne, E.; Clegg, A. The Clinical and Cost-effectiveness of Donepezil, Rivastigmine, Galantamine and Memantine for Alzheimer’s Disease; Health Technology Assessment: Winchester, UK, 2006. [Google Scholar]
  108. Braak, H.; Braak, E. Neuropahological staging of Alzheimer-related changes. Acta Neuropath. Appl. Neurobiol. 1991, 14, 39–44. [Google Scholar] [CrossRef]
  109. Fox, N.C.; Warrington, E.K.; Freeborough, P.; Hartikainen, P.; Kennedy, A.M.; Stevens, J.M.; Rossor, M. Presymptomatic hippocampal atrophy in Alzheimer’s disease. A longitudinal MRI study. Brain 1996, 119, 2001–2007. [Google Scholar] [CrossRef]
  110. Rose, S. Loss of connectivity in Alzheimer’s disease: An evaluation of white matter tract integrity with color coded MR diffusion tensorimaging. J. Neurol. Neurosurg. Psychiatry 2000, 69, 528–530. [Google Scholar] [CrossRef] [Green Version]
  111. Petersen, R.; Doody, R.; Kurz, A.; Mohs, R.; Morris, J.; Rabins, P.; Ritchie, K.; Rossor, M.; Thal, L.; Winblad, B. Current concepts in mild cognitive impairment Archives of neurology. Arch. Neurol. 2002, 58, 1985–1992. [Google Scholar] [CrossRef]
  112. Perri, R.; Serra, L.; Carlesimo, G.; Caltagirone, C. Preclinical dementia: An Italian multicenter study on amnestic mild cognitive impairment Dementia and geriatric cognitive disorders. Dement. Geriatr. Cogn. Disord. 2007, 23, 289–300. [Google Scholar] [CrossRef] [PubMed]
  113. Bozzali, M.; Franceschi, M.; Falini, A.; Pontesilli, S.; Cercignani, M.; Magnani, G.; Scotti, G.; Comi, G.; Filippi, M. Quantification of tissue damage in AD using diffusion tensor and magnetization transfer MRI. Neurology 2001, 57, 1135–1137. [Google Scholar] [CrossRef] [PubMed]
  114. Bozzali, M.; Falini, A.; Franceschi, M.; Cercignani, M.; Zuffi, M.; Scotti, G.; Comi, G.; Filippi, M. White matter damage in Alzheimer’s disease assessed in vivo using diffusion tensor magnetic resonance imaging. J. Neurol. Neurosurg. Psychiatry 2002, 72, 742–746. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  115. Kloppel, S.; Stonnington, C.; Chu, C.; Draganski, B.; Scahill, R.; Rohrer, J.; Fox, N.; Jack, C.; Ashburner, J.; Frackowiak, R. Automatic classification of MR scans in Alzheimer’s disease. Brain 2008, 131, 681–689. [Google Scholar] [CrossRef]
  116. Jack, C.R.; Wiste, H.J.; Vemuri, P.; Weigand, S.D.; Senjem, M.L.; Zeng, G.; Bernstein, M.A.; Gunter, J.L.; Pankratz, V.S.; Aisen, P.S.; et al. Brain beta-amyloid measures and magnetic resonance imaging atrophy both predict time-to-progression from mild cognitive impairment to Alzheimer’s disease. Brain 2010, 133, 3336–3348. [Google Scholar] [CrossRef]
  117. Klein-Koerkamp, Y.; Heckemann, R.; Baraly, K.; Moreaud, O.; Keignart, S.; Krainik, A.; Hammers, A.; Baciu, M.; Hot, P. Amygdalar atrophy in early Alzheimer’s disease. Curr. Alzheimer Res. 2014, 11, 239–252. [Google Scholar] [CrossRef] [Green Version]
  118. Colliot, O.; Chételat, G.; Chupin, M.; Desgranges, B.; Magnin, B.; Benali, H.; Dubois, B.; Garnero, L.; Eustache, F.; Lehéricy, S. Discrimination between Alzheimer Disease, Mild Cognitive Impairment, and Normal Aging by Using Automated Segmentation of the Hippocampus. Radiology 2008, 248, 194–201. [Google Scholar] [CrossRef] [Green Version]
  119. Dickerson, B.; Goncharova, I.; Sullivan, M.; Forchetti, C.; Wilson, R.; Bennett, D.; Beckett, L.; Morrell, L. MRI-derived entorhinal and hippocampal atrophy in incipient and very mild Alzheimer’s disease. Neurobiol. Aging 2001, 22, 747–754. [Google Scholar] [CrossRef]
  120. Wang, L.; Swank, J.; Glick, I.; Gado, M.; Miller, M.; Morris, J.; Csernansky, J. Changes in hippocampal volume and shape across time distinguish dementia of the Alzheimer’s type from healthy aging. NeuroImage 2003, 20, 667–682. [Google Scholar] [CrossRef]
  121. Barnes, J.; Foster, J.; Boyes, R.; Pepple, T.; Moore, E.; Schott, J.; Frost, C.; Scahill, R.; Fox, N. A comparison of methods for the automated calculation of volumes and atrophy rates in the hippocampus. NeuroImage 2008, 40, 1655–1671. [Google Scholar] [CrossRef]
  122. Wolz, R.; Heckemann, R.A.; Aljabar, P.; Hajnal, J.V.; Hammers, A.; Lötjönen, J.M.P.; Rueckert, D. Measurement of hippocampal atrophy using 4D graph-cut segmentation: Application to ADNI. NeuroImage 2010, 52, 109–118. [Google Scholar] [CrossRef] [PubMed]
  123. Leung, K.K.; Barnes, J.; Ridgway, G.; Bartlett, J.; Clarkson, M.J.; Macdonald, K.; Schuff, N.; Fox, N.; Ourselin, S.; The Alzheimer’s Disease Neuroimaging Initiative. Automated cross-sectional and longitudinal hippocampal volume measurement in mild cognitive impairment and Alzheimer’s disease. NeuroImage 2010, 51, 1345–1359. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  124. Shen, D.; Moffat, S.; Resnick, S.M.; Davatzikos, C. Measuring Size and Shape of the Hippocampus in MR Images Using a Deformable Shape Model. NeuroImage 2002, 15, 422–434. [Google Scholar] [CrossRef] [Green Version]
  125. Chupin, M.; Gerardin, E.; Cuingnet, R.; Boutet, C.; Lemieux, L.; Lehéricy, S.; Benali, H.; Garnero, L.; Colliot, O.; The Alzheimer’s Disease Neuroimaging Initiative. Fully automatic hippocampus segmentation and classification in Alzheimer’s disease and mild cognitive impairment applied on data from ADNI. Hippocampus 2009, 19, 579–587. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  126. Lötjönen, J.M.P.; Wolz, R.; Koikkalainen, J.; Thurfjell, L.; Waldemar, G.; Soininen, H.; Rueckert, D.; The Alzheimer’s Disease Neuroimaging Initiative. Fast and robust multi-atlas segmentation of brain magnetic resonance images. NeuroImage 2010, 49, 2352–2365. [Google Scholar] [CrossRef] [PubMed]
  127. Wolz, R.; Aljabar, P.; Hajnal, J.V.; Hammers, A.; Rueckert, D.; The Alzheimer’s Disease Neuroimaging Initiative. LEAP: Learning embeddings for atlas propagation. NeuroImage 2009, 49, 1316–1325. [Google Scholar] [CrossRef] [Green Version]
  128. Coupe, P.; Eskildsen, S.F.; Manjón, J.V.; Fonov, V.S.; Pruessner, J.C.; Allard, M.; Collins, D.L.; The Alzheimer’s Disease Neuroimaging Initiative. Scoring by nonlocal image patch estimator for early detection of Alzheimer’s disease. NeuroImage Clin. 2012, 1, 141–152. [Google Scholar] [CrossRef] [PubMed]
  129. Tong, T.; Wolz, R.; Gao, Q.; Hajnal, J.; Rueckert, D. Multiple instance learning for classification of dementia in brain MRI. Med Image Anal. 2013, 16, 599–606. [Google Scholar]
  130. Davatzikos, C.; Bhatt, P.; Shaw, L.; Batmanghelich, K.; Trojanowski, J. Prediction of MCI to AD conversion, via MRI, CSF biomarkers, and pattern classification. Neurobiol. Aging 2011, 32, 2322.e19–2322.e27. [Google Scholar] [CrossRef] [Green Version]
  131. Eskildsen, S.F.; Coupe, P.; García-Lorenzo, D.; Fonov, V.S.; Pruessner, J.C.; Collins, D.L.; The Alzheimer’s Disease Neuroimaging Initiative. Prediction of Alzheimer’s disease in subjects with mild cognitive impairment from the ADNI cohort using patterns of cortical thinning. NeuroImage 2012, 65, 511–521. [Google Scholar] [CrossRef] [Green Version]
  132. Ashburner, J.; Friston, K. Voxel-based morphometry-the methods. NeuroImage 2000, 11, 805–821. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  133. Ashburner, J.; Friston, K.J. Nonlinear spatial normalization using basis functions. Hum. Brain Mapp. 1999, 7, 254–266. [Google Scholar] [CrossRef]
  134. Koikkalainen, J.; Lötjönen, J.; Thurfjell, L.; Rueckert, D.; Waldemar, G.; Soininen, H.; The Alzheimer’s Disease Neuroimaging Initiative. Multi-template tensor-based morphometry: Application to analysis of Alzheimer’s disease. NeuroImage 2011, 56, 1134–1144. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  135. Good, C.D.; Johnsrude, I.S.; Ashburner, J.; Henson, R.N.; Friston, K.J.; Frackowiak, R. A Voxel-Based Morphometric Study of Ageing in 465 Normal Adult Human Brains. NeuroImage 2001, 14, 21–36. [Google Scholar] [CrossRef] [Green Version]
  136. Bozzali, M.; Filippi, M.; Magnani, G.; Cercignani, M.; Franceschi, M.; Schiatti, E.; Falini, A. The contribution of voxelbased morphometry in staging patients with mild cognitive impairment. Neurology 2006, 67, 453–460. [Google Scholar] [CrossRef]
  137. Chételat, G.; Landeau, B.; Eustache, F.; Mézenge, F.; Viader, F.; De La Sayette, V.; Desgranges, B.; Baron, J.C. Using voxel-based morphometry to map the structural changes associated with rapid conversion in MCI: A longitudinal MRI study. NeuroImage 2005, 27, 934–946. [Google Scholar] [CrossRef]
  138. Kinkingnéhun, S.; Sarazin, M.; Lehericy, S.; Guichart-Gomez, E.; Hergueta, T.; Dubois, B. VBM anticipates the rate of progression of Alzheimer disease: A 3-year longitudinal study. Neurology 2008, 70, 2201–2211. [Google Scholar] [CrossRef]
  139. Serra, L.; Cercignani, M.; Lenzi, D.; Perri, R.; Fadda, L.; Caltagirone, C.; Macaluso, E.; Bozzali, M. Grey and White Matter Changes at Different Stages of Alzheimer’s Disease. J. Alzheimer’s Dis. 2010, 19, 147–159. [Google Scholar] [CrossRef]
  140. Yang, X.; Tan, M.Z.; Qiu, A. CSF and Brain Structural Imaging Markers of the Alzheimer’s Pathological Cascade. PLoS ONE 2012, 7, e47406. [Google Scholar] [CrossRef] [Green Version]
  141. Liu, M.; Li, F.; Yan, H.; Wang, K.; Ma, Y.; Shen, L.; Xu, M. A multi-model deep convolutional neural network for automatic hippocampus segmentation and classification in Alzheimer’s disease. NeuroImage 2020, 208, 116459. [Google Scholar] [CrossRef]
  142. Cao, L.; Li, L.; Zheng, J.; Fan, X.; Yin, F.; Shen, H.; Zhang, J. Multi-task neural networks for joint hippocampus segmentation and clinical score regression. Multimedia Tools Appl. 2018, 77, 29669–29686. [Google Scholar] [CrossRef]
  143. Choi, H.; Jin, K. Predicting cognitive decline with deep learning of brain metabolism and amyloid imaging. Behav. Brain Res. 2018, 344, 103–109. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  144. Vansteenkiste, E. Quantitative Analysis of Ultrasound Images of the Preterm Brain. Ph.D. Thesis, Ghent University, Ghent, Belgium, 2007. [Google Scholar]
  145. Carass, A.; Roy, S.; Jog, A.; Cuzzocreo, J. Longitudinal multiple sclerosis lesion segmentation: Resourceand challenge. NeuroImage 2017, 148, 72–102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  146. Dice, L.R. Measures of the Amount of Ecologic Association between Species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  147. Crum, W.; Camara, O.; Hill, D.L. Generalized overlap measures for evaluation and validation in medical image analysis. IEEE Trans. Med. Imaging 2006, 25, 1451–1461. [Google Scholar] [CrossRef]
  148. Rathore, S.; Habes, M.; Iftikhar, M.A.; Shacklett, A.; Davatzikos, C. A review on neuroimaging-based classification studies and associated feature extraction methods for Alzheimer’s disease and its prodromal stages. NeuroImage 2017, 155, 530–548. [Google Scholar] [CrossRef]
  149. Hinrichs, C.; Singh, V.; Mukherjee, L.; Xu, G.; Chung, M.K.; Johnson, S.C.; The Alzheimer’s Disease Neuroimaging Initiative. Spatially augmented LPboosting for AD classification with evaluations on the ADNI dataset. NeuroImage 2009, 48, 138–149. [Google Scholar] [CrossRef] [Green Version]
  150. Salvatore, C.; Cerasa, A.; Battista, P.; Gilardi, M.C.; Quattrone, A.; Castiglioni, I. Magnetic resonance imaging biomarkers for the early diagnosis of Alzheimer’s disease: A machine learning approach. Front. Mol. Neurosci. 2015, 9, 270. [Google Scholar] [CrossRef] [Green Version]
  151. Kleesiek, J.; Urban, G.; Hubert, A.; Schwarz, D.; Maier-Hein, K.; Bendszus, M.; Biller, A. Deep MRI brain extraction: A 3D convolutional neural network for skull stripping. NeuroImage 2016, 129, 460–469. [Google Scholar] [CrossRef]
  152. Islam, J.; Zhang, Y. An Ensemble of Deep Convolutional Neural Networks for Alzheimer’s Disease Detection and Classification. In Proceedings of the 31st Conference on Neural Information Processing System (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  153. Lavin, A.; Gray, S. Fast algorithms for convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  154. Vemuri, P.; Gunter, J.L.; Senjem, M.; Whitwell, J.L.; Kantarci, K.; Knopman, D.S.; Boeve, B.F.; Petersen, R.C.; Jack, C.R. Alzheimer’s disease diagnosis in individual subjects using structural MR images: Validation studies. NeuroImage 2007, 39, 1186–1197. [Google Scholar] [CrossRef] [Green Version]
  155. Paine, T.; Jin, H.; Yang, J.; Lin, Z.; Huang, T. GPU asynchronous stochastic gradient descent to speed up neural network training. arXiv 2013, arXiv:1312.6186. [Google Scholar]
  156. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  157. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the MM 2014-Proceedings of the 2014 ACM Conference on Multimedia, New York, NY, USA, 3–7 November 2014. [Google Scholar]
  158. Collobert, R.; Kavukcuoglu, K.; Farabet, C. Torch7: A Matlab-Like Environment for Machine Learning. In Proceedings of the NIPS 2011 Workshop on Algorithms, Systems, and Tools for Learning at Scale, Sierra Nevada, Spain, 16–17 December 2011. [Google Scholar]
  159. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv 2015, arXiv:1603.04467. [Google Scholar]
  160. Bastien, F.; Lamblin, P.; Pascanu, R.; Bergstra, J.; Goodfellow, I.; Bergeron, A.; Bouchard, N.; Warde-Farley, D.; Bengio, Y. Theano: New features and speed improvements, CoRR. arXiv 2012, arXiv:1211.5590. [Google Scholar]
  161. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV 2015), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  162. Cho, J.; Lee, K.; Shin, E.; Choy, G.; Do, S. How much data is needed to train a medical image deep learning system to achieve necessary high accuracy. arXiv 2015, arXiv:1511.06348. [Google Scholar]
  163. Lekadir, K.; Galimzianova, A.; Betriu, A.; Vila, M.D.M.; Igual, L.; Rubin, D.L.; Fernandez, E.; Radeva, P.; Napel, S. A Convolutional Neural Network for Automatic Characterization of Plaque Composition in Carotid Ultrasound. IEEE J. Biomed. Health Inform. 2017, 21, 48–55. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The regions of interest (outlined using yellow color) illustrating the boundaries of the left precuneus (a) in the sagittal plane and right hippocampus (b) head (c) body and (d) tail.
Figure 1. The regions of interest (outlined using yellow color) illustrating the boundaries of the left precuneus (a) in the sagittal plane and right hippocampus (b) head (c) body and (d) tail.
Sensors 20 03243 g001
Figure 2. General pipeline for brain MRI analysis.
Figure 2. General pipeline for brain MRI analysis.
Sensors 20 03243 g002
Figure 3. Pre-processing of MRI. (a) T1-W original MRI, (b) Brain tissue image after removal of the nonbrain structure, (c) The bias field, (d) Brain tissue image after bias field correction.
Figure 3. Pre-processing of MRI. (a) T1-W original MRI, (b) Brain tissue image after removal of the nonbrain structure, (c) The bias field, (d) Brain tissue image after bias field correction.
Sensors 20 03243 g003
Figure 4. Generally used architecture of convolutional neural networks (CNNs).
Figure 4. Generally used architecture of convolutional neural networks (CNNs).
Sensors 20 03243 g004
Figure 5. The overall block diagram of AD diagnosis.
Figure 5. The overall block diagram of AD diagnosis.
Sensors 20 03243 g005
Table 1. Details of the OASIS, ADNI, IBSR, and MICCAI datasets.
Table 1. Details of the OASIS, ADNI, IBSR, and MICCAI datasets.
DatasetClass# of SubjectsSexAgeMMSE# of MRI Scans
MFMeanStdMeanStd
OASISAD100415976.767.1124.324.16100
HC31611919745.0923.1129.630.83316
ADNIAD1921019175.37.523.32.1530
MCI39825714174.77.427.01.81126
HC22912010975.85.029.11.0877
IBSRHC1814471---18
MICCAIHC35------35
Table 2. Categorization of segmentation methods using CNN architecture on brain MRI.
Table 2. Categorization of segmentation methods using CNN architecture on brain MRI.
StrategiesAuthorsDescription
Semantic-wiseDong [38]
Brosch [39]
Shakeri [40]
Zhenglun [41]
Milletari [42]
Raghav [43]
The main objective of the semantic-wise segmentation is to link each pixel of an image with its class label. It is called dense prediction because every pixel is predicted from the whole input image. Later, segmentation labels are mapped with the input image in a way that minimizes the loss function.
Patch–wiseKamnitsas [44]
Pereira [45]
Havaei [46]
Zhang [35]
Brebisson [36]
Moeskops [37]
Patch-wise segmentation handles high-resolution images, and the input images are split as local patches. An N×N patch is extracted from the input image. These patches are trained and provide class labels to identify normal or abnormal brain images. The design network consists of convolution layers, transfer functions, pooling, and sub-sampling layers, and fully connected layers.
CascadedDou [47]The cascaded architecture types are used to combine two different CNN architectures. The output of the first architecture is fed into the second architecture to get classification results. The first architecture is used to train the model with the initial prediction of class labels, and later for fine-tuning.
Single-modalityMoeskops [37]
Brebisson [36]
Raghav [43]
Milletari [42]
Shakeri [40]
This type of modality refers to single-source information and is adaptable to different scenarios. The single modality commonly used in the public dataset for tissue-type segmentation in brain MRI (mainly T1-W images).
Multi-modalityZhang [35]
Chen [48]
Lyksborg [49]
Multi-source information can be used, and it might require a larger number of parameters than using a single modality. The advantage of using multi-modality is to gain valuable contrast information. Furthermore, using multi-path configurations, the imaging sequences can be processed in parallel (e.g., T1 and T2, fluid-attenuated inversion recovery (FLAIR)).
Table 3. The brain structure segmentation methods based on deep learning.
Table 3. The brain structure segmentation methods based on deep learning.
No.AuthorsYearStrategiesDimensionKey MethodClassifierDataset
1Zhang [35]2015Patch-wise2DCNNSoft-maxClinical data
2Brebisson [36]2015Patch-wise2D/3DCNNSoft-maxMICCAI 2012
3Moeskops [37]2016Patch-wise2D/3DCNNSoft-maxNeoBrainS12
4Bao [69]2016Patch-wise2DCNNSoft-maxIBSR/LPBA40
5Dong [38]2016Semantic-wise2DCNNSoft-maxClinical data
6Shakeri [40]2016Semantic-wise2DFCNNSoft-maxIBSR data
7Raghav [43]2017Semantic-wise2D/3DM-Net + CNNSoft-maxIBSR/ MICCAI 2012
8Milletari [42]2017Semantic-wise2D/3DHough-CNNSoft-maxMICCAI 2012
9Dolz [70]2018Semantic-wise3DCNNSoft-maxIBSR/ABIDE
10Wachinger [71]2018Patch-Based3DCNNSoft-maxMICCAI 2012
11Zhenglun [41]2018Semantic-wise2DWavelet + CNNSoft-maxClinical data
12Khagi [72]2018Semantic-wise2DSegNet + CNNSoft-maxOASIS Dataset
13Bernal [73]2019Patch-Based2D/3DFCNNSoft-maxIBSR, MICCAI2012 & iSeg2017
14Jiong [74]2019Semantic-wise2DU-netSoft-maxMICCAI2017
15Chen [75]2019Semantic-wise2DLCMV-BrainWeb
16Pengcheng [76]2020Semantic-wise3D/4DFuzzy C-mean-BLSA
Table 4. Overview of papers using deep learning techniques for the segmentation of brain MRI.
Table 4. Overview of papers using deep learning techniques for the segmentation of brain MRI.
AuthorsMethodsApplication: Key Features
Zhang [35]CNNTissue segmentation: multi-modal 2D segmentation for isointense brain tissues using the deep CNN architecture.
Brebisson [36]CNNAnatomical segmentation: fusing multi-scale 2D patches with a 3D patch using a CNN.
Moeskops [37]CNNTissue segmentation: CNN trained on multiple patches and kernel sizes to extract information from each voxel.
Bao [69]CNNAnatomical segmentation: multi-scale late fusion CNN with a random walker as a novel label consistency method.
Dong [38]CNNTissue segmentation: FCN with a late fusion method on different modalities.
Shakeri [40]FCNNAnatomical segmentation: FCN followed by Markov random fields, whose topology corresponds to a volumetric grid.
Raghav [43]M-Net + CNNTissue segmentation: the 3D contextual information of a given slice is converted into a 2D slice using CNN.
Milletari [42]Hough-CNNAnatomical segmentation: Hough-voting to acquire mapping from CNN features to full patch segmentations.
Dolz [70]CNNAnatomical segmentation: 3D CNN architecture for the segmentation of subcortical MRI brain structure.
Wachinger [71]CNNAnatomical segmentation: neuroanatomy in T1-W MRI segmentation using deep CNN.
Zhenglun [41]Wavelet + CNNTissue segmentation: pre-processing is performed with the wavelet multi-scale transformation, and then, CNN is applied for the segmentation of brain MRI.
Bernal [73]FCNNTissue segmentation: the quantitative analysis of patch-based FCNN.
Jiong [74]U-netTissue segmentation: skip-connection U-net for WM hyper intensities segmentation.
Chen [75]LCMVTissue segmentation: new iterative linearly constrained minimum variance (LCMV) classification-based method developed for hyperspectral classification.
Pengcheng [76]Fuzzy C-meanTissue segmentation: fuzzy C-means framework to improve the temporal consistency of adults’ brain tissue segmentation.
Table 5. Comparison of the state-of-the-art methods in the field of AD diagnosis.
Table 5. Comparison of the state-of-the-art methods in the field of AD diagnosis.
No.AuthorsYearContentModalitiesKey MethodClassifierData (Size)
1Siqi [81]2014Full brainMRIAuto-encoderSoft-maxADNI (311)
2Suk [82]2015Full brainMRI + PETCNNSoft-maxADNI (204)
3Payan [83]2015Full brainMRICNNSoft-maxADNI (755)
4Andres [84]2016Gray matterMRI + PETDeep Belief NetworkNNADNI (818)
5Hosseini [85]2016Full brainfMRICNNSoft-maxADNI (210)
6Saraf [86]2016Full brainfMRICNNSoft-maxADNI (58)
7Mingxia [87]2017Full brainMRICNNSoft-maxADNI (821)
8Aderghal [88]2017HippocampusMRI + DTICNNSoft-maxADNI (1026)
9Shi [89]2017Full brainMRI + PETAuto-encoderSoft-maxADNI (207)
10Korolev [90]2017Full brainMRICNNSoft-maxADNI (821)
11Jyoti [91]2018Full brainMRICNNSoft-maxOASIS (416)
12Donghuan [92]2018Full brainMRICNNSoft-maxADNI (626)
13Khvostikov [93]2018HippocampusMRI + DTICNNSoft-maxADNI (214)
14Aderghal [94]2018HippocampusMRI + DTICNNSoft-maxADNI (815)
15Lian [95]2018Full brainMRIFCNSoft-maxADNI (821)
16Liu [96]2018Full brainMRI + PETCNNSoft-maxADNI (397)
17Lee [97]2019Full brainMRICNNAlex-NetADNI (843),
OASIS (416)
18Feng [98] 2019Full brain MRI + PETCNNSoft-maxADNI (397)
19Mefraz [99]2019Full brainMRITransfer learningSoft-maxADNI (50)
20Ruoxuan [100]2019HippocampusMRICNNSoft-maxADNI (811)
21Ahmed [101]2019Full brainMRICNNSoft-maxADNI (352)
GARD (326)
22Fung [102]2020Full brainMRI + PETCNNAdaboostADNI (352)
23Kam [103]2020Full brainMRICNNSoft-maxADNI (352)
24Shi [104]2020Full brainMRI + PET + CSFMachine learningAdaboost ADNI (202)
Table 6. Overview of existing methods using deep learning for the classification of AD.
Table 6. Overview of existing methods using deep learning for the classification of AD.
AuthorsMethodsApplications: Key Features
Siqi [81]Auto-encoderAD/HC classification: deep learning architecture contains sparse auto-encoders and a softmax regression layer for the classification of AD
Suk [82]CNNAD/MCI/HC classification: neuroimaging modalities for latent hierarchical feature representation from extracted patches using CNN
Payan [83]CNNAD/MCI/HC classification: 3D CNN pre-trained with sparse auto-encoders
Andres [84]Deep Belief NetworkAD/HC classification: automated anatomical labeling brain regions for the construction of classification techniques using deep learning architecture
Hosseini [85]CNNAD/MCI/HC classification: 3D CNN pre-trained with a 3D convolutional auto-encoder on MRI data
Saraf [86]CNNAD/HC classification: adapted Lenet-5 architecture on fMRI data
Mingxia [87]CNNAD/MCI/HC classification: landmark-based deep multi-instance learning framework for brain disease diagnosis
Aderghal [88]CNNAD/HC classification: separate CNN base classifier to form an ensemble of CNNs, each trained with a corresponding plane of MRI brain data
Shi [89]Auto-encoderAD/MCI/HC classification: multi-modal stacked deep polynomial networks with an SVM classifier on top layer using MRI and PET
Korolev [90]CNNAD/MCI/HC classification: residual and plain CNNs for 3D brain MRI
Jyoti [91]CNNAD/HC classification: deep CNN model for resolving an imbalanced dataset to identify AD and recognize the disease stages.
Donghuan [92]CNNAD/MCI classification: early diagnosis of AD by combing the multiple different modalities using multiscale and multimodal deep neural networks.
Khvostikov [93]CNNAD/HC classification: multi-modality fusion on hippocampal ROI using CNN
Aderghal [94]CNNAD/HC classification: diffusion tensor imaging modality from MRI using the transfer learning method
Lian [95]FCNAD/MCI/HC classification: CNN to discriminate the local patches in the brain MRI and multi-scale features are fused to construct hierarchical classification models for the diagnosis of AD.
Liu [96]CNNAD/MCI/HC classification: CNN to learn multi-level and multimodal features of MRI and PET brain images.
Lee [97]CNNAD/MCI/HC classification: data permutation scheme for the classification of AD in MRI using deep CNN.
Feng [98] CNNAD/MCI/HC classification: 3D-CNN designed to extract deep feature representation from both MRI and PET. Fully stacked bidirectional long short-term memory (FSBi-LSTM) applied to the hidden spatial information from deep feature maps to improve the performance.
Mefraz [99]Transfer learningAD/MCI/HC classification: transfer learning with intelligent training data selection for the prediction of AD and CNN pre-trained with VGG architecture.
Ruoxuan [100]CNNAD/MCI/HC classification: a new hippocampus analysis method combining the global and local features of the hippocampus by 3D densely connected CNN.
Ahmed [101]CNNAD/HC classification: ensembles of patch-based classifiers for the diagnosis of AD.
Fung [102]CNNAD/MCI/HC classification: an ensemble of deep CNNs with multi-modality images for the diagnosis of AD.
Kam [103]CNNAD/MCI/HC classification: CNN framework to simultaneously learn embedded features from brain functional networks (BFNs).
Shi [104]Machine LearningAD/MCI/HC classification: MRI, PET, and CSF are used as multimodal data. Coupled boosting and coupled metric ensemble scheme to model and learn an informative feature projection form the different modalities.
Table 7. Summary of the validation measures of brain segmentation and their mathematical formula regarding the number of true positives (TP), false positives and false-negative (FN) at voxel and lesion levels (TPL, FPL, and FNL).
Table 7. Summary of the validation measures of brain segmentation and their mathematical formula regarding the number of true positives (TP), false positives and false-negative (FN) at voxel and lesion levels (TPL, FPL, and FNL).
Metrics of Segmentation QualityMathematical Description
True positive rate, TPR (Sensitivity) T P R = T P T P + F N
Positive predictive rate, PPV (Precision) P P V = T P T P + F P
Negative predictive rate, NPV T N T N + F N
Dice similarity coefficient, DSC T P R = 2 T P 2   T P +   FP   + F N
Volume difference rate, VDR T P R = | F P F N | T P + F N
Lesion-wise true positive rate, LTPR L T P R = T P L T P L + F N L
Lesion-wise positive predictive value, LPPV T P P V = T P L T P L + F P L
Specificity T N T N + F P
F1 score 2   T P 2   T P +   FP   + F N
Accuracy T P + T N T P +   TN   +   FP   + F N
Balanced Accuracy ( S e n s i t i v i t y + S p e c i f i c i t y ) 2
Table 8. Summary of results in the existing methods using deep learning approaches for brain structure segmentation. (†: DSC, *: JI) (Unit: %).
Table 8. Summary of results in the existing methods using deep learning approaches for brain structure segmentation. (†: DSC, *: JI) (Unit: %).
AuthorsMICCAI [17]OASIS [15]Clinical/IBSR [18]
DSC and JIDSC and JIDSC and JI
CSFGMWMCSFGMWMCSFGMWM
1Zhang [35]------83.5 85.2 86.4
2Brebisson [36]72.5 72.5 72.5 ------
3Moeskops [37]73.5 73.5 73.5 ------
4Bao [69]------82.2 85.0 82.2
5Dong [38]------85.5 87.3 88.7
6Zhenglun [41]------94.3 *90.2 *91.4 *
7Khagi [72]---72.2
73.0 *
74.6
74.0 *
81.9
85.0 *
---
8Shakeri [40]------82.4 82.4 82.4
9Raghav [43]74.3 74.3 74.3 ---84.4 84.4 84.4
10Milletari [42]------77.0 77.0 77.0
11Dolz [70]------90.0 90.0 90.0
12Wachinger [71]90.6 90.6 90.6 ------
13Chen [75]------93.6 94.8 97.5
Table 9. A brief review of the state-of-the-art methods for AD classification (AD vs. HC) and MCI conversion prediction (pMCI vs. sMCI) using MRI data. (The best results obtained for different metrics are highlighted in bold).
Table 9. A brief review of the state-of-the-art methods for AD classification (AD vs. HC) and MCI conversion prediction (pMCI vs. sMCI) using MRI data. (The best results obtained for different metrics are highlighted in bold).
AuthorsSubjectsAD vs. HCpMCI vs. sMCI
ACCSENSPEAUCACCSENSPEAUC
1Siqi [81]204HC + 180AD0.790.830.870.78----
2Suk [82]101HC + 128sMCI + 76pMCI + 93AD0.920.920.950.970.720.370.910.73
3Korolev [90]61HC + 77sMCI + 43pMCI + 50AD0.80--0.870.52--0.52
4Khvostikov [93]58HC + 48AD0.850.880.90-----
5Lian [95]429HC + 465sMCI + 205pMCI + 358AD0.900.820.970.950.810.530.850.78
6Mingxia [87]229HC + 226sMCI + 167pMCI + 203AD0.910.880.930.950.760.420.820.77
7Andres [84]68HC + 70AD0.900.860.940.95----
8Adherghal [84]228HC + 188AD0.850.840.87-----
9Donghuan [92]360HC + 409sMCI + 217pMCI----0.750.730.76-
10Shi [89]52 NC + 56 sMCI + 43 pMCI + 51AD0.950.940.960.960.750.630.850.72
11Payan [83]755 subjects (AD, MCI, HC)0.95-------
12Hosseini [85]70HC + 70AD0.99-0.98-----
13Lee [97]843 subjects (AD, MCI, HC)0.980.960.97-----
14Liu [96]397 subjects (AD, MCI, HC)0.930.920.930.95----
15Feng [98] 397 subjects (AD, MCI, HC)0.940.970.920.96----
16Ruoxuan [100]811 subjects (AD, MCI, HC)0.900.860.920.920.730.690.750.76
Table 10. Summary of the hardware and software details required for the segmentation of brain MRI and classification of AD using deep learning methods.
Table 10. Summary of the hardware and software details required for the segmentation of brain MRI and classification of AD using deep learning methods.
AuthorDatasetScannerHardwareSoftwareTraining Time
Zhang [35]Clinical data3T SiemensTesla K20c GPU with 2496 coresiBEAT toolbox
ITK-SNAP toolbox
less than one day
Brebisson [36]MICCAI 2012-NVIDIA Tesla K40 GPU with 12 GB memory.Python with Theano framework-
Moeskops [37]NeoBrainS123T Philips Achieva-BET toolbox
FMRIB software library
-
Bao [69]IBSR
LPBA40
1.5 T GE-FLIRT toolbox-
Dong [38]Clinical data3T Siemens-Python with Caffe framework
iBEAT toolbox
-
Raghav [43]IBSR
MICCAI 2012
LPBA40
Hammers67n20
-
-
1.5 T GE
1 T Philips HPQ
NVIDIA K40 GPU, with 12 GB of RAM.Python with Keras packages
BET toolbox
-
Milletari [42]Clinical data-Intel i7 quad-core workstations with 32 GB of RAM and Nvidia GTX 980 (4 GB -RAM).Python with Caffe framework-
Dolz [70]IBSR
ABIDE
-Intel(R) Core(TM) i7-6700 K 4.0 GHz CPU and NVIDIA GeForce GTX 960 GPU with 2 GB of memory.Python with Theano framework
FreeSurfer 5.1 tool
Medical Interaction Tool Kit
2 days and a half
Wachinger [71]MICCAI 2012-NVIDIA Tesla K40 and TITAN X with 12 GB GPU memoryPython with Caffe framework
FreeSurfer tool
1 day(train)
1 h(test)
Bernal [73]IBSR
MICCAI2012 iSeg2017
-Ubuntu 16.04, with 128 GB RAM and TITAN-X PASCAL GPU with 8 GB RAMPython with Keras packages-
Jiong [74]MICCAI2017 -Ubuntu 16.04 with 32 GB RAM and GTX 1080 Ti GPUs.Python with Keras packages-
Chen [75]BrainWeb1.5 T SiemensWindows 7 computer with CPU Intel R Xeon R E5-2620 v3 @ 2.40 GHz processor and 32 GB RAM--
Pengcheng [76]BLSA--FSL software
3D-Slicer
-
Hosseini [85]ADNI1.5 T Siemens TrioAmazon EC2 g 2.8 x large with GPU GRID K520Python with Theano framework
-
Saraf [86]ADNI3T Siemens TrioNVIDIA GPUPython with Caffe framework
BET toolbox
FMRIB Software Library v 5.0
-
Mingxia [87]ADNI-1
ADNI-2
MIRIAD
1.5 T Siemens Trio
3 T Siemens Trio
1.5 T Signa GE
NVIDIA GTX TITAN 12 GB GPUMIPAV software
FSL software
27 h
<1 s (test)
Aderghal [88]ADNI1.5 T Siemens TrioIntel® Xeon® CPU E5-2680 v2 with 2.80 GHz and Tesla K20Xm with 2496 CUDA cores GPUPython with Caffe framework 2 h, 3 min
Jyoti [91]OASIS1.5 T SiemensLinux X86-64 with AMD A8 CPU, 16 GB RAM and NVIDIA GeForce GTX 770Python with Tensorflow and Keras library-
Khvostikov [93]ADNI1.5 T Siemens TrioIntel Core i7-6700 HQ with Nvidia GeForce
GTX 960 M and Intel Core i7-7700 HQ CPU with Nvidia GeForce GTX 1070 GPU
Python with Tensorflow framework
BET toolbox
-
Lian [95]ADNI-1
ADNI-2
1.5 T Siemens Trio
3 T Siemens Trio
NVIDIA GTX TITAN 12 GB GPUPython with Keras packages-
Liu [96]ADNI1.5 T Siemens TrioGPU NVIDIA GTX1080.Python with Theano framework and Keras packages
Lee [97]ADNI
OASIS
1.5 T Siemens Trio
1.5 T Siemens
Nvidia GTX 1080Ti GPU--
Feng [98] ADNI1.5 T Siemens TrioWindows with NVIDIA TITA- Xt GPUMIPAV Software
Keras library with Tensorflow as backend
-
Ruoxuan [100]ADNI1.5 T Siemens TrioUbuntu14.04-x64/ GPU of NVIDIA GeForce GTX 1080 TiFreeSurfer tool
Keras library with Tensorflow as backend
-
Ahmed [101]ADNI
GARD
1.5 T Siemens TrioIntel(R) Xeon (R) CPU E5-1607 v4 @ 3.10 GHz, 32 GB RAM NVIDIA Quadro M4000Keras library with Tensorflow as backend-
Fung [102]ADNI1.5 T Siemens TrioDesktop PC equipped with Intel Core i7, 8 GB memory and GPU with 16 G NVIDIA P100 × 8Ubuntu 16.04, Keras library with Tensorflow
MATLAB 2014b with SPM
-

Share and Cite

MDPI and ACS Style

Yamanakkanavar, N.; Choi, J.Y.; Lee, B. MRI Segmentation and Classification of Human Brain Using Deep Learning for Diagnosis of Alzheimer’s Disease: A Survey. Sensors 2020, 20, 3243. https://doi.org/10.3390/s20113243

AMA Style

Yamanakkanavar N, Choi JY, Lee B. MRI Segmentation and Classification of Human Brain Using Deep Learning for Diagnosis of Alzheimer’s Disease: A Survey. Sensors. 2020; 20(11):3243. https://doi.org/10.3390/s20113243

Chicago/Turabian Style

Yamanakkanavar, Nagaraj, Jae Young Choi, and Bumshik Lee. 2020. "MRI Segmentation and Classification of Human Brain Using Deep Learning for Diagnosis of Alzheimer’s Disease: A Survey" Sensors 20, no. 11: 3243. https://doi.org/10.3390/s20113243

APA Style

Yamanakkanavar, N., Choi, J. Y., & Lee, B. (2020). MRI Segmentation and Classification of Human Brain Using Deep Learning for Diagnosis of Alzheimer’s Disease: A Survey. Sensors, 20(11), 3243. https://doi.org/10.3390/s20113243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop