Next Article in Journal
A High-Dimensional and Small-Sample Submersible Fault Detection Method Based on Feature Selection and Data Augmentation
Next Article in Special Issue
Two-Stage Classification Model for the Prediction of Heart Disease Using IoMT and Artificial Intelligence
Previous Article in Journal
Mammography Image-Based Diagnosis of Breast Cancer Using Machine Learning: A Pilot Study
Previous Article in Special Issue
Advances in Thermal Image Analysis for the Detection of Pregnancy in Horses Using Infrared Thermography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Analysis of Deep-Neural-Network-Based Automatic Diagnosis of Diabetic Retinopathy

1
Department of Electrical Engineering, School of Engineering, University of Management and Technology (UMT), Lahore 54770, Pakistan
2
Department of Computer Engineering, Umm Al-Qura University, Makkah 21955, Saudi Arabia
3
Department of Information Systems, Umm Al-Qura University, Makkah 21955, Saudi Arabia
4
Telecommunications Engineering School, University of Malaga, 29010 Malaga, Spain
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 205; https://doi.org/10.3390/s22010205
Submission received: 20 November 2021 / Revised: 13 December 2021 / Accepted: 22 December 2021 / Published: 29 December 2021
(This article belongs to the Special Issue Biomedical Image and Signals for Treatment Monitoring)

Abstract

:
Diabetic retinopathy (DR) is a human eye disease that affects people who are suffering from diabetes. It causes damage to their eyes, including vision loss. It is treatable; however, it takes a long time to diagnose and may require many eye exams. Early detection of DR may prevent or delay the vision loss. Therefore, a robust, automatic and computer-based diagnosis of DR is essential. Currently, deep neural networks are being utilized in numerous medical areas to diagnose various diseases. Consequently, deep transfer learning is utilized in this article. We employ five convolutional-neural-network-based designs (AlexNet, GoogleNet, Inception V4, Inception ResNet V2 and ResNeXt-50). A collection of DR pictures is created. Subsequently, the created collections are labeled with an appropriate treatment approach. This automates the diagnosis and assists patients through subsequent therapies. Furthermore, in order to identify the severity of DR retina pictures, we use our own dataset to train deep convolutional neural networks (CNNs). Experimental results reveal that the pre-trained model Se-ResNeXt-50 obtains the best classification accuracy of 97.53% for our dataset out of all pre-trained models. Moreover, we perform five different experiments on each CNN architecture. As a result, a minimum accuracy of 84.01% is achieved for a five-degree classification.

1. Introduction

Diabetic retinopathy (DR) is a human eye infection in people with diabetes. It is initiated due to retinal vascular damage, which is caused by diabetes mellitus for a long-duration [1]. This disease is one of the most common reasons behind blindness [2]. Therefore, its detection in the early stages is critical [3]. There are many treatments for this disease; however, they take plenty of time and may even include many eye tests such as photo-coagulation and vitrectomy [4].
According to a survey in Europe, almost 60 million people are diabetes patients and they are most prone to DR. In the United States, 10.2 million people with an age of 40 or above have diabetes. Furthermore, 40% of these people are at risk of some vision-threatening disease [5]. Moreover, the survey of the Center for Disease Control in 2020 revealed that 3.3 million people are suffering from DR [6]. According to the World Health Organization, diabetes has affected 422 million people to date and this number will become 629 million by 2045 [7,8].
DR is normally categorized into five different groups: Normal-0, Mild-1, Moderate-2, Severe-3 and Proliferative-4 as listed in Table 1. The disease starts with small changes in the blood vessels of the eyes, which could be labeled as Mild DR. Concerning the case of Mild DR, the patient could defeat this disease and complete recovery is possible. If this condition of the disease is left untreated, then it will convert into Moderate DR. The leakage in the blood vessels may start in the case of Moderate DR. In the next stage, if the disease increases further then it changes to Severe and Proliferative DR and it could cause complete vision loss.
The current detection of DR is made through a dilated eye exam in which the doctors put some eye drops into the patient’s eyes. Subsequently, an image of the eye is taken with the help of various medical instruments. This technique is manual and therefore there are always some errors in diagnosis. Another way of detecting DR is examining through ophthalmoscopy. In one study, approximately 16% of patients were diagnosed as DR patients using ophthalmoscopy in respect of 442 right eyes [10].
Image processing is also used to identify DR based on highlights; for example, veins, radiates, hemorrhages and small-scale aneurysms. During this process, digital fundus cameras are used to obtain accurate eye images. Techniques like image enhancement, fusion, morphology detection and image segmentation help medical doctors to obtain more information from the data of medical images [10]. In the case of DR, people are not aware of the disease unless a manual detection is made. Due to the lack of related treatment, according to the specific level of the disease, chances of losing eyesight may increase [11].

1.1. State-of-the-Art on DR Detection Dsing Deep Learning Techniques

Numerous techniques have been proposed to detect DR. This section focuses on multi-class classification using deep learning and neural network techniques. Some studies have classified the fundus images into two categories: diabetic, which includes average to extreme conditions of non-proliferative DR; and non-diabetic, where the person is not affected with DR) [12]. Based on this, they proposed a technique to accurately appoint the class where a fundus image could be labeled, utilizing one principal classifier and back propagation neural organization (BPNN) procedures.
Similarly, a deep-learning-based method has been proposed to classify the fundus photographs for human ophthalmologist diagnostics. Authors have built a novel Siamese-like CNN (convolutional neural network) binocular model based on Inception V3 that can acknowledge fundus pictures of both eyes and yield the output of each eye at the same time [13]. A hybrid approach for diagnosing DR has been proposed that uses histogram equalization (HE) and contrast limited adaptive histogram equalization (CLAHE) to assist the deep learning model [14]. It provides more accentuation and effectiveness by way of the intelligent enhancement of the image during the diagnosis process. The authors exploited five CNN architectures to evaluate the performance parameters for the dataset of DR patients. Their classification methodology classifies images into three different groups based on the condition of the disease [15].
The authors developed a novel ResNet18-based CNN architecture to diagnose DR patients. This approach helps in solving a strong class imbalance problem and generates region scoring maps (RSMs) [16]. Furthermore, it indicates the severity level by highlighting the semantic regions of the fundus image. The authors proposed a technique only for the detection of DR regardless of the severity of DR. They classified images as normal and abnormal for the targeted dataset [17]. Similarly, the authors proposed a deep-learning-based CNN to classify a small dataset of DR images, using Cohen’s kappa as an accuracy metric [18].
In addition to the aforementioned research works, many datasets of fundus images have been developed for DR-related diagnoses. For example, TeleOphta uses a tele-ophthalmology network for diabetic retinopathy screening [19]. Other examples are Digital Retinal Images for Vessel Extraction (DRIVE) and Structured Analysis of the Retina (STARE), which are used to segment the vessel network using local and global vessel features [20,21]. Similarly, the SVM (support vector machine) provides 95% and Bayesian provides 90% accuracy [11]. In this technique, images are segmented, outliers are detected, image analysis is performed and the brightness is controlled. In an another technique, SVM provides 86% accuracy and KNN (K-nearest neighbor) provides 55% accuracy [22]. In KNN, images are clustered with the help of pixel clusters. The fundus image mask is removed with the help of pixel clustering [22].
There is another technique known as the extreme learning machine (ELM) design for detecting a disease in eye blood vessels. This technique is mainly used for the detection of diseased blood vessels. Some of the blood vessels are injured in diabetic retinopathy. In this technique, an image is provided to the ELM. The provided algorithm calculates the grayscale value and chooses some features that provide more information than other pixels. Consequently, researchers can achieve 90% accuracy [10]. Similarly, the authors analyzed various blood vessel segmentation techniques in [23,24]. They further identified the lesions for the detection of diabetic retinopathy. The results were compared with the neural network technique.
Finally, by integrating microaneurysms, haemorrhages and exudates, the authors described a method for detecting non-proliferative diabetic retinopathy [25]. They developed a novel convolutional layer that automatically determines the number of extracted features. Each category is then placed into different folders so that there exist a small number of patches for the model to process at runtime. Subsequently, six convolutional layers are added to the model to obtain a validation accuracy of 72% and a training accuracy of 75% .

1.2. Research Gap

Although pre-trained CNNs have been used previously for different diseases, there is a need to enhance the accuracy of classification using a custom dataset and deep transfer learning. A dataset composed of low-resolution DR images, as employed in the conventional methods of Section 1.1, may cause low accuracy or incorrect classification. At the same time, a high-risk patient in the proliferate category requires immediate cure and diagnosis. Keeping this view, the diagnosis procedure requires high accuracy with adequate images of the posterior pole. In a nutshell, there should be an efficient, immediate and autonomous method that can recognize retinopathy with accurate outcomes. This implies that there should be a methodology to evaluate the classification performance parameters on recent CNN architectures.

1.3. Contributions

In this article we propose a methodology to classify the DR images using five different pre-trained CNNs. The contributions are summarized in the following points:
  • Our proposed methodology is flexible and automatically detects the classified pictures of patients with a higher accuracy. It classifies the dataset based on the severity of the disease in different stages/categories. Moreover, it helps doctors to select one or more CNN architectures for the diagnosis.
  • We have analyzed the robustness of CNN architectures on our constructed (customized) dataset for the diagnosis of DR patients. A brief description of the customized dataset is provided in Section 1.4. It highlights how both CNN and dataset directly or indirectly affect performance evaluation. It implies that deep transfer learning techniques have been used with some pre-trained models and customized datasets to obtain high-accuracy results.
  • We have also analyzed how the previously made architectures will perform on our dataset and how these architectures can be fine-tuned to obtain the best results on our dataset.
  • To the best of our knowledge, the proposed work in this article is the first effort to consider the evaluation of recent CNNs, using a customized dataset.
The objective is to provide accurate and less time-consuming results (as compared to the manual methods) by applying different deep neural network algorithms for the classification of different eyes infected by the illness. This helps to obtain more information from the classified images. Consequently, doctors will be able to detect diabetic retinopathy levels more accurately.

1.4. Customized Dataset for Performance Evaluation

The classification accuracy of the DR mainly depends upon the size of the dataset. This implies that a higher accuracy requires a huge amount of training data using a machine learning algorithm. Moreover, the data should be collected from reliable sources with accurate tags. The following datasets are most widely used for DR detection: Digital Retinal Images for Vessel Extraction (DRIVE) dataset [20], Structured Analysis of the Retina (STARE) dataset [21], E-ophtha dataset [19] and Kaggle Diabetic Retinopathy dataset [26,27].
In this study, we created our custom dataset as explained in Section 4.1. The created dataset was built from different resources which are based on different severity levels. It also includes EyePacs [26], which has collected approximately 5 million images from 75,000 patients. Another dataset from Kaggle, which consists of 53,594 images for testing and 35,126 images for training, is also available for analysis. The Kaggle dataset includes a significant number of pictures (72,743) from DR patients. Furthermore, it has pictures for all DR categories in a single folder. Moreover, it also contains categories of various images and their descriptions in the form of comma separated value (CSV) files.
The corresponding enhancements and preprocessing of data are explained in Section 2.1, where all the images are oriented, resized and horizontally flipped. Moreover, the intensity of images is also enhanced.Furthermore, an augmentation is performed where all the images are made consistent in terms of size and intensity. The aforementioned enhancements and preprocessing techniques help CNN for the robust classification. Based on the aforementioned databases, we constructed a dataset of 5333 images, where 1421 are normal, 954 are mild, 1210 are moderate, 308 are severe and 1440 are high-risk patients (see Section 4.1).

1.5. Organization

The organization of this paper is as follows. The proposed methodology is described in Section 2. Section 3 explains the pre-trained CNN architectures and different performance matrices used in the results. Section 4 reports the results and implementation in light of the proposed methodology. The article is concluded in Section 5.

2. Proposed Approach

The proposed methodology is illustrated in Figure 1. The entire process comprises five steps. First, the retina pictures are pre-processed and supplemented using pre-trained models. Deep transfer learning (DTL) is then used during the training phase. During classification, feature extraction and precise prediction of models are employed. The retina prediction is made using a machine learning algorithm. Subsequently, it is classified into five different groups based on the severity of DR as described in Table 1.
The following steps are involved during the prediction process: dataset, data pre-processing, model setup and evaluation. In the dataset, the method of data generation for training and testing purposes is described. In data pre-processing, the pipeline for bringing the pictures from various sources is portrayed. Similarly, the model setup describes multiple convolution layers for the classification of images. Finally, the results are evaluated and analyzed.
The data were collected from different resources to construct a new dataset. Furthermore, Python visualization libraries were used to visualize our data [27]. The proposed method in this article employs deep neural networks and a supervised learning architecture (CNN) for image detection. The supervised learning is used for model training. After model training, the sample data are tested and verified with the given training data. Moreover, some evaluation techniques are applied for the classification of results. After executing classification techniques, results are classified on the basis of training data. Finally, the model accuracy is measured in comparison to training data.

2.1. Pre-Processing and Enhancement of DR Dataset

Actual pre-trained CNN models are too large to handle the retina images dataset, resulting in overfitting issues. To address this problem, a variation can be introduced to the dataset. Adding variation at the early point (input) of a neural network causes significant changes in the dataset generalization. A variation refers to the fact that the noise addition task augments the dataset in some way. The dataset constraint is one of the critical challenges faced by researchers in the healthcare field. As a result, we have employed some additional augmentation approaches. The retina image dataset was created as follows. After resizing the photos to 224 × 224 × 3, we used the following augmentation methods: random horizontal flip (aids in the detection of DR based on severity level), random resized crop (the last stage of DR, i.e., proliferate) and, last, picture enhancement by altering picture intensities.

2.2. CNN Architecture

Deep neural networks based on CNN models have recently been employed to handle computer vision challenges. To categorize the DR dataset among normal and various levels of DR patients, we employed deep-CNN-model-based AlexNet [28], GoogleNet [29], Inception V4 [30], Inception ResNet V2 [31] and ResNeXt-50 [32] models, as well as transfer learning approaches. Transfer learning may also aid with class imbalance and model execution time. The employed CNN models, as well as AlexNet, GoogleNet, Inception V4, Inception ResNet V2 and ResNeXt-50 models, are presented schematically in Figure 2. Pre-trained models work quite well on a new dataset before being used for classification.
The DTL is a useful approach for solving the issue of unfit training data. The goal of this strategy is to extract the information from a process (issue). The extracted information is then utilized over comparable tasks by overcoming isolated learning issues. This understanding provides an incentive to tackle the problems in a variety of disciplines where the development is hard. It has resulted in insufficient or partial training data. Figure 3 depicts the DTL process.
We utilize nine pre-trained architectures to deal with the retina image dataset, rather than using the long training process from scratch. The weights of existing pre-trained model layers are re-used for model training in a different domain, as illustrated in Figure 4. The DTL methodology has yielded beneficial and significant achievements in a variety of computer vision areas [33,34,35,36]. We used CNN architectural weights that had already been learned. Moreover, the entire model was fine-tuned with some appropriate learning rates.

3. Pre-Trained CNN Architectures and Performance Matrices

We selected five distinct pre-trained CNN architectures: AlexNet [28], GoogleNet [29], Inception V4 [30], Inception ResNet V2 [31] and ResNeXt-50 [32]. These models are used to classify the DR image dataset. In order to modify the classification layer, fine tuning is employed. The fine-tuning process extracts features for the targeted tasks. Since pre-trained models are utilized, only the previously diagnosed diabetic retinopathy images are used to make the models more accurate. The model training process is given as follows:
  • Load the pictures from every type of folder.
  • Use cv2 to resize images in (80, 80) and transmit images to array.
  • Label every picture with type.
  • Transform pictures and labels to numpy array.
  • Split the images in half, and in an 80–20 split, the labels change into category labels.
  • Set parameters of the trained model (e.g., epochs = 100, batch size = 32, etc.).
  • Pickle may be used to save both the model and the label.
  • In the end, we can visualize loss and accuracy.

3.1. AlexNet Architecture

AlexNet is the name of a convolutional neural network that has made a significant contribution to the field of machine learning. This is particularly true for in-depth learning in machine vision. The AlexNet architecture has 5 layers of flexibility, 3 layers of merging, 2 layers of standardization, 2 fully connected layers, and 1 softmax tile. The convolutional filters and the nonlinear activation function ReLU are included in each convolutional layer. Blending layers are used to create a variety of combinations. Due to the existence of completely linked layers, the input size is modified. Convolutional neural networks are a key component of neural networks. They are made up of neurons with a readable weight and bias. Each specific neuron receives a number of inputs. Subsequently, it takes a weight-bearing amount on top of it. Finally, it is transmitted by activating, turning and releasing. The complete architecture of AlexNet is illustrated in Figure 5.

3.2. GoogleNet Architecture

GoogleNet is a 22-level deep congenital neural network. Its salient feature is to work very fast. It has less memory usage and less power consumption. This neural network utilizes the averaged value of global pooling and maximum pooling. For our pre-trained model, it consists of four parallel paths. The inception blocks perform a convolution (1 × 1, 3 × 3, 5 × 5 window sizes) for spatial sizes and information extraction. The ReLU is also included in the convolution layer. The inception block is utilized three times. The first two inception blocks are used for 3 × 3 maximum pooling, while the third is used as a global average pool linked by a thick layer. The complete architecture of GoogleNet is illustrated in Figure 6.

3.3. Inception V4 Architecture

Concerning the deep CNN architectures, Inception is considered for a good performance with a low execution cost. It was initially introduced in [31] as Inception v1. Then, this architecture was improved with the concept of batch normalization to a new variant named Inception v2. Next, factorization was introduced during iterations to form different variants, i.e., Inception v4, Inception ResNet V1 and Inception ResNet V2. Inception v4 is a slightly modified version of Inception v3. The Inception model and Inception ResNet model are residual and non-residual variants; this is the main difference. Moreover, batch normalization is only used on top of the traditional layer rather than residual summations. The architecture of Inception v4 consists of the initial set of layers that were modified to make it uniform. This is referred to as the “stem of the architecture” and is used in front of the Inception block in the architecture. This does not require the partition of the replicas, which enables a training feature. However, the previous versions of the Inception architecture require a replica to fit in the memory. This also reduces the memory requirement because it uses memory optimization during backpropagation. In our paper, we use Inception v4 and Inception ResNet V2. The explanation of Inception ResNet V2 is given in the next section. The complete architecture of Inception V4 is illustrated in Figure 7.

3.4. Inception ResNet V2 Architecture

Inception ResNet V2 is a decisive neural structure built into the Inception family of architectures. It incorporates residual connections (changes the filter concatenation stage of Inception construction). It has an ability to split images into 1000 objects, e.g., mouse, keyboard, pencil. The network has a read rich property to accept presentations of various images. The network has an input image size of 299 × 299. The output is a vector form of measurable probability. The complete build of the network is based on a combination of the original structure and the remaining connections. Moreover, multiple heavy filters are integrated with the remaining connections. The use of residual connections not only prevents degradation (caused by deep structures) but also reduces the training time. The complete architecture of Inception ResNet V2 is illustrated in Figure 8.

3.5. ResNeXt-50 Architecture

ResNeXt-50 uses a squeeze and excitation (SE) block for each non-identity branch of a residual block. It comprises 5 different sections, including convolution and identity blocks. A single convolution block has three layers of convolution and each ID has 3 stages of conversion. The SE block acts as a computational unit that performs transformations from inputs to feature maps. It can be attached with different CNN architectures and residual networks. The SE block is placed before summation, which increases the computational cost. However, it enables ResNext-50 to achieve a higher accuracy as compared to ResNet-50. The complete architecture of Inception ResNet V2 is illustrated in Figure 9.

3.6. Performance Matrices

All of the previously mentioned CCNs were utilized in our experiments. We considered five parameters to evaluate the aforementioned CNN architectures to classify the retina images. All these parameters were calculated using four important terms from the confusion matrix, which are True Positive (TP), True Negative (TN), False Positive (FP) and False Negative (FN). Therefore, the corresponding values for these parameters (accuracy, error rate, precision, recall and Fscore) are given in Equations (1) and (4)–(6), subsequently.
A c c u r a c y = T P + T N F N + T P + T N + F P
R e c a l l / S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
E r r o r = 1 N n = 1 j | y j y j ^ |
R e c a l l = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
The accuracy of the classifier depends on different parameters, as given in Equation (1). Moreover, the rate of sensitivity interprets the ability of a classifier to correctly form the target class, as given in Equation (2). Similarly, the rate of specificity illustrates the capability of a classifier for separation, as shown in Equation (3). The precision rate evaluates the determination of a certain class. Finally, FScore is the harmonic mean sensitivity (recall) and accuracy value as set forth in Equation (6). The analytical average error value may be determined using Equation (5). In our research, all associated evaluation parameters for CNNs were computed. Consequently, the findings are presented in the next section based on the above parameters.

4. Results and Implementation

This section provides the description of the proposed custom dataset, experimental setup and obtained results.

4.1. Creation of Custom Dataset

We created our custom dataset of fundus images to grade the severity level of DR. The proposed approach contrasts with the existing grading (as mentioned in Section 1.1), which grades fundus images based on the pathological changes in the retina. In addition to this, we consider the clinical practice; that is, we categorize a fundus picture of the foundation of abnormalities and the treatment technique. For training and testing, the pictures are divided and placed in different files. A custom script is created to determine the kind of picture based on its tags. The pictures are then cropped and the essential characteristics are separated. Furthermore, a filtering technique is employed to equalize and contrast the picture modification. To increase the variety of data, data augmentation is used. Finally, flipping, cropping and padding are performed. To summarize, the created dataset comprises 1440 images for positive DR patients (high-risk).

4.2. Experimental Setup

We developed some fine-tuned CNN architectures to classify DR pictures. These architectures are AlexNet, GoogleNet, Inception V4, Inception ResNet V2 and ResNeXt-50. Each CNN architecture uses fully connected (FC) layers with a classification criticality of the final FC layer. The number of neurons in the final FC layer is calculated using the target dataset. It is necessary to set and optimize these parameters as the CNN architecture itself is not able to define parameters for the fine-tuning method. Therefore, the parameters are defined using the results of training for the improvement of performance. We utilized the Adam optimizer for the training of every network architecture with 30 epochs (maximum). The batch size and initial learning rate for training and testing are 32, 8 and 1e-5, respectively. We used Python (a programming language) to train the CNN models. All the experiments were executed on an NVIDIA GPU (NVIDIA CUDA Version: 10.1 with Tesla P100) using a Google Colaboratory. PyTorch version 1.5 was utilized to execute experimentation on pretrained CNN models (AlexNet, GoogleNet, Inception V4, Inception ResNet V2 and ResNeXt-50) using weights (This is a random value of initial weights for our pre-trained CNN architectures). The aforementioned CNN architecture takes advantage of hyper-parameters for DTL, as shown in Table 2.

4.3. Results and Analysis

The proposed methodology in this work was evaluated using the performance metrics (see Section 3.6), which were calculated during experiments. This methodology also explores the fine-tuning of DTL, which includes the extraction of the features from pretrained CNN networks. The experimental study was conducted using our custom dataset which is completely based on 4 publicly available datasets as described in Section 1.4. The complete experimental process was based on five pre-trained CNN networks, i.e., AlexNet, GoogleNet, Inception V4, Inception ResNet V2 and ResNeXt-50.
The process starts by evaluating the accuracy for a multi-class dataset. Next, k-fold is utilized to evaluate the average classification accuracy. The value of the average accuracy is calculated using the values of individual accuracy. The results illustrate the exploration of the DTL method using feature extraction on pre-trained CNN networks. The results are listed in Table 3 using our custom dataset (The performance of the experiments was obtained using finetuned and pre-trained architectures for all k-folds). It is noteworthy that the CNN classifies the images and reports a confusion matrix for each severity level of disease.
Table 3 for different k-folds indicates the accuracy value for the employed pre-trained models. The most accurate ResNeXt-50 model is 97.53% for fold-5 (fold-1 and fold-2 also reach the same accuracy). It is clear that the accuracy of the model increases when we increase the value of the fold. Our unique data package achieves the highest precision of 95.98% for fold-5 and the lowest precision of 84.01% for fold-1 for AlexNet models. The same precision is achieved by fold-3, fold-4 and fold-5, for Alexnet and GoogleNet. As regards the maximum individual accuracy of the pretrained models, we have AlexNet: 87.84%, GoogleNet: 91.16%, V4: 90.31%, ResNet V2: 91.61% and ResNeXt-50: 97.53%. For all AlexNet and GoogleNet models pretrained we observe that V4, ResNet V2 and ResNeXt-50 achieve comparable accuracy after fold-3; GoogleNet after fold-1, changing after fold-5; ResNet V4 after fold-1, changing after fold-5; ResNet V2 for fold-2 and fold-3; and ResNeXt-50 for fold-1, fold-2 and fold-5.

4.4. Comparison and Discussion

The comparison of our evaluation with the recent publications is presented in Table 4. We found two recent publications and analyzed our implementation results. The authors in [15] evaluated Alexnet and ResNet architectures using a small dataset. This collection includes 4476 pictures from three clinical departments of the Sichuan Provincial People’s Hospital. The authors in [11] implemented Inception V4 and ResNeXt-50 architectures with a dataset that is moderate in size. They utilized a pre-processing pipeline that converts a set of fundus images into a uniform format. They used a modified version of the Inception-V3 network. Subsequently, the performance is compared with several mainstream CNN models. The authors in [11] classified the dataset into five different classes with AlexNet and ResNet. The utilized dataset of 35,126 images exercised three types of ensembles.
In our work, we implemented several mainstream CNN architectures and evaluated five different parameters with DTL using our custom data. The authors in [24] exploited a modified version of Lenet-5 architectures that is quite old. However, in our study, we exploited recent CNNs. Regarding evaluation scores, the proposed methodology efficiently classifies images in different classes as DR, normal, mild, moderate, severe and high-risk. From our analysis, we conclude that the fine-tuning of a pre-trained CNN architecture with DTL could be employed as one of the efficient techniques in the medical field for the classification of DR images.
It is noteworthy that high-risk DR patients lie in the proliferate category. They require an immediate cure and diagnosis. In our diagnosis method, we exploit DR images of patients that reflect the posterior pole. Using a high resolution ultimately elevates the size of the dataset, which may increase the execution time for the classification. However, the use of low-resolution DR images could affect classification due to the lack of clear media. In this research work, we used high-resolution datasets from different sources as mentioned in Section 1.4. To handle this issue, we executed pre-processing of data where the DR images were resized and augmentation was performed for the picture enhancement. Last but not least, there is still a possibility of an error during classification that could be reduced, but the accuracy of the dataset reflects the correctness of the diseases. From the comparison of the state-of-the-art, we conclude that our datasets achieve better accuracy and this was the main goal of this research work.

5. Conclusions and Future Work

In this article we have provided a deep-transfer-learning technique, based on convolutional neural networks, for the categorization of diabetic retinopathy patients. In order to investigate the proposed deep-transfer-learning technique, five pre-trained convolutional neural network models were employed. It was observed that the fine tweaking of pre-trained models may be used effectively on a multi-class dataset. As a result, the diagnosis efficiency for diabetic retinopathy patients has been improved. Across all the five employed models, the ResNeXt-50 architecture achieved a maximum accuracy of 97.53 % for our dataset. Our high-accuracy findings can assist doctors and researchers in making clinical judgments. Our work includes a few limitations that can be addressed in future research. A more in-depth examination is needed, which requires more patient data. Future study should also focus on differentiating the accuracy of individuals with normal symptoms from those with non-proliferate symptoms. The non-proliferate symptoms may not be properly visible on retina images, or may not be visualized at all. Another probable direction is to apply the proposed method to larger datasets. It may address other medical problems such as cancer and tumors, as well being applicable in other computer vision industries such as energy, agriculture and transportation.

Author Contributions

Conceptualization, H.T. and M.R.; methodology, M.R. and E.Z.; software, A.J. and E.Z.; validation, H.T. and A.J.; formal analysis, M.R. and M.Y.I.Z.; investigation, A.J. and S.S.A.; resources, M.R. and M.Y.I.Z.; data curation, H.T.; writing—original draft preparation, H.T. and A.J.; writing—review and editing, M.R. and M.Y.I.Z.; visualization, E.Z.; supervision, M.R. and S.S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mayo Clinic. Diabetic Retinopathy. 2018. Available online: https://www.mayoclinic.org/diseases-conditions/diabetic-retinopathy/symptoms-causes/syc-20371611 (accessed on 20 September 2021).
  2. Lee, R.; Wong, T.Y.; Sabanayagam, C. Epidemiology of diabetic retinopathy, diabetic macular edema and related vision loss. Eye Vis. 2015, 2, 1–17. [Google Scholar] [CrossRef] [Green Version]
  3. National Eye Institute. Facts about Diabetic Eye Disease; National Eye Institute: Bethesda, MD, USA, 2020. Available online: https://www.nei.nih.gov/learn-about-eye-health/eye-conditions-and-diseases/diabetic-retinopathy (accessed on 22 September 2021).
  4. Michael, W.S. Treatment of diabetic retinopathy: Recent advances and unresolved challenges. World J. Diabetes 2016, 7, 333–341. [Google Scholar] [CrossRef]
  5. National Eye Institute. Diabetic Retinopathy Data and Statistics; National Eye Institute: Bethesda, MD, USA, 2019. Available online: https://www.nei.nih.gov/learn-about-eye-health/resources-for-health-educators/eye-health-data-and-statistics/diabetic-retinopathy-data-and-statistics (accessed on 27 September 2021).
  6. Centers for Disease Control and Prevention. National Diabetes Statistics Report; Centers for Disease Control and Prevention: Atlanta, GA, USA, 2020. Available online: https://www.cdc.gov/diabetes/data/statistics-report/index.html (accessed on 1 October 2021).
  7. World Health Organization. Diabetes; World Health Organization: Geneva, Switzerland, 2021; Available online: https://www.who.int/health-topics/diabetes (accessed on 1 October 2021).
  8. Cheloni, R.; Gandolfi, S.A.; Signorelli, C.; Odone, A. Global prevalence of diabetic retinopathy: Protocol for a systematic review and meta-analysis. BMJ Open 2019, 9, e022188. [Google Scholar] [CrossRef]
  9. American Optometric Association. Diabetic Retinopathy; American Optometric Association: New York, NY, USA, 2021; Available online: https://www.aoa.org/healthy-eyes/eye-and-vision-conditions/diabetic-retinopathy (accessed on 2 October 2021).
  10. Khan, W. Diabetic Retinopathy Detection using Image Processing: A Survey. Int. J. Emerg. Technol. Res. 2013, 1, 16–20. [Google Scholar]
  11. Kumar, S. Diabetic Retinopathy Diagnosis with Ensemble Deep-Learning. In Proceedings of the 3rd International Conference on Vision, Image and Signal Processing, Vancouver, BC, Canada, 26–28 August 2019; Association for Computing Machinery: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  12. Prasad, D.K.; Vibha, L.; Venugopal, K.R. Early detection of diabetic retinopathy from digital retinal fundus images. In Proceedings of the 2015 IEEE Recent Advances in Intelligent Computational Systems (RAICS), Trivandrum, India, 10–12 December 2015; pp. 240–245. [Google Scholar] [CrossRef]
  13. Zeng, X.; Chen, H.; Luo, Y.; Ye, W. Automated Diabetic Retinopathy Detection Based on Binocular Siamese-Like Convolutional Neural Network. IEEE Access 2019, 7, 30744–30753. [Google Scholar] [CrossRef]
  14. Hemanth, D.J.; Deperlioglu, O.; Kose, U. An enhanced diabetic retinopathy detection and classification approach using deep convolutional neural network. Neural Comput. Appl. 2020, 32, 707–721. [Google Scholar] [CrossRef]
  15. Gao, Z.; Li, J.; Guo, J.; Chen, Y.; Yi, Z.; Zhong, J. Diagnosis of Diabetic Retinopathy Using Deep Neural Networks. IEEE Access 2019, 7, 3360–3370. [Google Scholar] [CrossRef]
  16. Junjun, P.; Zhifan, Y.; Dong, S.; Hong, Q. Diabetic Retinopathy Detection Based on Deep Convolutional Neural Networks for Localization of Discriminative Regions. In Proceedings of the 2018 International Conference on Virtual Reality and Visualization (ICVRV), Qingdao, China, 22–24 October 2018; pp. 46–52. [Google Scholar] [CrossRef]
  17. Chetoui, M.; Akhloufi, M.A.; Kardouchi, M. Diabetic Retinopathy Detection Using Machine Learning and Texture Features. In Proceedings of the 2018 IEEE Canadian Conference on Electrical Computer Engineering (CCECE), Quebec, QC, Canada, 13–16 May 2018; pp. 1–4. [Google Scholar] [CrossRef]
  18. Samanta, A.; Saha, A.; Satapathy, S.C.; Fernandes, S.L.; Zhang, Y.D. Automated detection of diabetic retinopathy using convolutional neural networks on a small dataset. Pattern Recognit. Lett. 2020, 135, 293–298. [Google Scholar] [CrossRef]
  19. Decencière, E.; Cazuguel, G.; Zhang, X.; Thibault, G.; Klein, J.C.; Meyer, F.; Marcotegui, B.; Quellec, G.; Lamard, M.; Danno, R.; et al. TeleOphta: Machine learning and image processing methods for teleophthalmology. IRBM 2013, 34, 196–203. [Google Scholar] [CrossRef]
  20. Staal, J.; Abramoff, M.D.; Niemeijer, M.; Viergever, M.A.; van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef]
  21. Hoover, A.D.; Kouznetsova, V.; Goldbaum, M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 2000, 19, 203–210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Grissworld Home Care. The 4 Stages of Diabetic Retinopathy: What You Can Expect. 2019. Available online: https://www.griswoldhomecare.com/blog/2015/january/the-4-stages-of-diabetic-retinopathy-what-you-ca/ (accessed on 3 October 2021).
  23. Gupta, A.; Chhikara, R. Diabetic Retinopathy: Present and Past. Procedia Comput. Sci. 2018, 132, 1432–1440. [Google Scholar] [CrossRef]
  24. Wang, H.; Yuan, G.; Zhao, X.; Peng, L.; Wang, Z.; He, Y.; Qu, C.; Peng, Z. Hard exudate detection based on deep model learned information and multi-feature joint representation for diabetic retinopathy screening. Comput. Methods Prog. Biomed. 2020, 191, 105398. [Google Scholar] [CrossRef]
  25. Amin, J.; Sharif, M.; Yasmin, M. A Review on Recent Developments for Detection of Diabetic Retinopathy. Scientifica 2016, 132, 1432–1440. [Google Scholar] [CrossRef] [Green Version]
  26. EyePACS. News, EyePACS. 14 June 2019. Available online: http://www.eyepacs.com/blog/news (accessed on 3 October 2021).
  27. California Healthcare Foundation. Diabetic Retinopathy Detection; California Healthcare Foundation: Oakland, CA, USA, 2020; Available online: https://www.kaggle.com/c/diabetic-retinopathy-detection/data (accessed on 4 October 2021).
  28. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  29. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  30. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  31. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  32. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar] [CrossRef] [Green Version]
  33. Butt, C.; Gill, J.; Chun, D.; Babu, B.A. RETRACTED ARTICLE: Deep learning system to screen coronavirus disease 2019 pneumonia. Appl. Intell. 2020. [Google Scholar] [CrossRef]
  34. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How Transferable Are Features in Deep Neural Networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar]
  35. Dai, W.; Chen, Y.; Xue, G.R.; Yang, Q.; Yu, Y. Translated Learning: Transfer Learning across Different Feature Spaces. In Proceedings of the 21st International Conference on Neural Information Processing Systems, Kuching, Malaysia, 3–6 November 2014; Curran Associates Inc.: Red Hook, NY, USA, 2008; pp. 353–360. [Google Scholar]
  36. Ravishankar, H.; Sudhakar, P.; Venkataramani, R.; Thiruvenkadam, S.; Annangi, P.; Babu, N.; Vaidya, V. Understanding the Mechanisms of Deep Transfer Learning for Medical Images. In Deep Learning and Data Labeling for Medical Applications; Carneiro, G., Mateus, D., Peter, L., Bradley, A., Tavares, J.M.R.S., Belagiannis, V., Papa, J.P., Nascimento, J.C., Loog, M., Lu, Z., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 188–196. [Google Scholar]
Figure 1. Proposed framework for the detection of DR.
Figure 1. Proposed framework for the detection of DR.
Sensors 22 00205 g001
Figure 2. Schematics of CNN model for the detection of different DR stages.
Figure 2. Schematics of CNN model for the detection of different DR stages.
Sensors 22 00205 g002
Figure 3. The adopted DTL process.
Figure 3. The adopted DTL process.
Sensors 22 00205 g003
Figure 4. The DTL with pre-trained and learnable weights.
Figure 4. The DTL with pre-trained and learnable weights.
Sensors 22 00205 g004
Figure 5. The pre-trained architecture of AlexNet.
Figure 5. The pre-trained architecture of AlexNet.
Sensors 22 00205 g005
Figure 6. The pre-trained architecture of GoogleNet.
Figure 6. The pre-trained architecture of GoogleNet.
Sensors 22 00205 g006
Figure 7. The pre-trained architecture of Inception V4.
Figure 7. The pre-trained architecture of Inception V4.
Sensors 22 00205 g007
Figure 8. The pre-trained architecture of Inception ResNet V2.
Figure 8. The pre-trained architecture of Inception ResNet V2.
Sensors 22 00205 g008
Figure 9. The pre-trained architecture of ResNeXt-50.
Figure 9. The pre-trained architecture of ResNeXt-50.
Sensors 22 00205 g009
Table 1. Different stages of diabetic retinopathy with the passage of time [9].
Table 1. Different stages of diabetic retinopathy with the passage of time [9].
StageNormalNon-ProliferateProliferate
Years03–55–1010–15>15
Type of DRN/AMildModerateSevereHigh-risk
Condition of retinaHealthyA few tiny bulges in the blood vesselsLittle lumps in the veins with noticeable spots of blood spillage that stores the cholesterol.Larger areas of blood leakage. Beading in veins that is unpredictable. The formation of new blood vessels at the optic circle. Vein occlusion.High bleeding and the formation of new blood vessels elsewhere in the retina. Complete blindness.
Table 2. The setting of parameters for CNN architectures.
Table 2. The setting of parameters for CNN architectures.
ParametersAlexNetGoogleNetInception V4Inception ResNet V2ResNeXt-50
OptimizerADAMADAMADAMADAMADAM
Base learning rate1e-51e-51e-51e-51e-5
Learning decay rate0.10.10.10.10.1
Momentum β 1 0.90.90.90.90.9
RMSprop β 2 0.9990.9990.9990.9990.999
Dropout rate0.50.50.50.50.5
# of epochs3030303030
Train batch size3232323232
Test batch size88888
Total number of parameters60 M4 M43 M56 M27.56 M
Table 3. Results and performance obtained using pre-trained CNN architectures.
Table 3. Results and performance obtained using pre-trained CNN architectures.
ClassifierFoldsTPTNFPFNAccuracy (%)Specificity (%)Precision (%)Recall (%)Fscore (%)
AlexNetF137210351284.0185.7151.3875.5161.15
F238210371283.5085.0250.6676.060.80
F33821427887.8088.7958.4682.6068.46
F43721627887.8488.8857.8182.2267.89
F53721627887.8488.8857.8182.2267.89
GoogleNetF13821922789.8690.8763.3384.4472.38
F24022219790.9792.1167.7985.1075.47
F33822118890.8792.4667.8582.6174.51
F43722018890.8192.4367.2782.2274.00
F53822018791.1692.4367.8584.4475.24
Inception V4F13922421790.3791.4265.0084.7873.58
F23922417891.3292.9469.6482.9775.72
F33922516891.6693.3670.9082.9776.47
F43922618891.0692.6268.4282.9875.00
F53922220890.3191.7366.1082.9873.58
Inception ResNet V2F14022018691.5592.4468.9686.9676.92
F24022114692.8894.0474.0786.9680.00
F34022714792.7194.1974.0785.1179.21
F44122613593.6894.5675.9289.1382.00
F53922318691.6192.5368.4286.6776.47
ResNeXt-50F1412338595.4796.6883.6789.1386.31
F2412347595.8297.0985.4189.1387.23
F3422346496.5097.5087.5091.3089.36
F4422365397.2097.9289.3693.3391.30
F5412365297.5397.9289.1395.3592.13
Table 4. Comparison with state-of-the-art classifiers.
Table 4. Comparison with state-of-the-art classifiers.
ClassifiersAlexnetInception V4ResNet/ResNeXt-50
Acc (%)Pre (%)Rec (%)Acc (%)Pre (%)Rec (%)Acc (%)Pre (%)Rec (%)
Our Work87.8457.8182.2290.3166.1082.9897.5389.1395.35
S. Kumar et al. [11]60.1055.70
Z. Gao et al. [15]88.7295.77948487.6195.7695.52
Acc: Accuracy, Pre: Precision, Rec: Recall.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tariq, H.; Rashid, M.; Javed, A.; Zafar, E.; Alotaibi, S.S.; Zia, M.Y.I. Performance Analysis of Deep-Neural-Network-Based Automatic Diagnosis of Diabetic Retinopathy. Sensors 2022, 22, 205. https://doi.org/10.3390/s22010205

AMA Style

Tariq H, Rashid M, Javed A, Zafar E, Alotaibi SS, Zia MYI. Performance Analysis of Deep-Neural-Network-Based Automatic Diagnosis of Diabetic Retinopathy. Sensors. 2022; 22(1):205. https://doi.org/10.3390/s22010205

Chicago/Turabian Style

Tariq, Hassan, Muhammad Rashid, Asfa Javed, Eeman Zafar, Saud S. Alotaibi, and Muhammad Yousuf Irfan Zia. 2022. "Performance Analysis of Deep-Neural-Network-Based Automatic Diagnosis of Diabetic Retinopathy" Sensors 22, no. 1: 205. https://doi.org/10.3390/s22010205

APA Style

Tariq, H., Rashid, M., Javed, A., Zafar, E., Alotaibi, S. S., & Zia, M. Y. I. (2022). Performance Analysis of Deep-Neural-Network-Based Automatic Diagnosis of Diabetic Retinopathy. Sensors, 22(1), 205. https://doi.org/10.3390/s22010205

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop