Next Article in Journal
A Pathfinding Algorithm for Lowering Infection Exposure of Healthcare Personnel Working in Makeshift Hospitals
Next Article in Special Issue
Text Mining and Determinants of Sentiments towards the COVID-19 Vaccine Booster of Twitter Users in Malaysia
Previous Article in Journal
Persistent Changes of Peripheral Blood Lymphocyte Subsets in Patients with Oral Squamous Cell Carcinoma
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of COVID-19 Based on Chest X-rays Using Deep Learning

1
Department of Computer Engineering and Network, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Aljouf, Saudi Arabia
2
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Aljouf, Saudi Arabia
3
School of Computer Science and Engineering (SCE), Taylor’s University, Subang Jaya 47500, Selangor, Malaysia
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(2), 343; https://doi.org/10.3390/healthcare10020343
Submission received: 14 December 2021 / Revised: 27 January 2022 / Accepted: 30 January 2022 / Published: 10 February 2022
(This article belongs to the Special Issue Artificial Intelligence in the Management of the Pandemic)

Abstract

:
The coronavirus disease (COVID-19) is rapidly spreading around the world. Early diagnosis and isolation of COVID-19 patients has proven crucial in slowing the disease’s spread. One of the best options for detecting COVID-19 reliably and easily is to use deep learning (DL) strategies. Two different DL approaches based on a pertained neural network model (ResNet-50) for COVID-19 detection using chest X-ray (CXR) images are proposed in this study. Augmenting, enhancing, normalizing, and resizing CXR images to a fixed size are all part of the preprocessing stage. This research proposes a DL method for classifying CXR images based on an ensemble employing multiple runs of a modified version of the Resnet-50. The proposed system is evaluated against two publicly available benchmark datasets that are frequently used by several researchers: COVID-19 Image Data Collection (IDC) and CXR Images (Pneumonia). The proposed system validates its dominance over existing methods such as VGG or Densnet, with values exceeding 99.63% in many metrics, such as accuracy, precision, recall, F1-score, and Area under the curve (AUC), based on the performance results obtained.

1. Introduction

Around the world, COVID-19 is wreaking havoc on people’s lives and healthcare systems. It is a new virus strain discovered in 2019 that has never been seen by humans before. The first COVID-19-positive case was discovered in Wuhan, China, in December 2019, and it quickly spread to a number of other Chinese cities as well as several other countries around the world [1,2]. According to preliminary polls, COVID-19 causes minor symptoms in about 99%, while the remainder of cases are serious or critical. The number of people dying from pneumonia caused by the COVID-19 virus is rising every day [3].
The rapid global spread of COVID-19 put healthcare systems under tremendous pressure; this spread could be significantly slowed if a reliable screening method for patients with COVID-19 infections is established. Doctors and researchers found themselves facing a daunting challenge to find ways to diagnose the disease quickly [4]. A COVID-19 infection can cause serious problems such as acute kidney failure, septic shock, heart attack, and pulmonary edema [5]. The early detection and isolation of patients with infection is critical in combating and addressing the COVID-19 pestilence [6,7]. The prevalence of reported COVID-19 occurrence in the most affected nations around the world is depicted in Figure 1. The United States leads the world in terms of reported illnesses, accounting for 63,390,876 cases out of a total of 185,039,249 cases.
The most common COVID-19 detection technique is real-time polymerase chain reaction (RT-PCR). It has a high percentage of false-negative findings and may take up to two days to receive results, while having a sensitivity range of 70 to 90 [9]; it may also produce a quite high number of false-negative effects and may take up to two days to obtain results. In some countries, it may take up to five days or more due to the overwhelming number of tests that need to be analyzed [4].
Additionally, COVID-19 is detected and diagnosed using radiological screening tests such as CXR and computed tomography (CT). It has been noticed that CXR is one of the most effective methods for diagnosing pneumonia around the world because it is a rapid, inexpensive, and popular clinical method that exposes the patient to less radiation than CTs [10,11].
However, radiologists are needed to look for the radiological signs that show COVID-19 symptoms on a CXR. To save time and effort, it is important to automate the CXR analysis, which is a long and error-prone process that takes a lot of time and effort [12].
Figure 2 shows a CXR scan image and a CT scan image.
As a result, fully automated and real-time radiography image analyses are required to assist physicians in accurately detecting COVID-19 infection. Physicians may use computer-aided diagnosis (CAD) systems based on DL methods to help them perceive and understand the information in CXR images as well as to overcome the limitations of the imaging acquisition techniques used, rapidly and correctly. DL methods are becoming more common in medical imaging because of their ability to deal with massive datasets that surpass human capabilities. Combining CAD techniques with radiologist medical diagnostics decreases physicians’ stress as well as improves their accuracy and statistical analysis [11].
This paper propose a crossbred DL system for COVID-19 classification and prognosis that uses two unique DL approaches to accurately detect early COVID-19 symptoms from CXR images. The proposed system has two significant phases: preprocessing and classification. The preprocessing phase is used to improve the overall contrast of the image in order to reduce inconsistencies between images obtained from various X-ray devices. The image is also resized and normalized to suit the size of the training model throughout that process. The classification stage, on the other hand, involves a variety of classifiers, and the most effective classifier are chosen based on the classification error for each case.
The following are the key contributions of this research:
1.
To determine the feasibility of the proposed scheme, detailed comparative analyses are conducted using various measurement criteria such as accuracy, precision, recall, specificity, F1-score, and AUC.
2.
COVID-19 shows radiological indications that are readily detectable on CXR. As a result, DL-based methods can be used to automatically analyze CXR, significantly reducing the analysis time.
3.
The COV-PEN dataset is developed, which is a large-scale CXR image dataset. Among those currently publicly available, it includes a large number of CXR images with reported COVID-19 disease.
4.
To fine-tune the weights of pre-trained networks on small datasets as well as to train the weights of networks on large datasets, a modified version of Resnet-50 is used.
5.
To improve the generalized effectiveness of the suggested method and to prevent over-fitting, a different training protocol assisted by different combinations of training policies (e.g., validation patience and data augmentation) is used.
The following are the remaining parts of this paper: Section 2 summarizes recent related articles, Section 3 explains the methodology used to create the COVID-19 dataset and the proposed system’s design requirements, Section 4 introduces the study findings, and Section 5 ends with the conclusion and possible research opportunities.

2. Related Works

This section provided an overview of some related studies for a better understanding of the area under study and to provide the state-of-the-art picture. Convolutional neural networks (CNNs), which are one of the most effective DL models, have successfully proved their mastery over conventional methods in several disciplines, including image classification and pattern recognition [13,14]. Currently, it has indeed been successfully implemented in the field of medicine with impressive outcomes and outstanding performance in different challenging settings. Various medical imaging systems using DL techniques have also been developed to assist physicians and specialists in effective COVID-19 diagnosis, care, and follow-up examination [15,16]. Narin et al. [11] used five-fold cross validation to enforce various binary classifications. With an accuracy equal to 98%, specificity value of 100%, and a recall with 96%, the pre-trained ResNet-50 method gives the best efficiency. On the other hand, Wang et al. [17] have suggested using CXR images to automatically establish a new deep architecture called COVID-Net to detect COVID-19 instances. Using a database containing 13,975 CXR images, this model has the highest classification accuracy of 93.3%. The key strength of this approach is that the conceptual composition could create a balance between different goals such as accuracy and computational costs through architectural design choices. Hemdan et al. [18] introduced COVIDXNet, a DL framework for detecting COVID-19 infections in CXR images. A small dataset of 50 images was used to compare seven DL techniques (e.g., MobileNetV2, ResNetV2, VGG19, DenseNet201, InceptionV3, Inception, and Xception). DenseNet201 had the best performance, with a 91% accuracy score. While Zhang et al. [19] derived useful feature representations from CXR image using ResNet-18 model as a feature vector. Those derived features were then entered as an input into a multi-layer perception. A dataset of hundred images taken from seventy patients yielded the highest accuracy rate of 96%. A further supervised transfer-learning method for COVID-19 infection detection in CXR using an extreme version of the Xception model was developed by Das et al. [12], which achieved accuracy of 97.4%. Furthermore, Ozturk et al. [20] introduced a new system for COVID-19 identification using CXR for automatic detection. It was created to provide consistent and reliable diagnostics for multi-class classifications (COVID-19, mild, and pneumonia) and binary classifications (COVID vs. non-COVID). Using the DarkNet model, they were able to achieve a classification performance of 98.08% for binary classification and 87.02% for the classification of multi-class.
Many studies have tried to find COVID-19 infections in CXR images by using different DL methods [21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37], as indicated in Table 1. The investigation of COVID-19 identification and diagnostic systems that rely on CXR images indicated that there are still a number of vulnerabilities that need additional investigation. For starters, the majority of current systems have been validated with limited CXR datasets as well as a small presence of positive COVID-19 cases. The size of the datasets is insufficient to indicate the true output of the proposed systems. Furthermore, despite the fact that several researchers have achieved high reliability values using pre-trained models through transfer-learning, there has been little focus on developing and training a customized DL model from scratch due to a shortage of a large dataset including a substantial number of CXR images with reported COVID-19 infection.
Eventually, almost all of those studies only concentrated on training DL models used on original images instead of preprocessed images, restricting the capacity of the last classification network to generalize. To address these issues, the current study develops a lightweight COVID-19 detection system that alters the architecture of pre-trained models by inserting several layers, resulting in an optimized proposed system with greater satisfaction.

3. Proposed System

Figure 3 depicts the schematic methodology for the COVID19 detection system, which requires retraining a transfer DL approach (Resnet-50) over preprocessed images in the image datastore to learn discriminative and useful feature representations as illustrated in Algorithm 1. First, the procedure for constructing the datastore is described briefly. The proposed system’s implementation specifics are then discussed, including the proposed preprocessing algorithms, the main design, and the adopted approach’s training methodology.
Algorithm 1 Proposed System steps
  • Let ζ ϵ = chest X-ray, α =Augmentation, pp= pre-Processing, i= image, IEA= Image enhancement algorithm, r= rotation, s=scaling, rf=reflection, sm= shifting methods
  • Step 1: Read( ζ ϵ )
  • Step 2: α ( i m a g e ) w.r.t. r, s, rf, sm
  • Step 3: Perform (pp (i))
  •     3.1. Apply(IEA)
  •     3.2. Resize (i)/224*24*3
  •     3.3. Normalize pixelvalue(i)/interval [0, 1]
  •         3.3.1. Conversion
  •         3.3.2. Calculation (mean)
  •         3.3.3. Scaling(i)
  •         3.3.4. Convesrionback
  •     3.4. Split (dataset)/training, testing
  •     3.5. Extract(features)/Resnet-50 pre-trained model
  •     3.6. Optimize (Freeze layers, epochs, learning weights,
  •     batch size)/optimization methods
  • Step 4: Calculate VPM (accuracy, confusion matrix, ROC, AUC, Precision, recall, F1)
  • Step 5: Comparison (recent studies)

3.1. COV-PEN Image Datasets

Data are at the heart of DL, and it is used as a fuel for these learning models. COVID-19 is a novel disease, and a plethora of datasets are currently available. In this work, we gathered CXR images from two publicly accessible image databases of reported infected cases to create a dataset. There are 2790 CXR images in the dataset used to train and test the proposed system, which we refer to as COV-PEN. To build the COV-PEN dataset, we combined two publicly available data repositories:
  • COVID-19 Image Data Processing [41].
  • CXR Images (Pneumonia) [42].
COVID-19 X-ray files were created by Joseph et al. [41] and are available in an open source Github repository. The authors gathered radiology photographs from various authentic records of COVID-19 incidents for analysis purposes, and most COVID-19 studies use images from this source. A free archive of COVID-19 cases of CXR or CT images is available in the registry, which is maintained on a regular basis. At the time of publication, the archive had about 930 COVID-19 chest radiography files. Pneumonia and standard CXR images were gathered from the Kaggle list “CXR Images (Pnemonia)” [42]. There are 1583 mild cases and 4273 pneumonia cases in the dataset. The COV-PEN dataset, in particular, includes 2790 CXR images from these two sources. Figure 4 summarizes some CXR image samples from the COV-PEN dataset, showing the variety of patient cases in the dataset. Since both databases are open access and publicly accessible to the scientific community and the general public, they were chosen to create COV-PEN.

3.2. Image Preprocessing Step

This step includes data augmentation, image enhancement, image rescaling, and normalization, among other things. Since the model’s network becomes more sophisticated, the number of parameters to learn increases as well, leading to overfitting. After dividing the COV-PEN dataset into three mutually exclusive sets (e.g., preparing, verification, and evaluating sets) to overcome the overfitting issue created by the small number of training photos, data augmentation was used to prevent skewed prediction outcomes. Augmented images with corresponding masks such as rotation, reflection, shifting, and scaling are generated for each image in the dataset. The accuracy of a raw CXR image produced by an electronic detector is simply inadequate, reducing the availability for detection and diagnosis. To improve the quality of CXR images, image enhancement techniques should be used. Furthermore, training deep neural networks (DNNs) on top of preprocessed images rather than raw image data will significantly reduce the DNNs’ generalization error and training time. As a result, an appropriate image enhancement technique was proposed to improve the low quality of the CXR image before feeding it into the proposed system. First, the CXR image’s small details, textures, and low contrast were improved using adaptive contrast enhancement based on redistribution of the input image’s lightness values by taking the image as an input and to give out an enhanced image based on redistributing the histogram of the image, as shown in Figure 5. As a consequence, this approach improves the visibility of the edges and curves in each part of an image while also enhancing the image’s local contrast. Since the images in the dataset come from multiple datasets and could also come from various cameras, the parameters of image acquisition often vary because a portion of images have small pixel sizes and all of the images must be rescaled. Therefore, there are significant changes in the image’s brightness and size. Moreover, the images in the Kaggel dataset are grayscale; we must replicate the image three times to obtain an RGB image. Most of the images in the CXR image dataset almost certainly originated from various acquisition devices, each with its own set of requirements. The intensity of the pixel of each image can vary significantly, so the pixels intensity of all images is normalized between [−1, 1] to ensure that the data are within specific ranges and noise is removed. Normalization has the benefit of ensuring that the model is less vulnerable to slight variations in weights, making it easier to optimize.

3.3. Proposed Transfer Learning for COV-PEN Detection

The proposed system’s main architecture is based on the Resnet-50 model. The massive number of structures and hyper-parameters to be determined is the most difficult challenge when using DL models (e.g., learning rate, number of batch size, number of frozen layers, and number of epochs, etc.). The effects of various hyper-parameters value on the performance of the proposed systems is investigated. In this section, we describe in detail the potential solution based on a modified version of one of the Resnet-50 [43] model. In 2015, He K. et al. [43] developed Resnet-50, a residual learning component to the CNN architecture. A standard layer with a skipped connection compensates the residual unit. The skip connection enables a layer’s input signal to traverse the network by linking it to that layer’s output. As a result of the residual units, an extremely deep 152-layer model was trained, which won the 2015 LSVRC2015 competition. Its innovative residual structure allows for a more straightforward gradient flow and more efficient training. It has a top-five error rate of less than 3.6 percent. ResNet has 34, 50, and 101 layers in other versions. In this work, we investigate two modified versions of the Resnet-50 model as well as the original model, which are illustrated in Figure 6. The original Resnet-50 model is shown in Figure 6a. In order to build the proposed two versions, we modify the latest layers by adding one fully connected (FC) Layer with a size of 512 and two FC layers with sizes of 2048 and 1024. We also replace the original FC layer and softmax layer in both versions, as shown in Figure 6b and Figure 6c, respectively. The original layers of the Resnet-50 model are pre-trained on the ImageNet dataset [44]. Consequently, initially, the new added layers are assigned random weights. Then, during training, all model weights are updated using the back-propagation algorithm, which is the main algorithm for training neural network models.
Experiments using Resnet-50 without adding an additional FC was performed and the results were very poor, so that we took this trend of adding more FC at the end of the Resnet. In the first modified version of Resnet-50, shown in Figure 6b, the first FC was replaced with a new FC layer with a size of 512 and one FC layer with a size of 3; the number of classes was added after the replaced FC layer and before softmax layer, which also has been replaced with a new softmax layer. Based on what was mentioned by Basha, S.S et al. [45], when dealing with small datasets, the network needs more FC layers than when dealing with larger datasets. Any neuron from the previous layer is connected to every other neuron in the next layer in the fully connected layer, and each value contributes to predicting how well a value fits a given class. The output of the final FC layer is then redirected to an activation function, which calculates the class scores. One of DNN’s most common classifiers is Softmax which computes the probability distribution of the n output groups through its equations. The only drawback to adding a single FC layer is that it is extremely computationally intensive.
In the second modified version depicted in Figure 6c, we added two FC layers connected between the first FC layer and the softmax layer, which also has been replaced by a new FC layer and a new softmax layer, respectively. The size of the first FC layer is 2048, the size of the second FC layer is 1024, and the size of the third FC layer is 3. We use batch normalization because it is effective at combating network overfitting because of the fact that overfitting occurs when the model learns the training data extremely well but does not generalize well to other testing data. It is a common problem in DL models, and the risk of falling into this problem increases in situations where the training dataset is small, which is the case in this study. DNNs algorithms always produce results with a degree of variability [46] because, in such algorithms, many steps involve a degree of randomness. Thus, one way to improve the performance of DNNs algorithms is to use ensemble learning. Ensemble has multiple definitions, one of this definitions used here is to run the network for n times using the same network parameters (epochs, batch size, optimizer, etc.). In our work, we propose to implement stacked generalization by performing multiple training runs of the same model, which we refer to this as the multiple-runs ensemble.

4. Experimental Results

Several sufficiently large experiments were performed on the COV-PEN dataset to demonstrate the efficiency of the proposed DL systems and to equate their results to the existing state-of-the-art approaches. The proposed system’s code was written throughout MATLAB R2020b and evaluated on a Windows 10 machine with a Core i7-4650U CPU and 8 GB of RAM. All tests were carried out using an 80 percent random array of CXR images as a training collection for the proposed DL systems, according to the proposed training scheme. During the learning process, ten percent of the training data were chosen at random and used as a validation set to assess their abilities and to save the weight combinations with the highest accuracy value. The proposed framework is pre-trained on the COV-PEN dataset using the Adam and sigmoid optimizer with a learning rate strategy that decreases the learning rate when learning becomes stagnant for a period (i.e., validation patience). The following hyper-parameters were used for training in the Adam optimizer: Number of epochs = 15; batch size varying from 32 to 128, with a move of double its previous value; patience = 6; loss function = categorical cross-entropy; and momentum= 0.95. Finally, we incorporate a batch re-balancing strategy to enhance infection form distribution at the batch stage.

4.1. Assessment Methods

To evaluate performance, we compared our proposed system with other systems using the performance metrics listed in Equations (1)–(4),:
Accuracy = TP + TN TP + TN + FP + FN
Precision = TP TP + FP
Sensitivity = Recall = TP TP + FN
F 1 - Score = 2 × Precision × Recall Precision + Recall = 2 × TP 2 × TP + FP + FN
Here, TP denotes true positives (patients correctly identified as having COVID-19), TN denotes true negatives (patients correctly identified as not having COVID-19), FP denotes false positives (patients with lung diseases other than COVID-19 or mild lung identified as having COVID-19), and FN denotes false negatives (patients with COVID-19 identified as not having the disease).

4.2. Results of the Proposed Systems

In this section, we report the different experiments’ results of the proposed systems using the COV-PEN dataset with a 80–20% train–test split. That split is selected to ensure that execution times were not prohibitive. In the first experiment, we trained the Resnet-50 first version and second version models for 10 epochs using 10% of the training set as a validation set, a batch size of 128, and a learning rate ranging from 0.0002 up to 0.001 and froze the weights of the first 50 layers of the model. We executed the training three times and monitored the average accuracy measures over the validation set. Table 2, Table 3, Table 4 and Table 5 show the average accuracy of an ensemble of the modified models. As mentioned previously, we built the model ensemble using multiple runs (three runs for the same parameters) to train the same model with the same parameters (Runs 1–3, Table 2, Table 3, Table 4 and Table 5). An observation can be made regarding the accuracy, which varies from run to run as the weights are initialized randomly each run; only the best run result is saved. Comparing the two versions, the best achieved accuracy for the first and second version models are 97.84% and 94.26%, respectively.
We focus on the first version model in the second experiment because it outperformed the second version model. Therefore, in the second experiment, we trained the Resnet-50 first version model only for 10 epochs using 10% of the training set as a validation set, a batch size of 128, and a learning rate ranging from 0.0002 up to 0.001 with no freezing for the weights. We executed the training three times and monitored the average accuracy measures over the validation set. Table 6 and Table 7 show the average accuracies of an ensemble of the first version model, which are 86.63% and 94.24% using the Adam and sgmd optimizers, respectively. Eventually, Figure 6, Figure 7 and Figure 8 demonstrate the three confusion matrices of COVID-19 infected and non-COVID-19 test results using the first version model with freeze = 0 and freeze = 50 and the second version model with freeze = 50, respectively. Figure 7 indicates that two COVID-19-infected images were misclassified as non-COVID-19 images while two non-COVID-19 images were misclassified as COVID-19 images, and Figure 8 reveals that two COVID-19-infected images were misclassified as non-COVID-19 images whereas no non-COVID-19 images were misclassified as COVID-19 images. Finally, Figure 9 shows that three COVID-19-infected images were misclassified as non-COVID-19 images, and two non-COVID-19 images were misclassified as COVID-19 images.
The AUC receiver operating characteristic (ROC) curve, which is a graph that shows a model’s classification performance on two parameters: true positives and false positives, was used to evaluate diagnostic efficiency. By incorporating the areas of small trapezoidal fragments under the ROC curve, the AUC can be determined. The proposed method’s AUC evaluations as well as the first edition (with freeze = 0 and freeze = 50) and 2nd version models are seen in Figure 10 and Figure 11. In Figure 12, the AUC is also similar, which is the best-case scenario. The proposed first edition framework classification model (with AUC = 1) performs slightly better than existing COVID-19 classification models.
Table 8 shows the best results obtained using the first version with freeze = 0 and freeze = 50 and for the second version models. The disparity in performance can be explained by the fact that the first version with freeze = 0 and the second version models have an extra layer that starts with random weights rather than pre-trained weights, as in the other layers. These random weights add extra degrees of freedom to the model and should improve its generalization ability.
Figure 13 illustrates the best results obtained in terms of overall accuracy, precision, recall, F1-score, and AUC. It reveals that the best result is obtained when using the first version model and freeze = 50.
The first version model is used in the efficiency comparison of the proposed COV-PEN scheme with current state-of-the-art approaches, in addition to its usefulness in leveraging the great strengths of each classifier. These findings bolstered the case for implementing the proposed COV-PEN method in real-world environments to help radiologists diagnose COVID-19 infection more accurately using CXR images while also reducing their workload.

4.3. Comparison with State-of-the-Art Methods

The proposed system’s performance and reliability are compared with the most recent state-of-the-art COVID-19 detection systems. In this section, we present the proposed first version system’s outcomes and compare them with existing methods (see Table 9). As revealed in Figure 14, the proposed system demonstrates remarkable results that are more accurate than existing methods. Furthermore, compared with other models such as VGG16 or DenseNet, the proposed improved Resnet-50 system is lightweight. In terms of precision, accuracy, sensitivity, and F1-score, our proposed system outperformed existing methods.

5. Conclusions and Future Work

A reliable and automatic mechanism for COVID-19 diagnosis is presented using chest radiography images to differentiate between patients with mild, pneumonia, and COVID-19 infections. Image enhancement techniques were used in the proposed system to improve the intensity of the CXR image and to eliminate noise. To avoid overfitting and to improve the overall capabilities of the proposed DL systems, two different DL approaches (first and second proposed versions of Resnet-50) were trained on top of preprocessed chest medical imaging. To evaluate the reliability of the proposed system, a CXR image dataset labeled the COV-PEN dataset was built. With an overall accuracy of 99.63%, a precision of 100%, a recall of 98.89%, an F1-score of 99.44%, and an AUC of 100%, the proposed system performs equally well for expert radiologists. The proposed system outperforms current models, per the comparative studies. Experiments with a huge and challenging dataset containing multiple COVID-19 cases are required to demonstrate the efficacy of the proposed system. Other methods, such as Densenet, VGG, or Inception-Resnet network, may be used on the COV-PEN dataset as a future work.

Author Contributions

Conceptualization, W.G., M.A., M.H. and N.Z.J.; methodology, W.G., M.A., M.H. and N.Z.J.; software, W.G., M.A., M.H. and N.Z.J.; validation, W.G., M.A., M.H. and N.Z.J.; formal analysis, W.G., M.A., M.H. and N.Z.J.; investigation, W.G., M.A., M.H. and N.Z.J.; resources, W.G., M.A., M.H. and N.Z.J.; data curation, W.G., M.A., M.H. and N.Z.J.; writing—original draft preparation, W.G., M.A., M.H. and N.Z.J.; writing—review and editing, W.G., M.A., M.H. and N.Z.J.; visualization, W.G., M.A., M.H. and N.Z.J.; supervision, W.G., M.A., M.H. and N.Z.J.; project administration, W.G., M.A., M.H. and N.Z.J.; funding acquisition, W.G., All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Scientific Research at Jouf University under grant No (DSR-2021-02-0370).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia (accessed on 13 December 2021); https://github.com/ieee8023/covid-chestxray-dataset/tree/master/images] (accessed on 13 December 2021).

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research at Jouf University for funding this work through Research Grant No (DSR-2021-02-0370).

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

References

  1. Humayun, M.; Alsayat, A. Prediction Model for Coronavirus Pandemic Using Deep Learning. Comput. Syst. Sci. Eng. 2022, 40, 947–961. [Google Scholar] [CrossRef]
  2. Gorbalenya, A.E.; Baker, S.C.; Baric, R.; Groot, R.J.d.; Drosten, C.; Gulyaeva, A.A.; Haagmans, B.L.; Lauber, C.; Leontovich, A.M.; Neuman, B.W.; et al. Severe acute respiratory syndrome-related coronavirus: The species and its viruses—A statement of the Coronavirus Study Group. bioRxiv 2020. [Google Scholar]
  3. Stoecklin, S.B.; Rolland, P.; Silue, Y.; Mailles, A.; Campese, C.; Simondon, A.; Mechain, M.; Meurice, L.; Nguyen, M.; Bassi, C.; et al. First cases of coronavirus disease 2019 (COVID-19) in France: Surveillance, investigations and control measures, January 2020. Eurosurveillance 2020, 25, 2000094. [Google Scholar]
  4. Khalil, M.I.; Tehsin, S.; Humayun, M.; Jhanjhi, N.; Zain, M.A. Multi-Scale Network for Thoracic Organs Segmentation. Comput. Mater. Contin. 2022, 70, 3251–3265. [Google Scholar] [CrossRef]
  5. Chavez, S.; Long, B.; Koyfman, A.; Liang, S.Y. Coronavirus Disease (COVID-19): A primer for emergency physicians. Am. J. Emerg. Med. 2021, 44, 220–229. [Google Scholar] [CrossRef] [PubMed]
  6. Al-Waisy, A.; Mohammed, M.A.; Al-Fahdawi, S.; Maashi, M.; Garcia-Zapirain, B.; Abdulkareem, K.H.; Mostafa, S.; Kumar, N.M.; Le, D.N. Covid-deepnet: Hybrid multimodal deep learning system for improving covid-19 pneumonia detection in chest X-ray images. Comput. Mater. Contin. 2021, 67, 2409–2429. [Google Scholar] [CrossRef]
  7. Elghamrawy, S.M.; Hassnien, A.E.; Snasel, V. Optimized Deep Learning-Inspired Model for the Diagnosis and Prediction of COVID-19. Comput. Mater. Contin. 2021, 67, 2353–2371. [Google Scholar] [CrossRef]
  8. Armstrong, M. The Countries with the Most COVID-19 Cases. 2021. Available online: https://www.statista.com/chart/21467/coutries-most-covid-19-cases/ (accessed on 13 December 2021).
  9. Asai, T. COVID-19: Accurate Interpretation of Diagnostic Tests—A Statistical Point of View. 2021. Available online: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7729143/ (accessed on 13 December 2021).
  10. Gaál, G.; Maga, B.; Lukács, A. Attention u-net based adversarial architectures for chest X-ray lung segmentation. arXiv 2020, arXiv:2003.10304. [Google Scholar]
  11. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (Covid-19) using X-ray images and deep convolutional neural networks. Pattern Anal. Appl. 2021, 24, 1207–1220. [Google Scholar] [CrossRef] [PubMed]
  12. Das, N.N.; Kumar, N.; Kaur, M.; Kumar, V.; Singh, D. Automated deep transfer learning-based approach for detection of COVID-19 infection in chest X-rays. Irbm 2020, in press. [Google Scholar]
  13. Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the devil in the details: Delving deep into convolutional nets. arXiv 2014, arXiv:1405.3531. [Google Scholar]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Shen, D.; Wu, G.; Suk, H.I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Wang, L.; Lin, Z.Q.; Wong, A. Covid-net: A tailored deep convolutional neural network design for detection of covid-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 1–12. [Google Scholar] [CrossRef]
  18. Hemdan, E.E.-D.; Shouman, M.A.; Karar, M.E. COVIDXnet: A framework of deep learning classifiers to diagnose COVID-19 in X-ray images. arXiv 2020, arXiv:2003.11055. [Google Scholar]
  19. Zhang, J.; Xie, Y.; Li, Y.; Shen, C.; Xia, Y. Covid-19 screening on chest X-ray images using deep learning based anomaly detection. arXiv 2020, arXiv:2003.12338. [Google Scholar]
  20. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef]
  21. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  22. Ucar, F.; Korkmaz, D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from X-ray images. Med Hypotheses 2020, 140, 109761. [Google Scholar] [CrossRef]
  23. Apostolopoulos, I.D.; Aznaouridis, S.I.; Tzani, M.A. Extracting possibly representative COVID-19 biomarkers from X-ray images with deep learning approach and image data related to pulmonary diseases. J. Med Biol. Eng. 2020, 40, 462–469. [Google Scholar] [CrossRef] [PubMed]
  24. Sahinbas, K.; Catak, F.O. Transfer learning-based convolutional neural network for COVID-19 detection with X-ray images. In Data Science for COVID-19; Elsevier: Amsterdam, The Netherlands, 2021; pp. 451–466. [Google Scholar]
  25. Jamil, M.; Hussain, I. Automatic detection of COVID-19 infection from chest X-ray using deep learning. medRxiv 2020. [Google Scholar] [CrossRef]
  26. Alazab, M.; Awajan, A.; Mesleh, A.; Abraham, A.; Jatana, V.; Alhyari, S. COVID-19 prediction and detection using deep learning. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2020, 12, 168–181. [Google Scholar]
  27. Joaquin, A. Using Deep Learning to Detect Pneumonia Caused by NCOV-19 from X-ray Images. Available online: https://towardsdatascience.com/using-deep-learning-to-detect-ncov-19-from-X-ray-images-1a89701d1acd (accessed on 13 December 2021).
  28. Sethy, P.K.; Behera, S.K. Detection of Coronavirus Disease (COVID-19) Based on Deep Features. 2020. Available online: https://www.preprints.org/manuscript/202003.0300/v1 (accessed on 13 December 2021).
  29. Houssein, E.H.; Abohashima, Z.; Elhoseny, M.; Mohamed, W.M. Hybrid quantum convolutional neural networks model for COVID-19 prediction using chest X-ray images. arXiv 2021, arXiv:2102.06535. [Google Scholar]
  30. Saad, W.; Shalaby, W.A.; Shokair, M.; Abd El-Samie, F.; Dessouky, M.; Abdellatef, E. COVID-19 classification using deep feature concatenation technique. J. Ambient. Intell. Humaniz. Comput. 2021, 1–19. [Google Scholar] [CrossRef] [PubMed]
  31. Apostolopoulos, I.D.; Mpesiana, T.A. Covid-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [Green Version]
  32. Oh, Y.; Park, S.; Ye, J.C. Deep learning COVID-19 features on cxr using limited training data sets. IEEE Trans. Med Imaging 2020, 39, 2688–2700. [Google Scholar] [CrossRef]
  33. Brunese, L.; Mercaldo, F.; Reginelli, A.; Santone, A. Explainable deep learning for pulmonary disease and coronavirus COVID-19 detection from X-rays. Comput. Methods Programs Biomed. 2020, 196, 105608. [Google Scholar] [CrossRef]
  34. Islam, M.Z.; Islam, M.M.; Asraf, A. A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. Inform. Med. Unlocked 2020, 20, 100412. [Google Scholar] [CrossRef]
  35. Ezzat, D.; Hassanien, A.E.; Ella, H.A. An optimized deep learning architecture for the diagnosis of COVID-19 disease based on gravitational search optimization. Appl. Soft Comput. 2020, 98, 106742. [Google Scholar] [CrossRef] [PubMed]
  36. Sahlol, A.T.; Yousri, D.; Ewees, A.A.; Al-Qaness, M.A.; Damasevicius, R.; Abd Elaziz, M. COVID-19 image classification using deep features and fractional-order marine predators algorithm. Sci. Rep. 2020, 10, 15364. [Google Scholar] [CrossRef]
  37. Toraman, S.; Alakus, T.B.; Turkoglu, I. Convolutional capsnet: A novel artificial neural network approach to detect COVID-19 disease from X-ray images using capsule networks. Chaos Solitons Fractals 2020, 140, 110122. [Google Scholar] [CrossRef] [PubMed]
  38. Rajaraman, S.; Antani, S. Training deep learning algorithms with weakly labeled pneumonia chest X-ray data for COVID-19 detection. medRxiv 2020. [Google Scholar] [CrossRef]
  39. Afshar, P.; Heidarian, S.; Naderkhani, F.; Oikonomou, A.; Plataniotis, K.N.; Mohammadi, A. Covid-caps: A capsule network-based framework for identification of covid-19 cases from X-ray images. Pattern Recognit. Lett. 2020, 138, 638–643. [Google Scholar] [CrossRef] [PubMed]
  40. Elshennawy, N.M.; Ibrahim, D.M. Deep-pneumonia framework using deep learning models based on chest X-ray images. Diagnostics 2020, 10, 649. [Google Scholar] [CrossRef] [PubMed]
  41. Cohen, J.P.; Morrison, P.; Dao, L.; Roth, K.; Duong, T.Q.; Ghassemi, M. Covid-19 image data collection: Prospective predictions are the future. arXiv 2020, arXiv:2006.11988. [Google Scholar]
  42. Mooney, P. Chest X-ray Images (Pneumonia). 2018. Available online: https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia (accessed on 13 December 2021).
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  44. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  45. Basha, S.S.; Dubey, S.R.; Pulabaigari, V.; Mukherjee, S. Impact of fully connected layers on performance of convolutional neural networks for image classification. Neurocomputing 2020, 378, 112–119. [Google Scholar] [CrossRef] [Green Version]
  46. Renard, F.; Guedria, S.; De Palma, N.; Vuillerme, N. Variability and reproducibility in deep learning for medical image segmentation. Sci. Rep. 2020, 10, 1–16. [Google Scholar] [CrossRef]
Figure 1. Confirmed COVID-19 cases globally (15 January 2022) [8].
Figure 1. Confirmed COVID-19 cases globally (15 January 2022) [8].
Healthcare 10 00343 g001
Figure 2. Samples of CXR images and CT images (a), and CXR image scan (b) CT image scan.
Figure 2. Samples of CXR images and CT images (a), and CXR image scan (b) CT image scan.
Healthcare 10 00343 g002
Figure 3. A schematic methodology for the COVID19 detection system.
Figure 3. A schematic methodology for the COVID19 detection system.
Healthcare 10 00343 g003
Figure 4. CXR images from COV-PEN dataset: (a) COVID-19, (b) pneumonia, and (c) mild.
Figure 4. CXR images from COV-PEN dataset: (a) COVID-19, (b) pneumonia, and (c) mild.
Healthcare 10 00343 g004
Figure 5. Output of the proposed image enhancement process: (a) raw CXR image and (b) enhanced image.
Figure 5. Output of the proposed image enhancement process: (a) raw CXR image and (b) enhanced image.
Healthcare 10 00343 g005
Figure 6. Modified versions of the proposed Resnet-50 model: (a) original pre-trained model, (b) adding one FC layer, and (c) adding two FC layers and one Sigmoid.
Figure 6. Modified versions of the proposed Resnet-50 model: (a) original pre-trained model, (b) adding one FC layer, and (c) adding two FC layers and one Sigmoid.
Healthcare 10 00343 g006
Figure 7. The best result’s confusion matrix for the first version model freeze = 0.
Figure 7. The best result’s confusion matrix for the first version model freeze = 0.
Healthcare 10 00343 g007
Figure 8. The best result’s confusion matrix for the first version model freeze = 50.
Figure 8. The best result’s confusion matrix for the first version model freeze = 50.
Healthcare 10 00343 g008
Figure 9. The best result’s confusion matrix for the second version model.
Figure 9. The best result’s confusion matrix for the second version model.
Healthcare 10 00343 g009
Figure 10. ROC curve for the first version model freeze = 0.
Figure 10. ROC curve for the first version model freeze = 0.
Healthcare 10 00343 g010
Figure 11. ROC curve for the first version model freeze = 50.
Figure 11. ROC curve for the first version model freeze = 50.
Healthcare 10 00343 g011
Figure 12. ROC curve for the second version model.
Figure 12. ROC curve for the second version model.
Healthcare 10 00343 g012
Figure 13. Best result for all models.
Figure 13. Best result for all models.
Healthcare 10 00343 g013
Figure 14. The proposed first version model performance in comparison with current systems.
Figure 14. The proposed first version model performance in comparison with current systems.
Healthcare 10 00343 g014
Table 1. Literature comparison of COVID-19 diagnostic methods using CXR images.
Table 1. Literature comparison of COVID-19 diagnostic methods using CXR images.
Recent WorkTechniques UsedNumber of ClassesAccuracy
Khan et al. [21]CoroNet489.6%
Ucar and Korkmaz [22]Bayes-SqueezeNet398.3%
Apostopolus et al. [23]VGG-19393.48%
Sahinbas & Catak [24]VGG-16, VGG-19, ResNet, DenseNet, InceptionV3280%
Jamil and Hussain [25]Deep CNN293%
Alzab et al. [26]VGG-162-
Joaquin. [27]ResNet-50296.2%
Sethy et al. [28]ResNet-50 + SVM395.33%
Houssein et al. [29]hybrid quantum classical CNNs388.6%
Saad et al. [30]CNN, GoogleNet, ResNet-18299.3%
Apostolopoulos, & Mpesiana [31]MobileNetV2396.78%
Oh et al. [32]ResNet-18388.9%
Brunese et al. [33]VGG-16396%
slam et al. [34]CNN+LSTM299.4%
Ezzat et al. [35]DenseNet121+GSA298.3%
Sahlol et al. [36]Inception + FO-MPA299.6%
Toraman et al. [37]Capsule Network297.24%
Rajaraman, S. and Antani, S. [38]VGG16293.0%
Afshar, P. et al. [39]capsule network297.2%
Elshennawy, N. & Ibrahim, D. [40]ResNet152V2, MobileNetV2299.22%
Table 2. Average accuracy for the first version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = Adam, and batch size = 128.
Table 2. Average accuracy for the first version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = Adam, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.9498207890.9247311830.955197133
0.00040.9354838710.9193548390.928315412
0.00060.9641577060.9784946240.935483871
0.00080.9569892470.9301075270.944444444
0.0010.9354838710.9193548390.928315412
Table 3. Average accuracy for the first version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = sgmd, and batch size = 128.
Table 3. Average accuracy for the first version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = sgmd, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.9534050180.9551971330.962365591
0.00040.9318996420.9103942650.949820789
0.00060.9498207890.9480286740.931899642
0.00080.9193548390.9444444440.9390681
0.0010.9318996420.9103942650.949820789
Table 4. Average accuracy for the second version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = Adam, and batch size = 128.
Table 4. Average accuracy for the second version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = Adam, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.8530465950.9039426520.808960573
0.00040.7498207890.9075268820.892473118
0.00060.6781362010.8089605730.747311828
0.00080.887813620.624372760.716845878
0.0010.6835125450.6799283150.617845867
Table 5. Average accuracy for the second version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = sgmd, and batch size = 128.
Table 5. Average accuracy for the second version model using the COV-PEN dataset with the first 50 layers frozen, epochs = 15, optimizer = sgmd, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.9336917560.9372759860.933691756
0.00040.9390681000.9354838710.903942652
0.00060.9426523290.9175627240.907526882
0.00080.8924731180.7168458780.808960573
0.0010.7473118280.9211469530.808960573
Table 6. Average accuracy for the first version model using the COV-PEN dataset freeze = 0, epochs = 15, optimizer = Adam, and batch size = 128.
Table 6. Average accuracy for the first version model using the COV-PEN dataset freeze = 0, epochs = 15, optimizer = Adam, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.8154121860.8430107530.842175627
0.00040.6075268820.7724014340.607526882
0.00060.8663082440.7982078850.81172043
0.00080.6103942650.6788530470.607526882
0.0010.7566308240.7333333330.734050179
Table 7. Average accuracy for the first version model using the COV-PEN dataset freeze = 0, epochs = 15, optimizer = sgmd, and batch size = 128.
Table 7. Average accuracy for the first version model using the COV-PEN dataset freeze = 0, epochs = 15, optimizer = sgmd, and batch size = 128.
Learning RateEnsemble Using Several Runs
Run 1Run 2Run 3
0.00020.942652330.9372759860.933691756
0.00040.9336917560.9086021510.919002151
0.00060.93906810.881720430.843010753
0.00080.9354838710.8691756270.81172043
0.0010.7340501790.7903225810.756630824
Table 8. Best result for all models.
Table 8. Best result for all models.
Quantitative Measures1st Version Model (freeze = 0)1st Version Model (freeze = 50)2nd Version Model
Overall Accuracy99.0599.6399.05
Precision98.8910098.92
Recall98.3998.8998.39
F1-score98.6399.4498.65
AUC99.9910099.98
Table 9. Comparison of the proposed 1st version model with existing systems.
Table 9. Comparison of the proposed 1st version model with existing systems.
ReferenceDatasetClassifierAccuracyPrecisionRecallF1-ScoreAUC
Proposed 1st versionCOV-PENModified version of Resnet-5099.6310098.8999.44100
Afshar, P. et. al. [39]- COVID-19 Image Data Collection97.297.6797.597.70
- Chest X-ray Images (Pneumonia)
- COVID-19 Image Data Collection
Rajaraman, S. and Antani, S. [38]- Pediatric CXR datasetVGG1693.093.1597.5394.5795.0
- RSNA CXR datasetInceptionV3
- CheXpert CXR datasetXception
- NIH CXR-14 datasetDensenet121
- Twitter COVID-19 CXR datasetNASNet-Mobile
- COVID-19 Image Data Collection
Narin, A., et. al. [11]- Chest X-ray Images (Pneumonia)ResNet-5099.599.499.598.098.7
- COVID-19 Image Data CollectionResNet-101
ResNet-152
InceptionV3
InceptionRes
net-V2
- Chest X-ray Images (Pneumonia)
Hemdan, E. et. al. [18]COVID-19 Image Data CollectionDenseNet201908391.00
Elshennawy, N. & Ibrahim, D. [40]Chest X-ray Images (Pneumonia)ResNet152V299.2299.4399.4499.4499.77
MobileNetV2
Wang et al. [17]- COVID-19 Image Data CollectionVGG-1993.391
- COVID-19 Chest X-ray DatasetResnet-50
- ActualMed COVID-19 Chest X-ray DatasetCOVID-Net
- RSNA Pneumonia Detection Challenge dataset
- COVID-19 radiography database
Zhang et al. [19]- COVID-19 Image Data CollectionResNet-189695.18
- Chest X-ray Images (Pneumonia)
Das et al. [12]- COVID-19 Image Data Collectionextreme version of the Inception (Xception) model97.4097.0996.96
- ChestX-ray8 database (PneumoniaNormal)
Ozturk et. al. [20]- COVID-19 Image Data CollectionDarkNet98.0898.395.196.5
- ChestX-ray8 database (PneumoniaNormal)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gouda, W.; Almurafeh, M.; Humayun, M.; Jhanjhi, N.Z. Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare 2022, 10, 343. https://doi.org/10.3390/healthcare10020343

AMA Style

Gouda W, Almurafeh M, Humayun M, Jhanjhi NZ. Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare. 2022; 10(2):343. https://doi.org/10.3390/healthcare10020343

Chicago/Turabian Style

Gouda, Walaa, Maram Almurafeh, Mamoona Humayun, and Noor Zaman Jhanjhi. 2022. "Detection of COVID-19 Based on Chest X-rays Using Deep Learning" Healthcare 10, no. 2: 343. https://doi.org/10.3390/healthcare10020343

APA Style

Gouda, W., Almurafeh, M., Humayun, M., & Jhanjhi, N. Z. (2022). Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare, 10(2), 343. https://doi.org/10.3390/healthcare10020343

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop