Next Article in Journal
Optical Modelling of Planar and Fibre Perovskite Solar Cells
Next Article in Special Issue
A Novel Tightly Coupled Information System for Research Data Management
Previous Article in Journal
Simulation of Authentication in Information-Processing Electronic Devices Based on Poisson Pulse Sequence Generators
Previous Article in Special Issue
A Hybrid Framework for Lung Cancer Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ROENet: A ResNet-Based Output Ensemble for Malaria Parasite Classification

School of Computing and Mathematical Sciences, University of Leicester, East Midlands, Leicester LE1 7RH, UK
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(13), 2040; https://doi.org/10.3390/electronics11132040
Submission received: 7 June 2022 / Revised: 21 June 2022 / Accepted: 25 June 2022 / Published: 29 June 2022
(This article belongs to the Special Issue Data-Driven Processing from Complex Systems Perspective)

Abstract

:
(1) Background: People may be infected with an insect-borne disease (malaria) through the blood input of malaria-infected people or the bite of Anopheles mosquitoes. Doctors need a lot of time and energy to diagnose malaria, and sometimes the results are not ideal. Many researchers use CNN to classify malaria images. However, we believe that the classification performance of malaria parasites can be improved. (2) Methods: In this paper, we propose a novel method (ROENet) to automatically classify malaria parasite on the blood smear. The backbone of ROENet is the pre-trained ResNet-18. We use randomized neural networks (RNNs) as the classifier in our proposed model. Three RNNs are used in ROENet, which are random vector functional link (RVFL), Schmidt neural network (SNN), and extreme learning machine (ELM). To improve the performance of ROENet, the results of ROENet are the ensemble outputs from three RNNs. (3) Results: We evaluate the proposed ROENet by five-fold cross-validation. The specificity, F1 score, sensitivity, and accuracy are 96.68 ± 3.81%, 95.69 ± 2.65%, 94.79 ± 3.71%, and 95.73 ± 2.63%, respectively. (4) Conclusions: The proposed ROENet is compared with other state-of-the-art methods and provides the best results of these methods.

1. Introduction

People may be infected with an insect-borne disease (malaria) through the blood input of malaria-infected people or the bite of Anopheles mosquitoes. Colds, periodic and regular attacks, fever, and sweating are several characteristics of patients infected with malaria. After many attacks for a long time, it can cause anemia and splenomegaly. Thus far, the harmful impact of malaria in the world is still very serious. People living in malaria-endemic areas account for about 40% of the total human population. Malaria is one of the most feared diseases for people living on the African continent. Every year, malaria patients on the African continent account for 90% of malaria patients in the world. Malaria is the cause of death of more than two million people worldwide every year. Malaria is also prevalent in Southeast and Central Asia. The infection sources of malaria are roughly divided into two categories: (i) malaria patients and (ii) asymptomatic carriers containing gametophytes. The transmission probability of malaria increases with the increase in gametophyte density.
The earlier malaria is diagnosed, the more conducive it is for treatment of patients. Thus far, the diagnostic methods of malaria in the hospital are as follows: (i) clinical manifestation, (ii) molecular biological technology diagnosis, and (iii) therapeutic diagnosis. However, these diagnostic methods require doctors to spend a long period of time when diagnosing. This long diagnosis process may delay the timely treatment of patients. Moreover, doctors are easily disturbed by some factors in the diagnosis process, such as lack of sleep, illness, and so on. Therefore, the process of diagnosing malaria by doctors is very inefficient.
Many researchers are willing to apply computer technology to malaria diagnosis. Manescu et al. [1] introduced a new model (DeepMCNN), which was based on the convolution neural network. The introduced DeepMCNN obtained 0.92 sensitivity and 0.90 specificity. Yang et al. [2] introduced a novel method to detect malaria. This method combined IGMS and CNN and obtained 93.46 ± 0.32% accuracy, 92.59 ± 1.27% sensitivity, 94.33 ± 1.25% specificity, and 94.25 ± 1.13% precision. Shoohi and Saud [3] introduced a method for malaria image classification. This method (Deep Convolutional Generative Adversarial Network) was based on CNN. Mehanian et al. [4] used deep learning to classify malaria images. Mukherjee et al. [5] introduced a novel method to classify malaria images. The proposed method was based on the convolution neural network and obtained a 0.95 dice score. Khadim et al. [6] evaluated several different activation functions to detect malaria in the CNN model. These several different activation functions were Sigmoid, Tanh, Leaky ReLU, ReLU, and Swish. Based on the experimental results, the Swish activation function yielded better results than other activation functions. Magotra and Rohil [7] introduced a novel model (lightweight CNN) to detect malaria. Two classical models were used to compare with the lightweight CNN, which were VGG-19 and Inception-v3. This lightweight CNN model could obtain 96% accuracy. At the same time, this model can reduce training time and computations. Marques et al. [8] proposed a new model to detect malaria based on the CNN model. In this model, EfficientNet was used as the backbone model. This model achieved 97.74% precision, 99.76% ROC, 98.82% recall, and 98.28% F1. Sarkar et al. [9] presented a shallow-approach CNN model. The proposed model could reduce the run time. Raihan and Nahid [10] introduced an explainable CNN model to classify malaria. This explainable CNN model was composed of CNN, wavelet packet 2d, and Whale Optimization Algorithm. This model obtained 94.39% precision, 94.80% F1, 94.78% accuracy, and 95.21% recall.
From the above description of malaria diagnosis by computer technology [11], it can be concluded that most scholars use the CNN model for experiments. However, we believe that the classification performance of malaria parasite can be improved. We propose a new model (ROENet) to automatically classify malaria parasites on the blood smears. ROENet means that the model is a ResNet-based output ensemble for malaria parasite classification. The main contributions of this study are the following:
  • A novel method (ROENet) is proposed to automatically classify malaria parasite on the blood smear.
  • The fine-tuned ResNet-18 is the feature extraction.
  • Three RNNs are selected to replace the last five layers of the fine-tuned ResNet-18.
  • Three RNNs are selected as the classifier of the proposed ROENet.
  • The final outputs of ROENet are the ensemble outputs from three RNNs.
The remainder of the study is organized as follows: the public dataset is demonstrated in Section 2; Section 3 details the method; the experiment settings and results are presented in Section 4; Section 5 is the conclusion.

2. Materials

The malaria images are available on the NIH website, which the Chittagong Medical College Hospital prepared. This dataset is public. Giemsa-stained thin blood smears were from 150 patients infected with Plasmodium falciparum and 50 healthy patients in Chittagong Medical College Hospital, Bangladesh. These stained thin blood smears were photographed. The slide image of each micro field of view was taken through the built-in camera in the smartphone. These images were manually marked by experts from Oxford Tropical Medicine Research Center in Mahilon, Bangkok, Thailand. There are 27,558 malaria images in total, including 13,779 images of parasitized and 13,779 images of uninfected. Some of the parasitized and uninfected images are provided in Figure 1. The image processing method is used in this open dataset to find parasites in the digital image of blood film. The typical shape, data, and visual appearance of parasites are marked manually by experts. If there is no expert mark, the image is uninfected.

3. Methods’ Results

3.1. Proposed ROENet

The acronym and full explanation table is provided in Table 1. As more and more scholars research image analysis, image analysis technology continues to progress [12]. One of the most significant steps in the analysis of the image is to extract features from images [13]. However, each image contains too much content. Extracting useful features in the image quickly and accurately has been perplexing to scholars. Previously, scholars manually extracted features from the image [14]. However, the process of manually extracting features was very time consuming, and the results were often not ideal. More and more scholars were applying computer technology to image analysis and proposed many CNN models [15], such as AlexNet, ResNet, etc. In the CNN model, the convolution layers and pooling layers can reduce the number of parameters. In this situation, calculation and experimental time are greatly reduced. This is one of the reasons why the CNN model is so popular [16].
This paper proposes a novel method (ROENet) to automatically classify malaria parasite on the blood smear. The public dataset can be downloaded on the NIH website. The backbone of ROENet is the pre-trained ResNet-18. ResNet-18 was pre-trained on the ImageNet. Therefore, the output nodes of the pre-trained ResNet-18 are 1000. However, the output nodes are two in this paper. Therefore, we perform some modifications to the backbone. We chose randomized neural networks (RNNs) as the classifier in our proposed model. Three RNNs are used in ROENet, which include random vector functional link (RVFL) [17], Schmidt neural network (SNN) [18], and extreme learning machine (ELM) [19]. The results of ROENet are the ensemble outputs from three RNNs to improve the performance. Table 2 demonstrates the pseudocode of ROENet. The flowchart of ROENet is provided in Figure 2.

3.2. Backbone of ROENet

The depth of the network has a great influence on the performance of the CNN model. In theory, with the deepening of the network depth, the model should achieve better performance [20]. However, the performance of deeper networks is not necessarily good. When the depth of the CNN model continues to increase, the CNN model may encounter the degradation problem [21]. The performance of the model stagnates or even decreases when the number of layers of the CNN model increases. This is the problem of degradation [22].
The degradation problem at least shows that the network model is difficult to train. By adding new layers, the network model increases the depth continuously [23]. Sometimes, the newly added layers learn nothing and simply copy the features of the previous layer. This is identity mapping [24]. This can ensure that the performance of the network model will not degrade. Residual learning solves the degradation problem in this way [25]. For a CNN structure (formed by stacking several layers), X is the input, P ( X ) is the learned feature, and the feature extracted by residual learning is recorded as Q ( X ) .
Q ( X ) = P ( X ) X
P ( X ) = Q ( X ) + X
It can be seen from the above formula that the worst case for residual learning is that the residual is 0, so the stacking layer just completes the identity mapping [26]. When the residual is not 0, the stacking layer will learn new features to improve the network’s performance. Therefore, residual learning can cope with the degradation problem [27]. The residual learning structure is demonstrated in Figure 3.
The backbone of ROENet is the pre-trained ResNet-18. The output nodes of the pre-trained ResNet-18 are 1000. However, the output nodes are two in this paper. Therefore, we perform some modifications to the backbone. The modifications of the backbone of the proposed ROENet are provided in Figure 4. FC1000, softmax, and the classification layer are removed. We add FC128, ReLU, BN, FC2, softmax, and the classification layer.

3.3. Classifier of ROENet

There are many layers in the CNN model, and each layer has many parameters. The randomized neural networks (RNNs) have only three simple layers: input layer, hidden layer, and output layer. Only the shallow structure of the three-layer RNN model can effectively alleviate the overfitting problem. The parameters (the randomized weights and biases) in the RNN model are also trained quickly. Because RNN has good classification performance, it has been applied to many machine learning tasks, such as geography, big data analysis, chemistry, and so on. Three RNNs are used in this paper, which are ELM, RVFL, and SNN. ELM projects the input features into the hidden space randomly and does not need gradient-based backpropagation to adjust the weights [28]. The most obvious structural difference between RVFL and ELM is that there is a quick connection between input and output in RVFL [29]. This quick connection can effectively improve the classification performance of RVFL and the robustness of the model. SNN [30] was an RNN model proposed by Schmidt, Kraijveld, and Duin 30 years ago. The structure of SNN is consistent with that of ELM. However, in the SNN model, the output layer has a learnable output bias. These three RNN models are very classical and have achieved excellent classification performance since they were proposed. Their structures are provided in Figure 5.
As can be seen from the above figure, the structures of these three RNNs used in this paper are different. RNNs have only a three-layer structure; thus, the calculation method is almost identical. Suppose there is a dataset ( t i , y i ) and the dataset contains i-th sample:
t i = ( t i 1 , , t i n ) T R n ,   i = 1 , ,   N ,
y i = ( y i 1 , ,   y i m ) T R m ,   i = 1 , ,   N ,
where N is the number of different samples, the input dimension is represented by n , and m is the output dimension.
For ELM, we chave the following:
M E L M ( i ) = j = 1 u l ( v j t i + K j ) , i = 1 ,   ,   N .
where v j represents the weight from the input node to the j-th node in the hidden layer, K j represents the bias of the j-th node in the hidden layer, the sigmoid function is demonstrated by l , and u is the number of hidden nodes in the hidden layer.
For RVFL, this calculation is a step further:
M R V F L ( i ) = concat ( T ,   E ) ,
where the input matrix is T = ( t 1 , , t N ) T .
E R V F L ( i ) = j = 1 u l ( v j t i + K j ) , i = 1 ,   ,   N .
For SNN, we have the following.
M S N N ( i ) = j = 1 u l ( v j t i + K j ) , i = 1 ,   ,   N .
For ELM and RVFL, the final output weights are calculated as follows:
r = M n e t + Y
where r is the final output weight, the pseudo-inverse matrix of M n e t is M n e t + , and the ground-truth label of the dataset is Y = ( y 1   , ,   y N ) T .
For SNN, there are biases ( b ) between the hidden layer and output layer:
( r , b ) = M n e t + Y ,
where M n e t + is the pseudo-inverse matrix of ( M n e t 1 ) .
Although the RNN model is simple, bad weights and biases will seriously affect the classification performance. Therefore, in this paper, we combine the results of three RNN models to obtain the final classification model based on majority voting. Because the three RNN models used in this paper have some differences, it is more helpful to obtain diversified information in order to further improve the performance and robustness of the system.

3.4. Evaluation

The parasitized images are defined as the positive, and the uninfected images are defined as the negative. We evaluate the proposed ROENet by five-fold cross-validation. Five measures are selected, which are sensitivity (Se), accuracy (Ac), F1 score (F1), and specificity (Sp).
{ Se = TP TP + FN Ac = TP + TN TP + TN + FP + FN F 1 = 2 × TP 2 TP + FP + FN   Sp = TN TN + FP ,

4. Experiment Settings and Results

4.1. Experiment Settings

We set the max-epoch to 4 to prevent the overfitting problem. The learning rate is set as 10 4 . The minibatch size is 128. The dataset used in this paper is small and the batch size is large; thus, the convergence is fast. The number of the hidden nodes ( u ) in the hidden layer is 400. Table 3 provides the hyper-parameters of the proposed ROENet.

4.2. The Performance of ROENet

The classification performance of ROENet is provided in Table 4. In this paper, we evaluate the proposed ROENet by five-fold cross-validation. The F1 score (F1), specificity (Sp), accuracy (Ac), and sensitivity (Se) are 95.69 ± 2.65%, 96.68 ± 3.81%, 95.73 ± 2.63%, and 94.79 ± 3.71%, respectively. The average values of all results are greater than 94%. These results prove that the model in this paper is a good choice to classify malaria parasite on the blood smear.

4.3. Comparison of Different Backbones

We test different backbones, which are AlexNet and ResNet-50. The classification performances of these different backbones are presented in Table 4. The different backbones comparison figure is shown in Figure 6. Because ResNet-18 can achieve the best classification results when it is used as the backbone model based on the experimental results, ResNet-18 is selected as the preferred architecture in this paper.
Our model achieves the best results in accuracy, specificity, and F1 score in comparison to the other two models. AlexNet contains too many parameters, which could cause degradation problems. ResNet-50 has more layers than ResNet-18. Therefore, ResNet-50 may be more likely to meet gradient vanishing problems. Therefore, our model obtains better results than other models.

4.4. Effects of Output Ensemble

In this paper, the results of ROENet are the ensemble outputs from three RNNs. To verify the superiority of the proposed ROENet, the proposed model is compared with three individual models. The classification performances of three individual models are provided in Table 4. For a clearer comparison, the comparison figure is presented in Figure 7. Compared with the other three individual networks, our proposed network achieves the best results in accuracy, sensitivity, and F1 score. Although it is not the best in specificity, it is only 0.05% lower than the best.

4.5. Comparison with the Fine-Tuned ResNet-18

We compare the ROENet with the fine-tuned ResNet-18. The results of the fine-tuned model are presented in Table 4. The comparison of the proposed model with the fine-tuned model is provided in Figure 8. Our proposed model achieves better results than the fine-tuned ResNet-18. This proves that our model is an effective tool to classify malaria parasite on the blood smear.
The parameters and layers of RNNs are less than those of ResNet-18. Because the dataset in this paper is not very large, RNN is unlikely to have the problem of overfitting. Therefore, the proposed ROENet overperforms the fine-tuned ResNet-18.

4.6. Comparison with Other State-of-the-Art Methods

ROENet is compared with other state-of-the-art (SOTA) methods, which are DeepMCNN [1], Customized CNN [2], DCGAN [3], Computer-Automated CNN [4], and three-layer CNN [9], respectively. DCGAN, Computer-Automated CNN, and three-layer CNN used the same dataset as this paper. Other SOTA methods used different datasets. The comparison is provided in Table 5. The comparison figure is presented in Figure 9. From the table and figure, we can see that our model obtains the best results in comparison with other SOTA methods.
There are three reasons why our model can achieve better results than other SOTA methods. (i) ResNet-18 is the backbone of our model, which can accurately extract features. (ii) We use RNN as the classifier, which can avoid overfitting problems. (iii) The results of ROENet are the ensemble outputs from three RNNs, which can improve the classification performance.

5. Conclusions

This paper proposes a new method (ROENet) to classify malaria parasite on the blood smear automatically. The backbone of the ROENet is the pre-trained ResNet-18. The output nodes of the pre-trained ResNet-18 are 1000. However, the output nodes are two in this paper. Therefore, we perform some modifications to the backbone. We use randomized neural networks (RNNs) as the classifier in our proposed model, because the structure of RNN is simpler than ResNet-18. Three RNNs are used in ROENet, which are random vector functional link (RVFL), extreme learning machine (ELM), and Schmidt neural network (SNN). To improve the performance of ROENet, the results of ROENet are the ensemble outputs from three RNNs. We evaluate the proposed ROENet by five-fold cross-validation. The specificity (Sp), F1 score (F1), sensitivity (Se), and accuracy (Ac) are 96.68 ± 3.81%, 95.69 ± 2.65%, 94.79 ± 3.71%, and 95.73 ± 2.63%, respectively. The proposed ROENet is compared with other SOTA methods and provides the best classification performance among these methods, which proves that our model is an effective tool to classify malaria parasite on the blood smear.
Even though our model obtains excellent classification performance, there are still some limitations. (i) The dataset is still small. (ii) We only tested on one public dataset.
In future work, we will collect more datasets to test our model and continue to improve the performance and robustness of our model to better classify malaria parasite on the blood smear. Furthermore, we will try other latest methods to classify malaria parasite on the blood smear, such as VIT.

Author Contributions

Z.Z.: Conceptualization; software; data curation; writing—original draft; writing—review and editing; visualization. S.W.: methodology; software, validation; investigation; resources; writing—review and editing; supervision; funding acquisition. Y.Z.: methodology; formal analysis; investigation; data curation; writing—original draft; writing—review and editing; visualization; supervision; project administration; funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

The paper is partially supported by Hope Foundation for Cancer Research, UK (RM60G0680); Royal Society International Exchanges Cost Share Award, UK (RP202G0230); Medical Research Council Confidence in Concept Award, UK (MC_PC_17171); British Heart Foundation Accelerator Award, UK (AA/18/3/34220); Sino-UK Industrial Fund, UK (RP202G0289); Global Challenges Research Fund (GCRF), UK (P202PF11); LIAS Pioneering Partnerships award, UK (P202ED10); Data Science Enhancement Fund, UK (P202RE237); Guangxi Key Laboratory of Trusted Software, CN (kx201901).

Informed Consent Statement

The following information was supplied relating to ethical approvals: the Institutional Review Board (IRB#12972) at the National Library of Medicine (NLM), National Institutes of Health (NIH).

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://lhncbc.nlm.nih.gov/LHC-downloads/dataset.html (accessed on 16 January 2022).

Conflicts of Interest

The authors declare that the research was conducted without any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Manescu, P.; Shaw, M.J.; Elmi, M.; Neary-Zajiczek, L.; Claveau, R.; Pawar, V.; Kokkinos, I.; Oyinloye, G.; Bendkowski, C.; Oladejo, O.A. Expert-level automated malaria diagnosis on routine blood films with deep neural networks. Am. J. Hematol. 2020, 95, 883–891. [Google Scholar] [CrossRef] [PubMed]
  2. Yang, F.; Poostchi, M.; Yu, H.; Zhou, Z.; Silamut, K.; Yu, J.; Maude, R.J.; Jaeger, S.; Antani, S. Deep learning for smartphone-based malaria parasite detection in thick blood smears. IEEE J. Biomed. Health Inform. 2019, 24, 1427–1438. [Google Scholar] [CrossRef] [PubMed]
  3. Shoohi, L.M.; Saud, J.H. Dcgan for handling imbalanced malaria dataset based on over-sampling technique and using cnn. Med. Leg. Update 2020, 20, 1079–1085. [Google Scholar]
  4. Mehanian, C.; Jaiswal, M.; Delahunt, C.; Thompson, C.; Horning, M.; Hu, L.; Ostbye, T.; McGuire, S.; Mehanian, M.; Champlin, C. Computer-automated malaria diagnosis and quantitation using convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 116–125. [Google Scholar]
  5. Mukherjee, S.; Chatterjee, S.; Bandyopadhyay, O.; Biswas, A. Detection of Malaria Parasites in Thin Blood Smears Using CNN-Based Approach. In Computational Intelligence and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2021; pp. 19–27. [Google Scholar]
  6. Khadim, E.U.; Shah, S.A.; Wagan, R.A. Evaluation of Activation Functions in CNN Model for Detection of Malaria Parasite using Blood Smear Images. In Proceedings of the 2021 International Conference on Innovative Computing (ICIC), Lahore, Pakistan, 9–10 November 2021; pp. 1–6. [Google Scholar]
  7. Magotra, V.; Rohil, M.K. Malaria Diagnosis Using a Lightweight Deep Convolutional Neural Network. Int. J. Telemed. Appl. 2022, 2022, 4176982. [Google Scholar] [CrossRef] [PubMed]
  8. Marques, G.; Ferreras, A.; de la Torre-Diez, I. An ensemble-based approach for automated medical diagnosis of malaria using EfficientNet. Multimed. Tools Appl. 2022, 1–18. [Google Scholar] [CrossRef] [PubMed]
  9. Sarkar, S.; Sharma, R.; Shah, K. Malaria detection from RBC images using shallow Convolutional Neural Networks. arXiv 2020, arXiv:2010.11521. preprint. [Google Scholar]
  10. Raihan, M.; Nahid, A.-A. Malaria cell image classification by explainable artificial intelligence. Health Technol. 2022, 12, 47–58. [Google Scholar] [CrossRef]
  11. Abubakar, M.M.; Tuncer, T. Malaria microscopic cell image classification using custom-built cnn architecture. Int. J. Signal Process. 2021, 2, 1–8. [Google Scholar]
  12. Babikir, A.K.O.; Thron, C. Malaria Detection Using Machine Learning. In Artificial Intelligence for Data Science in Theory and Practice; Springer: Berlin/Heidelberg, Germany, 2022; pp. 139–153. [Google Scholar]
  13. Rinky, H.K.; Bhuiyan, R.R.; Rahman, H.T. Performance Comparison of CNN Architectures for Detecting Malaria Diseases. Bachelor’s Thesis, Brac University, Dhaka, Bangladesh, 2020. [Google Scholar]
  14. Balaram, A.; Silparaj, M.; Gajula, R. Detection of malaria parasite in thick blood smears using deep learning. Mater. Today Proc. 2022, in press. [Google Scholar] [CrossRef]
  15. Liang, Z.; Huang, J.X. Adaptive Cycle-consistent Adversarial Network for Malaria Blood Cell Image Synthetization. In Proceedings of the 2021 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 12–14 October 2021; pp. 1–7. [Google Scholar]
  16. Jiang, X.; Li, D.; Chen, H.; Zheng, Y.; Zhao, R.; Wu, L. Uni6D: A Unified CNN Framework without Projection Breakdown for 6D Pose Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11174–11184. [Google Scholar]
  17. Pao, Y.-H.; Park, G.-H.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 1994, 6, 163–180. [Google Scholar] [CrossRef]
  18. Schmidt, W.F.; Kraaijveld, M.A.; Duin, R.P. Feed forward neural networks with random weights. In Proceedings of the International Conference on Pattern Recognition, The Hague, The Netherlands, 30 August–3 September 1992; p. 1. [Google Scholar]
  19. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  20. Zhang, S.; Yu, Z.; Liu, L.; Wang, X.; Zhou, A.; Chen, K. Group R-CNN for Weakly Semi-supervised Object Detection with Points. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 9417–9426. [Google Scholar]
  21. Fang, J.; Lin, H.; Chen, X.; Zeng, K. A Hybrid Network of CNN and Transformer for Lightweight Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 1103–1112. [Google Scholar]
  22. Yasarla, R.; Patel, V.M. CNN-based Restoration of a Single Face Image Degraded by Atmospheric Turbulence. IEEE Trans. Biom. Behav. Identity Sci. 2022, 4, 222–233. [Google Scholar] [CrossRef]
  23. Wu, G.; Song, R.; Zhang, M.; Li, X.; Rosin, P.L. LiTMNet: A deep CNN for efficient HDR image reconstruction from a single LDR image. Pattern Recognit. 2022, 127, 108620. [Google Scholar] [CrossRef]
  24. Kumar, K.A.; Prasad, A.; Metan, J. A hybrid deep CNN-Cov-19-Res-Net Transfer learning architype for an enhanced Brain tumor Detection and Classification scheme in medical image processing. Biomed. Signal Process. Control. 2022, 76, 103631. [Google Scholar] [CrossRef]
  25. Chen, P.; Zhou, Z.; Yu, H.; Chen, K.; Yang, Y. Computerized-Assisted Scoliosis Diagnosis Based on Faster R-CNN and ResNet for the Classification of Spine X-Ray Images. Comput. Math. Methods Med. 2022, 2022, 3796202. [Google Scholar] [CrossRef] [PubMed]
  26. Showkat, S.; Qureshi, S. Efficacy of Transfer Learning-based ResNet models in Chest X-ray image classification for detecting COVID-19 Pneumonia. Chemom. Intell. Lab. Syst. 2022, 224, 104534. [Google Scholar] [CrossRef]
  27. Yan, Z.; Younes, R.; Forsyth, J. ResNet-Like CNN Architecture and Saliency Map for Human Activity Recognition. In Proceedings of the International Conference on Mobile Computing, Applications, and Services, Hangzhou, China, 18–19 September 2021; pp. 129–143. [Google Scholar]
  28. Quezada-Gaibor, D.; Torres-Sospedra, J.; Nurmi, J.; Koucheryavy, Y.; Huerta, J. Lightweight Hybrid CNN-ELM Model for Multi-building and Multi-floor Classification. arXiv 2022, arXiv:2204.10418. preprint. [Google Scholar]
  29. Zhou, Z.; Deng, W.; Wang, Y.; Zhu, Z. Classification of clothing images based on a parallel convolutional neural network and random vector functional link optimized by the grasshopper optimization algorithm. Text. Res. J. 2022, 92, 1415–1428. [Google Scholar] [CrossRef]
  30. Lu, J.; Zeng, W.; Zhang, L.; Shi, Y. A Novel Key Features Screening Method Based on Extreme Learning Machine for Alzheimer’s Disease Study. Front. Aging Neurosci. 2022, 14, 888575. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Some of the parasitized and uninfected images. (a) Parasitized images; (b) uninfected images.
Figure 1. Some of the parasitized and uninfected images. (a) Parasitized images; (b) uninfected images.
Electronics 11 02040 g001
Figure 2. The flowchart of the proposed ROENet.
Figure 2. The flowchart of the proposed ROENet.
Electronics 11 02040 g002
Figure 3. The structure of the residual learning.
Figure 3. The structure of the residual learning.
Electronics 11 02040 g003
Figure 4. The modifications of the backbone.
Figure 4. The modifications of the backbone.
Electronics 11 02040 g004
Figure 5. The structures of three RNNs. (a) ELM; (b) RVFL; (c) SNN.
Figure 5. The structures of three RNNs. (a) ELM; (b) RVFL; (c) SNN.
Electronics 11 02040 g005
Figure 6. The comparison of different backbones.
Figure 6. The comparison of different backbones.
Electronics 11 02040 g006
Figure 7. Effects of output ensemble.
Figure 7. Effects of output ensemble.
Electronics 11 02040 g007
Figure 8. The comparison of the proposed model with the fine-tuned model.
Figure 8. The comparison of the proposed model with the fine-tuned model.
Electronics 11 02040 g008
Figure 9. The comparison with other state-of-the-art methods.
Figure 9. The comparison with other state-of-the-art methods.
Electronics 11 02040 g009
Table 1. Acronym and Full Explanation.
Table 1. Acronym and Full Explanation.
AcronymFull Explanation
AcAccuracy
AvrAverage
BNBatch normalization
CNNConvolution neural network
ELMExtreme learning machine
F1F1 score
FCFully connected
MLMachine learning
RVFLRandom vector functional link
RNNsRandomized neural networks
SeSensitivity
SNNSchmidt neural network
SpSpecificity
StdStandard deviation
Table 2. The pseudocode of ROENet.
Table 2. The pseudocode of ROENet.
Pseudocode of ROENet
Electronics 11 02040 i001
Table 3. The hyper-parameters of the ROENet.
Table 3. The hyper-parameters of the ROENet.
Hyper-ParametersValue
Max-epoch4
Learning rate 10 4
Minibatch size128
Number of the hidden nodes u 400
Table 4. The classification performance.
Table 4. The classification performance.
MethodsFoldAcSeSpF1
ROENet (Ours)F 195.4194.3496.4895.36
F 295.5995.1496.0495.57
F 396.0395.1496.9295.99
F 496.0694.9997.1396.02
F 595.5794.3496.8195.52
Avr95.7394.7996.6895.69
Std±2.63±3.71±3.81±2.65
AlexNet-OEF 195.6195.7195.4395.62
F 295.6396.0195.2595.64
F 395.1795.7994.5695.20
F 494.9795.4694.4894.99
F 595.6695.7995.5395.67
Avr95.4195.7595.0595.42
Std±2.84±1.77±4.43±2.77
ResNet50-OEF 195.3494.7495.9495.31
F 294.9994.5695.4394.97
F 395.5794.9996.1595.55
F 495.4194.9995.8395.39
F 595.2894.8895.6895.26
Avr95.3294.8395.8195.30
Std±1.91±1.64±2.43±1.90
ResNet-18-ELMF 194.9994.2395.7594.95
F 295.1794.6695.6895.15
F 396.0595.4696.6396.02
F 495.6594.7496.5595.61
F 595.5294.6396.4195.48
Avr95.4894.7496.2095.44
Std±3.72±3.99±4.06±3.72
ResNet-18-RVFLF 195.1093.7296.4895.03
F 295.6895.0396.3395.65
F 395.9595.2196.7095.92
F 496.0194.8897.1395.96
F 595.5694.0997.0295.49
Avr95.6694.5996.7395.61
Std±3.26±5.78±3.06±3.38
ResNet-18-SNNF 195.3794.4896.2695.33
F 295.3794.9995.7595.35
F 395.8394.9296.7395.79
F 495.8395.0796.5995.80
F 595.1794.1996.1595.13
Avr95.5194.7396.3095.48
Std±2.68±3.39±3.45±2.68
Fine-tuned ResNet-18F 195.2394.3896.0895.19
F 295.4494.8196.0895.42
F 395.9495.3696.5295.91
F 495.8394.8596.8195.79
F 595.3694.0996.6395.30
Avr95.5694.7096.4295.52
Std±2.76±4.35±2.96±2.80
OE means output ensemble.
Table 5. The comparison with other SOTA methods.
Table 5. The comparison with other SOTA methods.
MethodsAcSeSpF1
DeepMCNN [1]-92.0090.00-
Customized CNN [2]93.4692.5994.33-
DCGAN [3]94.5090.0099.00-
Computer-Automated-CNN [4]-91.6094.10-
3-layer CNN [9]95.3294.3096.3495.27
ROENet (Ours)95.7394.7996.6895.69
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, Z.; Wang, S.; Zhang, Y. ROENet: A ResNet-Based Output Ensemble for Malaria Parasite Classification. Electronics 2022, 11, 2040. https://doi.org/10.3390/electronics11132040

AMA Style

Zhu Z, Wang S, Zhang Y. ROENet: A ResNet-Based Output Ensemble for Malaria Parasite Classification. Electronics. 2022; 11(13):2040. https://doi.org/10.3390/electronics11132040

Chicago/Turabian Style

Zhu, Ziquan, Shuihua Wang, and Yudong Zhang. 2022. "ROENet: A ResNet-Based Output Ensemble for Malaria Parasite Classification" Electronics 11, no. 13: 2040. https://doi.org/10.3390/electronics11132040

APA Style

Zhu, Z., Wang, S., & Zhang, Y. (2022). ROENet: A ResNet-Based Output Ensemble for Malaria Parasite Classification. Electronics, 11(13), 2040. https://doi.org/10.3390/electronics11132040

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop