A Weighted Voting Ensemble Self-Labeled Algorithm for the Detection of Lung Abnormalities from X-Rays
Abstract
:1. Introduction
2. Related Work
3. A New Weighted Voting Ensemble Self-Labeled Algorithm
Algorithm 1: WvEnSL |
Input: L − Set of labeled instances (Training labeled set). U − Set of unlabeled instances (Training unlabeled set). T − Set of unlabeled test instances (Testing set). D − Set of instances for evaluation (Evaluation set). − Set of self-labeled classifiers which constitute the ensemble. Output: The labels of instances in the testing set. /* Phase I: Training */ Step 1: for to N do Step 2: Train using the labeled L and the unlabeled dataset U. Step 3: end for /* Phase II: Evaluation */ Step 4: for to N do Step 5: Apply on the evaluation set D. Step 6: for to M do Step 7: Calculate the weight Step 8: end for Step 9: end for /* Phase III: Weighted-Voting Prediction */ Step 10: for each do Step 11: for to N do Step 12: Apply classifier on x. Step 13: end for Step 14: Predict the label of x using |
4. Experimental Methodology
4.1. Datasets
- Chest X-ray (Pneumonia) dataset: The dataset contains 5830 chest X-ray images (anterior-posterior) which were selected from retrospective cohorts of pediatric patients of one to five years old from Guangzhou Women and Children’s Medical Center, Guangzhou. All chest X-ray imaging was performed as part of patients’ routine clinical care. For the analysis of chest X-ray images, all chest radiographs were initially screened for quality control by removing all low quality or unreadable scans. The diagnoses for the images were then graded by two expert physicians before being cleared for training the artificial intelligence system. In order to account for any grading errors, the evaluation set was also checked by a third expert. The dataset was partitioned into two sets (training/testing). The training set consisting of 5216 examples (1341 normal, 3875 pneumonia) and the testing set with 624 examples (234 normal, 390 pneumonia) as in [28].
- Shenzhen lung mask (Tuberculosis) dataset: Shenzhen Hospital is one of the largest hospitals in China for infectious diseases with a focus both on their prevention, as well as treatment. The X-rays were collected within a one-month period, mostly in September 2012, as a part of the daily routine, using a Philips DR Digital Diagnost system. The dataset was constructed by manually-segmented lung masks for the Shenzhen Hospital X-ray set as presented in [29]. These segmented lung masks were originally utilized for the description of the lung segmentation technique in combination with lossless and lossy data augmentation. The segmentation masks for the Shenzhen Hospital X-ray set were manually prepared by students and teachers of the Computer Engineering Department, Faculty of Informatics and Computer Engineering, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute” [29]. The set contained 279 normal CXRs and 287 abnormal ones with tuberculosis. All classification algorithms were evaluated using the stratified ten-fold cross-validation.
- CT Medical images dataset: This data collectioncontains 100 images [30] which constitute part of a much larger effort, focused on connecting cancer phenotypes to genotypes by providing clinical images matched to subjects from the cancer genome Atlas [31]. The images consist of the middle slice of all Computed Tomography (CT) images taken from 69 different patients. The dataset is designed to allow different methods to be evaluated for examining the trends in CT image data associated with using contrast and patient age. The basic idea is to identify image textures, statistical patterns and features correlating strongly with these traits and possibly build simple tools for automatically classifying these images when they have been misclassified (or finding outliers which could be suspicious cases, bad measurements, or poorly calibrated machines). All classification algorithms were evaluated using the stratified ten-fold cross-validation.
4.2. Performance Evaluation of WvEnSL against Ensemble Self-Labeled Algorithms
- “CST-Voting (SMO)” stands for an ensemble of Co-training, Self-training and Tri-training with SMO as base learner using majority voting [2].
- “WvEnSL (SMO)” stands for Algorithm WvEnSL using the same components classifiers as CST-Voting (SMO).
- “CST-Voting (C4.5)” stands for an ensemble of Co-training, Self-training and Tri-training with C4.5 as base learner using majority voting [2].
- “WvEnSL (C4.5)” stands for Algorithm WvEnSL using the same components classifiers as CST-Voting (C4.5).
- “CST-Voting (kNN)” stands for an ensemble of Co-training, Self-training and Tri-training with kNN as base learner using majority voting [2].
- “WvEnSL (kNN)” stands for Algorithm WvEnSL using the same components classifiers as CST-Voting (kNN).
- “DTCo” stands for an ensemble of Democratic-Co learning, Tri-training and Co-Bagging with C4.5 as base learner using majority voting [16].
- “WvEnSL” stands for Algorithm WvEnSL using the same components classifiers as DTCo.
- “EnSL” stands for an ensemble of Self-training, Democratic-Co learning, Tri-training and Co-Bagging with C4.5 as base learner using majority voting [17].
- “WvEnSL” stands for Algorithm WvEnSL using the same components classifiers as EnSL.
4.3. Performance Evaluation of WvEnSL against Classical Supervised Algorithms
- we selected WvEnSL from all versions of the proposed algorithm since it presented the best overall performance.
- all supervised algorithms were trained using with 100% of the training set while WvEnSL was trained using of the training set.
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Van Ginneken, B.; Stegmann, M.B.; Loog, M. Segmentation of anatomical structures in chest radiographs using supervised methods: A comparative study on a public database. Medical Image Anal. 2006, 10, 19–40. [Google Scholar] [CrossRef] [PubMed]
- Livieris, I.; Kanavos, A.; Tampakas, V.; Pintelas, P. An ensemble SSL algorithm for efficient chest X-ray image classification. J. Imaging 2018, 4, 95. [Google Scholar] [CrossRef]
- Zhu, X.; Goldberg, A. Introduction to semi-supervised learning. Synth. Lect. Artif. Intell. Mach. Learn. 2009, 3, 1–130. [Google Scholar] [CrossRef]
- Chapelle, O.; Scholkopf, B.; Zien, A. Semi-supervised learning. IEEE Trans. Neural Netw. 2009, 20, 542. [Google Scholar] [CrossRef]
- Levatic, J.; Dzeroski, S.; Supek, F.; Smuc, T. Semi-supervised learning for quantitative structure-activity modeling. Informatica 2013, 37, 173–179. [Google Scholar]
- Levatić, J.; Ceci, M.; Kocev, D.; Džeroski, S. Semi-supervised classification trees. J. Intell. Inf. Syst. 2017, 49, 461–486. [Google Scholar] [CrossRef]
- Livieris, I.; Kanavos, A.; Tampakas, V.; Pintelas, P. An auto-adjustable semi-supervised self-training algorithm. Algorithm 2018, 11, 139. [Google Scholar] [CrossRef]
- Livieris, I.; Kiriakidou, N.; Kanavos, A.; Tampakas, V.; Pintelas, P. On ensemble SSL algorithms for credit scoring problem. Informatics 2018, 5, 40. [Google Scholar] [CrossRef]
- Triguero, I.; García, S.; Herrera, F. Self-labeled techniques for semi-supervised learning: Taxonomy, software and empirical study. Knowl. Inf. Syst. 2015, 42, 245–284. [Google Scholar] [CrossRef]
- Yarowsky, D. Unsupervised word sense disambiguation rivaling supervised methods. In Proceedings of the 33rd Annual Meeting of the Association For Computational Linguistics, Cambridge, MA, USA, 26–30 June 1995; pp. 189–196. [Google Scholar]
- Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the 11th Annual Conference on Computational Learning Theory, Madison, WI, USA, 24–26 July 1998; pp. 92–100. [Google Scholar]
- Zhou, Y.; Goldman, S. Democratic co-learning. In Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence (ICTAI), Boca Raton, FL, USA, 15–17 November 2014; IEEE: Piscataway, NI, USA, 2004; pp. 594–602. [Google Scholar]
- Zhou, Z.; Li, M. Tri-training: Exploiting unlabeled data using three classifiers. IEEE Trans. Knowl. Data Eng. 2005, 17, 1529–1541. [Google Scholar] [CrossRef]
- Li, M.; Zhou, Z. Improve computer-aided diagnosis with machine learning techniques using undiagnosed samples. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2007, 37, 1088–1098. [Google Scholar] [CrossRef]
- Hady, M.; Schwenker, F. Combining committee-based semi-supervised learning and active learning. J. Comput. Sci. Technol. 2010, 25, 681–698. [Google Scholar] [CrossRef]
- Livieris, I.; Kotsilieris, T.; Anagnostopoulos, I.; Tampakas, V. DTCo: An ensemble SSL algorithm for X-rays classification. In Advances in Experimental Medicine and Biology; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
- Livieris, I.; Kanavos, A.; Pintelas, P. Detecting lung abnormalities from X-rays using and improved SSL algorithm. Electron. Notes Theor. Comput. Sci. 2019. accepted for publication. [Google Scholar]
- Livieris, I. A new ensemble self-labeled semi-supervised algorithm. Informatica 2018. accepted for publication. [Google Scholar]
- Jaeger, S.; Karargyris, A.; Candemir, S.; Folio, L.; Siegelman, J.; Callaghan, F.; Xue, Z.; Palaniappan, K.; Singh, R.; Antani, S.; et al. Automatic tuberculosis screening using chest radiographs. IEEE Trans. Med. Imaging 2014, 33, 233–245. [Google Scholar] [CrossRef]
- Melendez, J.; van Ginneken, B.; Maduskar, P.; Philipsen, R.; Reither, K.; Breuninger, M.; Adetifa, I.; Maane, R.; Ayles, H.; Sánchez, C. A novel multiple-instance learning-based approach to computer-aided detection of tuberculosis on chest X-rays. IEEE Trans. Med. Imaging 2015, 34, 179–192. [Google Scholar] [CrossRef]
- Alam, J.; Alam, S.; Hossan, A. Multi-Stage Lung Cancer Detection and Prediction Using Multi-class SVM Classifier. In Proceedings of the 2018 International Conference on Computer, Communication, Chemical, Material and Electronic Engineering, Rajshahi, Bangladesh, 8–9 February 2018; IEEE: Piscataway, NI, USA, 2018; pp. 1–4. [Google Scholar]
- Madani, A.; Moradi, M.; Karargyris, A.; Syeda-Mahmood, T. Semi-supervised learning with generative adversarial networks for chest X-ray classification with ability of data domain adaptation. In Proceedings of the 15th IEEE International Symposium on Biomedical Imaging, Washington, DC, USA, 4–7 April 2018; IEEE: Piscataway, NI, USA, 2018; pp. 1038–1042. [Google Scholar]
- Guan, Q.; Huang, Y. Multi-label chest X-ray image classification via category-wise residual attention learning. Pattern Recognit. Lett. 2018. [Google Scholar] [CrossRef]
- Dietterich, T. Ensemble methods in machine learning. In Multiple Classifier Systems; Kittler, J., Roli, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2001; Volume 1857, pp. 1–15. [Google Scholar]
- Rokach, L. Pattern Classification Using Ensemble Methods; World Scientific Publishing Company: Singapore, 2010. [Google Scholar]
- Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
- Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I. The WEKA data mining software: An update. SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar] [CrossRef]
- Kermany, D.; Goldbaum, M.; Cai, W.; Valentim, C.; Liang, H.; Baxter, S.; McKeown, A.; Yang, G.; Wu, X.; Yan, F. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef]
- Stirenko, S.; Kochura, Y.; Alienin, O.; Rokovyi, O.; Gang, P.; Zeng, W.; Gordienko, Y. Chest X-ray analysis of tuberculosis by deep learning with segmentation and augmentation. arXiv, 2018; arXiv:1803.01199. [Google Scholar]
- Albertina, B.; Watson, M.; Holback, C.; Jarosz, R.; Kirk, S.; Lee, Y.; Lemmerman, J. Radiology data from the cancer Genome Atlas Lung Adenocarcinoma [TCGA-LUAD] collection. Cancer Imaging Arch. 2016. [Google Scholar]
- Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M. The Cancer Imaging Archive (TCIA): maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
- Wang, Y.; Xu, X.; Zhao, H.; Hua, Z. Semi-supervised learning based on nearest neighbor rule and cut edges. Knowl.-Based Syst. 2010, 23, 547–554. [Google Scholar] [CrossRef]
- Platt, J. Advances in Kernel Methods—Support Vector Learning; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
- Quinlan, J. C4.5: Programs for Machine Learning; Morgan Kaufmann: San Francisco, CA, USA, 1993. [Google Scholar]
- Aha, D. Lazy Learning; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1997. [Google Scholar]
- Wu, X.; Kumar, V.; Quinlan, J.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.; Ng, A.; Liu, B.; Yu, P.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef]
- Hodges, J.; Lehmann, E. Rank methods for combination of independent experiments in analysis of variance. Ann. Math. Stat. 1962, 33, 482–497. [Google Scholar] [CrossRef]
- Finner, H. On a monotonicity problem in step-down multiple test procedures. J. Am. Stat. Assoc. 1993, 88, 920–923. [Google Scholar] [CrossRef]
- Li, S.; Wang, Z.; Zhou, G.; Lee, S. Semi-supervised learning for imbalanced sentiment classification. In Proceedings of the IJCAI Proceedings-International Joint Conference on Artificial Intelligence, Barcelona, Catalonia, Spain, 16–22 July 2011; Volume 22, p. 1826. [Google Scholar]
- Jeni, L.A.; Cohn, J.F.; De La Torre, F. Facing imbalanced data—Recommendations for the use of performance metrics. In Proceedings of the Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; IEEE: Piscataway, NI, USA, 2013; pp. 245–251. [Google Scholar]
- Levatić, J.; Ceci, M.; Kocev, D.; Džeroski, S. Self-training for multi-target regression with tree ensembles. Knowl.-Based Syst. 2017, 123, 41–60. [Google Scholar] [CrossRef]
- Levatić, J.; Kocev, D.; Džeroski, S. The importance of the label hierarchy in hierarchical multi-label classification. J. Intell. Inf. Syst. 2015, 45, 247–271. [Google Scholar] [CrossRef]
- Levatić, J.; Kocev, D.; Ceci, M.; Džeroski, S. Semi-supervised trees for multi-target regression. Inf. Sci. 2018, 450, 109–127. [Google Scholar] [CrossRef]
Algorithm | Parameters | |
---|---|---|
SMO | Supervised base learner | , |
, | ||
Pearson VII function-based kernel, | ||
, | ||
. | ||
C4.5 | Supervised base learner | Confidence level: , |
Minimum number of item-sets per leaf: , | ||
Prune after the tree building. | ||
kNN | Supervised base learner | , |
Euclidean distance. | ||
Self-training | Self-labeled (single classifier) | , |
. | ||
Co-training | Self-labeled (multiple classifier) | , |
Tri-training | Self-labeled (multiple classifier) | No parameters specified. |
Co-Bagging | Self-labeled (multiple classifier) | , |
Ensemble learning = Bagging. | ||
Democratic-Co | Self-labeled (multiple classifier) | Classifiers = kNN, C4.5, NB. |
CST-Voting | Ensemble of self-labeled | No parameters specified. |
DTCo | Ensemble of self-labeled | No parameters specified. |
EnSL | Ensemble of self-labeled | No parameters specified. |
Algorithm | Ratio = 10% | Ratio = 20% | Ratio = 30% | Ratio = 40% | ||||
---|---|---|---|---|---|---|---|---|
Acc | Acc | Acc | Acc | |||||
CST-Voting (SMO) | 83.08% | 75.32% | 83.26% | 75.64% | 83.39% | 75.80% | 83.39% | 75.80% |
WvEnSL (SMO) | 83.39% | 75.80% | 83.48% | 75.96% | 83.76% | 76.44% | 83.85% | 76.60% |
CST-Voting (C4.5) | 85.52% | 79.97% | 85.85% | 80.45% | 86.68% | 81.73% | 86.58% | 81.57% |
WvEnSL (C4.5) | 85.65% | 80.13% | 86.08% | 80.77% | 86.78% | 81.89% | 86.92% | 82.05% |
CST-Voting (kNN) | 82.91% | 75.48% | 83.09% | 75.80% | 83.15% | 75.96% | 83.73% | 76.76% |
WvEnSL (kNN) | 83.63% | 76.60% | 83.73% | 76.76% | 83.95% | 77.08% | 84.23% | 77.56% |
DTCo | 86.79% | 81.41% | 87.21% | 82.05% | 87.21% | 82.05% | 87.74% | 82.85% |
WvEnSL | 87.12% | 81.89% | 87.44% | 82.37% | 87.54% | 82.53% | 87.97% | 83.17% |
EnSL | 87.19% | 82.05% | 86.92% | 81.57% | 87.34% | 82.21% | 87.61% | 82.69% |
WvEnSL | 87.51% | 82.53% | 87.70% | 82.69% | 88.23% | 83.49% | 88.17% | 83.49% |
Algorithm | Ratio = 10% | Ratio = 20% | Ratio = 30% | Ratio = 40% | ||||
---|---|---|---|---|---|---|---|---|
Acc | Acc | Acc | Acc | |||||
CST-Voting (SMO) | 69.27% | 69.43% | 68.65% | 68.37% | 69.50% | 69.61% | 70.42% | 70.32% |
WvEnSL (SMO) | 69.73% | 69.79% | 69.73% | 69.79% | 70.32% | 70.32% | 71.00% | 70.85% |
CST-Voting (C4.5) | 66.67% | 67.31% | 68.19% | 68.02% | 67.51% | 68.20% | 69.52% | 69.79% |
WvEnSL (C4.5) | 67.86% | 68.20% | 69.26% | 69.26% | 69.63% | 69.79% | 69.98% | 70.14% |
CST-Voting (kNN) | 65.71% | 66.08% | 66.43% | 66.96% | 68.21% | 68.55% | 68.93% | 69.26% |
WvEnSL (kNN) | 65.83% | 66.25% | 67.14% | 67.49% | 68.57% | 68.90% | 69.40% | 69.61% |
DTCo | 69.73% | 69.79% | 69.96% | 69.96% | 71.45% | 71.20% | 71.80% | 71.55% |
WvEnSL | 69.73% | 69.79% | 70.19% | 70.14% | 71.58% | 71.38% | 71.80% | 71.55% |
EnSL | 69.73% | 69.79% | 69.96% | 69.96% | 71.00% | 70.85% | 71.58% | 71.38% |
WvEnSL | 69.73% | 69.79% | 70.19% | 70.14% | 71.58% | 71.38% | 72.03% | 71.73% |
Algorithm | Ratio = 10% | Ratio = 20% | Ratio = 30% | Ratio = 40% | ||||
---|---|---|---|---|---|---|---|---|
Acc | Acc | Acc | Acc | |||||
CST-Voting (SMO) | 66.67% | 66.00% | 70.00% | 70.00% | 73.08% | 72.00% | 75.00% | 74.00% |
WvEnSL (SMO) | 68.00% | 68.00% | 71.29% | 71.00% | 73.79% | 73.00% | 75.73% | 75.00% |
CST-Voting (C4.5) | 67.96% | 67.00% | 71.84% | 71.00% | 73.79% | 73.00% | 73.79% | 73.00% |
WvEnSL (C4.5) | 69.90% | 69.00% | 73.79% | 73.00% | 75.00% | 74.00% | 75.73% | 75.00% |
CST-Voting (kNN) | 66.00% | 66.00% | 69.90% | 69.00% | 73.79% | 73.00% | 73.27% | 73.00% |
WvEnSL (kNN) | 66.67% | 67.00% | 70.59% | 70.00% | 72.00% | 72.00% | 74.75% | 75.00% |
DTCo | 66.02% | 65.00% | 69.90% | 69.00% | 72.55% | 72.00% | 74.29% | 73.00% |
WvEnSL | 67.33% | 67.00% | 71.29% | 71.00% | 72.55% | 72.00% | 76.92% | 76.00% |
EnSL | 64.08% | 63.00% | 71.84% | 71.00% | 74.29% | 73.00% | 74.29% | 73.00% |
WvEnSL | 69.90% | 69.00% | 75.73% | 75.00% | 76.47% | 76.00% | 77.67% | 77.00% |
Algorithm | FAR | Finner Post-Hoc Test | |
---|---|---|---|
-Value | Null Hypothesis | ||
WvEnSL | 15.667 | - | - |
WvEnSL | 34.958 | 0.174312 | accepted |
WvEnSL (C4.5) | 44.208 | 0.049863 | rejected |
EnSL | 47.958 | 0.029437 | rejected |
DTCo | 51.125 | 0.018734 | rejected |
CST-Voting (C4.5) | 64.042 | 0.001184 | rejected |
WvEnSL (SMO) | 71.417 | 0.000194 | rejected |
CST-Voting (SMO) | 88.292 | 0.000001 | rejected |
WvEnSL (kNN) | 89.083 | 0.000001 | rejected |
CST-Voting (kNN) | 98.250 | 0.000001 | rejected |
Algorithm | Pneumonia | Tuberculosis | CT Medical | |||
---|---|---|---|---|---|---|
Acc | Acc | Acc | ||||
SMO | 74.03% | 76.76% | 71.41% | 71.37% | 74.91% | 75.00% |
C4.5 | 72.41% | 74.83% | 62.32% | 62.36% | 79.82% | 80.00% |
3NN | 72.32% | 74.51% | 67.51% | 67.49% | 67.08% | 67.00% |
Voting | 73.34% | 76.12% | 71.00% | 71.02% | 74.07% | 74.00% |
WvEnSL | 88.17% | 83.49% | 72.03% | 71.73% | 77.67% | 77.00% |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Livieris, I.E.; Kanavos, A.; Tampakas, V.; Pintelas, P. A Weighted Voting Ensemble Self-Labeled Algorithm for the Detection of Lung Abnormalities from X-Rays. Algorithms 2019, 12, 64. https://doi.org/10.3390/a12030064
Livieris IE, Kanavos A, Tampakas V, Pintelas P. A Weighted Voting Ensemble Self-Labeled Algorithm for the Detection of Lung Abnormalities from X-Rays. Algorithms. 2019; 12(3):64. https://doi.org/10.3390/a12030064
Chicago/Turabian StyleLivieris, Ioannis E., Andreas Kanavos, Vassilis Tampakas, and Panagiotis Pintelas. 2019. "A Weighted Voting Ensemble Self-Labeled Algorithm for the Detection of Lung Abnormalities from X-Rays" Algorithms 12, no. 3: 64. https://doi.org/10.3390/a12030064
APA StyleLivieris, I. E., Kanavos, A., Tampakas, V., & Pintelas, P. (2019). A Weighted Voting Ensemble Self-Labeled Algorithm for the Detection of Lung Abnormalities from X-Rays. Algorithms, 12(3), 64. https://doi.org/10.3390/a12030064