A Deep Modality-Specific Ensemble for Improving Pneumonia Detection in Chest X-rays
Abstract
:1. Introduction
1.1. Related Works
1.2. Rationale for the Study
1.3. Contributions of the Study
- (i).
- To the best of our knowledge, this is the first study that studies the impact of using CXR modality-specific classifier backbones in a RetinaNet-based object detection model, particularly applied to detecting pneumonia-consistent findings in CXRs.
- (ii).
- We train state-of-the-art DL classifiers on large collections of CXR images to develop CXR modality-specific models. Next, we use these models as the classifier backbone in the RetinaNet object detection network. We also initialize this backbone using random weights and ImageNet-pretrained weights to compare detection performance. Finally, we construct an ensemble of the aforementioned models resulting in improved detection of pneumonia-consistent findings.
- (iii).
- Through this approach, we aim to study the combined benefits of various weight initializations for classifier backbones and construct an ensemble of the best-performing models to improve detection performance. The models’ performance is evaluated in terms of mAP and statistical significance is reported in terms of confidence intervals (CIs) and p-values.
2. Materials and Methods
2.1. Data Collection and Preprocessing
- (i).
- CheXpert CXR [18]: The dataset includes 223,648 frontal and lateral CXR images that are collected from 65,240 patients at Stanford Hospital, California, USA. The CXRs are labeled for 14 cardiopulmonary disease manifestations, the details are extracted from the associated radiology reports using an automated labeling algorithm.
- (ii).
- TBX11K CXR [19]: This collection includes 11,200 CXRs collected from normal patients and those with other cardiopulmonary abnormalities. The abnormal CXRs are collected from patients tested with the microbiological gold standard. There are 5000 CXRs showing no abnormalities and 6200 CXRs showing other abnormal findings including those collected from sick patients (n = 5000), active Tuberculosis (TB) (n = 924), latent Tuberculosis (n = 212), active and latent TB (n = 54), and other uncertain (n = 10) cases. The regions showing TB-consistent manifestations are labeled for the abnormal regions using coarse rectangular bounding boxes.
- (iii).
- RSNA CXR [20]: This CXR collection is released by RSNA for the RSNA Kaggle Pneumonia detection challenge. The collection consists of 26,684 CXRs that include 6012 CXR images showing pneumonia-consistent manifestations, 8851 CXRs showing no abnormal findings, and 11,821 CXRs showing other cardiopulmonary abnormalities. The CXRs showing pneumonia-consistent findings are labeled for abnormal regions using rectangular bounding boxes and are made available for the detection challenge.
2.2. Model Architecture
2.2.1. CXR Modality-Specific Retraining
2.2.2. RetinaNet Architecture
2.2.3. Ensemble of RetinaNet Models with Various Backbones
2.2.4. Loss Functions and Evaluation Metrics
CXR Image Modality-Specific Retraining
RetinaNet-Based Detection of Pneumonia-Consistent Findings
2.3. Statistical Analysis
3. Results and Discussion
3.1. Classification Performance during CXR Image Modality-Specific Retraining
3.2. Detection Performance Using RetinaNet Models and Their Ensembles
4. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell 2018, 172, 1122–1131.e9. [Google Scholar] [CrossRef] [PubMed]
- Rajaraman, S.; Candemir, S.; Thoma, G.; Antani, S. Visualizing and Explaining Deep Learning Predictions for Pneumonia Detection in Pediatric Chest Radiographs. Proc. SPIE 2019, 10950, 109500S. [Google Scholar]
- Nishio, M.; Noguchi, S.; Matsuo, H.; Murakami, T. Automatic Classification between COVID-19 Pneumonia, Non-COVID-19 Pneumonia, and the Healthy on Chest X-Ray Image: Combination of Data Augmentation Methods. Sci. Rep. 2020, 10, 17532. [Google Scholar] [CrossRef]
- Balabanova, Y.; Coker, R.; Fedorin, I.; Zakharova, S.; Plavinskij, S.; Krukov, N.; Atun, R.; Drobniewski, F. Variability in Interpretation of Chest Radiographs among Russian Clinicians and Implications for Screening Programmes: Observational Study. BMJ 2005, 331, 379–382. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Rajaraman, S.; Candemir, S.; Kim, I.; Thoma, G.; Antani, S.; Rajaraman, S.; Candemir, S.; Kim, I.; Thoma, G.; Antani, S. Visualization and Interpretation of Convolutional Neural Network Predictions in Detecting Pneumonia in Pediatric Chest Radiographs. Appl. Sci. 2018, 8, 1715. [Google Scholar] [CrossRef] [Green Version]
- Mouhafid, M.; Salah, M.; Yue, C.; Xia, K. Deep Ensemble Learning-Based Models for Diagnosis of COVID-19 from Chest CT Images. Healthcare 2022, 10, 166. [Google Scholar] [CrossRef] [PubMed]
- Pham, V.T.; Tran, C.M.; Zheng, S.; Vu, T.M.; Nath, S. Chest X-Ray Abnormalities Localization via Ensemble of Deep Convolutional Neural Networks. Int. Conf. Adv. Technol. Commun. 2021, 2021, 125–130. [Google Scholar] [CrossRef]
- Xie, X.; Liao, Q.; Ma, L.; Jin, X. Gated Feature Pyramid Network for Object Detection. Lect. Notes Comput. Sci. 2018, 11259, 199–208. [Google Scholar] [CrossRef]
- Mao, L.; Yumeng, T.; Lina, C. Pneumonia detection in chest X-rays: A deep learning approach based on ensemble RetinaNet and mask R-CNN. In Proceedings of the 8th International Conference on Advanced Cloud and Big Data, Taiyuan, China, 19–20 September 2020; pp. 213–218. [Google Scholar] [CrossRef]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ko, H.; Ha, H.; Cho, H.; Seo, K.; Lee, J. Pneumonia detection with weighted voting ensemble of CNN models. In Proceedings of the 2nd International Conference on Artificial Intelligence and Big Data (ICAIBD), Chengdu, China, 26–29 May 2019; pp. 306–310. [Google Scholar] [CrossRef]
- Fei-Fei, L.; Deng, J.; Li, K. ImageNet: Constructing a Large-Scale Image Database. J. Vis. 2010, 9, 1037. [Google Scholar] [CrossRef]
- Suzuki, K. Overview of Deep Learning in Medical Imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef] [PubMed]
- Rajaraman, S.; Sornapudi, S.; Kohli, M.; Antani, S. Assessment of an ensemble of machine learning models toward abnormality detection in chest radiographs. In Proceedings of the 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019. [Google Scholar]
- Rajaraman, S.; Folio, L.R.; Dimperio, J.; Alderson, P.O.; Antani, S.K. Improved Semantic Segmentation of Tuberculosis—Consistent Findings in Chest x-Rays Using Augmented Training of Modality-Specific u-Net Models with Weak Localizations. Diagnostics 2021, 11, 616. [Google Scholar] [CrossRef] [PubMed]
- Yadav, O.; Passi, K.; Jain, C.K. Using deep learning to classify x-ray images of potential tuberculosis patients. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018. [Google Scholar]
- Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilcus, S.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.; Shpanskaya, K.; et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. Proc. Conf. AAAI Artif. Intell. 2019, 33, 590–597. [Google Scholar] [CrossRef]
- Liu, Y.; Wu, Y.H.; Ban, Y.; Wang, H.; Cheng, M.M. Rethinking computer-aided tuberculosis diagnosis. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
- Shih, G.; Wu, C.C.; Halabi, S.S.; Kohli, M.D.; Prevedello, L.M.; Cook, T.S.; Sharma, A.; Amorosa, J.K.; Arteaga, V.; Galperin-Aizenberg, M.; et al. Augmenting the National Institutes of Health Chest Radiograph Dataset with Expert Annotations of Possible Pneumonia. Radiol. Artif. Intell. 2019, 1, e180041. [Google Scholar] [CrossRef] [PubMed]
- Rajaraman, S.; Jaeger, S.; Thoma, G.R.; Antani, S.K.; Silamut, K.; Maude, R.J.; Hossain, M.A. Understanding the Learned Behavior of Customized Convolutional Neural Networks toward Malaria Parasite Detection in Thin Blood Smear Images. J. Med. Imaging 2018, 5, 034501. [Google Scholar] [CrossRef] [PubMed]
- Ganesan, P.; Rajaraman, S.; Long, R.; Ghoraani, B.; Antani, S. Assessment of data augmentation strategies toward performance improvement of abnormality classification in chest radiographs. In Proceedings of the 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019. [Google Scholar]
- Rajaraman, S.; Antani, S.K.; Poostchi, M.; Silamut, K.; Hossain, M.A.; Maude, R.J.; Jaeger, S.; Thoma, G.R.; Hossain, A.; Maude, R.J.; et al. Pre-Trained Convolutional Neural Networks as Feature Extractors toward Improved Malaria Parasite Detection in Thin Blood Smear Images. PeerJ 2018, 6, e4568. [Google Scholar] [CrossRef] [PubMed]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
- Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef] [Green Version]
- Liu, C.; Yu, S.; Yu, M.; Wei, B.; Li, B.; Li, G.; Huang, W. Adaptive smooth L1 loss: A better way to regress scene texts with extreme aspect ratios. In Proceedings of the IEEE Symposium on Computers and Communications (ISCC), Athens, Greece, 5–8 September 2021. [Google Scholar] [CrossRef]
- Abraham, N.; Khan, N.M. A Novel Focal Tversky Loss Function with Improved Attention U-Net for Lesion Segmentation. In Proceedings of the IEEE 16th International Symposium on Biomedical Imaging (ISBI), Venice, Italy, 8–11 April 2019. [Google Scholar]
- Qi, W.; Su, H. A Cybertwin Based Multimodal Network for ECG Patterns Monitoring Using Deep Learning. IEEE Trans. Ind. Inform. 2022, 3203, 1–9. [Google Scholar] [CrossRef]
- Su, H.; Hu, Y.; Karimi, H.R.; Knoll, A.; Ferrigno, G.; De Momi, E. Improved Recurrent Neural Network-Based Manipulator Control with Remote Center of Motion Constraints: Experimental Results. Neural Netw. 2020, 131, 291–299. [Google Scholar] [CrossRef]
- Qi, W.; Aliverti, A. A Multimodal Wearable System for Continuous and Real-Time Breathing Pattern Monitoring during Daily Activity. IEEE J. Biomed. Health Inform. 2020, 24, 2199–2207. [Google Scholar] [CrossRef]
- Su, H.; Mariani, A.; Ovur, S.E.; Menciassi, A.; Ferrigno, G.; De Momi, E. Toward Teaching by Demonstration for Robot-Assisted Minimally Invasive Surgery. IEEE Trans. Autom. Sci. Eng. 2021, 18, 484–494. [Google Scholar] [CrossRef]
Dataset | Train | Validation | Test | |||
---|---|---|---|---|---|---|
Abnormal | Normal | Abnormal | Normal | Abnormal | Normal | |
CXR Modality-specific retraining | ||||||
CheXpert | 13,600 | 13,600 | 1700 | 1700 | 1700 | 1700 |
TBX11k | 3040 | 3040 | 380 | 380 | 380 | 380 |
RetinaNet-based object detection | ||||||
Dataset | Train | Validation | Test | |||
RSNA | 4212 | 600 | 1200 |
ResNet-50 Backbone and Classification Loss Functions | VGG-16 Backbone and Classification Loss Functions |
---|---|
ResNet-50 with random weights + focal loss | VGG-16 with random weights + focal loss |
ResNet-50 with random weights + focal Tversky loss | VGG-16 with random weights + focal Tversky loss |
ResNet-50 with ImageNet pretrained weights + focal loss | VGG-16 with ImageNet pretrained weights + focal loss |
ResNet-50 with ImageNet pretrained weights + focal Tversky loss | VGG-16 with ImageNet pretrained weights + focal Tversky loss |
ResNet-50 with CXR image modality-specific weights + focal loss | VGG-16 with CXR image modality-specific weights + focal loss |
ResNet-50 with CXR image modality-specific weights + focal Tversky loss | VGG-16 with CXR image modality-specific weights + focal Tversky loss |
Models | Accuracy | AUROC | AUPRC | Sensitivity | Precision | F-Score | MCC | Kappa |
---|---|---|---|---|---|---|---|---|
VGG-16 | 0.7834 | 0.8701 | 0.8777 | 0.8303 | 0.7591 | 0.7931 | 0.5693 (0.5542, 0.5844) | 0.5668 |
VGG-19 | 0.7743 | 0.8660 | 0.8727 | 0.8389 | 0.7429 | 0.7880 | 0.5532 (0.5380, 0.5684) | 0.5486 |
DenseNet-121 | 0.7738 | 0.8582 | 0.8618 | 0.8264 | 0.7477 | 0.7851 | 0.5507 (0.5355, 0.5659) | 0.5476 |
ResNet-50 | 0.7685 | 0.8586 | 0.8646 | 0.8207 | 0.7431 | 0.7800 | 0.5400 (0.5248, 0.5552) | 0.5370 |
EfficientNet-B0 | 0.7553 | 0.8568 | 0.8612 | 0.8678 | 0.7084 | 0.7800 | 0.5240 (0.5088, 0.5392) | 0.5106 |
MobileNet | 0.7584 | 0.8609 | 0.8655 | 0.8726 | 0.7104 | 0.7832 | 0.5309 (0.5157, 0.5461) | 0.5168 |
Models | AUPRC (mAP) |
---|---|
ResNet-50 with random weights + focal loss | 0.2763 (0.2509, 0.3017) |
ResNet-50 with random weights + focal Tversky loss | 0.2627 (0.2377, 0.2877) |
ResNet-50 with ImageNet pretrained weights + focal loss | 0.2719 (0.2467, 0.2971) |
ResNet-50 with ImageNet pretrained weights + focal Tversky loss | 0.2737 (0.2484, 0.2990) |
ResNet-50 with CXR image modality-specific weights + focal loss | 0.2865 (0.2609, 0.3121) |
ResNet-50 with CXR image modality-specific weights + focal Tversky loss | 0.2859 (0.2603, 0.3115) |
VGG-16 with random weights + focal loss | 0.2549 (0.2302, 0.2796) |
VGG-16 with random weights + focal Tversky loss | 0.2496 (0.2251, 0.2741) |
VGG-16 with ImageNet pretrained weights + focal loss | 0.2734 (0.2481, 0.2987) |
VGG-16 with ImageNet pretrained weights + focal Tversky loss | 0.2666 (0.2415, 0.2917) |
VGG-16 with CXR image modality-specific weights + focal loss | 0.2686 (0.2435, 0.2937) |
VGG-16 with CXR image modality-specific weights + focal Tversky loss | 0.2648 (0.2398, 0.2898) |
Ensemble Method | mAP |
---|---|
Top-3 model ensemble (ResNet-50 with CXR image modality-specific weights + focal loss, ResNet-50 with CXR image modality-specific weights + focal Tversky loss, and ResNet-50 with random weights + focal loss | 0.3272 (0.3006, 0.3538) |
Ensemble of the top-3 snapshots for each model | |
ResNet-50 with random weights + focal loss | 0.2777 (0.2523, 0.3031) |
ResNet-50 with random weights + focal Tversky loss | 0.2630 (0.2380, 0.2880) |
ResNet-50 with ImageNet pretrained weights + focal loss | 0.2788 (0.2534, 0.3042) |
ResNet-50 with ImageNet pretrained weights + focal Tversky loss | 0.2812 (0.2557, 0.3067) |
ResNet-50 with CXR image modality-specific weights + focal loss | 0.2973 (0.2714, 0.3232) |
ResNet-50 with CXR image modality-specific weights + focal Tversky loss | 0.2901 (0.2644, 0.3158) |
VGG-16 with random weights + focal loss | 0.2633 (0.2383, 0.2883) |
VGG-16 with random weights + focal Tversky loss | 0.2556 (0.2309, 0.2803) |
VGG-16 with ImageNet pretrained weights + focal loss | 0.2823 (0.2568, 0.3078) |
VGG-16 with ImageNet pretrained weights + focal Tversky loss | 0.2715 (0.2463, 0.2967) |
VGG-16 with CXR image modality-specific weights + focal loss | 0.2813 (0.2558, 0.3068) |
VGG-16 with CXR image modality-specific weights + focal Tversky loss | 0.2698 (0.2446, 0.2950) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Rajaraman, S.; Guo, P.; Xue, Z.; Antani, S.K. A Deep Modality-Specific Ensemble for Improving Pneumonia Detection in Chest X-rays. Diagnostics 2022, 12, 1442. https://doi.org/10.3390/diagnostics12061442
Rajaraman S, Guo P, Xue Z, Antani SK. A Deep Modality-Specific Ensemble for Improving Pneumonia Detection in Chest X-rays. Diagnostics. 2022; 12(6):1442. https://doi.org/10.3390/diagnostics12061442
Chicago/Turabian StyleRajaraman, Sivaramakrishnan, Peng Guo, Zhiyun Xue, and Sameer K. Antani. 2022. "A Deep Modality-Specific Ensemble for Improving Pneumonia Detection in Chest X-rays" Diagnostics 12, no. 6: 1442. https://doi.org/10.3390/diagnostics12061442
APA StyleRajaraman, S., Guo, P., Xue, Z., & Antani, S. K. (2022). A Deep Modality-Specific Ensemble for Improving Pneumonia Detection in Chest X-rays. Diagnostics, 12(6), 1442. https://doi.org/10.3390/diagnostics12061442