Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets
Abstract
:1. Introduction
- We developed and evaluated a three-stage training framework specifically for the pediatric chest X-ray dataset.
- Our study compares this three-stage training framework to other approaches, demonstrating that the ViT-Base/16 model, pre-trained on CXR, fine-tuned on CheXpert, and further fine-tuned on PediCXR, outperforms the otheres.
- We examined the top-performing model’s ability to detect common diseases accurately.
2. Materials and Methods
2.1. Dataset
2.1.1. Adult CXR Dataset
2.1.2. Pedatric CXR Dataset
2.2. Vision Transformer, Masked Autoencoder and Transfer Learning
2.3. Training Strategy and Details
2.3.1. Pre-Training Stage and Fine-Tuning Stage on Adult CXR
2.3.2. Knowledge Transfer from Adult to Pediatric CXR
2.3.3. Model Evaluation
3. Results
3.1. Model Performance with Supervised Learning
3.2. Transfer Learning via Linear Evaluation
3.3. Transfer Learning via Fine-Tuning
3.4. Model Interpretation
3.5. Embedding Visualization
3.6. Error Analysis with the Best-Performance Model
4. Discussion
- Introducing prior knowledge through masked autoencoder (MAE) pre-training on adult CXRs significantly boosts model performance on pediatric CXRs.
- Fine-tuning on adult CXRs further enhances the model’s ability to learn the intricate characteristics of thoracic disease distributions.
- Larger vision transformer models fine-tuned on the CheXpert dataset exhibit the best performance, both quantitatively and qualitatively.
- We developed a three-stage training system and assessed its effectiveness in classifying thoracic diseases using the most recent and extensive publicly accessible pediatric CXR dataset, PediCXR.
- Our study involved extensive quantitative experiments including (a) comparisons with direct training with a supervised learning strategy; (b) the use of CXR or non-medical images for pre-training; (c) the utilization of various adult CXR datasets for fine-tuning; and (d) an evaluation with linear and fine-tuning settings. Our findings demonstrate quantitatively and qualitatively that the top-performing model is the ViT-Base/16. This model was pre-trained on CXR, fine-tuned on CheXpert, and then further fully fine-tuned on PediCXR.
- We performed a detailed error analysis on PediCXR using the best-performing model, thoroughly examining its performance in identifying common diseases.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Reyes, M.A.; Etinger, V.; Hronek, C.; Hall, M.; Davidson, A.; Mangione-Smith, R.; Kaiser, S.V.; Parikh, K. Pediatric respiratory illnesses: An update on achievable benchmarks of care. Pediatrics 2023, 152, e2022058389. [Google Scholar] [CrossRef] [PubMed]
- World Health Organization. Stakeholder Consultative Meeting on Prevention and Management of Childhood Pneumonia and Diarrhoea: Report, 12–14 October 2021; World Health Organization: Geneva, Switzerland, 2022. [Google Scholar]
- Rahman, T.; Chowdhury, M.E.; Khandakar, A.; Islam, K.R.; Islam, K.F.; Mahbub, Z.B.; Kadir, M.A.; Kashem, S. Transfer learning with deep convolutional neural network (CNN) for pneumonia detection using chest X-ray. Appl. Sci. 2020, 10, 3233. [Google Scholar] [CrossRef]
- Banerjee, A.; Sarkar, A.; Roy, S.; Singh, P.K.; Sarkar, R. COVID-19 chest X-ray detection through blending ensemble of CNN snapshots. Biomed. Signal Process. Control 2022, 78, 104000. [Google Scholar] [CrossRef]
- Chen, S.; Ren, S.; Wang, G.; Huang, M.; Xue, C. Interpretable cnn-multilevel attention transformer for rapid recognition of pneumonia from chest X-ray images. IEEE J. Biomed. Health Inform. 2023, 28, 753–764. [Google Scholar] [CrossRef]
- Wollek, A.; Graf, R.; Čečatka, S.; Fink, N.; Willem, T.; Sabel, B.O.; Lasser, T. Attention-based saliency maps improve interpretability of pneumothorax classification. Radiol. Artif. Intell. 2022, 5, e220187. [Google Scholar] [CrossRef] [PubMed]
- Chetoui, M.; Akhloufi, M.A. Explainable vision transformers and radiomics for covid-19 detection in chest X-rays. J. Clin. Med. 2022, 11, 3013. [Google Scholar] [CrossRef] [PubMed]
- Wang, B.; Pan, H.; Aboah, A.; Zhang, Z.; Keles, E.; Torigian, D.; Turkbey, B.; Krupinski, E.; Udupa, J.; Bagci, U. GazeGNN: A Gaze-Guided Graph Neural Network for Chest X-Ray Classification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 1–6 January 2024; pp. 2194–2203. [Google Scholar]
- Mahapatra, D.; Bozorgtabar, B.; Ge, Z.; Reyes, M. GANDALF: Graph-based transformer and Data Augmentation Active Learning Framework with interpretable features for multi-label chest Xray classification. Med. Image Anal. 2024, 93, 103075. [Google Scholar] [CrossRef]
- Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
- Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9650–9660. [Google Scholar]
- He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
- Gazda, M.; Plavka, J.; Gazda, J.; Drotar, P. Self-supervised deep convolutional neural network for chest X-ray classification. IEEE Access 2021, 9, 151972–151982. [Google Scholar] [CrossRef]
- Tiu, E.; Talius, E.; Patel, P.; Langlotz, C.P.; Ng, A.Y.; Rajpurkar, P. Expert-level detection of pathologies from unannotated chest X-ray images via self-supervised learning. Nat. Biomed. Eng. 2022, 6, 1399–1406. [Google Scholar] [CrossRef]
- VanBerlo, B.; Hoey, J.; Wong, A. A survey of the impact of self-supervised pretraining for diagnostic tasks in medical X-ray, CT, MRI, and ultrasound. BMC Med. Imaging 2024, 24, 79. [Google Scholar] [CrossRef]
- Gidaris, S.; Singh, P.; Komodakis, N. Unsupervised representation learning by predicting image rotations. arXiv 2018, arXiv:1803.07728. [Google Scholar]
- Doersch, C.; Gupta, A.; Efros, A.A. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1422–1430. [Google Scholar]
- Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar] [CrossRef]
- Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning. PMLR, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
- Johnson, A.E.; Pollard, T.J.; Berkowitz, S.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.y.; Mark, R.G.; Horng, S. MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 2019, 6, 317. [Google Scholar] [CrossRef] [PubMed]
- Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilcus, S.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.; Shpanskaya, K.; et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 590–597. [Google Scholar]
- Padash, S.; Mohebbian, M.R.; Adams, S.J.; Henderson, R.D.; Babyn, P. Pediatric chest radiograph interpretation: How far has artificial intelligence come? A systematic literature review. Pediatr. Radiol. 2022, 52, 1568–1580. [Google Scholar] [CrossRef] [PubMed]
- Ravi, V.; Narasimhan, H.; Pham, T.D. A cost-sensitive deep learning-based meta-classifier for pediatric pneumonia classification using chest X-rays. Expert Syst. 2022, 39, e12966. [Google Scholar] [CrossRef]
- Pham, H.H.; Tran, T.T.; Nguyen, H.Q. VinDr-PCXR: An open, large-scale pediatric chest X-ray dataset for interpretation of common thoracic diseases. PhysioNet 2022. [Google Scholar] [CrossRef]
- Wang, L.; Lin, Z.Q.; Wong, A. Covid-net: A tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
- Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
- Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef]
- Wu, C.; Zhang, X.; Wang, Y.; Zhang, Y.; Xie, W. K-Diag: Knowledge-enhanced Disease Diagnosis in Radiographic Imaging. arXiv 2023, arXiv:2302.11557. [Google Scholar]
- Xiao, J.; Bai, Y.; Yuille, A.; Zhou, Z. Delving into masked autoencoders for multi-label thorax disease classification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 3588–3600. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do vision transformers see like convolutional neural networks? Adv. Neural Inf. Process. Syst. 2021, 34, 12116–12128. [Google Scholar]
- Reed, C.J.; Gupta, R.; Li, S.; Brockman, S.; Funk, C.; Clipp, B.; Keutzer, K.; Candido, S.; Uyttendaele, M.; Darrell, T. Scale-mae: A scale-aware masked autoencoder for multiscale geospatial representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 4088–4099. [Google Scholar]
- Zhou, L.; Liu, H.; Bae, J.; He, J.; Samaras, D.; Prasanna, P. Self Pre-training with Masked Autoencoders for Medical Image Classification and Segmentation. arXiv 2022, arXiv:2203.05573. [Google Scholar]
- Almalki, A.; Latecki, L.J. Self-Supervised Learning With Masked Autoencoders for Teeth Segmentation From Intra-Oral 3D Scans. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 7820–7830. [Google Scholar]
- Ericsson, L.; Gouk, H.; Hospedales, T.M. How well do self-supervised models transfer? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5414–5423. [Google Scholar]
- Chen, H.; Lundberg, S.M.; Erion, G.; Kim, J.H.; Lee, S.I. Forecasting adverse surgical events using self-supervised transfer learning for physiological signals. NPJ Digit. Med. 2021, 4, 167. [Google Scholar] [CrossRef] [PubMed]
- Truong, T.; Mohammadi, S.; Lenga, M. How transferable are self-supervised features in medical image classification tasks? Proc. Mach. Learn. Health PMLR 2021, 158, 54–74. [Google Scholar]
- Tran, T.T.; Pham, H.H.; Nguyen, T.V.; Le, T.T.; Nguyen, H.T.; Nguyen, H.Q. Learning to automatically diagnose multiple diseases in pediatric chest radiographs using deep convolutional neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3314–3323. [Google Scholar]
- Kirolos, A.; Manti, S.; Blacow, R.; Tse, G.; Wilson, T.; Lister, M.; Cunningham, S.; Campbell, A.; Nair, H.; Reeves, R.M.; et al. A systematic review of clinical practice guidelines for the diagnosis and management of bronchiolitis. J. Infect. Dis. 2020, 222, S672–S679. [Google Scholar] [CrossRef] [PubMed]
- Shin, H.J.; Son, N.H.; Kim, M.J.; Kim, E.K. Diagnostic performance of artificial intelligence approved for adults for the interpretation of pediatric chest radiographs. Sci. Rep. 2022, 12, 10215. [Google Scholar] [CrossRef]
- Kohne, J.G.; Farzaneh, N.; Barbaro, R.P.; Mahani, M.G.; Ansari, S.; Sjoding, M.W. Deep learning model performance for identifying pediatric acute respiratory distress syndrome on chest radiographs. Intensive Care-Med. Paediatr. Neonatal 2024, 2, 5. [Google Scholar] [CrossRef]
- Kjærgaard, J.; Anastasaki, M.; Stubbe Østergaard, M.; Isaeva, E.; Akylbekov, A.; Nguyen, N.Q.; Reventlow, S.; Lionis, C.; Sooronbaev, T.; Pham, L.A.; et al. Diagnosis and treatment of acute respiratory illness in children under five in primary care in low-, middle-, and high-income countries: A descriptive FRESH AIR study. PLoS ONE 2019, 14, e0221389. [Google Scholar] [CrossRef] [PubMed]
- Friedman, J.N.; Rieder, M.J.; Walton, J.M.; Society, C.P.; Committee, A.C.; Therapy, D.; Committee, H.S. Bronchiolitis: Recommendations for diagnosis, monitoring and management of children one to 24 months of age. Paediatr. Child Health 2014, 19, 485–491. [Google Scholar] [CrossRef] [PubMed]
- Taussig, L.M.; Smith, S.M.; Blumenfeld, R. Chronic bronchitis in childhood: What is it? Pediatrics 1981, 67, 1–5. [Google Scholar] [CrossRef] [PubMed]
- Cui, C.; Yang, H.; Wang, Y.; Zhao, S.; Asad, Z.; Coburn, L.A.; Wilson, K.T.; Landman, B.A.; Huo, Y. Deep multimodal fusion of image and non-image data in disease diagnosis and prognosis: A review. Prog. Biomed. Eng. 2023, 5, 022001. [Google Scholar] [CrossRef] [PubMed]
- Jabbour, S.; Fouhey, D.; Kazerooni, E.; Wiens, J.; Sjoding, M.W. Combining chest X-rays and electronic health record (EHR) data using machine learning to diagnose acute respiratory failure. J. Am. Med. Inform. Assoc. 2022, 29, 1060–1068. [Google Scholar] [CrossRef]
- Wang, Y.; Yin, C.; Zhang, P. Multimodal risk prediction with physiological signals, medical images and clinical notes. Heliyon 2024, 10, e26772. [Google Scholar] [CrossRef]
Strategies | Methods (Ref.) |
---|---|
Innate relationship | Image rotation prediction [16]; Image context prediction [17] |
Generative | Autoencoder [18]; GAN [19] |
Contrastive | SimCLR [20]; BYOL [10]; DINO [11] |
Self-prediction | MAE [12] |
Dataset | Samples (n) | Brocho-Pneumonia (n) | Bronchiolitis (n) | Bronchitis (n) | Pneumonia (n) | No Finding (n) | Other Diseases (n) |
---|---|---|---|---|---|---|---|
Training | 7728 | 545 | 497 | 842 | 392 | 5143 | 463 |
Test | 1387 | 84 | 90 | 174 | 89 | 907 | 81 |
Encoder | mAUC | AUC for Every Class | ||||||
---|---|---|---|---|---|---|---|---|
Broncho-Pneumonia | Bronchiolitis | Bronchitis | No Finding | Pneumonia | Others | p-Value | ||
DenseNet121 † | 0.709 | 0.696 | 0.638 | 0.691 | 0.776 | 0.802 | 0.703 | <0.0001 |
Densenet121 | 0.714 | 0.781 | 0.710 | 0.698 | 0.726 | 0.744 | 0.625 | <0.0001 |
[0.709–0.719] | [0.780–0.782] | [0.703–0.717] | [0.696–0.700] | [0.721–0.731] | [0.732–0.756] | [0.613–0.637] | ||
ResNet50 | 0.685 | 0.736 | 0.690 | 0.686 | 0.696 | 0.709 | 0.592 | <0.0001 |
[0.679–0.691] | [0.727–0.745] | [0.676–0.704] | [0.685–0.687] | [0.691–0.701] | [0.693–0.725] | [0.577–0.607] | ||
ViT–Small/16 | 0.646 | 0.698 | 0.627 | 0.646 | 0.655 | 0.652 | 0.600 | <0.0001 |
[0.641–0.652] | [0.696–0.699] | [0.621–0.632] | [0.644–0.649] | [0.648–0.662] | [0.643–0.661] | [0.580–0.621] | ||
ViT–Base/16 | 0.654 | 0.704 | 0.649 | 0.648 | 0.657 | 0.655 | 0.609 | <0.0001 |
[0.648–0.659] | [0.694–0.715] | [0.632–0.665] | [0.643–0.654] | [0.655–0.658] | [0.636–0.674] | [0.607–0.612] |
Encoder | Pretrained | Finetuned | mAUC | AUC for Every Class | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Broncho-Pneumonia | Bronchiolitis | Bronchitis | No Finding | Pneumonia | Others | p-Value | ||||
ViT-Small/16 | X-ray | COVIDx | 0.643 | 0.670 | 0.611 | 0.679 | 0.695 | 0.667 | 0.538 | <0.0001 |
[0.631–0.656] | [0.645–0.695] | [0.557–0.665] | [0.676–0.682] | [0.690–0.701] | [0.605–0.729] | [0.512–0.565] | ||||
X-ray | 0.653 | 0.700 | 0.637 | 0.647 | 0.696 | 0.695 | 0.545 | <0.0001 | ||
([0.648–0.658] | [0.689–0.710] | [0.615–0.659] | [0.641–0.653] | [0.689–0.703] | [0.664–0.727] | [0.518–0.572] | ||||
CheXpert | 0.662 | 0.707 | 0.632 | 0.655 | 0.690 | 0.734 | 0.554 | <0.0001 | ||
[0.662–0.662] | [0.675–0.740] | [0.625–0.640] | [0.647–0.663] | [0.683–0.697] | [0.723–0.744] | [0.518–0.590] | ||||
ViT-Base/16 | X-ray | COVIDx | 0.675 | 0.710 | 0.657 | 0.674 | 0.704 | 0.711 | 0.597 | <0.0001 |
[0.662–0.689] | [0.682–0.739] | [0.616–0.698] | [0.649–0.698] | [0.699–0.708] | [0.674–0.747] | [0.551–0.643] | ||||
X-ray | 0.678 | 0.739 | 0.658 | 0.646 | 0.702 | 0.735 | 0.586 | <0.0001 | ||
[0.673–0.682] | [0.723–0.755] | [0.648–0.669] | [0.628–0.664] | [0.698–0.705] | [0.727–0.744] | [0.553–0.620] | ||||
CheXpert | 0.690 | 0.745 | 0.677 | 0.670 | 0.708 | 0.737 | 0.604 | <0.0001 | ||
[0.682–0.698] | [0.728–0.762] | [0.642–0.712] | [0.654–0.686] | [0.703–0.712] | [0.722–0.752] | [0.578–0.631] |
Encoder | Pretrained | Finetuned | mAUC | AUC for Every Class | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Broncho-Pneumonia | Bronchiolitis | Bronchitis | No Finding | Pneumonia | Others | p-Value | ||||
DenseNet121 | X–ray | – | 0.749 | 0.823 | 0.732 | 0.722 | 0.776 | 0.827 | 0.615 | <0.0001 |
[0.747–0.752] | [0.817–0.830] | [0.729–0.734] | [0.720–0.724] | [0.770–0.783] | [0.818–0.836] | [0.610–0.620] | ||||
X–ray | 0.713 | 0.779 | 0.709 | 0.700 | 0.725 | 0.744 | 0.621 | <0.0001 | ||
[0.708–0.717] | [0.772–0.786] | [0.701–0.717] | [0.695–0.704] | [0.724–0.725] | [0.735–0.754] | [0.602–0.639] | ||||
ViT–Small/16 | Imagenet | – | 0.719 | 0.787 | 0.709 | 0.711 | 0.729 | 0.761 | 0.618 | <0.0001 |
[0.716–0.723] | [0.780–0.794] | [0.709–0.710] | [0.707–0.715] | [0.725–0.732] | [0.758–0.765] | [0.613–0.624] | ||||
– | 0.729 | 0.808 | 0.721 | 0.708 | 0.744 | 0.770 | 0.626 | <0.0001 | ||
[0.727–0.732] | [0.806–0.810] | [0.717–0.725] | [0.707–0.709] | [0.740–0.748] | [0.764–0.776] | [0.617–0.636] | ||||
X–ray | COVIDx | 0.746 | 0.825 | 0.725 | 0.729 | 0.760 | 0.797 | 0.639 | <0.005 | |
[0.741–0.751] | [0.818–0.833] | [0.716–0.733] | [0.726–0.732] | [0.756–0.764] | [0.779–0.815] | [0.632–0.645] | ||||
X–ray | 0.748 | 0.824 | 0.725 | 0.737 | 0.761 | 0.798 | 0.642 | <0.0001 | ||
[0.747–0.748] | [0.820–0.827] | [0.719–0.731] | [0.732–0.741] | [0.760–0.762] | [0.797–0.800] | [0.640–0.643] | ||||
CheXpert | 0.748 | 0.818 | 0.719 | 0.740 | 0.758 | 0.805 | 0.650 | <0.0001 | ||
[0.748–0.749] | [0.817–0.819] | [0.718–0.720] | [0.737–0.742] | [0.757–0.759] | [0.800–0.810] | [0.645–0.654] | ||||
ViT–Base/16 | Imagenet | – | 0.746 | 0.824 | 0.729 | 0.722 | 0.759 | 0.809 | 0.633 | <0.005 |
[0.745–0.747] | [0.823–0.826] | [0.727–0.732] | [0.720–0.725] | [0.756–0.761] | [0.807–0.810] | [0.631–0.634] | ||||
X–ray | – | 0.743 | 0.818 | 0.728 | 0.722 | 0.757 | 0.800 | 0.633 | 0.2 | |
[0.740–0.746] | [0.812–0.824] | [0.723–0.733] | [0.721–0.724] | [0.754–0.759] | [0.792–0.808] | [0.628–0.639] | ||||
COVIDx | 0.750 | 0.833 | 0.732 | 0.725 | 0.762 | 0.814 | 0.634 | <0.005 | ||
[0.749–0.751] | [0.830–0.835] | [0.730–0.734] | [0.720–0.730] | [0.760–0.765] | [0.812–0.816] | [0.628–0.641] | ||||
X–ray | 0.760 | 0.825 | 0.733 | 0.726 | 0.767 | 0.831 | 0.678 | <0.0001 | ||
[0.758–0.761] | [0.824–0.827] | [0.729–0.736] | [0.721–0.730] | [0.765–0.769] | [0.827–0.835] | [0.673–0.683] | ||||
CheXpert | 0.761 | 0.831 | 0.711 | 0.741 | 0.766 | 0.835 | 0.683 | |||
[0.760–0.763] | [0.829–0.833] | [0.709–0.714] | [0.737–0.745] | [0.764–0.767] | [0.833–0.837] | [0.672–0.695] |
Label | Accuracy | Sensitivity | Precision | Specificity | F1 Score |
---|---|---|---|---|---|
Brocho-pneumonia | 0.801 | 0.691 | 0.187 | 0.801 | 0.294 |
Bronchiolitis | 0.765 | 0.500 | 0.137 | 0.784 | 0.215 |
Bronchitis | 0.632 | 0.724 | 0.213 | 0.619 | 0.329 |
Pneumonia | 0.894 | 0.618 | 0.325 | 0.913 | 0.426 |
Other diseases | 0.852 | 0.309 | 0.142 | 0.885 | 0.195 |
mean | 0.770 | 0.639 | 0.279 | 0.683 | 0.376 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, Y.; Kohne, J.; Wittrup, E.; Najarian, K. Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets. Diagnostics 2024, 14, 1634. https://doi.org/10.3390/diagnostics14151634
Zhang Y, Kohne J, Wittrup E, Najarian K. Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets. Diagnostics. 2024; 14(15):1634. https://doi.org/10.3390/diagnostics14151634
Chicago/Turabian StyleZhang, Yufeng, Joseph Kohne, Emily Wittrup, and Kayvan Najarian. 2024. "Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets" Diagnostics 14, no. 15: 1634. https://doi.org/10.3390/diagnostics14151634
APA StyleZhang, Y., Kohne, J., Wittrup, E., & Najarian, K. (2024). Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets. Diagnostics, 14(15), 1634. https://doi.org/10.3390/diagnostics14151634