Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies
Abstract
:1. Introduction
2. Results
2.1. Combination of Biologically Inspired Indices and Machine Learning
2.1.1. Discrimination between Positive and Negative Samples with Full Image-Related Indices
- We tried a simple neural network (multilayer perceptron) as a more elaborate model: classification efficiency was not improved (testing cpa = 0.59, test auc = 0.83), in accordance with our earlier conclusion that simpler ML models were better suited to processing limited datasets [34].
- Since ML is considered fairly “data hungry” [39], it was of interest to ask whether an insufficient dataset size (137 samples) might be an important cause of prediction errors. This question was addressed by measuring the dependence of LR classification efficiency on sample number. As shown in Figure 3, index-based classification efficiency was only weakly dependent on the dataset size.
- The behavior of ML algorithms is dependent on so-called hyperparameters that are often ignored, since default values are usually satisfactory. It was checked that the classification efficiency of LR could not be improved by changing LR regularization parameter C. As expected, the default value (C = 1) was found satisfactory. Reducing regularization resulted in significant increase of training cpa, with a decrease of testing cpa, which was indicative of overfitting. Increasing regularization resulted in concomitant decrease of cpa on training and testing datasets.
- Aforementioned results strongly suggested that classification efficiency was limited by the intrinsic capacity of indices used to quantify images, in line with conventional wisdom [19]. Since the first index was derived from our experience of automatic detection of anti-nuclear antibodies [15,31], we tested the discrimination provided by this sole index, based on empirical determination of a threshold value separating positive from negative samples. Our dataset was randomly split 100 times between a training set (102 samples) and a testing set (35 samples). The average cpa parameters obtained on the training and testing sets were, respectively, 0.705 +/− 0.004 SE and 0.701 +/− 0.013 SE, which were slightly but significantly (p = 0.0016) higher than efficiency parameters shown in Table 1. This supports the well-known fact that addition of improper features may hamper LR efficiency.
2.1.2. Automatic Discrimination between Several Fluorescence Patterns
2.2. Use of AI for Autonomous Analysis of Fluorescence Images
2.2.1. Use of Data Reduction to Process Individual Cell Images
2.2.2. Analysis of Full Images with Neural Networks
- Efficiency parameters displayed limited change in response to fairly extensive variation of hyperparameters, suggesting a moderate dependence of classification efficiency on the model settings.
- Parameter cpa calculated on testing sets varied between a minimum value of 0.38 and a maximum of 0.51 (with kappa score and auc, respectively, equal to 0.67 and 0.85). Neural network performance was thus better than that achieved with standard ML models (shown in Table 4).
- Plots displayed in Figure 6C,D clearly confirmed the risk of overfitting as a consequence of insufficient regularization (C) or excessive number of features (D) as compared to the number of samples, leading to a high cpa training/cpa testing ratio.
2.2.3. Combination of Controlled Splitting of Training and Testing Datasets and Serum Rather than Image Classification
3. Discussion
- The information provided by individual serum samples might be enhanced by performing additional fixation or staining procedures. Indeed, it has long been reported that the localization of ANCA-related fluorescence is not the same on ethanol- and paraformaldehyde-fixed cells [9], and it might be more informative to use datasets including two fluorescence images. Also, nuclear localization provided by DAPI labelling could also be inserted in an additional channel. CNNs would be well suited to analyzing image stacks associated to individual cells, and DAPI staining was used in recent attempts at ANA classification with ML [17]. Also, it might not be warranted to increase the complexity and cost of immunofluorescence testing if this did not result in a very substantial increase of information content.
- Different image preprocessing procedures might be considered, such as filtering to remove noise or replacing image resizing by embedding into larger areas to retain information relevant to absolute distances.
- A common way of increasing ML power consists of increasing feature diversity [45,46]. Thus, it might be rewarding to combine images with other patients’ features. However, the need for additional parameters that might not be immediately available in hospital laboratories would delay computer-assisted analysis, thus hampering ML-generated rapidity gain. Therefore, specific clinical trials would be needed to validate feature extension.
- Training a model with a restricted dataset might be improved with data augmentation, which consists of creating “realistic” data with suitable algorithms. As an example, the classification of macrophages from microscopic images with simple geometrical features such as area or circularity was reported to display an accuracy increase from 0.3 to 0.93 when the dataset size was increased one hundred-fold with a custom-made image generator [47].
3.1. Model Choice
- As shown in the first part of this report (Table 1), the use of biologically inspired indices is an attractive way of combining biological expertise and AI. Indeed, many commercially available systems successfully use ML algorithms to process extensive sets of texture parameters. However, the development and continuous improvement of an algorithm involving more than 1000 parameters [40] may be more difficult to perform than the autonomous building of ML models. Accordingly, recent comparisons between deep learning and a combination of hand-crafted features and simple ML models such as SVM or random forests supported the superiority of neural networks [18,20]. However, it would be an attractive prospect to use ML to improve the power of selected parameters. While neural networks are often compared to “black boxes”, theoretical effort is currently underway regarding “interpreting” their behavior [48]. These endeavors might in the future help improve biological intuition and thereby allow for substantial improvement of so-called hand-crafted features.
- Results presented in this report revealed a significant but insufficient efficiency of a combination of data reduction with PCA and simple ML methods to classify 50 × 50 pixel images. Indeed, neural networks may now be considered the gold standard for image analysis [12], and they are currently the basis of many current reports on ML classification for medical purposes [49]. However, while more and more powerful network architectures are continually being reported and tested [17,18,50], model setting and training quality are essential determinants of final performance. Available strategies will be rapidly listed below.
3.2. Hyperparameter Setting
3.3. Improving the Training Process
- A common means of reducing overfitting consists of stopping the training phase as soon as the validation error reaches a minimum. While this early stopping procedure is widely used and intuitively considered reasonable, the identification of an optimal training duration may warrant further studies [52,53].
- Another procedure facilitating the training of very complex models, dubbed transfer learning, consists of using a pretrained model and training only the outer layers to fit a specific dataset. This method permits the use of highly successful models trained on public image datasets such as ImageNet [54] with a reasonable computing load for ANA classification with Hep-2 cells [50,55].
- An attractive prospect might consist of driving the development of a complex model through what might be dubbed smart training and yield unexpected performance. Thus, the development of a convolutional structure in a fully connected network was achieved by training this network with translation-invariant data [56]. Also, a neural network was claimed to acquire increased capacity through a special learning method dubbed meta-learning [57].
4. Materials and Methods
4.1. Patients
4.2. Immunofluorescence
4.3. ELISA Testing
4.4. Image Processing
4.4.1. Calculation of Overall Quantitative Indices
- -
- Index i1 is the ratio between the mean intensity on FITC images of pixels classified as “inside” and “outside”. This was expected to permit discrimination between positive and negative samples.
- -
- Index i2 is the ratio between the mean FITC intensity of pixels defined as “inside” and the first peak intensity of the histogram of FITC image.
- -
- Index i3 is similar to i2, but “inside” is defined on DAPI histograms as pixels with an intensity 16 times higher than that of the first background peak. It was expected that this region might be closer to actual nuclear regions.
- -
- Index i4 is the correlation between FITC and DAPI pixel intensities in regions defined as “inside” on DAPI images. It might be hoped that the correlation would be highest with ANA, lowest with C-ANCA and intermediate with P-ANCA.
4.4.2. Building Individual Cell Images
4.5. Machine Learning
4.5.1. Classification Based on “Hand-Crafted” Parameters
4.5.2. Analysis of Individual Cell Images
- Individual cell images (50 × 50 pixels) were first subjected to a scaling procedure (scikit-learn RobustScaler method) to ensure that all parameters displayed similar median and quartile distributions. In some cases, data reduction was performed with principal component analysis (PCA).
- Images were then analyzed with the aforementioned standard algorithms (logical regression, k nearest neighbors, decision tree and neural networks). In addition to aforementioned MLP, we used convolutive networks (CNNs), since they are thought to be well suited to image analysis [12,25] and are currently considered the gold standard [61]. In addition to conventional so-called dense layers, CNNs involve convolutional layers where each neuron is stimulated by a restricted set of neurons belonging to the underlying layer through a translation-independent set of weights (kernels). Also, a given layer may directly stimulate numerous upper layers (feature maps) with different sets of weights (filters). This architecture allows the model to identify motives of growing complexity in a fairly hierarchical way. Further, so-called dropout layers appeared to be a powerful means of reducing overfitting. The Tensorflow platform was used, taking advantage of the keras application programming interface. A number of architectures were tested by modification of a number of parameters (number of filters, kernel size, addition of an input channel for simultaneous processing of FITC and DAPI images, activation and loss parameter) starting from a suggested simple architecture ([25] p. 496). However, due to the high number of parameters and long training time, these attempts remained preliminary.
- Under all conditions, efficiency parameters were calculated by random splitting of datasets between 10 and 100 times into a training set (about 75% of samples) and a testing set (about 25% of samples). Classification efficiency was then calculated on the training and testing set after training models on training sets.
4.6. Statistics
5. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
AI | artificial intelligence |
ANA | anti-nuclear antibody |
ANCA | anti-neutrophil cytoplasmic antibody |
auc | area under ROC curve |
C-ANCA | cytoplasmic type ANCA |
CNN | Convolutional neural network |
cpa | corrected prediction accuracy |
DAPI | 4,6-diaminophenylindol (considered as a fluorescent nucleus marker) |
DT | decision tree classifier |
FITC | fluorescein isothiocyanate |
KNN | k nearest neighbors classifier |
LR | logistic regression classifier |
ML | machine learning |
MLP | multilayer perceptron |
P-ANCA | perinuclear-type ANCA |
pa | prediction accuracy |
ROC | receiver operator curve |
References
- Haug, C.J.; Drazen, J.M. Artificial Intelligence and Machine Learning in Clinical Medicine, 2023. N. Engl. J. Med. 2023, 388, 1201–1208. [Google Scholar] [CrossRef]
- Beam, A.L.; Drazen, J.M.; Kohane, I.S.; Leong, T.-Y.; Manrai, A.K.; Rubin, E.J. Artificial Intelligence in Medicine. N. Engl. J. Med. 2023, 388, 1220–1221. [Google Scholar] [CrossRef]
- Attia, Z.I.; Noseworthy, P.A.; Lopez-Jimenez, F.; Asirvatham, S.J.; Deshmukh, A.J.; Gersh, B.J.; Carter, R.E.; Yao, X.; Rabinstein, A.A.; Erickson, B.J.; et al. An Artificial Intelligence-Enabled ECG Algorithm for the Identification of Patients with Atrial Fibrillation during Sinus Rhythm: A Retrospective Analysis of Outcome Prediction. Lancet 2019, 394, 861–867. [Google Scholar] [CrossRef] [PubMed]
- Shen, B.; Yi, X.; Sun, Y.; Bi, X.; Du, J.; Zhang, C.; Quan, S.; Zhang, F.; Sun, R.; Qian, L.; et al. Proteomic and Metabolomic Characterization of COVID-19 Patient Sera. Cell 2020, 182, 59–72.e15. [Google Scholar] [CrossRef] [PubMed]
- Shrock, E.; Fujimura, E.; Kula, T.; Timms, R.T.; Lee, I.-H.; Leng, Y.; Robinson, M.L.; Sie, B.M.; Li, M.Z.; Chen, Y.; et al. Viral Epitope Profiling of COVID-19 Patients Reveals Cross-Reactivity and Correlates of Severity. Science 2020, 370, eabd4250. [Google Scholar] [CrossRef] [PubMed]
- Rigon, A.; Infantino, M.; Merone, M.; Iannello, G.; Tincani, A.; Cavazzana, I.; Carabellese, N.; Radice, A.; Manfredi, M.; Soda, P.; et al. The Inter-Observer Reading Variability in Anti-Nuclear Antibodies Indirect (ANA) Immunofluorescence Test: A Multicenter Evaluation and a Review of the Literature. Autoimmun. Rev. 2017, 16, 1224–1229. [Google Scholar] [CrossRef]
- Cinquanta, L.; Bizzaro, N.; Pesce, G. Standardization and Quality Assessment Under the Perspective of Automated Computer-Assisted HEp-2 Immunofluorescence Assay Systems. Front. Immunol. 2021, 12, 638863. [Google Scholar] [CrossRef]
- Kallenberg, C.G.M. Usefulness of Antineutrophil Cytoplasmic Autoantibodies in Diagnosing and Managing Systemic Vasculitis. Curr. Opin. Rheumatol. 2016, 28, 8–14. [Google Scholar] [CrossRef]
- Csernok, E.; Moosig, F. Current and Emerging Techniques for ANCA Detection in Vasculitis. Nat. Rev. Rheumatol. 2014, 10, 494–501. [Google Scholar] [CrossRef]
- Schmitt, W.H.; Csernok, E.; Gross, W.L. ANCA and Infection. Lancet 1991, 337, 1416–1417. [Google Scholar] [CrossRef]
- Durand, J.M.; Mege, J.L.; Velut, J.G.; Escallier, J.C.; Kaplanski, G.; Quiles, N.; Bongrand, P.; Soubeyrand, J. Antineutrophil Cytoplasmic Antibodies and Infection. Autoimmunity 1993, 15, 81–83. [Google Scholar] [CrossRef]
- Rahman, S.; Wang, L.; Sun, C.; Zhou, L. Deep Learning Based HEp-2 Image Classification: A Comprehensive Review. Med. Image Anal. 2020, 65, 101764. [Google Scholar] [CrossRef] [PubMed]
- Nagy, G.; Csípő, I.; Tarr, T.; Szűcs, G.; Szántó, A.; Bubán, T.; Sipeki, N.; Szekanecz, Z.; Papp, M.; Kappelmayer, J.; et al. Anti-Neutrophil Cytoplasmic Antibody Testing by Indirect Immunofluorescence: Computer-Aided versus Conventional Microscopic Evaluation of Routine Diagnostic Samples from Patients with Vasculitis or Other Inflammatory Diseases. Clin. Chim. Acta 2020, 511, 117–124. [Google Scholar] [CrossRef] [PubMed]
- Zarur, E.B.; Rodrigues, S.H.; Ambrogini, O., Jr.; Ferraz, M.L.G.; Mastroianni-Kirsztajn, G.; Andrade, L.E.C.; Souza, A.W.S. de Improvement of Indirect Immunofluorescence Technique to Detect Antineutrophil Cytoplasmic Antibodies and Its Impact on Test Positivity Rate. Braz. J. Med. Biol. Res. 2023, 56, e12636. [Google Scholar] [CrossRef] [PubMed]
- Bertin, D.; Mouhajir, Y.; Bongrand, P.; Bardin, N. ICARE Improves Antinuclear Antibody Detection by Overcoming the Barriers Preventing Accreditation. Clin. Chim. Acta 2016, 454, 57–61. [Google Scholar] [CrossRef]
- Hiemann, R.; Hilger, N.; Michel, J.; Nitschke, J.; Böhm, A.; Anderer, U.; Weigert, M.; Sack, U. Automatic Analysis of Immunofluorescence Patterns of HEp-2 Cells. Ann. N. Y. Acad. Sci. 2007, 1109, 358–371. [Google Scholar] [CrossRef]
- Wu, Y.-D.; Sheu, R.-K.; Chung, C.-W.; Wu, Y.-C.; Ou, C.-C.; Hsiao, C.-W.; Chang, H.-C.; Huang, Y.-C.; Chen, Y.-M.; Lo, W.-T.; et al. Application of Supervised Machine Learning to Recognize Competent Level and Mixed Antinuclear Antibody Patterns Based on ICAP International Consensus. Diagnostics 2021, 11, 642. [Google Scholar] [CrossRef]
- Xie, Q.; Chen, P.; Li, Z.; Xie, R. Automatic Segmentation and Classification for Antinuclear Antibody Images Based on Deep Learning. Comput. Intell. Neurosci. 2023, 2023, 1353965. [Google Scholar] [CrossRef]
- Vununu, C.; Lee, S.-H.; Kwon, K.-R. A Strictly Unsupervised Deep Learning Method for HEp-2 Cell Image Classification. Sensors 2020, 20, 2717. [Google Scholar] [CrossRef]
- Vohra, S.K.; Prodanov, D. The Active Segmentation Platform for Microscopic Image Classification and Segmentation. Brain Sci. 2021, 11, 1645. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Jiahui, G.; Feilong, M.; Visconti di Oleggio Castello, M.; Nastase, S.A.; Haxby, J.V.; Gobbini, M.I. Modeling Naturalistic Face Processing in Humans with Deep Convolutional Neural Networks. Proc. Natl. Acad. Sci. USA 2023, 120, e2304085120. [Google Scholar] [CrossRef]
- Müller, A.C.; Guido, S. Introduction to Machine Learning with Python: A Guide for Data Scientists, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2016; ISBN 978-1-4493-6941-5. [Google Scholar]
- Chollet, F. Deep Learning with Python, 2nd ed.; Manning: Shelter Island, NY, USA, 2021; ISBN 978-1-61729-686-4. [Google Scholar]
- Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, 3rd ed.; O’Reilly: Boston, MA, USA, 2023; ISBN 978-1-09-812597-4. [Google Scholar]
- Hu, Z.; Tang, A.; Singh, J.; Bhattacharya, S.; Butte, A.J. A Robust and Interpretable End-to-End Deep Learning Model for Cytometry Data. Proc. Natl. Acad. Sci. USA 2020, 117, 21373–21380. [Google Scholar] [CrossRef] [PubMed]
- Achar, S.R.; Bourassa, F.X.P.; Rademaker, T.J.; Lee, A.; Kondo, T.; Salazar-Cavazos, E.; Davies, J.S.; Taylor, N.; François, P.; Altan-Bonnet, G. Universal Antigen Encoding of T Cell Activation from High-Dimensional Cytokine Dynamics. Science 2022, 376, 880–884. [Google Scholar] [CrossRef] [PubMed]
- Hubert, L.; Arabie, P. Comparing Partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
- Landis, J.R.; Koch, G.G. The Measurement of Observer Agreement for Categorical Data. Biometrics 1977, 33, 159. [Google Scholar] [CrossRef]
- Knütter, I.; Hiemann, R.; Brumma, T.; Büttner, T.; Großmann, K.; Cusini, M.; Pregnolato, F.; Borghi, M.; Anderer, U.; Conrad, K.; et al. Automated Interpretation of ANCA Patterns—A New Approach in the Serology of ANCA-Associated Vasculitis. Arthritis Res. Ther. 2012, 14, R271. [Google Scholar] [CrossRef]
- Bertin, D.; Jourde-Chiche, N.; Bongrand, P.; Bardin, N. Original Approach for Automated Quantification of Antinuclear Autoantibodies by Indirect Immunofluorescence. Clin. Dev. Immunol. 2013, 2013, 182172. [Google Scholar] [CrossRef] [PubMed]
- Jameson, J.L. (Ed.) Harrison’s Principles of Internal Medicine, 20th ed.; McGraw-Hill Education: New York, NY, USA, 2018; ISBN 978-1-259-64404-7. [Google Scholar]
- de Hond, A.A.H.; Steyerberg, E.W.; van Calster, B. Interpreting Area under the Receiver Operating Characteristic Curve. Lancet Digit. Health 2022, 4, e853–e855. [Google Scholar] [CrossRef]
- Bertin, D.; Bongrand, P.; Bardin, N. Can Artificial Intelligence Help a Clinical Laboratory to Draw Useful Information from Limited Data Sets? Application to Mixed Connective Tissue Disease. J. Autonom. Intell. 2023, 6, 664. [Google Scholar] [CrossRef]
- Shovman, O.; Agmon-Levin, N.; Gilburd, B.; Martins, T.; Petzold, A.; Matthias, T.; Shoenfeld, Y. A Fully Automated IIF System for the Detection of Antinuclear Antibodies and Antineutrophil Cytoplasmic Antibodies. Immunol. Res. 2015, 61, 135–140. [Google Scholar] [CrossRef] [PubMed]
- Sack, U.; Knoechner, S.; Warschkau, H.; Pigla, U.; Emmrich, F.; Kamprad, M. Computer-Assisted Classification of HEp-2 Immunofluorescence Patterns in Autoimmune Diagnostics. Autoimmun. Rev. 2003, 2, 298–304. [Google Scholar] [CrossRef] [PubMed]
- Bizzaro, N.; Antico, A.; Platzgummer, S.; Tonutti, E.; Bassetti, D.; Pesente, F.; Tozzoli, R.; Tampoia, M.; Villalta, D. Automated Antinuclear Immunofluorescence Antibody Screening: A Comparative Study of Six Computer-Aided Diagnostic Systems. Autoimmun. Rev. 2014, 13, 292–298. [Google Scholar] [CrossRef] [PubMed]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2016. [Google Scholar] [CrossRef]
- van der Ploeg, T.; Austin, P.C.; Steyerberg, E.W. Modern Modelling Techniques Are Data Hungry: A Simulation Study for Predicting Dichotomous Endpoints. BMC Med. Res. Methodol. 2014, 14, 137. [Google Scholar] [CrossRef] [PubMed]
- Hiemann, R.; Büttner, T.; Krieger, T.; Roggenbuck, D.; Sack, U.; Conrad, K. Challenges of Automated Screening and Differentiation of Non-Organ Specific Autoantibodies on HEp-2 Cells. Autoimmun. Rev. 2009, 9, 17–22. [Google Scholar] [CrossRef] [PubMed]
- Xiao, H.; Rasul, K.; Vollgraf, R. Fashion-MNIST: A Novel Image Dataset for Benchmarking Machine Learning Algorithms. arXiv 2017, arXiv:1708.07747. [Google Scholar]
- Finlayson, S.G.; Subbaswamy, A.; Singh, K.; Bowers, J.; Kupke, A.; Zittrain, J.; Kohane, I.S.; Saria, S. The Clinician and Dataset Shift in Artificial Intelligence. N. Engl. J. Med. 2021, 385, 283–286. [Google Scholar] [CrossRef]
- Chekroud, A.M.; Hawrilenko, M.; Loho, H.; Bondar, J.; Gueorguieva, R.; Hasan, A.; Kambeitz, J.; Corlett, P.R.; Koutsouleris, N.; Krumholz, H.M.; et al. Illusory Generalizability of Clinical Prediction Models. Science 2024, 383, 164–167. [Google Scholar] [CrossRef]
- Halevy, A.; Norvig, P.; Pereira, F. The Unreasonable Effectiveness of Data. IEEE Intell. Syst. 2009, 24, 8–12. [Google Scholar] [CrossRef]
- Sweeney, S.M.; Hamadeh, H.K.; Abrams, N.; Adam, S.J.; Brenner, S.; Connors, D.E.; Davis, G.J.; Fiore, L.; Gawel, S.H.; Grossman, R.L.; et al. Challenges to Using Big Data in Cancer. Cancer Res. 2023, 83, 1175–1182. [Google Scholar] [CrossRef]
- Xu, C.; Prete, M.; Webb, S.; Jardine, L.; Stewart, B.J.; Hoo, R.; He, P.; Meyer, K.B.; Teichmann, S.A. Automatic Cell-Type Harmonization and Integration across Human Cell Atlas Datasets. Cell 2023, 186, 5876–5891.e20. [Google Scholar] [CrossRef]
- Selig, M.; Poehlman, L.; Lang, N.C.; Völker, M.; Rolauffs, B.; Hart, M.L. Prediction of Six Macrophage Phenotypes and Their IL-10 Content Based on Single-Cell Morphology Using Artificial Intelligence. Front. Immunol. 2024, 14, 1336393. [Google Scholar] [CrossRef]
- He, H.; Su, W.J. A Law of Data Separation in Deep Learning. Proc. Natl. Acad. Sci. USA 2023, 120, e2221704120. [Google Scholar] [CrossRef] [PubMed]
- Ao, J.; Shao, X.; Liu, Z.; Liu, Q.; Xia, J.; Shi, Y.; Qi, L.; Pan, J.; Ji, M. Stimulated Raman Scattering Microscopy Enables Gleason Scoring of Prostate Core Needle Biopsy by a Convolutional Neural Network. Cancer Res. 2023, 83, 641–651. [Google Scholar] [CrossRef] [PubMed]
- Boral, B.; Togay, A. Automatic Classification of Antinuclear Antibody Patterns With Machine Learning. Cureus 2023, 15, e45008. [Google Scholar] [CrossRef] [PubMed]
- Hanin, B.; Zlokapa, A. Bayesian Interpolation with Deep Linear Networks. Proc. Natl. Acad. Sci. USA 2023, 120, e2301345120. [Google Scholar] [CrossRef] [PubMed]
- Belkin, M.; Hsu, D.; Ma, S.; Mandal, S. Reconciling Modern Machine-Learning Practice and the Classical Bias–Variance Trade-Off. Proc. Natl. Acad. Sci. USA 2019, 116, 15849–15854. [Google Scholar] [CrossRef] [PubMed]
- Papyan, V.; Han, X.Y.; Donoho, D.L. Prevalence of Neural Collapse during the Terminal Phase of Deep Learning Training. Proc. Natl. Acad. Sci. USA 2020, 117, 24652–24663. [Google Scholar] [CrossRef] [PubMed]
- Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. arXiv 2014, arXiv:1409.0575. [Google Scholar] [CrossRef]
- Phan, H.T.H.; Kumar, A.; Kim, J.; Feng, D. Transfer Learning of a Convolutional Neural Network for HEp-2 Cell Image Classification. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 1208–1211. [Google Scholar]
- Ingrosso, A.; Goldt, S. Data-Driven Emergence of Convolutional Structure in Neural Networks. Proc. Natl. Acad. Sci. USA 2022, 119, e2201854119. [Google Scholar] [CrossRef]
- Lake, B.M.; Baroni, M. Human-like Systematic Generalization through a Meta-Learning Neural Network. Nature 2023, 623, 115–121. [Google Scholar] [CrossRef] [PubMed]
- Schneider, C.A.; Rasband, W.S.; Eliceiri, K.W. NIH Image to ImageJ: 25 Years of Image Analysis. Nat. Methods 2012, 9, 671–675. [Google Scholar] [CrossRef] [PubMed]
- André, P.; Benoliel, A.-M.; Capo, C.; Foa, C.; Buferne, M.; Boyer, C.; Schmitt-Verhulst, A.-M.; Bongrand, P. Use of Conjugates Made between a Cytolytic T Cell Clone and Target Cells to Study the Redistribution of Membrane Molecules in Cell Contact Areas. J. Cell Sci. 1990, 97, 335–347. [Google Scholar] [CrossRef] [PubMed]
- Hastie, T.; Tibshirani, R.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd ed.; Springer Series in Statistics; Springer: New York, NY, USA, 2009; ISBN 978-0-387-84857-0. [Google Scholar]
- Azad, R.; Kazerouni, A.; Heidari, M.; Aghdam, E.K.; Molaei, A.; Jia, Y.; Jose, A.; Roy, R.; Merhof, D. Advances in Medical Image Analysis with Vision Transformers: A Comprehensive Review. Med. Image Anal. 2024, 91, 103000. [Google Scholar] [CrossRef]
- Snedecor, G.W.; Cochran, W.G. Statistical Methods, 7th ed.; Iowa State University Press: Ames, IA, USA, 1980; ISBN 978-0-8138-1560-2. [Google Scholar]
- Liao, S.E.; Sudarshan, M.; Regev, O. Deciphering RNA Splicing Logic with Interpretable Machine Learning. Proc. Natl. Acad. Sci. USA 2023, 120, e2221165120. [Google Scholar] [CrossRef]
Analytic Tool | Dataset | Prediction Accuracy (pa) | Corrected Prediction Accuracy (cpa) | Area under ROC Curve (auc) |
---|---|---|---|---|
Logistic regression | train | 0.92 +/− 0.002 | 0.68 +/− 0.006 | 0.95 +/− 0.001 |
test | 0.91 +/− 0.004 | 0.64 +/− 0.014 | 0.95 +/− 0.004 | |
Nearest neighbors (3 neighbors) | training | 0.93 +/− 0.002 | 0.73 +/− 0.007 | 0.88 +/− 0.003 |
testing | 0.89 +/− 0.005 | 0.56 +/− 0.015 | 0.81 +/− 0.007 | |
Decision tree (maximum depth: 3) | training | 0.96+/− 0.002 | 0.83 +/− 0.006 | 0.94 +/− 0.003 |
testing | 0.89 +/− 0.005 | 0.56 +/− 0.016 | 0.84 +/− 0.007 |
Analytic Tool | Dataset | Prediction Accuracy | Corrected Accuracy | Area under ROC Curve (auc) |
---|---|---|---|---|
Logistic regression | training | 0.84 +/− 0.006 | 0.32 +/− 0.024 | 0.79 +/− 0.006 |
testing | 0.77 +/− 0.013 | 0.17 +/− 0.030 | 0.78 +/− 0.030 | |
Nearest neighbors (3 neighbors) | training | 0.85 +/− 0.004 | 0.39 +/− 0.016 | 0.71 +/− 0.008 |
testing | 0.79 +/− 0.011 | 0.23 +/− 0.028 | 0.66 +/− 0.015 | |
Decision tree (maximum depth: 3) | training | 0.94 +/− 0.004 | 0.73 +/− 0.016 | 0.91 +/− 0.006 |
testing | 0.68 +/− 0.014 | 0.04 +/− 0.018 | 0.57 +/− 0.017 |
Analytic Tool | Dataset | Prediction Accuracy (pa) | Corrected Prediction Accuracy (cpa) |
---|---|---|---|
Logistic regression | training | 0.87 +/− 0.002 | 0.66 +/− 0.005 |
testing | 0.82 +/− 0.005 | 0.61 +/− 0.012 | |
Nearest neighbors (3 neighbors) | training | 0.90 +/− 0.002 | 0.75 +/− 0.006 |
testing | 0.81 +/− 0.006 | 0.56 +/− 0.014 | |
Decision tree (maximum depth: 3) | training | 0.90 +/− 0.002 | 0.80 +/− 0.005 |
testing | 0.79 +/− 0.007 | 0.57 +/− 0.013 |
Number of Parameters | Discrimination Parameter | Logistic Regression | Nearest Neighbors (3 Neighbors/Scaling) | Decision Tree (Maximum Depth 3) |
---|---|---|---|---|
2500 (no pca) | cpa training | 1.0 +/− 0.0 | 0.65 +/− 0.019 | 0.64 +/− 0.030 |
cpa testing | 0.31 +/− 0.038 | 0.35 +/− 0.046 | 0.35 +/− 0.047 | |
auc testing | 0.86 +/− 0.014 | 0.76 +/− 0.025 | 0.78 +/− 0.030 | |
20 | cpa training | 0.48 +/− 0.015 | 0.68 +/− 0.019 | 0.58 +/− 0.018 |
cpa testing | 0.45+/− 0.043 | 0.42 +/− 0.045 | 0.38 +/− 0.047 | |
auc testing | 0.92 +/− 0.011 | 0.80 +/− 0.021 | 0.80 +/− 0.025 | |
5 | cpa training | 0.46 +/− 0.015 | 0.66 +/− 0.018 | 0.54 +/− 0.024 |
cpa testing | 0.45 +/− 0.044 | 0.39 +/− 0.041 | 0.40 +/− 0.050 | |
auc testing | 0.92 +/− 0.001 | 0.79 +/− 0.020 | 0.81 +/− 0.031 | |
2 | cpa training | 0.43 +/− 0.016 | 0.60 +/− 0.017 | 0.49 +/− 0.018 |
cpa testing | 0.43 +/− 0.043 | 0.35 +/− 0.039 | 0.38 +/− 0.0045 | |
auc testing | 0.91 +/− 0.012 | 0.77 +/− 0.021 | 0.80 +/− 0.0027 |
ML Algorithm | Dataset | Prediction Accuracy (pa) | Corrected Accuracy (cpa) | Area under ROC Curve (auc) |
---|---|---|---|---|
Logistic regression | training full | 1.00 +/− 0.00 SD | 1.00 +/− 0.00 SD | 1.00 +/− 0.00 SD |
testing full | 0.67 +/− 0.024 SD | 0.04 +/− 0.022 SD | 0.54 +/− 0.019 SD | |
training 20c | 0.76 +/− 0.008 SD | 0.06 +/− 0.018SD | 0.53 +/− 0.009 SD | |
testing 20c | 0.75 +/− 0.020 SD | 0.03 +/− 0.021 SD | 0.52 +/− 0.011 SD | |
Nearest neighbors (3 neighbors, scaling) | training full | 0.84 +/− 0.006 SD | 0.40 +/− 0.017 SD | 0.74 +/− 0.011 SD |
testing full | 0.71 +/− 0.022 SD | 0.08 +/− 0.034 SD | 0.55 +/− 0.025 SD | |
training 20c | 0.85 +/− 0.007 SD | 0.42 +/− 0.023 SD | 0.74 +/− 0.012 SD | |
testing 20c | 0.73 +/− 0.019 SD | 0.11 +/− 0.033 SD | 0.58 +/− 0.022 SD | |
Decision tree (maximum depth: 5) | training full | 0.83 +/− 0.015 SD | 0.34 +/− 0.032 SD | 0.68 +/− 0.33 SD |
testing full | 0.73 +/− 0.023 SD | 0.09 +/− 0.038 SD | 0.56 +/− 0.023 SD | |
training 20c | 0.81 +/− 0.013 SD | 0.28 +/− 0.048 SD | 0.65 +/− 0.036 SD | |
testing 20c | 0.55 +/− 0.023 SD | 0.07 +/− 0.036 SD | 0.55 +/− 0.023 SD |
Model | Feature Number | Controlled Cell Splitting | Prediction Accuracy (pa) | Corrected Accuracy (cpa) | Cohen Kappa Score |
---|---|---|---|---|---|
KNN | 2500 | No | 0.77 +/− 0.01 SE | 0.28 +/− 0.03 SE | 0.45 +/− 0.02 SE |
Yes | 0.77 +/− 0.02 SE | 0.28 +/− 0.05 SE | 0.44 +/− 0.05 SE | ||
20 | No | 0.79 +/− 0.01 SE | 0.32 +/− 0.02 SE | 0.49 +/− 0.02 SE | |
Yes | 0.89 +/− 0.02 SE | 0.57 +/− 0.05 SE | 0.71 +/− 0.04 SE | ||
MLP | 2500 | No | 082 +/− 0.01 SE | 0.39 +/− 0.03 SE | 0.54 +/− 0.03 SE |
Yes | 0.91 +/− 0.02 SE | 0.67 +/− 0.06 SE | 0.79 +/− 0.04 SE | ||
20 | No | 0.82 +/− 0.01 SE | 0.39 +/− 0.02 SE | 0.58 +/− 0.02 SE | |
Yes | 0.94 +/− 0.01 SE | 0.74 +/− 0.05 SE | 0.84 +/− 0.03 SE |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Bertin, D.; Bongrand, P.; Bardin, N. Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies. Int. J. Mol. Sci. 2024, 25, 3270. https://doi.org/10.3390/ijms25063270
Bertin D, Bongrand P, Bardin N. Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies. International Journal of Molecular Sciences. 2024; 25(6):3270. https://doi.org/10.3390/ijms25063270
Chicago/Turabian StyleBertin, Daniel, Pierre Bongrand, and Nathalie Bardin. 2024. "Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies" International Journal of Molecular Sciences 25, no. 6: 3270. https://doi.org/10.3390/ijms25063270
APA StyleBertin, D., Bongrand, P., & Bardin, N. (2024). Comparison of the Capacity of Several Machine Learning Tools to Assist Immunofluorescence-Based Detection of Anti-Neutrophil Cytoplasmic Antibodies. International Journal of Molecular Sciences, 25(6), 3270. https://doi.org/10.3390/ijms25063270