Optimized Screening for At-Risk Students in Mathematics: A Machine Learning Approach
Abstract
:1. Introduction
1.1. Gated Screening Approaches
1.2. Classification Accuracy
1.3. Current Study
- What is the difference in classification accuracy outcomes between single-measure and multi-measure (e.g., gated) screening frameworks?
- To what extent does the cut-score parameter influence the classification accuracy of the screening approach used to identify at-risk students?
2. Methods
2.1. Sample
2.2. Measures
2.2.1. Star Early Literacy
2.2.2. Star Reading
2.2.3. Star Math
2.3. Data Analysis
2.3.1. Single-Stage Screening
2.3.2. Gated Screening
2.3.3. Evaluation Criteria
3. Results
4. Discussion
4.1. Screening for Mathematics Difficulties
4.1.1. Single vs. Gated Screening Approaches
4.1.2. Weighing Evaluation Metrics
4.1.3. The Influence of Cut-Score Parameters
4.2. Limitations and Future Directions
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
FN | False negative classifications |
FP | False positive classifications |
MCC | Matthews correlation coefficient |
RF | Random Forest algorithm |
TN | True negative classifications |
TP | True positive classifications |
References
- Christ, T.J.; Nelson, P.M. Developing and evaluating screening systems: Practical and psychometric considerations. In Universal Screening in Educational Settings: Evidence-Based Decision Making for Schools; Kettler, R.J., Glover, T.A., Albers, C.A., Feeney-Kettler, K.A., Eds.; American Psychological Association: Washington DC, USA, 2014; pp. 79–110. [Google Scholar] [CrossRef]
- Mellard, D.F.; McKnight, M.; Woods, K. Response to intervention screening and progress-monitoring practices in 41 local schools. Learn. Disabil. Res. Pract. 2009, 24, 186–195. [Google Scholar] [CrossRef]
- Franklin, K. The Best Predictor of Future Behavior Is … Past Behavior. 2013. Available online: https://www.psychologytoday.com/us/blog/witness/201301/the-best-predictor-future-behavior-is-past-behavior (accessed on 15 March 2022).
- Cormier, D.C.; Bulut, O.; Niileksela, C.R.; Singh, D.; Funamoto, A.; Schneider, J. Revisiting the relationship between CHC abilities and academic achievement. Presented at the Annual Conference of the National Association of School Psychologists, New Orleans, LA, USA, 17–20 February 2016. [Google Scholar]
- Cormier, D.C.; Yeo, S.; Christ, T.J.; Offrey, L.D.; Pratt, K. An examination of the relationship between computation, problem solving, and reading. Exceptionality 2016, 24, 225–240. [Google Scholar] [CrossRef]
- Kan, A.; Bulut, O.; Cormier, D.C. The impact of item stem format on the dimensional structure of mathematics assessments. Educ. Assess. 2019, 24, 13–32. [Google Scholar] [CrossRef]
- Compton, D.L.; Fuchs, D.; Fuchs, L.S.; Bouton, B.; Gilbert, J.K.; Barquero, L.A.; Cho, E.; Crouch, R.C. Selecting at-risk first-grade readers for early intervention: Eliminating false positives and exploring the promise of a two-stage gated screening process. J. Educ. Psychol. 2010, 102, 327. [Google Scholar] [CrossRef]
- Van Norman, E.R.; Nelson, P.M.; Klingbeil, D.A.; Cormier, D.C.; Lekwa, A.J. Gated screening frameworks for academic concerns: The influence of redundant information on diagnostic accuracy outcomes. Contemp. Sch. Psychol. 2019, 23, 152–162. [Google Scholar] [CrossRef]
- Van Norman, E.R.; Nelson, P.M.; Klingbeil, D.A. Single measure and gated screening approaches for identifying students at-risk for academic problems: Implications for sensitivity and specificity. Sch. Psychol. Q. 2017, 32, 405. [Google Scholar] [CrossRef]
- Glover, T.A.; Albers, C.A. Considerations for evaluating universal screening assessments. J. Sch. Psychol. 2007, 45, 117–135. [Google Scholar] [CrossRef]
- Elliott, S.N.; Huai, N.; Roach, A.T. Universal and early screening for educational difficulties: Current and future approaches. J. Sch. Psychol. 2007, 45, 137–161. [Google Scholar] [CrossRef]
- Fuchs, D.; Fuchs, L.S.; Compton, D.L. Smart RTI: A next-generation approach to multilevel prevention. Except. Child. 2012, 78, 263–279. [Google Scholar] [CrossRef]
- Catts, H.W.; Petscher, Y.; Schatschneider, C.; Sittner Bridges, M.; Mendoza, K. Floor effects associated with universal screening and their impact on the early identification of reading disabilities. J. Learn. Disabil. 2009, 42, 163–176. [Google Scholar] [CrossRef]
- Johnson, E.S.; Jenkins, J.R.; Petscher, Y. Improving the accuracy of a direct route screening process. Assess. Eff. Interv. 2010, 35, 131–140. [Google Scholar] [CrossRef]
- Catts, H.W.; Nielsen, D.C.; Bridges, M.S.; Liu, Y.S.; Bontempo, D.E. Early identification of reading disabilities within an RTI framework. J. Learn. Disabil. 2015, 48, 281–297. [Google Scholar] [CrossRef] [PubMed]
- Nelson, P.M.; Van Norman, E.R.; Lackner, S.K. A comparison of methods to screen middle school students for reading and math difficulties. Sch. Psychol. Rev. 2016, 45, 327–342. [Google Scholar] [CrossRef]
- Klingbeil, D.A.; Van Norman, E.R.; Nelson, P.M.; Birr, C. Interval likelihood ratios: Applications for gated screening in schools. J. Sch. Psychol. 2019, 76, 107–123. [Google Scholar] [CrossRef]
- Poulsen, M.; Nielsen, A.M.V.; Juul, H.; Elbro, C. Early identification of reading difficulties: A screening strategy that adjusts the sensitivity to the level of prediction accuracy. Dyslexia 2017, 23, 251–267. [Google Scholar] [CrossRef] [PubMed]
- Aguiar, E.; Lakkaraju, H.; Bhanpuri, N.; Miller, D.; Yuhas, B.; Addison, K.L. Who, when, and why: A machine learning approach to prioritizing students at risk of not graduating high school on time. In Proceedings of the Fifth International Conference on Learning Analytics And Knowledge, Poughkeepsie, NY, USA, 16–20 March 2015; pp. 93–102. [Google Scholar] [CrossRef]
- Lakkaraju, H.; Aguiar, E.; Shan, C.; Miller, D.; Bhanpuri, N.; Ghani, R.; Addison, K.L. A machine learning framework to identify students at risk of adverse academic outcomes. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, 10–13 August 2015; pp. 1909–1918. [Google Scholar] [CrossRef]
- Márquez-Vera, C.; Cano, A.; Romero, C.; Noaman, A.Y.M.; Mousa Fardoun, H.; Ventura, S. Early dropout prediction using data mining: A case study with high school students. Expert Syst. 2016, 33, 107–124. [Google Scholar] [CrossRef]
- Hamsa, H.; Indiradevi, S.; Kizhakkethottam, J.J. Student academic performance prediction model using decision tree and fuzzy genetic algorithm. Procedia Technol. 2016, 25, 326–332. [Google Scholar] [CrossRef]
- Tamhane, A.; Ikbal, S.; Sengupta, B.; Duggirala, M.; Appleton, J. Predicting student risks through longitudinal analysis. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 1544–1552. [Google Scholar] [CrossRef]
- Hintze, J.M.; Silberglitt, B. A longitudinal examination of the diagnostic accuracy and predictive validity of R-CBM and high-stakes testing. Sch. Psychol. Rev. 2005, 34, 372–386. [Google Scholar] [CrossRef]
- Kilgus, S.P.; Chafouleas, S.M.; Riley-Tillman, T.C. Development and initial validation of the Social and Academic Behavior Risk Screener for elementary grades. Sch. Psychol. Q. 2013, 28, 210. [Google Scholar] [CrossRef]
- Renaissance. Star Assessments™ for Reading Technical Manual; Technical Report; Renaissance: Wisconsin Rapids, WI, USA, 2018. [Google Scholar]
- Renaissance. Star Assessments™ for Math Technical Manual; Technical Report; Renaissance: Wisconsin Rapids, WI, USA, 2018. [Google Scholar]
- Renaissance. Star Early Literacy™ for Early Literacy Technical Manual; Technical Report; Renaissance: Wisconsin Rapids, WI, USA, 2018. [Google Scholar]
- Dumont, R.; Willis, J.O.; Veizel, K.; Zibulsky, J. Wechsler Individual Achievement Test–Third Edition. In Encyclopedia of Special Education: A Reference for the Education of Children, Adolescents, and Adults with Disabilities and Other Exceptional Individuals; Wiley: Hoboken, NJ, USA, 2013. [Google Scholar]
- Fletcher, J.M.; Barth, A.E.; Stuebing, K.K. A response to intervention (RTI) approach to SLD identification. In Essentials of Specific Learning Disability Identification; Flanagan, D.P., Alfonso, V.C., Eds.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2011; pp. 115–144. [Google Scholar]
- Kuhn, M. caret: Classification and Regression Training. R Package Version 6.0-90. 2021. Available online: https://cran.r-project.org/web/packages/caret/index.html (accessed on 5 March 2022).
- R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2021. [Google Scholar]
- Friedman, J.; Hastie, T.; Tibshirani, R. Additive logistic regression: A statistical view of boosting (with discussion and a rejoinder by the authors). Ann. Stat. 2000, 28, 337–407. [Google Scholar] [CrossRef]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Couronné, R.; Probst, P.; Boulesteix, A.L. Random forest versus logistic regression: A large-scale benchmark experiment. BMC Bioinform. 2018, 19, 270. [Google Scholar] [CrossRef] [PubMed]
- Menardi, G.; Torelli, N. Training and assessing classification rules with imbalanced data. Data Min. Knowl. Discov. 2014, 28, 92–122. [Google Scholar] [CrossRef]
- Lunardon, N.; Menardi, G.; Torelli, N. ROSE: A Package for Binary Imbalanced Learning. R Journal 2014, 6, 79–89. [Google Scholar] [CrossRef]
- Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef]
- American Educational Research Association; American Psychological Association; National Council on Measurement in Education. Standards for Educational and Psychological Testing; American Educational Research Association: Washington, DC, USA, 2014. [Google Scholar]
- Bulut, O.; Cormier, D.C. Validity evidence for progress monitoring with Star Reading: Slope estimates, administration frequency, and number of data points. Front. Educ. 2018, 3. [Google Scholar] [CrossRef]
- Lambert, R.; Algozzine, B.; McGee, J. Effects of progress monitoring on math performance of at-risk students. Br. J. Educ. Soc. Behav. Sci. 2014, 4, 527–540. [Google Scholar] [CrossRef]
- Shapiro, E.S.; Dennis, M.S.; Fu, Q. Comparing computer adaptive and curriculum-based measures of math in progress monitoring. Sch. Psychol. Q. 2015, 30, 470. [Google Scholar] [CrossRef]
- Klingbeil, D.A.; Van Norman, E.R.; Nelson, P.M. Using interval likelihood ratios in gated screening: A direct replication study. Assess. Eff. Interv. 2021, 47, 14–25. [Google Scholar] [CrossRef]
- Turner, M.I.; Van Norman, E.R.; Hojnoski, R.L. An Independent Evaluation of the Diagnostic Accuracy of a Computer Adaptive Test to Predict Proficiency on an End of Year High-Stakes Assessment. J. Psychoeduc. Assess. 2022. [Google Scholar] [CrossRef]
- Fernández, Á.; Bella, J.; Dorronsoro, J.R. Supervised outlier detection for classification and regression. Neurocomputing 2022, 486, 77–92. [Google Scholar] [CrossRef]
- Bulut, O.; Cormier, D.C.; Shin, J. An intelligent recommender system for personalized test administration scheduling with computerized formative assessments. Front. Educ. 2020, 5, 182. [Google Scholar] [CrossRef]
- Bulut, O.; Shin, J.; Cormier, D.C. Learning analytics and computerized formative Assessments: An application of Dijkstra’s shortest path algorithm for personalized test scheduling. Mathematics 2022, 10, 2230. [Google Scholar] [CrossRef]
- Shin, J.; Bulut, O. Building an intelligent recommendation system for personalized test scheduling in computerized assessments: A reinforcement learning approach. Behav. Res. Methods 2022, 54, 216–232. [Google Scholar] [CrossRef] [PubMed]
Screening Method | Algorithm | Prediction Model | Sampling | Sensitivity | Specificity | Accuracy |
---|---|---|---|---|---|---|
Single-Stage | RF | Model 0 (SM) | Oversampling | 0.797 | 0.717 | 0.731 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Oversampling | 0.685 | 0.839 | 0.812 |
Single-Stage | LogitBoost | Model 0 (SM) | Oversampling | 0.713 | 0.790 | 0.777 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Oversampling | 0.835 | 0.726 | 0.744 |
Single-Stage | RF | Model 0 (SM) | Undersampling | 0.799 | 0.706 | 0.722 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Undersampling | 0.829 | 0.737 | 0.753 |
Single-Stage | LogitBoost | Model 0 (SM) | Undersampling | 0.711 | 0.792 | 0.778 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Undersampling | 0.788 | 0.781 | 0.782 |
Single-Stage | RF | Model 0 (SM) | ROSE | 0.635 | 0.793 | 0.766 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | ROSE | 0.882 | 0.667 | 0.704 |
Single-Stage | LogitBoost | Model 0 (SM) | ROSE | 0.856 | 0.690 | 0.719 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | ROSE | 0.866 | 0.665 | 0.699 |
Gated (Standard) | - | 30% Threshold | - | 0.468 | 0.923 | 0.844 |
Gated (Standard) | - | 40% Threshold | - | 0.624 | 0.863 | 0.822 |
Gated (Standard) | - | 50% Threshold | - | 0.778 | 0.781 | 0.780 |
Gated (Mixed) | RF | 30% Threshold | Oversampling | 0.698 | 0.788 | 0.773 |
Gated (Mixed) | RF | 40% Threshold | Oversampling | 0.781 | 0.739 | 0.746 |
Gated (Mixed) | RF | 50% Threshold | Oversampling | 0.863 | 0.670 | 0.703 |
Gated (Mixed) | LogitBoost | 30% Threshold | Oversampling | 0.570 | 0.868 | 0.816 |
Gated (Mixed) | LogitBoost | 40% Threshold | Oversampling | 0.892 | 0.568 | 0.624 |
Gated (Mixed) | LogitBoost | 50% Threshold | Oversampling | 0.896 | 0.594 | 0.646 |
Gated (Mixed) | RF | 30% Threshold | Undersampling | 0.703 | 0.787 | 0.773 |
Gated (Mixed) | RF | 40% Threshold | Undersampling | 0.784 | 0.736 | 0.744 |
Gated (Mixed) | RF | 50% Threshold | Undersampling | 0.876 | 0.663 | 0.699 |
Gated (Mixed) | LogitBoost | 30% Threshold | Undersampling | 0.627 | 0.848 | 0.810 |
Gated (Mixed) | LogitBoost | 40% Threshold | Undersampling | 0.704 | 0.804 | 0.787 |
Gated (Mixed) | LogitBoost | 50% Threshold | Undersampling | 0.885 | 0.609 | 0.657 |
Gated (Mixed) | RF | 30% Threshold | ROSE | 0.737 | 0.767 | 0.762 |
Gated (Mixed) | RF | 40% Threshold | ROSE | 0.791 | 0.712 | 0.726 |
Gated (Mixed) | RF | 50% Threshold | ROSE | 0.879 | 0.641 | 0.682 |
Gated (Mixed) | LogitBoost | 30% Threshold | ROSE | 0.909 | 0.504 | 0.574 |
Gated (Mixed) | LogitBoost | 40% Threshold | ROSE | 0.679 | 0.808 | 0.786 |
Gated (Mixed) | LogitBoost | 50% Threshold | ROSE | 0.930 | 0.529 | 0.598 |
Screening Method | Algorithm | Prediction Model | Sampling | Sensitivity | Specificity | Accuracy |
---|---|---|---|---|---|---|
Single-Stage | RF | Model 0 (SM) | Oversampling | 0.784 | 0.714 | 0.727 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Oversampling | 0.678 | 0.831 | 0.803 |
Single-Stage | LogitBoost | Model 0 (SM) | Oversampling | 0.731 | 0.775 | 0.767 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Oversampling | 0.843 | 0.710 | 0.734 |
Single-Stage | RF | Model 0 (SM) | Undersampling | 0.779 | 0.715 | 0.727 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Undersampling | 0.838 | 0.722 | 0.743 |
Single-Stage | LogitBoost | Model 0 (SM) | Undersampling | 0.701 | 0.790 | 0.774 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Undersampling | 0.842 | 0.670 | 0.701 |
Single-Stage | RF | Model 0 (SM) | ROSE | 0.683 | 0.738 | 0.728 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | ROSE | 0.873 | 0.664 | 0.702 |
Single-Stage | LogitBoost | Model 0 (SM) | ROSE | 0.847 | 0.678 | 0.708 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | ROSE | 0.898 | 0.607 | 0.659 |
Gated (Standard) | - | 30% Threshold | - | 0.461 | 0.914 | 0.832 |
Gated (Standard) | - | 40% Threshold | - | 0.622 | 0.851 | 0.809 |
Gated (Standard) | - | 50% Threshold | - | 0.758 | 0.764 | 0.763 |
Gated (Mixed) | RF | 30% Threshold | Oversampling | 0.713 | 0.776 | 0.764 |
Gated (Mixed) | RF | 40% Threshold | Oversampling | 0.779 | 0.728 | 0.737 |
Gated (Mixed) | RF | 50% Threshold | Oversampling | 0.849 | 0.661 | 0.695 |
Gated (Mixed) | LogitBoost | 30% Threshold | Oversampling | 0.729 | 0.682 | 0.690 |
Gated (Mixed) | LogitBoost | 40% Threshold | Oversampling | 0.717 | 0.776 | 0.765 |
Gated (Mixed) | LogitBoost | 50% Threshold | Oversampling | 0.824 | 0.679 | 0.705 |
Gated (Mixed) | RF | 30% Threshold | Undersampling | 0.710 | 0.776 | 0.764 |
Gated (Mixed) | RF | 40% Threshold | Undersampling | 0.783 | 0.720 | 0.731 |
Gated (Mixed) | RF | 50% Threshold | Undersampling | 0.858 | 0.651 | 0.688 |
Gated (Mixed) | LogitBoost | 30% Threshold | Undersampling | 0.615 | 0.831 | 0.792 |
Gated (Mixed) | LogitBoost | 40% Threshold | Undersampling | 0.736 | 0.763 | 0.758 |
Gated (Mixed) | LogitBoost | 50% Threshold | Undersampling | 0.821 | 0.694 | 0.717 |
Gated (Mixed) | RF | 30% Threshold | ROSE | 0.716 | 0.762 | 0.753 |
Gated (Mixed) | RF | 40% Threshold | ROSE | 0.808 | 0.690 | 0.711 |
Gated (Mixed) | RF | 50% Threshold | ROSE | 0.864 | 0.637 | 0.678 |
Gated (Mixed) | LogitBoost | 30% Threshold | ROSE | 0.698 | 0.793 | 0.776 |
Gated (Mixed) | LogitBoost | 40% Threshold | ROSE | 0.678 | 0.803 | 0.780 |
Gated (Mixed) | LogitBoost | 50% Threshold | ROSE | 0.867 | 0.619 | 0.664 |
Screening Method | Algorithm | Prediction Model | Sampling | 2016–2017 | 2017–2018 | ||
---|---|---|---|---|---|---|---|
MCC | MCC | ||||||
Single-Stage | RF | Model 0 (SM) | Oversampling | 0.755 | 0.414 | 0.747 | 0.418 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Oversampling | 0.754 | 0.450 | 0.747 | 0.437 |
Single-Stage | LogitBoost | Model 0 (SM) | Oversampling | 0.750 | 0.417 | 0.752 | 0.418 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Oversampling | 0.777 | 0.438 | 0.771 | 0.453 |
Single-Stage | RF | Model 0 (SM) | Undersampling | 0.750 | 0.393 | 0.746 | 0.413 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | Undersampling | 0.780 | 0.445 | 0.776 | 0.439 |
Single-Stage | LogitBoost | Model 0 (SM) | Undersampling | 0.749 | 0.416 | 0.743 | 0.412 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | Undersampling | 0.784 | 0.462 | 0.746 | 0.399 |
Single-Stage | RF | Model 0 (SM) | ROSE | 0.705 | 0.343 | 0.709 | 0.355 |
Single-Stage | RF | Model 1 (SM + SR + SEL) | ROSE | 0.760 | 0.432 | 0.754 | 0.414 |
Single-Stage | LogitBoost | Model 0 (SM) | ROSE | 0.764 | 0.411 | 0.753 | 0.410 |
Single-Stage | LogitBoost | Model 1 (SM + SR + SEL) | ROSE | 0.752 | 0.442 | 0.724 | 0.388 |
Gated (Standard) | - | 30% Threshold | - | 0.621 | 0.420 | 0.613 | 0.399 |
Gated (Standard) | - | 40% Threshold | - | 0.724 | 0.444 | 0.719 | 0.428 |
Gated (Standard) | - | 50% Threshold | - | 0.779 | 0.448 | 0.761 | 0.429 |
Gated (Mixed) | RF | 30% Threshold | Oversampling | 0.740 | 0.401 | 0.743 | 0.403 |
Gated (Mixed) | RF | 40% Threshold | Oversampling | 0.756 | 0.413 | 0.753 | 0.406 |
Gated (Mixed) | RF | 50% Threshold | Oversampling | 0.754 | 0.415 | 0.743 | 0.400 |
Gated (Mixed) | LogitBoost | 30% Threshold | Oversampling | 0.688 | 0.405 | 0.705 | 0.397 |
Gated (Mixed) | LogitBoost | 40% Threshold | Oversampling | 0.694 | 0.438 | 0.745 | 0.409 |
Gated (Mixed) | LogitBoost | 50% Threshold | Oversampling | 0.714 | 0.427 | 0.745 | 0.402 |
Gated (Mixed) | RF | 30% Threshold | Undersampling | 0.743 | 0.406 | 0.742 | 0.404 |
Gated (Mixed) | RF | 40% Threshold | Undersampling | 0.759 | 0.411 | 0.750 | 0.400 |
Gated (Mixed) | RF | 50% Threshold | Undersampling | 0.755 | 0.411 | 0.740 | 0.395 |
Gated (Mixed) | LogitBoost | 30% Threshold | Undersampling | 0.721 | 0.424 | 0.707 | 0.397 |
Gated (Mixed) | LogitBoost | 40% Threshold | Undersampling | 0.751 | 0.426 | 0.749 | 0.417 |
Gated (Mixed) | LogitBoost | 50% Threshold | Undersampling | 0.722 | 0.374 | 0.752 | 0.401 |
Gated (Mixed) | RF | 30% Threshold | ROSE | 0.752 | 0.406 | 0.738 | 0.404 |
Gated (Mixed) | RF | 40% Threshold | ROSE | 0.749 | 0.418 | 0.744 | 0.400 |
Gated (Mixed) | RF | 50% Threshold | ROSE | 0.741 | 0.405 | 0.733 | 0.395 |
Gated (Mixed) | LogitBoost | 30% Threshold | ROSE | 0.648 | 0.417 | 0.742 | 0.397 |
Gated (Mixed) | LogitBoost | 40% Threshold | ROSE | 0.738 | 0.376 | 0.735 | 0.417 |
Gated (Mixed) | LogitBoost | 50% Threshold | ROSE | 0.674 | 0.423 | 0.722 | 0.401 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Bulut, O.; Cormier, D.C.; Yildirim-Erbasli, S.N. Optimized Screening for At-Risk Students in Mathematics: A Machine Learning Approach. Information 2022, 13, 400. https://doi.org/10.3390/info13080400
Bulut O, Cormier DC, Yildirim-Erbasli SN. Optimized Screening for At-Risk Students in Mathematics: A Machine Learning Approach. Information. 2022; 13(8):400. https://doi.org/10.3390/info13080400
Chicago/Turabian StyleBulut, Okan, Damien C. Cormier, and Seyma Nur Yildirim-Erbasli. 2022. "Optimized Screening for At-Risk Students in Mathematics: A Machine Learning Approach" Information 13, no. 8: 400. https://doi.org/10.3390/info13080400
APA StyleBulut, O., Cormier, D. C., & Yildirim-Erbasli, S. N. (2022). Optimized Screening for At-Risk Students in Mathematics: A Machine Learning Approach. Information, 13(8), 400. https://doi.org/10.3390/info13080400