Predicting Students’ Outcome in an Introductory Programming Course: Leveraging the Student Background
Abstract
:1. Introduction
2. Related Work
3. Data and Methods
- Passed the course in extra-ordinary instances, such as intensive courses (theory courses conducted between semesters that replace the final theory grade) or special exams (additional sufficiency exams that allow students to improve their final theory grade).
- Were repeating the course after a previous failure.
- Entered their study programme before the observation period.
- Were not studying an engineering programme (e.g., the College programme).
- Had an incomplete record.
- Pearson correlation test for two numeric variables.
- Independent samples t-test (or Wilcoxon rank sums if conditions are not met) for a dichotomous and a numeric variable.
- Independent samples ANOVA (or Kruskal–Wallis if conditions are not met) for a numeric and a categorical variable with more than two levels.
- association test for two categorical variables.
- Two hyper-parameters must be tuned for radial kernel SVMs: cost (C), which penalises wrong classification, and sigma, which regulates the curvature of the decision border. Values of powers of two ranging from to were considered for both parameters.
- The selected implementation of a Random Forest only allows the mtry hyper-parameter to be tuned, which regulates how many of the input features are to be considered when building a decision tree, ranging from one to the number of available features.
- For XGB, the nrounds parameter determines the number of trees in the final model, for which the range 100 to 2000 was considered, increasing by 50. eta prevents over-fitting by adjusting feature weights and considers the following possible values: 0.025 and 0.05 to 0.5, increasing by 0.05. max_depth regulates the maximum depth of a tree, ranging from one to the number of available features. min_child_weight regulates the minimum number of observations in each node, ranging from 1 to 10. colsample_bytree is the sub-sample ratio of columns when building each tree, ranging from 0.1 to 1 and increasing by 0.1. gamma regulates the minimum loss reduction needed to further partition a node, considering the same values listed for eta. subsample regulates the subset of observations sampled before growing the trees. This parameter was set to one since cross-validation already separates instances for model assessment.
4. Results
5. Discussion
6. Conclusions and Future Work
- By removing correlated variables, it was possible to obtain results that were closely comparable to those achieved in prior work [27] with simpler and more explainable models.
- Models were built using all the students in the sample, as well as subsets based on programme duration and the department responsible for the programme. Except for MLR, all of the methods outperform the others in certain scenarios, with SVMs surpassing the others in the majority of cases.
- With only two features—the students’ GPA and the result on the entrance language test—a model was created that can accurately predict whether students pass or fail the theory part of FCYP with an accuracy of 67.71%.
- Specific models were built for each subset of students. With the exception of Geography and Metallurgy specialisations, all departmental models improved the accuracy of the base model, with Informatics, Industry, Civil Engineering, and Electricity showing the best results.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
ANN | Artificial Neural Networks |
ATS | ANOVA-type statistic |
CI | Confidence interval |
DT | Decision Tree |
FCYP | Fundamentos de Computación y Programación, an initial programming course |
FING | Facultad de Ingeniería, Engineering Faculty |
GPA | Grade Point Average |
LR | Linear Regression |
MLR | Multivariate Logistic Regression |
NB | Naive Bayes |
RF | Random Forest |
RFE | Recursive Feature Elimination |
STEM | Science, Technology, Engineering y Mathematics |
SVM | Support Vector Machine |
XGB | Extreme Gradient Boosting |
Appendix A. Supplementary Confusion Matrices
References
- World Economic Forum. The Future of Jobs: Employment, Skills and Workforce Strategy for the Fourth Industrial Revolution; World Economic Forum: Geneva, Switzerland, 2016. [Google Scholar]
- Leidl, K.D.; Bers, M.U.; Mihm, C. Programming with ScratchJr: A review of the first year of user analytics. In Proceedings of the International Conference on Computational Thinking Education, Wanchai, Hong Kong, 13–15 July 2017; pp. 116–121. [Google Scholar]
- De Kereki, I.F.; Manataki, A. “Code Yourself” and “A Programar”: A bilingual MOOC for teaching computer science to teenagers. In Proceedings of the 2016 IEEE Frontiers in Education Conference (FIE), Erie, PA, USA, 12–15 October 2016; IEEE: New York, NY, USA, 2016; pp. 1–9. [Google Scholar]
- Kalelioğlu, F. A new way of teaching programming skills to K-12 students: Code. org. Comput. Hum. Behav. 2015, 52, 200–210. [Google Scholar] [CrossRef]
- Chen, X.; Liu, W. The Value of Python Programming in General Education and Comprehensive Quality Improvement of Medical Students Based on a Retrospective Cohort Study. J. Healthc. Eng. 2022, 2022, 4043992. [Google Scholar] [CrossRef]
- dos Santos, M.T.; Vianna, A.S., Jr.; Le Roux, G.A. Programming skills in the industry 4.0: Are chemical engineering students able to face new problems? Educ. Chem. Eng. 2018, 22, 69–76. [Google Scholar] [CrossRef]
- Hansen, S.M. Deconstruction/Reconstruction: A pedagogic method for teaching programming to graphic designers. In Proceedings of the Generative Arts Conference 2017, Ravenna, Italy, 13–15 December 2017; Generative Art Conference: Milan, Italy, 2017; pp. 419–431. [Google Scholar]
- Lee, Y.J.; Lien, K.W. Reconstruct Programming 101 for Social Science Preference Students. In Proceedings of the 2019 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), Yilan, Taiwan, 20–22 May 2019; IEEE: New York, NY, USA, 2019; pp. 1–2. [Google Scholar]
- Piteira, M.; Costa, C. Computer programming and novice programmers. In Proceedings of the Workshop on Information Systems and Design of Communication, Lisbon, Portugal, 11 June 2012; pp. 51–53. [Google Scholar]
- Cheah, C.S. Factors contributing to the difficulties in teaching and learning of computer programming: A literature review. Contemp. Educ. Technol. 2020, 12, ep272. [Google Scholar] [CrossRef] [PubMed]
- Medeiros, R.P.; Ramalho, G.L.; Falcão, T.P. A Systematic Literature Review on Teaching and Learning Introductory Programming in Higher Education. IEEE Trans. Educ. 2019, 62, 77–90. [Google Scholar] [CrossRef]
- Tsai, C.Y. Improving students’ understanding of basic programming concepts through visual programming language: The role of self-efficacy. Comput. Hum. Behav. 2019, 95, 224–232. [Google Scholar] [CrossRef]
- Emerson, A.; Rodríguez, F.J.; Mott, B.; Smith, A.; Min, W.; Boyer, K.E.; Smith, C.; Wiebe, E.; Lester, J. Predicting Early and Often: Predictive Student Modeling for Block-Based Programming Environments. In Proceedings of the 12th International Conference on Educational Data Mining, Montreal, QC, Canada, 2–5 July 2019. [Google Scholar]
- Sobral, R.; Oliveira, F. Predicting students performance in introductory programming courses: A literature review. In INTED2021 Proceedings, Proceedings of the 15th International Technology, Education and Development Conference, Online, 8–9 March 2021; IATED: Valencia, Spain, 2021; pp. 7402–7412. [Google Scholar] [CrossRef]
- Biamonte, A.J. Predicting success in programmer training. In Proceedings of the Second SIGCPR Conference on Computer Personnel Research, New York, NY, USA, 20–21 July 1964; pp. 9–12. [Google Scholar]
- Leeper, R.R.; Silver, J.L. Predicting success in a first programming course. ACM SIGCSE Bull. 1982, 14, 147–150. [Google Scholar] [CrossRef]
- Bergin, S.; Reilly, R. Programming: Factors that influence success. In Proceedings of the 36th SIGCSE technical Symposium on Computer Science Education, St. Louis, MO, USA, 23–27 February 2005; pp. 411–415. [Google Scholar]
- Loksa, D.; Ko, A.J. The role of self-regulation in programming problem solving process and success. In Proceedings of the 2016 ACM Conference on International Computing Education Research, Melbourne, VIC, Australia, 8–12 September 2016; pp. 83–91. [Google Scholar]
- Prather, J.; Pettit, R.; McMurry, K.; Peters, A.; Homer, J.; Cohen, M. Metacognitive difficulties faced by novice programmers in automated assessment tools. In Proceedings of the 2018 ACM Conference on International Computing Education Research, Espoo, Finland, 13–15 August 2018; pp. 41–50. [Google Scholar]
- Lopez, M.; Whalley, J.; Robbins, P.; Lister, R. Relationships between reading, tracing and writing skills in introductory programming. In Proceedings of the Fourth International Workshop on Computing Education Research, Sydney, Australia, 6–7 September 2008; pp. 101–112. [Google Scholar]
- Qian, Y.; Lehman, J.D. Correlates of success in introductory programming: A study with middle school students. J. Educ. Learn. 2016, 5, 73–83. [Google Scholar] [CrossRef]
- Hinckle, M.; Rachmatullah, A.; Mott, B.; Boyer, K.E.; Lester, J.; Wiebe, E. The relationship of gender, experiential, and psychological factors to achievement in computer science. In Proceedings of the 2020 ACM Conference on Innovation and Technology in Computer Science Education, Trondheim, Norway, 15–19 June 2020; pp. 225–231. [Google Scholar]
- Álvarez, C.; Fajardo, C.; Meza, F.; Vásquez, A. An exploration of STEM freshmen’s attitudes, engagement and autonomous learning in introductory computer programming. In Proceedings of the 2019 38th International Conference of the Chilean Computer Science Society (SCCC), Concepcion, Chile, 4–9 November 2019; IEEE: New York, NY, USA, 2019; pp. 1–8. [Google Scholar]
- Alvarez, C.; Wise, A.; Altermatt, S.; Aranguiz, I. Predicting academic results in a modular computer programming course. In Proceedings of the 2nd Latin American Conference on Learning Analytics, LALA, Valdivia, Chile, 18–19 March 2019; Volume 2425, pp. 21–30. [Google Scholar]
- Bellino, A.; Herskovic, V.; Hund, M.; Munoz-Gama, J. A real-world approach to motivate students on the first class of a computer science course. ACM Trans. Comput. Educ. (TOCE) 2021, 21, 1–23. [Google Scholar] [CrossRef]
- Moonsamy, D.; Naicker, N.; Adeliyi, T.T.; Ogunsakin, R.E. A meta-analysis of educational data mining for predicting students performance in programming. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 97–104. [Google Scholar] [CrossRef]
- Köhler, J.; Hidalgo, L.; Jara, J.L. Using machine learning techniques to predict academic success in an introductory programming course. In Proceedings of the 2022 41st International Conference of the Chilean Computer Science Society (SCCC), Santiago, Chile, 21–25 November 2022; IEEE: New York, NY, USA, 2022; pp. 1–8. [Google Scholar]
- Alturki, S.; Hulpuș, I.; Stuckenschmidt, H. Predicting Academic Outcomes: A Survey from 2007 Till 2018. Technol. Knowl. Learn. 2022, 27, 275–307. [Google Scholar] [CrossRef]
- Alsariera, Y.A.; Baashar, Y.; Alkawsi, G.; Mustafa, A.; Alkahtani, A.A.; Ali, N. Assessment and Evaluation of Different Machine Learning Algorithms for Predicting Student Performance. Comput. Intell. Neurosci. 2022, 2022, 4151487. [Google Scholar] [CrossRef] [PubMed]
- Abu Saa, A.; Al-Emran, M.; Shaalan, K. Factors Affecting Students’ Performance in Higher Education: A Systematic Review of Predictive Data Mining Techniques. Technol. Knowl. Learn. 2019, 24, 567–598. [Google Scholar] [CrossRef]
- Ismail, N.; Yusof, U.K. A systematic literature review: Recent techniques of predicting STEM stream students. Comput. Educ. Artif. Intell. 2023, 5, 100141. [Google Scholar] [CrossRef]
- Al-Fairouz, E.; Al-Hagery, M. Students performance: From detection of failures and anomaly cases to the solutions-based mining algorithms. Int. J. Eng. Res. Technol. 2020, 13, 2895–2908. [Google Scholar] [CrossRef]
- Sandoval, A.; Gonzalez, C.; Alarcon, R.; Pichara, K.; Montenegro, M. Centralized student performance prediction in large courses based on low-cost variables in an institutional context. Internet High. Educ. 2018, 37, 76–89. [Google Scholar] [CrossRef]
- Falát, L.; Piscová, T. Predicting GPA of University Students with Supervised Regression Machine Learning Models. Appl. Sci. 2022, 12, 8403. [Google Scholar] [CrossRef]
- Beaulac, C.; Rosenthal, J.S. Predicting University Students’ Academic Success and Major Using Random Forests. Res. High. Educ. 2019, 60, 1048–1064. [Google Scholar] [CrossRef]
- Gil, P.D.; da Cruz Martins, S.; Moro, S.; Costa, J.M. A data-driven approach to predict first-year students’ academic success in higher education institutions. Educ. Inf. Technol. 2021, 26, 2165–2190. [Google Scholar] [CrossRef]
- Aluko, R.O.; Daniel, E.I.; Shamsideen Oshodi, O.; Aigbavboa, C.O.; Abisuga, A.O. Towards reliable prediction of academic performance of architecture students using data mining techniques. J. Eng. Des. Technol. 2018, 16, 385–397. [Google Scholar] [CrossRef]
- Barker, R.J.; Unger, E.A. A Predictor for Success in an Introductory Programming Class Based upon Abstract Reasoning Development. SIGCSE Bull. 1983, 15, 154–158. [Google Scholar] [CrossRef]
- Costa, E.B.; Fonseca, B.; Santana, M.A.; de Araújo, F.F.; Rego, J. Evaluating the effectiveness of educational data mining techniques for early prediction of students’ academic failure in introductory programming courses. Comput. Hum. Behav. 2017, 73, 247–256. [Google Scholar] [CrossRef]
- Sivasakthi, M.; Padmanabhan, K.R.A. Prediction of Students Programming Performance Using Naïve Bayesian and Decision Tree. In Soft Computing for Security Applications; Ranganathan, G., Fernando, X., Piramuthu, S., Eds.; Springer: Singapore, 2023; pp. 97–106. [Google Scholar]
- Shen, G.; Yang, S.; Huang, Z.; Yu, Y.; Li, X. The prediction of programming performance using student profiles. Educ. Inf. Technol. 2023, 28, 725–740. [Google Scholar] [CrossRef]
- Van Petegem, C.; Deconinck, L.; Mourisse, D.; Maertens, R.; Strijbol, N.; Dhoedt, B.; De Wever, B.; Dawyndt, P.; Mesuere, B. Pass/Fail Prediction in Programming Courses. J. Educ. Comput. Res. 2023, 61, 68–95. [Google Scholar] [CrossRef]
- Araya Sánchez, V.; Fuentes Bravo, F.; Salazar Loyola, J.; Melo Fuenzalida, P.; Rickmers Blamey, B. Characterization of Assessments on a First Programming Course in Higher Education. In Proceedings of the 2022 41st International Conference of the Chilean Computer Science Society (SCCC), Santiago, Chile, 21–25 November 2022; pp. 1–8. [Google Scholar] [CrossRef]
- Bello, F.A.; Köhler, J.; Hinrechsen, K.; Araya, V.; Hidalgo, L.; Jara, J.L. Using machine learning methods to identify significant variables for the prediction of first-year Informatics Engineering students dropout. In Proceedings of the 2020 39th International Conference of the Chilean Computer Science Society (SCCC), Coquimbo, Chile, 16–20 November 2020; IEEE: New York, NY, USA, 2020; pp. 1–5. [Google Scholar]
- Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene Selection for Cancer Classification using Support Vector Machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
- Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef] [PubMed]
- Menard, S. Coefficients of determination for multiple logistic regression analysis. Am. Stat. 2000, 54, 17–24. [Google Scholar]
- Cutler, A.; Cutler, D.R.; Stevens, J.R. Random forests. In Ensemble Machine Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 157–175. [Google Scholar]
- Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 2016 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
- Friedrich, S.; Konietschke, F.; Pauly, M. The R Journal: Resampling-Based Analysis of Multivariate Data and Repeated Measures Designs with the R Package MANOVA.RM. R J. 2019, 11, 380–400. [Google Scholar] [CrossRef]
- Ivanova, A.A.; Srikant, S.; Sueoka, Y.; Kean, H.H.; Dhamala, R.; O’Reilly, U.M.; Bers, M.U.; Fedorenko, E. Comprehension of computer code relies primarily on domain-general executive brain regions. eLife 2020, 9, e58906. [Google Scholar] [CrossRef] [PubMed]
- Prat, C.S.; Madhyastha, T.M.; Mottarella, M.J.; Kuo, C.H. Relating natural language aptitude to individual differences in learning programming languages. Sci. Rep. 2020, 10, 3817. [Google Scholar] [CrossRef]
Variable | Type | Description |
---|---|---|
PROG | Categorical | Programme code (20 levels). |
P_TYPE | Dichotomous | Indicates the programme duration. |
DPT | Categorical | Indicates which department manages the programme. |
PREF | Categorical | Preference order of the programme in application process. |
PSU_SCI | Integer | Score in Science admission test. |
PSU_LAN | Integer | Score in Language admission test. |
PSU_MAT | Integer | Score in Maths admission test. |
PSU_GPA | Integer | Normalised high-school grade point average. |
PSU_RAN | Integer | Normalised ranking score. |
PSU_AVG | Integer | Weighted average admission score. |
FEE_EX | Dichotomous | Indicates if the student has been granted the fee-exemption benefit. |
SCHOOL | Categorical | Type of high school of origin (municipal, private, or subsidised). |
QUINTILE | Categorical | Family income quintile at entrance. |
PREC_HDI | Real | Human development index of the student’s municipality of residence. |
CAL | Real | Final grade after taking Calculus I for the first time . |
PHY | Real | Final grade after taking Physics I for the first time . |
ALG | Real | Final grade after taking Algebra I for the first time . |
MET | Real | Final grade after taking Study Methods for the first time . |
INT | Real | Final grade after taking Introduction to Engineering for the first time . |
GPA | Real | Grade point average after the first semester . |
L_THEO | Categorical | Lecturer with whom the student took the theory part of FCYP. |
L_LAB | Categorical | Lecturer with whom the student took the laboratory part of FCYP. |
CLASS | Dichotomous | Indicates if a student passed the theory component of FCYP. |
Variable 1 | Variable 2 | Test | Statistic | p |
---|---|---|---|---|
ALG | PROG | Kruskal–Wallis | H = 70.663 | p < 0.001 |
ALG | P_TYPE | Wilcoxon | W = 620,035.500 | p < 0.001 |
ALG | DPT | Kruskal–Wallis | H = 25.887 | p = 0.001 |
MET | PREF | Kruskal–Wallis | H = 39.730 | p < 0.001 |
GPA | FEE_EX | Wilcoxon | W = 529,103.500 | p < 0.001 |
ALG | SCHOOL | Kruskal–Wallis | H = 19.481 | p < 0.001 |
GPA | QUINTILE | Kruskal–Wallis | H = 21.502 | p < 0.001 |
CAL | L_THEO | Kruskal–Wallis | H = 68.523 | p < 0.001 |
ALG | L_LAB | Kruskal–Wallis | H = 84.681 | p < 0.001 |
GPA | PREC_HDI | Correlation | t = 2738 | p = 0.006 |
Variable | Type | Description |
---|---|---|
P_TYPE | Dichotomous | Indicates the programme duration. |
DPT | Categorical | Indicates which department manages the programme. |
PSU_SCI | Integer | Score in Science admission test. |
PSU_LAN | Integer | Score in Language admission test. |
PSU_MAT | Integer | Score in Maths admission test. |
PSU_GPA | Integer | Normalised high-school grade point average. |
PSU_AVG | Integer | Weighted average admission score. |
CAL | Real | Final grade after taking Calculus I for the first time [1.0, 7.0]. |
PHY | Real | Final grade after taking Physics I for the first time [1.0, 7.0]. |
ALG | Real | Final grade after taking Algebra I for the first time [1.0, 7.0]. |
MET | Real | Final grade after taking Study Methods for the first time [1.0, 7.0]. |
INT | Real | Final grade after taking Introduction to Engineering for the first time [1.0, 7.0]. |
GPA | Real | Grade point average after the first semester [1.0, 7.0]. |
CLASS | Dichotomous | Indicates if a student passed the theory component of FCYP. |
Subset | n | % Pass | % Fail | Imbalance | ||
---|---|---|---|---|---|---|
All | 2372 | 1191 | 1181 | 50.21% | 49.79% | No |
6-year programmes | 1726 | 893 | 833 | 51.74% | 48.26% | No |
4-year programmes | 646 | 298 | 348 | 46.13% | 53.87% | No |
Electricity | 372 | 149 | 223 | 40.05% | 59.95% | Yes |
Geography | 172 | 67 | 105 | 38.95% | 61.05% | Yes |
Industry | 444 | 267 | 177 | 60.14% | 39.86% | Yes |
Informatics | 248 | 163 | 85 | 65.73% | 34.27% | Yes |
Mechanics | 330 | 190 | 140 | 57.58% | 42.42% | Yes |
Metallurgy | 126 | 56 | 70 | 44.44% | 55.56% | No |
Mining | 260 | 103 | 157 | 39.62% | 60.38% | Yes |
Civil | 180 | 71 | 109 | 39.44% | 60.56% | Yes |
Chemistry | 240 | 125 | 115 | 52.08% | 47.92% | No |
Subset | Selected Features |
---|---|
All | GPA, PSU_LAN |
6-year programmes | PSU_LAN, INT, ALG, GPA, PSU_MAT, CAL, PHY, PSU_SCI |
4-year programmes | GPA, PHY, PSU_MAT, ALG, PSU_SCI, CAL, MET, PSU_AVG, PSU_GPA |
Electricity | GPA, CAL |
Geography | CAL |
Industry | GPA, CAL, PSU_SCI, PSU_LAN, PHY |
Informatics | PSU_LAN, GPA, INT |
Mechanics | GPA, PSU_MAT, PSU_AVG, CAL |
Metallurgy | PHY, GPA, PSU_LAN, CAL |
Mining | GPA |
Civil | GPA, ALG, PHY, CAL, INT |
Chemistry | GPA, CAL, PSU_AVG, PHY, PSU_GPA, INT, ALG, PSU_SCI, PSU_LAN |
Subset | Model | Acc | Sens | Spec | ROC |
---|---|---|---|---|---|
All | RF | 62.24% | 60.84% | 63.65% | 66.98% |
MLR | 67.60% | 70.34% | 64.84% | 73.23% | |
SVM | 67.71% | 63.21% | 72.24% | 73.08% | |
XGB | 66.37% | 64.36% | 68.41% | 72.21% | |
6-year programmes | RF | 66.29% | 66.08% | 66.51% | 71.56% |
MLR | 68.30% | 72.24% | 64.07% | 73.56% | |
SVM | 68.66% | 70.29% | 66.90% | 73.72% | |
XGB | 67.24% | 67.83% | 66.61% | 73.20% | |
4-year programmes | RF | 64.92% | 57.53% | 71.25% | 69.43% |
MLR | 66.79% | 60.44% | 72.22% | 71.57% | |
SVM | 67.50% | 62.00% | 72.22% | 71.43% | |
XGB | 66.08% | 56.16% | 74.58% | 70.27% | |
Electricity | RF | 71.66% | 78.00% | 65.33% | 76.70% |
MLR | 67.37% | 72.15% | 62.58% | 73.52% | |
SVM | 73.98% | 62.62% | 85.34% | 76.18% | |
XGB | 70.32% | 77.28% | 63.35% | 71.22% | |
Geography | RF | 67.23% | 64.13% | 70.42% | 71.61% |
MLR | 66.35% | 66.44% | 66.39% | 72.58% | |
SVM | 68.60% | 61.06% | 76.25% | 72.74% | |
XGB | 66.02% | 53.17% | 78.94% | 71.44% | |
Industry | RF | 73.80% | 69.65% | 77.98% | 83.25% |
MLR | 64.39% | 67.19% | 61.63% | 70.33% | |
SVM | 78.28% | 100.00% | 56.57% | 78.29% | |
XGB | 69.78% | 64.74% | 74.84% | 74.04% | |
Informatics | RF | 82.32% | 77.82% | 86.78% | 90.18% |
MLR | 65.62% | 67.31% | 63.91% | 75.01% | |
SVM | 87.14% | 98.11% | 76.13% | 88.33% | |
XGB | 80.74% | 75.85% | 85.61% | 85.85% | |
Mechanics | RF | 75.05% | 70.11% | 80.00% | 82.85% |
MLR | 67.03% | 70.68% | 63.37% | 75.01% | |
SVM | 72.95% | 74.74% | 71.16% | 77.84% | |
XGB | 73.39% | 68.11% | 78.68% | 79.21% | |
Metallurgy | RF | 66.15% | 53.23% | 76.57% | 70.19% |
MLR | 61.06% | 44.23% | 74.43% | 63.65% | |
SVM | 63.46% | 41.57% | 81.14% | 66.10% | |
XGB | 66.71% | 56.03% | 75.14% | 68.38% | |
Mining | RF | 73.27% | 78.91% | 67.69% | 77.22% |
MLR | 69.68% | 71.23% | 68.19% | 76.70% | |
SVM | 72.31% | 65.58% | 79.07% | 71.76% | |
XGB | 72.95% | 68.19% | 77.74% | 75.68% | |
Civil | RF | 75.87% | 79.68% | 72.06% | 82.89% |
MLR | 66.25% | 64.66% | 67.80% | 71.22% | |
SVM | 78.30% | 56.61% | 100.00% | 77.77% | |
XGB | 70.82% | 76.09% | 65.62% | 71.47% | |
Chemistry | RF | 71.50% | 72.75% | 70.16% | 75.43% |
MLR | 71.33% | 73.38% | 69.12% | 75.91% | |
SVM | 73.64% | 76.37% | 70.67% | 76.93% | |
XGB | 74.84% | 74.58% | 75.14% | 77.21% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Köhler, J.; Hidalgo, L.; Jara, J.L. Predicting Students’ Outcome in an Introductory Programming Course: Leveraging the Student Background. Appl. Sci. 2023, 13, 11994. https://doi.org/10.3390/app132111994
Köhler J, Hidalgo L, Jara JL. Predicting Students’ Outcome in an Introductory Programming Course: Leveraging the Student Background. Applied Sciences. 2023; 13(21):11994. https://doi.org/10.3390/app132111994
Chicago/Turabian StyleKöhler, Jacqueline, Luciano Hidalgo, and José Luis Jara. 2023. "Predicting Students’ Outcome in an Introductory Programming Course: Leveraging the Student Background" Applied Sciences 13, no. 21: 11994. https://doi.org/10.3390/app132111994
APA StyleKöhler, J., Hidalgo, L., & Jara, J. L. (2023). Predicting Students’ Outcome in an Introductory Programming Course: Leveraging the Student Background. Applied Sciences, 13(21), 11994. https://doi.org/10.3390/app132111994