Credit Risk Prediction Using Machine Learning and Deep Learning: A Study on Credit Card Customers
Abstract
:1. Introduction
A lucrative credit card client portfolio demands effective credit card customer segmentation. It serves as a risk management tool and enables the company to provide clients with appropriate offers and gain a better understanding of their credit card portfolios. Credit card users can be classified into five categories based on how they use their cards: max or full payers, revolvers, non-payers, traders, and non-users. According to Sajumo (Sajumon 2015), among these five categories, revolvers are the key clients for businesses, as credit card corporations can make profits at their expense. They only make the minimum payment required, if any, and continue using their cards as usual to make purchases. They are not influenced by high-interest rates. Any credit card provider would prefer to avoid having non-payers as customers. They obtain all accessible credit cards, utilize all available credit on those cards, but fail to make any payments.
1.1. Aims and Objectives
- Determine the most essential characteristics that can be used to anticipate the defaulting status.
- Implement balancing techniques to enhance the appropriateness of the data for identifying and examining credit card data.
- Investigate various machine-learning and deep-learning methods and use credit card data as a predictive base.
- Identify and analyze the performance metrics that are the most appropriate for measuring the classification problem.
- Assess the robustness of the selected models over time to ensure sustained accuracy and reliability in varying economic conditions.
1.2. Research Contributions
- Performing data analysis and visualization to gain insights, as well as summarizing significant data features to achieve a better understanding of the dataset. The approved and disbursed loan data from a given time period were studied, and the performance window was then selected and used to predict the performance of future periods (Han et al. 2012).
- Among six machine-learning algorithms, XGBoost was identified as the best-performing model and was chosen as the final model. Based on credit card customer information, a model was developed to classify good and bad customers. Refer to Section 5.3: Summary of Performance Metrics (Chen and Guestrin 2016).
- To understand the relationship between the dependent variables, a correlation matrix was employed, and feature importance was used to identify the critical features that determine the classifier’s performance. Age, income, employment duration, and the number of family members are the primary predictors for the best-performing XGBoost model. Refer to Section 5.2: Feature Importance of the Best Performing Model XGBoost (Lundberg and Lee 2017).
- Evaluating the robustness of the models over time to ensure that accuracy remains consistent across different datasets and economic conditions. This ensures that the models are not only accurate on the initial dataset but also maintain performance as new data become available (Krawczyk 2016).
2. Related Literature
2.1. Importance of Credit Risk Analysis
2.2. Methodologies for Credit Risk Analysis
2.3. Reinforcement Learning in Finance
2.4. Optimization Techniques
3. Data Analysis
3.1. Dataset Description and Methodology Overview
3.2. Exploratory Data Analysis (EDA)
3.2.1. Performance Window and Target Variable Creation
3.2.2. Target Distribution
3.2.3. Handling Outliers
3.2.4. Data Visualization
- The second graph shows that people with less experience typically pose a greater risk than those with more experience. When issuing a card, banks and other financial organizations should consider experience as one of the factors.
- As per the third graph, the red square refers to the average income, showing that the average income of the bad customers is below the average income of the good customers.
- It is evident from the pie charts shown in Figure 10 below that good customers have a higher proportion of property ownership compared to bad customers.
- The income total shows the highest distribution between 50,000 and 275,000.
- Age shows the highest distribution between 300 months and 600 months, which is 25 to 50 years.
- Employment months have the highest distribution between 0 and 50 months, less than 1 year to 4 years.
3.3. Data Pre-Processing
4. Machine-Learning Algorithms
4.1. Random Forest
Algorithm 1. Random Forest Algorithm |
Precondition: A training set , features , and number of trees in forest .
|
- = the probability of picking the data point with the class i;
- c = number of classes.
4.2. Logistic Regression
Algorithm 2. Logistic Regression |
Logistic Regression
|
Algorithm 3. Sigmoid Function |
Sigmoid Function Require: Training data , number of epochs , learning rate η, standard deviation σ Ensure: Weights
|
- x = input value;
- y = predicted output;
- = bias or intercept term;
- = coefficient for input (x).
4.3. Neural Network
- W(t) = vector of the weights at iteration step t;
- ∇E(t) = current gradient of the error function;
- E = sum of the squared errors;
- μ = learning rate.
- ΔW(t) = current adjustment of the weights;
- ΔW(t − 1) = previous change to the weights;
- β = momentum.
4.4. XGBoost Algorithm
- αi—regularization parameters;
- r—residuals computed with the ith tree;
- hi—function trained to predict residuals;
- ri—using X for the ith tree.
- Residuals and Arg(minα) have to be computed in order to compute the α.
- L(Y, F(X)) is the differentiable loss function.
4.5. LightGBM Algorithm
4.6. AdaBoost Algorithm
Algorithm 4. AdaBoost |
AdaBoost Algorithm Given: . Initialize: For :
|
- Good generalization skills: Creating compositions for real-world issues of a higher caliber is more feasible than using the fundamental algorithms. As the number of fundamental algorithms rises, the generalization ability may become more effective (in some missions).
- Own boosting expenses are minimal: The training time of the fundamental algorithms nearly entirely determines the amount of time needed to construct the final image.
- Several drawbacks of AdaBoost can be discussed as follows:
- Boosting technology develops gradually: It is crucial to guarantee high-quality data.
- Tactful to unorganized data and outliers: Hence, it is intensely advised to avoid these before using AdaBoost.
- Slower than XGBoost: It has also been indicated that AdaBoost is slower than XGBoost.
5. Implementation and Results
5.1. Confusion Matrix
- True positive—When good customers are accurately predicted as good customers.
- False positive—When bad customers are improperly predicted as good customers.
- True negative—When bad customers are accurately predicted are bad customers.
- False negative—When good customers are improperly predicted as bad customers.
- True positive = Number of customers accurately predicted as good/Actual number of good customers
- False positive = Number of customers improperly predicted as good/Actual number of bad customers
- True negative = Number of customers accurately predicted as bad/Actual number of bad customers
- False negative = Number of customers improperly predicted as bad/Actual number of good customers
5.2. Implementation and Comparison of Machine-Learning Algorithms
5.2.1. Accuracy
5.2.2. Recall and Precision
5.2.3. F1 Score
5.2.4. ROC Curve and AUC Score
- Scale is unimportant to AUC: It evaluates how well predictions are ranked rather than the absolute values of the predictions.
- AUC is independent of the classification threshold: It evaluates how well the model predicts regardless of the classification threshold.
5.2.5. MCC
5.3. Comparison of the Best- and Worst-Performing Models
6. Further Discussion
7. Conclusions
7.1. Summary
7.2. Recommendations
7.3. Limitations of the Study and Future Work
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A
References
- Adha, Risfian, Siti Nurrohmah, and Sarini Abdullah. 2018. Multinomial Logistic Regression and Spline Regression for Credit Risk Modelling. Journal of Physics: Conference Series 1108: 012019. [Google Scholar] [CrossRef]
- Ali, Mohsin, Abdul Razaque, Joon Yoo, Uskenbayeva R. Kabievna, Aiman Moldagulova, Satybaldiyeva Ryskhan, Kalpeyeva Zhuldyz, and Aizhan Kassymova. 2024. Designing an Intelligent Scoring System for Crediting Manufacturers and Importers of Goods in Industry 4.0. Logistics 8: 33. [Google Scholar] [CrossRef]
- Alonso Robisco, Andrés, and José Manuel Carbó Martínez. 2022. Measuring the model risk-adjusted performance of machine learning algorithms in credit default prediction. Financial Innovation 8: 70. [Google Scholar] [CrossRef]
- Al-qerem, Ahmad, Ghazi Al-Naymat, and Mays Alhasan. 2019. Loan Default Prediction Model Improvement through Comprehensive Preprocessing and Features Selection. Paper presented at the 2019 International Arab Conference on Information Technology (ACIT), Al Ain, United Arab Emirates, December 3–5. [Google Scholar]
- Aswini, Ravichandran, Balamurugan Muruganantham, Santosh Kumar, and Arumugam Murugan. 2020. Exploratory Data Analysis for Social Big Data Using Regression and Recurrent Neural Networks. Webology 17: 922–36. [Google Scholar] [CrossRef]
- Bao, Wang, Ning Lianju, and Kong Yue. 2019. Integration of unsupervised and supervised machine learning algorithms for credit risk assessment. Expert Systems with Applications 128: 301–15. [Google Scholar] [CrossRef]
- Behera, Sudersan, Sarat Chandra Nayak, and Annadanapu Pavan Kumar. 2023. A Comprehensive Survey on Higher Order Neural Networks and Evolutionary Optimization Learning Algorithms in Financial Time Series Forecasting. Archives of Computational Methods in Engineering 30: 4401–48. [Google Scholar] [CrossRef]
- Beheshti, Nima. 2022. Random Forest Classification. Available online: https://towardsdatascience.com/random-forest-classification-678e551462f5 (accessed on 31 October 2024).
- Bindal, Anirudh, and Sandeep Chaurasia. 2018. Predictive Risk Analysis For Loan Repayment of Credit Card Clients. Paper presented at the 2018 3rd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, May 18–19. [Google Scholar]
- Buchanan, Bonnie G., and Danika Wright. 2021. The impact of machine learning on UK financial services. Oxford Review of Economic Policy 37: 537–63. [Google Scholar] [CrossRef]
- Chang, Victor, Raul Valverde, Muthu Ramachandran, and Chung-Sheng Li. 2020. Toward business integrity modeling and analysis framework for risk measurement and analysis. Applied Sciences 10: 3145. [Google Scholar] [CrossRef]
- Chen, Tianqi, and Carlos Guestrin. 2016. XGBoost: A Scalable Tree Boosting System. Paper presented at the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13–17; pp. 785–94. [Google Scholar]
- Chen, Xihui, Xuyuan You, and Victor Chang. 2021. FinTech and commercial banks’ performance in China: A leap forward or survival of the fittest? Technological Forecasting and Social Change 166: 120645. [Google Scholar] [CrossRef]
- Dm, Obare, and Muraya Mm. 2018. Comparison of Accuracy of Support Vector Machine Model and Logistic Regression Model in Predicting Individual Loan Defaults. American Journal of Applied Mathematics and Statistics 6: 266–71. [Google Scholar]
- Donges, Niklas. 2021. A Complete Guide to the Random Forest Algorithm. Available online: https://builtin.com/data-science/random-forest-algorithm (accessed on 31 October 2024).
- Duan, Jing. 2019. Financial system modeling using deep neural networks (DNNs) for effective risk assessment and prediction. Journal of the Franklin Institute 356: 4716–31. [Google Scholar] [CrossRef]
- Education, IBM Cloud. 2020. What Is Exploratory Data Analysis (EDA)? Available online: https://www.ibm.com/topics/exploratory-data-analysis (accessed on 31 October 2024).
- GeeksforGeeks. n.d. LightGBM (Light Gradient Boosting Machine). Available online: https://www.geeksforgeeks.org/lightgbm-light-gradient-boosting-machine/ (accessed on 31 October 2024).
- Guégan, Dominique, and Bertrand Hassani. 2018. Regulatory learning: How to supervise machine learning models? An application to credit scoring. The Journal of Finance and Data Science 4: 157–71. [Google Scholar] [CrossRef]
- Han, Jiawei, Micheline Kamber, and Jian Pei. 2012. Data Mining Concepts and Techniques, 3rd ed. Waltham: Morgan Kaufmann Publishers. [Google Scholar]
- Kharwal, Aman. 2020. Boosting Algorithms in Machine Learning. Available online: https://thecleverprogrammer.com/2020/10/30/boosting-algorithms-in-machine-learning/ (accessed on 31 October 2024).
- Krawczyk, Bartosz. 2016. Learning from imbalanced data: Open challenges and future directions. Progress in Artificial Intelligence 5: 221–32. [Google Scholar]
- Lin, Cian, Chih-Fong Tsai, and Wei-Chao Lin. 2023. Towards hybrid over- and under-sampling combination methods for class imbalanced datasets: An experimental study. Artificial Intelligence Review 56: 845–63. [Google Scholar] [CrossRef]
- Liu, Lulu. 2022. A Self-Learning BP Neural Network Assessment Algorithm for Credit Risk of Commercial Bank. Wireless Communications and Mobile Computing 2022: 9650934. [Google Scholar] [CrossRef]
- Lucarelli, Giorgio, and Matteo Borrotti. 2020. A deep Q-learning portfolio management framework for the cryptocurrency market. Neural Computing and Applications 32: 17229–44. [Google Scholar] [CrossRef]
- Lundberg, Scott M., and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems. Long Beach: Curran Associates Inc., pp. 4768–77. [Google Scholar]
- Ma, Xiaojun, Jinglan Sha, Dehua Wang, Yuanbo Yu, Qian Yang, and Xueqi Niu. 2018. Study on a prediction of P2P network loan default based on the machine learning LightGBM and XGboost algorithms according to different high dimensional data cleaning. Electronic Commerce Research and Applications 31: 24–39. [Google Scholar] [CrossRef]
- Maldonado, Sebastián, Juan Pérez, and Cristián Bravo. 2017. Cost-based feature selection for Support Vector Machines: An application in credit scoring. European Journal of Operational Research 261: 656–65. [Google Scholar] [CrossRef]
- Malibari, Nadeem, Iyad Katib, and Rashid Mehmood. 2023. Systematic Review on Reinforcement Learning in the Field of Fintech. Available online: https://arxiv.org/pdf/2305.07466 (accessed on 31 October 2024).
- Meltzer, Rachel. 2023. What Is Random Forest? Available online: https://careerfoundry.com/en/blog/data-analytics/what-is-random-forest/ (accessed on 31 October 2024).
- Naik, K. S. 2021. Predicting Credit Risk for Unsecured Lending: A Machine Learning Approach. arXiv arXiv:2110.02206. [Google Scholar]
- Nazarenko, E., V. Varkentin, and T. Polyakova. 2019. Features of Application of Machine Learning Methods for Classification of Network Traffic (Features, Advantages, Disadvantages). Paper presented at the 2019 International Multi-Conference on Industrial Engineering and Modern Technologies (FarEastCon), Vladivostok, Russia, October 1–4. [Google Scholar]
- Sajumon, Akshatha. 2015. 5 Types of Credit Card Users You Definitely Fall Under. Available online: https://blog.bankbazaar.com/5-types-of-credit-card-users-you-definitely-fall-under/ (accessed on 31 October 2024).
- Sariannidis, Nikolaos, Stelios Papadakis, Alexandros Garefalakis, Christos Lemonakis, and Tsioptsia Kyriaki-Argyro. 2020. Default avoidance on credit card portfolios using accounting, demographical and exploratory factors: Decision making based on machine learning (ML) techniques. Annals of Operations Research 294: 715–39. [Google Scholar] [CrossRef]
- Sayjadah, Yashna, Ibrahim Abaker Targio Hashem, Faiz Alotaib, and Khairl Azhar Kasmiran. 2018. Credit Card Default Prediction using Machine Learning Techniques. Paper presented at the 2018 Fourth International Conference on Advances in Computing, Communication & Automation (ICACCA), Subang Jaya, Malaysia, October 26–28. [Google Scholar]
- Sharma, Pranshu. 2023. Beginner’s Guide To Decision Tree Classification Using Python. Available online: https://www.analyticsvidhya.com/blog/2021/04/beginners-guide-to-decision-tree-classification-using-python/ (accessed on 31 October 2024).
- Shi, Si, Rita Tse, Wuman Luo, Stefano D’Addona, and Giovanni Pau. 2022. Machine learning-driven credit risk: A systemic review. Neural Computing and Applications 34: 14327–39. [Google Scholar] [CrossRef]
- Sumiea, Ebrahim Hamid, Said Jadid Abdulkadir, Hitham Seddig Alhussian, Safwan Mahmood Al-Selwi, Alawi Alqushaibi, Mohammed Gamal Ragab, and Suliman Mohamed Fati. 2024. Deep deterministic policy gradient algorithm: A systematic review. Heliyon 10: e30697. [Google Scholar] [CrossRef] [PubMed]
- Sun, Ting, and Miklos A. Vasarhelyi. 2018. Predicting credit card delinquencies: An application of deep neural networks. Intelligent Systems in Accounting, Finance and Management 25: 174–89. [Google Scholar] [CrossRef]
- Tian, Zhenya, Jialiang Xiao, Haonan Feng, and Yutian Wei. 2020. Credit Risk Assessment based on Gradient Boosting Decision Tree. Procedia Computer Science 174: 150–60. [Google Scholar] [CrossRef]
- Ullah, Mohammad Aman, Mohammad Manjur Alam, Shamima Sultana, and Rehana Sultana Toma. 2018. Predicting Default Payment of Credit Card Users: Applying Data Mining Techniques. Paper presented at the 2018 International Conference on Innovations in Science, Engineering and Technology (ICISET), Chittagong, Bangladesh, October 27–28. [Google Scholar]
- Wang, Kui, Meixuan Li, Jingyi Cheng, Xiaomeng Zhou, and Gang Li. 2022. Research on personal credit risk evaluation based on XGBoost. Procedia Computer Science 199: 1128–35. [Google Scholar] [CrossRef]
- Xia, Yufei, Chuanzhe Liu, Bowen Da, and Fangming Xie. 2018. A novel heterogeneous ensemble credit scoring model based on bstacking approach. Expert Systems with Applications 93: 182–99. [Google Scholar] [CrossRef]
- Xu, Che, and Shuwen Zhang. 2024. A Genetic Algorithm-based sequential instance selection framework for ensemble learning. Expert Systems with Applications 236: 121269. [Google Scholar] [CrossRef]
- Zhu, Lin, Dafeng Qiu, Daji Ergu, Cai Ying, and Kuiyi Liu. 2019. A study on predicting loan default based on the random forest algorithm. Procedia Computer Science 162: 503–13. [Google Scholar] [CrossRef]
Study | Model | Accuracy |
---|---|---|
Zhu et al. (2019) | Random forest | 98.00% |
Sayjadah et al. (2018) | Random forest | 82.00% |
Tian et al. (2020) | Gradient boosting decision tree | 92.19% |
Sun and Vasarhelyi (2018) | Neural network | 99.50% |
Duan (2019) | Neural network backpropagation (3 hidden layers) | 93.00% |
Wang et al. (2022) | XGBoost | 87.10% |
Naik (2021) | LGBM | 95.50% |
Adha et al. (2018) | Logistic regression | 95.30% |
Ullah et al. (2018) | AdaBoost | 88.00% |
Dm and Mm (2018) | SVM | 86.12% |
Methodology | Hyperparameter | Value |
---|---|---|
Random Forest | n_estimators | 250 |
min_samples_leaf | 16 | |
maximum depth | 12 | |
random State | 42 | |
XGBoost | max_depth | 12 |
n_estimators | 250 | |
min_child_weight | 8 | |
subsample | 0.8 | |
learning_rate | 0.02 | |
seed | 60 | |
gamma | 0 | |
colsample_bytree | 0.8 | |
objective | binary: logistic | |
LGBM | num_leaves | 50 |
learning_rate | 0.02 | |
n_estimators | 250 | |
subsample | 0.8 | |
colsample_bytree | 0.8 | |
AdaBoost | max_depth | 1 |
n_estimators | 100 | |
learning_rate | 1 | |
Neural Network | hidden_layer_sizes | 400,800 |
max_iter | 1000 | |
random_state | 25 | |
shuffle | TRUE | |
Logistic Regression | random_state | 42 |
Algorithm | Accuracy | Precision | Recall | F1 Score | ROC–AUC | MCC |
---|---|---|---|---|---|---|
Random Forest | 97.9% | 0.979 | 0.979 | 0.979 | 0.996 | 0.958 |
Neural Network | 87.2% | 0.872 | 0.872 | 0.872 | 0.942 | 0.744 |
XGB | 99.3% | 0.993 | 0.993 | 0.993 | 0.997 | 0.986 |
LGBM | 99.2% | 0.992 | 0.992 | 0.993 | 0.997 | 0.744 |
AdaBoost | 0.920 | 0.925 | 0.920 | 0.920 | 0.976 | 0.845 |
Logistic Regression | 0.843 | 0.846 | 0.843 | 0.843 | 0.910 | 0.690 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chang, V.; Sivakulasingam, S.; Wang, H.; Wong, S.T.; Ganatra, M.A.; Luo, J. Credit Risk Prediction Using Machine Learning and Deep Learning: A Study on Credit Card Customers. Risks 2024, 12, 174. https://doi.org/10.3390/risks12110174
Chang V, Sivakulasingam S, Wang H, Wong ST, Ganatra MA, Luo J. Credit Risk Prediction Using Machine Learning and Deep Learning: A Study on Credit Card Customers. Risks. 2024; 12(11):174. https://doi.org/10.3390/risks12110174
Chicago/Turabian StyleChang, Victor, Sharuga Sivakulasingam, Hai Wang, Siu Tung Wong, Meghana Ashok Ganatra, and Jiabin Luo. 2024. "Credit Risk Prediction Using Machine Learning and Deep Learning: A Study on Credit Card Customers" Risks 12, no. 11: 174. https://doi.org/10.3390/risks12110174
APA StyleChang, V., Sivakulasingam, S., Wang, H., Wong, S. T., Ganatra, M. A., & Luo, J. (2024). Credit Risk Prediction Using Machine Learning and Deep Learning: A Study on Credit Card Customers. Risks, 12(11), 174. https://doi.org/10.3390/risks12110174