Evaluation of Cost-Sensitive Learning Models in Forecasting Business Failure of Capital Market Firms
Abstract
:1. Introduction
2. Literature Review
2.1. Imbalance Datasets Solution
2.2. Imbalanced Learning in Finance
3. Methods
3.1. CorrOV-CSEn
3.1.1. Correlation-Based Oversampling
- represents the Linear Covariance Matrix (LCM);
- is a minority class instance;
- denotes the k-nearest neighbors (K-NN) of ;
- is the matrix of K-NN instances of ;
- is the centroid of the Y matrix.
- Oversampling rate determination: Higher LCM values, particularly among the K-NN of the same minority class, indicate stronger correlation and guide a higher oversampling rate. This strategy reduces variance and generates synthetic instances in regions with higher minority class correlations, especially near borderline instances.
- Oversampling region optimization: For each minority instance, oversampling is performed only if its LCM with respect to the K-NN of the same class label is greater than its LCM with instances from other classes. This ensures that synthetic data are generated in the most relevant regions, enhancing both model robustness and the quality of the generated samples.
3.1.2. Cost-Sensitive Ensemble Learning
Multi-Layer Perceptron (MLP)
- An MLP consists of an input layer, one or more hidden layers, and an output layer. Each layer is composed of several neurons (nodes). If the dataset contains features, the input layer will have neurons. The number of neurons in the hidden layers can be chosen based on the complexity of the task. Each neuron applies a weighted sum of inputs followed by a non-linear activation function such as ReLU or sigmoid. Mathematically, the output of a neuron can be expressed as
- During forward propagation, inputs pass through the network from the input layer to the output layer. Each hidden layer neuron processes the weighted sum of inputs and applies the activation function. The final output layer provides predictions, which can be either Classification or Regression.
- The loss function quantifies the error between the predicted output and the actual target. For regression, the Mean Squared Error (MSE) is often used.
- Backpropagation and Weight Update: The gradient of the loss function is calculated using the chain rule, and weights are updated using gradient descent.
3.2. Random Forest
- Bootstrap Sampling: For each of the T trees in the forest, a random subset of the data is drawn with replacement. If there are N total samples, then each tree is built from a subset of N samples drawn randomly with replacement, resulting in different training sets for each tree:
- Feature Selection: At each node of the decision tree, a random subset of features is chosen, typically equal to the square root of the total number of features in classification tasks (i.e., ). This helps reduce the correlation between trees and improve model variance. For regression, the number of selected features is often . This features minimizes correlations among the trees [60].
- Splitting Criterion: From the selected subset of features at each node, the feature that best splits the data is chosen using a splitting criterion, often the Gini index or entropy. For example, the Gini index for a split can be calculated as
- Building the Forest: Each tree is grown to its full depth without pruning, resulting in a collection of deep, unpruned trees. By default, 500 trees are built, though this number can be adjusted for specific applications.
- Prediction Aggregation: For classification tasks, the final prediction for each data point is determined by majority voting across all trees. Let represent the prediction of the tree for a data point . Then, the final prediction is given by
3.3. Gradient Boosted Trees
- Initialize the model: Start with an initial estimate, often taken as the mean value of the target variable for regression tasks or a single weak classifier for classification.
- Iterative Model Updates: For each iteration :
- Compute the Residuals: Calculate the residuals for each sample based on the current model :
- Fit a New Tree: Train a new decision tree to predict the residuals .
- Update the Model: Add the new tree to the model with a learning rate (to control the contribution of each tree), yielding an updated model:
- Final Prediction: After iterations, the final model is an ensemble of the trees, each adjusted to reduce the error from prior steps. For regression, the final prediction is
3.4. XGBoost
- Objective Function: The objective of XGBoost is to minimize a regularized loss function that combines the traditional loss function with regularization terms for complexity control. For trees, the objective function is defined as
- is the loss function, such as mean squared error for regression or log-loss for classification;
- is the regularization term with parameters and , controlling the complexity of each tree.
- Tree Structure and Growth: Each tree in XGBoost is built to minimize the residuals from the previous trees, following the same general structure as GBT. However, XGBoost introduces a tree-pruning technique, where trees are pruned based on their impact on the objective function rather than growing to full depth. The parameter controls the maximum depth of each tree, preventing the model from overfitting by limiting tree complexity.
- Update Process: In each iteration, the algorithm calculates the best tree structure to minimize the residuals of the previous ensemble. The updates are computed using second-order gradients (Hessian) of the loss function, making it more efficient. The model update at each step is given by
- Handling Missing Data: XGBoost automatically manages missing data by learning optimal paths for instances with missing values during training. It assigns missing values to the most suitable branch, improving model accuracy when dealing with incomplete datasets.
- Final Prediction: The final prediction is an aggregation of all trees, represented as
3.5. AdaBoost
- Initialize sample weights: Each sample i in the training set receives an initial weight:
- Train a weak classifier: In each round , a weak classifier is trained on the weighted samples, and its error rate is calculated as
- Calculate the classifier’s weight: The weight of the weak classifier is determined based on its accuracy:
- Update sample weights: Sample weights are updated to reflect the classifier’s performance, giving more weight to misclassified samples:
- Combine weak classifiers: The final strong classifier is a weighted sum of all weak classifiers:
3.6. CatBoost
- Ordered Boosting to Avoid Target Leakage: In standard GBT, future data points might unintentionally influence earlier predictions, leading to target leakage. Ordered boosting solves this by using a permutation-based scheme, ensuring that only past information influences each iteration. This ordered approach is particularly useful in datasets where feature-target relationships are complex and dynamic, and it enhances CatBoost’s accuracy.
- Handling of Categorical Variables: CatBoost automatically handles categorical features without requiring extensive preprocessing. It converts categorical features into numeric representations through a series of random permutations, using them to guide the splitting criteria for each decision tree.
- Objective Function: CatBoost minimizes a regularized loss function similar to other boosting methods, but with an emphasis on ordered boosting:
- is the loss function (e.g., cross-entropy or log-loss for classification tasks);
- is the regularization term for tree complexity, helping to control overfitting.
- Tree Structure and Decision Rule: CatBoost uses binary decision trees as base learners. For each input , the decision tree assigns it to one of the leaf regions based on a series of splits. The function for each tree can be represented as
- represents the decision function for each sample ;
- is the disjoint region corresponding to each leaf in the tree;
- is the predicted output value for region .
- Final Prediction: The final prediction is the aggregation of all the trees in the ensemble. For a dataset with trees, the final output is given by
- Regularization and Overfitting Prevention: CatBoost uses random permutations when selecting tree splits, which reduces overfitting and enhances model generalization. This, combined with ordered boosting, allows CatBoost to outperform traditional GBT methods on many complex tasks.
3.7. Business Failure
3.8. Evaluating Methods
3.9. Statistical Significance Test
- is the number of data sets (blocks);
- is the number of models (groups);
- is the sum of ranks for each model.
4. Case Study
5. Experimental Discussion
5.1. Evaluation Among Models
- ❖
- CorrOV-CSEn Results:
- •
- Multi-Layer Perceptron (MLP) shows good sensitivity (0.84). However, it struggles with precision (0.33), meaning a relatively small proportion of the predicted failure cases are actual failures. This imbalance results in a moderate F1 score of 0.47.
- •
- Random forest demonstrates strong sensitivity (0.89), meaning it effectively detects failure cases. However, it struggles with precision (0.38), indicating that only a relatively small portion of the firms predicted as failures are actually failures. This results in a moderate F1 score of (0.53). On the other hand, when using SMOTE, it records (0.80) for sensitivity and loses much of its success rate for identifying default firms. However, precision got better ((0.60) and (0.69)).
- •
- Gradient boosting offers balanced performance, with a sensitivity of (0.80) and higher precision (0.44), resulting in an F1 score of (0.57). This indicates better overall handling of both false positives and false negatives.
- •
- XGBoost performs similarly to gradient boosting, with the same sensitivity (0.80) but slightly lower precision (0.39), resulting in an F1 score of (0.53). While still robust, it is slightly outperformed by gradient boosting in terms of precision.
- •
- AdaBoost has the lowest sensitivity (0.75) but the highest precision (0.48), resulting in a competitive F1 score of (0.58). This indicates that while its failure predictions are more accurate, it misses some failure cases.
- •
- CatBoost exhibits the highest sensitivity (0.91) but struggles the most with precision (0.20), leading to the weakest F1 score (0.33). This suggests that while CatBoost is highly effective at detecting failures, which is our primary objective, it produces more false positives.
- ❖
- SMOTE Results:
- •
- Multi-Layer Perceptron (MLP) maintains a similar performance pattern. Sensitivity remains high at 0.84, effectively capturing failure cases, while precision stays relatively low at 0.33, indicating that many predicted failure cases were not actual failures.
- •
- Random forest sensitivity drops to 0.80 while precision improves to 0.60, leading to an F1 score of 0.69. However, the sensitivity reduction indicates some missed failure cases.
- •
- Gradient boosting shows lower sensitivity (0.73) with a slight precision increase (0.60), resulting in an F1 score of 0.66, suggesting a modest trade-off.
- •
- XGBoost sees a minor decrease in sensitivity (0.77) and an increase in precision (0.55), with an F1 score of 0.65.
- •
- AdaBoost under SMOTE shows a significant drop in sensitivity (0.57) with minimal gain in precision (0.56), reducing its F1 score to 0.56.
- •
- CatBoost improves precision (0.72) but its sensitivity remains lower than CorrOV-CSEn, with an F1 score of 0.73, showing more balanced results but still lower sensitivity.
5.2. Significance Differences
- AdaBoost vs. CatBoost: This is the only comparison with a significant difference (p-value = 0.030), showing that CatBoost performs significantly better than AdaBoost in terms of sensitivity.
- Gradient boosting vs. CatBoost, AdaBoost vs. CatBoost, and MLP vs. CatBoost: All comparisons show significant differences with p-values of 0.030, indicating that CatBoost has significantly lower precision compared to gradient boosting, AdaBoost, and MLP.
5.3. Feature Importance
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Usmani, S.; Shamsi, J.A. LSTM based stock prediction using weighted and categorized financial news. PLoS ONE 2023, 18, e0282234. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Z.; Liu, X.; Niu, H. Financial crisis early warning of Chinese listed companies based on MD&A text-linguistic feature indicators. PLoS ONE 2023, 18, e0291818. [Google Scholar] [CrossRef]
- Jezeie, F.V.; Sadjadi, S.J.; Makui, A. Constrained portfolio optimization with discrete variables: An algorithmic method based on dynamic programming. PLoS ONE 2022, 17, e0271811. [Google Scholar] [CrossRef] [PubMed]
- Bi, W.; Zhang, Q. Forecasting mergers and acquisitions failure based on partial-sigmoid neural network and feature selec-tion. PLoS ONE 2021, 16, e0259575. [Google Scholar] [CrossRef]
- Li, M. Financial investment risk prediction under the application of information interaction Firefly Algorithm combined with Graph Convolutional Network. PLoS ONE 2023, 18, e0291510. [Google Scholar] [CrossRef]
- Dahal, K.R.; Pokhrel, N.R.; Gaire, S.; Mahatara, S.; Joshi, R.P.; Gupta, A.; Banjade, H.R.; Joshi, J. A comparative study on effect of news sentiment on stock price prediction with deep learning architecture. PLoS ONE 2023, 18, e0284695. [Google Scholar] [CrossRef] [PubMed]
- Javid, I.; Ghazali, R.; Syed, I.; Zulqarnain, M.; Husaini, N.A. Study on the Pakistan stock market using a new stock crisis prediction method. PLoS ONE 2022, 17, e0275022. [Google Scholar] [CrossRef]
- Cui, Y.; Liu, L. Investor sentiment-aware prediction model for P2P lending indicators based on LSTM. PLoS ONE 2022, 17, e0262539. [Google Scholar] [CrossRef]
- Zhu, C.; Liu, X.; Chen, D. Prediction of digital transformation of manufacturing industry based on interpretable machine learning. PLoS ONE 2024, 19, e0299147. [Google Scholar] [CrossRef] [PubMed]
- Khan, A.H.; Shah, A.; Ali, A.; Shahid, R.; Zahid, Z.U.; Sharif, M.U.; Jan, T.; Zafar, M.H. A performance comparison of machine learning models for stock market prediction with novel investment strategy. PLoS ONE 2023, 18, e0286362. [Google Scholar] [CrossRef] [PubMed]
- Wei, X.; Ouyang, H.; Liu, M. Stock index trend prediction based on TabNet feature selection and long short-term memory. PLoS ONE 2022, 17, e0269195. [Google Scholar] [CrossRef] [PubMed]
- Tran, T.; Nguyen, N.H.; Le, B.T.; Vu, N.T.; Vo, D.H. Examining financial distress of the Vietnamese listed firms using accounting-based models. PLoS ONE 2023, 18, e0284451. [Google Scholar] [CrossRef] [PubMed]
- Laghari, F.; Ahmed, F.; López García, M.D.L.N. Cash flow management and its effect on firm performance: Empirical ev-idence on non-financial firms of China. PLoS ONE 2023, 18, e0287135. [Google Scholar] [CrossRef] [PubMed]
- Almustafa, H.; Nguyen, Q.K.; Liu, J.; Dang, V.C. The impact of COVID-19 on firm risk and performance in MENA countries: Does national governance quality matter? PLoS ONE 2023, 18, e0281148. [Google Scholar] [CrossRef] [PubMed]
- Tian, X.; Wang, Y.; Kohar, U.H.A. Capital structure, business model innovation, and firm performance: Evidence from Chinese listed corporate based on system GMM model. PLoS ONE 2024, 19, e0306054. [Google Scholar] [CrossRef]
- Samour, A.; AlGhazali, A.; Gadoiu, M.; Banuta, M. Capital structure and financial performance of China’s energy industry: What can we infer from COVID-19? PLoS ONE 2024, 19, e0300936. [Google Scholar]
- Berloco, C.; Morales, G.D.F.; Frassineti, D.; Greco, G.; Kumarasinghe, H.; Lamieri, M.; Massaro, E.; Miola, A.; Yang, S. Predicting corporate credit risk: Network contagion via trade credit. PLoS ONE 2021, 16, e0250115. [Google Scholar] [CrossRef] [PubMed]
- Hlongwane, R.; Ramaboa, K.K.K.M.; Mongwe, W. Enhancing credit scoring accuracy with a comprehensive evaluation of alternative data. PLoS ONE 2024, 19, e0303566. [Google Scholar] [CrossRef] [PubMed]
- Ma, Z.; Hou, W.; Zhang, D. A credit risk assessment model of borrowers in P2P lending based on BP neural network. PLoS ONE 2021, 16, e0255216. [Google Scholar] [CrossRef] [PubMed]
- Wang, H.; Liu, X. Undersampling bankruptcy prediction: Taiwan bankruptcy data. PLoS ONE 2021, 16, e0254030. [Google Scholar] [CrossRef] [PubMed]
- Japkowicz, N. Learning from imbalanced data sets: A comparison of various strategies. In AAAI Workshop on Learning from Imbalanced Data Sets; AAAI Press: Menlo Park, CA, USA, 2000. [Google Scholar]
- Groccia, M.C.; Guido, R.; Conforti, D.; Pelaia, C.; Armentaro, G.; Toscani, A.F.; Miceli, S.; Succurro, E.; Hribal, M.L.; Sciacqua, A. Cost-Sensitive Models to Predict Risk of Cardiovascular Events in Patients with Chronic Heart Failure. Information 2023, 14, 542. [Google Scholar] [CrossRef]
- Natha, P.; RajaRajeswari, P. Advancing Skin Cancer Prediction Using Ensemble Models. Computers 2024, 13, 157. [Google Scholar] [CrossRef]
- Devi, D.; Biswas, S.K.; Purkayastha, B. Correlation-based Oversampling aided Cost Sensitive Ensemble learning technique for Treatment of Class Imbalance. J. Exp. Theor. Artif. Intell. 2022, 34, 143–174. [Google Scholar] [CrossRef]
- Alloway, B.T.; Weisenthal, J. What’s Been Happening with the Iranian Stock Market; Bloomberg: New York, NY, USA, 2023. [Google Scholar]
- Rawat, S.S.; Mishra, A.K. Review of Methods for Handling Class-Imbalanced in Classification Problems. arXiv 2022, arXiv:2211.05456. [Google Scholar]
- Tomek, I. Two Modifications of CNN. IEEE Trans. Syst. Man Cybern. 1976, 11, 769–772. [Google Scholar]
- Kubat, M.; Matwin, S. Addressing the curse of imbalanced data sets: One-sided sampling. In Proceedings of the Fourteenth International Conference on Machine Learning, Nashville, TN, USA, 8–12 July 1997. [Google Scholar]
- Japkowicz, N. The class imbalance problem: Significance and strategies. In Proceedings of the International Conference on Artificial Intelligence, Las Vegas, NV, USA, 26–29 June 2000. [Google Scholar]
- Laurikkala, J. Improving identification of difficult small classes by balancing class distribution. In Proceedings of the Artificial Intelligence in Medicine: 8th Conference on Artificial Intelligence in Medicine in Europe, AIME 2001, Cascais, Portugal, 1–4 July 2001; Proceedings 8. Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
- Hoyos-Osorio, J.; Alvarez-Meza, A.; Daza-Santacoloma, G.; Orozco-Gutierrez, A.; Castellanos-Dominguez, G. Relevant information undersampling to support imbalanced data classification. Neurocomputing 2021, 436, 136–146. [Google Scholar] [CrossRef]
- Lee, W.; Seo, K. Downsampling for Binary Classification with a Highly Imbalanced Dataset Using Active Learning. Big Data Res. 2022, 28, 100314. [Google Scholar] [CrossRef]
- Laveti, R.N.; Mane, A.A.; Pal, S.N. Dynamic Stacked Ensemble with Entropy based Undersampling for the Detection of Fraudulent Transactions. In Proceedings of the 2021 6th International Conference for Convergence in Technology (I2CT), Maharashtra, India, 2–4 April 2021; pp. 1–7. [Google Scholar]
- Solberg, A.S.; Solberg, R. A large-scale evaluation of features for automatic detection of oil spills in ERS SAR images. In Proceedings of the IGARSS ’96. 1996 International Geoscience and Remote Sensing Symposium, Lincoln, NB, USA, 21–26 May 1996; pp. 1484–1486. [Google Scholar]
- Mathew, J.; Pang, C.K.; Luo, M.; Leong, W.H. Classification of Imbalanced Data by Oversampling in Kernel Space of Support Vector Machines. IEEE Trans. Neural Networks Learn. Syst. 2017, 29, 4065–4076. [Google Scholar] [CrossRef] [PubMed]
- Bennin, K.E.; Keung, J.; Phannachitta, P.; Monden, A.; Mensah, S. MAHAKIL: Diversity Based Oversampling Approach to Alleviate the Class Imbalance Issue in Software Defect Prediction. IEEE Trans. Softw. Eng. 2017, 44, 534–550. [Google Scholar] [CrossRef]
- Cheng, K.; Zhang, C.; Yu, H.; Yang, X.; Zou, H.; Gao, S. Grouped SMOTE With Noise Filtering Mechanism for Classifying Imbalanced Data. IEEE Access 2019, 7, 170668–170681. [Google Scholar] [CrossRef]
- Tarawneh, A.S.; Hassanat, A.B.A.; Almohammadi, K.; Chetverikov, D.; Bellinger, C. SMOTEFUNA: Synthetic Minority Over-Sampling Technique Based on Furthest Neighbour Algorithm. IEEE Access 2020, 8, 59069–59082. [Google Scholar] [CrossRef]
- Almomani, I.; Qaddoura, R.; Habib, M.; Alsoghyer, S.; Al Khayer, A.; Aljarah, I.; Faris, H. Android ransomware detection based on a hybrid evolutionary approach in the context of highly im-balanced data. IEEE Access 2021, 9, 57674–57691. [Google Scholar] [CrossRef]
- Juez-Gil, M.; Arnaiz-González, Á.; Rodríguez, J.J.; López-Nozal, C.; García-Osorio, C. Approx-SMOTE: Fast SMOTE for Big Data on Apache Spark. Neurocomputing 2021, 464, 432–437. [Google Scholar] [CrossRef]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Li, C. Data Mining for Direct Marketing: Problems and Solutions; National Library of Canada= Bibliothèque nationale du Canada: Ottawa, ON, Canada, 2000. [Google Scholar]
- Xu, Z.; Shen, D.; Nie, T.; Kou, Y. A hybrid sampling algorithm combining M-SMOTE and ENN based on Random forest for medical imbalanced data. J. Biomed. Informatics 2020, 107, 103465. [Google Scholar] [CrossRef]
- Dong; Xiao, H.; Dong, Y. SA-CGAN: An oversampling method based on single attribute guided conditional GAN for multi-class imbalanced learning. Neurocomputing 2022, 472, 326–337. [Google Scholar] [CrossRef]
- Sharma, A.; Singh, P.K.; Chandra, R. SMOTified-GAN for Class Imbalanced Pattern Classification Problems. IEEE Access 2022, 10, 30655–30665. [Google Scholar] [CrossRef]
- Puri, A.; Gupta, M.K. Improved Hybrid Bag-Boost Ensemble With K-Means-SMOTE–ENN Technique for Handling Noisy Class Imbalanced Data. Comput. J. 2021, 65, 124–138. [Google Scholar] [CrossRef]
- Seiffert, C.; Khoshgoftaar, T.M.; Van Hulse, J.; Napolitano, A. RUSBoost: A Hybrid Approach to Alleviating Class Imbalance. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 2009, 40, 185–197. [Google Scholar] [CrossRef]
- Czarnowski, I. Weighted Ensemble with one-class Classification and Over-sampling and Instance selection (WECOI): An approach for learning from imbalanced data streams. J. Comput. Sci. 2022, 61, 101614. [Google Scholar] [CrossRef]
- Wang, H.; Xu, Q.; Zhou, L. Large Unbalanced Credit Scoring Using Lasso-Logistic Regression Ensemble. PLoS ONE 2015, 10, e0117844. [Google Scholar] [CrossRef] [PubMed]
- Ariza-Garzón, M.-J.; Arroyo, J.; Segovia-Vargas, M.-J.; Caparrini, A. Profit-sensitive machine learning classification with explanations in credit risk: The case of small businesses in peer-to-peer lending. Electron. Commer. Res. Appl. 2024, 67, 101428. [Google Scholar] [CrossRef]
- Turney, P.D. Cost-Sensitive Classification: Empirical Evaluation of a Hybrid Genetic Decision Tree Induction Algorithm. J. Artif. Intell. Res. 1994, 2, 369–409. [Google Scholar] [CrossRef]
- Ling, C.X.; Yang, Q.; Wang, J.; Zhang, S. Decision trees with minimal costs. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004. [Google Scholar]
- Drummond, C.; Holte, R.C. Exploiting the cost (in) sensitivity of decision tree splitting criteria. In Proceedings of the International Conference on Machine Learning, Stanford, CA, USA, 29 June 29–2 July 2000. [Google Scholar]
- Domingos, P. Metacost: A general method for making classifiers cost-sensitive. In Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 15–18 August 1999. [Google Scholar]
- Witten, I.H.; Frank, E. Data mining: Practical machine learning tools and techniques with Java implementations. Acm Sigmod Rec. 2002, 31, 76–77. [Google Scholar] [CrossRef]
- Chai, X.; Deng, L.; Yang, Q.; Ling, C.X. Test-cost sensitive naive bayes classification. In Proceedings of the Fourth IEEE International Conference on Data Mining (ICDM’04), Brighton, UK, 1–4 November 2004; IEEE: Piscataway, NJ, USA. [Google Scholar]
- Sheng, V.S.; Ling, C.X. Thresholding for making classifiers cost-sensitive. In Proceedings of the Association for the Advancement of Artificial Intelligence, Boston, MA, USA, 16–20 July 2006. [Google Scholar]
- Khan, S.H.; Hayat, M.; Bennamoun, M.; Sohel, F.A.; Togneri, R. Cost-Sensitive Learning of Deep Feature Representations From Imbalanced Data. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 3573–3587. [Google Scholar] [CrossRef] [PubMed]
- Lu, H.; Xu, Y.; Ye, M.; Yan, K.; Gao, Z.; Jin, Q. Learning misclassification costs for imbalanced classification on gene expression data. BMC Bioinform. 2019, 20, 1–10. [Google Scholar] [CrossRef] [PubMed]
- Feng, F.; Li, K.C.; Shen, J.; Zhou, Q.; Yang, X. Using cost-sensitive learning and feature selection algorithms to improve the performance of imbalanced clas-sification. IEEE Access 2020, 8, 69979–69996. [Google Scholar] [CrossRef]
- Khandani, A.E.; Kim, A.J.; Lo, A.W. Consumer credit-risk models via machine-learning algorithms. J. Bank. Financ. 2010, 34, 2767–2787. [Google Scholar] [CrossRef]
- Barboza, F.; Kimura, H.; Altman, E. Machine learning models and bankruptcy prediction. Expert Syst. Appl. 2017, 83, 405–417. [Google Scholar] [CrossRef]
- Yıldırım, M.; Okay, F.Y.; Özdemir, S. Big data analytics for default prediction using graph theory. Expert Syst. Appl. 2021, 176, 114840. [Google Scholar] [CrossRef]
- Peykani, P.; Sargolzaei, M.; Sanadgol, N.; Takaloo, A.; Kamyabfar, H. The application of structural and machine learning models to predict the default risk of listed companies in the Iranian capital market. PLoS ONE 2023, 18, e0292081. [Google Scholar] [CrossRef] [PubMed]
- Chen, N.; Ribeiro, B. A consensus approach for combining multiple classifiers in cost-sensitive bankruptcy prediction. In Proceedings of the Adaptive and Natural Computing Algorithms: 11th International Conference, ICANNGA 2013, Lausanne, Switzerland, 4–6 April 2013; Proceedings 11. Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
- Bahnsen, A.C.; Aouada, D.; Ottersten, B. Example-dependent cost-sensitive decision trees. Expert Syst. Appl. 2015, 42, 6609–6619. [Google Scholar] [CrossRef]
- Zakaryazad, A.; Duman, E. A profit-driven Artificial Neural Network (ANN) with applications to fraud detection and direct marketing. Neurocomputing 2016, 175, 121–131. [Google Scholar] [CrossRef]
- Xia, Y.; Liu, C.; Liu, N. Cost-sensitive boosted tree for loan evaluation in peer-to-peer lending. Electron. Commer. Res. Appl. 2017, 24, 30–49. [Google Scholar] [CrossRef]
- Fiore, U.; De Santis, A.; Perla, F.; Zanetti, P.; Palmieri, F. Using generative adversarial networks for improving classification effectiveness in credit card fraud detection. Inf. Sci. 2017, 479, 448–455. [Google Scholar] [CrossRef]
- Papouskova, M.; Hajek, P. Two-stage consumer credit risk modelling using heterogeneous ensemble learning. Decis. Support Syst. 2019, 118, 33–45. [Google Scholar] [CrossRef]
- De Bock, K.W.; Coussement, K.; Lessmann, S. Cost-sensitive business failure prediction when misclassification costs are uncertain: A heterogeneous ensemble selection approach. Eur. J. Oper. Res. 2020, 285, 612–630. [Google Scholar] [CrossRef]
- Hou, W.-H.; Wang, X.-K.; Zhang, H.-Y.; Wang, J.-Q.; Li, L. A novel dynamic ensemble selection classifier for an imbalanced data set: An application for credit risk assessment. Knowl.-Based Syst. 2020, 208, 106462. [Google Scholar] [CrossRef]
- Li, Z.; Zhang, J.; Yao, X.; Kou, G. How to identify early defaults in online lending: A cost-sensitive multi-layer learning framework. Knowl.-Based Syst. 2021, 221, 106963. [Google Scholar] [CrossRef]
- Barbaglia, L.; Manzan, S.; Tosetti, E. Forecasting Loan Default in Europe with Machine Learning. J. Financ. Econ. 2021, 21, 569–596. [Google Scholar] [CrossRef]
- Gramegna, A.; Giudici, P. SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk. Front. Artif. Intell. 2021, 4, 752558. [Google Scholar] [CrossRef]
- Zou, Y.; Gao, C.; Gao, H. Business Failure Prediction Based on a Cost-Sensitive Extreme Gradient Boosting Machine. IEEE Access 2022, 10, 42623–42639. [Google Scholar] [CrossRef]
- Xing, J.; Chi, G.; Pan, A. Instance-dependent misclassification cost-sensitive learning for default prediction. Res. Int. Bus. Financ. 2024, 69, 102265. [Google Scholar] [CrossRef]
- Wang, S.; Chi, G. Cost-sensitive stacking ensemble learning for company financial distress prediction. Expert Syst. Appl. 2024, 255, 124525. [Google Scholar] [CrossRef]
- Correa Bahnsen, A.; Aouada, D.; Ottersten, B. Ensemble of Example-Dependent Cost-Sensitive Decision Trees. arXiv 2015, arXiv:1505.04637. [Google Scholar]
- Pandove, D.; Rani, R.; Goel, S. Local graph based correlation clustering. Knowl.-Based Syst. 2017, 138, 155–175. [Google Scholar] [CrossRef]
- Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
- Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
- Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: San Francisco, Ca, USA; pp. 785–794. [Google Scholar]
- Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Con-ference on Neural Information Processing Systems, Montréal, Canada, 3–8 December 2018; Curran Associates Inc.: Montréal, QC, Canada; pp. 6639–6649. [Google Scholar]
- Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
- Kumar, V.; Kedam, N.; Sharma, K.V.; Mehta, D.J.; Caloiero, T. Advanced Machine Learning Techniques to Improve Hydrological Prediction: A Comparative Analysis of Streamflow Prediction Models. Water 2023, 15, 2572. [Google Scholar] [CrossRef]
- Charoenwong, B.; Reddy, P. Using forensic analytics and machine learning to detect bribe payments in regime-switching environments: Evidence from the India demonetization. PLoS ONE 2022, 17, e0268965. [Google Scholar] [CrossRef] [PubMed]
- Nandi, A.K.; Randhawa, K.K.; Chua, H.S.; Seera, M.; Lim, C.P. Credit card fraud detection using a hierarchical behavior-knowledge space model. PLoS ONE 2022, 17, e0260579. [Google Scholar] [CrossRef] [PubMed]
- Carbo-Valverde, S.; Cuadros-Solas, P.; Rodríguez-Fernández, F. A machine learning approach to the digitalization of bank customers: Evidence from random and causal forests. PLoS ONE 2020, 15, e0240362. [Google Scholar] [CrossRef]
- Hlongwane, R.; Ramabao, K.; Mongwe, W. A novel framework for enhancing transparency in credit scoring: Leveraging Shapley values for interpretable credit scorecards. PLoS ONE 2024, 19, e0308718. [Google Scholar] [CrossRef]
- Quach, A.C. A Extensions and Improvements to Random Forests for Classification; Utah State University: Logan, Utah, 2017. [Google Scholar]
- Wyrobek, J.; Kluza, K. Efficiency of Gradient Boosting Decision Trees Technique in Polish Companies’ Bankruptcy Prediction. In Proceedings of the Information Systems Architecture and Technology: Proceedings of 39th International Conference on Information Systems Architecture and Technology–ISAT 2018: Part III, Wrocław, Poland, 16–18 September 2019; pp. 24–35. [Google Scholar]
- Freund, Y. Boosting a Weak Learning Algorithm by Majority. Inf. Comput. 1995, 121, 256–285. [Google Scholar] [CrossRef]
- Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
- Lu, M.; Hou, Q.; Qin, S.; Zhou, L.; Hua, D.; Wang, X.; Cheng, L. A Stacking Ensemble Model of Various Machine Learning Models for Daily Runoff Forecasting. Water 2023, 15, 1265. [Google Scholar] [CrossRef]
- Ainan, U.H.; Por, L.Y.; Chen, Y.-L.; Yang, J.; Ku, C.S. Advancing Bankruptcy Forecasting with Hybrid Machine Learning Techniques: Insights from an Unbalanced Polish Dataset. IEEE Access 2024, 12, 1. [Google Scholar] [CrossRef]
- Aiken, J.M.; De Bin, R.; Hjorth-Jensen, M.; Caballero, M.D. Predicting time to graduation at a large enrollment American university. PLoS ONE 2020, 15, e0242334. [Google Scholar] [CrossRef] [PubMed]
- Du, H.; Lv, L.; Wang, H.; Guo, A. A novel method for detecting credit card fraud problems. PLoS ONE 2024, 19, e0294537. [Google Scholar] [CrossRef]
- Jabeur, S.B.; Gharib, C.; Mefteh-Wali, S.; Arfi, W.B. CatBoost model and artificial intelligence techniques for corporate failure prediction. Technol. Fore-Cast. Soc. Chang. 2021, 166, 120658. [Google Scholar] [CrossRef]
- Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
- Lu, H.; Hu, X. Enhancing Financial Risk Prediction for Listed Companies: A Catboost-Based Ensemble Learning Approach. J. Knowl. Econ. 2023, 15, 1–17. [Google Scholar] [CrossRef]
- Enkhtuya, T.; Kang, D.K. Bankruptcy Prediction with Explainable Artificial Intelligence for Early-Stage Business Models. Int. J. Internet Broadcast. Commun. 2023, 15, 58–65. [Google Scholar]
- Peykani, P.; Sargolzaei, M.; Botshekan, M.H.; Oprean-Stan, C.; Takaloo, A. Optimization of Asset and Liability Management of Banks with Minimum Possible Changes. Mathematics 2023, 11, 2761. [Google Scholar] [CrossRef]
- Peykani, P.; Sargolzaei, M.; Takaloo, A.; Sanadgol, N. Investigating the monetary policy risk channel based on the dynamic stochastic general equilibrium model: Empirical evidence from Iran. PLoS ONE 2023, 18, e0291934. [Google Scholar] [CrossRef] [PubMed]
- Marino, M.J. Chapter 3—Statistical Analysis in Preclinical Biomedical Research. In Research in the Biomedical Sciences; Williams, M., Curtis, M.J., Mullane, K., Eds.; Academic Press: Cambridge, MA, USA, 2018; pp. 107–144. [Google Scholar]
- Riffenburgh, R.H. Chapter Summaries. In Statistics in Medicine, 2nd ed.; Riffenburgh, R.H., Ed.; Academic Press: Burlington, MA, USA, 2006; pp. 533–580. [Google Scholar]
- Friedman, M. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
- Hull, J. Machine Learning in Business: An Introduction to the World of Data Science; Amazon Distribution: London, UK, 2020. [Google Scholar]
- Altman, E.I. Financial ratios, discriminant analysis and the prediction of corporate bankruptcy. J. Financ. 1968, 23, 589–609. [Google Scholar] [CrossRef]
- Carton, R.B.; Hofer, C.W. Measuring Organizational Performance: Metrics for Entrepreneurship and Strategic Management Research; Edward Elgar Publishing: Cheltenham, UK, 2006. [Google Scholar]
- Peykani, P.; Sargolzaei, M.; Takaloo, A.; Valizadeh, S. The Effects of Monetary Policy on Macroeconomic Variables through Credit and Balance Sheet Channels: A Dynamic Stochastic General Equilibrium Approach. Sustainability 2023, 15, 4409. [Google Scholar] [CrossRef]
Year | Research | Method of Imbalanced Data | Machine Learning Model | Dataset |
---|---|---|---|---|
2013 | Chen & Ribeiro [65] | Cost-sensitive | KNN Support Vector Machines Decision Trees | 37 French firms |
2015 | Bahnsen [79] | Cost-sensitive | Decision Trees | Credit card transactions and customer data |
2016 | Zakaryazad and Duman [67] | Cost-sensitive | ANN | Credit card fraud detection |
2017 | Xia et al. [68] | Cost-sensitive | XGBoost | Two real-world P2P lending datasets |
2017 | Fiore et al. [69] | Resampling | GAN | credit card fraud |
2019 | Papouskova and Hajek [70] | Cost-sensitive | Decision Tree (C4.5) Logistic regression SVM Random forest AdaBoost | P2P lending consumer loans |
2020 | De Bock et al. [71] | Cost-sensitive | Bagging Random forests | 21 datasets across various industries |
2020 | Hou et al. [72] | Resampling | XGBoost LightGBM | P2P loan |
2021 | Li et al. [73] | Cost-sensitive | LightGBM | 1.6 million online loans |
2021 | Barbaglia et al. [74] | Cost-sensitive | XGBoost Gradient Boosted tree Logistic Regression | 12 million loans |
2021 | Gramegna and Giudici [75] | Resampling | XGBoost | Italian small and medium enterprises |
2022 | Zou et al. [76] | Cost-sensitive | XGBoost | Chinese capital market |
2022 | Chi et al. [77] | Cost-sensitive | Logistic Regression Neural Network | |
2024 | Wang and Chi [78] | Cost-sensitive | Ensemble learning method | 3425 Chinese companies from 2000 to 2020 |
2024 | Our Research | Cost-sensitive and Resampling (CorrOV-CSEn) | Random forest Gradient Boosted tree AdaBoost XGBoost CatBoost | Iranian capital market firms |
Variable | Formula |
---|---|
X1 | |
X3 | |
X4 | |
X5 | |
OM | |
GA | |
GS | |
CROE | |
CPB |
Training Set | X1 | X3 | X4 | X5 | OM | GA | GS | CROE | CPB |
---|---|---|---|---|---|---|---|---|---|
count | 2987 | 2987 | 2987 | 2987 | 2987 | 2987 | 2987 | 2987 | 2987 |
mean | 0.083 | 0.129 | 19.821 | 0.724 | −2.688 | 0.374 | 349.054 | 0.698 | 0.057 |
std | 0.682 | 0.182 | 104.381 | 0.720 | 129.034 | 1.536 | 19,013.508 | 8.601 | 3.794 |
min | −16.681 | −2.109 | 0.002 | −0.192 | −6824.769 | −0.786 | −203.866 | −181.728 | −112.889 |
25% | −0.046 | 0.026 | 1.339 | 0.219 | 0.061 | 0.038 | −0.014 | −0.266 | −0.077 |
50% | 0.145 | 0.106 | 4.532 | 0.577 | 0.192 | 0.176 | 0.257 | 0.143 | 0.013 |
75% | 0.341 | 0.222 | 13.310 | 1.001 | 0.463 | 0.429 | 0.671 | 1.505 | 0.124 |
max | 0.982 | 0.842 | 4133.761 | 7.780 | 230.176 | 68.611 | 1,039,154.000 | 190.281 | 125.772 |
Test set | X1 | X3 | X4 | X5 | OM | GA | GS | CROE | CPB |
count | 1240 | 1240 | 1240 | 1240 | 1240 | 1240 | 1240 | 1240 | 1240 |
mean | 0.224 | 0.192 | 1407.757 | 0.805 | 0.463 | 0.581 | 0.970 | −0.771 | −0.111 |
std | 0.365 | 0.187 | 18,561.378 | 0.787 | 4.646 | 1.882 | 11.161 | 4.969 | 4.079 |
min | −3.494 | −0.781 | 0.001 | −0.579 | −18.486 | −0.637 | −27.413 | −99.452 | −129.227 |
25% | 0.057 | 0.061 | 3.283 | 0.271 | 0.123 | 0.174 | 0.162 | −1.218 | −0.143 |
50% | 0.232 | 0.179 | 7.290 | 0.633 | 0.286 | 0.366 | 0.479 | −0.173 | −0.013 |
75% | 0.402 | 0.313 | 16.479 | 1.132 | 0.622 | 0.612 | 0.828 | 0.425 | 0.083 |
max | 1.000 | 0.838 | 387,142.019 | 7.467 | 159.588 | 44.695 | 385.756 | 32.016 | 47.076 |
Skewness | Kurtosis | |||
---|---|---|---|---|
Training Set | Test Set | Training Set | Test Set | |
X1 | −10.814 | −2.181 | 199.295 | 17.969 |
X3 | −1.193 | −0.008 | 15.695 | 0.939 |
X4 | 25.938 | 17.379 | 897.031 | 319.686 |
X5 | 2.584 | 2.417 | 12.077 | 10.372 |
OM | −50.328 | 32.369 | 2629.754 | 1113.145 |
GA | 32.133 | 19.269 | 1337.749 | 426.320 |
GS | 54.653 | 33.171 | 2986.988 | 1142.797 |
CROE | 1.569 | −7.078 | 172.100 | 136.141 |
CPB | 5.855 | −24.427 | 725.704 | 826.716 |
Training Set | X1 | X3 | X4 | X5 | OM | GA | GS | CROE | CPB |
---|---|---|---|---|---|---|---|---|---|
X1 | 1.000 | 0.529 | 0.103 | 0.097 | 0.255 | 0.040 | −0.004 | −0.173 | 0.008 |
X3 | 0.529 | 1.000 | 0.130 | 0.281 | 0.173 | 0.075 | −0.015 | −0.139 | 0.016 |
X4 | 0.103 | 0.130 | 1.000 | −0.032 | 0.005 | 0.025 | −0.003 | 0.201 | 0.002 |
X5 | 0.097 | 0.281 | −0.032 | 1.000 | 0.022 | −0.018 | −0.016 | 0.003 | 0.008 |
OM | 0.255 | 0.173 | 0.005 | 0.022 | 1.000 | 0.013 | 0.000 | −0.013 | 0.001 |
GA | 0.040 | 0.075 | 0.025 | −0.018 | 0.013 | 1.000 | 0.003 | −0.242 | 0.005 |
GS | −0.004 | −0.015 | −0.003 | −0.016 | 0.000 | 0.003 | 1.000 | −0.001 | −0.003 |
CROE | −0.173 | −0.139 | 0.201 | 0.003 | −0.013 | −0.242 | −0.001 | 1.000 | 0.001 |
CPB | 0.008 | 0.016 | 0.002 | 0.008 | 0.001 | 0.005 | −0.003 | 0.001 | 1.000 |
Test set | X1 | X3 | X4 | X5 | OM | GA | GS | CROE | CPB |
X1 | 1.000 | 0.445 | 0.127 | 0.017 | −0.001 | −0.046 | −0.025 | −0.029 | 0.024 |
X3 | 0.445 | 1.000 | −0.115 | 0.321 | 0.026 | −0.003 | 0.112 | −0.045 | 0.042 |
X4 | 0.127 | −0.115 | 1.000 | −0.086 | 0.009 | −0.028 | −0.016 | 0.012 | −0.001 |
X5 | 0.017 | 0.321 | −0.086 | 1.000 | −0.049 | −0.027 | 0.009 | 0.054 | 0.035 |
OM | −0.001 | 0.026 | 0.009 | −0.049 | 1.000 | −0.007 | −0.001 | 0.004 | −0.008 |
GA | −0.046 | −0.003 | −0.028 | −0.027 | −0.007 | 1.000 | −0.001 | −0.606 | −0.023 |
GS | −0.025 | 0.112 | −0.016 | 0.009 | −0.001 | −0.001 | 1.000 | −0.007 | 0.037 |
CROE | −0.029 | −0.045 | 0.012 | 0.054 | 0.004 | −0.606 | −0.007 | 1.000 | 0.005 |
CPB | 0.024 | 0.042 | −0.001 | 0.035 | −0.008 | −0.023 | 0.037 | 0.005 | 1.000 |
Model | Sensitivity | Precision | F1 Score |
---|---|---|---|
CorrOV-CSEn | |||
Multi-Layer Perceptron (MLP) | 0.841 | 0.327 | 0.471 |
Random Forest | 0.886 | 0.375 | 0.527 |
Gradient Boosting | 0.795 | 0.443 | 0.569 |
XGBoost | 0.795 | 0.393 | 0.526 |
AdaBoost | 0.750 | 0.478 | 0.584 |
CatBoost | 0.909 | 0.201 | 0.329 |
SMOTE | |||
Multi-Layer Perceptron (MLP) | 0.841 | 0.327 | 0.471 |
Random Forest | 0.795 | 0.603 | 0.686 |
Gradient Boosting | 0.727 | 0.603 | 0.660 |
XGBoost | 0.772 | 0.554 | 0.645 |
AdaBoost | 0.568 | 0.555 | 0.561 |
CatBoost | 0.750 | 0.717 | 0.733 |
Dataset-I | Dataset-II | |||||
---|---|---|---|---|---|---|
Model | Sensitivity | Precision | F1 Score | Sensitivity | Precision | F1 Score |
Multi-Layer Perceptron (MLP) | 0.693 | 0.455 | 0.550 | 0.844 | 0.371 | 0.516 |
Random Forest | 0.950 | 0.593 | 0.731 | 0.714 | 0.192 | 0.303 |
Gradient Boosting | 1.000 | 0.666 | 0.800 | 0.571 | 0.500 | 0.533 |
XGBoost | 1.000 | 0.606 | 0.755 | 0.714 | 0.385 | 0.500 |
AdaBoost | 0.800 | 0.640 | 0.711 | 0.429 | 0.429 | 0.429 |
CatBoost | 1.000 | 0.339 | 0.506 | 0.857 | 0.188 | 0.308 |
Dataset-III | Dataset-IV | |||||
Model | Sensitivity | Precision | F1 Score | Sensitivity | Precision | F1 Score |
Multi-Layer Perceptron (MLP) | 0.773 | 0.370 | 0.500 | 0.670 | 0.451 | 0.540 |
Random Forest | 0.666 | 0.240 | 0.353 | 0.875 | 0.368 | 0.519 |
Gradient Boosting | 0.556 | 0.227 | 0.323 | 1.000 | 0.444 | 0.615 |
XGBoost | 0.667 | 0.300 | 0.414 | 1.000 | 0.333 | 0.500 |
AdaBoost | 0.556 | 0.313 | 0.4 | 0.875 | 0.389 | 0.538 |
CatBoost | 0.889 | 0.138 | 0.239 | 1.000 | 0.116 | 0.208 |
Precision | ||||||
Friedman Test Statistic | 12.00 | |||||
p-value | 0.03479 | |||||
Random Forest | Multi-Layer Perceptron (MLP) | Gradient Boosting | XGBoost | AdaBoost | CatBoost | |
Random Forest | - | 0.854075 | 0.635776 | 0.900000 | 0.635776 | 0.744925 |
Multi-Layer Perceptron (MLP) | 0.854075 | - | 0.900000 | 0.900000 | 0.900000 | 0.136905 |
Gradient Boosting | 0.635776 | 0.900000 | - | 0.900000 | 0.900000 | 0.052161 |
XGBoost | 0.900000 | 0.900000 | 0.900000 | - | 0.900000 | 0.410222 |
AdaBoost | 0.635776 | 0.900000 | 0.900000 | 0.900000 | - | 0.052161 |
CatBoost | 0.744925 | 0.136905 | 0.052161 | 0.410222 | 0.052161 | - |
Sensitivity | ||||||
Friedman Test Statistic | 10.04 | |||||
p-value 1 | 0.07413 | |||||
No significant difference was found by the Friedman test because the p-value is greater than the significance level of 0.05. | ||||||
F1 Score | ||||||
Friedman Test Statistic | 10.43 | |||||
p-value 2 | 0.06396 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Peykani, P.; Peymany Foroushany, M.; Tanasescu, C.; Sargolzaei, M.; Kamyabfar, H. Evaluation of Cost-Sensitive Learning Models in Forecasting Business Failure of Capital Market Firms. Mathematics 2025, 13, 368. https://doi.org/10.3390/math13030368
Peykani P, Peymany Foroushany M, Tanasescu C, Sargolzaei M, Kamyabfar H. Evaluation of Cost-Sensitive Learning Models in Forecasting Business Failure of Capital Market Firms. Mathematics. 2025; 13(3):368. https://doi.org/10.3390/math13030368
Chicago/Turabian StylePeykani, Pejman, Moslem Peymany Foroushany, Cristina Tanasescu, Mostafa Sargolzaei, and Hamidreza Kamyabfar. 2025. "Evaluation of Cost-Sensitive Learning Models in Forecasting Business Failure of Capital Market Firms" Mathematics 13, no. 3: 368. https://doi.org/10.3390/math13030368
APA StylePeykani, P., Peymany Foroushany, M., Tanasescu, C., Sargolzaei, M., & Kamyabfar, H. (2025). Evaluation of Cost-Sensitive Learning Models in Forecasting Business Failure of Capital Market Firms. Mathematics, 13(3), 368. https://doi.org/10.3390/math13030368