Next Article in Journal
Optimizing Care Across the Continuum for Older Adults with Lung Cancer: A Review
Previous Article in Journal
Heterogeneous Formation of DNA Double-Strand Breaks and Cell-Free DNA in Leukemia T-Cell Line and Human Peripheral Blood Mononuclear Cells in Response to Topoisomerase II Inhibitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mortality Prediction Modeling for Patients with Breast Cancer Based on Explainable Machine Learning

1
Department of Medical Informatics, School of Medicine, Kangwon National University, Chuncheon 24341, Republic of Korea
2
Institute of Medical Science, School of Medicine, Kangwon National University, Chuncheon 24341, Republic of Korea
3
Cancer Data Center, National Cancer Control Institute, National Cancer Center, Goyang 10408, Republic of Korea
4
Department of Surgery, Center of Breast Cancer, National Cancer Center, Goyang 10408, Republic of Korea
5
Department of Medical Oncology, Center for Breast Cancer, National Cancer Center, Goyang 10408, Republic of Korea
6
Department of Internal Medicine, Kangwon National University Hospital, Chuncheon 24289, Republic of Korea
7
Department of Internal Medicine, School of Medicine, Kangwon National University, Chuncheon 24341, Republic of Korea
8
Targeted Therapy Branch, Research Institute, National Cancer Center, Goyang 10408, Republic of Korea
9
Department of Laboratory Medicine, Hospital, National Cancer Center, Goyang 10408, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2024, 16(22), 3799; https://doi.org/10.3390/cancers16223799
Submission received: 19 September 2024 / Revised: 6 November 2024 / Accepted: 9 November 2024 / Published: 12 November 2024
(This article belongs to the Section Clinical Research of Cancer)

Simple Summary

Breast cancer is the most common cancer in women worldwide, and strategic efforts are required to reduce its mortality. Past studies have usually focused on a limited number of clinical or demographic factors to predict breast cancer prognosis. However, we used thirty-one features, including demographic characteristics, laboratory results, pathology, and treatment information, to predict breast cancer mortality. In addition, the Shapley Additive Explanation (SHAP) method, an explainable artificial intelligence technique, was used. This approach allows us to identify and interpret the key features that have a significant impact on breast cancer mortality. Key predictors of the mortality classification model included occurrence in other organs, age at diagnosis, N stage, T stage, curative radiation treatment, and Ki-67(%). Accurate breast cancer mortality prediction and detection of risk factors based on machine learning may provide opportunities for appropriate therapeutic interventions such as early chemotherapy, surgery, and other measures that may reduce mortality.

Abstract

Background/Objectives: Breast cancer is the most common cancer in women worldwide, requiring strategic efforts to reduce its mortality. This study aimed to develop a predictive classification model for breast cancer mortality using real-world data, including various clinical features. Methods: A total of 11,286 patients with breast cancer from the National Cancer Center were included in this study. The mortality rate of the total sample was approximately 6.2%. Propensity score matching was used to reduce bias. Several machine learning models, including extreme gradient boosting, were applied to 31 clinical features. To enhance model interpretability, we used the SHapley Additive exPlanations method. ML analyses were also performed on the samples, excluding patients who developed other cancers after breast cancer. Results: Among the ML models, the XGB model exhibited the highest discriminatory power, with an area under the curve of 0.8722 and a specificity of 0.9472. Key predictors of the mortality classification model included occurrence in other organs, age at diagnosis, N stage, T stage, curative radiation treatment, and Ki-67(%). Even after excluding patients who developed other cancers after breast cancer, the XGB model remained the best-performing, with an AUC of 0.8518 and a specificity of 0.9766. Additionally, the top predictors from SHAP were similar to the results for the overall sample. Conclusions: Our models provided excellent predictions of breast cancer mortality using real-world data from South Korea. Explainable artificial intelligence, such as SHAP, validated the clinical applicability and interpretability of these models.

1. Introduction

Breast cancer in women is a major public health problem with the highest incidence among cancers and the highest cancer disability-adjusted life-years [1,2]. According to the World Health Organization (WHO), over 2.3 million women were diagnosed with breast cancer in 2020, resulting in 685,000 deaths [3]. The mortality rate of breast cancer is expected to increase, reaching 1 million worldwide by 2040 [4]. Therefore, at the time of initial diagnosis, it is very important to identify the risk factors related to recurrence or survival for each patient and to provide appropriate management plans and treatments, such as surgery, radiation therapy, and chemotherapy. Therefore, predicting mortality and identifying risk factors for breast cancer patients are essential for improving clinical decision-making, controlling the patient’s environment, and enabling appropriate interventions. At the time of initial diagnosis, identifying risk factors related to recurrence or survival for each patient is crucial, allowing for tailored management plans and treatments, such as surgery, radiation therapy, and chemotherapy [5,6].
Traditionally, statistical analyses have been used to investigate breast cancer mortality rates. Recently, novel machine learning (ML) models have demonstrated superior predictive capabilities compared to traditional methods such as logistic regression or Cox regression analysis [7,8,9]. Several studies have demonstrated that ML models have good predictive abilities in identifying patients with breast cancer [10,11,12,13]. Many previous studies have mainly focused on considering only one risk factor, such as BRCA1/2 pathogenic variants, neo-adjuvant chemotherapy, or mammography, to predict breast cancer prognosis [5,14,15]. However, considering the differences in individual mortality rates, even for the same type of cancer, it is important to predict the prognosis using various clinical characteristics for personalized treatment [16,17,18,19,20]. For this reason, more recent studies have reported models for predicting personalized prognosis using various clinical characteristics [17,18,19,20,21,22,23]. In addition, many studies using large cohort datasets such as the Surveillance, Epidemiology, and End Results Program (SEER) [24,25,26,27,28] and the National Cancer Data Base (NCDB) 10–14 [29,30,31] have been instrumental in studying predicting prognoses and proposing treatment methods for breast cancer. These studies have demonstrated risk variables by integrating data on demographic information with clinical and pathologic factors in breast cancer patients. Despite many research advancements, applying findings from Western datasets to the Korean population is limited by genetic, environmental, and other differences, requiring cautious generalization.
Thus, in this study, we developed predictive classification models for breast cancer mortality, identified risk factors, and confirmed the most optimized ML model by using real-world data. We used 31 risk factors, including occurrence in other organs, N stage, chemotherapy, histologic grade, and tumor subtype, to predict breast cancer survival rates.
In addition, we used the SHapley Additive exPlanation (SHAP) method, an explainable artificial intelligence (XAI) technique, for more detailed model interpretation. Consequently, we can overcome the limitations of ML characteristics, which are black-box models, and discover and interpret key features that significantly affect breast cancer mortality (Figure 1).

2. Materials and Methods

2.1. Materials

2.1.1. Study Population and Data Collection

This retrospective study analyzed data from patients diagnosed with breast cancer between January 2005 and December 2020, sourced from the National Cancer Center (NCC), Korea database, a prospectively collected archive of breast cancer treatments at the NCC [32]. Mortality data, including causes of death until 2021, were obtained from Statistics Korea and linked to the NCC database to identify survival status. Initially, 13,165 patients were screened for eligibility, as shown in Figure 2. Given that our study focused on female patients diagnosed with breast cancer, we excluded 1879 patients based on the following criteria: male patients (N = 1), patients without a breast cancer disease code (N = 1), foreign patients (N = 166), patients diagnosed with other cancers before breast cancer (N = 324), patients with benign tumors (N = 858), patients with missing information on histologic grade and TNM stage (N = 167), and patients receiving radiation therapy for palliative treatment (N = 362). After these exclusions, 11,286 patients with a first diagnosis of breast cancer remained, with 695 deaths (6%) and 10,591 survivors (94%). We analyzed the data and found that the occurrence of cancer in other organs after breast cancer diagnosis had the greatest impact on predicting breast cancer mortality. Accordingly, this study was divided into two groups to understand the effect of other variables on mortality prediction. One group included all patients with breast cancer, while the other excluded patients who developed other cancers after a breast cancer diagnosis (n = 1071), leaving a total of 10,215 patients. Among them, 426 (4%) died and 9789 (96%) survived. These were calculated using covariates such as age at diagnosis, height, alcohol consumption, p53(%), Ki-67(%), and human epidermal growth factor receptor 2 (HER2).

2.1.2. Variables

Of the 64 raw data features constructed in the NCC registry shown in Table S1, 33 were excluded. Date-related variables were deemed unnecessary for this study’s model implementation (N = 7), and there were a large number of missing values and data imbalances between groups (N = 10). Furthermore, variables were excluded to prevent multicollinearity based on clinical experts’ opinions and data (N = 16). Thus, we used 31 features in five main groups, including patient’s health and demographic information, laboratory results, treatment types, pathology, and other variables, as summarized in Table S2. Operative features refer to the types of surgical procedures performed for cancer removal, such as breast-conserving surgery (BCS) and mastectomy. Radiotherapy treatment status was determined based on its use for purely curative purposes. Chemotherapy was classified as adjuvant, neoadjuvant, or none. Additional treatments were categorized based on hormone levels and targeted therapy. Experienced breast surgeons from the NCC Pathology Department evaluated the pathological data, which included tumor location (bilateral or unilateral) and type (synchronous or metachronous). Tumor subtypes were classified by hormone receptor and HER2 status as follows: Luminal A (ER+ and/or PR+, and HER2−), Luminal B (ER+ and/or PR+, and HER2+), HER2-overexpressing (ER− and PR−, and HER2+), and basal-like (ER− and PR−, and HER2−). The p53(%) and Ki-67(%) molecular pathology results were used as prognostic markers for early breast cancer to determine the need for further adjuvant chemotherapy. TNM stage was assigned according to the American Joint Committee on Cancer (AJCC) Cancer Staging Manual, 7th edition. The T and N stages, not described in the TNM stage classification, were defined under the guidance of an experienced clinician according to the AJCC criteria, using the size of the primary tumor and the number of lymph nodes.

2.2. Methods

2.2.1. Data Pre-Processing

Prior knowledge about raw data can significantly impact the performance of optimized classifiers, highlighting the importance of data pre-processing in effective mortality classification using machine learning algorithms. The dataset used in this study comprised clinical patient treatment records, where a few outliers—resulting from incorrect entries by medical staff in the electronic medical records (EMR)—were removed. Missing values for features such as age at diagnosis, height, alcohol consumption, smoking status, age at menarche and menopause, and family history were imputed. Continuous variables were imputed with means, while categorical variables were imputed using their respective modes. This approach ensured consistent and efficient handling of missing data. Although mean and mode imputation may not capture all data complexities, they offer a straightforward and effective method for minimizing the risk of introducing additional biases. The number of cases where imputation was performed for each feature is detailed in Table S3. Given the relatively high missing data rate, we selected the mean for continuous variables and the mode for categorical variables [33,34,35]. Continuous variables such as height (mean = 157.35), age at menarche (mean = 14.79), and age at menopause (mean = 49.51) were imputed with their respective means. Categorical variables such as alcohol consumption (mode = 0 [No]), smoking status (mode = 0 [No]), and family history (mode = 0 [No]) were imputed with their respective modes. Mode-based imputation, which fills missing values with the most frequent category, maintains consistency when a majority of observations share a common feature. Sensitivity analyses showed that this strategy did not adversely impact model performance. Furthermore, data from pathology reports were used to impute T- and N-stage data. The T stage was supplemented with pathologic stage and tumor size, while the N stage was supplemented with pathologic stage information. Additionally, skewness and kurtosis methods were employed to identify variables with biased distributions and potential outliers. Log transformation was applied to fasting glucose levels and white blood cell counts, followed by normality tests for each variable. Finally, min-max normalization was conducted on all input variables to prevent issues arising from differences in data scale, enhancing model performance. This normalization step was crucial, as features with larger scales could disproportionately influence the machine learning model compared to those with smaller values [34].

2.2.2. Machine Learning

Four ML models, such as LR, random forest (RF), extreme gradient boosting (XGB), and light gradient boosting machine (LGB), were constructed to classify mortality in breast cancer patients. LR is a traditional probabilistic statistical model widely used in the medical sciences [35]. RF is an ensemble method in ML that constructs numerous decision trees during training and outputs a class, with the final classification determined by the mode or mean of individual tree outcomes [36]. XGB is a tree-based ensemble ML algorithm that combines multiple decision trees and employs classification and regression techniques based on gradient descent. It expands decision trees horizontally (i.e., level-wise) to reduce their depth and works well on imbalanced datasets with excellent accuracy and speed [37]. LGB is an algorithm similar to XGB but can learn faster on large datasets [38]. These classifiers were trained using 31 features, including demographic characteristics, laboratory results, pathology, and treatment information. Mortality prediction was based on the presence or absence of other carcinomas after breast cancer diagnosis. All ML programming for this study was performed using Python (version 3.8.10), with the models constructed using Scikit-learn (version 1.2.0). Hyperparameter tuning was performed using Optuna (version 3.3.0), a hyperparameter optimization framework.

2.2.3. Model Algorithm Development and Evaluation

Of the total data, 80% were allocated to training, while the remaining 20% were used for testing. Validation data were used for 20% of the total training set. For each of the four classification methods, 20% of the selected participant data were completely removed from the cross-validation (CV)-based hyper-parameter value estimates. Stratified k-fold CV (k = 5) was conducted to avoid label distortions during model generation and to maintain stability. The performance measures included accuracy, precision, recall, F1 score, area under the receiver operating characteristic curve (AUC), specificity, Brier score, Matthews correlation coefficient (MCC), and area under the precision–recall curve (AUPRC) [39,40,41,42]. Recall was calculated as true positive (TP)/(TP + false negative (FN)) and specificity as true negative (TN)/(TN + false positive (FP)). The F1 score is the harmonic mean of precision and recall, where precision equals TP/(TP + FP). These criteria are commonly used to report model evaluations [43,44,45]. Additionally, the Brier score offers a direct measure of the accuracy of probabilistic predictions by assessing both calibration and sharpness. It ranges from 0 to 1, with lower values indicating a model that produces more precise predictions [46]. The MCC provides a balanced metric that accounts for true and false positives and negatives, calculated using the formula (TP × TN − FP × FN)/((TP + FP)(TP + FN)(TN + FP)(TN + FN))1/2. The AUPRC emphasizes performance in imbalanced datasets by focusing on precision and recall. In addition, model classification predictions were elucidated using SHAP, a model explanation method based on feature importance. SHAP utilizes cooperative game theory to determine how each feature contributes to ML model predictions, allowing for the interpretation of model performance [47,48]. SHAP assumes independence between features, and correlated features could potentially affect the reliability of SHAP values. To address this, we analyzed the correlations among features used in the model. A correlation heatmap was generated to visually inspect feature relationships, showing that the features were not highly correlated with each other, allowing us to reasonably assume independence. The correlation heatmap is provided in the Supplementary Material (Figure S4). We used Pearson correlation coefficients and observed that no pairs of features exceeded a correlation threshold of 0.7, indicating minimal multicollinearity.

2.2.4. Statistical Analysis

To enhance the robustness and generalizability of our primary findings and balance the baseline covariates, we employed an exposure-driven 1:3 propensity score matching (PSM) analysis while minimizing a logistic regression model with the implications of potential confounders [49]. The quality of the match was evaluated through the standardized mean difference (SMD), with an SMD < 0.1 indicating a negligible difference between the groups (Table S4) [50]. All patient characteristics are presented as means with standard deviations or counts (%). An independent t-test was performed to assess the differences in the mean between the alive and death groups for continuous variables. Categorical variables are compared between two groups using Chi-squared (χ2) tests. Considering the multiple testing of the same cohort, we used a Bonferroni correction based on the number of tested samples in the same cohort [significance threshold of α = 0.05/2(number of tested samples in the same cohort) = 0.025]. In addition, the Cox proportional hazards regression model was fitted after verifying the proportional hazards assumption using the log-log plot (Figure S5). We assessed the hazard ratio (HR) of all variables, using univariate and multivariate Cox regression analysis. We selected the high-impact variables identified consistently across the four ML models and used them in Kaplan–Meier (KM) analysis to assess their effects on survival. For survivors, the duration was calculated from the initial breast cancer diagnosis to the end of the observation period, while for those who passed away, it was calculated from the diagnosis to the date of death. The KM analysis provided insights into survival rates over the entire observation period, tracking changes in breast cancer patient survival. To determine the statistical significance of the top variables identified by SHAP, we performed KM analysis, allowing us to both visually and statistically evaluate the influence of each variable on survival and mortality. The analysis yielded significant p-values for each variable [51], with all statistical tests conducted at a significance level of p < 0.001. All analyses were performed using R software (version 3.6.3).

3. Results

3.1. Patient Characteristics

In this study, we developed predictive ML classification models to predict mortality in patients with breast cancer. We formed two groups: (1) all patients with breast cancer (overall breast cancer group) and (2) excluding patients who developed other cancers after breast cancer (breast-cancer-only group). The first group comprised 11,286 patients (10,591 alive and 695 deceased), and the second group included 10,215 patients (9789 alive and 426 deceased), as shown in Table S5. Using 1:3 PSM, we analyzed 2780 cases (2085 alive and 695 deceased) in the overall breast cancer group and 1704 cases (1278 alive and 426 deceased) in the breast-cancer-only group.
Table 1 presents the characteristics of the two groups. In the overall group, the incidence of cancer in areas other than the breast was 38.7% (p < 0.001) among the deceased, which was approximately five times higher than that among the survivors. The T stage (p < 0.001) and N stage (p < 0.001) showed higher rates among the deceased as the disease progressed. For tumor subtypes (p < 0.05), Luminal A (63.0%), basal (16.1%), Luminal B (13.8%), and HER2 overexpression (7.1%) were observed among the deceased. Additionally, 78.8% (p < 0.001) of survivors received adjuvant chemotherapy at a higher rate than the deceased. In the second group, T stages 2–4 and N stages 1–3 showed higher rates among the deceased as the disease progressed compared to the survivors. Adjuvant chemotherapy was administered to 78.7% (p < 0.001) of survivors, which was higher than that of the deceased. Furthermore, the rate of hormone treatment (p < 0.001) among the survivors was high at 79.2%.

3.2. Cox Proportional Hazard Model

Table 2 presents the results of the univariate and multivariate analyses. In the univariable results for overall breast cancer, age at diagnosis was associated with the outcome (HR 1.03, 95% confidence incidence (CI) [1.03, 1.04]). Other significant predictors included height (HR 0.96, 95% CI [0.95, 0.97]), age at menarche (HR 1.08, 95% CI [1.03, 1.13]), fasting glucose (HR 1.01, 95% CI [1.00, 1.01]), and occurrence in other organs (HR 3.84, 95% CI [3.30, 4.48]). For the breast-cancer-only group, age at diagnosis was also associated with the outcome (HR 1.05, 95% CI [1.04, 1.06]). Other significant predictors included height (HR 0.94, 95% CI [0.92, 0.96]), age at menarche (HR 1.17, 95% CI [1.09, 1.26]), fasting glucose (HR 1.00, 95% CI [1.00, 1.01]), and WBC (HR 1.07, 95% CI [1.03, 1.11]).
In the multivariable results for overall breast cancer, age at diagnosis was associated with the outcome (HR 1.03, 95% CI [1.02, 1.04]). Other significant predictors included height (HR 0.99, 95% CI [0.97, 1.00]), age at menarche (HR 0.96, 95% CI [0.91, 1.02]), fasting glucose (HR 1.00, 95% CI [1.00, 1.00]), and occurrence in other organs (HR 2.57, 95% CI [2.17, 3.04]). For the breast-cancer-only group, age at diagnosis was also associated with the outcome (HR 1.04, 95% CI [1.03, 1.05]). Other significant predictors included height (HR 1.00, 95% CI [0.98, 1.02]), age at menarche (HR 1.08, 95% CI [1.01, 1.14]), fasting glucose (HR 1.00, 95% CI [1.00, 1.01]), and WBC (HR 1.01, 95% CI [0.97, 1.05]).

3.3. Model Performance and Evaluation

Four ML algorithms (LR, RF, XGB, and LGB) were used to construct predictive models for breast cancer mortality. The results for both groups are presented separately. The model was trained using the 31 finally selected features (Table S2). Model performance is presented in Table 3.
In the overall breast cancer group, XGB exhibited the highest discriminative ability, with an AUC of 0.8722, an F1 score of 0.6121, and an AUPRC of 0.7130, followed by LGB with an AUC of 0.8677, an F1 score of 0.6345, and an AUPRC of 0.7059, as shown in Figure 3A. XGB also had the highest accuracy of 0.8381 among the four ML models, followed by LGB with an accuracy of 0.8094 and RF with an accuracy of 0.8058.
In the breast-cancer-only group, XGB showed the highest discriminative ability, with an AUC of 0.8518, an F1 score of 0.6107, and an AUPRC of 0.7013, followed by LGB with an AUC of 0.8482 and an accuracy of 0.8270, as shown in Figure 3B. In addition, XGB demonstrated the highest accuracy of 0.8504, followed by LGB with an accuracy of 0.8270 and RF with 0.7859.
However, although XGB presents the best performance with regards to its AUC, accuracy, and AUPRC, when considering model performance for imbalanced data and minority classes in both groups, it does not show the best performance, with an F1 score lower than LGB for both groups.

3.4. Feature Importance and Interpretation

SHAP values indicate the contributions of individual variables to the predictive classification model results. They help interpret the influence and importance of each feature in the model’s decision-making process. The x-axis represents the SHAP value, which indicates the effect of a feature on the prediction outcome, and the y-axis lists the features. Figure 4 illustrates the XAI results for XGB for each group. The other models’ results are presented in Figure S1. The SHAP summary plot provides an information-dense summary of the effects of the top features of a dataset on the model results (Figure 4). For each instance, the explanation is represented by a single dot for each feature flow. The distribution of points on the plot for each variable across the SHAP value axis indicates the degree of the impact of this feature on the output of the model. Each point represents an individual data point in the dataset.
In the overall breast cancer group, variables such as occurrence in other organs, age at diagnosis, N stage, radiation treatment for curative purposes, height, Ki-67(%), and T stage had a significant impact on the XGB model, which exhibited the highest performance among all variables.
The top four variables, occurrence in other organs, N stage, age at diagnosis, and radiation treatment for curative purposes, were found to have a significant influence on individual model results, which was consistent across the other models as well. The interpretation of SHAP values suggests that the occurrence of other cancers after breast cancer, older age at diagnosis, higher N stage, performance of radiation treatment for curative purposes, and higher Ki-67(%) values are more important in determining the model results. These features are meaningful indicators of mortality predictions.
Similarly, in the breast-cancer-only group, variables such as N stage, age at diagnosis, T stage, p53(%), and Ki-67(%) had a significant impact on the XGB model, which exhibited the highest accuracy among all variables. The variables of T and N stages, age at diagnosis, and Ki-67(%) were found to have a significant influence on the individual model results, consistent with other models. In particular, p53(%) and WBC were included as important variables, compared to the overall breast cancer group.
SHAP analysis revealed that certain high-impact variables significantly contribute to the model’s predictions, indicating their influence on patient outcomes. Specifically, these findings suggest a positive correlation between these variables and survival rates, implying that an increase in their frequency or grade may positively affect survival outcomes. Figure S2 shows the KM survival curves for high-impact variables in the ML models applied to the overall breast cancer group. Furthermore, Figure S3 shows the KM survival curves for the high-impact variables in the ML models applied to the breast-cancer-only group.

4. Discussion

In this study, we developed and evaluated ML models to predict mortality in patients with breast cancer using a comprehensive dataset from the NCC in Korea. Our findings demonstrate the efficacy of ML-based predictive models in accurately classifying patient mortality and highlight the importance of XAI in enhancing the interpretability of these models for clinical decision support systems. Breast cancer remains a significant public health challenge with high incidence rates globally [1,2,3]. Accurate prediction of mortality in patients with breast cancer is crucial for timely intervention and personalized treatment planning. Traditional methods, such as LR and Cox regression analysis, have been widely used but often fall short in terms of predictive performance. Our study showed that advanced ML techniques, including XGB and LGB, outperform traditional methods, providing higher accuracy and specificity for mortality prediction. In our study, the results indicate that the XGB model exhibited the highest discriminative ability, with an AUC of 0.8722 for the overall breast cancer group and 0.8518 for the breast-cancer-only group. The LGB model also showed robust performance, particularly in the second group, with an accuracy of 0.8270. Compared with the existing literature, we utilized advanced ML models and comprehensive data. Hou et al. [52], Nguyen et al. [17], and Allugunti [53] used advanced ML models to improve predictive survival performance. Similar to these studies, we conducted predictive classification for the mortality of patients with breast cancer using advanced models, such as XGB and LGB, achieving high predictive accuracy. A notable strength of our study is the feature selection and importance analysis. Similar to previous studies [54], we emphasized feature importance analysis using SHAP values, providing deeper insights into how individual variables influence mortality predictions with high accuracy. Additionally, KM analysis was used to confirm the survival rates for each category of high-impact variables identified by the ML models. An essential aspect of our study is the application of SHAP values to interpret the model predictions. SHAP values provide insights into the contribution of individual variables, such as age at diagnosis, N stage, T stage, and treatment modalities (e.g., radiation and chemotherapy), to the model’s decision-making process [55]. Interestingly, height was included as an important variable, which may somewhat support the association between height and breast cancer risk reported in previous studies [56,57]. In particular, p53 (%) and WBC were identified as important variables in the breast-cancer-only group. In general, somatic mutations in the p53 gene are common in triple-negative breast cancer with a poor prognosis. However, since the p53 used in this study was based on immunohistochemistry results, it was difficult to identify the association with gene expression or mutation. Further studies are needed to prove the link between them. In addition, our WBC results suggest that additional studies may be needed to identify the relationship between the neutrophil/lymphocyte ratio and poor prognoses. This level of interpretability is crucial for clinicians to understand the factors driving mortality predictions and make informed decisions regarding subject management.
This study had several strengths and unique contributions. First, we used PSM to create a balanced dataset to ensure robust model training and evaluation [58,59]. PSM helped minimize the bias commonly associated with observational studies and improved the reliability of our model predictions. Rajendran et al. [59] addressed class imbalance in breast cancer prediction using techniques such as the synthetic minority oversampling technique [60], under-sampling [61], and hybrid methods. However, although these data augmentation and reduction methods can increase the learning effect of the model, they can be detrimental to the error values that exist in real data, such as outliers. Therefore, we applied the PSM method to ensure that data features and labels were adequately balanced [62,63,64]. Unlike simple oversampling or under-sampling techniques, PSM allowed us to match patients based on key covariates, minimizing selection bias and creating a realistic dataset for training. This approach handled real-world variability more effectively and improved model generalizability by reducing class overrepresentation, resulting in a more balanced view of patient outcomes. In summary, PSM improved model performance while mitigating biases and maintaining data integrity, leading to more reliable predictions, especially in observational datasets.
Second, unlike other studies that primarily focused on predictive performance, our study emphasized the interpretability of model predictions using SHAP values [65]. This approach provides a transparent view of how each feature influences the model’s decisions, which is essential for clinical applications [58]. SHAP values offer a detailed explanation by considering the offsets among all variables and using permutation calculations. This approach ensures that our ML models are accurate and interpretable, making them practical for real-world clinical implementation. Third, we used a dataset that included a wide range of features, including demographic, clinical, pathological, and treatment-related variables, ensuring a holistic approach to mortality prediction. Fourth, our study utilizes real-world data from the NCC, providing a realistic and practical basis for model development and evaluation. Using real-world clinical data offers several advantages, including greater diversity in subject demographics, treatment protocols, and disease progression patterns. In addition, real-world data reflect actual clinical conditions and variations, making the models more applicable and reliable for routine clinical decision-making. This enhances the applicability and relevance of our findings to clinical practice.
However, this study had a few limitations. First, the retrospective nature of this study and data from a single NCC may limit the generalizability of our findings. Moreover, the imbalance between survivors and deaths could introduce bias, although we addressed this issue using PSM. To the best of our knowledge, there is currently no established standard for the optimal ratio of PSM, as long as the case-to-control ratio is not too large (e.g., 1:5 or more). In general, PSM has been performed with a 1:1 ratio, which is also the default value for most PSM tools. However, in this study, the number of deaths among breast cancer patients was relatively small (n = 695, 6% of the total); therefore, we thought that the number of samples included in the machine learning model would be very small if PSM is performed with a ratio of 1:1. For this reason, we considered a 1:3 PSM ratio, referring to previous studies that used a 1:3 PSM ratio for analyses related to the survival of breast cancer patients [66,67,68]. We used approximately 20% of the total data for ML training. Future studies should consider a multimodal approach that combines structural and nonstructural data, such as biosignals and images. Second, this study was cross-sectional, with a focus on mortality prediction and without longitudinal data. Longitudinal studies are needed to design more precise and personalized treatments and validate the utility of these models in continuous subject monitoring. Second, we were unable to perform an external validation of our models, which is crucial for confirming the generalizability and robustness of our findings across different populations. Third, this study included a large number of imputed data using the mean or mode of the variables due to the high rate of missing variables in the EMR data. In addition to the large amount of missing data, this simple imputation approach for missing data could potentially be a major limitation of this study. Recently, various statistical inference methods for missing data (e.g., inference based on maximum likelihood estimates) have been reported, and further research is needed on how actual model results differ between various imputation approaches. In addition, future studies need to make additional efforts to reduce the rate of missing data in datasets. Fourth, the recall of the XGB model was notably lower compared to other performance metrics. Recall was calculated as TP/(TP + FN). In other words, recall is the ratio of the correctly predicted positive class to all classes. In imbalanced datasets, where the positive class (1:death) is often the minority, models tend to predict most data points as belonging to the majority class (0:alive). This leads to an increase in false negatives, as the model misses a significant portion of the minority class. As a result, the sensitivity (or recall) tends to decrease. Therefore, in imbalanced datasets, models are more likely to overlook positive cases, causing sensitivity to drop. To address this limitation, we introduced the Area Under the Precision–Recall Curve (AUPRC) as an additional performance metric. The AUPRC can effectively complement the F1 score, as it evaluates the model’s precision and recall performance across all thresholds rather than just one. This additional measure is particularly useful in imbalanced datasets, where understanding model performance across a range of decision thresholds can reveal nuances that the single-point F1 score may miss. In future studies, to improve this limitation, techniques such as class weighting or using loss functions that focus on recall (such as focal loss) will be used [69]. Fifth, interpretation through SHAP has potential issues because the attribution of feature importance is typically based on random permutations. High correlations among the predictors can lead to SHAP values that may not accurately represent the importance of each feature. In future studies, we plan to implement Kernel SHAP, a method that employs a weighted sampling strategy to calculate each SHAP value, considering the intercorrelations among features. Finally, we used data classified based on whether cancer had occurred in other organs after breast cancer diagnosis to understand its impact on survivors. However, we could not definitively determine whether the malignancies in other organs were new cancers that developed after the breast cancer diagnosis or metastases from the original breast cancer. Future studies are required to establish more precise criteria for patient classification to ensure a clearer distinction between metastatic diseases and new primary malignancies. Given the current data limitations, we included patients with the involvement of other organs; however, future work should aim to more meticulously exclude these cases to differentiate between metastasis and new cancer development accurately.

5. Conclusions

This study demonstrated the predictive classification of breast cancer mortality using cause of death information from Statistics Korea. By utilizing real-world data, such as clinical diagnosis and treatment information, our results are more reliable. We constructed four ML models and achieved high accuracy and AUC using 31 risk factors for breast cancer. Furthermore, through XAI, we identified key variables affecting breast cancer mortality, such as occurrence in other organs, age at diagnosis, N stage, and T stage. Our study results may help physicians determine treatment for the prognostic management of breast cancer patients by predicting the prognosis related to survival. However, a multi-center extension study on representative large-scale data is needed to reduce missing values in EMR data and to develop a more powerful model.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers16223799/s1, Figure S1. SHAP summary plots for each model: (A) overall breast cancer group, logistic regression, (B) overall breast cancer group, random forest, (C) overall breast cancer group, light gradient boosting, (D) breast-cancer-only group, logistic regression, (E) breast-cancer-only group, random forest, (F) breast-cancer-only group, light gradient boosting. SHAP, Shapley additive explanations. Figure S2. Kaplan–Meier plots for the overall breast cancer group: (A) occurrence in other organs, (B) N stage, (C) age at diagnosis, and (D) curative radiation treatment. Figure S3. Kaplan–Meier plots for the breast-cancer-only group: (A) T stage, (B) N stage, (C) age at diagnosis, and (D) Ki-67(%). Figure S4. Heatmap of variable correlations: (A) overall breast cancer group, (B) breast-cancer-only group. Figure S5. Log-Log plot for each variable: (A) overall breast cancer group, (B) breast-cancer-only group. Table S1. All features from the hospital registry database. Table S2. Primary features from the hospital registry database. Table S3. Details of imputed feature information. Table S4. Standardized mean differences before and after PSM in each group. Table S5. Participant characteristics before propensity score matching.

Author Contributions

Conceptualization, S.-Y.J. and H.-J.K.; Data curation, S.W.P., Y.-L.P., E.-G.L., H.C., P.P., D.-W.C. and Y.H.C.; Formal analysis, S.W.P., Y.-L.P., S.A. and W.J.K.; Funding acquisition, H.-J.K.; Investigation, D.-W.C. and J.H.; Methodology, S.W.P., E.-G.L., H.C., S.-Y.J. and H.-J.K.; Visualization, S.W.P. and Y.-L.P.; Writing—original draft, S.W.P. and Y.-L.P.; Writing—review & editing, E.-G.L., H.C., P.P., D.-W.C., Y.H.C., J.H., S.A., K.K., W.J.K., S.-Y.K., S.-Y.J. and H.-J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Cancer Center grant funded by the Korean Government of the Republic of Korea (grant number NCC-2210542-3).

Institutional Review Board Statement

This study was approved by the Institutional Research Board of the National Cancer Center (NCC2022-0129).

Informed Consent Statement

The requirement for informed consent was waived by the IRB.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Global Burden of Disease Cancer Collaboration; Fitzmaurice, C.; Abate, D.; Abbasi, N.; Abbastabar, H.; Abd-Allah, F.; Abdel-Rahman, O.; Abdelalim, A.; Abdoli, A.; Abdollahpour, I.; et al. Global, Regional, and National Cancer Incidence, Mortality, Years of Life Lost, Years Lived with Disability, and Disability-Adjusted Life-Years for 29 Cancer Groups, 1990 to 2017: A Systematic Analysis for the Global Burden of Disease Study. JAMA Oncol. 2019, 5, 1749–1768. [Google Scholar] [CrossRef] [PubMed]
  2. Pfeiffer, R.M.; Park, Y.; Kreimer, A.R.; Lacey, J.V.; Pee, D.; Greenlee, R.T.; Buys, S.S.; Hollenbeck, A.; Rosner, B.; Gail, M.H.; et al. Risk Prediction for Breast, Endometrial, and Ovarian Cancer in White Women Aged 50 y or Older: Derivation and Validation from Population-Based Cohort Studies. PLoS Med. 2013, 10, e1001492. [Google Scholar] [CrossRef] [PubMed]
  3. DeSantis, C.E.; Bray, F.; Ferlay, J.; Lortet-Tieulent, J.; Anderson, B.O.; Jemal, A. International Variation in Female Breast Cancer Incidence and Mortality Rates. Cancer Epidemiol. Biomark. Prev. 2015, 24, 1495–1506. [Google Scholar] [CrossRef] [PubMed]
  4. Arnold, M.; Morgan, E.; Rumgay, H.; Mafra, A.; Singh, D.; Laversanne, M.; Vignat, J.; Gralow, J.R.; Cardoso, F.; Siesling, S.; et al. Current and Future Burden of Breast Cancer: Global Statistics for 2020 and 2040. Breast 2022, 66, 15–23. [Google Scholar] [CrossRef] [PubMed]
  5. Antunes Meireles, P.; Fragoso, S.; Duarte, T.; Santos, S.; Bexiga, C.; Nejo, P.; Luís, A.; Mira, B.; Miguel, I.; Rodrigues, P.; et al. Comparing Prognosis for BRCA1, BRCA2, and Non-BRCA Breast Cancer. Cancers 2023, 15, 5699. [Google Scholar] [CrossRef]
  6. Zhou, S.; Blaes, A.; Shenoy, C.; Sun, J.; Zhang, R. Risk Prediction of Heart Diseases in Patients with Breast Cancer: A Deep Learning Approach with Longitudinal Electronic Health Records Data. iScience 2024, 27, 110329. [Google Scholar] [CrossRef]
  7. Du, M.; Haag, D.G.; Lynch, J.W.; Mittinty, M.N. Comparison of the Tree-Based Machine Learning Algorithms to Cox Regression in Predicting the Survival of Oral and Pharyngeal Cancers: Analyses Based on SEER Database. Cancers 2020, 12, 2802. [Google Scholar] [CrossRef]
  8. Cristofanilli, M.; Budd, G.T.; Ellis, M.J.; Stopeck, A.; Matera, J.; Miller, M.C.; Reuben, J.M.; Doyle, G.V.; Allard, W.J.; Terstappen, L.W.M.M.; et al. Circulating Tumor Cells, Disease Progression, and Survival in Metastatic Breast Cancer. N. Engl. J. Med. 2004, 351, 781–791. [Google Scholar] [CrossRef]
  9. Liu, J.; Zhu, Z.; Hua, Z.; Lin, W.; Weng, Y.; Lin, J.; Mao, H.; Lin, L.; Chen, X.; Guo, J. Radiotherapy Refusal in Breast Cancer with Breast-Conserving Surgery. Radiat. Oncol. 2023, 18, 130. [Google Scholar] [CrossRef]
  10. Nasser, M.; Yusof, U.K. Deep Learning Based Methods for Breast Cancer Diagnosis: A Systematic Review and Future Direction. Diagnostics 2023, 13, 161. [Google Scholar] [CrossRef]
  11. Jabbar, M.A. Breast Cancer Data Classification Using Ensemble Machine Learning. Eng. Appl. Sci. Res. 2021, 48, 65–72. [Google Scholar] [CrossRef]
  12. Chen, H.; Wang, N.; Du, X.; Mei, K.; Zhou, Y.; Cai, G. Classification Prediction of Breast Cancer Based on Machine Learning. Comput. Intell. Neurosci. 2023, 2023, 6530719. [Google Scholar] [CrossRef] [PubMed]
  13. Zhong, X.; Lin, Y.; Zhang, W.; Bi, Q. Predicting Diagnosis and Survival of Bone Metastasis in Breast Cancer Using Machine Learning. Sci. Rep. 2023, 13, 18301. [Google Scholar] [CrossRef]
  14. Gentile, D.; Sagona, A.; De Carlo, C.; Fernandes, B.; Barbieri, E.; Di Maria Grimaldi, S.; Jacobs, F.; Vatteroni, G.; Scardina, L.; Biondi, E.; et al. Pathologic Response and Residual Tumor Cellularity after Neo-Adjuvant Chemotherapy Predict Prognosis in Breast Cancer Patients. Breast 2023, 69, 323–329. [Google Scholar] [CrossRef] [PubMed]
  15. Kim, H.; Lim, J.; Kim, H.-G.; Lim, Y.; Seo, B.K.; Bae, M.S. Deep Learning Analysis of Mammography for Breast Cancer Risk Prediction in Asian Women. Diagnostics 2023, 13, 2247. [Google Scholar] [CrossRef]
  16. Ahn, J.S.; Shin, S.; Yang, S.-A.; Park, E.K.; Kim, K.H.; Cho, S.I.; Ock, C.-Y.; Kim, S. Artificial Intelligence in Breast Cancer Diagnosis and Personalized Medicine. J. Breast Cancer 2023, 26, 405–435. [Google Scholar] [CrossRef]
  17. Nguyen, Q.T.N.; Nguyen, P.; Wang, C.; Phuc, P.T.; Lin, R.; Hung, C.; Kuo, N.; Cheng, Y.; Lin, S.; Hsieh, Z.; et al. Machine Learning Approaches for Predicting 5-year Breast Cancer Survival: A Multicenter Study. Cancer Sci. 2023, 114, 4063–4072. [Google Scholar] [CrossRef]
  18. Kalafi, E.Y.; Nor, N.A.M.; Taib, N.A.; Ganggayah, M.D.; Town, C.; Dhillon, S.K. Machine Learning and Deep Learning Approaches in Breast Cancer Survival Prediction Using Clinical Data. Folia Biol. 2019, 65, 212–220. [Google Scholar] [CrossRef]
  19. Lou, S.-J.; Hou, M.-F.; Chang, H.-T.; Lee, H.-H.; Chiu, C.-C.; Yeh, S.-C.J.; Shi, H.-Y. Breast Cancer Surgery 10-Year Survival Prediction by Machine Learning: A Large Prospective Cohort Study. Biology 2021, 11, 47. [Google Scholar] [CrossRef]
  20. Song, X.; Chu, J.; Guo, Z.; Wei, Q.; Wang, Q.; Hu, W.; Wang, L.; Zhao, W.; Zheng, H.; Lv, X.; et al. Prognostic Prediction of Breast Cancer Patients Using Machine Learning Models: A Retrospective Analysis. Gland. Surg. 2024, 13, 1575–1587. [Google Scholar] [CrossRef]
  21. Sun, J.; Sun, C.-K.; Tang, Y.-X.; Liu, T.-C.; Lu, C.-J. Application of SHAP for Explainable Machine Learning on Age-Based Subgrouping Mammography Questionnaire Data for Positive Mammography Prediction and Risk Factor Identification. Healthcare 2023, 11, 2000. [Google Scholar] [CrossRef] [PubMed]
  22. Escala-Garcia, M.; Morra, A.; Canisius, S.; Chang-Claude, J.; Kar, S.; Zheng, W.; Bojesen, S.E.; Easton, D.; Pharoah, P.D.P.; Schmidt, M.K. Breast Cancer Risk Factors and Their Effects on Survival: A Mendelian Randomisation Study. BMC Med. 2020, 18, 327. [Google Scholar] [CrossRef] [PubMed]
  23. Zhong, X.; Luo, T.; Deng, L.; Liu, P.; Hu, K.; Lu, D.; Zheng, D.; Luo, C.; Xie, Y.; Li, J.; et al. Multidimensional Machine Learning Personalized Prognostic Model in an Early Invasive Breast Cancer Population-Based Cohort in China: Algorithm Validation Study. JMIR Med. Inform. 2020, 8, e19069. [Google Scholar] [CrossRef] [PubMed]
  24. Manikandan, P.; Durga, U.; Ponnuraja, C. An Integrative Machine Learning Framework for Classifying SEER Breast Cancer. Sci. Rep. 2023, 13, 5362. [Google Scholar] [CrossRef]
  25. Wu, Y.; Zhang, Y.; Duan, S.; Gu, C.; Wei, C.; Fang, Y. Survival Prediction in Second Primary Breast Cancer Patients with Machine Learning: An Analysis of SEER Database. Comput. Methods Programs Biomed. 2024, 254, 108310. [Google Scholar] [CrossRef]
  26. Li, X.; Yang, J.; Peng, L.; Sahin, A.A.; Huo, L.; Ward, K.C.; O’Regan, R.; Torres, M.A.; Meisel, J.L. Triple-Negative Breast Cancer Has Worse Overall Survival and Cause-Specific Survival than Non-Triple-Negative Breast Cancer. Breast Cancer Res. Treat. 2017, 161, 279–287. [Google Scholar] [CrossRef]
  27. Narod, S.A.; Iqbal, J.; Giannakeas, V.; Sopik, V.; Sun, P. Breast Cancer Mortality After a Diagnosis of Ductal Carcinoma In Situ. JAMA Oncol. 2015, 1, 888–896. [Google Scholar] [CrossRef]
  28. Nelson, D.R.; Brown, J.; Morikawa, A.; Method, M. Breast Cancer-Specific Mortality in Early Breast Cancer as Defined by High-Risk Clinical and Pathologic Characteristics. PLoS ONE 2022, 17, e0264637. [Google Scholar] [CrossRef]
  29. Dhungana, A.; Vannier, A.; Zhao, F.; Freeman, J.Q.; Saha, P.; Sullivan, M.; Yao, K.; Flores, E.M.; Olopade, O.I.; Pearson, A.T.; et al. Development and Validation of a Clinical Breast Cancer Tool for Accurate Prediction of Recurrence. npj Breast Cancer 2024, 10, 46. [Google Scholar] [CrossRef]
  30. Lara, O.D.; Wang, Y.; Asare, A.; Xu, T.; Chiu, H.-S.; Liu, Y.; Hu, W.; Sumazin, P.; Uppal, S.; Zhang, L.; et al. Pan-Cancer Clinical and Molecular Analysis of Racial Disparities. Cancer 2020, 126, 800–807. [Google Scholar] [CrossRef]
  31. Vannier, A.G.L.; Dhungana, A.; Zhao, F.; Chen, N.; Shubeck, S.; Hahn, O.M.; Nanda, R.; Jaskowiak, N.T.; Fleming, G.F.; Olopade, O.I.; et al. Validation of the RSClin Risk Calculator in the National Cancer Data Base. Cancer 2024, 130, 1210–1220. [Google Scholar] [CrossRef] [PubMed]
  32. Cha, H.S.; Jung, J.M.; Shin, S.Y.; Jang, Y.M.; Park, P.; Lee, J.W.; Chung, S.H.; Choi, K.S. The Korea Cancer Big Data Platform (K-CBP) for Cancer Research. Int. J. Environ. Res. Public Health 2019, 16, 2290. [Google Scholar] [CrossRef] [PubMed]
  33. Jones, J.A.; Farnell, B. Missing and Incomplete Data Reduces the Value of General Practice Electronic Medical Records as Data Sources in Research. Aust. J. Prim. Health 2007, 13, 74–80. [Google Scholar] [CrossRef]
  34. Patro, S.G.K.; Sahu, K.K. Normalization: A Preprocessing Stage. Int. Adv. Res. J. Sci. Eng. Technol. 2015, 2, 20–22. [Google Scholar] [CrossRef]
  35. Feng, J.; Xu, H.; Mannor, S.; Yan, S. Robust Logistic Regression and Classification. In Proceedings of the NIPS 2014, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  36. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  37. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: San Francisco, CA, USA, 2016; pp. 785–794. [Google Scholar]
  38. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  39. Kim, M.; Hwang, K.-B. An Empirical Evaluation of Sampling Methods for the Classification of Imbalanced Data. PLoS ONE 2022, 17, e0271260. [Google Scholar] [CrossRef]
  40. Saito, T.; Rehmsmeier, M. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef]
  41. Liu, Z.; Bondell, H.D. Binormal Precision–Recall Curves for Optimal Classification of Imbalanced Data. Stat. Biosci. 2019, 11, 141–161. [Google Scholar] [CrossRef]
  42. Movahedi, F.; Antaki, J.F. Limitation of ROC in Evaluation of Classifiers for Imbalanced Data. J. Heart Lung Transplant. 2021, 40, S413. [Google Scholar] [CrossRef]
  43. Seyedtabib, M.; Kamyari, N. Predicting Polypharmacy in Half a Million Adults in the Iranian Population: Comparison of Machine Learning Algorithms. BMC Med. Inform. Decis. Mak. 2023, 23, 84. [Google Scholar] [CrossRef]
  44. Hossin, M.; Sulaiman, M.N. A Review on Evaluation Metrics for Data Classification Evaluations. Int. J. Data Min. Knowl. Manag. Process. 2015, 5, 01–11. [Google Scholar] [CrossRef]
  45. Zuo, D.; Yang, L.; Jin, Y.; Qi, H.; Liu, Y.; Ren, L. Machine Learning-Based Models for the Prediction of Breast Cancer Recurrence Risk. BMC Med. Inform. Decis. Mak. 2023, 23, 276. [Google Scholar] [CrossRef] [PubMed]
  46. Schinkel, M.; Boerman, A.W.; Paranjape, K.; Wiersinga, W.J.; Nanayakkara, P.W.B. Detecting Changes in the Performance of a Clinical Machine Learning Tool over Time. eBioMedicine 2023, 97, 104823. [Google Scholar] [CrossRef]
  47. Lundberg, S.M.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  48. Cordova, C.; Muñoz, R.; Olivares, R.; Minonzio, J.-G.; Lozano, C.; Gonzalez, P.; Marchant, I.; González-Arriagada, W.; Olivero, P. HER2 Classification in Breast Cancer Cells: A New Explainable Machine Learning Application for Immunohistochemistry. Oncol. Lett. 2022, 25, 44. [Google Scholar] [CrossRef]
  49. Austin, P.C. An Introduction to Propensity Score Methods for Reducing the Effects of Confounding in Observational Studies. Multivar. Behav. Res. 2011, 46, 399–424. [Google Scholar] [CrossRef]
  50. Choi, Y.; Kim, H.J.; Park, J.; Lee, M.; Kim, S.; Koyanagi, A.; Smith, L.; Kim, M.S.; Rahmati, M.; Lee, H.; et al. Acute and Post-Acute Respiratory Complications of SARS-CoV-2 Infection: Population-Based Cohort Study in South Korea and Japan. Nat. Commun. 2024, 15, 4499. [Google Scholar] [CrossRef]
  51. Li, R.; Shinde, A.; Liu, A.; Glaser, S.; Lyou, Y.; Yuh, B.; Wong, J.; Amini, A. Machine Learning–Based Interpretation and Visualization of Nonlinear Interactions in Prostate Cancer Survival. JCO Clin. Cancer Inform. 2020, 4, 637–646. [Google Scholar] [CrossRef]
  52. Hou, C.; Zhong, X.; He, P.; Xu, B.; Diao, S.; Yi, F.; Zheng, H.; Li, J. Predicting Breast Cancer in Chinese Women Using Machine Learning Techniques: Algorithm Development. JMIR Med. Inform. 2020, 8, e17364. [Google Scholar] [CrossRef]
  53. Allugunti, V.R. Breast Cancer Detection Based on Thermographic Images Using Machine Learning and Deep Learning Algorithms. Int. J. Eng. Comput. Sci. 2022, 4, 49–56. [Google Scholar] [CrossRef]
  54. Ganggayah, M.D.; Taib, N.A.; Har, Y.C.; Lio, P.; Dhillon, S.K. Predicting Factors for Survival of Breast Cancer Patients Using Machine Learning Techniques. BMC Med. Inform. Decis. Mak. 2019, 19, 48. [Google Scholar] [CrossRef]
  55. Li, X.; Zhou, Y.; Dvornek, N.C.; Gu, Y.; Ventola, P.; Duncan, J.S. Efficient Shapley Explanation for Features Importance Estimation Under Uncertainty. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020; Martel, A.L., Abolmaesumi, P., Stoyanov, D., Mateus, D., Zuluaga, M.A., Zhou, S.K., Racoceanu, D., Joskowicz, L., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 12261, pp. 792–801. ISBN 978-3-030-59709-2. [Google Scholar]
  56. van den Brandt, P.A.; Ziegler, R.G.; Wang, M.; Hou, T.; Li, R.; Adami, H.-O.; Agnoli, C.; Bernstein, L.; Buring, J.E.; Chen, Y.; et al. Body Size and Weight Change over Adulthood and Risk of Breast Cancer by Menopausal and Hormone Receptor Status: A Pooled Analysis of 20 Prospective Cohort Studies. Eur. J. Epidemiol. 2021, 36, 37–55. [Google Scholar] [CrossRef] [PubMed]
  57. Kapoor, P.M.; Lindström, S.; Behrens, S.; Wang, X.; Michailidou, K.; Bolla, M.K.; Wang, Q.; Dennis, J.; Dunning, A.M.; Pharoah, P.D.P.; et al. Assessment of Interactions between 205 Breast Cancer Susceptibility Loci and 13 Established Risk Factors in Relation to Breast Cancer Risk, in the Breast Cancer Association Consortium. Int. J. Epidemiol. 2020, 49, 216–232. [Google Scholar] [CrossRef] [PubMed]
  58. Hussain, S.; Ali, M.; Naseem, U.; Nezhadmoghadam, F.; Jatoi, M.A.; Gulliver, T.A.; Tamez-Peña, J.G. Breast Cancer Risk Prediction Using Machine Learning: A Systematic Review. Front. Oncol. 2024, 14, 1343627. [Google Scholar] [CrossRef] [PubMed]
  59. Rajendran, K.; Jayabalan, M.; Thiruchelvam, V. Predicting Breast Cancer via Supervised Machine Learning Methods on Class Imbalanced Data. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 54–63. [Google Scholar] [CrossRef]
  60. Sorayaie Azar, A.; Babaei Rikan, S.; Naemi, A.; Bagherzadeh Mohasefi, J.; Pirnejad, H.; Bagherzadeh Mohasefi, M.; Wiil, U.K. Application of Machine Learning Techniques for Predicting Survival in Ovarian Cancer. BMC Med. Inform. Decis. Mak. 2022, 22, 345. [Google Scholar] [CrossRef]
  61. Lu, S.-C.; Xu, C.; Nguyen, C.H.; Geng, Y.; Pfob, A.; Sidey-Gibbons, C. Machine Learning-Based Short-Term Mortality Prediction Models for Patients With Cancer Using Electronic Health Record Data: Systematic Review and Critical Appraisal. JMIR Med. Inform. 2022, 10, e33182. [Google Scholar] [CrossRef]
  62. Lee, J.; Yoo, S.K.; Kim, K.; Lee, B.M.; Park, V.Y.; Kim, J.S.; Kim, Y.B. Machine Learning-based Radiomics Models for Prediction of Locoregional Recurrence in Patients with Breast Cancer. Oncol. Lett. 2023, 26, 422. [Google Scholar] [CrossRef]
  63. Ma, X.; Wu, S.; Zhang, X.; Chen, N.; Yang, C.; Yang, C.; Cao, M.; Du, K.; Liu, Y. Adjuvant Chemotherapy and Survival Outcomes in Older Women with HR+/HER2− Breast Cancer: A Propensity Score-Matched Retrospective Cohort Study Using the SEER Database. BMJ Open 2024, 14, e078782. [Google Scholar] [CrossRef]
  64. Li, C.; Liu, M.; Zhang, Y.; Wang, Y.; Li, J.; Sun, S.; Liu, X.; Wu, H.; Feng, C.; Yao, P.; et al. Novel Models by Machine Learning to Predict Prognosis of Breast Cancer Brain Metastases. J. Transl. Med. 2023, 21, 404. [Google Scholar] [CrossRef]
  65. Taraniya, I.; PV, B.R.; Divyasri, Y.; Chaithra, V.; Raviteja, N.L. Machine Learning Based Breast Cancer Detection Using Logistic Regression. AIP Conf. Proc. 2024, 2742, 020084. [Google Scholar] [CrossRef]
  66. Cheung, T.T.; Chok, K.S.; Chan, A.C.; Tsang, S.H.; Dai, W.C.; Yau, T.C.; Kwong, A.; Lo, C.M. Survival Analysis of Breast Cancer Liver Metastasis Treated by Hepatectomy: A Propensity Score Analysis for Chinese Women in Hong Kong. Hepatobiliary Pancreat. Dis. Int. 2019, 18, 452–457. [Google Scholar] [CrossRef] [PubMed]
  67. Lee, J.; Kim, J.-Y.; Bae, S.-J.; Cho, Y.; Ji, J.-H.; Kim, D.; Ahn, S.-G.; Park, H.-S.; Park, S.; Kim, S.-I.; et al. The Impact of Post-Mastectomy Radiotherapy on Survival Outcomes in Breast Cancer Patients Who Underwent Neoadjuvant Chemotherapy. Cancers 2021, 13, 6205. [Google Scholar] [CrossRef] [PubMed]
  68. Scomersi, S.; Giudici, F.; Cacciatore, G.; Losurdo, P.; Fracon, S.; Cortinovis, S.; Ceccherini, R.; Zanconati, F.; Tonutti, M.; Bortul, M. Comparison between Male and Female Breast Cancer Survival Using Propensity Score Matching Analysis. Sci. Rep. 2021, 11, 11639. [Google Scholar] [CrossRef] [PubMed]
  69. Wang, K.; Xue, Q.; Lu, J.J. Risky Driver Recognition with Class Imbalance Data and Automated Machine Learning Framework. Int. J. Environ. Res. Public Health 2021, 18, 7534. [Google Scholar] [CrossRef]
Figure 1. The scheme of the study flow.
Figure 1. The scheme of the study flow.
Cancers 16 03799 g001
Figure 2. Flow chart of clinical variables with participants.
Figure 2. Flow chart of clinical variables with participants.
Cancers 16 03799 g002
Figure 3. (A) ROC curve for the overall breast cancer group, (B) ROC curve for the breast-cancer-only group. ROC, Receiver operating characteristic.
Figure 3. (A) ROC curve for the overall breast cancer group, (B) ROC curve for the breast-cancer-only group. ROC, Receiver operating characteristic.
Cancers 16 03799 g003
Figure 4. The SHAP summary plots of the XGB: (A) overall breast cancer group and (B) breast-cancer-only group. SHAP, Shapley additive explanations; XGB, extreme gradient boosting.
Figure 4. The SHAP summary plots of the XGB: (A) overall breast cancer group and (B) breast-cancer-only group. SHAP, Shapley additive explanations; XGB, extreme gradient boosting.
Cancers 16 03799 g004
Table 1. Description of participant attributes after propensity score matching.
Table 1. Description of participant attributes after propensity score matching.
Overall Breast Cancer GroupBreast-Cancer-Only Group
Total (N = 2780)Death (N = 695)Alive (N = 2085)pTotal (N = 1704)Death (N = 426)Alive (N = 1278)p
Age at diagnosis51.4 ± 11.454.7 ± 14.050.3 ± 10.1<0.00152.1 ± 12.057.0 ± 14.650.5 ± 10.5<0.001
Height157.1 ± 5.8155.6 ± 6.2157.5 ± 5.6<0.001156.8 ± 6.0155.2 ± 6.3157.4 ± 5.8<0.001
BMI23.9 ± 3.524.3 ± 3.623.8 ± 3.5<0.0523.9 ± 3.624.4 ± 3.723.8 ± 3.6<0.05
Smoking 0.058 0.518
No2605 (93.7)638 (91.8)1967 (94.3) 1607 (94.3)397 (93.2)1210 (94.7)
Yes175 (6.3)57 (8.2)118 (5.7) 97 (5.7)29 (6.8)68 (5.3)
Drinking <0.001 <0.001
No2214 (79.6)600 (86.3)1614 (77.4) 1335 (78.3)373 (87.6)962 (75.3)
Yes566 (20.4)95 (13.7)471 (22.6) 369 (21.7)53 (12.4)316 (24.7)
Age at menarche14.9 ± 1.515.1 ± 1.514.8 ± 1.5<0.00114.9 ± 1.615.2 ± 1.514.8 ± 1.5<0.001
Age at menopause49.5 ± 3.349.3 ± 3.849.5 ± 3.10.16049.5 ± 3.549.3 ± 4.049.5 ± 3.30.566
Parturition experience 0.907 0.969
No327 (11.8)85 (12.2)242 (11.6) 218 (12.8)53 (12.4)165 (12.9)
Yes2453 (88.2)610 (87.8)1843 (88.4) 1486 (87.2)373 (87.6)1113 (87.1)
Experience of oral contraceptives 0.522 0.963
No2507 (90.2)619 (89.1)1888 (90.6) 1526 (89.6)380 (89.2)1146 (89.7)
Yes273 (9.8)76 (10.9)197 (9.4) 178 (10.4)46 (10.8)132 (10.3)
Hormone Replacement Therapy 0.399 0.821
No1943 (93.2)637 (91.7)2580 (92.8) 1547 (90.8)390 (91.5)1157 (90.5)
Yes142 (6.8)58 (8.3)200 (7.2) 157 (9.2)36 (8.5)121 (9.5)
Family history <0.05 <0.001
No2587 (93.1)667 (96.0)1920 (92.1) 1575 (92.4)414 (97.2)1161 (90.8)
Yes193 (6.9)28 (4.0)165 (7.9) 129 (7.6)12 (2.8)117 (9.2)
Parents’ cancer history <0.001 <0.001
Paternity273 (9.8)40 (5.8)233 (11.2) 145 (8.5)23 (5.4)122 (9.5)
Maternal line158 (5.7)16 (2.3)142 (6.8) 124 (7.3)10 (2.3)114 (8.9)
Parental53 (1.9)6 (0.9)47 (2.3) 32 (1.9)3 (0.7)29 (2.3)
None2296 (82.6)633 (91.1)1663 (79.8) 1403 (82.3)390 (91.5)1013 (79.3)
Cancer history 0.986 0.884
No2665 (95.9)667 (96.0)1998 (95.8) 1654 (97.1)415 (97.4)1239 (96.9)
Yes115 (4.1)28 (4.0)87 (4.2) 50 (2.9)11 (2.6)39 (3.1)
Total cholesterol194.8 ± 36.6194.3 ± 39.2195.0 ± 35.80.907195.0 ± 36.7196.4 ± 40.2194.5 ± 35.50.671
Fasting glucose113.2 ± 41.6122.8 ± 54.4110.0 ± 35.7<0.001113.7 ± 40.6124.2 ± 54.6110.2 ± 34.0<0.001
WBC6.5 ± 2.26.8 ± 2.56.3 ± 2.1<0.0016.5 ± 2.36.9 ± 2.86.4 ± 2.1<0.05
Surgical type <0.001 <0.001
None40 (1.4)9 (1.3)31 (1.5) 23 (1.3)5 (1.2)18 (1.4)
BCS2395 (86.2)529 (76.1)1866 (89.5) 1450 (85.1)322 (75.6)1128 (88.3)
Mastectomy345 (12.4)157 (22.6)188 (9.0) 231 (13.6)99 (23.2)132 (10.3)
Tumor location 0.920 0.330
Left1359 (48.9)350 (50.4)1009 (48.4) 833 (48.9)224 (52.6)609 (47.7)
Right1296 (46.6)313 (45.0)983 (47.1) 796 (46.7)189 (44.4)607 (47.5)
Both125 (4.5)32 (4.6)93 (4.5) 75 (4.4)13 (3.1)62 (4.9)
Tumor location index 0.831 0.613
Single2655 (95.5)663 (95.4)1992 (95.5) 1629 (95.6)413 (96.9)1216 (95.1)
Bilateral and synchronous83 (3.0)24 (3.5)59 (2.8) 47 (2.8)9 (2.1)38 (3.0)
Bilateral and metachronous42 (1.5)8 (1.2)34 (1.6) 28 (1.6)4 (0.9)24 (1.9)
Bilateral 0.764 0.348
No2688 (96.7)675 (97.1)2013 (96.5) 1645 (96.5)416 (97.7)1229 (96.2)
Yes92 (3.3)20 (2.9)72 (3.5) 59 (3.5)10 (2.3)49 (3.8)
Occurrence in other organs <0.001----
No2348 (84.5)426 (61.3)1922 (92.2) ----
Yes432 (15.5)269 (38.7)163 (7.8) ----
Histologic grade <0.001 <0.001
1241 (8.7)32 (4.6)209 (10.0) 153 (9.0)24 (5.6)129 (10.1)
21282 (46.1)246 (35.4)1036 (49.7) 786 (46.1)153 (35.9)633 (49.5)
31257 (45.2)417 (60.0)840 (40.3) 765 (44.9)249 (58.5)516 (40.4)
p53(%)29.9 ± 24.733.4 ± 26.528.8 ± 24.0<0.00129.8 ± 24.333.1 ± 26.228.7 ± 23.5<0.05
Ki-67(%)27.3 ± 22.533.1 ± 26.525.3 ± 20.6<0.00128.1 ± 23.032.7 ± 26.326.6 ± 21.5<0.001
T stage <0.001 <0.001
0134 (4.8)17 (2.4)117 (5.6) 70 (4.1)11 (2.6)59 (4.6)
11569 (56.4)278 (40.0)1291 (61.9) 987 (57.9)166 (39.0)821 (64.2)
2950 (34.2)316 (45.5)634 (30.4) 567 (33.3)196 (46.0)371 (29.0)
395 (3.4)58 (8.3)37 (1.8) 60 (3.5)33 (7.7)27 (2.1)
432 (1.2)26 (3.7)6 (0.3) 20 (1.2)20 (4.7)0 (0.0)
N stage <0.001 <0.001
01862 (67.0)315 (45.3)1547 (74.2) 1140 (66.9)199 (46.7)941 (73.6)
1641 (23.1)213 (30.6)428 (20.5) 410 (24.1)131 (30.8)279 (21.8)
2200 (7.2)110 (15.8)90 (4.3) 105 (6.2)61 (14.3)44 (3.4)
377 (2.8)57 (8.2)20 (1.0) 49 (2.9)35 (8.2)14 (1.1)
Tumor subtype <0.05 <0.001
Luminal A1850 (66.5)438 (63.0)1412 (67.7) 1136 (66.7)260 (61.0)876 (68.5)
Luminal B363 (13.1)96 (13.8)267 (12.8) 205 (12.0)52 (12.2)153 (12.0)
Basal347 (12.5)112 (16.1)235 (11.3) 213 (12.5)83 (19.5)130 (10.2)
HER2 overexpressing220 (7.9)49 (7.1)171 (8.2) 150 (8.8)31 (7.3)119 (9.3)
Radiation treatment for curative <0.001 <0.001
No724 (26.0)267 (38.4)457 (21.9) 479 (28.1)178 (41.8)301 (23.6)
Yes2056 (74.0)428 (61.6)1628 (78.1) 1225 (71.9)248 (58.2)977 (76.4)
Chemotherapy <0.001 <0.001
None211 (7.6)45 (6.5)166 (8.0) 146 (8.6)39 (9.2)107 (8.4)
Adjuvant2056 (74.0)412 (59.3)1644 (78.8) 1266 (74.3)260 (61.0)1006 (78.7)
Neoadjuvant513 (17.1)238 (34.2)275 (13.2) 292 (17.2)127 (29.8)165 (12.9)
Anti-HER2 <0.001 <0.001
No2253 (81.0)479 (68.9)1774 (85.1) 1440 (84.5)334 (78.4)1106 (86.5)
Yes527 (19.0)216 (31.1)311 (14.9) 264 (15.5)92 (21.6)172 (13.5)
Anti-hormone <0.001 <0.001
No656 (23.6)210 (30.2)446 (21.4) 416 (24.4)150 (35.2)266 (20.8)
Yes2124 (76.4)485 (69.8)1639 (78.6) 1288 (75.6)276 (64.8)1012 (79.2)
BMI, body mass index; WBC, white blood cell; BCS, breast-conserving surgery; HER2, human epidermal growth factor receptor 2.
Table 2. Results of univariate and multivariate analyses.
Table 2. Results of univariate and multivariate analyses.
Overall Breast Cancer GroupBreast-Cancer-Only Group
UnivariableMultivariableUnivariableMultivariable
HR (95% CI)p-ValueHR (95% CI)p-ValueHR (95% CI)p-ValueHR (95% CI)p-Value
Age at diagnosis1.03 [1.03, 1.04] <0.0011.03 [1.02, 1.04]<0.0011.05 [1.04, 1.06]<0.0011.04 [1.03, 1.05]<0.001
Height0.96 [0.95, 0.97]<0.0010.99 [0.97, 1.00]0.0560.94 [0.92, 0.96]<0.0011.00 [0.98, 1.02]0.785
BMI1.04 [1.02, 1.06]<0.0010.98 [0.96, 1.01]0.1381.05 [1.02, 1.08]0.0021.01 [0.98, 1.03]0.717
Smoking
No- - -
Yes1.38 [1.05, 1.81]0.0191.36 [1.02, 1.82]0.0371.25 [0.86, 1.82]0.251
Drinking
No- - - -
Yes0.63 [0.51, 0.78]<0.0010.77 [0.61, 0.97]0.0270.52 [0.39, 0.69]<0.0010.52 [0.39, 0.69]<0.001
Age at menarche1.08 [1.03, 1.13]0.0010.96 [0.91, 1.02]0.1681.17 [1.09, 1.26]<0.0011.08 [1.01, 1.14]0.017
Age at menopause0.99 [0.97, 1.01]0.387 0.98 [0.95, 1.01]0.287
Parturition experience
No- -
Yes1.03 [0.82, 1.29]0.801 1.12 [0.84 1.50]0.426
Experience of oral contraceptives
No- -
Yes1.09 [0.86, 1.38]0.492 1.05 [0.73, 1.49]0.784
Hormone replacement therapy
No- -
Yes1.13 [0.86, 1.48]0.367 0.86 [0.61, 1.22]0.400
Family history
No- -
Yes0.73 [0.50, 1.06]0.101 0.44 [0.25, 0.78]0.0050.60 [0.33, 1.09]0.095
Parents’ cancer history
Paternity- - -
Maternal line0.58 [0.32, 1.03]0.065 0.45 [0.22, 0.95]0.0370.61 [0.28, 1.31]0.202
Parental0.70 [0.30, 1.64]0.408 0.66 [0.20, 2.19]0.4950.33 [0.00, 1.15]0.082
None1.26 [0.92, 1.74]0.154 1.25 [0.82, 1.90]0.3091.07 [0.69, 1.65]0.776
Cancer history
No- -
Yes1.35 [0.93, 1.98]0.118 1.11 [0.61, 2.01]0.743
Total cholesterol1.00 [1.00, 1.00]0.471 1.00 [1.00, 1.00]0.092
Fasting glucose1.01 [1.00, 1.01]<0.0011.00 [1.00, 1.00]0.0051.00 [1.00, 1.01]<0.0011.00 [1.00, 1.01]0.140
WBC1.09 [1.06, 1.12]<0.0011.04 [1.00, 1.07]0.0281.07 [1.03, 1.11]<0.0011.01 [0.97, 1.05]0.691
Surgical type
None- -
BCS0.90 [0.47, 1.75]0.7640.62 [0.29, 1.32]0.2121.34 [0.56, 3.25]0.5131.34 [0.56, 3.25]0.513
Mastectomy1.99 [1.02, 3.89]0.0450.53 [0.24, 1.15]0.1082.74 [1.12, 6.73]0.0282.74 [1.12, 6.73]0.028
Tumor location
Left- -
Right0.94 [0.81, 1.09]0.425 0.89 [0.74, 1.08]0.2510.62 [0.23, 1.65] 0.340
Both0.93 [0.65, 1.33]0.685 0.61 [0.35, 1.07]0.0860.49 [0.18, 1.32]0.158
Tumor location index
Single- -
Bilateral & synchronous1.26 [0.84, 1.89]0.268 0.87 [0.45, 1.69]0.687
Bilateral & metachronous0.55 [0.28, 1.11]0.098 0.41 [0.15, 1.09]0.075
Bilateral
No- -
Yes0.88 [0.57, 1.38]0.583 0.70 [0.38, 1.32]0.273
Occurrence in other organs
No- -
Yes3.84 [3.30, 4.48]<0.0012.57 [2.17, 3.04]<0.001
Histologic grade
1- - - -
21.49 [1.03, 2.16]0.0330.89 [0.59, 1.34]0.5811.37 [0.89, 2.11]0.1520.98 [0.62, 1.56]0.936
32.79 [1.95, 3.99]<0.0011.14 [0.75, 1.74]0.5272.45 [1.61, 3.73]<0.0011.00 [0.61, 1.63]0.991
p53(%)1.01 [1.00, 1.01]<0.0011.00 [0.99, 1.00]0.1521.01 [1.00, 1.01]<0.0011.00 [1.00, 1.00]0.825
Ki-67(%)1.02 [1.02, 1.02]<0.0011.01 [1.01, 1.02]<0.0011.02 [1.01, 1.02] <0.0011.01 [1.00, 1.01]<0.001
T stage
0- - - -
11.57 [0.96, 2.56]0.0721.28 [0.75, 2.17]0.3631.02 [0.56, 1.89]0.9370.88 [0.46, 1.68]0.699
22.92 [1.79, 4.75]<0.0011.62 [0.94, 2.77]0.0822.15 [1.17, 3.95]0.0131.20 [0.62, 2.33]0.596
36.52 [3.80, 11.20]<0.0012.38 [1.30, 4.34]0.0054.08 [2.06, 8.08]<0.0011.53 [0.71, 3.27]0.276
415.89 [8.60, 29.33]<0.0012.65 [1.34, 5.22]0.00518.15 [8.66, 38.03]<0.0012.85 [1.25, 6.53]0.013
N stage
0- - - -
12.05 [1.72, 2.44]<0.0011.47 [1.22, 1.78]<0.0012.07 [1.66, 2.59]<0.0011.64 [1.28, 2.09]<0.001
23.50 [2.82, 4.35]<0.0011.99 [1.55, 2.56]<0.0013.90 [2.93, 5.19]<0.0012.83 [2.03, 3.96]<0.001
36.07 [4.57, 8.05]<0.0013.25 [2.40, 4.40]<0.0014.90 [3.42, 7.02]<0.0012.89 [1.92, 4.37]<0.001
Tumor subtype
Luminal A- - - -
Luminal B1.09 [0.87, 1.35]0.4700.62 [0.48, 0.81]<0.0011.18 [0.87, 1.58]0.2860.95 [0.67, 1.35]0.772
Basal1.53 [1.25, 1.89]<0.0010.46 [0.33, 0.66]<0.0012.08 [1.62, 2.66]<0.0010.93 [0.60, 1.45]0.745
HER2 overexpressing1.05 [0.78, 1.41]0.7350.26 [0.17, 0.40]<0.0011.00 [0.69, 1.45]0.9890.52 [0.30, 0.91]0.021
Radiation treatment for curative
Yes- - -
No1.89 [1.62, 2.20]<0.0011.80 [1.53, 2.13]<0.0012.03 [1.67, 2.46]<0.0012.21 [1.70, 2.64]<0.001
Chemotherapy
None- -
Adjuvant0.85 [0.63, 1.16]0.3150.62 [0.48, 0.81]<0.0010.70 [0.50, 0.98]0.0401.35 [0.87, 2.08]0.177
Neoadjuvant2.88 [1.62, 5.06]<0.0010.26 [0.17, 0.40]<0.0012.33 [1.62, 3.35]<0.0013.43 [2.10, 5.61]<0.001
Anti-HER2
No- - -
Yes2.22 [1.89, 2.61]<0.0011.48 [1.19, 1.85]<0.0011.73 [1.37, 2.18]<0.0011.09 [0.80, 1.49]0.566
Anti-hormone
No- - -
Yes0.60 [0.51, 0.71]<0.0010.34 [0.24, 0.47]<0.0010.49 [0.40, 0.60]<0.0010.50 [0.33, 0.76]0.001
HR, hazard ratio; CI, confidence incidence; BMI, body mass index; WBC, white blood cell; BCS, breast-conserving surgery; HER2, Human epidermal growth factor receptor 2.
Table 3. Performance of the four machine learning algorithms.
Table 3. Performance of the four machine learning algorithms.
AccuracyPrecisionRecallF1 ScoreAUCSpecificityBrier ScoreMCCAUPRC
Overall
breast cancer group
LR0.79500.56910.74100.64380.84670.81290.20500.51190.6662
RF0.80580.65660.46760.54620.86240.91850.19420.43700.6679
XGB0.83810.76340.51080.61210.87220.94720.16190.53140.7130
LGB0.80940.60930.66190.63450.86770.85850.19060.50660.7059
Breast-Cancer-only groupLR0.75070.50000.75290.60090.83380.75000.24930.44930.6928
RF0.78590.65790.29410.40650.80430.94920.21410.33450.5648
XGB0.85040.86960.47060.61070.85180.97660.14960.56620.7013
LGB0.82700.68570.56470.61940.84820.91410.17300.51280.6676
AUC, area under the curve; MCC, Matthews correlation coefficient; AUPRC, area under the precision–recall curve; LR, logistic regression; RF, random forest; XGB, extreme gradient boosting; LGB, light gradient boosting machine.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, S.W.; Park, Y.-L.; Lee, E.-G.; Chae, H.; Park, P.; Choi, D.-W.; Choi, Y.H.; Hwang, J.; Ahn, S.; Kim, K.; et al. Mortality Prediction Modeling for Patients with Breast Cancer Based on Explainable Machine Learning. Cancers 2024, 16, 3799. https://doi.org/10.3390/cancers16223799

AMA Style

Park SW, Park Y-L, Lee E-G, Chae H, Park P, Choi D-W, Choi YH, Hwang J, Ahn S, Kim K, et al. Mortality Prediction Modeling for Patients with Breast Cancer Based on Explainable Machine Learning. Cancers. 2024; 16(22):3799. https://doi.org/10.3390/cancers16223799

Chicago/Turabian Style

Park, Sang Won, Ye-Lin Park, Eun-Gyeong Lee, Heejung Chae, Phillip Park, Dong-Woo Choi, Yeon Ho Choi, Juyeon Hwang, Seohyun Ahn, Keunkyun Kim, and et al. 2024. "Mortality Prediction Modeling for Patients with Breast Cancer Based on Explainable Machine Learning" Cancers 16, no. 22: 3799. https://doi.org/10.3390/cancers16223799

APA Style

Park, S. W., Park, Y. -L., Lee, E. -G., Chae, H., Park, P., Choi, D. -W., Choi, Y. H., Hwang, J., Ahn, S., Kim, K., Kim, W. J., Kong, S. -Y., Jung, S. -Y., & Kim, H. -J. (2024). Mortality Prediction Modeling for Patients with Breast Cancer Based on Explainable Machine Learning. Cancers, 16(22), 3799. https://doi.org/10.3390/cancers16223799

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop