Skip to main content

Machine learning models for predicting the onset of chronic kidney disease after surgery in patients with renal cell carcinoma

Abstract

Background

Patients with renal cell carcinoma (RCC) have an elevated risk of chronic kidney disease (CKD) following nephrectomy. Therefore, continuous monitoring and subsequent interventions are necessary. It is recommended to evaluate renal function postoperatively. Therefore, a tool to predict CKD onset is essential for postoperative follow-up and management.

Methods

We constructed a cohort using data from eight tertiary hospitals from the Korean Renal Cell Carcinoma (KORCC) database. A dataset of 4389 patients with RCC was constructed for analysis from the collected data. Nine machine learning (ML) models were used to classify the occurrence and nonoccurrence of CKD after surgery. The final model was selected based on the area under the receiver operating characteristic (AUROC), and the importance of the variables constituting the model was confirmed using the shapley additive explanation (SHAP) value and Kaplan-Meier survival analyses.

Results

The gradient boost algorithm was the most effective among the various ML models tested. The gradient boost model demonstrated superior performance with an AUROC of 0.826. The SHAP value confirmed that preoperative eGFR, albumin level, and tumor size had a significant impact on the occurrence of CKD after surgery.

Conclusions

We developed a model to predict CKD onset after surgery in patients with RCC. This predictive model is a quantitative approach to evaluate post-surgical CKD risk in patients with RCC, facilitating improved prognosis through personalized postoperative care.

Peer Review reports

Introduction

Detection of kidney cancer has improved because of the proliferation of imaging tests such as ultrasound and computed tomography. The kidney is ninth among the most common primary cancer sites, with renal cell carcinomas (RCCs) constituting approximately 90% of these cases [1]. Metastatic diseases affect approximately 20–30% of those diagnosed with RCC, which eventually claims the lives of over 40% of these patients [1,2,3].

Most patients with localized RCC undergo radical or partial nephrectomy [4,5,6]. Although radical nephrectomy offers excellent oncological outcomes for confined malignancies, it is associated with complications [7]. The literature features prominent concerns about the risk of developing chronic kidney disease (CKD) postoperatively following a reduction in nephron mass [8]. Individuals diagnosed with CKD, which is characterized by an estimated glomerular filtration rate (eGFR) of < 60 ml/min/1.73 m2, frequently experience progression to kidney failure, complications due to diminished renal function, or cardiovascular disease and, in some cases, death [9, 10].

Current guidelines from the European Association of Urology and American Urological Association recommend assessing renal function in patients who have undergone surgery for RCC [11]. Renal function in some patients progressively declines following nephrectomy; 1–2% of patients who undergo this procedure ultimately require renal replacement therapy as a result of developing end-stage kidney disease (ESKD) [12, 13]. Hence, it is crucial to identify patients at an elevated risk of developing CKD following nephrectomy.

Research on leveraging machine learning (ML) has emerged with advances in computer technology. Currently, research on medical data for cancer prediction using ML is being intensively pursued [14,15,16,17]. Research on ML in patients with CKD is intensifying. As of 2021, 28 studies have focused on utilizing ML for disease prognosis analysis, and 21 studies have been dedicated to disease diagnostic analysis [18]. However, there is a notable lack of studies predicting CKD in patients who have undergone RCC surgery. Most existing studies target the general patient population or focus on utilizing ML for prognosis rather than CKD prediction [19,20,21]. Risk and odds ratios have been assessed in studies on CKD onset in RCC surgery patients using regression and survival analyses; however, they cannot provide direct probability calculations [4, 22,23,24,25]. If ML is employed, which calculates probabilities and identifies influential factors, it can serve as a diagnostic tool.

In this study, we assessed whether the risk factors identified in prior research are consistent in Korean patients with RCC. Further, we introduced an ML model that forecasts the likelihood of CKD based on a combination of these factors. As our algorithm was formulated using a multicenter dataset sourced from a top-tier Korean hospital, it represents the features of Korean RCC patients without any distortion. Moreover, to our knowledge, our study pioneered the use of ML to predict CKD in patients with RCC. By employing the algorithm we crafted, detecting CKD in patients at an early postoperative stage becomes feasible, allowing for timely and appropriate treatment, enhancing their overall outcome.

Materials and methods

Study population

A web-based database system called the Korean Renal Cell Carcinoma (KORCC) was created to gather basic demographic and clinicopathological data on a sizable cohort of RCC patients in Korea. It was retrospectively built using data from eight hospitals from 1990 to the present [26]. This database was approved by the Ethics Committee of Seoul National University Bundang Hospital (IRB No.: B1202/145 − 102). Using this database, we obtained data from 9598 patients with RCC, encompassing 214 variables. These variables comprised basic demographic factors such as age, sex, height, and weight and clinicopathological features, including clinical stage, pre- and post-surgical test results, and pathological stage. Furthermore, the outcomes were established based on assessing postoperative eGFR levels. CKD was defined as a postoperative eGFR value dropping below 60 ml/min/1.73 m2 [23], which we designated as the outcome. Notably, the eGFR value referred to here was computed using the CKD-EPI equation. The study protocol was approved by the Institutional Review Board (IRB) of the Catholic University of Korea (IRB No. KC23ZIDI0683). The IRB of the Catholic University of Korea waived the requirement for informed consent because this study was retrospective, and personal information in the data were blinded.

Variable selection

The variable selection process was meticulously structured into three sequential phases. Initially, we focused on existing literature addressing postoperative CKD in patients with RCC, leading us to undertake a comprehensive meta-analysis [4, 22,23,24,25]. This study identified 38 notable variables from the previous studies. From this pool, 17 were found to be compatible with our dataset. In the second phase, rigorous statistical evaluations were performed to verify the authenticity of each variable and monitor any missing data points. A p-value threshold of 0.05 was our yardstick for validation. To differentiate between the non-post-CKD and post-CKD cohorts, continuous variables underwent t-test evaluations, whereas categorical counterparts were scrutinized using chi-squared tests, ensuring that we highlighted statistically significant variances. In the last phase, with valuable input from urology specialists, we integrated an additional variable, “smoking,” culminating in a final tally of 12 pivotal variables for our analysis. The final variables included sex, smoking status, history of diabetes mellitus (DM), presence of hypertension (HTN), European Cooperative Oncology Group (ECOG) performance score, hemoglobin (Hb) level, creatinine level, albumin level, calcium level, preoperative eGFR value, tumor size, and tumor location.

Gender was categorized into two groups: male and female, and utilized as a variable in all associated research. Diabetes and hypertension were classified based on their presence or absence. Smoking status was divided into three categories: non-smokers, current smokers, and former smokers who do not currently smoke. The ECOG status, with scores ranging from 0 to 4, indicated physical activity ability; a higher score signifies greater functioning. The ‘Tumor Location’ variable reflects the extent of the tumor’s occupancy on the kidney’s surface. Variables such as Hb, Creatinine, Albumin, Calcium, eGFR, and Tumor Size were continuous, derived from test measurements. All variables, with the exception of Hb, Tumor Size, and Tumor Location, were identified as significant factors in prior research [4, 22,23,24,25].

Data screening

Of the 9,598 patients with RCC, 201 who did not undergo surgery were excluded. Subsequently, we excluded 2,846 patients diagnosed with CKD before surgery, 917 patients with preoperative eGFR levels below 60, 273 patients lacking postoperative eGFR readings, and 972 patients with missing data. Of the remaining 4389 patients, 1,076 developed CKD post-surgery, while 3313 did not exhibit postoperative CKD (Fig. 1).

Fig. 1
figure 1

Data flowchart for analysis

Data splitting and SMOTE for imbalanced datasets

We divided the data at a 7:3 ratio to train and evaluate our model to create two distinct datasets. Given the relatively small number of patients in the post-CKD group, there was a potential for class imbalance. In other words, if trained as is, the model might heavily lean toward the larger no-post-CKD group, ultimately forming a model that predominantly predicts the no-post-CKD group. Techniques such as oversampling, undersampling, and SMOTE can be employed [27,28,29]. We opted for the SMOTE method, known for its widespread use and efficacy, to equalize the training data to a patient group ratio of 1:1 (Table 1).

Table 1 Dataset distribution before to and following the SMOTE application

Model development and validation

In this study, we evaluated the efficacy of several prominent ML classifiers: kernel support vector machine (SVM) [30], logistic regression [31], decision tree [32], k-nearest neighbor (KNN) [33], random forest [34], gradient boost [35], AdaBoost [34, 36], XGBoost [37], and LightGBM [38]. For each classifier, we derived four metrics: sensitivity, specificity, accuracy, F1-score, and area under the receiver operating characteristic (AUROC) [39]. The model delivering the superior performance was chosen based on its AUROC score, which is a crucial metric for assessing classifier effectiveness. All statistical analyses and algorithmic developments were performed using Python (version 3.9.7).

Using the final selected model, we examined the primary factors using the SHapley Additive exPlanations (SHAP) values. SHAP values serve as a representative element of explainable AI, offering a method to gauge the contribution of each variable within the model [40]. Thus, we could discern which variables played a significant role in the model and whether their impact on the outcome was positive or negative. This study aimed to determine the influence of each variable.

Results

Patient characteristics

We assessed the differences in patient attributes and the distribution of individual variables between the post-CKD and non-post-CKD cohorts (Table 2). A higher proportion of patients who developed CKD after surgery were male (73.4% vs. 68.2% in the non-CKD group). The percentage of patients with DM was 20.7% in the postoperative CKD group and 12.9% in the no-post-CKD group. Similarly, the incidences of hypertension (HTN) were 47.9% and 34.4% in the post-CKD and non-post-CKD groups, respectively. These results are in line with those of previous research [4, 22, 24, 25]. The average eGFR values (ml/min/1.73 m2) in the no-post-CKD and post-CKD groups after surgery were 94.7 and 77.7, respectively. Additionally, the average tumor size(cm) before surgery was 38.6% in the no post-CKD group and 52.4 for the post-CKD group. This suggests that patients in the post-CKD group had lower eGFR and larger tumor size before surgery. The distributions of the other variables are listed in Table 2.

Table 2 Baseline characteristics of the patients (N = 4389)

Model performance

We applied 12 variables to nine ML models and calculated their accuracy, specificity, sensitivity, AUROC, and F1 scores to compare their performances. In this study, the final model was selected based on the AUROC value and F1 score, considering data imbalance, sensitivity, and specificity. A grid search was conducted for each ML model to determine the optimal hyperparameters. Table 3 lists each model’s intended hyperparameters. Using the best-selected hyperparameters, we assessed the degree to which each model performed; the results are listed in Table 4.

Table 3 Hyperparameter optimization using grid search

We assessed the performance of each ML model using the optimized parameters, and the results comparing the sensitivity, specificity, accuracy, AUROC, and F1 scores are presented in Table 4. Figure 2 shows the receiver operating characteristic (ROC) curves for the nine ML models. A larger area under the curve indicates a higher AUROC value, indicating a superior performance. Based on the AUROC, the Gradient boost was the top-performing model. Gradient boost demonstrated a predictive performance with a sensitivity of 0.594, specificity of 0.877, accuracy of 0.808, AUROC of 0.826, and F1 score of 0.603.

Table 4 Performance of the ML models
Fig. 2
figure 2

Receiver operating characteristic (ROC) curves of the ML models

To ascertain the significance of the performance of the trained models, post-hoc McNemar’s Chi-Squared testing was conducted [41]. The models were compared against the Gradient boost model, which exhibited the highest AUROC. This comparison was based on the prediction results for the test set. It was observed that the Kernel SVM (p < 0.001), Logistic Regression (p < 0.001), Decision Tree (p < 0.001), and KNN (p < 0.001) all showed statistically significant differences when compared to the Gradient boost model. However, no significant differences were found with the Random forest (p = 0.921), AdaBoost (p = 0.555), XGBoost (p = 0.935), and LightGBM (p = 0.793) models. These models, all belonging to the same family of tree-based ensemble methods, presented challenges in differentiation, leading us to verify the outcomes through the analysis of confusion matrices [42].

Fig. 3
figure 3

Confusion matrix of the top models: a Random forest; b Gradient boost

As depicted in Fig. 2, the performance of the Gradient boost, which showed the highest efficacy, was compared with that of the Random forest, the model with the second-highest performance, using their respective confusion matrices (Fig. 3). In terms of True Negatives (TN), Gradient boost identifies approximately 47 more instances correctly compared to Random forest. However, it is observed that Gradient boost predicts 19 fewer True Positives (TP) than Random forest. Given that the primary objective is to accurately identify diseases within imbalanced clinical data, models with higher AUROC and recall values can be considered to perform better [43]. The recall rates for each model are 55.5% for Random forest and 61.2% for Gradient boost, respectively. Consequently, it can be concluded that Gradient boost demonstrated superior predictive accuracy over Random forest.

Visualization of feature importance

SHAP was used to visually explain the variables constituting the model and verify their impact on CKD after surgery in patients with RCC. Figure 4 shows the SHAP values. The y-axis shows the importance of the model, with the most important variables at the top. The x-axis represents the exponent that responds to the effects of each variable. Here, the red points indicate high-risk values, and the blue points indicate low-risk values. Thus, it was related to the prediction that the occurrence of CKD after surgery would be higher when the eGFR value was lower, albumin was lower, tumor size was larger, calcium level was lower, creatinine level was lower, and Hb level was higher.

Fig. 4
figure 4

SHAP value of the Gradient boost model

We conducted Kaplan-Meier (K-M) survival analyses for the most influential variables: eGFR, Albumin, and Tumor Size [25]. To facilitate a clear comparison between managed and unmanaged groups, we established criteria based on demographic data. Specifically, eGFR was divided into higher and lower groups using a threshold of 86.2, Albumin using 4.35, and Tumor Size using 45.5. In Fig. 5, K-M survival curves were then drawn to ascertain the actual impact of these variables on CKD development. While the survival rate generally dropped below 20% within a year post RCC surgery, it was observed that groups with an eGFR value higher than 86.2 maintained a survival rate above 20% for nearly two years.

Fig. 5
figure 5

Kaplan–Meier curves for the top variables: a eGFR; b Albumin; c Tumor size

Discussion

To our knowledge, this is the first study to predict postoperative CKD in patients with RCC using an ML model, underlining its significance. Nephrectomy and CKD share a pronounced correlation [22]. Consequently, preliminary research has been conducted to ascertain these relationships and associated risk factors using statistical methods [4, 8, 22,23,24,25]. However, no prior studies have harnessed ML techniques to predict postoperative CKD, primarily because of the challenge of amassing adequate data from a single institution for ML applications. In our endeavor, we analyzed data from 4389 patients with RCC gathered from eight top-tier hospitals in our country using the KORCC database. We successfully created an algorithm that uses 12 factors for predicting the probability of postoperative CKD in patients with RCC. Gradient boost demonstrated the best performance among the nine models employed, with an AUROC of 0.826 and an f1 score of 0.603.

Nevertheless, a salient limitation of such ML models is their intricate interpretability [44]. To address this, we utilized the SHAP value to discern feature importance. Our results indicated that preoperative eGFR was the most important variable, followed by albumin level, tumor size, and calcium level (Fig. 4). According to prior studies, preoperative eGFR levels were identified as a significant factor in predicting postoperative CKD risk among patients [22,23,24]. Our data distribution showed a distinct difference in the average preoperative eGFR levels between the no-post-CKD group at 94.7 and the post-CKD group at 77.7. The results of the SHAP values also underscore their prominence, showing the highest significance among the variables. Additionally, albumin and creatinine, which emerged as the second and fifth most influential variables, respectively, have been underscored in prior studies as contributing factors to the onset of postoperative CKD [22]. In our dataset, we observed that the average albumin level was notably lower in the post-CKD group than in the no-post-CKD group. Conversely, the average creatinine levels were slightly elevated in the post-CKD group compared to the no-post-CKD group. The influence of preoperative tumor size on the occurrence of postoperative CKD was verified using SHAP values in our study. Previous research has also indicated that preoperative tumor sizes are notably larger in the post-CKD group than in the no-post-CKD group [22]. In line with these findings, our data showcased that the average tumor sizes in the no-post-CKD and post-CKD groups were 38.6 and 52.4 cm, respectively, highlighting a discernible difference. Our study highlights the significant influence of calcium levels. While previous research has indicated calcium as a significant variable related to preoperative CKD occurrence, it did not highlight the same correlation with postoperative CKD onset [22]. Nonetheless, in our study, slightly higher calcium levels were observed in the no-post-CKD group than in the control group.

In this analysis, a preliminary examination of albumin and calcium levels revealed no apparent significant difference between the post-CKD group and the no-post-CKD group, with an average difference of merely 0.1. However, a deeper investigation using the t-test statistical method uncovered significant disparities. The influence of sample size on statistical significance becomes evident here [45, 46]. Despite the nominal difference in mean albumin levels—merely 0.1—between the two groups, the t-test yielded a t-value exceeding 4, indicative of a significantly low p-value. This outcome illustrates that the significance of variables cannot be adequately assessed by the difference in sample means alone. In a similar vein, calcium levels, despite also presenting a mean difference of 0.1, were found to be significant upon t-test analysis. Furthermore, the machine learning-based SHAP value analysis also identified albumin as a highly influential factor. This underscores the importance of utilizing both statistical and machine learning approaches to fully understand the subtleties in data, especially when initial observations might suggest otherwise [47].

Furthermore, to gain an intuitive understanding of the impact on CKD development, Kaplan-Meier survival analyses were performed on the identified variables [25]. This analysis particularly focused on eGFR, Albumin, and Tumor Size, which emerged as the most influential factors. It was observed that the presence or absence of management for these factors significantly affects the incidence of CKD following RCC surgery.

Our dataset mirrored the distribution characteristics of the risk factors identified in most previous studies. Another strength of our research is the utilization of data from eight different institutions, which helps mitigate potential bias. Additionally, missing values were not permitted for the chosen variables. Employing missing-value imputation techniques for the 972 excluded patients may pave the way for developing a more robust model [48, 49].

Conclusions

We developed a predictive model using ML algorithms to predict the onset of CKD in patients after partial or radical nephrectomy. Gradient boost exhibited the highest performance among the ML models, with an AUROC of 0.826. Using this predictive model, we calculated the likelihood of postoperative CKD occurrence in each patient with RCC. Moreover, this model can improve the prognosis of CKD in patients through tailored postoperative care and appropriate treatment.

Data availability

The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.

References

  1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: A Cancer Journal for Clinicians. 2021;71(3):209–49.

  2. Chin AI, Lam JS, Figlin RA, Belldegrun AS. Surveillance strategies for renal cell carcinoma patients following nephrectomy. Rev Urol. 2006;8(1):1–7.

    PubMed  PubMed Central  Google Scholar 

  3. Jemal A, Murray T, Ward E, Samuels A, Tiwari RC, Ghafoor A et al. Cancer statistics, 2005. CA: A Cancer Journal for Clinicians. 2005;55(1):10–30.

  4. Choi YS, Park YH, Kim YJ, Kang SH, Byun SS, Hong SH. Predictive factors for the development of chronic renal insufficiency after renal surgery: a multicenter study. Int Urol Nephrol. 2014;46(4):681–6.

    Article  CAS  PubMed  Google Scholar 

  5. Ta AD, Bolton DM, Dimech MK, White V, Davis ID, Coory M, et al. Contemporary management of renal cell carcinoma (RCC) in Victoria: implications for longer term outcomes and costs: contemporary management of RCC in Victoria. BJU Int. 2013;112:36–43.

    Article  PubMed  Google Scholar 

  6. Patel MI, Strahan S, Bang A, Vass J, Smith DP. Predictors of surgical approach for the management of renal cell carcinoma: a population-based study from New South Wales. ANZ J Surg. 2017;87(11):E193–8.

    Article  PubMed  Google Scholar 

  7. Sun M, Thuret R, Abdollah F, Lughezzani G, Schmitges J, Tian Z, et al. Age-adjusted incidence, mortality, and survival rates of stage-specific renal cell carcinoma in North America: a trend analysis. Eur Urol. 2011;59(1):135–41.

    Article  PubMed  Google Scholar 

  8. Kim SP, Thompson RH, Boorjian SA, Weight CJ, Han LC, Murad MH, et al. Comparative effectiveness for survival and renal function of partial and radical nephrectomy for localized renal tumors: a systematic review and meta-analysis. J Urol. 2012;188(1):51–7.

    Article  PubMed  Google Scholar 

  9. Lucas SM, Stern JM, Adibi M, Zeltser IS, Cadeddu JA, Raj GV. Renal function outcomes in patients treated for renal masses smaller than 4 cm by ablative and extirpative techniques. J Urol. 2008;179(1):75–80.

    Article  PubMed  Google Scholar 

  10. Go AS, Chertow GM, Fan D, McCulloch CE, Hsu C. yuan. Chronic kidney disease and the risks of death, cardiovascular events, and hospitalization. New England Journal of Medicine. 2004;351(13):1296–305.

  11. Donat SM, Diaz M, Bishoff JT, Coleman JA, Dahm P, Derweesh IH, et al. Follow-up for clinically localized renal neoplasms: AUA guideline. J Urol. 2013;190(2):407–16.

    Article  PubMed  Google Scholar 

  12. Yap SA, Finelli A, Urbach DR, Tomlinson GA, Alibhai SMH. Partial nephrectomy for the treatment of renal cell carcinoma (RCC) and the risk of end-stage renal disease (ESRD). BJU Int. 2015;115(6):897–906.

    Article  PubMed  Google Scholar 

  13. Yokoyama M, Fujii Y, Iimura Y, Saito K, Koga F, Masuda H, et al. Longitudinal change in renal function after radical nephrectomy in Japanese patients with renal cortical tumors. J Urol. 2011;185(6):2066–71.

    Article  PubMed  Google Scholar 

  14. Barlow H, Mao S, Khushi M. Predicting high-risk prostate cancer using machine learning methods. Data. 2019;4(3):129.

    Article  Google Scholar 

  15. Mohammed MA, Abd Ghani MK, Hamed RI, Ibrahim DA. Review on nasopharyngeal carcinoma: concepts, methods of analysis, segmentation, classification, prediction and impact: a review of the research literature. J Comput Sci. 2017;21:283–98.

    Article  Google Scholar 

  16. Mohammed MA, Ghani MKA, Hamed RI, Ibrahim DA. Analysis of an electronic methods for nasopharyngeal carcinoma: prevalence, diagnosis, challenges and technologies. J Comput Sci. 2017;21:241–54.

    Article  Google Scholar 

  17. Mohammed MA, Abd Ghani MK, Arunkumar N, Mostafa SA, Abdullah MK, Burhanuddin MA. Trainable model for segmenting and identifying nasopharyngeal carcinoma. Comput Electr Eng. 2018;71:372–87.

    Article  Google Scholar 

  18. Sanmarchi F, Fanconi C, Golinelli D, Gori D, Hernandez-Boussard T, Capodici A. Predict, diagnose, and treat chronic kidney disease with machine learning: a systematic literature review. J Nephrol. 2023;36(4):1101–17.

    Article  PubMed  PubMed Central  Google Scholar 

  19. Krisanapan P, Tangpanithandee S, Thongprayoon C, Pattharanitima P, Cheungpasitporn W. Revolutionizing chronic kidney disease management with machine learning and artificial intelligence. J Clin Med. 2023;12(8):3018.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Ferguson T, Ravani P, Sood MM, Clarke A, Komenda P, Rigatto C, et al. Development and external validation of a machine learning model for progression of CKD. Kidney Int Rep. 2022;7(8):1772–81.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Schena FP, Anelli VW, Abbrescia DI, Di Noia T. Prediction of chronic kidney disease and its progression by artificial intelligence algorithms. J Nephrol. 2022;35(8):1953–71.

    Article  PubMed  Google Scholar 

  22. Kim SH, Lee SE, Hong SK, Jeong CW, Park YH, Kim YJ, et al. Incidence and risk factors of chronic kidney disease in Korean patients with T1a renal cell carcinoma before and after radical or partial nephrectomy. Jpn J Clin Oncol. 2013;43(12):1243–8.

    Article  PubMed  Google Scholar 

  23. Ellis RJ, White VM, Bolton DM, Coory MD, Davis ID, Francis RS, et al. Incident chronic kidney disease after radical nephrectomy for renal cell carcinoma. Clin Genitourin Cancer. 2019;17(3):e581–91.

    Article  PubMed  Google Scholar 

  24. Ahn T, Ellis RJ, White VM, Bolton DM, Coory MD, Davis ID, et al. Predictors of new-onset chronic kidney disease in patients managed surgically for T1a renal cell carcinoma: an Australian population-based analysis. J Surg Oncol. 2018;117(7):1597–610.

    Article  PubMed  Google Scholar 

  25. Kawamura N, Yokoyama M, Fujii Y, Ishioka J, Numao N, Matsuoka Y, et al. Recovery of renal function after radical nephrectomy and risk factors for postoperative severe renal impairment: a Japanese multicenter longitudinal study. Int J Urol. 2016;23(3):219–23.

    Article  PubMed  Google Scholar 

  26. Byun SS, Hong SK, Lee S, Kook HR, Lee E, Kim HH, et al. The establishment of KORCC (KOrean renal cell carcinoma) database. Invest Clin Urol. 2016;57(1):50–7.

    Article  Google Scholar 

  27. Li K, Zhang W, Lu Q, Fang X. An improved SMOTE, imbalanced data classification method based on support degree. In: 2014 International conference on identification, information and knowledge in the internet of things. 2014. p. 34–8.

  28. Xu Z, Shen D, Nie T, Kou Y, Yin N, Han X. A cluster-based oversampling algorithm combining SMOTE and k-means for imbalanced medical data. Inf Sci. 2021;572:574–89.

    Article  Google Scholar 

  29. Xu Z, Shen D, Nie T, Kou Y. A hybrid sampling algorithm combining M-SMOTE and ENN based on random forest for medical imbalanced data. J Biomed Inform. 2020;107:103465.

    Article  PubMed  Google Scholar 

  30. Huang MW, Chen CW, Lin WC, Ke SW, Tsai CF. SVM and SVM ensembles in breast cancer prediction. PLoS ONE. 2017;6(1):e0161501.

    Article  Google Scholar 

  31. Liao JG, Chin KV. Logistic regression for disease classification using microarray data: model selection in a large p and small n case. Bioinformatics. 2007;23(15):1945–51.

    Article  CAS  PubMed  Google Scholar 

  32. Song YY, Lu Y. Decision tree methods: applications for classification and prediction. Shanghai Arch Psychiatry. 2015;27(2):130–5.

    PubMed  PubMed Central  Google Scholar 

  33. Deng Z, Zhu X, Cheng D, Zong M, Zhang S. Efficient kNN classification algorithm for big data. Neurocomputing. 2016;195:143–8.

    Article  Google Scholar 

  34. Chan JCW, Paelinckx D. Evaluation of random forest and adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens Environ. 2008;112(6):2999–3011.

    Article  Google Scholar 

  35. Chang YC, Chang KH, Wu GJ. Application of eXtreme gradient boosting trees in the construction of credit risk assessment models for financial institutions. Appl Soft Comput. 2018;73:914–20.

    Article  Google Scholar 

  36. Freund Y, Schapire RE. A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci. 1997;55(1):119–39.

    Article  Google Scholar 

  37. Sahin EK. Assessing the predictive capability of ensemble tree methods for landslide susceptibility mapping using XGBoost, gradient boosting machine, and random forest. SN Appl Sci. 2020;2(7):1308.

    Article  CAS  Google Scholar 

  38. Liao H, Zhang X, Zhao C, Chen Y, Zeng X, Li H. LightGBM: an efficient and accurate method for predicting pregnancy diseases. J Obstet Gynaecol. 2022;42(4):620–9.

    Article  CAS  PubMed  Google Scholar 

  39. Huang J, Ling CX. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans Knowl Data Eng. 2005;17(3):299–310.

    Article  CAS  Google Scholar 

  40. Lundberg SM, Lee SI. A unified approach to interpreting model predictions. In: Proceedings of the 31st international conference on neural information processing systems. Red Hook, NY, USA: Curran Associates Inc.; 2017. p. 4768–77. (NIPS’17).

  41. Grzenda A, Speier W, Siddarth P, Pant A, Krause-Sorio B, Narr K et al. Machine learning prediction of treatment outcome in late-life depression. Frontiers in Psychiatry. 2021 [cited 2024 Jan 30];12. Available from: https://www.frontiersin.org/articles/10.3389/fpsyt.2021.738494

  42. Kim HM, Byun SS, Kim JK, Jeong CW, Kwak C, Hwang EC, et al. Machine learning-based prediction model for late recurrence after surgery in patients with renal cell carcinoma. BMC Med Inf Decis Mak. 2022;22(1):241.

    Article  Google Scholar 

  43. Hicks SA, Strümke I, Thambawita V, Hammou M, Riegler MA, Halvorsen P, et al. On evaluation metrics for medical applications of artificial intelligence. Sci Rep. 2022;12(1):5979.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  44. Azodi CB, Tang J, Shiu SH. Opening the black box: interpretable machine learning for geneticists. Trends Genet. 2020;36(6):442–55.

    Article  CAS  PubMed  Google Scholar 

  45. Schuele CM, Justice LM. The ASHA Leader. American Speech-Language-Hearing Association; 2018 [cited 2024 Feb 20]. The Importance of Effect Sizes in the Interpretation of Research. Available from: https://leader.pubs.asha.org/doi/10.1044/leader.FTR4.11102006.14

  46. Thompson B, Statistical. Practical, and clinical: how many kinds of significance do counselors need to consider? Jour Couns Develop. 2002;80(1):64–71.

    Article  Google Scholar 

  47. Statistics versus machine learning.| Nature Methods. [cited 2024 Jan 30]. Available from: https://www.nature.com/articles/nmeth.4642

  48. Nijman S, Leeuwenberg A, Beekers I, Verkouter I, Jacobs J, Bots M, et al. Missing data is poorly handled and reported in prediction model studies using machine learning: a literature review. J Clin Epidemiol. 2022;142:218–29.

    Article  PubMed  Google Scholar 

  49. Zhang X, Fei N, Zhang X, Wang Q, Fang Z. Machine learning prediction models for postoperative stroke in elderly patients: analyses of the MIMIC database. Frontiers in Aging Neuroscience. 2022 [cited 2023 Aug 20];14. Available from: https://www.frontiersin.org/articles/10.3389/fnagi.2022.897611

Download references

Acknowledgements

We appreciate the help that the KOrean Renal Cell Carcinoma (KORCC) group provided for data analysis.

Funding

This research was supported by the Korea Medical Device Development Fund grant funded by the Korean government (Ministry of Science and ICT, Ministry of Trade, Industry and Energy, Ministry of Health & Welfare, Republic of Korea, Ministry of Food and Drug Safety) (Project Number: KMDF_PR_20200901_0096).

Author information

Authors and Affiliations

Authors

Contributions

Conceptualization: S.W.O. and S.-H.H.Methodology: S.-S.B., J.K.K., C.W.J., E.C.H., S.H.K., and S.-H.H.Software: S.W.O.Validation: J.C., C.K., Y.-J.K., Y.-S.H., and S.-H.H.Formal Analysis: S.W.O.Investigation: S.-S.B., J.K.K., C.W.J., E.C.H., S.H.K., J.C., C.K., Y.-J.K., and Y.-S.H.Data curation: S.-H.H.Writing—Original Draft Preparation: S.W.O.Writing—Review and Editing: All authors.Visualization: S.W.O.Supervision: S.-H.H.All authors read and approved the final manuscript.

Corresponding author

Correspondence to Sung-Hoo Hong.

Ethics declarations

Ethics approval and consent to participate

The study protocol was approved by the Institutional Review Board (IRB) of the Catholic University of Korea (IRB No. KC23ZIDI0683). The IRB of the Catholic University of Korea waived the requirement for informed consent because this study was retrospective nature of the study. All experiments were performed in accordance with relevant guidelines and regulations (such as the Declaration of Helsinki).

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Oh, S.W., Byun, SS., Kim, J.K. et al. Machine learning models for predicting the onset of chronic kidney disease after surgery in patients with renal cell carcinoma. BMC Med Inform Decis Mak 24, 85 (2024). https://doi.org/10.1186/s12911-024-02473-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12911-024-02473-8

Keywords