Next Article in Journal
Hepatic Ly6CLo Non-Classical Monocytes Have Increased Nr4a1 (Nur77) in Murine Biliary Atresia
Previous Article in Journal
Reconstructive Surgery and Joint-Sparing Surgery in Valgus and Varus Ankle Deformities: A Comprehensive Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Mortality Using Machine Learning Algorithms in Patients Who Require Renal Replacement Therapy in the Critical Care Unit

1
Division of Nephrology, Department of Internal Medicine, Antai Medical Care Corporation Antai Tian-Sheng Memorial Hospital, Donggang 928, Taiwan
2
Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701, Taiwan
3
Renal-Electrolyte Division, Department of Medicine, University of Pittsburgh, Pittsburgh, PA 15213, USA
4
Division of Nephrology, Department of Internal Medicine, Changhua Christian Hospital, Changhua 500, Taiwan
5
Department of Hospitality Management, MingDao University, Changhua 500, Taiwan
6
Division of Nephrology and Hypertension, Vanderbilt University Medical Center, Nashville, TN 37011, USA
7
Vanderbilt Center for Kidney Disease (VCKD) and Integrated Program for AKI Research (VIP-AKI), Nashville, TN 37011, USA
8
Tennessee Valley Health Systems (TVHS), Veteran’s Health Administration, Nashville, TN 37212, USA
9
Division of General Internal Medicine, Department of Medicine, University of Pittsburgh, Pittsburgh, PA 15213, USA
10
Program for Critical Care Nephrology, CRISMA, Department of Critical Care Medicine, University of Pittsburgh School of Medicine, Pittsburgh, PA 15213, USA
11
Department of Critical Care Medicine, University of Pittsburgh, Pittsburgh, PA 15213, USA
12
Kidney Medicine Section, VA Pittsburgh Healthcare System, Pittsburgh, PA 15240, USA
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(18), 5289; https://doi.org/10.3390/jcm11185289
Submission received: 15 July 2022 / Revised: 4 September 2022 / Accepted: 5 September 2022 / Published: 8 September 2022
(This article belongs to the Section Nephrology & Urology)

Abstract

:
Background: General severity of illness scores are not well calibrated to predict mortality among patients receiving renal replacement therapy (RRT) for acute kidney injury (AKI). We developed machine learning models to make mortality prediction and compared their performance to that of the Sequential Organ Failure Assessment (SOFA) and HEpatic failure, LactatE, NorepInephrine, medical Condition, and Creatinine (HELENICC) scores. Methods: We extracted routinely collected clinical data for AKI patients requiring RRT in the MIMIC and eICU databases. The development models were trained in 80% of the pooled dataset and tested in the rest of the pooled dataset. We compared the area under the receiver operating characteristic curves (AUCs) of four machine learning models (multilayer perceptron [MLP], logistic regression, XGBoost, and random forest [RF]) to that of the SOFA, nonrenal SOFA, and HELENICC scores and assessed calibration, sensitivity, specificity, positive (PPV) and negative (NPV) predicted values, and accuracy. Results: The mortality AUC of machine learning models was highest for XGBoost (0.823; 95% confidence interval [CI], 0.791–0.854) in the testing dataset, and it had the highest accuracy (0.758). The XGBoost model showed no evidence of lack of fit with the Hosmer–Lemeshow test (p > 0.05). Conclusion: XGBoost provided the highest performance of mortality prediction for patients with AKI requiring RRT compared with previous scoring systems.

1. Introduction

Acute kidney injury (AKI) is a common and significant problem in intensive care units (ICU), with incidence rates reportedly as high as 50% of patients admitted [1]. Up to 25% of AKI patients in the ICU require renal replacement therapy (RRT) [1,2]. Despite advancements in the performance and technology of RRT, the mortality rate of those patients remains 30% to 50% [3,4]. Although the outcomes of these patients are likely related partly to the severity of their underlying diseases, having clinical tools that can accurately and reliably provide prognostic predictions is important to aid in clinical decision-making.
General severity of illness scores have been used to predict ICU mortality. For example, the Acute Physiology And Chronic Health Evaluation (APACHE) and Simplified Acute Physiology Score (SAPS) have been developed since the 1980s. They provide adequate prediction of in-hospital and ICU mortality of all ICU patients regardless of ICU type [5,6,7,8]. The Sequential Organ Failure Assessment (SOFA) score is also used for hospital and ICU mortality prediction [9]. However, studies have suggested that these traditional models are not reliable for the AKI populations who need RRT in ICU [10,11,12]. Instead, models using data at RRT initiation have performed better at mortality prediction. Some of these models have shown good performance for mortality prediction but have limited results during external validation [13,14,15,16].
Recently, machine learning models have been broadly applied from disease diagnosis to mortality prediction. They are expected to capture nonlinear interactions from high complexity data and consider all data points for continuous data, thus providing more accurate risk prediction than traditional models. We developed machine learning algorithms using data collected from the Medical Information Mart for Intensive Care (MIMIC-III) [17] and eICU Collaborative Research (eICU-CRD) databases [18] and compared the performance of the results to that of the SOFA [19]; nonrenal SOFA [20]; and HEpatic failure, LactatE, NorepInephrine, medical Condition, and Creatinine (HELENICC) [16] scores in 30-day mortality prediction for AKI patients requiring RRT.

2. Materials and Methods

Data Sources

This retrospective observational cohort study was performed using two publicly available ICU datasets (MIMIC-III and eICU-CRD). MIMIC-III (53,423 ICU admissions between 2001 and 2012) was released by the Massachusetts Institute of Technology Laboratory for Computational Physiology (MIT-LCP) from a single tertiary care hospital (Beth Israel Deaconess Medical Center) in 2016 [17]. eICU-CRD (approximately 200,000 ICU admissions between 2014 and 2015) is a multicenter critical care database from rural/nonacademic hospitals across the United States made available by Philips Healthcare with the help of researchers from MIT-LCP in 2018 [18]. There is no overlap of patients included in these two databases [18].
We included adults ≥18 years old who received RRT (intermittent hemodialysis or continuous RRT [CRRT]) for AKI in the ICU. AKI was defined by the creatinine change level and diagnosis codes in this study. We only used creatinine criteria due to the unreliable urine data in the retrospective databases. Patients were included when they did not have at least 2 creatinine data but had AKI as a diagnosis using ICD-9 codes (Supplementary Table S1) or by maxium–miminum change of creatinine ≥ 0.3 mg/dL from ICU admission to RRT. If a patient had been admitted to the ICU multiple times during one hospitalization course, data from the first ICU admission were extracted for study. Patients with a history of end-stage kidney disease who underwent chronic peritoneal dialysis (PD) or hemodialysis (HD) were excluded from the study, as were those with chronic kidney disease (CKD) stages 4 and 5 based on ICD-9 codes (Supplementary Table S1), because advanced CKD patients who develop AKI are more likely to survive the episode of AKI [21]. Patients with a history of any organ transplant were also excluded, as they may have other confounding risk variables that affect mortality.

3. Model Development

3.1. Predictors

The variables of our models consisted of demographics, medical history, mechanical ventilation use, FiO2, vital signs, laboratory tests, and medications (diuretics, vasopressors). The mechanical ventilation, vital signs, lab tests, and medications were recorded within 24 h before RRT initiation. Relevant past medical history, extracted from database records using ICD-9 codes (Supplementary Table S1), included diabetes mellitus (DM), CKD, hypertension (HTN), congestive heart failure (CHF), liver cirrhosis (LC), and cancer. We used mean values of lab tests, FiO2, Glasgow Coma Scales (GCS), mean arterial pressure (MAP), and respiratory and heart rates (HR). FiO2 was from a laboratory test in MIMIC and from a respiratory chart in eICU.
For laboratory tests, we used mean values of all variables recorded within 24 h before the first dialysis therapy initiation date, because some laboratory data values would have been influenced by dialysis. Supplementary Table S2 reveals the percentage of missing data in laboratory tests. We excluded variables with >30% missing values. Multiple imputation by chained equations (MICE) with five imputed datasets was used to derive the missing values of the laboratory tests and vital signs, and the results were pooled using the MICE package [22]. The missing values were completed, handled by MICE, and then the imputed data were used to build models.
We modified the codes found in the public domain at https://github.com/nus-mornin-lab/oxygenation_kc (accessed on 5 February 2020) and https://github.com/MIT-LCP/mimic-code/tree/master/concepts/severityscores (accessed on 5 February 2020) to calculate a SOFA score using variables collected within 24 h before RRT start in eICU and MIMIC based on methods in the original study [19]. We also calculated the nonrenal SOFA score, which was calculated by the total SOFA score minus the points for the renal system [20]. For patients with missing variables, SOFA and nonrenal SOFA scores were imputed using MICE, as described above.
The primary outcome was all-cause mortality within 30 days of RRT initiation.

3.2. Prediction Machine Learning Algorithms

Predicting mortality problem belongs to a classification topic in supervised machine learning. Four machine learning classification methods were applied in this study: logistic regression (LR), XGBoost, random forest (RF), and multilayer perceptron (MLP). We used grid search with tenfold cross-validation to find the best hyperparameters for all models. Our machine learning modeling strategy followed TRIPOD statement recommendations for the reporting of predictive models [23].
  • LR is the fundamental algorithm for machine learning development. In scikit-learn, the LR uses regularization by default. The advantage of regularization is to improve numerical stability.
  • XGBoost [24] is an implementation of the gradient-boosted decision trees ensemble algorithm. The implementation of XGBoost is optimized for performance and provides the best available solutions in many fields. It reduces variance and bias by using multiple models and adjusting the subsequent trees by the errors the previous trees made.
  • RF [25] is a bagging ensemble machine learning model that also includes several decision trees, but decisions made among trees are independent. It chooses the final model by voting for the most common class that reduces variance in decision trees. The advantages of RF are as follows: it is robust to overfitting and is more stable in high-dimensional data than other machine learning algorithms [26].
  • MLP [27] is a well known supervised learning implementation in artificial neural networks. Typically, it consists of one input layer, one or more hidden layers, and one output layer. It solves high-dimensional classification problems by dealing with the interactions among variables.

4. Model Validation

Models were validated using two strategies (Figure 1): Using the first strategy to assess validation, we used the MIMIC dataset as a development model and assessed external validity using the eICU dataset, which was based on the higher severity of comorbidities and more complete records. Using the second strategy to build more robust models that could be applied across institutions, we pooled eICU and MIMIC datasets containing more diverse and heterogeneous data so that the trained models would generalize across different hospitals and then randomly split them into training and testing datasets at a ratio of 8:2. We performed grid searches with tenfold cross-validation to obtain the best parameters using the training dataset and then tested models on the independent testing dataset.

5. Statistical Analyses

We compared the baseline characteristics between the survival and death groups. Categorical variables were presented as proportions, and the mean with standard deviation or median with interquartile range was used to summarize the results for continuous variables. Numeric variables of clinical characteristics with normal distribution tested by the Kolmogorov–Smirnov test between the two groups were compared using the Student’s t-test. Non-normally distributed continuous variables were tested by a Mann–Whitney U test. A Chi-squared test was used to compare the differences in categorical variables.
The overall performance of the prediction models on validation was assessed by the calculation of the area under the receiver operating characteristic curve (AUC) and the associated 95% confidence interval (CI) using the roc_auc_score function of scikit-learn. Calibration was assessed using the Hosmer–Lemeshow test and by constructing calibration curves. The differences between model AUCs were pairwise-compared using the DeLong test (p < 0.05 was considered statistically significant). Sensitivity, specificity, positive (PPV) and negative (NPV) predicted values, and accuracy were calculated for evaluation of model performance. To evaluate the impact of features on our best model, we used the SHAP framework (available in the public domain at https://github.com/slundberg/shap (accessed on 5 February 2020) [28]. We used decision curve analysis to assess the net benefits of our best machine learning model, SOFA, nonrenal SOFA, and HELENICC scores. In the decision curve analysis, SOFA, nonrenal SOFA, and HELENICC scores were converted to a logistic regression using probability theory [12].
Machine learning algorithms and statistical analyses were performed using Python version 3.6, scikit-learn version 0.22.1, keras version 2.3.1, and R version 3.6.1.

6. Results

6.1. General Demographics

Of 3357 patients in the MIMIC and 8201 patients in the eICU databases who required dialysis therapy, 1129 and 2283, respectively, met the criteria for study inclusion (Figure 2).
Baseline demographics, comorbidities, vital signs, and laboratory values for patients in the two datasets are grouped by survival status (Supplement Tables S3 and S4). Overall, the cohorts from the group who died were older and had a lower percentage of black race, a longer ICU stay before dialysis therapy initiation, and a higher percentage of mechanical ventilation use. Supplementary Table S5 reveals that the mortality rate and comorbidity are significantly different between the eICU and MIMIC datasets. The 30-day ICU mortality rate was 42.9% and 32.7% in the MIMIC and eICU datasets, respectively.
Table 1 shows the differences between the training and testing datasets of the pooled dataset. Regarding BUN, hemoglobin, and glucose, other variables were similar between the training and testing datasets.

6.2. SOFA, Nonrenal SOFA, and HELENICC Scores Performance in the MIMIC and eICU Datasets

The SOFA scores of the survival/death groups in the MIMIC and eICU cohorts were 8.7 ± 3.4/11.5 ± 3.6 (p < 0.001) and 10.6 ± 3.1/13.6 ± 3.2 (p < 0.001), respectively. The nonrenal SOFA scores of the survival/death groups in the MIMIC and eICU cohorts were 5.7 ± 3.3/8.8 ± 3.6 (p < 0.001) and 7.7 ± 3.2/11.1 ± 3.3 (p < 0.001), respectively. The performance of SOFA, nonrenal SOFA, and HELENICC scores in 30-day mortality prediction was modest in both datasets (Table 2). The nonrenal SOFA score performed better than the SOFA and HELENICC scores (0.728; 95% CI, 0.699–0.758 and 0.769; 95% CI, 0.749–0.789 in the MIMIC and eICU cohorts, respectively).

6.3. Machine Learning Algorithm Performance and Comparison with Other Predictive Models in the First Strategy

In the eICU (testing) dataset, the RF model achieved the highest AUC (0.816; 95% CI, 0.798–0.834) (Table 3), but there were no significant differences between those models (Supplementary Table S6). All four models performed significantly better than the SOFA, nonrenal SOFA, and HELENICC scores (p < 0.001). The Hosmer–Lemeshow test showed a poor model fit, except the MLP model. Figure 3 illustrates the ROC curves for our models, as well as for the SOFA, nonrenal SOFA, and HELENICC scores. Supplementary Figures S1 and S2 demonstrate the calibration curves of all models in the training and testing datasets.

6.4. Machine Learning Algorithm Performance and Comparison with Other Predictive Models in the Secondary Strategy

In the pooled dataset, there were 978 and 258 deaths in the training and testing datasets, respectively. The XGBoost model achieved the highest AUC value and accuracy (0.823; 95% CI, 0.791–0.854; 0.758) (Table 4). The MLP model performed worse than the other three models (Supplement Table S7). The XGBoost, LR, and RF models showed no evidence of lack of fit with a Hosmer–Lemeshow test (p > 0.05) in the testing dataset. All four models performed significantly better than the SOFA, nonrenal SOFA, and HELENICC scores (p < 0.001). Figure 4 shows the ROC curves for all models. Supplementary Figures S3 and S4 demonstrate the calibration curves of all models in the training and testing datasets. The decision curve analysis showed that the net benefit of the XGBoost model was superior to the previous scoring systems (Figure 5).

6.5. Important Features of Machine Learning Algorithm and Results of Multivariable Logistic Regression Analysis

Figure 6a shows the top 10 important features of the XGBoost model that were calculated by SHAP value in the training datasets. In the 80% pooled dataset, older age, higher FiO2 and RR, lower creatinine and HCO3, increased anion gap, lower GCS, lower BP, vasopressor use, and decreased platelet count were associated with an increased mortality rate. The predictor ranks of all the models by the mean absolute value of SHAP are shown in Figure 6b. The results of multivariable logistic regression analysis using stepwise variable selection are shown in Table S8 and were similar to those of the XGBoost model.

7. Discussion

SOFA, nonrenal SOFA, and HELENICC scores have only modest predictive value of 30-day mortality for ICU patients with AKI requiring RRT. In the first strategy, we found that the machine learning models performed better than SOFA, nonrenal SOFA, and HELENICC scores when validating models using an external validation dataset (eICU dataset). In the second strategy, the XGBoost model showed reasonable performance and a sufficiently good fit (p = 0.22) in the heterogenous dataset. Decision curve analysis indicated that the XGBoost model improved the net benefit for predicting the 30-day mortality compared with SOFA, nonrenal SOFA, and HELENICC scores. The reasons for why the XGBoost model performed better may be related to the application of regularization and high flexibility to tune hyperparameters.
General severity of illness scores, which use clinical and laboratory variables at ICU admission or even sequential data, predict mortality well for all ICU patients, like APACHE, SAPS, and SOFA scores. However, they showed poor mortality prediction for AKI patients requiring RRT [10,11,12,13]. The models targeted specifically at this population including the HELENICC score, ATN study, and Cleveland Clinic score revealed good performance in predicting mortality (AUC = 0.82, 0.85, 0.81, respectively) [13,16,29]. Those models either lacked external validation, did not perform well during external validation [15], or focused on patients with specific conditions. Notably, some variables were not readily available in the clinical datasets we used, thus limiting our ability to make a direct comparison of ML models with these scores. Prior studies have predicted mortality using machine learning algorithms, although these did not center on RRT. Brajer et al. [30] revealed excellent performance using XGBoost to predict the in-hospital mortality of adults (AUC~0.85). Another study developed models for patients with influenza infection requiring ICU admission and found that XGBoost achieved the highest AUC (0.842) [31]. Kang et al. [12] applied machine learning algorithms to predict mortality in patients requiring CRRT and found that the RF model achieved the highest AUC (0.768). This was a retrospective study in one hospital (n = 1094) and lacked external validation. The performance of our models was reasonable high for all AKI patients requiring RRT, either using the eICU dataset as an external validation dataset or using the independent part of the pooled dataset as a testing dataset to provide results that are more generalizable.
In this study, we found that the performance of validating models was better when using the eICU dataset. This may be related to the patient characteristics of the two datasets. Patients in the MIMIC dataset had a higher mortality rate and more severe comorbidities than those in the eICU dataset (Supplementary Table S5). We speculate that this could be due to demographic differences, as the eICU dataset included data from multiple ICUs in rural areas while the MIMIC dataset contained data from a single ICU in an urban medical center. Another reason may be that the data distribution of MIMIC was more complicated to classify than that of eICU (Supplementary Figure S5) using principal component analysis.
One challenge for medical researchers using machine learning algorithms is that it is difficult to assess or explain the individual contributing factors [32]. However, scientists are creating many advanced ways to make machine learning more transparent. We used the SHAP value to visualize the feature importance and determine the effect of different variables on the final output. SHAP offers not only the rank order of importance of variables but also how the variables impact the outcome, such as low creatinine associated with death risk. The results highly correlated with clinical outcome. Overall, the features generated by SHAP could be classified into hemodynamic status, central nerve system, coagulation, respiratory systems, kidney-related features, and age. The SHAP results were like the SOFA score parameters, but the performance of our models outperformed the SOFA score. That may be related to the loss of information when categorizing data, inaccurately allocating creatinine score for those AKI patients using the SOFA score, and capturing nonlinear interactions from high complexity data using machine learning algorithms. In this study, the XGBoost model used more and different variables, such as age and anion gap, which may lead to better performance than the SOFA and HELENICC scores. Besides, we retrained a new XGBoost model only using the top 10 features generated by SHAP and still achieved a good AUC (0.818; 95% CI: 0.786–0.849) in the testing dataset, which allows clinical physicians to use the model by inputting only 10 data points, can minimize the burden on them, and limit non-use in the case of missing data on a larger number of variables.
Given its improved performance over traditional severity-of-illness scoring measures, such a model or tool could potentially be used and further refined for several potential applications. For example, given its relatively high negative predictive value, it might help to enrich clinical trials for a targeted risk profile of patients. Another potential advantage of our model is that utilizes data that is easily available and routinely collected in clinical practice. This is a distinct advantage over some other prior risk scores, such as ATN score, that have been used in this population, as those included data collected as part of a research study and may not be available clinically. Having a model that can be calculated in real-time will allow clinicians to have prognostic data and help them have informed shared decision-making with the patient and their caregivers to decide whether to initiate RRT. This will also allow physicians to discuss the overall aggressiveness of care and help with medical decision-making
The strengths of the study include large sample size, external validation, curated datasets representing heterogeneous ICU populations, and the use of routinely collected clinical data. This study has several limitations. The MIMIC III dataset is old, and practice patterns may have changed. This dataset is only limited to labs when the patient is in the ICU, so it is possible that we may have missed AKI by creatinine values if they were admitted to the hospital in a non-ICU setting prior to their ICU admission. Although our datasets included a large amount of collected clinical data, some data had to be excluded due to poor-quality data recording or missing data. Using MICE to impute data may reduce predictive power. Due to many missing values, we were unable to calculate SAPS and APACHE at RTT start time. Thus, we could not make a head-to-head comparison of the performance of SAPS or APACHE with our models. In addition, the datasets did not have variables to allow for comparison with other scores that have looked specifically at this population, such as the ATN study score, which included research data. Moreover, the variables only capture data collected in the hospitals, and we are unable to capture patient mortality out of the hospitals in the eICU dataset. Our sample size may not be large enough for machine learning to have better performance. The interpretation of results should only focus on specific patients, since we excluded all transplant and advanced CKD patients in the United States. Finally, the causal relationship between the top 10 features of the XGBoost model and 30-day mortality was not clearly explored.

8. Conclusions

All machine learning models had a reasonable performance and were superior to the SOFA, nonrenal SOFA, and HELENICC scores in predicting 30-day mortality for AKI patients requiring RRT. XGBoost provided the highest performance in this study. Further prospective research is needed to validate these results prospectively and explore how they can be integrated into clinical decision-making.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm11185289/s1. Table S1: ICD-9 diagnosis codes used to identify acute kidney injury, transplant history, and comorbidities; Table S2: Percentages of missing data in the MIMIC and eICU datasets; Table S3: Baseline characteristics of the patients requiring renal replacement therapy in the MIMIC dataset; Table S4: Baseline characteristics of the patients requiring renal replacement therapy in the eICU dataset; Table S5: Comparison of baseline characteristics of the patients requiring renal replacement therapy in the eICU and MIMIC datasets; Table S6: Pairwise p value of area under ROC curves (AUROCs) of prediction models using the Delong test in the eICU dataset; Table S7: Pairwise p value of area under ROC curves (AUROCs) of prediction models using the Delong test in the 20% pooled dataset; Table S8: Significant variables in multivariate logistic regression model; Figure S1: Calibration curves of all models using MIMIC dataset as a training dataset; Figure S2: Calibration curves of all models using the eICU dataset as the testing dataset; Figure S3: Calibration curves of all models using the 80% pooled eICU and MIMIC dataset as the training dataset; Figure S4: Calibration curves of all models using the 20% pooled eICU and MIMIC datasets as the testing datasets; Figure S5: Principal component analysis (PCA) diagram reveals the similarity between the death (1) and survival (0) groups.

Author Contributions

Conceptualization, H.-H.C.; data curation, H.-H.C. and C.-S.W.; formal analysis, H.-H.C. and M.J.; funding acquisition, J.-H.C.; methodology, H.-H.C. and J.-H.C.; supervision, J.-H.C.; validation, J.-H.C.; visualization, H.-H.C.; writing—original draft, H.-H.C., J.-H.C., C.-S.W., P.-F.C., K.A.-K., H.C., J.Y. and M.J.; writing—review and editing, E.D.S., R.M., G.C. and P.M.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was supported in part by the Ministry of Science and Technology Research Grant MOST 109-2634-F-006-016. EDS was supported by the Vanderbilt O’Brien Kidney Center P30-DK114809 Clinical and Translational Core.

Institutional Review Board Statement

Both databases are previously de-identified and have been reviewed by the institutional review boards (IRB) of their hosting organizations and determined to be exempt from subsequent IRB.

Informed Consent Statement

The data descriptor for MIMIC-III states, “Requirement for individual patient consent was waived because the project did not impact clinical care and all protected health information was deidentified”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hoste, E.A.; Bagshaw, S.M.; Bellomo, R.; Cely, C.M.; Colman, R.; Cruz, D.N.; Edipidis, K.; Forni, L.G.; Gomersall, C.D.; Govil, D. Epidemiology of acute kidney injury in critically ill patients: The multinational AKI-EPI study. Intensive Care Med. 2015, 41, 1411–1423. [Google Scholar] [PubMed]
  2. Uchino, S.; Kellum, J.A.; Bellomo, R.; Doig, G.S.; Morimatsu, H.; Morgera, S.; Schetz, M.; Tan, I.; Bouman, C.; Macedo, E. Acute renal failure in critically ill patients: A multinational, multicenter study. Jama 2005, 294, 813–818. [Google Scholar] [CrossRef] [PubMed]
  3. Wald, R.; Adhikari, N.K.; Smith, O.M.; Weir, M.A.; Pope, K.; Cohen, A.; Thorpe, K.; McIntyre, L.; Lamontagne, F.; Soth, M. Comparison of standard and accelerated initiation of renal replacement therapy in acute kidney injury. Kidney Int. 2015, 88, 897–904. [Google Scholar] [CrossRef] [PubMed]
  4. Zarbock, A.; Kellum, J.A.; Schmidt, C.; Van Aken, H.; Wempe, C.; Pavenstädt, H.; Boanta, A.; Gerß, J.; Meersch, M. Effect of early vs. delayed initiation of renal replacement therapy on mortality in critically ill patients with acute kidney injury: The ELAIN randomized clinical trial. JAMA 2016, 315, 2190–2199. [Google Scholar]
  5. Moreno, R.P.; Metnitz, P.G.; Almeida, E.; Jordan, B.; Bauer, P.; Campos, R.A.; Iapichino, G.; Edbrooke, D.; Capuzzo, M.; Le Gall, J.-R. SAPS 3—From evaluation of the patient to evaluation of the intensive care unit. Part 2: Development of a prognostic model for hospital mortality at ICU admission. Intensive Care Med. 2005, 31, 1345–1355. [Google Scholar] [CrossRef] [PubMed]
  6. Nassar, A.P.; Malbouisson, L.S.; Moreno, R. Evaluation of simplified acute physiology score 3 performance: A systematic review of external validation studies. Crit. Care 2014, 18, R117. [Google Scholar] [CrossRef]
  7. Zimmerman, J.E.; Kramer, A.A.; McNair, D.S.; Malila, F.M. Acute Physiology and Chronic Health Evaluation (APACHE) IV: Hospital mortality assessment for today’s critically ill patients. Crit. Care Med. 2006, 34, 1297–1310. [Google Scholar]
  8. Falcão, A.L.E.; de Almeida Barros, A.G.; Bezerra, A.A.M.; Ferreira, N.L.; Logato, C.M.; Silva, F.P.; do Monte, A.B.F.O.; Tonella, R.M.; de Figueiredo, L.C.; Moreno, R. The prognostic accuracy evaluation of SAPS 3, SOFA and APACHE II scores for mortality prediction in the surgical ICU: An external validation study and decision-making analysis. Ann. Intensive Care 2019, 9, 18. [Google Scholar] [CrossRef]
  9. Soo, A.; Zuege, D.J.; Fick, G.H.; Niven, D.J.; Berthiaume, L.R.; Stelfox, H.T.; Doig, C.J. Describing organ dysfunction in the intensive care unit: A cohort study of 20,000 patients. Crit. Care 2019, 23, 186. [Google Scholar] [CrossRef]
  10. Czempik, P.; Cieśla, D.; Knapik, P.; Krzych, Ł. Mortality of patients with acute kidney injury requiring renal replacement therapy. Adv. Clin. Exp. Med. 2018, 27, 327–333. [Google Scholar] [CrossRef]
  11. Maccariello, E.; Valente, C.; Nogueira, L.; Bonomo, H.; Ismael, M.; Machado, J.E.; Baldotto, F.; Godinho, M.; Valença, R.; Rocha, E. SAPS 3 scores at the start of renal replacement therapy predict mortality in critically ill patients with acute kidney injury. Kidney Int. 2010, 77, 51–56. [Google Scholar] [PubMed]
  12. Kang, M.W.; Kim, J.; Kim, D.K.; Oh, K.-H.; Joo, K.W.; Kim, Y.S.; Han, S.S. Machine learning algorithm to predict mortality in patients undergoing continuous renal replacement therapy. Crit. Care 2020, 24, 42. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Demirjian, S.; Chertow, G.M.; Zhang, J.H.; O’Connor, T.Z.; Vitale, J.; Paganini, E.P.; Palevsky, P.M.; VA/NIH Acute Renal Failure Trial Network. Model to predict mortality in critically ill adults with acute kidney injury. Clin. J. Am. Soc. Nephrol. 2011, 6, 2114–2120. [Google Scholar] [CrossRef] [PubMed]
  14. Li, D.H.; Wald, R.; Blum, D.; McArthur, E.; James, M.T.; Burns, K.E.; Friedrich, J.O.; Adhikari, N.K.; Nash, D.M.; Lebovic, G. Predicting mortality among critically ill patients with acute kidney injury treated with renal replacement therapy: Development and validation of new prediction models. J. Crit. Care 2020, 56, 113–119. [Google Scholar]
  15. de Carvalho, G.M.C.; Leite, T.T.; Libório, A.B. Prediction of 60-day case fatality in critically ill patients receiving renal replacement therapy: External validation of a prediction model. Shock 2018, 50, 156–161. [Google Scholar] [CrossRef]
  16. da Hora Passos, R.; Ramos, J.G.R.; Mendonça, E.J.B.; Miranda, E.A.; Dutra, F.R.D.; Coelho, M.F.R.; Pedroza, A.C.; Correia, L.C.L.; Batista, P.B.P.; Macedo, E. A clinical score to predict mortality in septic acute kidney injury patients requiring continuous renal replacement therapy: The HELENICC score. BMC Anesthesiol. 2017, 17, 21. [Google Scholar] [CrossRef]
  17. Johnson, A.E.; Pollard, T.J.; Shen, L.; Li-wei, H.L.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Celi, L.A.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar]
  18. Pollard, T.J.; Johnson, A.E.; Raffa, J.D.; Celi, L.A.; Mark, R.G.; Badawi, O. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci. Data 2018, 5, 180178. [Google Scholar]
  19. Vincent, J.-L.; Moreno, R.; Takala, J.; Willatts, S.; De Mendonça, A.; Bruining, H.; Reinhart, C.; Suter, P.; Thijs, L.G. The SOFA (Sepsis-Related Organ Failure Assessment) Score to Describe Organ Dysfunction/Failure; Springer: Berlin, Germany, 1996. [Google Scholar]
  20. Palevsky, P.M.; O’Connor, T.; Zhang, J.H.; Star, R.A.; Smith, M.W. Design of the VA/NIH Acute Renal Failure Trial Network (ATN) Study: Intensive versus conventional renal support in acute renal failure. Clin. Trials 2005, 2, 423–435. [Google Scholar] [CrossRef]
  21. Hsu, C.-y.; Chertow, G.M.; McCulloch, C.E.; Fan, D.; Ordoñez, J.D.; Go, A.S. Nonrecovery of kidney function and death after acute on chronic renal failure. Clin. J. Am. Soc. Nephrol. 2009, 4, 891–898. [Google Scholar] [CrossRef]
  22. Royston, P.; White, I.R. Multiple imputation by chained equations (MICE): Implementation in Stata. J. Stat. Softw. 2011, 45, 1–20. [Google Scholar] [CrossRef]
  23. Moons, K.G.; Altman, D.G.; Reitsma, J.B.; Ioannidis, J.P.; Macaskill, P.; Steyerberg, E.W.; Vickers, A.J.; Ransohoff, D.F.; Collins, G.S. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): Explanation and elaboration. Ann. Intern. Med. 2015, 162, W1–W73. [Google Scholar]
  24. Chen, T.; Guestrin, C. Xgboost: A Scalable Tree Boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  25. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  26. Sarica, A.; Cerasa, A.; Quattrone, A. Random forest algorithm for the classification of neuroimaging data in Alzheimer’s disease: A systematic review. Front. Aging Neurosci. 2017, 9, 329. [Google Scholar] [CrossRef] [PubMed]
  27. Rana, A.; Rawat, A.S.; Bijalwan, A.; Bahuguna, H. Application of Multi Layer (Perceptron) Artificial Neural Network in the Diagnosis System: A Systematic Review. In Proceedings of the 2018 International Conference on Research in Intelligent and Computing in Engineering (RICE), San Salvador, El Salvador, 22–24 August 2018; pp. 1–6. [Google Scholar]
  28. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; Volume 30, pp. 4765–4774. [Google Scholar]
  29. Thakar, C.V.; Arrigain, S.; Worley, S.; Yared, J.-P.; Paganini, E.P. A clinical score to predict acute renal failure after cardiac surgery. J. Am. Soc. Nephrol. 2005, 16, 162–168. [Google Scholar] [CrossRef] [PubMed]
  30. Brajer, N.; Cozzi, B.; Gao, M.; Nichols, M.; Revoir, M.; Balu, S.; Futoma, J.; Bae, J.; Setji, N.; Hernandez, A. Prospective and external evaluation of a machine learning model to predict in-hospital mortality of adults at time of admission. JAMA Netw. Open 2020, 3, e1920733. [Google Scholar] [PubMed] [Green Version]
  31. Hu, C.-A.; Chen, C.-M.; Fang, Y.-C.; Liang, S.-J.; Wang, H.-C.; Fang, W.-F.; Sheu, C.-C.; Perng, W.-C.; Yang, K.-Y.; Kao, K.-C. Using a machine learning approach to predict mortality in critically ill influenza patients: A cross-sectional retrospective multicentre study in Taiwan. BMJ Open 2020, 10, e033898. [Google Scholar] [CrossRef]
  32. Price, W.N. Big data and black-box medical algorithms. Sci. Transl. Med. 2018, 10, eaao5333. [Google Scholar] [CrossRef]
Figure 1. Model development and validation flowchart.
Figure 1. Model development and validation flowchart.
Jcm 11 05289 g001
Figure 2. Participant flow diagram. n is patient unit encounter.
Figure 2. Participant flow diagram. n is patient unit encounter.
Jcm 11 05289 g002
Figure 3. Area under operating characteristic (ROC) curves for prediction models trained in MIMIC (a) and tested in eICU (b). Red, RF (0.783, 0.816); Green, XGBoost (0.793, 0.812); Blue, MLP (0.785, 0.810); Orange, LR (0.786, 0.815); Brown, nonrenal SOFA score (0.728, 0.769); Purple, SOFA score (0.717, 0.749); Pink, HELENICC score (0.694, 0.756).
Figure 3. Area under operating characteristic (ROC) curves for prediction models trained in MIMIC (a) and tested in eICU (b). Red, RF (0.783, 0.816); Green, XGBoost (0.793, 0.812); Blue, MLP (0.785, 0.810); Orange, LR (0.786, 0.815); Brown, nonrenal SOFA score (0.728, 0.769); Purple, SOFA score (0.717, 0.749); Pink, HELENICC score (0.694, 0.756).
Jcm 11 05289 g003
Figure 4. Area under operating characteristic (ROC) curves for prediction models trained in 80% pooled data (a) and tested in 20% pooled data (b). Red, RF (0.809, 0.821); Green, XGBoost (0.814, 0.823); Blue, MLP (0.818, 0.784); Orange, LR (0.814, 0.819); Brown, nonrenal SOFA score (0.735, 0.726); Purple, SOFA score (0.718, 0.710); Pink, HELENICC score (0.735, 0.752).
Figure 4. Area under operating characteristic (ROC) curves for prediction models trained in 80% pooled data (a) and tested in 20% pooled data (b). Red, RF (0.809, 0.821); Green, XGBoost (0.814, 0.823); Blue, MLP (0.818, 0.784); Orange, LR (0.814, 0.819); Brown, nonrenal SOFA score (0.735, 0.726); Purple, SOFA score (0.718, 0.710); Pink, HELENICC score (0.735, 0.752).
Jcm 11 05289 g004
Figure 5. Decision curve analysis of random forest, SOFA, nonrenal SOFA, and HELENICC for 30-day mortality prediction.
Figure 5. Decision curve analysis of random forest, SOFA, nonrenal SOFA, and HELENICC for 30-day mortality prediction.
Jcm 11 05289 g005
Figure 6. (a) Top 10 important features of the XGBoost model by SHAP value. The X-axis shows the distribution of the SHAP value, which reveals how much it impacts the outcome. The Y-axis is the feature value. The color represents the positive or negative contribution to outcome (red prone to death, blue prone to survival). (b) Ranks of the predictors by the mean absolute value of the SHAP values.
Figure 6. (a) Top 10 important features of the XGBoost model by SHAP value. The X-axis shows the distribution of the SHAP value, which reveals how much it impacts the outcome. The Y-axis is the feature value. The color represents the positive or negative contribution to outcome (red prone to death, blue prone to survival). (b) Ranks of the predictors by the mean absolute value of the SHAP values.
Jcm 11 05289 g006
Table 1. Baseline characteristics of the patients requiring renal replacement therapy between the training and testing datasets of pooled data.
Table 1. Baseline characteristics of the patients requiring renal replacement therapy between the training and testing datasets of pooled data.
VariablesTraining DatasetTesting Datasetp Value
Number of patients2729683
Death %35.8%37.8%0.369
Demographics
Age, years62.9 ± 15.062.7 ± 14.90.793
Male sex, % 59.5% 59.7% 0.962
Black race, % 16.8% 18% 0.149
Comorbidities (%)
Diabetes mellitus 21.6% 21.1%0.800
Hypertension 25.9% 25.0%0.677
CHF 24.8% 26.4%0.445
CKD 17.5% 16.7%0.667
Malignancy 6.7% 6.9%0.910
Liver cirrhosis 10.6% 9.7%0.523
Days of ICU stay before RRT initiation2.9 ± 4.83.1 ± 5.10.223
Diuretics, % 14.4% 12.0% 0.126
Vasopressor, % 36.5% 38.4% 0.381
Mechanical ventilation, % 72.4% 74.7%<0.254
Laboratory variables
BUN (mg/dL)56.0 (36.0–84.0)61.0 (38.0–89.0) 0.01 *
FiO2 (%) 49.5 ± 26.8 48.7 ± 26.3 0.501
HCO3 (mmol/L) 20.5 ± 5.7 20.3 ± 5.8 0.464
Hgb (mg/dL) 9.6 (8.5–10.8) 9.8 (8.6–11.1) 0.02 *
O2 Sat (%) 93.7 ± 8.0 93.6 ± 8.0 0.767
WBC count (×1000/μL) 15.8 ± 26.2 15.8 ± 14.6 0.976
Anion gap (mmol/L) 16.5 ± 6.7 16.9 ± 6.4 0.155
Calcium (mg/dL) 8.2 ± 1.1 8.2 ± 1.1 0.959
Creatinine (mg/dL) 4.6 ± 3.2 4.6 ± 2.9 0.713
Glucose (mg/dL) 146.0 ± 65.1 152.8 ± 87.7 0.024
Platelet count (×1000/μL) 182.1 ± 115.1 175.7 ± 106 0.193
Potassium (mmol/L) 4.7 ± 1.0 4.7 ± 1.0 0.349
Sodium (mmol/L) 137.4 ± 6.1 137.5 ± 6.1 0.537
GCS score 11.1 ± 4.1 11.0 ± 4.0 0.381
MAP (mmHg) 76.2 ± 15.0 75.3 ± 14.6 0.170
HR (beats per minute) 89.4 ± 18.4 90.0 ± 18.7 0.444
RR (breaths per minute) 20.9 ± 5.6 21.1 ± 5.4 0.374
Abbreviations: CHF, congestive heart failure; CKD, chronic kidney disease; BUN, blood urea nitrogen; FiO2, fraction of inspired oxygen; Hgb, hemoglobin; WBC, white blood cell; GCS, Glasgow Coma Scale; HR, heart rate; MAP, mean arterial pressure; RR, respiratory rate; SI, shock index; ICU, intensive care unit; RRT, renal replacement therapy; CRRT, continuous renal replacement therapy; IHD, intermittent hemodialysis; MV, mechanical ventilation. Data are expressed as n (%) for categorical data and as mean ± standard deviation or median (interquartile range) for continuous data. * Mann–Whitney U test.
Table 2. Performance of the SOFA, nonrenal SOFA, and HELENICC scores 1 day before the beginning of dialysis in the MIMIC and eICU datasets.
Table 2. Performance of the SOFA, nonrenal SOFA, and HELENICC scores 1 day before the beginning of dialysis in the MIMIC and eICU datasets.
DatasetMIMICeICU
ModelSOFANonrenal SOFAHELENICCSOFANonrenal SOFAHELENICC
AUC0.7170.7280.6940.7490.7690.756
95% CI0.687–0.7470.699–0.7580.664–0.7520.728–0.7700.749–0.7890.735–0.776
Sensitivity0.5140.5280.4010.3720.4460.341
Specificity0.7980.7920.8450.8840.8680.914
PPV0.6560.6560.6590.6120.6250.662
NPV0.6870.6910.6530.7410.7610.738
Accuracy0.6760.6790.6540.7150.7290.725
Abbreviations: PPV, positive predictive value; NPV, negative predictive value.
Table 3. Model performance measures in the MIMIC and eICU datasets.
Table 3. Model performance measures in the MIMIC and eICU datasets.
Training DatasetMIMIC
ModelLRXGBoostRFMLP
AUC0.7860.7930.7830.785
95% CI0.752–0.8200.760–0.8260.743–0.8220.752–0.819
Sensitivity0.5780.6190.6210.617
Specificity0.8090.8030.8000.779
PPV0.6940.7020.7000.678
NPV0.7190.7370.7380.731
Accuracy0.7100.7240.7230.710
Hosmer-Lemeshow test<0.050.02<0.050.44
Testing DataseteICU
ModelLRXGBoostRFMLP
AUC0.8150.8120.8160.810
95% CI0.797–0.8330.794–0.8300.798–0.8340.792–0.828
Sensitivity0.4400.4880.5950.489
Specificity0.9050.8920.8370.885
PPV0.6950.6910.6420.677
NPV0.7670.7800.8080.779
Accuracy0.7520.7590.7570.755
Hosmer-Lemeshow test<0.05<0.05<0.050.29
Table 4. Model performance measures in the pooled dataset.
Table 4. Model performance measures in the pooled dataset.
Training Dataset80% Pooled Data
ModelLRXGBoostRFMLP
AUC0.8140.8140.8090.818
95% CI0.797–0.8310.800–0.8280.796–0.8220.802–0.833
Sensitivity0.5740.5840.5560.644
Specificity0.8450.8330.8530.825
PPV0.6740.6620.6790.672
NPV0.7800.7820.7740.805
Accuracy0.7480.7440.7460.760
Hosmer-Lemeshow test<0.050.020.08<0.05
Testing Dataset20% Pooled Data
ModelLRXGBoostRFMLP
AUC0.8190.8230.8210.784
95% CI0.787–0.8510.791–0.8540.790–0.8520.750–0.817
Sensitivity0.6200.6350.5620.662
Specificity0.8040.8320.8630.785
PPV0.6580.6970.7140.652
NPV0.7770.7900.7640.793
Accuracy0.7340.7580.7490.739
Hosmer-Lemeshow test0.110.220.17<0.05
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chang, H.-H.; Chiang, J.-H.; Wang, C.-S.; Chiu, P.-F.; Abdel-Kader, K.; Chen, H.; Siew, E.D.; Yabes, J.; Murugan, R.; Clermont, G.; et al. Predicting Mortality Using Machine Learning Algorithms in Patients Who Require Renal Replacement Therapy in the Critical Care Unit. J. Clin. Med. 2022, 11, 5289. https://doi.org/10.3390/jcm11185289

AMA Style

Chang H-H, Chiang J-H, Wang C-S, Chiu P-F, Abdel-Kader K, Chen H, Siew ED, Yabes J, Murugan R, Clermont G, et al. Predicting Mortality Using Machine Learning Algorithms in Patients Who Require Renal Replacement Therapy in the Critical Care Unit. Journal of Clinical Medicine. 2022; 11(18):5289. https://doi.org/10.3390/jcm11185289

Chicago/Turabian Style

Chang, Hsin-Hsiung, Jung-Hsien Chiang, Chi-Shiang Wang, Ping-Fang Chiu, Khaled Abdel-Kader, Huiwen Chen, Edward D. Siew, Jonathan Yabes, Raghavan Murugan, Gilles Clermont, and et al. 2022. "Predicting Mortality Using Machine Learning Algorithms in Patients Who Require Renal Replacement Therapy in the Critical Care Unit" Journal of Clinical Medicine 11, no. 18: 5289. https://doi.org/10.3390/jcm11185289

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop