Next Article in Journal
Effects of Biodanza® SRT on Motor, Cognitive, and Behavioral Symptoms in Patients with Parkinson’s Disease: A Randomized Controlled Study
Previous Article in Journal
Shoulder Dystocia: A Comprehensive Literature Review on Diagnosis, Prevention, Complications, Prognosis, and Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Internal and External Validation of Machine Learning Models for Predicting Acute Kidney Injury Following Non-Cardiac Surgery Using Open Datasets

1
Department of Anesthesiology and Pain Medicine, Asan Medical Center, University of Ulsan College of Medicine, Seoul 05505, Republic of Korea
2
Biomedical Engineering Research Center, Biosignal Analysis & Perioperative Outcome Research (BAPOR) Laboratory, Asan Institute for Lifesciences, Seoul 05505, Republic of Korea
3
Department of Anesthesiology and Pain Medicine, Brain Korea 21 Project, University of Ulsan College of Medicine, Seoul 05505, Republic of Korea
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2024, 14(6), 587; https://doi.org/10.3390/jpm14060587
Submission received: 7 May 2024 / Revised: 27 May 2024 / Accepted: 28 May 2024 / Published: 30 May 2024
(This article belongs to the Section Clinical Medicine, Cell, and Organism Physiology)

Abstract

:
This study developed and validated a machine learning model to accurately predict acute kidney injury (AKI) after non-cardiac surgery, aiming to improve patient outcomes by assessing its clinical feasibility and generalizability. We conducted a retrospective cohort study using data from 76,032 adults who underwent non-cardiac surgery at a single tertiary medical center between March 2019 and February 2021, and used data from 5512 patients from the VitalDB open dataset for external model validation. The predictive variables for model training consisted of demographic, preoperative laboratory, and intraoperative data, including calculated statistical values such as the minimum, maximum, and mean intraoperative blood pressure. When predicting postoperative AKI, our gradient boosting machine model incorporating all the variables achieved the best results, with AUROC values of 0.868 and 0.757 for the internal and external validations using the VitalDB dataset, respectively. The model using intraoperative data performed best in internal validation, while the model with preoperative data excelled in external validation. In this study, we developed a predictive model for postoperative AKI in adult patients undergoing non-cardiac surgery using preoperative and intraoperative data, and external validation demonstrated the efficacy of open datasets for generalization in medical artificial modeling research.

1. Introduction

Acute kidney injury (AKI) is a condition characterized by a rapid decline in kidney function and can be caused by a variety of factors [1]. In particular, postoperative patients are known to be at a higher risk of developing AKI than the average non-surgical hospitalized patient [2]. The occurrence of postoperative AKI not only prolongs the patient’s hospitalization duration but also leads to escalated healthcare expenses [3,4]. Furthermore, postoperative AKI can impact the long-term outcome for surgical patients by elevating postoperative morbidity and mortality rates, potentially leading to the development of chronic kidney disease in the future [5,6,7]. Predicting the occurrence of postoperative AKI before surgery enables the allocation of adequate postoperative medical resources to surgical patients, ultimately enhancing patient outcomes and mitigating healthcare costs [8,9]. However, to date, a systematic approach for predicting postoperative AKI in patients undergoing non-cardiac surgery has not been well established. We intend to achieve this by developing a predictive model using machine learning (ML) techniques, which have recently been actively applied in the medical field. Whereas previous artificial intelligence (AI) studies have focused on predicting post-surgery AKI, most relied on data from single institutions, which limited their generalizability [9,10,11,12]. To overcome this limitation, our study aimed to explore the model’s feasibility and utility across different institutions by externally validating it using an openly accessible dataset. In medical AI research, the utilization of open datasets for external validation is a crucial aspect of model development, offering a highly practical and viable alternative. However, because of the challenges associated with sharing and accessing medical data over different time periods, obtaining external datasets for validation remains a formidable task. Therefore, we anticipate that research focused on leveraging open datasets for external model validation will continue to thrive in the future.

2. Materials and Methods

2.1. Ethical Statement and Study Data

This study utilized the data from 76,032 adult patients aged 18 years and older who underwent non-cardiac surgery at Asan Medical Center (AMC) from March 2019 to February 2021, with the data extracted from the hospital’s Electronic Health Record (EHR) system. In addition, we utilized the open dataset VitalDB for the external validation of the model. Out of the total of 6388 patients in VitalDB, we extracted data for 5512 patients to use as the external validation dataset for our prediction model (Figure S1). The study received approval from the Institutional Review Board (IRB) of Asan Medical Center (IRB No. 2024-0060), and informed consent from participants was waived based on the retrospective collection and secondary utilization of EHR data. We excluded pediatric patients under the age of 18, individuals who underwent cardiac surgery, and those who were diagnosed with end-stage renal disease or had preoperative creatinine levels exceeding 4.5 mg/dL (Figure 1).

2.2. Variables for Modeling and Preprocessing Data

The variables employed in the modeling of this study encompassed demographic information, preoperative data, and intraoperative data (Table S1). The demographic data included the patient’s age, gender, body mass index (BMI), and American Society of Anesthesiologists (ASA) class. The preoperative data included preoperative laboratory test results such as the white blood cell, hemoglobin, hematocrit, platelet, sodium, potassium, chloride, total bilirubin, albumin, aspartate aminotransferase (AST), and alanine aminotransferase (ALT) results, along with the estimated glomerular filtration rate (eGFR), glucose level, prothrombin time (PT), activated partial thromboplastin time (aPTT), blood urea nitrogen (BUN) level, creatinine level, and C-reactive protein (CRP) level. The intraoperative data included the arterial blood pressure values (systolic blood pressure (SBP), diastolic blood pressure (DBP), and mean blood pressure (MBP)) measured through arterial catheterization, non-invasive blood pressure values (SBP, DBP, and MBP), estimated blood loss, anesthetic time, and surgery time. In our study, anesthetic time refers to the total duration from the induction of anesthesia to the moment of extubation and the patient’s return to consciousness. Surgery time denotes the period from the initial skin incision by the surgeon to the completion of the surgery. The intraoperative blood pressure data were employed as input variables for the model by computing the maximum, minimum, mean, and standard deviation values of the blood pressure measurements, along with the sum of the changes in the blood pressure differences. Intraoperative monitoring measured blood pressure every 3 to 5 min using non-invasive methods, while arterial catheterization provided continuous real-time blood pressure data. However, the anesthesia records analyzed in our study only documented blood pressure readings at 5 min intervals, so more frequent measurements were not accessible. Intraoperative blood pressure data, which is automatically linked to the electronic medical record, often contains artifacts. We removed these artifacts by using the algorithm presented in our previous study [13]. Blood pressure measurements exceeding the following thresholds were excluded from the analysis: a systolic blood pressure greater than 300 mm Hg or less than 20 mm Hg, diastolic blood pressure greater than 225 mm Hg or less than 5 mm Hg, and systolic blood pressure less than the diastolic blood pressure + 5 mm Hg, in accordance with the criteria proposed in a previous study [13,14]. Since the missing values in the modeling dataset did not exhibit any specific pattern or correlation, we addressed them by inputting the missing values with the mean value of each variable in the dataset (Figures S2 and S3).

2.3. Open Dataset for External Validation

In our study, we utilized an open dataset known as VitalDB to externally validate the performance of our prediction model [15]. VitalDB comprises preoperative data from 6388 surgical patients who underwent non-cardiac surgery at Seoul National University Hospital, including intraoperative monitoring parameters. Since AKI, the primary outcome of our study, relies on creatinine blood test results, we chose VitalDB as our external validation dataset because it provides both preoperative and postoperative blood laboratory test data. Furthermore, VitalDB is publicly accessible, allowing researchers to download it from the internet without the need for institutional IRB approval. In this regard, it proves valuable as an external validation dataset for a range of clinical studies.

2.4. Primary Outcome

The primary outcome in our study was the occurrence of AKI within 7 days following surgery. AKI is defined as either a 1.5-fold increase in the creatinine level from the preoperative baseline or a rise of 0.3 mg/dL over a 48 h period, as measured within the first 7 days after surgery, in accordance with the definition from the Kidney Disease Improving Global Outcomes (KDIGO) guidelines. If no creatinine level was measured postoperatively, it was assumed that an AKI had not occurred.

2.5. Modeling and Model Evaluation

In this study, the dataset variables used for modeling were categorized into demographic, preoperative, and intraoperative datasets. We compared the performances of models based on the individual datasets as well as various combinations of dataset configurations. We divided the datasets for modeling into training, testing, and validation datasets in a ratio of 6:2:2. In order to mitigate any potential overfitting of the model performance stemming from the specific composition of the training dataset, we employed the bootstrapping method for model evaluations by resampling each dataset multiple times. We here present the results of evaluations of the models in the form of the mean values and statistical confidence intervals for the performance outcomes of the trained models on each resampled dataset. In our study, we employed four predictive modeling methods: logistic regression (LR), random forest (RF), gradient boosting machine (GBM), and a simple deep neural network (DNN) with five layers. We configured the hyperparameters for our modeling techniques as follows. For LR, default values were used for the L2 penalty and “lbfgs” solver. For RF and GBM, we used a grid search approach to optimize the hyperparameters. Our DNN model comprised five hidden layers, along with the incorporation of batch normalization and a dropout rate of 0.5. The dense layers utilized rectified linear unit (ReLU) activation functions, and the final output layer employed a sigmoid activation function for binary classification. We applied a learning rate of 0.001, employed binary cross-entropy as the loss function, and utilized the adam optimizer. In our study, we primarily assessed the prediction model’s performance using two key metrics: the area under the receiver operating characteristic (AUROC) curve and area under the precision–recall curve (AUPRC). We employed the DeLong test to establish the statistical significance of the performance differences. Furthermore, we identified and compared the crucial variables influencing the performance of each model using Shapley additive explanation (SHAP) values.

2.6. Statistical Analysis and Modeling Tools

To statistically describe the dataset characteristics, the data distributions of continuous variables are depicted using mean and standard deviation values, whereas categorical variables are presented using counts and proportions. To compare continuous variables across datasets, t-tests were employed, whereas chi-squared tests were used for categorical variables. The descriptive statistics were obtained using R 3.4.3, whereas ML predictive modeling and model performance evaluations were conducted using Python 3.11.

3. Results

3.1. Study Data Characteristics

Table 1 presents the results of a comparative analysis of the attributes of research data extracted from two distinct datasets: the AMC data (an internal dataset) and VitalDB data (an external dataset). Initially, there does not seem to be any significant disparities in demographic variables such as age, gender, and BMI between the two datasets (Table 1). Nevertheless, upon closer examination of the surgical department information, it becomes apparent that the external dataset exhibits a higher prevalence of general and thoracic surgeries in contrast to the internal dataset (Table 1). Among the 76,032 adult patients aged 18 years and older who underwent general anesthesia for non-cardiac surgery at AMCs, a total of 2314 (3.1%) patients experienced postoperative AKI. Conversely, for the external validation dataset from VitalDB, which included 5512 patients who did not meet the exclusion criteria, 78 (1.5%) developed postoperative AKI (Table 2). The percentage of patients with a hospital stay of 7 days or more was higher for the external dataset, at 57.2%, in contrast to 38.5% for the internal dataset (Table 2). Mortality within 30 days of surgery did not exhibit a significant difference between the two datasets; however, postoperative in-hospital mortality was higher for the internal dataset, registering at 2.1% compared to 0.9% for the external dataset (Table 2). Additionally, the rate of postoperative intensive care unit (ICU) admissions for the internal dataset was nearly twice as high, standing at 18.3% compared to 9.5% for the external dataset (Table 2). For our internal dataset, we conducted a comparison of clinical outcomes between the AKI and non-AKI groups, and as anticipated, the AKI group exhibited a higher incidence of extended hospital stays, postoperative in-hospital mortality, and postoperative ICU admissions compared to the non-AKI group.

3.2. Predictive Performance Results of Internal and External Validations

The internal validation results for the AKI prediction model showed that the GBM algorithm had the highest prediction performance when using all the datasets, with an AUROC of 0.868, AUPRC of 0.786, and F1-score of 0.723 (Figure 2 and Table 3). The GBM algorithm also had the highest performance when utilizing the intraoperative dataset, with an AUROC of 0.816 and AUPRC of 0.715, as compared to the prediction model performances when using the demographic and preoperative datasets (Table 3). The external validation results for the AKI model on the VitalDB dataset showed the highest prediction performance, with an AUROC of 0.769 and AUPRC of 0.696, was achieved by the GBM algorithm when using the preoperative dataset (Table 4). Furthermore, the LR algorithm exhibited the highest prediction performance when utilizing the entire dataset, with an F1-score of 0.627 (Table 4).

3.3. Feature Importance

An analysis of the feature importance of the AKI prediction model using SHAP values on the internal dataset showed that total surgery time and intraoperative blood pressure data were highly significant, whereas the gender status from the demographic data, and albumin and creatinine levels in the preoperative laboratory test data, played pivotal roles in the prediction performance (Figure 3).

4. Discussion

In our study, we developed a model for predicting postoperative AKI by utilizing both preoperative and intraoperative data. We subsequently conducted an external validation of its performance using a publicly available dataset from a distinct institution, thus highlighting the generalizability of our predictive model. Furthermore, by utilizing a substantial EHR dataset, our study demonstrated the potential to predict the occurrence of AKI—a significant postoperative complication. We achieved this using automatically generated EHR data that could be seamlessly integrated with a hospital’s electronic medical record (EMR) system. This integration would serve the dual purposes of alerting medical personnel to the risk in advance, thereby facilitating the provision of intensive care to patients at risk, and assisting in optimizing the allocation of hospital resources by appropriately identifying and reclassifying these at-risk patients.
AKI is a major postoperative complication, contributing significantly to prolonged hospital stays, increased postoperative morbidity, and mortality in patients [1,3,4,16,17,18]. The incidence of postoperative AKI varies widely based on the type of surgery, with surgical factors being some of the most common causes [2]. Whereas AKI is generally less frequent in non-cardiac surgeries as compared to cardiac procedures, its occurrence in the former can present greater prediction challenges, potentially leading to delayed detection and treatment, and consequently worsening the patient’s prognosis. Postoperative AKI is associated with prolonged postoperative mechanical ventilation, extended intensive care hospitalization, and increased treatment requirements, all of which negatively impact the patient’s postoperative outcome [1,3,4,6,7,16,17,18,19,20,21]. Therefore, the ability to predict an AKI before surgery is crucial for efficiently allocating hospital resources to at-risk patients and directing medical resources to high-risk patients, ultimately leading to improved patient outcomes [8,9,22]. Previous studies have explored the use of ML algorithms to predict the occurrence of postoperative AKI [8,9,11,23,24,25,26,27]. Lei et al. aimed to predict the risk of postoperative AKI in non-cardiac surgery by utilizing preoperative and intraoperative blood pressure data [9]. Nevertheless, a limitation of this study was its reliance solely on data from a single institution. This limitation hindered the ability to assess the generalizability of their findings to other healthcare facilities and also made it challenging to evaluate the potential overfitting issue in the prediction model. Overfitting is a common concern in AI research, particularly in studies that solely rely on data from a single institution. Therefore, for the developed ML prediction model to hold significance in terms of reproducibility and generalizability, it needed to undergo validation using data from other institutions to assess its predictive performance in an external context. An external validation of the effectiveness of our ML prediction model was performed by utilizing the VitalDB dataset, which is an open dataset containing information on surgical patients [15].
One noteworthy aspect of our study was our approach to validation. Unlike previous studies that predominantly relied on single-institution data for internal validation, we took a different path by extensively validating our model’s performance against external institutional datasets. Whereas it is recognized that external validation may yield slightly less predictive accuracy compared to internal validation, our research underscored that the predictive performance achieved through external validation remained notably strong, particularly when compared to findings from prior studies. This pivotal finding highlighted our study’s distinctive contribution in addressing a common limitation observed in AI model research, namely the limited generalizability of models trained on data from a single institution, which results in their suboptimal performance on datasets from other institutions.
One of the paramount considerations in AI prediction research involving medical data is to ensure that the developed model avoids overfitting to the training data. This step is crucial for enhancing the model’s applicability in real-world clinical scenarios, making the outcomes of such studies more practically relevant [28,29]. Consequently, we believe that external validation, employing data from external sources, is an essential and vital phase in medical AI prediction research. Nevertheless, acquiring data from external organizations often proves to be a substantial challenge. This difficulty stems from the various data regulations governing these organizations, which impose stringent restrictions on the exchange and sharing of large-scale data across different time periods. As a result, a practical alternative is to evaluate predictive models using single-institution data on publicly available open datasets. This approach offers a valuable means of conducting external validation in AI prediction research within real-world healthcare settings.
Our study had several strengths compared to prior research. First, we considered not only the basic statistical metrics of intraoperative blood pressure, which is a critical parameter, but also changes in blood pressure data. Additionally, we incorporated factors like intraoperative blood loss and surgery time information into our prediction model. Consequently, we enhanced the postoperative AKI prediction performance, achieving an AUC of 0.87, which surpassed the previous study’s AUC of 0.82 [8,9]. The external validation also demonstrated strong predictive performance, with an AUC of 0.769.
Our study had several limitations that warrant consideration in future research. First, we did not use the complete intraoperative blood pressure (BP) dataset for the predictive modeling; instead, we utilized statistical summaries of the BP data. Therefore, the modeling process did use all the BP data but substituted a statistical representation derived from predictive modeling using the complete blood pressure dataset. Future research should emphasize the application of predictive modeling techniques such as transformers and recurrent neural networks (RNN), which are known for their effectiveness in handling time series data like blood pressure readings. The goal should be to develop models capable of capturing all of the blood pressure data. Moreover, it is imperative to validate the applicability of predictive models across diverse real-world organizations using a more comprehensive set of publicly available datasets. Prospective studies are also essential to assess the practical value of these predictive models when deployed in real clinical settings. Ultimately, such prospective investigations can offer valuable insights into the clinical utility of AI predictive models.

5. Conclusions

In conclusion, we successfully developed and implemented a prediction model for postoperative AKI in adult patients undergoing non-cardiac surgery by using a dataset that could automatically be extracted from an EHR system. Furthermore, we conducted an external validation of our prediction model using a separate institutional dataset, which was an open dataset, thereby confirming the reproducibility and generalizability of our model. In addition, our study underscored the viability of open datasets as a valuable alternative to third-party datasets for the external validation of prediction models in the field of medical AI research. As additional diverse datasets become available in the future, it is anticipated that more researchers will leverage them for validating a range of AI prediction models.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jpm14060587/s1. Table S1. Variables for modeling; Figure S1. Study flow chart of open dataset; Figure S2. Distribution plots of missing values (A) Distribution plot of missing values in internal dataset. (B) Distribution plot of missing values in external dataset; Figure S3. Nullity correlation heatmaps (A) Nullity correlation heatmap for internal dataset. (B) Nullity correlation heatmap for external dataset; Figure S4. Distribution plots of various blood pressure variables in internal and external datasets; Figure S5. AUROC and AUPRC values of prediction models across different modeling methods according to the various feature selections (A) Demographic dataset. (B) Preoperative dataset. (C) Intraoperative dataset; Figure S6. SHAP value summary plots generated from the GBM models according to the various feature selections (A) Demographic dataset. (B) Preoperative dataset. (C) Intraoperative dataset.

Author Contributions

Conceptualization: S.-W.L. and S.-H.K.; data curation: S.-W.L. and J.J.; investigation: S.-W.L., J.J., W.-Y.S. and D.L.; methodology: S.-W.L., J.J., W.-Y.S. and S.-H.K.; supervision: S.-W.L. and S.-H.K.; writing—original draft: S.-W.L.; writing—review and editing: S.-W.L., J.J. and S.-H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health and Welfare, Republic of Korea (grant numbers: HR20C0026, HI22C1723). This study was also supported by a grant (2022IP0053, 2023IP0132) from the Asan Institute for Life Sciences, Asan Medical Center, Seoul, Republic of Korea.

Institutional Review Board Statement

The study received approval from the Institutional Review Board (IRB) of Asan Medical Center (IRB No. 2024-0060) and was conducted according to the guidelines of the Declaration of Helsinki. Informed consent from participants was waived based on the retrospective collection and secondary utilization of EHR data.

Informed Consent Statement

Informed consent from participants was waived based on the retrospective collection and secondary utilization of EHR data by the IRB of Asan Medical Center.

Data Availability Statement

The data presented in this study are available on reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hoste, E.A.J.; Kellum, J.A.; Selby, N.M.; Zarbock, A.; Palevsky, P.M.; Bagshaw, S.M.; Goldstein, S.L.; Cerda, J.; Chawla, L.S. Global epidemiology and outcomes of acute kidney injury. Nat. Rev. Nephrol. 2018, 14, 607–625. [Google Scholar] [CrossRef]
  2. Thakar, C.V. Perioperative acute kidney injury. Adv. Chronic Kidney Dis. 2013, 20, 67–75. [Google Scholar] [CrossRef]
  3. Chertow, G.M.; Burdick, E.; Honour, M.; Bonventre, J.V.; Bates, D.W. Acute kidney injury, mortality, length of stay, and costs in hospitalized patients. J. Am. Soc. Nephrol. 2005, 16, 3365–3370. [Google Scholar] [CrossRef]
  4. Hobson, C.; Ozrazgat-Baslanti, T.; Kuxhausen, A.; Thottakkara, P.; Efron, P.A.; Moore, F.A.; Moldawer, L.L.; Segal, M.S.; Bihorac, A. Cost and Mortality Associated with Postoperative Acute Kidney Injury. Ann. Surg. 2015, 261, 1207–1214. [Google Scholar] [CrossRef]
  5. Biteker, M.; Dayan, A.; Tekkesin, A.I.; Can, M.M.; Tayci, I.; Ilhan, E.; Sahin, G. Incidence, risk factors, and outcomes of perioperative acute kidney injury in noncardiac and nonvascular surgery. Am. J. Surg. 2014, 207, 53–59. [Google Scholar] [CrossRef]
  6. Cho, E.; Kim, S.C.; Kim, M.G.; Jo, S.K.; Cho, W.Y.; Kim, H.K. The incidence and risk factors of acute kidney injury after hepatobiliary surgery: A prospective observational study. BMC Nephrol. 2014, 15, 169. [Google Scholar] [CrossRef]
  7. Saran, R.; Robinson, B.; Abbott, K.C.; Agodoa, L.Y.C.; Bhave, N.; Bragg-Gresham, J.; Balkrishnan, R.; Dietrich, X.; Eckard, A.; Eggers, P.W.; et al. US Renal Data System 2017 Annual Data Report: Epidemiology of Kidney Disease in the United States. Am. J. Kidney Dis. 2018, 71, A7. [Google Scholar] [CrossRef]
  8. Karway, G.K.; Koyner, J.L.; Caskey, J.; Spicer, A.B.; Carey, K.A.; Gilbert, E.R.; Dligach, D.; Mayampurath, A.; Afshar, M.; Churpek, M.M. Development and external validation of multimodal postoperative acute kidney injury risk machine learning models. JAMIA Open 2023, 6, ooad109. [Google Scholar] [CrossRef]
  9. Lei, V.J.; Luong, T.; Shan, E.; Chen, X.; Neuman, M.D.; Eneanya, N.D.; Polsky, D.E.; Volpp, K.G.; Fleisher, L.A.; Holmes, J.H.; et al. Risk Stratification for Postoperative Acute Kidney Injury in Major Noncardiac Surgery Using Preoperative and Intraoperative Data. JAMA Netw. Open 2019, 2, e1916921. [Google Scholar] [CrossRef] [PubMed]
  10. Huang, C.; Murugiah, K.; Mahajan, S.; Li, S.X.; Dhruva, S.S.; Haimovich, J.S.; Wang, Y.; Schulz, W.L.; Testani, J.M.; Wilson, F.P.; et al. Enhancing the prediction of acute kidney injury risk after percutaneous coronary intervention using machine learning techniques: A retrospective cohort study. PLoS Med. 2018, 15, e1002703. [Google Scholar] [CrossRef] [PubMed]
  11. Koyner, J.L.; Carey, K.A.; Edelson, D.P.; Churpek, M.M. The Development of a Machine Learning Inpatient Acute Kidney Injury Prediction Model. Crit. Care Med. 2018, 46, 1070–1077. [Google Scholar] [CrossRef]
  12. Wu, L.; Hu, Y.; Liu, X.; Zhang, X.; Chen, W.; Yu, A.S.L.; Kellum, J.A.; Waitman, L.R.; Liu, M. Feature Ranking in Predictive Models for Hospital-Acquired Acute Kidney Injury. Sci. Rep. 2018, 8, 17298. [Google Scholar] [CrossRef] [PubMed]
  13. Salmasi, V.; Maheshwari, K.; Yang, D.; Mascha, E.J.; Singh, A.; Sessler, D.I.; Kurz, A. Relationship between Intraoperative Hypotension, Defined by Either Reduction from Baseline or Absolute Thresholds, and Acute Kidney and Myocardial Injury after Noncardiac Surgery: A Retrospective Cohort Analysis. Anesthesiology 2017, 126, 47–65. [Google Scholar] [CrossRef] [PubMed]
  14. Gregory, A.; Stapelfeldt, W.H.; Khanna, A.K.; Smischney, N.J.; Boero, I.J.; Chen, Q.; Stevens, M.; Shaw, A.D. Intraoperative Hypotension Is Associated with Adverse Clinical Outcomes after Noncardiac Surgery. Anesth. Analg. 2021, 132, 1654–1665. [Google Scholar] [CrossRef] [PubMed]
  15. Lee, H.C.; Park, Y.; Yoon, S.B.; Yang, S.M.; Park, D.; Jung, C.W. VitalDB, a high-fidelity multi-parameter vital signs database in surgical patients. Sci. Data 2022, 9, 279. [Google Scholar] [CrossRef] [PubMed]
  16. Chertow, G.M.; Levy, E.M.; Hammermeister, K.E.; Grover, F.; Daley, J. Independent association between acute renal failure and mortality following cardiac surgery. Am. J. Med. 1998, 104, 343–348. [Google Scholar] [CrossRef]
  17. Lok, C.E.; Austin, P.C.; Wang, H.; Tu, J.V. Impact of renal insufficiency on short- and long-term outcomes after cardiac surgery. Am. Heart J. 2004, 148, 430–438. [Google Scholar] [CrossRef]
  18. Machado, M.N.; Nakazone, M.A.; Maia, L.N. Prognostic value of acute kidney injury after cardiac surgery according to kidney disease: Improving global outcomes definition and staging (KDIGO) criteria. PLoS ONE 2014, 9, e98028. [Google Scholar] [CrossRef] [PubMed]
  19. Bihorac, A.; Yavas, S.; Subbiah, S.; Hobson, C.E.; Schold, J.D.; Gabrielli, A.; Layon, A.J.; Segal, M.S. Long-term risk of mortality and acute kidney injury during hospitalization after major surgery. Ann. Surg. 2009, 249, 851–858. [Google Scholar] [CrossRef]
  20. Grams, M.E.; Sang, Y.; Coresh, J.; Ballew, S.; Matsushita, K.; Molnar, M.Z.; Szabo, Z.; Kalantar-Zadeh, K.; Kovesdy, C.P. Acute Kidney Injury after Major Surgery: A Retrospective Analysis of Veterans Health Administration Data. Am. J. Kidney Dis. 2016, 67, 872–880. [Google Scholar] [CrossRef]
  21. Park, J.T. Postoperative acute kidney injury. Korean J. Anesthesiol. 2017, 70, 258–266. [Google Scholar] [CrossRef] [PubMed]
  22. Yu, X.; Ji, Y.; Huang, M.; Feng, Z. Machine learning for acute kidney injury: Changing the traditional disease prediction mode. Front. Med. 2023, 10, 1050255. [Google Scholar] [CrossRef]
  23. Tomasev, N.; Glorot, X.; Rae, J.W.; Zielinski, M.; Askham, H.; Saraiva, A.; Mottram, A.; Meyer, C.; Ravuri, S.; Protsyuk, I.; et al. A clinically applicable approach to continuous prediction of future acute kidney injury. Nature 2019, 572, 116–119. [Google Scholar] [CrossRef] [PubMed]
  24. Simonov, M.; Ugwuowo, U.; Moreira, E.; Yamamoto, Y.; Biswas, A.; Martin, M.; Testani, J.; Wilson, F.P. A simple real-time model for predicting acute kidney injury in hospitalized patients in the US: A descriptive modeling study. PLoS Med. 2019, 16, e1002861. [Google Scholar] [CrossRef] [PubMed]
  25. Koyner, J.L.; Adhikari, R.; Edelson, D.P.; Churpek, M.M. Development of a Multicenter Ward-Based AKI Prediction Model. Clin. J. Am. Soc. Nephrol. 2016, 11, 1935–1943. [Google Scholar] [CrossRef] [PubMed]
  26. Hodgson, L.E.; Sarnowski, A.; Roderick, P.J.; Dimitrov, B.D.; Venn, R.M.; Forni, L.G. Systematic review of prognostic prediction models for acute kidney injury (AKI) in general hospital populations. BMJ Open 2017, 7, e016591. [Google Scholar] [CrossRef] [PubMed]
  27. Hodgson, L.E.; Roderick, P.J.; Venn, R.M.; Yao, G.L.; Dimitrov, B.D.; Forni, L.G. Correction: The ICE-AKI study: Impact analysis of a Clinical prediction rule and Electronic AKI alert in general medical patients. PLoS ONE 2018, 13, e0203183. [Google Scholar] [CrossRef] [PubMed]
  28. Moons, K.G.; Altman, D.G.; Reitsma, J.B.; Ioannidis, J.P.; Macaskill, P.; Steyerberg, E.W.; Vickers, A.J.; Ransohoff, D.F.; Collins, G.S. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): Explanation and elaboration. Ann. Intern. Med. 2015, 162, W1–W73. [Google Scholar] [CrossRef]
  29. Debray, T.P.; Vergouwe, Y.; Koffijberg, H.; Nieboer, D.; Steyerberg, E.W.; Moons, K.G. A new framework to enhance the interpretation of external validation studies of clinical prediction models. J. Clin. Epidemiol. 2015, 68, 279–289. [Google Scholar] [CrossRef]
Figure 1. Study flow chart of the internal dataset.
Figure 1. Study flow chart of the internal dataset.
Jpm 14 00587 g001
Figure 2. AUROC and AUPRC values of prediction models across various modeling methods using all features in the internal and external datasets. (A) AUROC and (B) AUPRC values of prediction models across various modeling methods using all features in the internal datasets. (C) AUROC and (D) AUPRC values of prediction models across various modeling methods using all features in the external datasets. AUROC and AUPRC values are represented as 95% confidence intervals. AUROC, area under receiver operating characteristic; AUPRC, area under precision–recall curve; LR, logistic regression; RF, random forest; GBM, gradient boosting machine; DNN, deep neural network.
Figure 2. AUROC and AUPRC values of prediction models across various modeling methods using all features in the internal and external datasets. (A) AUROC and (B) AUPRC values of prediction models across various modeling methods using all features in the internal datasets. (C) AUROC and (D) AUPRC values of prediction models across various modeling methods using all features in the external datasets. AUROC and AUPRC values are represented as 95% confidence intervals. AUROC, area under receiver operating characteristic; AUPRC, area under precision–recall curve; LR, logistic regression; RF, random forest; GBM, gradient boosting machine; DNN, deep neural network.
Jpm 14 00587 g002
Figure 3. Feature importance of the GBM model for predicting postoperative AKI using SHAP values. GBM, gradient boosting machine; AKI, acute kidney injury; SHAP, Shapley additive explanation; SBP, systolic blood pressure; ABP, arterial blood pressure; DBP, diastolic blood pressure; BMI, body mass index; PT, prothrombin time; std, standard deviation; min, minimum.
Figure 3. Feature importance of the GBM model for predicting postoperative AKI using SHAP values. GBM, gradient boosting machine; AKI, acute kidney injury; SHAP, Shapley additive explanation; SBP, systolic blood pressure; ABP, arterial blood pressure; DBP, diastolic blood pressure; BMI, body mass index; PT, prothrombin time; std, standard deviation; min, minimum.
Jpm 14 00587 g003
Table 1. Characteristics of the internal and external datasets.
Table 1. Characteristics of the internal and external datasets.
Internal DatasetOpen Database
Total
(n = 76,032)
AKI
(n = 2314)
Non-AKI
(n = 73,718)
Total
(n = 5512)
Demographic data
 Age, years55.3 ± 15.259.0 ± 13.455.2 ± 15.358.2 ± 14.0
 Sex (female)34,491 (45.4)1647 (71.2)32,844 (44.6)2766 (50.2)
 BMI (kg/m2)24.2 ± 3.724.7 ± 3.824.2 ± 3.723.4 ± 3.6
 ASA
  17826 (10.3)149 (6.4)7677 (10.4)1612 (29.2)
  260,450 (79.5)1439 (62.2)59,011 (80.1)3337 (60.5)
  37092 (9.3)595 (25.7)6497 (8.8)516 (9.4)
  ≥4661 (0.8)130 (5.6)531 (0.7)47 (8.5)
 Surgery department
  GS32,092 (42.2)1347 (58.2)30,745 (41.7)4272 (77.5)
  URO7828 (10.3)677 (29.3)7151 (9.7)116 (2.1)
  OBY7785 (10.2)46 (1.9)7739 (10.5)197 (3.6)
  OS7237 (9.5)78 (3.4)7159 (9.7)-
  ENT5699 (7.5)32 (1.4)5667 (7.7)-
  NS5149 (6.8)20 (0.9)5129 (7.0)-
  CS4409 (5.8)93 (4.0)4316 (5.9)927 (16.8)
  PS2691 (3.5)15 (0.6)2676 (3.6)-
  OPH2254 (3.0)1 (0.0)2253 (3.1)-
  DNT878 (1.2)5 (0.2)873 (1.2)-
  DER10 (0.0)0 (0.0)10 (0.0)-
Preoperative laboratory results
 White blood cell, 103/μL6.5 ± 2.66.6 ± 3.36.5 ± 2.56.2 ± 2.4
 Hemoglobin, g/dL12.9 ± 1.812.3 ± 2.312.9 ± 1.812.9 ± 1.9
 Sodium, mmol/L140.2 ± 2.6139.8 ± 3.5140.3 ± 2.5140.1 ± 2.8
 Platelet, 103/μL241.8 ± 75.9208.0 ± 99.6242.9 ± 74.8242.87 ± 83.4
 Potassium, mmol/L4.2 ± 0.44.2 ± 0.54.3 ± 0.44.2 ± 0.4
 Chloride, mmol/L103.7 ± 2.9103.5 ± 3.9103.7 ± 2.9103.3 ± 4.0
 Total bilirubin, mg/dL0.6 ± 1.61.4 ± 3.90.6 ± 1.40.7 ± 1.7
 BUN, mg/dL15.3 ± 6.517.7 ± 10.115.2 ± 6.414.8 ± 6.9
 Creatinine, mg/dL0.8 ± 0.31.0 ± 0.60.8 ± 0.30.8 ± 0.3
 Albumin, g/dL3.8 ± 0.53.5 ± 0.63.8 ± 0.54.1 ± 0.5
 AST, IU/L24.8 ± 49.240.0 ± 186.124.3 ± 37.631.1 ± 140.0
 ALT, IU/L22.4 ± 40.832.3 ± 148.222.1 ± 32.129.1 ± 95.2
 Hematocrit, %38.9 ± 5.036.8 ± 6.638.9 ± 4.937.4 ± 6.1
 eGFR, mL/min/1.73 m292.5 ± 19.985.0 ± 26.692.7 ± 19.586.7 ± 26.4
 Glucose, mg/dL115.2 ± 40.4122.9 ± 49.0114.9 ± 40.1115.8 ± 41.7
 PT, INR1.0 ± 0.21.1 ± 0.41.0 ± 0.21.0 ± 0.2
 aPTT, s27.6 ± 4.429.5 ± 8.527.6 ± 4.232.9 ± 8.6
 CRP, mg/dL1.2 ± 3.41.9 ± 4.51.1 ± 3.31.2 ± 3.6
Intraoperative data
 EBL, mL112.9 ± 232.4145.0 ± 649.423.3 ± 1360.1365.86 ± 1176.3
 Anesthesia time, min171.5 ± 128.5301.9 ± 233.4167.4 ± 121.6203.9 ± 114.0
 Surgery time, min126.0 ± 114.3241.8 ± 210.7122.4 ± 108.0140.10 ± 100.3
Data represent mean ± standard deviation, median (interquartile range), or number (percentage). AKI, acute kidney injury; BMI, body mass index; ASA, American Society of Anesthesiologists classification; GS, general surgery; URO, urology; OBY, obstetrics and gynecology; OS, orthopedic surgery; ENT, ear, nose, and throat surgery; NS, neurosurgery; CS, cardiothoracic surgery; PS, plastic surgery; OPH, ophthalmology; DNT, dental surgery; DER, dermatology; BUN, blood urea nitrogen; AST, aspartate aminotransferase; ALT, alanine aminotransferase; PT, prothrombin time; aPTT, activated partial thromboplastin time; EBL, estimated blood loss.
Table 2. Clinical outcomes in the internal and external datasets.
Table 2. Clinical outcomes in the internal and external datasets.
Clinical OutcomeInternal DatasetOpen Database
Total
(n = 76,032)
AKI
(n = 2314)
Non-AKI
(n = 73,718)
Total
(n = 5512)
Acute kidney injury, n (%)2314 (3.1)--78 (1.4)
Length of hospital stay (days)9.0 ± 38.321.9 ± 34.58.6 ± 38.310.3 ± 13.7
Length of hospital stay ≥ 7 days, n (%)29,280 (38.5)1651 (71.3)27,629 (37.5)3153 (57.2)
In-hospital death, n (%)1595 (2.1)144 (6.2)1201 (1.6)47 (0.9)
30-day mortality, n (%)270 (0.4)54 (2.3)216 (0.3)28 (0.5)
Postoperative ICU care, n (%)7217 (9.5)704 (30.4)6513 (9.8)1008 (18.3)
Length of ICU stay (days)0.7 ± 6.73.8 ± 13.90.6 ± 6.40.49 ± 3.3
Data represent mean ± standard deviation, median (interquartile range), or number (percentage). AKI, acute kidney injury; ICU, intensive care unit.
Table 3. Predictive performances of machine learning techniques for predicting postoperative AKI in the internal dataset using combinations of various features.
Table 3. Predictive performances of machine learning techniques for predicting postoperative AKI in the internal dataset using combinations of various features.
FeaturesModelAUROCAUPRCF1-Score
Demographic dataLR0.6942 ± 0.00380.5714 ± 0.00700.3981 ± 0.0048
RF0.6549 ± 0.00480.5165 ± 0.00570.4931 ± 0.0053
GBM0.7137 ± 0.00380.5906 ± 0.00770.4573 ± 0.0074
DNN0.6794 ± 0.00920.5060 ± 0.01400.2919 ± 0.042
Preoperative dataLR0.6795 ± 0.00430.5942 ± 0.00800.4184 ± 0.0059
RF0.7224 ± 0.00430.6375 ± 0.00610.5023 ± 0.0052
GBM0.7280 ± 0.00400.6468 ± 0.00660.5072 ± 0.0047
DNN0.6281 ± 0.01610.5218 ± 0.01450.3707 ± 0.0290
Intraoperative dataLR0.7449 ± 0.00390.6267 ± 0.00660.4963 ± 0.0042
RF0.7221 ± 0.00340.5983 ± 0.00670.4595 ± 0.0051
GBM0.8161 ± 0.00370.7146 ± 0.00820.6562 ± 0.0069
DNN0.8105 ± 0.00700.6749 ± 0.01020.6539 ± 0.0109
All featuresLR0.7827 ± 0.00430.6609 ± 0.00770.5449 ± 0.0056
RF0.8547 ± 0.00290.7578 ± 0.00570.7101 ± 0.0058
GBM0.8679 ± 0.00390.7862 ± 0.00840.7226 ± 0.0062
DNN0.8279 ± 0.00690.6963 ± 0.000990.6355 ± 0.0277
Data are presented as means (95% confidence intervals). AKI, acute kidney injury; AUROC, area under receiver operating characteristic; AUPRC, area under precision–recall curve; LR, logistic regression; RF, random forest; GBM, gradient boosting machine; DNN, deep neural network.
Table 4. Predictive performance of machine learning techniques for predicting postoperative AKI in the external dataset using combinations of various features.
Table 4. Predictive performance of machine learning techniques for predicting postoperative AKI in the external dataset using combinations of various features.
FeaturesModelAUROCAUPRCF1-Score
Demographic dataLR0.6815 ± 0.01230.5491 ± 0.02140.4028 ± 0.0063
RF0.6031 ± 0.01220.4442 ± 0.01610.4176 ± 0.0179
GBM0.6995 ± 0.01020.5900 ± 0.01770.5042 ± 0.0153
DNN0.6600 ± 0.01080.5079 ± 0.01720.3171 ± 0.0360
Preoperative dataLR0.6808 ± 0.01140.6443 ± 0.01250.4211 ± 0.0153
RF0.6031 ± 0.01220.4442 ± 0.01610.4176 ± 0.0179
GBM0.7693 ± 0.01280.6958 ± 0.0.1850.5118 ± 0.0139
DNN0.6839 ± 0.002230.5978 ± 0.02390.4641 ± 0.0386
Intraoperative dataLR0.7674 ± 0.01010.4027 ± 0.02120.0292 ± 0.0009
RF0.6499 ± 0.02950.4748 ± 0.02620.4877 ± 0.0273
GBM0.7054 ± 0.01680.5836 ± 0.02660.5985 ± 0.0116
DNN0.5770 ± 0.37800.489 ± 0.04200.4804 ± 0.0196
All featuresLR0.7495 ± 0.01660.6541 ± 0.02690.6268 ± 0.0189
RF0.7342 ± 0.02950.6403 ± 0.03250.6098 ± 0.0228
GBM0.7572 ± 0.01500.6834 ± 0.02130.6129 ± 0.0205
DNN0.6715 ± 0.05300.5584 ± 0.05400.5292 ± 0.0303
Data are presented as means (95% confidence intervals). AKI, acute kidney injury; AUROC, area under receiver operating characteristic; AUPRC, area under precision–recall curve; LR, logistic regression; RF, random forest; GBM, gradient boosting machine; DNN, deep neural network.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, S.-W.; Jang, J.; Seo, W.-Y.; Lee, D.; Kim, S.-H. Internal and External Validation of Machine Learning Models for Predicting Acute Kidney Injury Following Non-Cardiac Surgery Using Open Datasets. J. Pers. Med. 2024, 14, 587. https://doi.org/10.3390/jpm14060587

AMA Style

Lee S-W, Jang J, Seo W-Y, Lee D, Kim S-H. Internal and External Validation of Machine Learning Models for Predicting Acute Kidney Injury Following Non-Cardiac Surgery Using Open Datasets. Journal of Personalized Medicine. 2024; 14(6):587. https://doi.org/10.3390/jpm14060587

Chicago/Turabian Style

Lee, Sang-Wook, Jaewon Jang, Woo-Young Seo, Donghee Lee, and Sung-Hoon Kim. 2024. "Internal and External Validation of Machine Learning Models for Predicting Acute Kidney Injury Following Non-Cardiac Surgery Using Open Datasets" Journal of Personalized Medicine 14, no. 6: 587. https://doi.org/10.3390/jpm14060587

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop