Next Article in Journal
An Energy Management System for Distributed Energy Storage System Considering Time-Varying Linear Resistance
Previous Article in Journal
Balancing Security and Efficiency: A Power Consumption Analysis of a Lightweight Block Cipher
Previous Article in Special Issue
Comparison of Machine Learning Models for Predicting Interstitial Glucose Using Smart Watch and Food Log
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diagnosing and Characterizing Chronic Kidney Disease with Machine Learning: The Value of Clinical Patient Characteristics as Evidenced from an Open Dataset

1
Department of Computer Science, St. Francis Xavier University, Antigonish, NS B2G 2W5, Canada
2
Nova Scotia Health Authority, Halifax, NS B3H 1V8, Canada
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(21), 4326; https://doi.org/10.3390/electronics13214326
Submission received: 6 September 2024 / Revised: 18 October 2024 / Accepted: 30 October 2024 / Published: 4 November 2024
(This article belongs to the Special Issue Machine Learning for Biomedical Applications)

Abstract

:
Applying artificial intelligence (AI) and machine learning for chronic kidney disease (CKD) diagnostics and characterization has the potential to improve the standard of patient care through accurate and early detection, as well as providing a more detailed understanding of the condition. This study employed reproducible validation of AI technology with public domain software applied to CKD diagnostics on a publicly available CKD dataset acquired from 400 patients. The approach presented includes patient-specific symptomatic variables and demonstrates performance improvements associated with this approach. Our best-performing AI models, which include patient symptom variables, achieve predictive accuracies ranging from 99.4 to 100% across both hold-out and 5-fold validation with the light gradient boosting machine. We demonstrate that the exclusion of patient symptom variables reduces model performance in line with the literature on the same dataset. We also provide an unsupervised learning cluster analysis to help interpret variability among, and characterize the population of, patients with CKD.

1. Introduction

The kidneys are a filtration system for removing waste and contaminants from the body through the urine. The kidneys remove excess acid from the body and maintain healthy levels of salts, water, and minerals in our blood, which are necessary for healthy tissue function. They also produce hormones that assist in blood pressure regulation, the creation of red blood cells, and maintaining healthy bones [1]. Chronic kidney disease (CKD) is a “permanent, gradual, and progressive loss of kidney function over several years caused by a variety of medical conditions” [2]. CKD does not have a known cure, so once it is present, it will last the full life of the individual, and without proper management, it can lead to tissue failure.
CKD is not restricted to a particular group or age. However, patients with diabetes mellitus, heart disease, or high blood pressure are most likely to suffer from CKD. Patients over the age of 60, or with a family history of CKD, or who are HIV positive are at an elevated risk of developing CKD [3]. CKD is a challenge for public health globally, in part due to the preliminary stages of the disease often going undetected, with patients unaware of their condition until it becomes critical or even terminal [4].
The World Health Organization estimates that CKD affects 10% of the global population [5]. CKD has a higher prevalence and morbidity in developing nations due to a lack of access to adequate healthcare among most of the population [5].

1.1. CKD Stages

Chronic kidney disease is categorized into five stages; this helps doctors to determine the best care for each patient. CKD progresses from stage 1 to stage 5; a patient’s stage is determined with the estimated glomerular filtration rare (eGFR) [6].
Patients with stage 1 CKD have mild kidney damage with an eGFR of 90 mL/min or greater. At stage 1, there are no symptoms to suggest that the kidneys are damaged as the kidneys appear to function normally at this stage, and most people with the condition are unaware that they have stage 1 CKD [7]. At stage 2, the patient has mild kidney damage with an eGFR of 60–89 mL/min, and as in stage 1 CKD, there are usually no symptoms noticed by the patient, as the kidneys still function normally. At these stages, if the patient discovers that they have CKD, it is usually because they were examined for another medical condition [8].
A patient with stage 3 CKD has moderate to severe kidney damage with an eGFR between 30 and 59 mL/min. This stage is subdivided into stage 3a (mild to moderate kidney damage) with an eGFR of 45–59 mL/min and stage 3b (moderate to severe kidney damage) with an eGFR of 30–44 mL/min. At this stage, the kidneys do not function effectively; as kidney function declines, waste products build up in the blood, resulting in uremia. At stage 3, the patient develops complications such as high blood pressure, anemia, and early bone disease. At stage 3, symptoms may include frequent or less frequent urination, urination changes (foamy; dark orange, brown, tea-colored, or red), fatigue, itchy or dry skin, nausea, loss of appetite that leads to weight loss, and swelling in the hands or feet [9].
A patient with stage 4 CKD has severe loss of kidney function with an eGFR within the range of 15–29 mL/min. In stage 4 CKD, the patient is at the highest risk of having kidney failure and an elevated risk of heart disease. At this stage, the patient will require hemodialysis, peritoneal dialysis, or a kidney transplant. It is characterized as the last stage before kidney failure. At stage 4, damage to the kidney is irreversible. However, under the guidance of a nephrologist and dietitian, efforts may be taken to slow down the kidney damage and mitigate it from progressing to stage 5 (total failure). Unfortunately, even with treatment, kidneys may still fail [10].
Patients with stage 5 kidney failure, commonly known as end-stage kidney disease (ESKD), have an estimated (eGFR) of <15 mL/min. At stage 5, the patient is at the highest risk of developing heart disease. Patients in stages 5 and 4 experience health complications such as anemia, metabolic acidosis, mineral and bone disorders, and hyperkalemia, as well as symptoms such as ammonia-smelling breath, changes in skin color, muscle cramps, difficulty breathing, and little to no urine. Patients diagnosed with stage 5 CKD should consult a nephrologist immediately. At this stage, the patient will need hemodialysis, peritoneal dialysis, or a kidney transplant [11].
Treatment for CKD is expensive and often unsuccessful. Therefore, early detection of CKD is particularly important for implementing preventative measures and managing the disease to slow its progression. Initially, the diagnosis of CKD by a health practitioner involves requesting the patient’s medical history, medications they are currently on, and a description of symptoms. Following that, the healthcare practitioner requests a blood test, also known as the estimated glomerular filtration rate (eGFR), and a urine test, also known as the urine albumin–creatinine ratio (uACR), to assess kidney function. Having an eGFR under 60 and a uACR over 30 for three months or more is a sign of kidney disease [12].

1.2. Machine Learning Applied to CKD

Machine learning research in this field has focused on a variety of applications, including predicting the progression of diabetic kidney disease [13], the creation of deep learning models for the prediction of intradialytic hypertension [14], and technology focused on patients with oliguric acute kidney injury [15]. Technologies focused specifically on the prediction of CKD have been developed, using the same public domain dataset relied upon in this study, including a rotation forest model yielding 99.2% accuracy [16], an approach based on the random forest yielding 97% accuracy [17], a support vector machine approach yielding 98.86% accuracy [18], an approach using the K-Nearest Neighbour classifier as well as the Extra Tree Classifier yielding 99% accuracy [19], and multiple approaches based on the XGBoost algorithm (https://github.com/dmlc/xgboost, accessed on 24 February 2024) achieving 94% accuracy [20] and 98.3% accuracy [21]. Additional studies have been conducted on alternative CKD datasets as well [22,23,24]. It should be noted that there are inevitably differences in the validation employed by each of these studies, as standard validation requires data sample randomization, which would have been performed multiple times independently across studies. As such, direct comparison of these performance metrics has inherent challenges, and coming to firm conclusions as to which algorithms perform the best for prediction of CKD should involve a controlled and fair comparison across all algorithms being considered. Performance metrics reported would ideally be part of an identical validation suite, rather than being compared across studies. Regardless, there is value in comparing machine learning performance reported across studies, even though differences always exist in the evaluation methods employed. It is noteworthy that all studies were able to achieve generally high accuracy results with clear room for further improvement of the technology’s predictive capacity.
Although a wide variety of machine learning approaches have been developed for CKD, our study attempted to address the potential value of clinical patient characteristics in this publicly available dataset. In this study, we hypothesized that the use of open-source software df-analyze, inclusive of the high-performing light gradient boosting machine (LGBM), would provide valuable improvements to existing technologies for diagnostics of CKD. We further hypothesized that the inclusion of patient clinical variables, not focused on in other studies based on this dataset, would help further improve the quality of diagnostic predictions in CKD. Finally, we hypothesized that unsupervised learning could assist in the characterization of CKD.

2. Materials and Methods

The dataset was collected from Kaggle (https://www.kaggle.com/datasets/mansoordaku/ckdisease) and is also available from the University of California Irvine Machine Learning Repository (https://archive.ics.uci.edu/dataset/336/chronic+kidney+disease), and was accessed on 28 January 2024. This dataset is publicly available and published under a Creative Commons Attribution 4.0 International (CC BY 4.0) license. This dataset was created from a population of 400 patients over a period of 2 months in India, with the acquisition of 24 feature measurements collected with the target variable CKD status [25]. It has 12 numerical columns and 13 categorical or nominal values. In this dataset, 250 of the samples were from patients with CKD, and 150 samples were acquired from patients without CKD. The machine learning task addressed in this analysis is to predict the CKD status from the set of feature measurements acquired from each patient, thus we are using AI technology to perform CKD diagnoses. To summarize our methods, we downloaded the above dataset and divided it into two parts, one with and one without our patient clinical characteristics, and then public domain software was used to fairly compare a collection of machine learning and feature selection algorithms on each of the two versions of the dataset considered. An unsupervised learning cluster analysis was also performed to help assess sample groupings in the dataset. Details are provided below.

2.1. Dataset Description

Table 1 and Table 2 below provide the different variables with their description, data type, and units along with the mean and standard deviation values for the numerical variables for both the CKD and not CKD groups.

2.2. Machine Learning

The current framework used for executing the machine learning algorithms considered is the public domain software package df-analyze (https://github.com/stfxecutables/df-analyze/), which is available for download and was accessed on 24 February 2024. “df-analyze is a command-line tool for performing AutoML on small to medium-sized tabular datasets. In particular, df-analyze attempts to automate:
  • Feature type inference.
  • Feature description (e.g., univariate associations and stats).
  • Data cleaning (e.g., NaN handling and imputation).
  • Training, validation, and test splitting.
  • Feature selection.
  • Hyperparameter tuning.
  • Model selection and validation.
and saves all key tables and outputs from this process.” [26]. This tool was first used in a study on schizophrenia [27] and has since had a variety of features added, including additional feature selection and analytics. Detailed instructions are provided for the use of this software (Instructions).
The software package df-analyze was used to fairly compare the following machine learning algorithms: K-Nearest Neighbour (KNN), light gradient boosting machine (LGBM), random forest, logistic regression, Stochastic Gradient Descent, Multi-layered Perceptron Artificial Neural Network, and the support vector machine. A dummy classifier that simply predicts the majority class was also included as a baseline for comparison. All algorithms were compared with two forms of validation: K-fold validation, and evaluation on a holdout dataset. Missing values in the dataset were imputed with the corresponding median value for the numerical attributes after an initial robust normalization scheme documented online, and missing values for categorical variables were converted to a single additional category, which was then clearly labeled as a NaN (Not a Number) indicator after the categorical variable was one-hot encoded (i.e., a new feature is created that is 1 for all samples missing that entry in their categorical variable, and 0 for all other samples). Entries in the dataset containing “?” symbols were converted to NaN, which is compatible with df-analyze alongside blank entries. Hyperparamater tuning was performed within the training set, with the search spaces outlined online for the KNN, LGBM, logistic regression, and Stochastic Gradient Descent. Validation runs were repeated 50 times.
The analysis was repeated two times, once with all the feature measurements available (both numerical features and patient clinical characteristic features), and the second run was performed while excluding the patient clinical characteristic features, which are further presented below to illustrate their predictive potential. Feature selection was included as part of each run of df-analyze, including wrapper, filter (both prediction and association-based), and embedded methods, each applied to each machine learning algorithm included in the analysis.

2.3. Statistical Analysis

For each machine learning and feature selection algorithm pairing trained for CKD prediction as part of df-analyze, we computed the overall accuracy (ACC), area under the receiver operating characteristic curve (AUROC), F1 score, negative predictive value (NPV), positive predictive value (PPV), as well as the sensitivity (sens) and specificity (spec) values. Comparisons of statistical metrics computed were performed on both the holdout set and 5-fold validation on the holdout set, for both the cases with and without the aforementioned clinical patient characteristic features.

3. Results

3.1. Patient Clinical Characteristics

The patient clinical characteristic variables, included in one of our runs of df-analyze but not the other, are outlined in Figure 1 and Figure 2 below, broken down between CKD and not CKD status. These features include whether the patient has the following symptoms: abnormal appetite (APPET), pedal edema (PE), anemia (ANE), hypertension (HTN), diabetes mellitus (DM), and coronary artery disease (CAD). These figures are provided to illustrate the difference between our two runs of our validation software (with and without these features), in order to assess their potential for informing CKD diagnosis, as they were not considered in this manner in previous analyses on this dataset.
It is noteworthy from these bar plots that 100% of the patients who do not have CKD also do not exhibit any of these symptoms in this dataset. Thus, these variables have predictive potential to inform multivariable machine learning technologies, which may contribute towards making more accurate predictions of CKD vs. not CKD.

3.2. Machine Learning with Patient Clinical Characteristics

Table 3 provides the results of the machine learning analysis for the complete dataset inclusive of patient clinical characteristics applied to the holdout set. Table 4 provides the machine learning results for the complete dataset inclusive of patient clinical characteristics as part of 5-fold cross-validation. Table 3 and Table 4 are provided to illustrate the results of our detailed comparative analysis across machine learning technologies for the entire dataset. The LGBM (light gradient boosting machine) with either no feature selection or with embedded (embed_lgbm)-based feature selection performed the best among all the machine learning techniques compared, producing 99.4 to 100% accuracy (inclusive of clinical patient characteristic features) across the two validation approaches employed. The feature selection reports for the embedded feature selection (embed_lgbm) are provided in Appendix A. The next best performing approach, using prediction-based feature selection (pred), is provided for comparative reference in Appendix B.

3.3. Machine Learning Without Patient Clinical Characteristics

Table 5 provides the results of the machine learning analysis for the reduced dataset, not inclusive of patient clinical characteristics, applied to the holdout set. Table 6 provides the machine learning results for the reduced dataset not inclusive of patient clinical characteristics as part of 5-fold cross-validation. Table 5 and Table 6 are provided to illustrate the results of our detailed comparative analysis across machine learning technologies for the reduced dataset that does not include the clinical variables outlined in Figure 1 and Figure 2. Note that there is no clear winning algorithm across the two validation approaches addressed and that the performance of the classifiers is generally degraded relative to the results from the complete dataset inclusive of patient clinical characteristics provided in Table 3 and Table 4.
Our leading performing models were consistently the light gradient boosting machine (LGBM) applied to the complete dataset inclusive of clinical patient characteristics. It is noteworthy that the predictive performance of the LGBM with no feature selection was nearly the same as the performance of the LGBM with embedded feature selection (embed_lgbm). The features selected for from the embedded feature selection (embed_lgbm) are summarized in Appendix A, while the features selected for from the filter-prediction-based method (pred) are provided in Appendix B. These appendices provide an indication, based on competing methods, of what feature measurements are most important to rely upon for accurate prediction of CKD.

3.4. Cluster Analysis Results

In order to help illustrate and improve understanding of the complete dataset, we also performed a cluster analysis with the unsupervised K-means algorithm. For the purpose of determining an optimal K value, both silhouette scores and Calinski–Harabasz scores [28] were applied, respectively, producing Figure 3 and Figure 4.
Based on both the silhouette score and the Calinski–Harabasz score, the highest K value occurred with K = 2. We were also separately interested in K = 6, as this value exhibited a remarkable drop-off in the silhouette score when K was further raised beyond 6, indicating possible value from this granularity of clustering. Another reason why we were interested in the K value of 6 is that there are five distinct stages of CKD, as outlined in the introduction, as well as a normal/healthy control group (6 total). The clustering results with K = 2 are visualized using a principal component analysis (PCA) projection of the data in Figure 5, showing loose alignment with the main group status (CKD vs. not CKD). Furthermore, the PCA-projected clustering with K = 6 produced the results shown in Figure 6, which appear to loosely align with the severity of CKD, implying that the feature measurements included in this analysis hold considerable value, a finding with relevance for possible future applications of CKD staging. This implies the potential for extending machine learning technology in this domain, to be used in stage predictions and monitoring of disease progression. Figure 7 is provided as a comparative reference for Figure 5 and Figure 6, demonstrating the ground truth diagnoses of CKD and not CKD. This is further reinforced by the resultant cluster centers, with centroids illustrating characteristic underlying feature profiles, the results of which are provided in Figure 8, which demonstrates group-wise differences between the learned clusters.

3.5. Results Summary

In summary, our findings demonstrate that the inclusion of patient clinical characteristics is associated with a broad improvement in predictive accuracy across algorithms when compared to an analysis where they were excluded, an approach not considered in pre-existing studies. Our findings also indicate that multiple configurations of the LGBM consistently produce high-performing models with accuracies ranging from 99.4 to 100%. Cluster analysis results demonstrate the potential for machine-learning-based characterization of CKD severity, with potential long-term applications in CKD staging.

4. Discussion

4.1. Patient Clinical Characteristics

It is noteworthy from Figure 1 and Figure 2 that non CKD patients are distributed in just one of the two possible values for these clinical variables. Thus, these clinical patient characteristic variables have predictive potential in informing multivariable machine learning technologies, which would allow them to contribute towards making more accurate predictions of CKD vs. not CKD.

4.2. Impact of Patient Clinical Characteristics on Machine Learning

Our results indicate that the approach developed achieved 99.4 to 100% accuracy across multiple types of validation. When applying the same analyses but on the reduced version of the dataset, not including the set of clinical patient characteristics, the performance degrades to be in line with the results presented in previous literature studies, outlined in the introduction. These results imply value from the inclusion of clinical patient characteristics in the prediction/diagnosis of CKD by machine learning. This information could be of potential value to clinicians involved in the management of patients with CKD. Our leading feature-reduced models, based on LGBM and embedded feature selection, indicate useful predictive value from knowledge of the patient’s diabetes mellitus status (see Appendix A). Our findings also demonstrate that an unsupervised machine learning analysis, inclusive of patient clinical characteristics, loosely aligns with the severity of CKD. These findings lead us to propose that future work should involve the development of CKD characterization/staging technologies and potentially disease progression and treatment monitoring technologies, which may be of future clinical interest.

4.3. Comparative Machine Learning Results Across Studies

The leading models were obtained with the light gradient boosting machine (LGBM), noteworthy in that it is a lightweight computationally efficient alternative to the extreme gradient boosting machine, which was the leading machine learning technique in several studies on this topic [20,21,24]. High-quality results were obtainable with the LGBM inclusive of all of our features, but we were also able to produce similarly performing models with reduced feature sets, outlined in Appendix A and Appendix B. These findings imply that feature measurements not selected for inclusion in Appendix A and Appendix B are either not useful for CKD diagnosis by machine learning, or are redundant to other features included in the selection process. Our results compare favorably with existing research focused on the same dataset, including a rotation forest model yielding 99.2% accuracy [16], an approach based on the random forest yielding 97% accuracy [17], a support vector machine approach yielding 98.86% accuracy [18], an approach using the K-Nearest Neighbour classifier as well as the Extra Tree Classifier yielding 99% accuracy [19], and multiple approaches based on the XGBoost algorithm (https://github.com/dmlc/xgboost, accessed on 24 February 2024) achieving 94% accuracy [20], and 98.3% accuracy [21]. Thus, our finding that multiple configurations of the LGBM can produce 99.4 to 100% accuracy across two validation methods potentially adds value to the machine learning literature on this topic. It should also be noted that the LGBM is a computationally efficient ‘lightweight’ machine learning algorithm, so it offers advantages in terms of computational complexity.

4.4. Limitations and Future Work

Limitations of this study include that it was performed on a dataset with relatively few samples (n = 400). Machine learning algorithms often achieve peak performance on balanced datasets (i.e., in the context of this work, a balanced dataset would have an equal number of samples in the group of interest—CKD—and in the control group). Our dataset is not balanced, and as such, this can cause a potential bias, possibly reducing the predictive accuracy of the technologies trained (on unbalanced data). Acquiring a larger, balanced dataset may help reduce associated machine learning class imbalance bias in this application. Noteworthy, however, is that if you sample a large population, people with CKD are expected to be a small minority. As such, a balanced dataset with equal numbers of CKD and controls would not be expected to be representative of the population that such a tool would be tasked with diagnosing in the real world. Training the algorithm on a non-representative population relative to how the tool will be used in the real world can cause a different type of bias, in that the performance estimates from validation on the balanced dataset will necessarily shift to align with the reality of the technology’s use on a population in which CKD is comparatively rare. Future work will involve evaluating the approaches developed herein on larger and independently acquired datasets. Future work should also involve a direct comparison of high-performing competing algorithms applied in similar studies on CKD, such as XGBoost [20,21,24] and the Rotation Forest [16]. Strengths of this study include the use of open-source machine learning technology (https://github.com/stfxecutables/df-analyze/), inclusive of a standardized validation suite and an array of feature selection technologies, the consideration of the light gradient boosting algorithm, and the consideration of patient clinical characteristics in the models developed, providing a thorough analysis with high-performing predictive results. The inclusion of an unsupervised clustering analysis is another strength of this study.
Many people live with CKD and are unaware of their status until it reaches an advanced stage of development [4]. As such, the lack of efficient diagnostics applied to asymptomatic populations results in many cases of CKD going untreated, and thus, likely progressing towards more severe manifestations of the condition. In this context, AI technologies that can be easily applied to large populations of individuals, which has a lot of potential to improve the clinical management of CKD by identifying the condition in asymptomatic people through screening and thus supporting earlier treatment, which is commonly associated with better prognoses/outcomes.
The advantages of the approaches taken in this study and in the literature [16,17,18,19,20,21,22,23,24] include the creation of objective learning machines that are approaching perfect accuracy for the diagnosis of the condition. The major limitations of using artificial intelligence, at this time, include limited datasets and samples upon which to train the learning machines, potentially limiting their robust applicability in the real world. The limited explainability of existing machine learning technologies is another limitation of the use of the current methods, as when an AI technology makes a diagnosis, we want to be able to report the specific logical reasons why the AI came to its diagnostic conclusion. Unfortunately, this is largely an unsolved problem in AI technologies generally.
The work presented herein has the potential to assist with early diagnoses, potentially accurately predicting patient CKD status in the early stages of disease progression, which could theoretically result in earlier interventions and thus an improved standard of patient care and improved prognoses/outcomes. While the population in this study is relatively small (n = 400), such high-performing technologies could be the subject of future research in CKD screening, potentially helping to identify the disease prior to patients exhibiting symptoms, which has tremendous potential for reducing the morbidity associated with the condition. Machine learning clustering results demonstrate the potential for further characterizing CKD into stages and creating technologies that can assist in the monitoring of disease and treatment progression. Future work will investigate the staging of CKD with a dataset that contains ground truth knowledge of CKD stages, established by clinical experts in the field, which was unavailable in our dataset.

5. Conclusions

In summary, the progression of chronic kidney disease can be managed if detected early, which can help prevent damage to the kidney, and thus the development of technologies that can assist in early detection has considerable potential for improving the standard of patient care in CKD. This study highlights the efficacy of artificial intelligence and machine learning in diagnosing chronic kidney disease. Using public domain software, we demonstrated that the light gradient boosting machine (LGBM) with embedded feature selection performed the best across all metrics, with accuracies ranging from 99.4 to 100% across multiple validation trials. Cluster analysis results demonstrate potential towards the creation of staging technology for disease characterization, as well as treatment and disease progression technologies.

Author Contributions

Conceptualization, J.F., P.E., A.K.S. and J.L.; methodology, J.F., P.E. and A.K.S.; df-analyze software, D.B. and J.L.; validation, D.B. and J.L.; formal analysis, J.F., P.E., A.K.S. and J.L.; investigation, J.F., P.E. and A.K.S.; resources, D.B. and J.L.; data curation, J.F., P.E. and A.K.S.; writing—original draft preparation, P.E., J.F. and A.K.S.; writing—review and editing, P.E., J.F. and J.L.; visualization, J.F.; supervision, J.L.; project administration, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financially supported by a Canada Foundation for Innovation grant, a Nova Scotia Research and Innovation Trust grant, an NSERC Discovery grant, a Compute Canada Resource Allocation, and a Nova Scotia Health Authority grant to J.L.

Data Availability Statement

The dataset used in this study is publicly available and can be accessed from Kaggle at https://www.kaggle.com/datasets/mansoordaku/ckdisease and the University of California Irvine (UCI) Machine Learning Repository at https://archive.ics.uci.edu/ml/datasets/Chronic_Kidney_Disease (accessed on 24 February 2024). No new data were created or collected specifically for this study. Since this was a retrospective analysis of public domain data, no institutional review board approval was necessary.

Acknowledgments

We would like to extend our heartfelt thanks to St. Francis Xavier University and the Department of Computer Science for providing the computational resources and facilities that made this research possible. We would like to express our sincere gratitude to Jacob Levman for his invaluable guidance and supervision throughout this project. His expertise and insight were instrumental in our research, and his support with both the technical aspects and report preparation was greatly appreciated; without him all of this would not have been possible. We are also grateful to Derek Berger and Jacob Levman for their key role in developing and maintaining the df-analyze software, which was fundamental to our analysis. Finally, we specifically extend our gratitude to all those who offered technical support and guidance throughout the project such as Xuchen Guo.

Conflicts of Interest

Levman is founder of Time Will Tell Technologies, Inc. The authors declare no relevant conflicts of interest. The funders had no role in the design of this study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A

The results of embedded feature selection (embed_lgbm) with the LGBM algorithm on the entire dataset.
# Wrapper-Based Feature Selection Summary
Wrapper model: LGBM
## Selected Features
[‘sod_NAN’, ‘rc_NAN’, ‘age’, ‘sg’, ‘al’, ‘bgr’, ‘sc’, ‘sod’, ‘pot’, ‘hemo’, ‘pcv’, ‘wc’, ‘dm_yes’, ‘pc_nan’, ‘rbc_normal’]
## Selection scores (Importances: Larger magnitude = More important)
| feature | score |
|:------------|----------:|
| age_NAN | 0.000e+00 |
| bp_NAN | 0.000e+00 |
| sg_NAN | 0.000e+00 |
| al_NAN | 0.000e+00 |
| su_NAN | 0.000e+00 |
| bgr_NAN | 0.000e+00 |
| bu_NAN | 0.000e+00 |
| sc_NAN | 0.000e+00 |
| sod_NAN | 1.700e+01 |
| pot_NAN | 1.200e+01 |
| hemo_NAN | 6.000e+00 |
| pcv_NAN | 2.000e+00 |
| wc_NAN | 3.000e+00 |
| rc_NAN | 2.400e+01 |
| age | 2.800e+01 |
| bp | 1.200e+01 |
| sg | 3.800e+01 |
| al | 3.800e+01 |
| su | 0.000e+00 |
| bgr | 2.500e+01 |
| bu | 1.100e+01 |
| sc | 8.000e+01 |
| sod | 2.600e+01 |
| pot | 2.700e+01 |
| hemo | 1.160e+02 |
| pcv | 2.400e+01 |
| wc | 2.000e+01 |
| rc | 1.100e+01 |
| ane_yes | 0.000e+00 |
| ane_nan | 0.000e+00 |
| appet_poor | 5.000e+00 |
| appet_nan | 0.000e+00 |
| ba_present | 0.000e+00 |
| ba_nan | 0.000e+00 |
| cad_yes | 0.000e+00 |
| cad_nan | 0.000e+00 |
| dm_yes | 1.800e+01 |
| dm_nan | 0.000e+00 |
| htn_yes | 1.300e+01 |
| htn_nan | 0.000e+00 |
| pc_normal | 3.000e+00 |
| pc_nan | 1.700e+01 |
| pcc_present | 0.000e+00 |
| pcc_nan | 0.000e+00 |
| pe_yes | 0.000e+00 |
| pe_nan | 0.000e+00 |
| rbc_normal | 1.100e+02 |
| rbc_nan | 4.000e+00 |

Appendix B

The results of filter-prediction-based feature selection (pred) with the LGBM algorithm on the entire dataset.
# Filter-Based Feature Selection Summary
## Selected Features
[‘hemo’, ‘pcv’, ‘rc’, ‘sc’, ‘sg’, ‘al’, ‘sod’, ‘rbc’, ‘htn’, ‘dm’, ‘pc’, ‘ba’]
## Selection Prediction Scores
### Continuous Features (Accuracy: Higher = More important)
| feature | acc |
|:----------|----------:|
| hemo | 2.750e-01 |
| pcv | 2.583e-01 |
| rc | 2.208e-01 |
| sc | 2.083e-01 |
| sg | 1.417e-01 |
| al | 1.417e-01 |
| sod | 1.333e-01 |
| bgr | 6.667e-02 |
| bp | 3.750e-02 |
| age | 3.333e-02 |
| bu | 2.917e-02 |
| su | 0.000e+00 |
| pot | 0.000e+00 |
| wc | 0.000e+00 |
### Categorical Features (Accuracy: Higher = More important)
| feature | acc |
|:----------|----------:|
| rbc | 2.000e-01 |
| htn | 1.125e-01 |
| dm | 8.333e-02 |
| pc | 2.917e-02 |
| ba | 1.250e-02 |
| pcc | 1.250e-02 |
| ane | 0.000e+00 |
| appet | 0.000e+00 |
| cad | 0.000e+00 |
| pe | 0.000e+00 |

References

  1. National Institute of Diabetes and Digestive and Kidney Diseases. Your Kidneys & How They Work—National Institute of Diabetes and Digestive and Kidney Diseases. Available online: https://www.niddk.nih.gov/health-information/kidney-disease/kidneys-how-they-work (accessed on 20 February 2024).
  2. Johns Hopkins Hospital. Chronic Kidney Disease | Johns Hopkins Medicine. Available online: https://www.hopkinsmedicine.org/health/conditions-and-diseases/chronic-kidney-disease (accessed on 24 February 2024).
  3. National Kidney Foundation. Chronic Kidney Disease (CKD)—Symptoms, Causes, Treatment | National Kidney Foundation. Available online: https://www.kidney.org/atoz/content/about-chronic-kidney-disease (accessed on 24 February 2024).
  4. Zhao, J.; Zhang, Y.; Qiu, J.; Zhang, X.; Wei, F.; Feng, J.; Chen, C.; Zhang, K.; Feng, S.; Li, W.D. An early prediction model for chronic kidney disease. Sci. Rep. 2022, 12, 2765. [Google Scholar] [CrossRef] [PubMed]
  5. World Health Organization. The Top 10 Causes of Death—World Health Organization. Available online: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death (accessed on 24 February 2024).
  6. Estimating Glomerular Filtration Rate. National Institute of Diabetes and Digestive and Kidney Diseases. Available online: https://www.niddk.nih.gov/health-information/professionals/clinical-tools-patient-management/kidney-disease/laboratory-evaluation/glomerular-filtration-rate/estimating (accessed on 3 February 2024).
  7. American Kidney Fund. Stage 1 of Chronic Kidney Disease CKD: Causes, Symptoms and Treatment. Available online: https://www.kidneyfund.org/all-about-kidneys/stages-kidney-disease/stage-1-chronic-kidney-disease (accessed on 24 February 2024).
  8. American Kidney Fund. Stage 2 Chronic Kidney Disease (CKD). Available online: https://www.kidneyfund.org/all-about-kidneys/stages-kidney-disease/stage-2-chronic-kidney-disease-ckd (accessed on 24 February 2024).
  9. American Kidney Fund. Stage 3 Chronic Kidney Disease (CKD). Available online: https://www.kidneyfund.org/all-about-kidneys/stages-kidney-disease/stage-3-chronic-kidney-disease-ckd (accessed on 24 February 2024).
  10. American Kidney Fund. Stage 4 Chronic Kidney Disease (CKD). Available online: https://www.kidneyfund.org/all-about-kidneys/stages-kidney-disease/stage-4-chronic-kidney-disease-ckd (accessed on 24 February 2024).
  11. American Kidney Fund. Stage 5 Chronic Kidney Disease (CKD). Available online: https://www.kidneyfund.org/all-about-kidneys/stages-kidney-disease/stage-5-chronic-kidney-disease-ckd (accessed on 24 February 2024).
  12. Cleveland Clinic. Chronic Kidney Disease (CKD): Symptoms & Treatment. Available online: https://my.clevelandclinic.org/health/diseases/15096-chronic-kidney-disease (accessed on 7 May 2023).
  13. Chen, T.K.; Knicely, D.H.; Grams, M.E. Chronic Kidney Disease Diagnosis and Management A Review. National Library of Medicine. Available online: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7015670 (accessed on 11 March 2024).
  14. Lee, H.; Yun, D.; Yoo, J.; Yoo, K.; Kim, Y.C.; Kim, D.K.; Oh, K.H.; Joo, K.W.; Kim, Y.S.; Kwak, N.; et al. Deep learning model for real-time prediction of intradialytic hypotension. Clin. J. Am. Soc. Nephrol. 2021, 16, 396–406. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, Z.; Ho, K.M.; Hong, Y. Machine learning for the prediction of volume responsiveness in patients with oliguric acute kidney injury in critical care. Crit. Care 2019, 23, 112. [Google Scholar] [CrossRef] [PubMed]
  16. Dritsas, E.; Trigka, M. Machine Learning Techniques for Chronic Kidney Disease Risk Prediction. Big Data Cogn. Comput. 2022, 6, 98. [Google Scholar] [CrossRef]
  17. Yashfi, S.; Islam, M.; Pritilata; Sakib, N.; Islam, T.; Shahbaaz, M.; Pantho, S. Risk Prediction Of Chronic Kidney Disease Using Machine Learning Algorithms. In Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 1–3 July 2020. [Google Scholar] [CrossRef]
  18. Chittora, P.; Sandeep, C.; Chakrabarti, P.; Kumawat, G.; Chakrabarti, T.; Leonowicz, Z.; Jasiński, M.; Jasiński, Ł.; Gono, R.; Jasińska, E.; et al. Prediction of Chronic Kidney Disease—A Machine Learning Perspective. IEEE Access 2021, 9, 17312–17334. [Google Scholar] [CrossRef]
  19. Baidya, D.; Umaima, U.; Islam, M.N.; Shamrat, F.M.J.M.; Pramanik, A.; Rahman, M.S. A Deep Prediction of Chronic Kidney Disease by Employing Machine Learning Method. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 28–30 April 2022; pp. 1305–1310. [Google Scholar] [CrossRef]
  20. Abdur, M.; Rahat, R.; Cao, D.M.; Tayaba, M.; Ghosh, B.P.; Ayon, H.; Nobe Nur Akter, T.; Rahman, M.; Bhuiyan, M.S. Comparing Machine Learning Techniques for Detecting Chronic Kidney Disease in Early Stage. J. Comput. Sci. Technol. Stud. 2024, 6, 20–32. [Google Scholar] [CrossRef]
  21. Islam, M.A.; Majumder, M.Z.H.; Hussein, M.A. Chronic kidney disease prediction based on machine learning algorithms. J. Pathol. Inform. 2023, 14, 100189. [Google Scholar] [CrossRef] [PubMed]
  22. Debal, D.A.; Sitote, T.M. Chronic kidney disease prediction using machine learning techniques. J. Big Data 2022, 9, 109. [Google Scholar] [CrossRef]
  23. Ghosh, S.K.; Khandoker, A.H. A machine learning driven nomogram for predicting chronic kidney disease stages 3–5. Sci. Rep. 2023, 13, 21613. [Google Scholar] [CrossRef] [PubMed]
  24. Xiao, J.; Ding, R.; Xu, X.; Guan, H.; Feng, X.; Sun, T.; Zhu, S.; Ye, Z. Comparison and development of machine learning tools in the prediction of chronic kidney disease progression. J. Transl. Med. 2019, 17, 119. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  25. Rubini, L.; Soundarapandian, P.; Eswaran, P. Chronic Kidney Disease; UCI Machine Learning Repository: Espoo, Finland, 2015. [Google Scholar] [CrossRef]
  26. DM-Berger. stfxecutables/df-analyze: Sensitivity/Specificity Update; Zenodo: Geneva, Switzerland, 2021. [Google Scholar] [CrossRef]
  27. Levman, J.; Jennings, M.; Rouse, E.; Berger, D.; Kabaria, P.; Nangaku, M.; Gondra, I.; Takahashi, E. A morphological study of schizophrenia with Magnetic Resonance Imaging, advanced analytics, and Machine Learning. Front. Neurosci. 2022, 16, 926426. [Google Scholar] [CrossRef] [PubMed]
  28. Wella, Y.; Okfalisa, O.; Insani, F.; Saeed, F.; Hussin, A. Service quality dealer identification: The optimization of K-Means clustering. SINERGI 2023, 27, 433. [Google Scholar] [CrossRef]
Figure 1. Bar plots of patient clinical characteristics (abnormal appetite, pedal edema, anemia) and how they are distributed in patients with chronic kidney disease (CKD) and patients without (NOTCKD).
Figure 1. Bar plots of patient clinical characteristics (abnormal appetite, pedal edema, anemia) and how they are distributed in patients with chronic kidney disease (CKD) and patients without (NOTCKD).
Electronics 13 04326 g001
Figure 2. Bar plots of patient characteristics (hypertension—HTN, diabetes mellitus—DM, and coronary artery disease—CAD) and how they are distributed in patients with chronic kidney disease (CKD) and patients without (NOTCKD).
Figure 2. Bar plots of patient characteristics (hypertension—HTN, diabetes mellitus—DM, and coronary artery disease—CAD) and how they are distributed in patients with chronic kidney disease (CKD) and patients without (NOTCKD).
Electronics 13 04326 g002
Figure 3. Silhouette score applied to the dataset. This plot was used to assess the apparent value of clustering results at varying numbers of clusters. The highest values and local peaks were of potential interest.
Figure 3. Silhouette score applied to the dataset. This plot was used to assess the apparent value of clustering results at varying numbers of clusters. The highest values and local peaks were of potential interest.
Electronics 13 04326 g003
Figure 4. Calinski–Harabasz score plot for the same dataset. This plot was used to assess the apparent value of clustering results at varying numbers of clusters. The highest values and local peaks were of potential interest.
Figure 4. Calinski–Harabasz score plot for the same dataset. This plot was used to assess the apparent value of clustering results at varying numbers of clusters. The highest values and local peaks were of potential interest.
Electronics 13 04326 g004
Figure 5. Plot of resulting clusters, using K-means algorithm with PCA and K = 2.
Figure 5. Plot of resulting clusters, using K-means algorithm with PCA and K = 2.
Electronics 13 04326 g005
Figure 6. Plot of resulting clusters, using K-means algorithm with PCA and K = 6.
Figure 6. Plot of resulting clusters, using K-means algorithm with PCA and K = 6.
Electronics 13 04326 g006
Figure 7. Visualization of PCA components’ projection based on the class CKD or not CKD.
Figure 7. Visualization of PCA components’ projection based on the class CKD or not CKD.
Electronics 13 04326 g007
Figure 8. Centroids of each cluster from K-means analysis of the CKD dataset, K = 6. Note that group numbers are arbitrarily assigned by the K-means algorithm.
Figure 8. Centroids of each cluster from K-means analysis of the CKD dataset, K = 6. Note that group numbers are arbitrarily assigned by the K-means algorithm.
Electronics 13 04326 g008
Table 1. CKD dataset description for numerical variables.
Table 1. CKD dataset description for numerical variables.
VariableDefinitionTypeValueClass of Interest = CKDClass of Not Interest = NOTCKD
MeanStandard DeviationMeanStandard Deviation
age—ageAge of the patientNumericalyears54.517.446.315.5
bp—blood pressure (Diastolic)Pressure of blood pumped by heart into wall’s vesselsNumericalmm/Hg79.615.271.48.6
sg—specific gravitySpecific indicator of renal function; varies between 1.005 and 1.025 for humansNumerical(1.005, 1.010, 1.015, 1.020,
1.025)
1.00.01.00.0
bgr—blood glucose randomMeasure of the glucose in the blood at the moment of the testNumericalmgs/dL175.492.1107.718.6
bu—blood ureaMeasurement of urea nitrogenic blood or serum, a major indicator of kidney failureNumericalmgs/dL72.458.632.711.4
sc—serum creatinineCreatinine is a waste component that is removed by the kidneys from the bloodNumericalmgs/dL4.47.00.90.3
sod—sodiumConcentration of sodium in bloodNumericalmEq/L133.912.4141.74.8
pot—potassiumConcentration of potassium in bloodNumericalmEq/L4.94.34.30.6
hemo—hemoglobinHemoglobin is a protein in charge of transport of oxygen in the red blood cellsNumericalgms10.62.215.21.3
pcv—packed cell volumeProportion of blood cells within serumNumericalpercentage32.97.246.34.1
wc—white blood cell countCount of white blood cellsNumericalcells/cmm9069.53580.57687.31833.2
rc—red blood cell countCount of red blood cellsNumericalmillions/cmm3.90.95.40.6
Table 2. CKD dataset description for categorical variables.
Table 2. CKD dataset description for categorical variables.
VariableDefinitionCounts for Each ValueValues as a Percentage (%) of All
Samples
al—albuminMeasurement of albumin protein in blood(199:0, 44:1, 43:2, 43:3, 24:4, 1:5, 46:undefined)(49.75% 0, 11% 1, 10.75% 2, 10.75% 3, 6% 4, 0.25% 5, 11.5% undefined)
su—sugarMeasurement of sugar (glucose) in blood(290:0, 13:1, 18:2, 14:3, 13:4, 3:5, 49:undefined)(72.5% 0, 3.25% 1, 4.5% 2, 3.5% 3, 3.25% 4, 0.75% 5, 12.25% undefined)
rbc—red blood cellsAssessment of red blood cells(201 normal, 47 abnormal, 152 undefined)(50.25% normal, 11.75% abnormal, 38% undefined)
pc—pus cellAccumulation of dead white blood cells in the urine(259 normal, 76 abnormal, 65 undefined)(64.75% normal, 19% abnormal, 16.25% undefined)
pcc—pus cell clumpsPresence of dead cells in urine, which indicates kidney infection, or a sexually transmitted disease(42 present, 354 not present, 4 undefined)(10.5% present, 88.5% not present, 1% undefined)
ba—bacteriaPresence of bacteria in urine (22 present, 374 not present, 4 undefined)(5.5% present, 93.5% not present, 1% undefined)
htn—hypertensionPatient with hypertension diagnosed(147 yes, 251 no, 2 undefined)(36.75% yes, 62.75% no, 0.50% undefined)
dm—diabetes mellitusFailure of the body to react to insulin to control blood glucose levels(137 yes, 261 no, 2 undefined)(34.25% yes, 65.25% no, 0.50% undefined)
cad—coronary artery diseaseNarrow arteries can cause obstructions of blood flow(34 yes, 364 no, 2 undefined)(8.5% yes, 91% no, 0.50% undefined)
appet—appetiteAbnormal appetite(317 good, 82 poor, 1 undefined)(79.25% good, 20.5% poor, 0.25% undefined)
pe—pedal edemaExcess fluid in the lower extremities or knees(76 yes, 323 no, 1 undefined)(19% yes, 80.75% no, 0.25% undefined)
ane—anemiaReduction in red blood cells(60 yes, 339 no, 1 undefined)(15% yes, 84.75% no, 0.25% undefined)
class—diagnosisThe target value to predict(250 CKD, 150 not CKD)(62.5% CKD, 37.5% not CKD)
Table 3. Holdout set performance for the complete dataset.
Table 3. Holdout set performance for the complete dataset.
ModelSelectionEmbed_SelectorACCAUROCBAL-ACCF1NPVPPVSensSpec
lgbmnonenone11111111
lgbmembed_lgbmlgbm11111111
lgbmembed_linearlinear0.99410.9950.9930.98410.9951
lgbmassocnone0.99410.9950.9930.98410.9951
rfassocnone0.99410.9920.99310.990.9920.983
lrwrapnone0.9880.9980.990.9870.96810.991
rfembed_lgbmlgbm0.9880.9990.9870.9870.9830.990.9870.983
lgbmprednone0.98810.990.9870.96810.991
mlpprednone0.98810.990.9870.96810.991
sgdwrapnone0.9880.9980.990.9870.96810.991
mlpwrapnone0.9810.9990.9850.980.95210.9851
rfprednone0.9810.9990.9780.980.9830.980.9780.967
lrprednone0.9810.9980.9820.980.9670.990.9820.983
rfnonenone0.9690.9970.9620.9660.9820.9610.9620.933
knnprednone0.9630.970.970.9610.90910.971
lgbmwrapnone0.9560.9960.9580.9540.9210.9790.9580.967
rfembed_linearlinear0.9560.9920.9620.9540.9080.9890.9620.983
rfwrapnone0.9380.9890.9370.9340.9030.9590.9370.933
sgdprednone0.9060.8880.8880.8970.9250.8970.8880.817
knnassocnone0.8630.9270.860.8550.7970.9060.860.85
knnembed_linearlinear0.8440.8350.8350.8340.7870.8790.8350.8
knnembed_lgbmlgbm0.8380.9110.830.8280.7740.8780.830.8
mlpembed_linearlinear0.8120.9190.7730.7860.8410.8020.7730.617
knnnonenone0.7750.770.770.7640.6820.840.770.75
knnwrapnone0.7690.8540.7850.7650.6460.8890.7850.85
mlpembed_lgbmlgbm0.7690.8820.7780.7630.6530.8710.7780.817
sgdembed_linearlinear0.6880.670.670.6690.5810.7550.670.6
sgdassocnone0.6810.7370.6480.6510.5850.7290.6480.517
lrnonenone0.6750.830.5830.5590.7220.6690.5830.217
sgdnonenone0.6620.6540.6270.6290.5580.7130.6270.483
lrembed_linearlinear0.6560.8210.5480.4920.7780.6490.5480.117
lrassocnone0.6440.8190.5320.4620.7140.6410.5320.083
lrembed_lgbmlgbm0.6310.7610.5150.430.60.6320.5150.05
mlpnonenone0.6250.610.50.385nan0.6250.50
mlpassocnone0.6250.660.50.385nan0.6250.50
dummyembed_lgbmlgbm0.6250.50.50.385nan0.6250.50
dummywrapnone0.6250.50.50.385nan0.6250.50
dummyprednone0.6250.50.50.385nan0.6250.50
sgdembed_lgbmlgbm0.6250.5930.5930.5940.50.6920.5930.467
dummynonenone0.6250.50.50.385nan0.6250.50
dummyembed_linearlinear0.6250.50.50.385nan0.6250.50
dummyassocnone0.6250.50.50.385nan0.6250.50
Table 4. Five-fold holdout set performance for complete dataset.
Table 4. Five-fold holdout set performance for complete dataset.
ModelSelectionEmbed_SelectorACCAUROCBAL-ACCF1NPVPPVSensSpec
lgbmprednone11111111
lgbmembed_lgbmlgbm0.99410.9920.99310.990.9920.983
lgbmassocnone0.99410.9920.99310.990.9920.983
lgbmnonenone0.99410.9920.99310.990.9920.983
lgbmembed_linearlinear0.99410.9920.99310.990.9920.983
sgdwrapnone0.98810.990.9870.97110.991
lgbmwrapnone0.9810.9970.9820.980.9690.990.9820.983
rfembed_linearlinear0.98110.9850.980.95410.9851
lrwrapnone0.98110.9820.980.9710.990.9820.983
lrprednone0.9750.9950.9770.9740.9550.990.9770.983
rfwrapnone0.9690.9940.9720.9670.9370.990.9720.983
rfnonenone0.9630.9950.9570.960.9660.9620.9570.933
rfassocnone0.9560.9980.9620.9540.9130.990.9620.983
rfembed_lgbmlgbm0.950.9930.9570.9480.9030.990.9570.983
knnprednone0.950.9570.9570.9480.9030.990.9570.983
rfprednone0.9440.9930.9480.9410.9010.980.9480.967
sgdprednone0.90.8970.8970.8940.8640.9320.8970.883
knnembed_lgbmlgbm0.8380.910.8330.830.7840.8830.8330.817
sgdembed_linearlinear0.8190.8150.8150.8060.7750.8850.8150.8
sgdassocnone0.8120.8890.8070.8010.7440.8730.8070.783
sgdnonenone0.8120.8720.7930.7970.7760.8370.7930.717
knnassocnone0.8120.9290.8330.8090.6930.9390.8330.917
knnembed_linearlinear0.7940.7850.7850.7810.7310.8490.7850.75
knnnonenone0.7810.7820.7820.7710.6890.8620.7820.783
sgdembed_lgbmlgbm0.7250.7150.710.7080.6480.7860.710.65
knnwrapnone0.7060.7910.7380.7020.570.9050.7380.867
lrnonenone0.650.9070.5330.45210.6410.5330.067
lrembed_lgbmlgbm0.6370.8550.520.4240.750.6360.520.05
lrembed_linearlinear0.6310.9080.5080.40210.6290.5080.017
dummyembed_lgbmlgbm0.6250.50.50.385nan0.6250.50
lrassocnone0.6250.9070.50.385nan0.6250.50
dummywrapnone0.6250.50.50.385nan0.6250.50
dummyprednone0.6250.50.50.385nan0.6250.50
dummynonenone0.6250.50.50.385nan0.6250.50
dummyembed_linearlinear0.6250.50.50.385nan0.6250.50
dummyassocnone0.6250.50.50.385nan0.6250.50
mlpassocnone0.5750.6420.50.3620.3750.6250.50.2
mlpprednone0.4750.7350.50.3170.3750.6250.50.6
mlpembed_linearlinear0.4750.6370.50.3170.3750.6250.50.6
mlpembed_lgbmlgbm0.4750.630.50.3170.3750.6250.50.6
mlpwrapnone0.4440.7240.4720.3680.350.5650.4720.583
mlpnonenone0.3750.6520.50.2730.375nan0.51
Table 5. Holdout set performance for dataset without clinical patient characteristics.
Table 5. Holdout set performance for dataset without clinical patient characteristics.
ModelSelectionEmbed_SelectorACCAUROCBAL-ACCF1NPVPPVSensSpec
lgbmnonenone11111111
rfwrapnone0.99410.9950.9930.98410.9951
lgbmwrapnone0.99410.9950.9930.98410.9951
lgbmprednone0.99410.9950.9930.98410.9951
lgbmassocnone0.99410.9950.9930.98410.9951
knnwrapnone0.9880.9990.990.9870.96810.991
mlpprednone0.98810.9870.9870.9830.990.9870.983
lrwrapnone0.9880.9990.9870.9870.9830.990.9870.983
rfassocnone0.98810.9870.9870.9830.990.9870.983
lgbmembed_lgbmlgbm0.98810.990.9870.96810.991
sgdwrapnone0.9880.9990.9870.9870.9830.990.9870.983
mlpwrapnone0.98810.990.9870.96810.991
rfprednone0.9880.9980.990.9870.96810.991
lrprednone0.9810.9990.9780.980.9830.980.9780.967
rfnonenone0.9810.9990.9820.980.9670.990.9820.983
knnprednone0.9750.9770.9770.9740.9520.990.9770.983
lgbmembed_linearlinear0.9630.9880.9570.960.9660.9610.9570.933
rfembed_lgbmlgbm0.9560.9930.9520.9530.9490.960.9520.933
rfembed_linearlinear0.9440.9880.9380.940.9320.950.9380.917
lrembed_lgbmlgbm0.9250.980.9170.9190.9140.9310.9170.883
sgdprednone0.9120.9350.9070.9070.8830.930.9070.883
sgdembed_lgbmlgbm0.9060.9690.8920.8980.9090.9050.8920.833
knnembed_linearlinear0.8690.9320.8550.8590.8420.8830.8550.8
knnembed_lgbmlgbm0.850.920.8470.8420.7810.8960.8470.833
knnassocnone0.8120.7970.7970.7990.7590.8430.7970.733
mlpassocnone0.80.8850.7470.7620.8890.7740.7470.533
knnnonenone0.7870.8770.7830.7770.6970.8510.7830.767
mlpembed_lgbmlgbm0.7750.970.70.70910.7350.70.4
mlpnonenone0.7750.8960.7830.7690.6620.8720.7830.817
mlpembed_linearlinear0.7620.8520.7770.7580.6410.8780.7770.833
lrnonenone0.7380.8650.670.6750.80.7230.670.4
lrassocnone0.7380.8650.670.6750.80.7230.670.4
lrembed_linearlinear0.7190.8660.6450.6450.7780.7070.6450.35
sgdassocnone0.6810.6320.6320.6350.6050.7090.6320.433
sgdembed_linearlinear0.6810.6350.6350.6390.60.7130.6350.45
sgdnonenone0.6620.790.6030.6030.5790.6890.6030.367
dummyembed_lgbmlgbm0.6250.50.50.385nan0.6250.50
dummywrapnone0.6250.50.50.385nan0.6250.50
dummyprednone0.6250.50.50.385nan0.6250.50
dummynonenone0.6250.50.50.385nan0.6250.50
dummyembed_linearlinear0.6250.50.50.385nan0.6250.50
dummyassocnone0.6250.50.50.385nan0.6250.50
Table 6. Five-fold holdout set performance for dataset without clinical patient characteristic variables.
Table 6. Five-fold holdout set performance for dataset without clinical patient characteristic variables.
ModelSelectionEmbed_SelectorACCAUROCBAL-ACCF1NPVPPVSensSpec
sgdwrapnone0.98110.9820.980.9690.990.9820.983
lrwrapnone0.98110.9820.980.9690.990.9820.983
knnwrapnone0.9750.9950.980.9740.94110.981
lgbmnonenone0.9750.9980.970.9730.9850.9720.970.95
lgbmassocnone0.9750.9970.970.9730.9850.9720.970.95
lgbmembed_lgbmlgbm0.9750.9980.970.9730.9850.9720.970.95
lgbmprednone0.9690.9970.9650.9660.9660.9710.9650.95
lgbmwrapnone0.9630.9960.9630.960.9380.9810.9630.967
rfassocnone0.9630.9970.960.960.950.9710.960.95
rfnonenone0.9560.9960.9580.9540.9220.9810.9580.967
lrprednone0.9560.9970.9580.9540.9220.9810.9580.967
lgbmembed_linearlinear0.9560.9920.9550.9530.9380.9710.9550.95
rfwrapnone0.9440.990.9420.940.920.960.9420.933
knnprednone0.9380.940.940.9340.8950.9690.940.95
rfprednone0.9310.9890.9280.9270.9090.9510.9280.917
rfembed_linearlinear0.9250.9820.9270.9210.8810.9590.9270.933
rfembed_lgbmlgbm0.9190.9920.9280.9150.8460.9810.9280.967
lrembed_lgbmlgbm0.9060.9730.9020.90.8730.930.9020.883
sgdprednone0.90.9220.8930.8940.8820.920.8930.867
sgdembed_lgbmlgbm0.8810.9550.8720.8730.8520.9030.8720.833
knnnonenone0.8440.9450.8720.8420.730.9880.8720.983
knnembed_linearlinear0.8440.950.8380.8340.7890.890.8380.817
knnembed_lgbmlgbm0.8310.9240.8350.8250.7650.9030.8350.85
sgdnonenone0.8060.8750.7950.7940.7460.850.7950.75
sgdassocnone0.7940.7850.7850.7820.7250.8450.7850.75
knnassocnone0.7870.7930.7930.780.6870.8790.7930.817
sgdembed_linearlinear0.7810.7620.7620.7640.7280.8150.7620.683
lrnonenone0.6310.8760.5120.4150.750.6310.5120.033
lrassocnone0.6310.8760.5120.4150.750.6310.5120.033
dummyembed_lgbmlgbm0.6250.50.50.385nan0.6250.50
dummywrapnone0.6250.50.50.385nan0.6250.50
dummyprednone0.6250.50.50.385nan0.6250.50
dummynonenone0.6250.50.50.385nan0.6250.50
dummyembed_linearlinear0.6250.50.50.385nan0.6250.50
dummyassocnone0.6250.50.50.385nan0.6250.50
lrembed_linearlinear0.6190.8720.4950.38200.6230.4950
mlpembed_lgbmlgbm0.5560.7240.5250.4020.2920.7090.5250.4
mlpprednone0.5370.8130.5430.40.4350.7020.5430.567
mlpwrapnone0.5250.7360.50.340.3750.6250.50.4
mlpnonenone0.4750.6190.50.3170.3750.6250.50.6
mlpembed_linearlinear0.4750.5790.50.3170.3750.6250.50.6
mlpassocnone0.3750.6330.50.2730.375nan0.51
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Figueroa, J.; Etim, P.; Shibu, A.K.; Berger, D.; Levman, J. Diagnosing and Characterizing Chronic Kidney Disease with Machine Learning: The Value of Clinical Patient Characteristics as Evidenced from an Open Dataset. Electronics 2024, 13, 4326. https://doi.org/10.3390/electronics13214326

AMA Style

Figueroa J, Etim P, Shibu AK, Berger D, Levman J. Diagnosing and Characterizing Chronic Kidney Disease with Machine Learning: The Value of Clinical Patient Characteristics as Evidenced from an Open Dataset. Electronics. 2024; 13(21):4326. https://doi.org/10.3390/electronics13214326

Chicago/Turabian Style

Figueroa, Juan, Patrick Etim, Adithyan Karanathu Shibu, Derek Berger, and Jacob Levman. 2024. "Diagnosing and Characterizing Chronic Kidney Disease with Machine Learning: The Value of Clinical Patient Characteristics as Evidenced from an Open Dataset" Electronics 13, no. 21: 4326. https://doi.org/10.3390/electronics13214326

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop