Next Article in Journal
The Effects of the Pericapsular Nerve Group Block on Postoperative Pain in Patients with Hip Fracture: A Multicenter Study
Next Article in Special Issue
The Role of Large Language Models (LLMs) in Providing Triage for Maxillofacial Trauma Cases: A Preliminary Study
Previous Article in Journal
Prenatal Diagnosis and Prognosis of Abdominal Arteriovenous Fistulae: A Comprehensive Case Series and Systematic Review
Previous Article in Special Issue
Performance Assessment of ChatGPT versus Bard in Detecting Alzheimer’s Dementia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analyzing Longitudinal Health Screening Data with Feature Ensemble and Machine Learning Techniques: Investigating Diagnostic Risk Factors of Metabolic Syndrome for Chronic Kidney Disease Stages 3a to 3b

1
Department of Healthcare Administration, College of Healthcare & Management, Asia Eastern University of Science and Technology, New Taipei City 220, Taiwan
2
Graduate Institute of Business Administration, Fu Jen Catholic University, New Taipei City 242, Taiwan
3
Department of Business Administration, Tamkang University, New Taipei City 251, Taiwan
4
Artificial Intelligence Development Center, Fu Jen Catholic University, New Taipei City 242, Taiwan
5
Department of Information Management, Fu Jen Catholic University, New Taipei City 242, Taiwan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2024, 14(8), 825; https://doi.org/10.3390/diagnostics14080825
Submission received: 27 February 2024 / Revised: 12 April 2024 / Accepted: 13 April 2024 / Published: 17 April 2024

Abstract

:
Longitudinal data, while often limited, contain valuable insights into features impacting clinical outcomes. To predict the progression of chronic kidney disease (CKD) in patients with metabolic syndrome, particularly those transitioning from stage 3a to 3b, where data are scarce, utilizing feature ensemble techniques can be advantageous. It can effectively identify crucial risk factors, influencing CKD progression, thereby enhancing model performance. Machine learning (ML) methods have gained popularity due to their ability to perform feature selection and handle complex feature interactions more effectively than traditional approaches. However, different ML methods yield varying feature importance information. This study proposes a multiphase hybrid risk factor evaluation scheme to consider the diverse feature information generated by ML methods. The scheme incorporates variable ensemble rules (VERs) to combine feature importance information, thereby aiding in the identification of important features influencing CKD progression and supporting clinical decision making. In the proposed scheme, we employ six ML models—Lasso, RF, MARS, LightGBM, XGBoost, and CatBoost—each renowned for its distinct feature selection mechanisms and widespread usage in clinical studies. By implementing our proposed scheme, thirteen features affecting CKD progression are identified, and a promising AUC score of 0.883 can be achieved when constructing a model with them.

1. Introduction

A sub-health condition (SHC) or sub-optimal health status refers to a condition characterized by decreased vitality, physiological function, and capacity for adaptation. However, it is yet to be medically diagnosed as a disease or functional somatic syndrome [1]. It is imperative to consider all SHC indicators to prevent chronic diseases and achieve better health outcomes. Metabolic syndrome (MetS) is a collection of indicators that define SHC risk and can assist in formulating strategies for preventing disease progression [2,3]. Metabolic factors, such as being overweight or obese and having hypertension, hyperlipidemia, and hyperglycemia, are critical metabolic changes that can increase the risk of chronic illness [4]. MetS increases the risk of developing various chronic diseases, such as a 2.5-fold higher risk of chronic kidney disease (CKD) [5], a 2.5-fold higher risk of myocardial infarction [4], a 2–4-fold higher risk of cardiovascular stroke, and a 5-fold higher risk of type II diabetes mellitus [6,7].
CKD is characterized by abnormal kidney function and is stratified into stages 1, 2, 3a, 3b, 4, and 5 according to the Kidney Disease Improving Global Outcomes’ (KDIGO) guideline [8]. Kidney diseases have become a major public health issue as they affect around 850 million individuals worldwide [9,10]. Patients with CKD often develop complications and MetS, accelerating their renal function deterioration, shortening the kidney lifespan, and ultimately increasing the incidence of CKD and intensifying its progression [11,12]. Both MetS and CKD are important risk factors for diverse complications [13,14]. Studies have demonstrated a positive correlation between MetS and CKD [15,16], and a MetS diagnosis effectively predicts CKD risk [16,17]. Among the existing studies, conventional statistical method usage is the approach that is commonly taken.
Machine learning (ML) approaches, being relatively unaffected by the limitations of conventional statistical methods that rely on predefined assumptions and hypotheses, have found widespread use in detecting and predicting diseases at various stages, demonstrating promising performance [18,19]. ML approaches can proficiently analyze latent and intricate relationships and information that underlie multiple predictor variables/risk factors and outcomes [20]. Based on the feature selection results obtained from ML methods, the employment of the variable ensemble rule (VER) can aid in assessing the predictor variables of models to improve analytical outcomes. The VER consolidates the selection results of different variables using various approaches or principles to enhance the robustness of variable selection outcomes [21].
Stage 3 CKD can be divided into Stages 3a and 3b, representing mild and moderate renal function impairment, respectively. Both substages are vital in assessing whether a patient should undergo kidney dialysis, and they are associated with different mortality risks and clinical features [22,23]. However, only a few studies have explored how metabolic syndrome (MetS) affects Stage 3a and 3b CKD in patients and their shared risk factors [17]. While clinical health data, in general, can accumulate into a substantial amount of big data, in practice, whether addressing preventive healthcare for chronic diseases, including MetS, or assisting in the assessment of CKD at stages 3a-3b for high-risk diagnosis, establishing a risk prediction model for clinical use requires the incorporation of more appropriate or rational limited longitudinal healthcare data into research analysis. Collecting such data is essential for building predictive models and thereby identifying relevant risk factors more accurately, facilitating specialized physicians in clinical diagnosis, and aiding in medical decision making.
When dealing with limited medical datasets that are small or medium-sized, establishing conditional data under various variable ensemble rules can be beneficial in model building. ML algorithms, when applied with various variable ensemble rules, can compensate for the limitation of a small sample size, achieving a simultaneous improvement in predictive capabilities. Existing CKD analyses are primarily based on the findings of cross-sectional studies [24,25], which have mainly discussed CKD and evaluated its risk factors. The analysis of health examination data requires consideration of trends in the continuous change of data. That is, when conducting longitudinal data analysis, it is essential to give precedence to scrutinizing the trend and variability of the data, rather than exclusively depending on baseline data. Therefore, we generated four extended variables to gather trend and variability information from each of the predictor variables. These extended variables can provide a wide range of information and can be used as predictor variables for constructing the ML prediction model.
Given the significance of MetS as a risk factor for CKD and its pertinent role in CKD development mechanisms over times, evaluating the risk factors for CKD in patients with MetS using longitudinal data is a vital step in effectively managing and preventing CKD. This study aimed to use ML methods and VER schemes to identify the important risk factors for CKD in longitudinal data for patients with MetS diagnosed with stages 3a or 3b CKD. It assesses six effective ML methods—random forest (RF), multivariate adaptive regression splines (MARS), least absolute shrinkage and selection operator (Lasso), extreme gradient boosting (XGBoost), gradient boosting with categorical features support (CatBoost), and light gradient boosting machine (LightGBM)—as they are already being successfully utilized in various healthcare and medical applications [26,27,28,29], and five VERs in feature engineering—maximum aggregation (MA), arithmetic mean aggregation (AMA), geometric mean aggregation (GMA), Borda count aggregation (BCA), and ranking mean aggregation (RMA). Using these methods, we develop an ML- and VER-based hybrid multiphase CKD prediction scheme for evaluating and consolidating the key risk factors for patients with MetS and CKD.
The proposed scheme first aggregates the scoring generated from corresponding ML methods via the five VERs. Because each machine learning model can provide feature importance scores on both numerical and categorical scales, the corresponding Variable Explanation Ratio (VER) is chosen based on these scores, allowing for the consideration of a wider range of information. Then, a union operation is employed to create a final selection of the most important features. By implementing the proposed scheme, we can reduce the complexity associated with a large number of features, thereby providing clinicians with crucial information to support medical decision making. Furthermore, as the proposed scheme can select important features, model performance can be improved when using these selected features.
The accumulation of clinical health data often leads to big data challenges. However, creating effective risk prediction models for chronic diseases, such as MetS and high-risk CKD stages 3a-3b requires, focused longitudinal healthcare data analysis, which is often limited to small datasets. The limited longitudinal healthcare data make it challenging to build effective models in health promotion fields. This study innovates by proposing an effective scheme to enhance predictive accuracy with traditional machine learning algorithms, even with smaller datasets. The proposed scheme can effectively find features influencing CKD progression while improving the performance of the model in classifying CKD progression transitioning from stage 3a to 3b.

2. Materials and Methods

2.1. Data

This study used the regular health examination records of 71,108 patients in the Mei Jhao (MJ) Health Checkup-Based Population Database (MJPD), a Taiwanese long-term and continuous patient follow-up database, from 2005 to 2017. This timeframe was chosen to accumulate a sufficient number of consecutive samples. This decision was driven by the focus on a high-risk population with both CKD stages 3a to 3b and MetS. They included 201,807 health examination indicators and questionnaire records. We identified patients with MetS and stage 3a or 3b CKD using the MetS definition of the Health Promotion Administration (HPA) of the Ministry of Health and Welfare of Taiwan [30] and the KDIGO guidelines and references. This study was approved by the Institutional Review Board of the Far Eastern Memorial Hospital (approval number: 110027-E; approval date: 3 March 2021) and the MJ Health Research Foundation (authorization code: MJHRF2023004A) and was registered at ClinicalTrials.gov ID:NCT05225454, https://beta.clinicaltrials.gov/study/NCT05225454 (accessed on 27 February 2024).

2.2. Definitions of the Longitudinal Variables and Subjects

This study constructed the longitudinal variables and their data by using each subject’s first two examination results to predict their third CKD examination results. We collected subjects’ 12-year examination records (from 2005 to 2017). Consistent with the prediction goals and conversion principles of the longitudinal data (each subject completed one examination annually), we excluded subjects with less than 3 or more than 12 examination records, leaving 33,533 subjects (125,641 health examination records). The data may include patients on dialysis, and they were not within the scope of this study, so we excluded 11 subjects whose estimated glomerular filtration rate (eGFR) was below 15 (42 records in total), leaving 33,522 subjects (125,599 records). We grouped these subjects based on the definition and eGFR criterion of CKD into an experimental group, a control group, and an “others” group. The experimental group comprised subjects with two consecutive eGFR values ≥60 in their health examination records; in total, 33 subjects met the definition of stage 3a CKD as their eGFR was ≥30 and <45. The control group comprised 302 subjects who met the definition of stage 3b CKD. The remaining 33,187 subjects were placed in the others group as they did not meet the criteria for the experimental or control group. After a multiphase processing of all subjects’ data, there were 335 eligible subjects. The process of identifying the longitudinal subjects is shown in Figure 1.
Among all 335 subjects, for each one, a total of three records were collected. As the aim of this study was to predict the relationship between each subject’s third CKD examination result and their risk factors, each subject’s previous two examination results were used as longitudinal predictor variables. Table 1 provides detailed descriptions and definitions of the predictor and target variables in the longitudinal data. The predictor variable V i , t represents the result of the i th variable at the t th examination (for example, the variable V 1 , 1 is the BF value at the first examination), and the objective variable Y represents the CKD result at the third examination. This study used 19 risk factors as predictor variables and can be further defined as Equation (1).
V i , t ,   i ,   t N where   i = 1 , 2 , , 19 ; t = 1 , 2 .
To generate extended variables, the four statistics involving the closest value, mean value, standard deviation (SD) value, and difference value of a predictor variable are considered. The closest value of a predictor variable uses the subjects’ latest examination results, which is the second examination in this study ( V i , 2 ). The predictor variable ( V i C ) generated based on the closest value is defined as Equation (2). For example, V 1 C is the BF record ( V 1 , 2 ) at the second examination, and it can be abbreviated as BF(C).The predictor variable ( V i M ) generated using the mean value of a predictor variable is the mean of the previous two examination results ( V i , 1 , V i , 2 ) , and it can be defined as Equation (3). For example, V 1 M is generated by obtaining the mean BF value of the last two examinations, and it can be abbreviated as BF(M). The predictor variable ( V i S ) is the SD of the last two examination results   ( V i , 1 , V i , 2 ) of a predictor variable, and it can be defined as Equation (4). For example, V 1 S is generated by obtaining the SD of the BF result ( V 1 S )   at the first and second examinations. V 1 S can be abbreviated as BF(S). A predictor variable ( V i D ) is the difference between the last two examination results   ( V i , 1 , V i , 2 ) of a predictor variable, and it can be defined as Equation (5). For example, V 1 D   is generated by subtracting the BF results of the first and second examinations. V 1 D can be abbreviated as BF(D).
All four of the statistical approaches to generating extended variables are applied to all 19 predictor variables to generate the predictor variables for analysis. Therefore, a total of 76 predictor variables are considered, and they are also used to construct the ML prediction model. The demographics of the 19 variables from the subjects’ latest examination ( V i C ) are shown in Table S1 in the Supplementary Materials.
V i C = V i , 2
V i M = V i , 1 + V i , 2 2
V i S = ( V i , 1 V i M ) 2 + ( V i , 2 V i M ) 2 2 1
V i D = V i , 2 V i , 1

2.3. Proposed Multiphase Hybrid Risk Factor Evaluation Scheme

In order to predict CKD outcomes and identify the key risk factors for CKD, this study proposes a multiphase hybrid CKD prediction scheme grounded in six ML algorithms (RF, MARS, Lasso, XGBoost, CatBoost, and LightGBM) that utilize the longitudinal variables generated in the previous section. RF is an ensemble learning method that consists of decision trees combined by bagging (bootstrap aggregation) [31]. MARS is a multivariate, nonlinear, nonparametric regression method combining recursive partitioning and piecewise polynomial functions [32]. Lasso shrinks predictor variables with weaker contributions to zero to control the trade-off between the bias and the variance in model fitting while reducing the likelihood of overfitting [33].
XGBoost is an ensemble learning method based on gradient boosting [34]. CatBoost is an improved decision tree algorithm that combines ordered boosting, gradient boosting, and classification features [35]. LightGBM is a histogram-based distributed gradient boosting framework algorithm that restricts the maximum depth of the decision trees [36]. These ML methods have been used successfully in various healthcare and medical applications [26,27,28,29], and all of them have the ability to select features while providing importance scoring to the input features. To evaluate the performance of the ML models, the balanced accuracy (BA), sensitivity (SEN), specificity (SPE), and area under the receiver operating characteristic (ROC) curve (AUC) are used. Furthermore, as it is widely utilized in many clinical-related studies, logistic regression (LGR) is also considered as the benchmark in this study to ensure all six of the ML models have reasonable performance.
The procedure of the hybrid multiphase CKD prediction scheme is shown in Figure 2. As shown in the figure, after obtaining the data with the generated longitudinal variables, ML models are constructed using the data. Additionally, an oversampling technique is utilized to address the class imbalance issue in the data. With the built ML models, the relevant importance value of each variable can be extracted from each ML model. Because each model has different hyper-parameters required to be tuned, 10-fold nested cross-validation (10f-NCV) is utilized for hyper-parameter tuning. Under the structure of 10f-NCV, in one iteration, the data are randomly split into 10 folds, where 1 fold is used for testing and the rest of the 9 folds are used for testing. During training, the 9 folds of the data will be further split into 8 folds for training and 1 fold for validation. Training ends when all of the 9 folds of data are used for validation once and the optimized hyper-parameter sets are found; then, the testing fold is used for evaluation. The entire 10f-NCV process is finished when all folds are used for testing once (a total of 10 iterations).
After constructing valid ML models, each can generate relevant information for each variable according to its model rules, thus generating two variable importance values: the ratio-scale-based relative importance value (RIV) and the ordinal-scale-based ordinal ranking value (ORV). In the RIV, the values of the most and least important variables are 100 and 0, respectively. The RIV is ordered from highest to lowest, and the given ranking value is the ORV. The most important variable, whose RIV is the highest or equal to 100, is placed at the top; the least important variable, whose RIV is the lowest or equal to 0, is placed at the bottom. Values can be repeated, which means that the variable importance of two or more variables may be similar. As each optimal ML method was repeated 10 times, there will be 10 corresponding variable importance values that are distinctive to each method. Each method’s mean importance was calculated to yield its single merged RIV and ORV.
Because a single selection variable algorithm has the propensity to choose a locally optimal solution, the ensemble variable has more opportunities to better approximate the optimal solution by averaging different assumptions [37]. To derive more stable results, different VER approaches are considered. VER approaches can provide more robust variable selection results than a single variable selection method and reduce bias and variance. It has shown excellent results across various research domains [38,39]. Hence, MA, AMA, GMA, BCA, and RMA are used in this study as they are widely utilized in many studies [37,40]. Moreover, because different VER approaches are only applicable to a specific variable measurement scale, the RIV variable integration is based on MA, AMA, and GMA, whereas ORV is based on AMA, BCA, and RMA. The equations of each VER approach used are as follows:
A M A F i = 1 j k = 1 j r i k = 1 j ( r i 1 + r i 2 + + r i j )
G M A F i = ( k = 1 j r i k ) 1 j = r i 1 r i 2 r i j j
M A F i = M a x ( r i 1 , r i 2 , , r i j )
R M A F i = M e d i a n ( r i 1 , r i 2 , , r i j )
B C A F i = Mode ( C o u n t ( r i 1 , r i 2 , , r i j ) )
where r i j is the RIV or ORV of the i th variable in the j th method. After aggregation via VER approaches, six sets of variable importance rankings are generated, namely RIV-AMA, RIV-GMA, RIV-MA, ORV-AMA, ORV-RMA, and ORV-BCA. Finally, union operation is used to integrate and compare the six importance ranking sets and to identify the most important risk factors for discussion.
This study used the R programming language (version 4.0.5; http://www.R-project.org (accessed on 27 February 2024)) and RStudio software (version 1.1.453; https://www.rstudio.com/products/rstudio/ (accessed on 27 February 2024)) to construct an effective ML model. All of the algorithm equations and estimated optimal hyperparameters of the models were built using R-related software packages. The package information is as following: The RF, LGR, MARS, Lasso, XGBoost, CatBoost, and LightGBM models were created using the randomForest (version 4.7-1.1) [41], stats (version 4.0.5), earth (version 5.3.1) [42], glmnet (version 4.1-7) [43], xgboost (version 1.6.0.1) [44], catboost (version 0.25.1) [45], and lightgbm (version 3.3.2) [46] packages, respectively. Lastly, the optimal hyperparameters were estimated for all models using the caret package (version 6.0-93) [47].

3. Results

Table 2 shows the average performance of each ML model after 10f-NCV. As shown in the table, Lasso had the best BA of 0.813, LightGBM had the best SEN of 0.791, and RF had the best SPE of 0.898. Lasso had the best AUC of 0.800, and all six ML models had greater scores of AUC than the benchmark LGR model (AUC 0.669). This can also be found in the ROC curve presented in Figure 3. Overall, according to the results in Table 2 and Figure 3, the usage of all six ML models for ensemble to identify important variables is reasonable.
The variable importance value of each variable generated from each ML model in terms of RIV (Table 3) and ORV (Table 4) can be found in the tables. In Table 3, the first 12 variables are presented. As different ML models analyze the data with different approaches and mechanisms, it can be seen that each ML model yields a different RIV for each variable. For example, both Lasso and LightGBM yield the lowest RIV of zero to V 1 ( S ) , whereas the other four ML methods yield a relatively higher RIV. The same concept can be found in Table 4. For example, V 3 ( M ) is ranked relatively lower by MARS (ORV 26) than the other five ML methods. Next, in order to derive more stable results and considerations when identifying important variables, the results of RIV and ORV are aggregated via the VER approaches, which are presented in Table 5.
Table 5 presents the first 12 variables’ aggregation results from RIVs and ORVs via different VER approaches. As the characteristic of RIV, the aggregated importance values are generated from the six ML models with the corresponding VER equations. The aggregated importance value ranges between 0 and 100, and more important variables will have higher values. This concept suits all of the aggregated RIVs (RIV-AMA, RIV-GMA, and RIV-MA). Both the aggregations of ORV-AMA and ORV-RMA have similar concepts as RIVs, but with slight differences due to the characteristics of ORV. As the most important variable will be assigned the rank of one under the structure of ORV, the more important variable will have a value closer to one after aggregation. Under ORV-BCA, when two or more ML models are assigned the same ranking to a variable, that specific ranking will be the aggregated value of the variable. Taking ORVs of V 1 ( C ) in Table 4 as an example, because both MARS and Lasso assigned the ranking of 70 to V 1 ( C ) , the aggregated value of V 1 ( C ) in ORV-BCA is 70, and this can also be seen in Table 5. Furthermore, if all six ML models have assigned separate rankings to a variable, the worst-case scenario will be taken into consideration, in which the aggregated value will be the worst-ranking value. For example, V 3 ( M ) in Table 4 can be found with separated rankings assigned from each ML model. Because the worst ranking of V 3 ( M ) is 26, the ORV-BCA of V 3 ( M ) is 26. Overall, as shown in Table 5, with VER approaches, different information regarding the variables analyzed by each ML method can be brought into consideration. To better compare and interpret the rankings of each variable via VER approaches, the results in Table 5 can be further organized into Table 6.
Table 6 presents the top 12 ranking variables of RIV and ORV based on the corresponding VER approaches. As shown in the table, BUN(S) is the most important variable across all six aggregation results utilizing VER approaches; BUN(M) is the second most important variable, which only ranked the third in RIV-MA. Overall, the top three ranking important variables are similar and begin to vary in lower rankings.
To examine the association between important variables using different VER approaches, union operation is performed on Table 6, and the results are organized into Table 7. As presented in the table, important variables identified after union operation in different ranking combinations can be seen. Five conditions of the combination for union operation are used, which are within the first 4 rankings, within the first 6 rankings, within the first 8 rankings, within the first 10 rankings, and within the first 12 rankings. Taking the first condition (within the first four rankings) for the RIV rule as an example, the first four ranking variables from each aggregation rule in Table 6 have to be identified first, then, with union operation, BUN(S), BUN(M), Hb(S), RBC(S), and r-GT(M) can be found, thus satisfying the condition in Table 7. The same process is performed for all of the other conditions for both rules.
To evaluate the stability of the union operation results of the selected variable sets under the two proposed aggregation rules, Lasso is constructed based on variables selected in each variable combination condition of RIV and ORV from Table 7, as the preliminary ML model performance results revealed that Lasso is the best one in this study. The performance of Lasso with different variable combination condition sets is shown in Table 8. According to the table, all AUCs of the union operations of the ranked variables under the two rules were greater than 0.804. Notably, variables within the first eight ranking conditions of RIV yield the best AUC of 0.883. Lasso uses 13 variables in total under the best variable combination condition; on the other hand, Lasso using all 76 variables has lower performance, with an AUC of 0.800. Therefore, the results are improved after variable selection, which greatly improves the overall prediction performance.
Figure 4 shows the AUC values of Lasso using different variable combination conditions with RIV and ORV. As shown in the figure, both RIV and ORV have increasing AUC values from the condition within the first four rankings to within the first eight rankings, and then both of their AUCs begin to decrease. ORV has better performance in AUC than RIV when the conditions are within the first four rankings and within the first six rankings; after that, RIV is superior to ORV in AUC when the amount of variables increases. In summary, the stability of the union operation is confirmed, and it indicates that the proposed risk factor evaluation scheme of this study can provide promising results.

4. Discussion

While clinical health data, in general, can accumulate into a substantial amount of big data, in practice, whether addressing preventive healthcare for chronic diseases, including metabolic syndrome (MetS), or assisting in the assessment of chronic kidney disease (CKD) at stages 3a-3b for high-risk diagnosis, the establishment of a risk prediction model for clinical use requires more appropriate or rational limited longitudinal healthcare data for research analysis. Collecting such data is essential to build predictive models, thereby identifying relevant risk factors more accurately, facilitating specialized physicians in clinical diagnosis, and aiding in medical decision making.
The analysis of clinical data often faces the challenge of dealing with limited sample sizes and complex variable interactions. Employing methods that consider multiple aspects can help compensate for these limitations. As the information can vary from different ML methods, consideration of VERs and union operation to aggregate them could enhance the overall predictive capability. Moreover, aggregated information, such as important features, can support healthcare or actual clinical scenarios. Furthermore, the analysis of health examination data requires consideration of the trends in the continuous change of data. Analyzing longitudinal data that meet the conditions is of greater research value and contributes to the subsequent predictive benefits of this study. Predicting CKD progression risk is a vital task in clinical management.
CKD is a progressive kidney disease characterized by deteriorating renal function. ML methods have been successfully used to predict CKD risk. This study used ML methods and feature engineering to construct a longitudinal variable set scheme to identify patients with MetS diagnosed with stages 3a or 3b CKD. Its results are highly significant for patients. For example, early CKD detection is conducive to providing effective interventions and measures to high-risk patients. Early treatment often leads to favorable treatment outcomes in the disease course. Next, in our longitudinal analysis, all six ML methods outperformed conventional LGR, and the Lasso model was the best, as reflected by its AUC of 0.800, which was 2.8% and 13.1% higher than that of conventional LGR. Our findings corroborate those of previous studies. Indeed, ML methods, especially Lasso [48], can be used to resolve the classification bias in several categories while showing strong prediction performance with imbalanced data [18].
In addition, this study used variable ensemble and union operation to integrate the ML-selected variable results and analyzed the Lasso-selected variables. Its experimental results demonstrated that the variable ensemble methods and union operation all effectively improved the Lasso model’s prediction performance. We offer reliable estimations of CKD risk factors based on different VERs. Specifically, we reduced the number of cross-sectional variables from 76 to 13 through variable selection, enhancing prediction performance. As reflected by the AUC, prediction performance increased from 5.6% to 8.3% after variable selection with Lasso. Like previous studies, the proposed hybrid scheme outperformed standalone schemes, as variable selection identified the important CKD variables and increased the model’s prediction performance [49,50,51]. The results with the selected variables were similar to those of existing clinical studies. For example, Lasso identified BUN, Hb, RBC, r-GT, HDL, UA, SBP, and FPG as key risk factors for CKD based on the cross-sectional results. The associations between these risk factors and CKD are elaborated based on previous studies.
Past related publications have not emphasized the analysis of longitudinal data in small to medium-sized samples. For health or clinical data with two or more real instances, the consideration of statistics like closest (C), mean (M), standard deviation (S), and difference (D) among variables has been lacking. In this study, not only did we identify the significance of these variables (Hb, RBC, BUN, r-GT, HDL, UA, SBP, and FPG) again, but we also delved deeper to explore which statistical values among these continuous data variables hold more meaning. Hb or RBCs are important indicators of the blood’s oxygen-carrying capacity. An excessively low value can lead to anemia, a common complication of CKD and a recognized risk factor for CKD deterioration [52,53]. A survey found that 41% of 209,311 patients with CKD had anemia [54]. Accordingly, study results found that in longitudinal data, the two variable statistical patterns, Hb(S), RBCs(M), and RBC(S), possess greater predictive capabilities.
BUN is an independent risk factor for CKD [27]. Increasing studies have examined the relationship between CKD and BUN. For example, BUN and CKD are positively correlated. Moreover, Seki et al. (2019) also reported that BUN may predict kidney disease development [55]. This study found that in longitudinal data, the important variable BUN may be more meaningfully assessed through the BUN(C), BUN(D), BUN(S), and BUN(M) values of each examination. Many studies have stressed the importance of UA in CKD. One of those studies identified the UA level as an important predictor of CKD [56]. A recent study observed that higher UA levels correlated significantly with CKD in middle-aged men regardless of their BMI [57]. Based on the findings of this study, it was discovered that in longitudinal data, the important variable UA may be more meaningfully assessed through the standard deviation “UA(S)” values of its multiple examinations.
r-GT is an enzyme found on the cell surface of all tissues. It is a typical indicator of alcohol consumption and hepatic impairment. Increasing studies have identified r-GT as an independent risk factor for CKD and ESRD [58]. For example, two Japanese studies concurred that increased r-GT correlates positively with CKD [59,60]. A study on male South Korean workers found that r-GT and CKD correlated positively, and r-GT may predict early CKD [61]. The research results indicate that exploring the mean value r-GT(M) of the variable r-GT(D) over longitudinal data may be more meaningful.
A recent 7472 person-years follow-up study in Korea showed that in patients with CKD, higher SBP and DBP levels were associated with a higher risk of a composite kidney outcome reflecting CKD progression. SBP had a greater association with adverse kidney outcomes than DBP [62]. Blood pressure control is undoubtedly an important risk factor for CKD, but for the CKD high-risk group during 3a to 3b, the variability of systolic blood pressure DBP(S) may be cause for concern. There exists strong evidence that HDL is associated with patients with impairment of kidney function and/or progression of CKD. HDL-C concentrations, the composition of HDL particles, disturbances in functionality, and especially the reverse cholesterol transport, might be different between various stages of kidney impairment, especially between patients with and without nephrotic syndrome [63]. Furthermore, a Mendelian randomization study showed HDL-C, LDL-C, and Triglycerides as CKD risk factors [64]. Therefore, the variance of HDL(S) may cause concern for CKD during 3a to 3b.
To summarize, blood sugar management is conducive to preventing diabetes, nephropathy, and other diabetic microvascular complications. Research has identified FPG as an important risk factor for CKD [52]. A retrospective study by Cao et al. (2022) [65] identified age, sex, BMI, T2DM, FPG, stroke, and hypertension as risk factors for CKD. Regarding the FPG values, the results of this study indicate that investigating the difference in “FPG(D)” from the previous test as a research variable may have a higher predictive value for risk assessment.

5. Limitations and Future Recommendations

While this study utilized innovative applications and a set of ensemble variable analyses for continuous data in small and medium-sized health data samples, clinical validation requires consideration of regional and ethnic differences. These variations may affect the construction of models and lead to differences in research outcomes across different regions and ethnic groups in subsequent studies. Due to the limitation of the research dataset, not every subject has completed, multi-year data, so only the most recent three data points from the longitudinal data are selected. Additionally, the proposed scheme is restricted to the ML methods’ mechanism having the ability to provide feature importance scoring. Some methods, such as neural network or k-nearest neighbor, may not be applicable to our scheme, as their algorithm design cannot provide feature importance scoring. Future research can consider more points of data analysis. Avoiding the biases in data collection or model assumptions would strengthen the validity of the findings. Exploring the potential for future research directions, such as validation of the predictive model in a clinical setting or investigating the impact of early intervention based on the identified risk factors, could add depth to the study’s implications. Regarding concerns about ethnic and regional differences and recommendations from clinical guidelines, follow-up research can apply the prediction model of this study to other healthcare settings or patient populations to improve scalability and generalizability; at the same time, based on the identified risk factors discovered in this study, an evaluation of the impact of early intervention could be performed.

6. Conclusions

The analysis of health screening data requires consideration of the trends in the continuous change of data. Analyzing longitudinal data that meet the conditions is of greater research value and contributes to the subsequent predictive benefits of this study. The hybrid multiphase scheme for predicting CKD in patients with MetS developed in this study through ML methods and feature engineering showed strong prediction performance. The limited longitudinal health screening data based on different feature ensembles demonstrated that the hybrid multiphase scheme effectively improved ML predictive performance. This study also examined common risk factors affecting CKD in patients with MetS using different models and ranked their importance for future reference. These rankings not only facilitate kidney condition assessment based on the risk factors but also the detection of other underlying diseases that patients with CKD might have. Moreover, our results are generalizable to a certain extent and may be used to enhance the understanding and treatment of other diseases by using the same ML methods and similar hybrid schemes. For healthcare professionals, information on how to incorporate the findings of this study into their clinical practice or decision-making processes would be beneficial. Based on the analysis results of this case study, for preventing CKD in the sub-healthy population with MetS, it is predictable that well-known factors like BUN, UA, and PFG play crucial roles. However, lesser-discussed factors, such as Hb, RBCs, or r-GT, should receive more attention in clinical practice or decision-making processes, which could be beneficial. Additionally, observing the mean, standard deviation, and difference of different variables across three consecutive data points carries distinct implications. The study’s findings may inform personalized medicine and targeted interventions for potential patients with high-risk CKD, thereby identifying clinical implications of risk factors and real-world healthcare applications. Validating the prediction model in a clinical setting and externally aligns with existing guidelines, potentially enhancing current CKD management practices.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics14080825/s1, Table S1: The demographics of the 19 variables from the subjects’ latest examination ( V i , C ).

Author Contributions

M.-S.C. and T.-C.L. made contributions equivalent to the first author. Conception and design, project administration, and funding acquisition, M.-S.C. and C.-J.L.; data collection, M.-S.C.; methods, T.-C.L., M.-J.J., C.-T.Y. and C.-J.L.; analysis and interpretation, T.-C.L., M.-S.C. and C.-J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Science and Technology Council, Taiwan (NSTC 111-2221-E-030-009), Fu Jen Catholic University (A0112181).

Institutional Review Board Statement

This study was approved by the Institutional Review Board of the Far Eastern Memorial Hospital (approval number: 110027-E; approval date: 3 March 2021) and the Mei Jhao Health Research Foundation (authorization code: MJHRF2022002A) and was registered at ClinicalTrials.gov (ID:NCT05225454, https://beta.clinicaltrials.gov/study/NCT05225454 (accessed on 27 February 2024)).

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analyzed during the current study are not publicly available due to ethical restrictions. The data acquisition process requires approval from the Institutional Review Board (IRB) and authorization from the MJ Health Research Foundation (MJHRF). For more details regarding the data application procedures, please refer to https://www.mjhrf.org/main/page/release1/en/#release01 (accessed on 27 February 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Alzain, M.A.; Asweto, C.O.; Hassan, S.U.; Saeed, M.E.; Kassar, A.; Alsaif, B. Psychometric Properties of Suboptimal Health Status Instruments: A Systematic Review. J. Pers. Med. 2023, 13, 299. [Google Scholar] [CrossRef] [PubMed]
  2. Gurka, M.J.; Ice, C.L.; Sun, S.S.; Deboer, M.D. A confirmatory factor analysis of the metabolic syndrome in adolescents: An examination of sex and racial/ethnic differences. Cardiovasc. Diabetol. 2012, 11, 128. [Google Scholar] [CrossRef] [PubMed]
  3. Lin, C.M. An Application of Metabolic Syndrome Severity Scores in the Lifestyle Risk Assessment of Taiwanese Adults. Int. J. Environ. Res. Public Health 2020, 17, 3348. [Google Scholar] [CrossRef] [PubMed]
  4. Hao, Y.; Zhu, Y.J.; Zou, S.; Zhou, P.; Hu, Y.W.; Zhao, Q.X.; Gu, L.N.; Zhang, H.Z.; Wang, Z.; Li, J. Metabolic Syndrome and Psoriasis: Mechanisms and Future Directions. Front. Immunol. 2021, 12, 711060. [Google Scholar] [CrossRef] [PubMed]
  5. Singh, A.K.; Kari, J.A. Metabolic syndrome and chronic kidney disease. Curr. Opin. Nephrol. Hypertens. 2013, 22, 198–203. [Google Scholar] [CrossRef] [PubMed]
  6. van Rooy, M.J.; Pretorius, E. Metabolic syndrome, platelet activation and the development of transient ischemic attack or thromboembolic stroke. Thromb. Res. 2015, 135, 434–442. [Google Scholar] [CrossRef] [PubMed]
  7. Ford, E.S.; Li, C.; Sattar, N. Metabolic syndrome and incident diabetes: Current state of the evidence. Diabetes Care 2008, 31, 1898–1904. [Google Scholar] [CrossRef]
  8. Kidney Disease Improving Global Outcomes. KDIGO 2012 Clinical Practice Guideline for the Evaluation and Management of Chronic Kidney Disease. Kidney Int. 2013, 3, 5–14. [Google Scholar]
  9. Jager, K.J.; Kovesdy, C.; Langham, R.; Rosenberg, M.; Jha, V.; Zoccali, C. A single number for advocacy and communication-worldwide more than 850 million individuals have kidney diseases. Nephrol. Dial. Transplant. 2019, 34, 1803–1805. [Google Scholar] [CrossRef]
  10. Lv, J.C.; Zhang, L.X. Prevalence and Disease Burden of Chronic Kidney Disease. Adv. Exp. Med. Biol. 2019, 1165, 3–15. [Google Scholar] [CrossRef]
  11. Perazella, M.A.; Khan, S. Increased mortality in chronic kidney disease: A call to action. Am. J. Med. Sci. 2006, 331, 150–153. [Google Scholar] [CrossRef] [PubMed]
  12. DeBoer, M.D.; Filipp, S.L.; Musani, S.K.; Sims, M.; Okusa, M.D.; Gurka, M.J. Metabolic Syndrome Severity and Risk of CKD and Worsened GFR: The Jackson Heart Study. Kidney Blood Press. Res. 2018, 43, 555–567. [Google Scholar] [CrossRef] [PubMed]
  13. Prasad, G.V. Metabolic syndrome and chronic kidney disease: Current status and future directions. World J. Nephrol. 2014, 3, 210–219. [Google Scholar] [CrossRef] [PubMed]
  14. Choe, W.S.; Choi, E.K.; Han, K.D.; Lee, E.J.; Lee, S.R.; Cha, M.J.; Oh, S. Association of metabolic syndrome and chronic kidney disease with atrial fibrillation: A nationwide population-based study in Korea. Diabetes Res. Clin. Pract. 2019, 148, 14–22. [Google Scholar] [CrossRef] [PubMed]
  15. Tozawa, M.; Iseki, C.; Tokashiki, K.; Chinen, S.; Kohagura, K.; Kinjo, K.; Takishita, S.; Iseki, K. Metabolic syndrome and risk of developing chronic kidney disease in Japanese adults. Hypertens. Res. 2007, 30, 937–943. [Google Scholar] [CrossRef] [PubMed]
  16. Thomas, G.; Sehgal, A.R.; Kashyap, S.R.; Srinivas, T.R.; Kirwan, J.P.; Navaneethan, S.D. Metabolic syndrome and kidney disease: A systematic review and meta-analysis. Clin. J. Am. Soc. Nephrol. 2011, 6, 2364–2373. [Google Scholar] [CrossRef] [PubMed]
  17. Jhou, M.J.; Chen, M.S.; Lee, T.S.; Yang, C.T.; Chiu, Y.L.; Lu, C.J. A Hybrid Risk Factor Evaluation Scheme for Metabolic Syndrome and Stage 3 Chronic Kidney Disease Based on Multiple Machine Learning Techniques. Healthcare 2022, 10, 2496. [Google Scholar] [CrossRef] [PubMed]
  18. Saberi-Karimian, M.; Khorasanchi, Z.; Ghazizadeh, H.; Tayefi, M.; Saffar, S.; Ferns, G.A.; Ghayour-Mobarhan, M. Potential value and impact of data mining and machine learning in clinical diagnostics. Crit. Rev. Clin. Lab. Sci. 2021, 58, 275–296. [Google Scholar] [CrossRef] [PubMed]
  19. Peiffer-Smadja, N.; Rawson, T.M.; Ahmad, R.; Buchard, A.; Georgiou, P.; Lescure, F.X.; Birgand, G.; Holmes, A.H. Machine learning for clinical decision support in infectious diseases: A narrative review of current applications. Clin. Microbiol. Infect. 2020, 26, 584–595. [Google Scholar] [CrossRef]
  20. Liu, Y.; Chen, P.C.; Krause, J.; Peng, L. How to Read Articles That Use Machine Learning: Users’ Guides to the Medical Literature. JAMA 2019, 322, 1806–1816. [Google Scholar] [CrossRef]
  21. Bolón-Canedo, V.; Alonso-Betanzos, A. Ensembles for feature selection: A review and future trends. Inf. Fusion 2019, 52, 1–12. [Google Scholar] [CrossRef]
  22. Abutaleb, N. Why we should sub-divide CKD stage 3 into early (3a) and late (3b) components. Nephrol. Dial. Transplant. 2007, 22, 2728–2729. [Google Scholar] [CrossRef] [PubMed]
  23. Zahran, A.; Shoker, A. About CKD stage-3 subdivision proposal. Nephrol. Dial. Transplant. 2008, 23, 1765–1766. [Google Scholar] [CrossRef]
  24. Chang, H.J.; Lin, K.R.; Chang, J.L.; Lin, M.T. Risk Factors for Chronic Kidney Disease in Older Adults with Hyperlipidemia and/or Cardiovascular Diseases in Taipei City, Taiwan: A Community-Based Cross-Sectional Analysis. Int. J. Environ. Res. Public Health 2020, 17, 8763. [Google Scholar] [CrossRef] [PubMed]
  25. Jeong, B.; Cho, H.; Kim, J.; Kwon, S.K.; Hong, S.; Lee, C.; Kim, T.; Park, M.S.; Hong, S.; Heo, T.Y. Comparison between Statistical Models and Machine Learning Methods on Classification for Highly Imbalanced Multiclass Kidney Data. Diagnostics 2020, 10, 415. [Google Scholar] [CrossRef] [PubMed]
  26. Qin, J.; Chen, L.; Liu, Y.; Liu, C.; Feng, C.; Chen, B. A Machine Learning Methodology for Diagnosing Chronic Kidney Disease. IEEE Access 2019, 8, 20991–21002. [Google Scholar] [CrossRef]
  27. Chiu, Y.L.; Jhou, M.J.; Lee, T.S.; Lu, C.J.; Chen, M.S. Health Data-Driven Machine Learning Algorithms Applied to Risk Indicators Assessment for Chronic Kidney Disease. Risk Manag. Healthc. Policy 2021, 14, 4401–4412. [Google Scholar] [CrossRef] [PubMed]
  28. Chang, C.C.; Yeh, J.H.; Chen, Y.M.; Jhou, M.J.; Lu, C.J. Clinical Predictors of Prolonged Hospital Stay in Patients with Myasthenia Gravis: A Study Using Machine Learning Algorithms. J. Clin. Med. 2021, 10, 4393. [Google Scholar] [CrossRef] [PubMed]
  29. Liao, P.C.; Chen, M.S.; Jhou, M.J.; Chen, T.C.; Yang, C.T.; Lu, C.J. Integrating Health Data-Driven Machine Learning Algorithms to Evaluate Risk Factors of Early Stage Hypertension at Different Levels of HDL and LDL Cholesterol. Diagnostics 2022, 12, 1965. [Google Scholar] [CrossRef]
  30. Health Promotion Administration Ministry of Health and Welfare Metabolic Syndrome Criteria. Available online: https://www.hpa.gov.tw/Pages/Detail.aspx?nodeid=639&pid=1219 (accessed on 3 March 2023).
  31. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  32. Friedman, J.H. Multivariate Adaptive Regression Splines. Ann. Stat. 1991, 19, 1–67. [Google Scholar] [CrossRef]
  33. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. 1996, 58, 267–288. Available online: https://www.jstor.org/stable/2346178 (accessed on 3 March 2023). [CrossRef]
  34. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2019; pp. 785–794. [Google Scholar]
  35. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar] [CrossRef]
  36. Ke, G.; Meng, Q.; Finley, T.W.; Wang, T.; Chen, W.; Ma, W.; Qiwei, Y.; Liu, T. LightGBM: A highly efficient gradient boosting decision tree. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 3147–3155. [Google Scholar]
  37. Pes, B. Ensemble feature selection for high-dimensional data: A stability analysis across multiple domains. Neural Comput. Appl. 2020, 32, 5951–5973. [Google Scholar] [CrossRef]
  38. Tuli, S.; Basumatary, N.; Gill, S.S.; Kahani, M.; Arya, R.C.; Wander, G.S.; Buyya, R. HealthFog: An ensemble deep learning based Smart Healthcare System for Automatic Diagnosis of Heart Diseases in integrated IoT and fog computing environments. Future Gener. Comput. Syst. 2020, 104, 187–200. [Google Scholar] [CrossRef]
  39. Moghimi, A.; Yang, C.; Marchetto, P.M. Ensemble Feature Selection for Plant Phenotyping: A Journey from Hyperspectral to Multispectral Imaging. IEEE Access 2018, 6, 56870–56884. [Google Scholar] [CrossRef]
  40. Wang, J.; Xu, J.; Zhao, C.; Peng, Y.; Wang, H. An ensemble feature selection method for high-dimensional data based on sort aggregation. Syst. Sci. Control Eng. 2019, 7, 32–39. [Google Scholar] [CrossRef]
  41. Breiman, L.; Cutler, A.; Liaw, A.; Wiener, M. randomForest: Breiman and Cutler’s Random Forests for Classification and Regression. R Package Version, 4.7-1.1. Available online: https://CRAN.R-project.org/package=randomForest (accessed on 3 March 2023).
  42. Milborrow, S. Derived from Mda: MARS by T. Hastie and R. Tibshirani. Earth: Multivariate Adaptive Regression Splines. R Package Version, 5.3.1. Available online: http://CRAN.R-project.org/package=earth (accessed on 3 March 2023).
  43. Friedman, J.; Hastie, T.; Tibshirani, R.; Narasimhan, B.; Tay, K.; Simon, N.; Qian, J.; Yang, J. Glmnet: Glmnet: Lasso and Elastic-Net Regularized Generalized Linear Models. 2023. R Package Version, 4.1-7. Available online: https://CRAN.R-project.org/package=glmnet (accessed on 3 March 2023).
  44. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K.; Mitchell, R.; Cano, I.; Zhou, T.; et al. Xgboost: Extreme Gradient Boosting. R Package Version, 1.6.0.1. Available online: https://CRAN.R-project.org/package=xgboost (accessed on 3 March 2023).
  45. Yandex Technologies. CatBoost: Unbiased Boosting with Categorical Features. R Package Version, 1.0.6. Available online: https://github.com/CatBoost/CatBoost/ (accessed on 3 March 2023).
  46. Microsoft. LightGBM: Light Gradient Boosting Machine. R Package Version, 3.3.2. Available online: https://github.com/microsoft/LightGBM (accessed on 3 March 2023).
  47. Kuhn, M. Caret: Classification and Regression Training. R Package Version, 6.0-93. Available online: https://CRAN.R-project.org/package=caret (accessed on 3 March 2023).
  48. Mansour, O.; Paik, J.M.; Wyss, R.; Mastrorilli, J.M.; Bessette, L.G.; Lu, Z.; Tsacogianis, T.; Lin, K.J. A Novel Chronic Kidney Disease Phenotyping Algorithm Using Combined Electronic Health Record and Claims Data. Clin. Epidemiol. 2023, 15, 299–307. [Google Scholar] [CrossRef] [PubMed]
  49. Imran Ali, S.; Ali, B.; Hussain, J.; Hussain, M.; Satti, F.A.; Park, G.H.; Lee, S. Cost-Sensitive Ensemble Feature Ranking and Automatic Threshold Selection for Chronic Kidney Disease Diagnosis. Appl. Sci. 2020, 10, 5663. [Google Scholar] [CrossRef]
  50. Ebiaredoh-Mienye, S.A.; Swart, T.G.; Esenogho, E.; Mienye, I.D. A Machine Learning Method with Filter-Based Feature Selection for Improved Prediction of Chronic Kidney Disease. Bioengineering 2022, 9, 350. [Google Scholar] [CrossRef]
  51. Ismail, W.N. Snake-Efficient Feature Selection-Based Framework for Precise Early Detection of Chronic Kidney Disease. Diagnostics 2023, 13, 2501. [Google Scholar] [CrossRef] [PubMed]
  52. Shih, C.C.; Lu, C.J.; Chen, G.D.; Chang, C.C. Risk Prediction for Early Chronic Kidney Disease: Results from an Adult Health Examination Program of 19,270 Individuals. Int. J. Environ. Res. Public Health 2020, 17, 4973. [Google Scholar] [CrossRef] [PubMed]
  53. Chang, Y.P.; Liao, C.M.; Wang, L.H.; Hu, H.H.; Lin, C.M. Static and Dynamic Prediction of Chronic Renal Disease Progression Using Longitudinal Clinical Data from Taiwan’s National Prevention Programs. J. Clin. Med. 2021, 10, 3085. [Google Scholar] [CrossRef] [PubMed]
  54. Inker, L.A.; Grams, M.E.; Levey, A.S.; Coresh, J.; Cirillo, M.; Collins, J.F.; Gansevoort, R.T.; Gutierrez, O.M.; Hamano, T.; Heine, G.H.; et al. Relationship of estimated GFR and albuminuria to concurrent laboratory abnormalities: An individual participant data meta-analysis in a global consortium. Am. J. Kidney Dis. 2019, 73, 206–217. [Google Scholar] [CrossRef] [PubMed]
  55. Seki, M.; Nakayama, M.; Sakoh, T.; Yoshitomi, R.; Fukui, A.; Katafuchi, E.; Tsuda, S.; Nakano, T.; Tsuruya, K.; Kitazono, T. Blood urea nitrogen is independently associated with renal outcomes in Japanese patients with stage 3–5 chronic kidney disease: A prospective observational study. BMC Nephrol. 2019, 20, 115. [Google Scholar] [CrossRef] [PubMed]
  56. Chou, Y.C.; Kuan, J.C.; Yang, T.; Chou, W.Y.; Hsieh, P.C.; Bai, C.H.; You, S.L.; Chen, C.H.; Wei, C.Y.; Sun, C.A. Elevated uric acid level as a significant predictor of chronic kidney disease: A cohort study with repeated measurements. J. Nephrol. 2015, 28, 457–462. [Google Scholar] [CrossRef] [PubMed]
  57. Kuma, A.; Mafune, K.; Uchino, B.; Ochiai, Y.; Enta, K.; Kato, A. Development of chronic kidney disease influenced by serum urate and body mass index based on young-to-middle-aged Japanese men: A propensity score-matched cohort study. BMJ Open 2022, 12, e049540. [Google Scholar] [CrossRef] [PubMed]
  58. Caravaca-Fontán, F.; Azevedo, L.; Bayo, M.Á.; Gonzales-Candia, B.; Luna, E.; Caravaca, F. High levels of both serum gamma-glutamyl transferase and alkaline phosphatase are independent preictors of mortality in patients with stage 4–5 chronic kidney disease. Niveles séricos elevados de gamma-glutamil transferasa y fosfatasa alcalina son predictores independientes de mortalidad en la enfermedad renal crónica estadio 4–5. Nefrologia 2017, 37, 267–275. [Google Scholar] [CrossRef]
  59. Ishigami, T.; Yamamoto, R.; Nagasawa, Y.; Isaka, Y.; Rakugi, H.; Iseki, K.; Yamagata, K.; Tsuruya, K.; Yoshida, H.; Fujimoto, S.; et al. An association between serum γ-glutamyltransferase and proteinuria in drinkers and non-drinkers: A Japanese nationwide cross-sectional survey. Clin. Exp. Nephrol. 2014, 18, 899–910. [Google Scholar] [CrossRef]
  60. Noborisaka, Y.; Ishizaki, M.; Yamazaki, M.; Honda, R.; Yamada, Y. Elevated Serum Gamma-Glutamyltransferase (GGT) Activity and the Development of Chronic Kidney Disease (CKD) in Cigarette Smokers. Nephro-Urol. Mon. 2013, 5, 967–973. [Google Scholar] [CrossRef]
  61. Ryu, S.; Chang, Y.; Kim, D.I.; Kim, W.S.; Suh, B.S. gamma-Glutamyltransferase as a predictor of chronic kidney disease in nonhypertensive and nondiabetic Korean men. Clin. Chem. 2007, 53, 71–77. [Google Scholar] [CrossRef] [PubMed]
  62. Lee, J.Y.; Park, J.T.; Joo, Y.S.; Lee, C.; Yun, H.R.; Yoo, T.H.; Kang, S.W.; Choi, K.H.; Ahn, C.; Oh, K.H.; et al. Association of blood pressure with the progression of CKD: Findings from KNOW-CKD study. Am. J. Kidney Dis. 2021, 78, 236–245. [Google Scholar] [CrossRef] [PubMed]
  63. Kronenberg, F. HDL in CKD—The devil is in the detail. J. Am. Soc. Nephrol. 2018, 29, 1356–1371. [Google Scholar] [CrossRef] [PubMed]
  64. Lanktree, M.B.; Thériault, S.; Walsh, M.; Paré, G. HDL cholesterol, LDL cholesterol, and triglycerides as risk factors for CKD: A Mendelian randomization study. Am. J. Kidney Dis. 2018, 71, 166–172. [Google Scholar] [CrossRef]
  65. Cao, X.; Yang, B.; Zhou, J. Scoring model to predict risk of chronic kidney disease in Chinese health screening examinees with type 2 diabetes. Int. Urol. Nephrol. 2022, 54, 1629–1639. [Google Scholar] [CrossRef]
Figure 1. The process of identifying the longitudinal subjects.
Figure 1. The process of identifying the longitudinal subjects.
Diagnostics 14 00825 g001
Figure 2. Proposed multiphase hybrid risk factor evaluation scheme.
Figure 2. Proposed multiphase hybrid risk factor evaluation scheme.
Diagnostics 14 00825 g002
Figure 3. ROC curves of each ML model.
Figure 3. ROC curves of each ML model.
Diagnostics 14 00825 g003
Figure 4. AUC of Lasso with different variable combination conditions.
Figure 4. AUC of Lasso with different variable combination conditions.
Diagnostics 14 00825 g004
Table 1. Definitions and descriptions of the predictor and target variables.
Table 1. Definitions and descriptions of the predictor and target variables.
VariableDescriptionUnit
V 1 , t Body Fat (BF)BF of subject at t th examination %
V 2 , t Body Mass Index (BMI)BMI of subject at t th examinationkg/m2
V 3 , t Blood Urea Nitrogen (BUN)BUN of subject at t th examinationmg/dL
V 4 , t Diastolic Blood Pressure (DBP)DBP of subject at t th examinationmmHg
V 5 , t Fasting Plasma Glucose (FPG)FPG of subject at t th examinationmg/dL
V 6 , t Hemoglobin (Hb)Hb of subject at t th examinationg/dL
V 7 , t Hip Circumference (HC)HC of subject at t th examinationcm
V 8 , t High-Density Lipoprotein Cholesterol (HDL)HDL of subject at t th examinationmg/dL
V 9 , t Intraocular Pressure (IOP)IOP of subject at t th examinationmmHg
V 10 , t Low-Density Lipoprotein Cholesterol (LDL)LDL of subject at t th examinationmg/dL
V 11 , t Mean Cell Volume (MCV)MCV of subject at t th examinationfl
V 12 , t Red Blood Cells (RBCs)RBCs of subject at t th examination106/μL
V 13 , t Gamma Glutamyl Transpeptidase (r-GT)r-GT of subject at t th examinationU/L
V 14 , t Systolic Blood Pressure (SBP)SBP of subject at t th examination mmHg
V 15 , t Serum Glutamic Oxaloacetic Transaminase (SGOT)SGOT of subject at t th examinationU/L
V 16 , t Serum Glutamic Pyruvic Transaminase (SGPT)SGPT of subject at t th examinationU/L
V 17 , t Triglyceride (TG)TG of subject at t th examinationmg/dL
V 18 , t Uric Acid (UA)UA of subject at t th examinationmg/dL
V 19 , t Waist Circumference (WC)WC of subject at t th examinationcm
Y Chronic Kidney Disease (CKD)CKD result of subject at the third examination
Table 2. Average performance of each ML model after 10f-NCV.
Table 2. Average performance of each ML model after 10f-NCV.
ModelBA (SD)SEN (SD)SPE (SD)AUC (SD)
LGR0.704 (0.09)0.656 (0.32)0.752 (0.19)0.669 (0.10)
RF0.798 (0.04)0.699 (0.12)0.898 (0.15)0.797 (0.04)
MARS0.766 (0.04)0.752 (0.09)0.780 (0.12)0.717 (0.07)
Lasso0.813 (0.05)0.769 (0.06)0.856 (0.12)0.800 (0.08)
XGBoost0.769 (0.05)0.741 (0.13)0.797 (0.18)0.763 (0.06)
CatBoost0.763 (0.06)0.777 (0.16)0.750 (0.14)0.708 (0.07)
LightGBM0.780 (0.10)0.791 (0.15)0.770 (0.17)0.781 (0.12)
Table 3. First 12 RIVs of each variable from the six used ML models.
Table 3. First 12 RIVs of each variable from the six used ML models.
VarsRFMARSLassoXGBoostCatBoostLightGBM
V 1 ( C ) 24.001.410.229.2210.433.78
V 1 ( M ) 23.3713.200.009.3014.070.43
V 1 ( S ) 9.361.120.002.8812.380.00
V 1 ( D ) 12.096.950.0010.3117.161.40
V 2 ( C ) 10.920.000.001.922.340.02
V 2 ( M ) 8.103.700.001.088.570.11
V 2 ( S ) 9.588.350.001.8116.790.12
V 2 ( D ) 15.970.000.008.5610.910.31
V 3 ( C ) 22.9724.220.003.0417.982.38
V 3 ( M ) 56.8048.1710.7340.2246.5011.82
V 3 ( S ) 99.03100.0053.95100.0094.22100.00
V 3 ( D ) 17.4431.610.004.856.830.66
Table 4. First 12 ORVs of each variable from the six used ML models.
Table 4. First 12 ORVs of each variable from the six used ML models.
VarsRFMARSLassoXGBoostCatBoostLightGBM
V 1 ( C ) 187070253934
V 1 ( M ) 236276163442
V 1 ( S ) 517076464171
V 1 ( D ) 376476184138
V 2 ( C ) 437676506267
V 2 ( M ) 577076614866
V 2 ( S ) 446476532164
V 2 ( D ) 297676254060
V 3 ( C ) 103776483450
V 3 ( M ) 22643611
V 3 ( S ) 213121
V 3 ( D ) 213676324347
Table 5. First 12 aggregated RIVs and ORVs via VER approaches of each variable.
Table 5. First 12 aggregated RIVs and ORVs via VER approaches of each variable.
VarsRIV-AMARIV-GMARIV-MAORV-AMAORV-RMAORV-BCA
V 1 ( C ) 8.183.7324.0042.6736.5070
V 1 ( M ) 10.060.0023.3742.1738.0076
V 1 ( S ) 4.290.0012.3859.1760.5076
V 1 ( D ) 7.980.0017.1645.6739.5076
V 2 ( C ) 2.530.0010.9262.3364.5076
V 2 ( M ) 3.590.008.5763.0063.5076
V 2 ( S ) 6.110.0016.7953.6758.5076
V 2 ( D ) 5.960.0015.9751.0050.0076
V 3 ( C ) 11.760.0024.2242.5042.5076
V 3 ( M ) 35.7129.4256.808.675.0026
V 3 ( S ) 91.2089.19100.001.671.501
V 3 ( D ) 10.230.0031.6142.5039.5076
Table 6. Top 12 ranking variables of RIV and ORV with different VER approaches.
Table 6. Top 12 ranking variables of RIV and ORV with different VER approaches.
RIVORV
Rule/Rank RIV-AMARIV-GMARIV-MAORV-AMAORV-RMAORV-BCA
1BUN(S)BUN(S)BUN(S)BUN(S)BUN(S)BUN(S)
2BUN(M)BUN(M)Hb(S)BUN(M)BUN(M)BUN(M)
3Hb(S)Hb(S)BUN(M)Hb(S)Hb(S)LDL(M)
4RBC(S)RBC(S)r-GT(M)RBC(S)FPG(D)HDL(S)
5r-GT(M)RBC(M)RBC(S)FPG(D)RBC(S)TG(S)
6HDL(S)UA(S)HDL(S)HDL(S)HDL(S)HC(S)
7BUN(C)SBP(S)r-GT(D)RBC(M)RBC(M)UA(S)
8RBC(M)FPG(D)BUN(D)SBP(S)Hb(M)BMI(S)
9r-GT(D)BF(C)RBC(M)r-GT(M)SBP(M)DBP(D)
10LDL(D)SBP(C)WC(C)SBP(M)BF(C)Hb(M)
11FPG(D)Hb(M)LDL(D)Hb(M)UA(S)SGOT(D)
12BUN(D)DBP(S)RBC(C)BF(M)BF(M)BF(C)
Table 7. Important Variables identified after union operation in different ranking combinations.
Table 7. Important Variables identified after union operation in different ranking combinations.
RulesVariable Combination ConditionsSelected Important Variables after Union Operation
RIVWithin the first 4 rankings BUN(S), BUN(M), Hb(S), RBC(S), r-GT(M)
Within the first 6 rankingsBUN(S), BUN(M), Hb(S), RBC(S), r-GT(M), RBC(M), HDL(S), UA(S)
Within the first 8 rankingsBUN(S), BUN(M), Hb(S), RBC(S), r-GT(M), RBC(M), HDL(S), UA(S), BUN(C), SBP(S), r-GT(D), FPG(D), BUN(D)
Within the first 10 rankingsBUN(S), BUN(M), Hb(S), RBC(S), r-GT(M), RBC(M), HDL(S), UA(S), BUN(C), SBP(S), r-GT(D), FPG(D), BUN(D), BF(C), LDL(D), SBP(C), WC(C)
Within the first 12 rankingsBUN(S), BUN(M), Hb(S), RBC(S), r-GT(M), RBC(M), HDL(S), UA(S), BUN(C), SBP(S), r-GT(D), FPG(D), BUN(D), BF(C), LDL(D), SBP(C), WC(C), Hb(M), DBP(S), RBC(C)
ORVWithin the first 4 rankings BUN(S), BUN(M), Hb(S), LDL(M), RBC(S), FPG(D), HDL(S)
Within the first 6 rankingsBUN(S), BUN(M), Hb(S), LDL(M), RBC(S), FPG(D), HDL(S), TG(S), HC(S)
Within the first 8 rankingsBUN(S), BUN(M), Hb(S), LDL(M), RBC(S), FPG(D), HDL(S), TG(S), HC(S), RBC(M), UA(S), SBP(S), Hb(M), BMI(S)
Within the first 10 rankingsBUN(S), BUN(M), Hb(S), LDL(M), RBC(S), FPG(D), HDL(S), TG(S), HC(S), RBC(M), UA(S), SBP(S), Hb(M), BMI(S), r-GT(M), SBP(M), DBP(D), BF(C)
Within the first 12 rankingsBUN(S), BUN(M), Hb(S), LDL(M), RBC(S), FPG(D), HDL(S), TG(S), HC(S), RBC(M), UA(S), SBP(S), Hb(M), BMI(S), r-GT(M), SBP(M), DBP(D), BF(C), SGOT(D), BF(M)
Table 8. AUC of Lasso with different variable combinations according to the results in Table 7.
Table 8. AUC of Lasso with different variable combinations according to the results in Table 7.
RuleVariable Combination Conditions (Number of the Selected Variable)AUC
RIVWithin the first 4 rankings (5)0.804
Within the first 6 rankings (8)0.843
Within the first 8 rankings (13) *0.883
Within the first 10 rankings (17)0.855
Within the first 12 rankings (20)0.825
ORVWithin the first 4 rankings (7)0.835
Within the first 6 rankings (9)0.850
Within the first 8 rankings (14)0.870
Within the first 10 rankings (20)0.844
Within the first 12 rankings (19)0.821
* represents the best AUC value.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, M.-S.; Liu, T.-C.; Jhou, M.-J.; Yang, C.-T.; Lu, C.-J. Analyzing Longitudinal Health Screening Data with Feature Ensemble and Machine Learning Techniques: Investigating Diagnostic Risk Factors of Metabolic Syndrome for Chronic Kidney Disease Stages 3a to 3b. Diagnostics 2024, 14, 825. https://doi.org/10.3390/diagnostics14080825

AMA Style

Chen M-S, Liu T-C, Jhou M-J, Yang C-T, Lu C-J. Analyzing Longitudinal Health Screening Data with Feature Ensemble and Machine Learning Techniques: Investigating Diagnostic Risk Factors of Metabolic Syndrome for Chronic Kidney Disease Stages 3a to 3b. Diagnostics. 2024; 14(8):825. https://doi.org/10.3390/diagnostics14080825

Chicago/Turabian Style

Chen, Ming-Shu, Tzu-Chi Liu, Mao-Jhen Jhou, Chih-Te Yang, and Chi-Jie Lu. 2024. "Analyzing Longitudinal Health Screening Data with Feature Ensemble and Machine Learning Techniques: Investigating Diagnostic Risk Factors of Metabolic Syndrome for Chronic Kidney Disease Stages 3a to 3b" Diagnostics 14, no. 8: 825. https://doi.org/10.3390/diagnostics14080825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop