Next Article in Journal
Optical Coherence Tomography Angiography, Elastography, and Attenuation Imaging for Evaluation of Liver Regeneration
Previous Article in Journal
ICU ‘Magic Numbers’: The Role of Biomarkers in Supporting Clinical Decision-Making
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity

1
Department of Smart Computing and Applied Mathematics, Tunghai University, Taichung 407224, Taiwan
2
Cardiovascular Center, Taichung Veterans General Hospital, Taichung 407219, Taiwan
3
Department of Post-Baccalaureate Medicine, National Chung Hsing University, Taichung 402202, Taiwan
4
Department of Life Science, Tunghai University, Taichung 407224, Taiwan
5
Department of Industrial Engineering and Enterprise Information, Tunghai University, Taichung 407224, Taiwan
6
Informatics Department, Krida Wacana University, Jakarta 11470, Indonesia
7
Department of Computer Science, Tunghai University, Taichung 407224, Taiwan
8
Research Center for Smart Sustainable Circular Economy, Tunghai University, Taichung 407224, Taiwan
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(8), 976; https://doi.org/10.3390/diagnostics15080976
Submission received: 4 January 2025 / Revised: 22 March 2025 / Accepted: 31 March 2025 / Published: 11 April 2025
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Background and Objective: Using echocardiogram data for cardiovascular disease (CVD) can lead to difficulties due to imbalanced datasets, leading to biased predictions. Machine learning models can enhance prognosis accuracy, but their effectiveness is influenced by optimal feature selection and robust classification techniques. This study introduces an event-based self-similarity approach to enhance automatic feature selection approach for imbalanced echocardiogram data. Critical features correlated with disease progression were identified by leveraging self-similarity patterns. This study used an echocardiogram dataset, visual presentations of high-frequency sound wave signals, and data of patients with heart disease who are treated using three treatment methods: catheter ablation, ventricular defibrillator, and drug control—over the course of three years. Methods: The dataset was classified into nine categories and Recursive Feature Elimination (RFE) was applied to identify the most relevant features, reducing model complexity while maintaining diagnostic accuracy. Machine learning classification models, including XGBoost and CATBoost, were trained and evaluated. Results: Both models achieved comparable accuracy values, 84.3% and 88.4%, respectively, under different normalization techniques. To further optimize performance, the models were combined into a voting ensemble, improving feature selection and predictive accuracy. Four essential features—age, aorta (AO), left ventricular (LV), and left atrium (LA)—were identified as critical for prognosis and were found in Random Forest (RF)-voting ensemble classifier. The results underscore the importance of feature selection techniques in handling imbalanced datasets, improving classification robustness, and reducing bias in automated prognosis systems. Conclusions: Our findings highlight the potential of machine learning-driven echocardiogram analysis to enhance patient care by providing accurate, data-driven assessments.

1. Introduction

Ensuring a healthy life is essential to support the realization of the United Nations’ Sustainable Development Goals (SDGs). A person’s lifestyle affects how healthy they are. Unhealthy eating patterns, smoking habits, lack of movement, excessive stress, high blood sugar, high cholesterol, and weather patterns can contribute to cardiovascular diseases (CVD), which is ranked the second-highest cause of mortality [1,2]. Bad weather can even tighten blood vessels and raise the chances of getting heart disease. As age increases, there is a linear increase in the risk of death. Older people are at a higher risk of cardiovascular and cerebrovascular diseases and are part of an age group that experiences a constant decline in heart health. Thus, the early detection of CVD is challenging. Additionally, between 1970 and 2020, the global population of seniors above the age of 60 years slightly increased from 10% to 13% [3], and the figure is projected to rise to 23% in 2050.
Invasive methods or non-invasive methods can be employed to diagnose heart disease [4]. Electrocardiography, phonocardiography, and echocardiography are non-invasive procedures commonly used to diagnose CVD. Regarding cardiac speed assessment, echocardiography facilitates a quick cardiac evaluation comprising the structure and heart function. Echocardiography features highly reliable in-resolution imaging and real-time feedback, affordability, and user-friendliness [4,5]. Although the non-invasive imaging detection approach has advantages, it also has drawbacks as it is lengthy, requires multiple evaluations thus increasing the complexity and user subjectivity, and is influenced by the physician’s experience. Moreover, the nature of a non-invasive approach requires a broader examination which creates varying complexities and medical opinions, even when conducted under standardized conditions [6,7]. To address these constraints, cutting-edge technology and artificial intelligence can usher in new atmospheres and breakthroughs to facilitate accurate, effective, and swift decision-making in healthcare.
A great interest has been focused on developing tools for predicting and managing many diseases with meaningful impacts on human health. The subset of artificial intelligence, machine learning, deals with endometriosis (EM) [8], breast cancer [9], lung cancer [10], arrhythmia [11], COVID-19 [12], Parkinson’s disease [13], chronic disease [14], acute chronic kidney disease [15,16], brain tumour [17],and CVD [1,3,4,18,19,20,21,22,23,24,25,26,27]. Table 1 shows some critical studies on CVD that utilize the methodology of machine learning and most studies used electrocardiogram data and a single dataset. The table contains the author’s name, the name of the classifier, experiment results, the number of features, the number of samples, and the name of the dataset used. Random Forest (RF) is used for general-purpose tasks; apart from its ability to handle mixed data, the model is also easy to use. Due to its high accuracy rates, Sumwiza [28] revealed that RF is a promising tool. Before implementing any feature selection approach, the accuracy of their proposed model reached 96%. This demonstrates the capability of the RF to handle dataset effectively. Even after using the correlation coefficient feature selection technique, the performance of the proposed model improved by approximately 3%. On the other hand, Wang [1] used data from patients in the coronary care units (CCU) and showed the performance of XGBoost outperformed Naïve Bayes, Logistics Regression, and Support Vector Machine (SVM) models. The accuracy score of XGBoost reached 66.3%, indicating that the XGBoost is a machine learning algorithm that offers specific characteristics: flexibility, efficiency in processing missing data, and accurate predictions even if there are weak prediction models. In other research, the performance results of XGBoost were not optimal. Sachdeva [19] used the UCI dataset of 299 heart failure patients and applied various classifiers like SVM, RF, Decision Tree (DT), and XGBoost to predict the survival of patients with cardiovascular diseases. They concluded that SVM worked better than the other classifiers and had an accuracy of 96.67%. LightGBM model offers a fast and efficient model, is memory-friendly, and is good with large datasets. Yang [29] proposed the LightGBM model based on an optional hyperparameter optimization framework validated with 4240 samples to mitigate the costs associated with the medical treatment of patients suffering from CVD. The XGBoost was also compared with the LightGBM. However, the performance of LightGBM achieved 97.8% on the area under the curve (AUC), outperforming other comparative models. The improved LightGBM model showed superior sensitivity, specificity, and accuracy. Data imbalance significantly affects data classification because it impacts model training convergence and generalization capabilities on unseen test data [30,31]. Regarding the early prevention of CVD, Boudali [32] evaluated the DT, RF, LightGBM, CATBoost, and XGBoost on the imbalanced dataset. The issue of imbalanced data is tackled by random under-sampling technique and LightGBM declares its ability to handle a large imbalanced dataset and emerge as the top performer with an accuracy rate of 77%. Among boosting approaches, the CATBoost classifier is also known as one of the most effective models with a high accuracy rate and a capability for versatility and adaptability [33]. The performance of CATBoost outperforms LightGBM and XGBoost, achieving an impressive accuracy rate of up to 98%. The CATBoost’s advantages in dealing with classified variables and producing reliable results were utilized by Wei [34] in CVD risk prediction. Their proposed CATBoost model reached its accuracy at 86.30%, thus, the robustness of CATBoost is suitable to support physicians in potential risks due to reducing the misdiagnosis rate. Individual and voting techniques can also be used to solve CVD issues; most previous studies used a single predictive model. Sen [35] proposed a soft voting meta classifier composed of CATBoost, XGBoost, LightGBM, Gaussian Naive Bayes, and RF, which was then trained with 918 samples of CVD electrocardiogram and 11 features. The accuracy of the soft voting ensemble model achieved 91.85% and a 0.9344 AUC score, indicating strong predictive performance and highly effective in distinguishing between CVD-positive and CVD-negative cases. The work presented in [36,37] used hard voting ensemble approach with more than 10 parameters. The performance of their proposed model reached an accuracy of 90%.
The result of an echocardiography procedure is an ultrasound image, also known as an echocardiogram in the US, which is the visual representation of the high-frequency sound wave signals propagating through an object. An echocardiogram and clinical dataset contain critical features that influence the decision-making. Many irrelevant and redundant features in a dataset cause model performance to decrease because of over-fitting and dimensionality issues. The number of features in any dataset can be reduced by using different approaches [38]. The Recursive Feature Elimination (RFE) algorithm for feature selection is commonly used in machine learning [39]. By iteratively selecting the most critical features and eliminating the weakest ones, this approach can improve the model’s performance and remove irrelevant or redundant features that may cause overfitting [40]. The resulting feature’s importance ranking can help to identify the most informative feature and optimize the model accuracy while reducing the computational complexity and training time. This technique is beneficial for analyzing complex and high-dimensional datasets, such as medical images or signals, where selecting relevant features can be challenging due to the large number of potential variables [9,13].
The novel contribution of this research work is as follows:
  • Addresses the issue of imbalanced echocardiogram datasets by integrating the RFE feature selection technique to identify the potential features of CVD, reducing model complexity while maintaining accuracy.
  • Demonstrates the effectiveness of ensemble voting strategies in enhancing prediction robustness, combining classifiers to improve feature selection and diagnosis accuracy.
  • Proposes a robust classification pipeline incorporating XGBoost, RF, LightGBM, CATBoost, and ensemble learning to mitigate bias and improve predictive performance.
  • Provides a clinically relevant feature subset facilitating better decision-making in healthcare
The proposed study aims to assist clinical decision-making with an automated prognosis system with effective prediction models and automated feature selection in the context of imbalanced datasets. The paper is organized as follows: Section 1 presents a comprehensive review of cardiac ultrasound and machine learning. Materials and methods are presented in Section 2 as part of the background research. The Section 3 describes the results. The Section 4 explains the experiment processing. The research conclusions and future work are given in the Section 5.

2. Materials and Methods

This section describes the data collection, the classification methods incorporating the supervised learning method, and the feature selection method. Four supervised methods, XGBoost, RF, LightGBM, and CATBoost, were evaluated in the model-construction phase. The supervised technique entails the utilization of predetermined independent variables as features or risk factors. That dependent variable is the response outcome data for a model to predict novel instances once trained. It comprises independent input variables and their dependent variables outputs. The CVD early diagnosis approach uses ultrasound heart exam equipped with machine learning methods. Figure 1 shows the research block diagram, and the pseudocode of the proposed model is shown in Appendix A.

2.1. Data Source

Taichung Veterans General Cardiologists provided two sample datasets: sample A consists of 547 records, and sample B consists of 1212 records; both datasets are taken from the diagnosis result and treatment section and the patients’ historical visitations section over three years. During three years, a patient may have several records with information about his diagnosis, visitation, treatment, and the result of a series of cardiac ultrasound tracking. The simple matching method and event-based self-similarity approach were used to find corresponding patients based on clinical characteristics such as age, gender, primary diagnosis, and medical histories to ensure these patients had similar ultrasound report numbers and clinical backgrounds. Dataset sample B contains all the records over three years and is employed to classify the task. To understand the impact of dataset size on machine learning model performance, dataset sample A was created. Dataset sample A is a subset of dataset sample B. These are new records of patients who did a series of tracking cardiac ultrasounds because their heart’s physiological condition worsened compared to previously recorded data. There are 745 records with CVD and 561 records without CVD. A patient may have several records, whereas the average is 1–2 records per patient. Figure 2a shows the frequencies of distribution patients—from one to five—corresponding with the number of records in sample A, with details as follows: 360, 65, 12, 4, 1, respectively. Figure 2b shows the frequencies of distribution patients corresponding with the number of records in sample B, with details as follows: 462, 172, 77, 24, 11, 4, respectively.
Heart disease treatments are performed through cardiac catheter ablation, ventricular defibrillator, and drug control. Regarding the diagnosis result categories, the dataset collection is organized into several statuses: died, one-time or more than one-time follow-up in cardiology and surgery, and no follow-up. Table 2 shows the grouping—Group 11, Group 13, Group 14, Group 21, Group 23, Group 24, Group 31, Group 33, and Group 34—based on treatments and diagnosis cohort.
Group 11 corresponds to patients who received cardiac catheter ablation treatment and were recorded as deceased. Group 13 corresponds to patients who received cardiac catheter ablation treatment and had more than one-time follow-up visits after treatment. Group 14 corresponds to patients who received cardiac catheter ablation treatment but did not return for continued treatment.
Group 21 corresponds to patients who received ventricular defibrillator treatment and were recorded as deceased. Group 23 corresponds to patients who received ventricular defibrillator treatment and continued with more than one-time follow-up visits after treatment. Group 24 corresponds to patients who received ventricular defibrillator treatment but did not return for continued treatment.
Group 31 corresponds to patients who underwent drug control treatment and were recorded as deceased. Group 33 corresponds to patients who undertook drug control treatment and proceeded with more than one-time follow-up visits after treatment. Group 34 corresponds to patients who underwent drug control treatment but did not return for further treatment.
Table 3 and Table 4 show the distribution of patients across diagnosis and treatment categories, which was imbalanced and highlights that the most significant number of patients belong to Group 23, which received ventricular defibrillator treatment and ongoing follow-up visits for diagnosis. The term imbalanced datasets describes when most classes outnumber the minority classes [41]. From sample A, 40.21% of the records were for patients who had a ventricular defibrillator treatment and were re-visited to follow up on their treatment (220 records). From sample B, 41.09% of the records were for patients who had a ventricular defibrillator treatment and received ongoing follow-up visits for diagnosis (498 records).
Table 5 shows the descriptive statistics result of Group 23 in sample A. It shows the age range of 26 years old to 96 years old in which the mean is 77 years old; the mean of Left Ventricular (LV) is 46.8 mm; the mean of Ventricular Septum (VS) is 13 mm; the mean of Left Ventricle Posterior Wall (LVPW) is 12.6 mm; the mean of left atrium (LA) is 46.5 mm; the mean of aorta (AO) is 36.2 mm; the mean of TR mean pressure gradient (TR_PG_Mean) is 24.2 mmHG; and the mean of Left Ventricular Ejection Fraction (LVEF) is 56.
A preview of the data distribution on continuous attributes is visualized through a histogram as shown in Figure 3. It can be noted that some continuous numerical features, such as age, LV, VS, AO, LVPW, and LA, show a bell-shaped distribution, indicating that it is beneficial for statistical modeling and machine learning algorithms.
To assess the model’s generalizability and robustness, the dataset was divided into training and testing sets, with proportionate allocations of 70% and 30%, respectively. The distribution of training and testing data composition is applied to samples A and B.

2.2. Random Forest (RF)

This classifier builds an ensemble of M decision trees and combines their outputs for the final prediction [42,43], represented as:
m M , n x ; Q 1 , , Q m , n = 1 M i = 1 m m n x ; Q i , n
where M can be any size but is limited to computing resources. Considering the i-th tree in a cluster of trees, the predicted value at every query point x is denoted by m n x ; Q i , n , where Q 1 , , Q m is the independent random variables and n is the training time.
Each tree in the forest is built from a random subset of the training data and a random subset of features to enhance model robustness and avoid overfitting. The RF classifier has the hyperparameter as follows: the number of estimators, the maximum depth for each tree, the number of minimal samples that is needed to split the nodes, the minimal number of samples of leaf.

2.3. XGBoost

XGBoost is a constructed tree ensemble algorithm under the Gradient Boosting Decision Tree (GBDT) framework. It combines the advantages of bagging and boosting, namely training loss and regularization, to prevent overfitting of the model, and it can be used in the regression. For example, Gradient Boosting refits the errors of all previous weak learners by adding a new weak learner iteratively, in which the final accuracy will be higher than what would be achieved using a single learner. XGBoost retains this feature so that each tree is related and keeps the original model unchanged in each operation. It then adds a new function to the model so that the tree generated later can correct the errors of the previous tree. XGBoost uses random feature extraction when generating trees; hence, not all features will be used in every decision-making process during tree generation. Each decision tree calculates the feature and threshold with the best branch effect and completes the split construction. The hyperparameter of XGBoost is as follows: the number of boosting rounds, known as the estimator number, the number of learning rates, the maximum depth of the tree, and the minimum weight of the child node.
In Gradient Boosting [44], each new model f m ( x ) is added to correct the errors of the previous models. The model is built iteratively by minimizing the loss function L ( y , y ^ ) over all training instances. The prediction at stage m is given by [43] :
y ^ ( m ) = y ^ ( m 1 ) + η f m ( x )
where: y ^ ( m ) is the prediction at stage m, η is learning rate, f m ( x ) is the m-th weak learner (a decision tree in this case)
The objective function for XGBoost includes a regularized loss function to balance model complexity and predictive accuracy [44]:
L = i = 1 n L ( y i , y ^ i ) + k = 1 K Ω ( f k )
where L ( y i , y ^ i ) is training loss function such as mean squared error for regression, Ω ( f k ) is the regularization term for tree complexity, with T as the number of leaves, γ controls the number of leaves, and λ controls the leaf weights. Total number of boosting rounds or estimators M. Each round introduces a new tree to improve the model. Learning rate η controls the contribution of each tree. A smaller learning rate requires more rounds. Maximum depth d max of each tree, controlling the model’s complexity and overfitting. Minimum sum of instance weights required in a child node w min , helping prevent overfitting on small samples.

2.4. LightGBM

LightGBM is a Gradient Boosting Decision Tree (GBDT) model optimized for high efficiency and low memory usage. It iteratively adds decision trees to the ensemble [45] and uses a histogram-based algorithm which reduces memory usage and improves speed by binning continuous features into discrete intervals [32], as demonstrated in Equation (4).
bin ( x i , j ) = x i , j min ( X j ) max ( X j ) min ( X j ) × B
Unlike Level-wise, which splits nodes at the same level in each step, LightGBM grows trees Leaf-wise. This approach allows each split to focus on the leaf with the highest error reduction. While it can improve accuracy, it may lead to overfitting, especially on small datasets. LightGBM introduces a maximum depth for leaves, d max , which limits how deep the tree can grow to prevent overfitting. This parameter provides a balance between model complexity and overfitting. LightGBM achieves high training speed, up to 10 times faster than traditional GBDT.

2.5. CATBoost

CATBoost is a machine learning algorithm developed by Yandex that optimizes gradient boosting specifically for datasets with categorical features [33,34]. It implements ordered boosting and dynamic partitioning to handle categorical variables more effectively, reducing overfitting and enhancing accuracy. Traditional gradient boosting often introduces overfitting, especially in datasets with categorical features. CATBoost mitigates this with ordered boosting, which ensures that at each boosting iteration, only the past data is used to calculate predictions, preventing target leakage and overfitting. CATBoost applies dynamic encoding to categorical features using feature combinations. It encodes categorical values by averaging target values for subsets, effectively managing categorical features in high-dimensional spaces. The prediction equation [17] is shown in Equation (5):
y ^ ( m ) = y ^ ( m 1 ) + η f m ( x )
where η is the learning rate, f m ( x ) is the decision tree added at iteration m. The number of iterations m is a hyperparameter controlling how many trees are added to the model, which affects accuracy. Higher values of iterations may improve the accuracy model, but increase the computational cost. Learning Rate η controls the contribution of each tree in the boosting process.

2.6. Normalization Techniques

Raw dataset is the initial, unprocessed data directly obtained from the source. Raw datasets often contain features with varying scales, which can lead to biased model performance if not appropriately scaled, especially for algorithms sensitive to feature magnitude, such as the k-nearest neighbors, SVM. The Z-score normalization transforms the features by removing the mean and scaling to unit variance, resulting in a distribution with a mean of 0 and a standard deviation of 1 for each feature. This is useful when features in the dataset are normally distributed and when we want to center data around zero which can improve model convergence during training. The transformation formula [37,46] is:
X scaled = X μ σ
where X is the feature value, μ is the mean, and σ is the standard deviation. The absolute maximum normalization approach scales each feature by its maximum absolute value, transforming the data to lie within the range [−1, 1]. This scaler preserves sparsity, making it suitable for sparse datasets, and is often used when feature values have no fixed bounds but differ in magnitude. Each of these scaling methods serves distinct purposes, depending on the dataset’s characteristics and the requirements of the chosen machine learning model.

2.7. Recursive Feature Elimination (RFE)

RFE is a wrapper-based feature selection method that recursively removes features and builds a model on the remaining features [47]. Then it ranks the features by importance and eliminates the least important ones. The resulting feature’s importance ranking can help to identify the most informative features and optimize the model accuracy while reducing the computational complexity and training time. The process is repeated until a predetermined number of features are reached. Based on this circumstance, the most critical features with the highest impact on the target variable are identified. The most critical features can be used for better model performance and efficiency. This approach can improve the model’s performance and remove irrelevant or redundant features that may cause overfitting. This technique is beneficial for analyzing complex and high-dimensional datasets such as medical images or signals, where selecting relevant features can be challenging due to the large number of potential variables.

3. Results

Table 6 shows the performance of various supervised learning models (XGBoost, RF, LightGBM, and CATBoost) in two distinct samples, sample A and sample B, each evaluated under three scaling conditions: raw dataset, Z-score normalization, and absolute maximum normalization in both cases. A significant accuracy gap is evident between sample A and sample B across all models, with sample B achieving much higher accuracy values. For instance, the CATBoost model shows an accuracy of 0.394 across all scaling methods in sample A, but this accuracy jumps to 0.884 in sample B under the same conditions. This discrepancy points to an imbalance in data quality or sample representativeness, leading to varying model performance depending on sample characteristics. The accuracy gap is particularly noteworthy in cases where Z-score normalization marginally improves model accuracy, especially in sample B. For example, RF’s accuracy in sample B improves from 0.851 (Raw dataset) to 0.868 (Z-score normalization), suggesting that scaling techniques can mitigate some of the effects of data imbalance but may not entirely address underlying distributional differences between samples. This unbalanced performance indicates that the dataset’s inherent characteristics, such as feature distributions and class proportions, may differ significantly between sample A and sample B. This imbalance could lead to biased predictions, particularly in scenarios where models trained on sample A might underperform due to a lack of feature scaling or sample diversity, thereby limiting the generalizability of the supervised learning models.
In sample A, the highest accuracy percentage of 41.2% was reached using the LightGBM approach with Z-score normalization, while in another case, its accuracy was dominant at 87.6%. Based on these results, LightGBM is an effective model where the performance may be limited by the preprocessing method and the dataset’s characteristics. The difference in the feature distribution in the dataset is a factor that may cause significant differences in accuracy values. The sensitivity of LightGBM to hyperparameters and data complexity means that the Z-score normalization process may not be optimal for this particular case. In sample B, the supervised approaches had an accuracy percentage of more than 80% and showed exemplary achievements. The superior performances demonstrate their robustness when trained on a larger dataset containing more representative and diverse features, allowing supervised models to generalize better. Increasing the amount of data significantly improves the accuracy of supervised learning, with an improvement of up to 40%. This result is consistent with those of the previous study [48]. Although the models applied normalization and unnormalization approaches, the difference in the performance was insignificant. As shown in Table 6, the XGBoost and the CATBoost models achieved accuracy values of 84.3% and 88.4%, respectively. Applying the normalization approach to data did not change the accuracy values significantly. In sample B, the CATBoost model was the most popular.
Table 7 and Table 8 show the evaluation metrics for the proposed machine learning models on training and testing datasets. Overall, all models achieved very high performance on the training set, with precision, recall, and F1-score exceeding 0.99 in most cases. However, the RF experienced a lack of variability or potential overfitting. In the test set, the models’ performances decreased slightly. XGBoost and LightGBM show a moderate decrease, but their scores are still relatively high, indicating a good balance between learning and generalization. CATBoost seems to generalize the best, as its training and testing scores remain very close. CATBoost maintains high consistency, making it a reliable choice for predicting CVD prognosis.
Table 9 shows the accuracy scores obtained from 10-fold cross-validation for different machine learning models that predict CVD prognosis. The accuracy values vary slightly across different folds. However, they remain within a reasonable range. Some variations are expected due to each fold’s different training and validation splits. LightGBM demonstrated strong performance with an accuracy mean of 0.892, followed closely by XGBoost (0.881), CATBoost (0.866), and RF (0.853). The standard deviation of accuracies (XGBoost: 0.023, RF: 0.022, LightGBM: 0.022, CATBoost: 0.020) indicated stable and consistent model performance across folds. The performance of LightGBM often reach its accuracy above 0.90; its achievement was consistent across most folds. LightGBM achieves the highest accuracy by multiple folds, indicating strong predictive performance. On the other hand, RF had a lower accuracy score than XGBoost and LightGBM, indicating it might not generalize as well for the proposed dataset. XGBoost achieved its accuracy values close to LightGBM in multiple folds, indicating it performed well. CATBoost provides moderate results but does not outperform LightGBM or XGBoost. Its accuracy values were slightly lower in some folds than XGBoost.
Table 10 shows the result of implementing the RFE technique with the voting ensemble-supervised learning approaches. The accuracy of the ensemble technique (A) is analyzed across feature sets (F) to optimize the performance of the model. Leveraging the models’ performance with the voting ensemble-supervised learning framework, mainly RF, reached a accuracy of 96.9%. The highest achievement of the RF-voting ensemble model was constructed with four features. Four dominant features—age, AO, LV, and LA—influence the model’s performance. As age increases, the LA value becomes higher. Table 10 highlights that many features do not determine the model performance. The accuracy score of the predictive LightGBM-XGBoost combination with 15 features was similar to that gained by the voting ensemble-XGBoost combination with 8 features. The relationship between the number of features and accuracy is not always proportional or inversely proportional. Therefore, the SHapley Additive Explanations (SHAP) plot was employed to analyze how the echocardiogram features affect disease prediction. The SHAP plot ranks the features from most impactful to least impactful based on their contribution to the prediction [17,49]. The negative values on the X-axis indicate a lower risk prediction, whereas the positive values on the right side indicate a higher risk prediction. The color represents the feature value; blue indicates the low feature value, and red indicates the high feature value. In the scenario with the XGboost classifier and RF-voting ensemble, the AO feature was placed in the first rank, indicating the feature is at a high risk of CVD. The features rank can be seen in Figure 4a,b. Figure 4a shows a more spread-out distribution, indicating that multiple features contribute to the model’s decision. Despite using more features, XGBoost may not effectively generalize due to feature interactions or overfitting.
Table 11 shows that combining the individual supervised learning model with the voting ensemble increases the model’s performance. The most significant improvement, almost 12%, occurred in the comparison of individual RF performance with the RF-voting ensemble. When combined with the voting ensemble, the individual CATBoost model and the LightGBM model experienced an increase of 2.5%, while other models only experienced an increase of 3%. Essentially, all the models showed improvement.

4. Discussion

The initial stage of this study involved data cleaning. Based on the number identification (ID) of patients in the repository, the 563 ultrasound data from the diagnosis result and treatment sections were compared with the 36,994 ultrasound data from the patients’ visitation history section. During data cleaning, we checked the columns for completeness and appropriate data types, categorizing columns as either numeric or categorical. Incomplete entries were handled by replacing null values with the median of each column. The gender column, containing categorical values, was encoded by assigning 1 to “M” instances and 0 to “F” instances. Next, an event-based self-similarity approach was used to identify patterns within time-series by comparing similar events across different instances in grouping patients. The term even refers to a significant change in echocardiographic characteristics. By quantifying self-similarity between events using euclidean distance metrics, unique patterns were detected, indicating high-risk patients. After preprocessing, the data was split into two samples (sample A and sample B). Each part was grouped into nine cohorts. Descriptive statistics were then applied to observe the data distribution of different clusters.
The normalized and unnormalized techniques are implemented to analyze the impact of the model’s performance. The subsequent data normalization processes applied the Z-score normalization technique and absolute maximum normalization technique to enhance data consistency and reduce the risk of overfitting. This normalization minimized dimensionality issues and bolstered dataset reliability for robust model training.
The model development framework entailed an exploratory search for optimal supervised learning models, each trained separately on samples A and B. Each model’s hyperparameters were fine-tuned for optimal performance. XGBoost builds trees sequentially to correct previous errors. XGBoost was configured with 100 estimators, considering the balance between accuracy and computational efficiency. Too few estimators can cause underfitting, while too many estimators can increase training time and the risk of overfitting. The learning rate controls how much each tree contributes to the final prediction. If the learning rate is below 0.3, more trees are needed for convergence, which results in increased computation time. The proposed XGBoost’s learning rate of 0.3 was moderate, allowing the model to learn effectively without requiring an excessive number of trees. On the other hand, RF builds trees independently and averages predictions to reduce variance. RF was set with nine estimators to keep computational costs low while maintaining sufficient model diversity for accurate predictions, whereas a random state of 1 to ensures the results’ reproducibility. CATBoost was set to a depth of 6 to ensure the balancing model interpretability and predictive power while keeping training efficient. Underfitting may occur when shallow trees have a depth below 4. Ensuring enough boosting steps for improved accuracy without excessive computation time, the iteration of CATBoost was up to 100 times per tree. Contrary to XGBoost, the LightGBM preferred a lower learning rate because it is more aggresive in growing trees. For LightGBM’s learning rate parameter, a lower learning rate allows for gradual optimization, reducing overfitting. The proposed LightGBM employed 31 leaves per tree with a learning rate of 0.1 for balancing generalization and learning capacity. Each model’s performance was assessed based on accuracy, precision, recall, F1-score. In addition, Cohen’s Kappa was utilized in this study to measure the interrater agreement. The Cohen’s Kappa metrics is frequently employed in classification problems to measure model performance in datasets with imbalanced values [50,51]. For further validation, 10-fold cross-validation was performed in sample B, achieving mean accuracy scores above 85% for all models.
Several features can help to identify the treatment and predict future diagnosis after treatment. Incorporating the supervised learning models XGBoost, RF, LightGBM, and CATBoost with the RFE technique facilitated the examination of possible feature combinations. The feature selection process was conducted using the RFE technique in Python 3.7.15. Figure 5 shows the compilation result of the RFE technique. The Explainable AI (XAI) technique offers an effective method for analyzing the importance of features within model predictive outcomes [49]. The importance of features was assessed using the SHAP method, which is one of the XAI. Key features influencing treatment outcomes and diagnostic predictions included age, AO, LV, and LA. These features, particularly impactful in the RF ensemble classifier, were consistent with prior findings, such as those from Wang [1], which identified age as an essential feature. These echocardiogram-based measurements provide clinicians with valuable diagnostic and prognostic insights. For example, in Group 23, we identified parameters associated with more prolonged survival among ventricular fibrillation patients. In contrast, Group 31 patients could benefit from more aggressive interventions, including defibrillator implantation or surgical treatments. While this study’s dataset size is limited, increasing the data volume in future work could expand the clinical utility of machine learning to guide treatment planning based on ultrasound results.
Afterward, to enhance the model’s performance, the single model of supervised learning framework was combined using a voting ensemble method and was employed in the dataset sample B. In this second framework, multi-model approaches can leverage the strengths of multiple models and reduce any single model’s bias. Multi-model approaches can also be made more robust and resistant to noise in the data. Figure 6 displays the results of executing the voting ensemble method with a single model in the form of a graph. The graph highlights the correspondence between the accuracy value and the number of features, the model name combined with voting, the accuracy value achieved when voting is combined, and the number of features supporting this accuracy.
Table 12 shows the comparison of accuracy between individual models, voting ensemble models, and previous studies. Figure 7 depicts the accuracy comparable between existing models and the proposed models. The proposed models have consistently high accuracy, above approximately 80%, indicating the proposed approaches are effective. The proposed models outperformed the earlier models by [1,32]. The performances of XGBoost and LightGBM from earlier studies show significantly lower accuracy. An accuracy of 66.30% was achieved when using XGBoost [1], whereas LightGBM [32] reached its accuracy at 77%. In [28], the RF model was employed, resulting in a 96% accuracy. Yang [29] demonstrated that LightGBM achieved an accuracy of 93%, whereas [33,34] reported the CATBoost’s accuracy scores of 86.30% and 98.08%, respectively. Some existing models appear comparable or even slightly higher than the proposed models, implying that the proposed models are still competitive. Figure 8 shows that each proposed voting ensemble model achieves a high accuracy, exceeding 85%, indicating strong predictive performance. The performance of the proposed RF-voting ensemble outperformed the model’s performance, which was studied by [35]. Using voting ensemble learning improves generalization, and its effectiveness in CVD prognosis remains promising and competitve with state-of-the-art models.

5. Conclusions and Future Work

Immersive technologies impact the development of data formats. Machine learning techniques can categorize types of heart disease by analyzing a patient’s echocardiographic data. This study demonstrated the utilizing event-based self-similarity for the effectiveness of machine learning in the automated prognosis of CVD, addressing the challenges posed by imbalanced echocardiogram data. The results indicate that class imbalance significantly affected model performance, with CATBoost model yielded the best individual results. However, ensemble techniques, particularly the voting ensemble approach, provided a more a robust solution, achieving an accuracy of 96.9% when combined with the RF model. This highlights the realibility of the ensemble voting approach in improving predictive performance. Furthermore, key features—age, AO, LV, and LA—were identified as significant predictors for CVD prognosis, reinforcing the model’s practical applicability in aiding physicians with early diagnosis and personalized treatment strategies. Even though the RFE technique was effective for future works, other feature selection methods, such as LASSO regression, SHAP values, or hybrid techniques, could be explored to enhance model interpretability and accuracy further. To address the class imbalance issue, the data augmentation approach, such as synthetic minority over-sampling, can be used on minority classes. For further advance, the deep learning method can also be complemented with an automation prognosis system. Overall, these findings highlight the potential of automated prognosis systems to assist physicians, with machine learning offering further improvements for personalized treatment of CVD patients, especially in the context of imbalanced datasets. By refining feature selection, optimizing class balance strategies, and leveraging advanced learning techniques. These findings pave the way for future advancements in AI-driven healthcare technologies.

Author Contributions

Conceptualization, W.-W.L., R.W., H.-N.H. and C.-T.Y.; methodology, W.-W.L. and C.-T.Y.; software, H.-M.C., Y.-C.C. and Y.-H.W.; validation, W.-W.L. and C.-T.Y.; investigation, Y.-H.W.; resources, W.-W.L.; writing—original draft preparation, H.-N.H., H.-M.C. and Y.-H.W.; writing—review and editing, R.W. and C.-T.Y.; supervision, W.-W.L. and C.-T.Y.; project administration, C.-T.Y.; funding acquisition, W.-W.L. and C.-T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Science and Technology Council research project grant, project number 113-2622-E-029-003 and 113-2221-E-029-028-MY3, and also by Taichung Veterans General Hospital, Taiwan, under Grant number: TCVGH-1127202C, TCVGH-T1117807, and TCVGH-T1137808.

Institutional Review Board Statement

The research materials of this study were mainly provided by the Department of Cardiology, Taichung Veterans General Hospital, and the Ethics Committee of their review committee had approved this study. (IRB number: CE23277C).

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are not openly available due to reasons of sensitivity and are available from the corresponding author upon reasonable request. Data are located in controlled access data storage at the Department of Cardiology, Taichung Veterans General Hospital.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Additional Figures

Figure A1. Pseudocode of proposed model.
Figure A1. Pseudocode of proposed model.
Diagnostics 15 00976 g0a1

References

  1. Wang, X.; Zhu, T.; Xia, M.; Liu, Y.; Wang, Y.; Wang, X.; Zhuang, L.; Zhong, D.; Zhu, J.; He, H.; et al. Predicting the Prognosis of Patients in the Coronary Care Unit: A Novel Multi-Category Machine Learning Model Using XGBoost. Front. Cardiovasc. Med. 2022, 9, 764629. [Google Scholar] [CrossRef] [PubMed]
  2. Sianga, B.E.; Mbago, M.C.; Msengwa, A.S. Predicting the prevalence of cardiovascular diseases using machine learning algorithms. Intell. Based Med. 2025, 11, 100199. [Google Scholar] [CrossRef]
  3. Damodharan, D.; Kumar Goel, A. A Novel Approach on classification in Machine Learning model-based USG Cardiogram Images. In Proceedings of the 2021 International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Greater Noida, India, 4–5 March 2021; pp. 947–950. [Google Scholar] [CrossRef]
  4. Kumar, N.; Kumar, D. Machine Learning based Heart Disease Diagnosis using Non-Invasive Methods: A Review. J. Phys. Conf. Ser. 2021, 1950, 012081. [Google Scholar] [CrossRef]
  5. Aly, I.; Rizvi, A.; Roberts, W.; Khalid, S.; Kassem, M.W.; Salandy, S.; Du Plessis, M.; Tubbs, R.S.; Loukas, M. Cardiac ultrasound: An Anatomical and Clinical Review. Transl. Res. Anat. 2021, 22, 100083. [Google Scholar] [CrossRef]
  6. Zhou, J.; Du, M.; Chang, S.; Chen, Z. Artificial intelligence in echocardiography: Detection, functional evaluation, and disease diagnosis. Cardiovasc. Ultrasound 2021, 19, 29. [Google Scholar] [CrossRef]
  7. Soares De Siqueira, V.; De Castro Rodrigues, D.; Nunes Dourado, C.; Marcos Borges, M.; Gomes Furtado, R.; Pereira Delfino, H.; Stelle, D.; Melgaco Barbosa, R.; Martins Da Costa, R. Machine Learning Applied to Support Medical Decision in Transthoracic Echocardiogram Exams: A Systematic Review. In Proceedings of the 2020 IEEE 44th Annual Computers, Software, and Applications Conference (COMPSAC), Madrid, Spain, 13–17 July 2020; pp. 400–407. [Google Scholar] [CrossRef]
  8. She, J.; Su, D.; Diao, R.; Wang, L. A Joint Model of Random Forest and Artificial Neural Network for the Diagnosis of Endometriosis. Front. Genet. 2022, 13, 848116. [Google Scholar] [CrossRef]
  9. Ono, Y.; Mitani, Y. Effect of the random forests with recursive feature elimination for breast cancer classification. In Proceedings of the 2021 6th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Oita, Japan, 25–27 November 2021; pp. 95–96. [Google Scholar] [CrossRef]
  10. Singh, D.; Khandelwal, A.; Bhandari, P.; Barve, S.; Chikmurge, D. Predicting Lung Cancer using XGBoost and other Ensemble Learning Models. In Proceedings of the 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), Delhi, India, 6–8 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  11. Bao, J. Multi-features Based Arrhythmia Diagnosis Algorithm Using XGBoost. In Proceedings of the 2020 International Conference on Computing and Data Science (CDS), Stanford, CA, USA, 1–2 August 2020; pp. 454–457. [Google Scholar] [CrossRef]
  12. Montomoli, J.; Romeo, L.; Moccia, S.; Bernardini, M.; Migliorelli. Machine learning using the extreme gradient boosting (XGBoost) algorithm predicts 5-day delta of SOFA score at ICU admission in COVID-19 patients. J. Intensive Med. 2021, 1, 110–116. [Google Scholar] [CrossRef]
  13. Goyal, J.; Khandnor, P.; Aseri, T.C. Analysis of Parkinson’s disease diagnosis using a combination of Genetic Algorithm and Recursive Feature Elimination. In Proceedings of the 2020 Fourth World Conference on Smart Trends in Systems, Security and Sustainability (WorldS4), London, UK, 27–28 July 2020; pp. 268–272. [Google Scholar] [CrossRef]
  14. Nanehkaran, Y.A.; Licai, Z.; Chen, J.; Zhongpan, Q.; Xiaofeng, Y.; Navaei, Y.D.; Einy, S. Diagnosis of Chronic Diseases Based on Patients’ Health Records in IoT Healthcare Using the Recommender System. Wirel. Commun. Mob. Comput. 2022, 2022, 5663001. [Google Scholar] [CrossRef]
  15. Alturki, N.; Altamimi, A.; Umer, M.; Saidani, O.; Alshardan, A.; Alsubai, S.; Omar, M.; Ashraf, I. Improving Prediction of Chronic Kidney Disease Using KNN Imputed SMOTE Features and TrioNet Model. CMES Comput. Model. Eng. Sci. 2024, 139, 3513–3534. [Google Scholar] [CrossRef]
  16. Liu, J.; Xu, L.; Zhu, E.; Han, C.; Ai, Z. Prediction of acute kidney injury in patients with femoral neck fracture utilizing machine learning. Front. Surg. 2022, 9, 928750. [Google Scholar] [CrossRef]
  17. Tanone, R.; Li, L.H.; Saifullah, S. ViT-CB: Integrating hybrid Vision Transformer and CatBoost to enhanced brain tumor detection with SHAP. Biomed. Signal Process. Control. 2025, 100, 107027. [Google Scholar] [CrossRef]
  18. Borkar, S.; Annadate, M.N. Supervised Machine Learning Algorithm for Detection of Cardiac Disorders. In Proceedings of the 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA), Pune, India, 16–18 August 2018; pp. 1–4. [Google Scholar] [CrossRef]
  19. Sachdeva, R.K.; Singh, K.D.; Sharma, S.; Bathla, P.; Solanki, V. An Organized Method for Heart Failure Classification. In Proceedings of the 2023 International Conference on Emerging Smart Computing and Informatics (ESCI), Pune, India, 1–3 March 2023; pp. 1–6. [Google Scholar] [CrossRef]
  20. Rath, A.; Mishra, D.; Panda, G. Imbalanced ECG signal-based heart disease classification using ensemble machine learning technique. Front. Big Data 2022, 5, 1021518. [Google Scholar] [CrossRef] [PubMed]
  21. Ogunpola, A.; Saeed, F.; Basurra, S.; Albarrak, A.M.; Qasem, S.N. Machine Learning-Based Predictive Models for Detection of Cardiovascular Diseases. Diagnostics 2024, 14, 144. [Google Scholar] [CrossRef] [PubMed]
  22. Li, J.P.; Haq, A.U.; Din, S.U.; Khan, J.; Khan, A.; Saboor, A. Heart Disease Identification Method Using Machine Learning Classification in E-Healthcare. IEEE Access 2020, 8, 107562–107582. [Google Scholar] [CrossRef]
  23. Guo, A.; Pasque, M.; Loh, F.; Mann, D.L.; Payne, P.R.O. Heart Failure Diagnosis, Readmission, and Mortality Prediction Using Machine Learning and Artificial Intelligence Models. Curr. Epidemiol. Rep. 2020, 7, 212–219. [Google Scholar] [CrossRef]
  24. Miao, L.; Wang, W. Cardiovascular Disease Prediction Based on Soft Voting Ensemble Model. J. Phys. Conf. Ser. 2023, 2504, 012021. [Google Scholar] [CrossRef]
  25. Upton, R.; Mumith, A.; Beqiri, A.; Parker, A.; Hawkes, W.; Gao, S.; Porumb, M.; Sarwar, R.; Marques, P.; Markham, D.; et al. Automated Echocardiographic Detection of Severe Coronary Artery Disease Using Artificial Intelligence. JACC Cardiovasc. Imaging 2022, 15, 715–727. [Google Scholar] [CrossRef]
  26. Baviskar, V.; Ojha, A.; Verma, I.; Mahla, S.; Verma, M.; Chatterjee, P.; Kumar, S. H2EMDCO: Design of a Hyper Ensemble Model for Evaluation of Heart Diseases from Clinical Observations. Procedia Comput. Sci. 2023, 218, 1288–1294. [Google Scholar] [CrossRef]
  27. Baranidharan, B.; Pal, A.; Muruganandam, P. Cardio-vascular disease prediction based on ensemble technique enhanced using extra tree classifier for feature selection. Int. J. Recent Technol. Eng. 2019, 8, 3236–3242. [Google Scholar] [CrossRef]
  28. Sumwiza, K.; Twizere, C.; Rushingabigwi, G.; Bakunzibake, P.; Bamurigire, P. Enhanced cardiovascular disease prediction model using random forest algorithm. Inform. Med. Unlocked 2023, 41, 101316. [Google Scholar] [CrossRef]
  29. Yang, H.; Chen, Z.; Yang, H.; Tian, M. Predicting Coronary Heart Disease Using an Improved LightGBM Model: Performance Analysis and Comparison. IEEE Access 2023, 11, 23366–23380. [Google Scholar] [CrossRef]
  30. Li, D.; Zheng, C.; Zhao, J.; Liu, Y. Diagnosis of heart failure from imbalance datasets using multi-level classification. Biomed. Signal Process. Control. 2023, 81, 104538. [Google Scholar] [CrossRef]
  31. Mullick, S.S.; Datta, S.; Das, S. Adaptive learning-based k-nearest neighbor classifiers with resilience to class imbalance. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5713–5725. [Google Scholar] [CrossRef] [PubMed]
  32. Boudali, I.; Chebaane, S.; Zitouni, Y. A predictive approach for myocardial infarction risk assessment using machine learning and big clinical data. Healthc. Anal. 2024, 5, 100319. [Google Scholar] [CrossRef]
  33. Aziz, S.; Afreen, N.; Akram, F.; Ahmed, M. A Framework for Cardiac Arrest Prediction via Application of Ensemble Learning Using Boosting Algorithms. Procedia Comput. Sci. 2024, 235, 3293–3304. [Google Scholar] [CrossRef]
  34. Wei, X.; Rao, C.; Xiao, X.; Chen, L.; Goh, M. Risk assessment of cardiovascular disease based on SOLSSA-CatBoost model. Expert Syst. Appl. 2023, 219, 119648. [Google Scholar] [CrossRef]
  35. Sen, K.; Verma, B. Heart Disease Prediction Using a Soft Voting Ensemble of Gradient Boosting Models, RandomForest, and Gaussian Naive Bayes. In Proceedings of the 2023 4th International Conference for Emerging Technology (INCET), Belgaum, India, 26–28 May 2023; pp. 1–7. [Google Scholar] [CrossRef]
  36. Atallah, R.; Al-Mousa, A. Heart Disease Detection Using Machine Learning Majority Voting Ensemble Method. In Proceedings of the International Conference on New Trends in Computing Sciences (ICTCS), Amman, Jordan, 9–11 October 2019. [Google Scholar] [CrossRef]
  37. Abidin, A.J.; Borhan, R.; Akhter, A.; Jahan, S.; Based, M.A.; Zilany, A.K. A Voting Ensemble based Machine Learning Approach to Predict Cardiovascular Disease. In Proceedings of the International Conference on Computer and Information Technology (ICCIT), Cox’s Bazar, Bangladesh, 13–15 December 2023. [Google Scholar] [CrossRef]
  38. Alsahaf, A.; Petkov, N.; Shenoy, V.; Azzopardi, G. A framework for feature selection through boosting. Expert Syst. Appl. 2022, 187, 115895. [Google Scholar] [CrossRef]
  39. Abdullahi, A.; Barre, M.A.; Elmi, A.H. A machine learning approach to cardiovascular disease prediction with advanced feature selection. Indones. J. Electr. Eng. Comput. Sci. 2024, 33, 1030. [Google Scholar] [CrossRef]
  40. Ogunleye, A.; Wang, Q.G. XGBoost Model for Chronic Kidney Disease Diagnosis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 17, 2131–2140. [Google Scholar] [CrossRef]
  41. Albert, A.J.; Murugan, R.; Sripriya, T. Diagnosis of heart disease using oversampling methods and decision tree classifier in cardiology. Res. Biomed. Eng. 2023, 39, 99–113. [Google Scholar] [CrossRef]
  42. Shobayo, O.; Zachariah, O.; Odusami, M.O.; Ogunleye, B. Prediction of Stroke Disease with Demographic and Behavioural Data Using Random Forest Algorithm. Analytics 2023, 2, 604–617. [Google Scholar] [CrossRef]
  43. Ahirwal, M.K.; Londhe, N.D.; Kumar, A. Artificial Intelligence Applications for Health Care, 1st ed.; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar] [CrossRef]
  44. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the Knowledge Discovery and Data Mining International Conference, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  45. Omotehinwa, T.O.; Oyewola, D.O.; Moung, E.G. Optimizing the light gradient-boosting machine algorithm for an efficient early detection of coronary heart disease. Informatics Health 2024, 1, 70–81. [Google Scholar] [CrossRef]
  46. Botros, J.; Mourad-Chehade, F.; Laplanche, D. Explainable multimodal data fusion framework for heart failure detection: Integrating CNN and XGBoost. Biomed. Signal Process. Control. 2025, 100, 106997. [Google Scholar] [CrossRef]
  47. Dai, Y.; Niu, L.; Wei, L.; Tang, J. Feature Selection in High Dimensional Biomedical Data Based on BF-SFLA. Front. Neurosci. 2022, 16, 854685. [Google Scholar] [CrossRef]
  48. Bailly, A.; Blanc, C.; Francis, É.; Guillotin, T.; Jamal, F.; Wakim, B.; Roy, P. Effects of dataset size and interactions on the prediction performance of logistic regression and deep learning models. Comput. Methods Progr. Biomed. 2022, 213, 106504. [Google Scholar] [CrossRef]
  49. Rezk, N.G.; Alshathri, S.; Sayed, A.; Hemdan, E.E.D.; El-Behery, H. XAI-Augmented Voting Ensemble Models for Heart Disease Prediction: A SHAP and LIME-Based Approach. Bioengineering 2024, 11, 1016. [Google Scholar] [CrossRef]
  50. Zhao, X.; Feng, G.C.; Ao, S.H.; Liu, P.L. Interrater reliability estimators tested against true interrater reliabilities. BMC Med. Res. Methodol. 2022, 22, 232. [Google Scholar] [CrossRef]
  51. Wang, T.W.; Tzeng, Y.H.; Wu, K.T.; Liu, H.R.; Hong, J.S.; Hsu, H.Y.; Fu, H.N.; Lee, Y.T.; Yin, W.H.; Wu, Y.T. Meta-analysis of deep learning approaches for automated coronary artery calcium scoring: Performance and clinical utility AI in CAC scoring: A meta-analysis. Comput. Biol. Med. 2024, 183, 109295. [Google Scholar] [CrossRef]
Figure 1. Block diagram.
Figure 1. Block diagram.
Diagnostics 15 00976 g001
Figure 2. Frequency of sample A and sample B. (a) Frequency of sample A, (b) Frequency of sample B.
Figure 2. Frequency of sample A and sample B. (a) Frequency of sample A, (b) Frequency of sample B.
Diagnostics 15 00976 g002
Figure 3. Histogram for the continous attributes.
Figure 3. Histogram for the continous attributes.
Diagnostics 15 00976 g003
Figure 4. The ranked of features in the SHAP plot. (a) SHAP of XGBoost ensemble with 8 features, (b) SHAP of RF-Voting ensemble with 4 features.
Figure 4. The ranked of features in the SHAP plot. (a) SHAP of XGBoost ensemble with 8 features, (b) SHAP of RF-Voting ensemble with 4 features.
Diagnostics 15 00976 g004
Figure 5. RFE technique implementation.
Figure 5. RFE technique implementation.
Diagnostics 15 00976 g005
Figure 6. Voting ensemble with CATBoost implementation.
Figure 6. Voting ensemble with CATBoost implementation.
Diagnostics 15 00976 g006
Figure 7. Comparison of the proposed study with the previous studies.
Figure 7. Comparison of the proposed study with the previous studies.
Diagnostics 15 00976 g007
Figure 8. Comparison of the proposed voting ensemble model with the previous studies.
Figure 8. Comparison of the proposed voting ensemble model with the previous studies.
Diagnostics 15 00976 g008
Table 1. Some recent studies on CVD.
Table 1. Some recent studies on CVD.
AuthorsBest ModelAccuracyTotal ParameterTotal SamplesRepository
Wang [1]XGBoost66.3%305360MIMIC-III
Sachdeva [19]SVM96.67%13299UCI Heart Failure Clinical
Sumwiza [28]RF96%141025Kaggle
Yang [29]LightGBM93%104240CHD Framingham Heart Institute
Boudali [32]LightGBM77%2147,786BRFSS
Aziz [33]CATBoost98.08%14n.aUCI Cleveland
Wei [34]CATBoost86.30%11918Kaggle
Sen [35]Soft voting91.85%11918UCI and Statlog
Atallah [36]Hard voting ensemble90%14303UCI
Abidin [37]Hard voting ensemble90%11303Kaggle
Table 2. Grouping patient treatments and diagnosis categories.
Table 2. Grouping patient treatments and diagnosis categories.
Treatments in Three-Year Follow-UpDiedFollow-UpNo Follow-Up
Cardiac catheter ablationGroup 11Group 13Group 14
Ventricular defibrillatorsGroup 21Group 23Group 24
Drug controlGroup 31Group 33Group 34
Table 3. The Sample A’s distribution in diagnosis and treatment categories.
Table 3. The Sample A’s distribution in diagnosis and treatment categories.
Treatments in Three-Year Follow-UpSample A
DiedFollow-UpNo Follow-Up
Cardiac Catheter Ablation6375
Ventricular Defibrillator6622077
Drug Control1011610
Table 4. The Sample B’s distribution in diagnosis and treatment categories.
Table 4. The Sample B’s distribution in diagnosis and treatment categories.
Treatments in Three-Year Follow-UpSample B
DiedFollow-UpNo Follow-Up
Cardiac Catheter Ablation2010022
Ventricular Defibrillator150498121
Drug Control2426017
Table 5. The descriptive statistics result of group 23.
Table 5. The descriptive statistics result of group 23.
Group23AgeLV (mm)VS (mm)LVPW (mm)LA (mm)AO (mm)TR_PG_Mean (mmHG)LVEF
count220220220220220220220220
mean76.746.81312.646.536.224.256
std13.37.52.11.69.64.58.18.1
min264870241021
25%694211.711.441331852
50%8145.71312.445.4362357
75%8651.41413.951.63927.861
max96752118.1100515769
Table 6. Supervised techniques accuracy.
Table 6. Supervised techniques accuracy.
ModelsSample ASample B
Raw DatasetZ-ScoreAbs MaximumRaw DatasetZ-ScoreAbs Maximum
XGBoost0.4060.4060.4060.8430.8430.843
RF0.4060.4060.4060.8510.8680.851
LightGBM0.3940.4120.3940.8680.8760.868
CATBoost0.3940.3940.3940.8840.8840.884
Table 7. The results of train evaluation metrics.
Table 7. The results of train evaluation metrics.
ModelPrecisionRecallF1-ScoreCohen’s Kappa
XGBoost0.9920.9920.9920.984
RF1.0001.0001.0001.000
LightGBM0.9900.9900.9900.980
CATBoost0.9920.9920.9920.984
Table 8. The results of testing evaluation metrics.
Table 8. The results of testing evaluation metrics.
ModelPrecisionRecallF1-ScoreCohen’s Kappa
XGBoost0.9490.9170.9320.874
RF0.9580.9590.9550.913
LightGBM0.9580.9420.9500.905
CATBoost0.9900.9900.9890.980
Table 9. The K-fold cross validation results.
Table 9. The K-fold cross validation results.
K-FoldXGBoostRFLightGBMCATBoost
10.9130.8350.9130.874
20.8740.8500.8980.874
30.9130.8820.9210.889
40.8820.8660.9060.889
50.8500.8190.8430.835
60.8660.8580.8740.866
70.9060.8660.8820.889
80.8890.8500.9060.850
90.8500.8190.8980.835
100.8650.8810.8810.857
Table 10. The accuracy (A) of comparison models (ensemble technique) with different number of features (F).
Table 10. The accuracy (A) of comparison models (ensemble technique) with different number of features (F).
Predictive Models FeaturesXGBoostRFLightGBMCATBoost
AFAFAFAF
XGBoost0.851120.87680.89370.8935
RF0.88480.87640.89370.8934
LightGBM0.876150.90180.90170.9017
CATBoost0.868110.89390.89390.8847
Voting Ensemble0.87680.96940.89370.9094
Table 11. The comparison accuracy between individual models and voting ensemble model.
Table 11. The comparison accuracy between individual models and voting ensemble model.
ModelIndividualVoting Ensemble
XGBoost0.8430.876
RF0.8510.969
LightGBM0.8680.893
CATBoost0.8840.909
Table 12. The comparison of accuracy between individual models, voting ensemble models, and previous studies.
Table 12. The comparison of accuracy between individual models, voting ensemble models, and previous studies.
ModelAccuracy
XGBoost [1]0.663
RF [28]0.960
LightGBM [29]0.930
LightGBM [32]0.770
CATBoost [33]0.980
CATBoost [34]0.863
Soft voting [35]0.918
Hard voting [36]0.900
Hard voting [37]0.900
XGBoost proposed model0.843
RF proposed model0.851
LightGBM proposed model0.868
CATBoost proposed model0.884
XGBoost-Voting ensemble proposed model0.876
RF-Voting ensemble proposed model0.969
LightGBM-Voting ensemble proposed model0.893
CATBoost-Voting ensemble proposed model0.909
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, H.-N.; Chen, H.-M.; Lin, W.-W.; Wiryasaputra, R.; Chen, Y.-C.; Wang, Y.-H.; Yang, C.-T. Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity. Diagnostics 2025, 15, 976. https://doi.org/10.3390/diagnostics15080976

AMA Style

Huang H-N, Chen H-M, Lin W-W, Wiryasaputra R, Chen Y-C, Wang Y-H, Yang C-T. Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity. Diagnostics. 2025; 15(8):976. https://doi.org/10.3390/diagnostics15080976

Chicago/Turabian Style

Huang, Huang-Nan, Hong-Min Chen, Wei-Wen Lin, Rita Wiryasaputra, Yung-Cheng Chen, Yu-Huei Wang, and Chao-Tung Yang. 2025. "Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity" Diagnostics 15, no. 8: 976. https://doi.org/10.3390/diagnostics15080976

APA Style

Huang, H.-N., Chen, H.-M., Lin, W.-W., Wiryasaputra, R., Chen, Y.-C., Wang, Y.-H., & Yang, C.-T. (2025). Automatic Feature Selection for Imbalanced Echocardiogram Data Using Event-Based Self-Similarity. Diagnostics, 15(8), 976. https://doi.org/10.3390/diagnostics15080976

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop