Next Article in Journal
Hospital Readmissions in Patients Supported with Durable Centrifugal-Flow Left Ventricular Assist Devices
Previous Article in Journal
Pulmonary Embolism in Pregnancy: A Review for Clinical Practitioners
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel CAD Diagnosis Method Based on Search, PCA, and AdaBoostM1 Techniques

1
Department of Computer Engineering, Turkish Air Force Academy, National Defence University, Istanbul 34149, Türkiye
2
School of Computer Science and Informatics, Cardiff University, Cardiff CF24 4AG, UK
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2024, 13(10), 2868; https://doi.org/10.3390/jcm13102868
Submission received: 24 March 2024 / Revised: 26 April 2024 / Accepted: 7 May 2024 / Published: 13 May 2024
(This article belongs to the Section Cardiology)

Abstract

:
Background: Cardiovascular diseases (CVDs) are the primary cause of mortality worldwide, resulting in a growing number of annual fatalities. Coronary artery disease (CAD) is one of the basic types of CVDs, and early diagnosis of CAD is crucial for convenient treatment and decreasing mortality rates. In the literature, several studies use many features for CAD diagnosis. However, due to the large number of features used in these studies, the possibility of early diagnosis is reduced. Methods: For this reason, in this study, a new method that uses only five features—age, hypertension, typical chest pain, t-wave inversion, and region with regional wall motion abnormality—and is a combination of eight different search techniques, principal component analysis (PCA), and the AdaBoostM1 algorithm has been proposed for early and accurate CAD diagnosis. Results: The proposed method is devised and tested on a benchmark dataset called Z-Alizadeh Sani. The performance of the proposed method is tested with a variety of metrics and compared with basic machine-learning techniques and the existing studies in the literature. The experimental results have shown that the proposed method is efficient and achieves the best classification performance, with an accuracy of 91.8%, ever reported on the Z-Alizadeh Sani dataset with so few features. Conclusions: As a result, medical practitioners can utilize the proposed approach for diagnosing CAD early and accurately.

1. Introduction

Cardiovascular diseases (CVDs) are a class of disorders that include the blood and heart vessels [1,2]. The main types and risk factors of CVDs are shown in Figure 1. Coronary artery disease (CAD) is an illness that influences the blood vessels providing blood to the heart and occurs when coronary arteries are blocked or narrowed. Additionally, it is associated with ischemic heart disease, coronary heart disease, atherosclerotic heart disease, heart failure, heart attack, sudden coronary death, and angina pectoris medical science. A stroke is disease-causing damage to a particular brain area and occurs when blood vessels are ruptured or blocked. Finally, peripheral artery disease affects the blood vessels that supply blood to the feet and legs [2]. Readers are referred to [2] for the main risk factors of CVDs and characteristics that can be associated with CVDs.
CVDs are the main cause of death all over the world. In 2021, the World Health Organization (WHO) [3] reported that nearly 17.9 million people died from CVDs in 2019. This number of mortalities constitutes 32% of the total deaths worldwide, and 85% of these deaths are caused by heart attack and stroke. In addition, the top ten global causes of mortality in 2019 are shown in Figure 2. As shown in the figure, CAD, also known as ischemic heart disease, is the leading reason for deaths [4].
Disease diagnosis is a highly complex process in medical science, and many tests are necessary for accurate diagnosis. In order to help medical doctors with the early detection of disease, machine learning and data mining techniques have been widely utilized recently [5]. Especially in CAD, with early detection, the possibility of treatment is greatly increased and patients’ lives can be saved.
In the literature, numerous methods have been developed to diagnose CAD on the Cleveland heart disease dataset [6] up to the present [7,8,9,10,11,12,13,14,15,16,17]. The prediction and diagnosis success of the existing studies tested on this dataset is quite satisfying. For this reason, in this study, the performance of the proposed approach is evaluated on a newer dataset called the Z-Alizadeh Sani [18], released in 2017. In the work introduced by Alizadehsani et al. [19], the Z-Alizadeh Sani dataset was collected and utilized for the first time for CAD diagnosis. From 2012 to 2016, Alizadehsani et al. employed various machine learning techniques such as sequential minimal optimization (SMO), artificial neural network (ANN), support vector machine (SVM), Naïve Bayes, bagging, C4.5 decision tree, information gain, and a genetic algorithm for CAD detection [19,20,21,22,23,24,25]. In a study using the same dataset, Qin et al. [26] presented a CAD detection method utilizing an ensemble algorithm based on multiple feature selection (EA-MFS) and SVM. Arabasadi et al. [27] proposed a hybrid CAD prediction approach combining a genetic algorithm and multilayer perceptron artificial neural network (MLP-ANN) on a subset of the Z-Alizadeh Sani dataset from which 22 features were selected. In order to diagnose CAD, Babič et al. [28] offered a predictive and descriptive analysis. They used four different classifiers such as decision trees, Naïve Bayes, SVM, and ANN.
In the work of Kılıç and Kaya Keleş [29], the artificial bee colony (ABC) algorithm and SMO technique were utilized for feature selection and classification, respectively. Sixteen features were selected by the ABC algorithm, and SMO was applied to these features. Hu et al. [30] proposed two methods, namely, minimum message length finite inverted Beta-Liouville mixture (MML-IBLMM) and variational finite inverted Beta-Liouville mixture (Var-IBLMM), and then tested the performances of these models on Z-Alizadeh Sani dataset. In the study introduced by Abdar et al. [31], a CAD detection technique called N2Genetic-nuSVM, which is based on a genetic optimizer and nu-support vector classification, was presented. In another work realized by Abdar et al. [32], a nested ensemble nu-support vector classification (NE-nu-SVC) approach was proposed to diagnose CAD accurately. In the feature selection step of the proposed approach, a genetic search method was utilized, and 16 features were selected. In the research of Joloudari et al. [33], the performances of SVM, chi-squared automatic interaction detection (CHAID) decision tree, C5.0 decision tree, and random trees were investigated for CAD diagnosis. The experimental results indicate that the random trees technique is better than the other classifiers. On the other hand, Nasarian et al. [34] presented a hybrid feature selection method named heterogeneous hybrid feature selection (2HFS) that utilizes the synthetic minority over-sampling technique (SMOTE) and adaptive synthetic (ADASYN) for handling the Z-Alizadeh Sani dataset and uses random forests, Gaussian Naïve Bayes, eXtreme Gradient Boosting (XGBoost), and decision tree for CAD classification. In another work proposed by Ashish et al. [35], a CAD detection method based on random forests, SVM, and XGBoost was introduced. In the data-dividing step of the method, the random forests technique was used for training and testing of the Z-Alizadeh Sani dataset. In the classification step, the SVM and XGBoost techniques were utilized together. In a recent study [36], an ensemble feature selection approach and seven classifiers were used, and the best classification accuracy rate was attained with 25 features and the MLP classifier.
The aforementioned studies adopted some combination of feature selection, feature extraction, and classification techniques such as information gain, genetic algorithm, ABC, bagging, decision trees, random trees, Naïve Bayes, SMO, SVM, SVC, ANN, CHAID, and XGBoost to diagnose CAD. Unlike the abovementioned methods, this work proposes a new CAD diagnosis method based on eight different search techniques, principal component analysis (PCA), and AdaBoostM1. To the best of the author’s knowledge, there is no other work in the literature utilizing PCA and AdaBoostM1 techniques together for CAD diagnosis in this framework and detecting CAD based on age, hypertension, typical chest pain, t-wave inversion, and region with regional wall motion abnormality features. The major findings and contributions of this research study are as follows:
  • Proposes a new method to diagnose CAD based on age, hypertension, typical chest pain, t-wave inversion, and region with regional wall motion abnormality features.
  • Explores attribute spaces using eight different search methods, namely, evolutionary, best first, genetic, harmony, particle swarm optimization (PSO), greedy stepwise, rank, and multi-objective evolutionary search.
  • Enhances the performance of CAD diagnosis by efficiently taking advantage of using PCA and AdaBoostM1 techniques together.
  • The performance of the proposed method is tested in terms of several metrics and compared with basic classifiers and existing studies in the literature.
  • Achieves the best classification performance ever reported on the Z-Alizadeh Sani dataset with so few features (five) with an accuracy rate of 91.80%.
  • The experimental results demonstrate that the proposed method is promising to be utilized by medical specialists for diagnosing CAD.
The rest of the paper is ordered as follows. The proposed CAD diagnosis method and the dataset used are introduced in Section 2. Section 3 shows the experimental results, comparing the proposed method’s performance to the existing studies in the literature. Finally, conclusions are summarized in Section 4.

2. Materials and Methods

2.1. Dataset Description

In this work, the Z-Alizadeh Sani dataset that is freely available from the University of California—Irvine Machine Learning Repository [18] was used to evaluate the proposed method. The dataset contains 303 records, of which 87 of them are healthy persons and 216 of them are CAD patients. Fifty-five attributes can be classified into four groups: symptom and examination (14 attributes), demographic (17 attributes), electrocardiography (ECG) (7 attributes), and laboratory and echocardiography (echo) (17 attributes). The overview of the Z-Alizadeh Sani dataset, including attribute name, category, and range, is shown in Table 1.

2.2. The Proposed CAD Diagnosis Method

This study presents a new CAD diagnosis method based on age, HTN, typical chest pain, t-wave inversion, and region-RWMA features. The proposed method comprises four basic steps, which are feature selection, feature extraction, data dividing, and classification. The flowchart of the proposed CAD diagnosis method is demonstrated in Figure 3. A correlation-based feature subset selection technique is utilized with evolutionary, best first, genetic, harmony, PSO, greedy stepwise, rank, and multi-objective evolutionary search methods in the feature selection step. Then, the PCA technique transforming the data into another space is used for feature extraction and size reduction on the data obtained after selecting common attributes. In the data-dividing step, the k-fold cross-validation technique is exploited to divide the whole dataset into k separate subsets, in which k-1 subsets are utilized for training and the remaining part is separated for testing. In the classification step, the AdaBoostM1 algorithm is performed for classifying coronary artery disease as healthy or patient. The techniques utilized to perform the proposed diagnostic method are described in the following subsections.

2.2.1. Feature Selection

In the feature selection step of the proposed method, a correlation-based feature subset selection technique [37] was used with eight different search methods, namely, evolutionary [38], best first [39], genetic [40], harmony [41], PSO [42], greedy stepwise [43], rank [44], and multi-objective evolutionary search [45]. To evaluate the worth of a subset of attributes, the feature subset selection technique considers the estimative ability of every feature associated with the redundancy degree between them. The evolutionary search method utilizes an evolutionary algorithm (EA) to discover the attribute space. The best first search method uses a greedy hillclimbing algorithm enhanced with a backtracking ability for searching the space of a subset of attributes. The genetic search method carries out a search utilizing Goldberg’s genetic algorithm. The greedy stepwise search method applies a greedy backward/forward search, along with the space of a subset of attributes. To rank all the attributes, the rank search method utilizes a subset or attribute evaluator. Finally, the harmony, PS,O and multi-objective evolutionary search methods explore the attribute space using the harmony, PSO, and multi-objective evolutionary algorithms, respectively. Interested readers can kindly refer to [38,39,40,41,42,43,44,45] for further details about the search methods.
In the initial phase of the feature selection process outlined in the proposed method, eight distinct search methods are employed to explore and identify useful attributes. Table 2 presents the attributes selected through these search methods, along with the respective counts and attribute numbers. Subsequently, in the second stage, attributes common to all search methods are retained. As indicated in Table 2 (common attributes highlighted in bold), the selected attributes are numbered 1, 7, 25, 35, and 54, corresponding to features such as Age, HTN, Typical Chest Pain, T-Wave Inversion, and Region-RWMA.

2.2.2. Feature Extraction

The data collected from a system often have dozens of related attributes. However, there may only be a few actual driving forces governing the behavior of a system, even though we have more attributes in the data measuring many system variables that provide redundant information [46]. It is usually possible to simplify problems containing redundancy by taking advantage of dimensionality reduction techniques. PCA is one of the most famous kinds of dimensionality reduction methods and has been widely used in various fields till now. It is intensely used for dimension reduction and feature extraction purposes as it decreases overfitting risk, reduces computational complexity, eliminates distracting noise, and so on [47].
PCA employs orthogonal transformations to condense multiple correlated variables into a reduced set of uncorrelated variables [47,48]. This technique establishes a new orthogonal-basis space where each axis represents a principal component, formed as a linear combination of the original data variables. By rigorously calculating these principal components, PCA ensures no redundancy of information within this new space [46]. Maximizing variance along each axis, PCA aligns the first axis with the highest variance of the data points, while the subsequent axes are orthogonal to the previous ones, sequentially maximizing the remaining variance [46]. Hence, in the transformed space, principal components are arranged in descending order of variance, with the first component explaining the most variance and subsequent components explaining progressively less [47,49].
The mathematical formulations required to compute the principal components are given hereafter. Let x ( t ) for t = 1 , 2 , , n be an arbitrary dataset containing its corresponding instances and features with zero mean. Its covariance matrix R is computed as follows:
R = 1 n 1 t = 1 n x ( t ) x ( t ) T
The next equation can be utilized to compute linear combinations of variables in the original data, i.e., the linear transformation from x ( t ) to y ( t ) ,
y ( t ) = M T x ( t )
where M denotes an orthogonal matrix of the size n × n , and the ith column of this matrix, also of the sample covariance matrix R, is essentially equal to the ith eigenvector. At this point, the eigenvalue problem is initially set to be solved by the following equation:
λ l q l = R q l
where q l represents the corresponding eigenvector, and λ l stands for an eigenvalue of the covariance matrix R (consider λ 1 > λ 2 > > λ n ) . Based on Equation (4), the principal component is computed by
y l ( t ) = q l T x ( t ) , l = 1 , , n
where y l ( t ) stands for the ith principle component. For additional information and further details, readers can refer to the references [47,48].
For the Z-Alizadeh Sani dataset with selected attributes, Figure 4 illustrates the variance values explained by each principal component generated and depicts only the first eight components, which account for around 95% of the total variance.

2.2.3. Data Dividing

A methodology known as k-fold cross-validation can be used to reduce the bias related to a random sampling of the holdout and training data samples when comparing the predicted accuracy of two or more methods. The entire data set is randomly separated into k mutually exclusive subsets of approximately similar size in k-fold cross-validation, also known as rotation estimation. The classification technique is trained and tested k times. k-1 of mutually exclusive subsets are utilized for training, while the remaining one is reserved for testing. With averaging the k individual accuracy measures, the prediction of a technique’s overall accuracy is computed as
C V a c c u r a c y = 1 k i = 1 k A i
where A represents the accuracy measure of each fold such as specificity, sensitivity, and hit rate, and k denotes the number of used folds [50,51].
Since it is the most widespread practice for k to have a value of 10, the k-fold cross-validation is also known as 10-fold cross-validation. Empirical studies have shown that the optimal number of folds seems to be 10 [50,51]. For this reason, in this study, 10-fold cross-validation was utilized for evaluating the proposed diagnosis method. Figure 5 shows a visualization of k-fold cross-validation with k = 10 [50,51].

2.2.4. Classification

In the classification step of the proposed method, the AdaBoostM1 algorithm [52] is utilized to classify coronary artery disease as patient or normal. The following is a description of the AdaBoostM1 algorithm. T n = { ( X 1 , Y 1 ) , ( X 2 , Y 2 ) , , ( X n , Y n ) } is a training set with Y values in 1 , 2 , , k . Each observation X i is given a weight w b ( i ) , which is originally set to 1 / n . After each step, this value is updated. The classifier’s error is denoted by ϵ b and is calculated as follows:
ϵ b = i = 1 n w b ( i ) ξ b ( i )
where
ξ b = 0 , C b ( x i ) = y i 1 , C b ( x i ) y i
The constant α b is calculated from the classifier’s error in the bth iteration, and this value is utilized for the weight update. Particularly, α b = 1 / 2 l n ( ( 1 ϵ b ) / ϵ b ) , and for the b + 1 th iteration, the new weight is
w b + 1 ( i ) = w b ( i ) exp { α b ξ b ( i ) }
The obtained weights are then normalized to the sum of one. As a result, the weight of incorrectly categorized observations increases while the weight of correctly classified observations reduces, driving the single classifier produced in the next iteration to focus on the most difficult examples. Furthermore, while the single classifier’s error is low, differences in weight updates are bigger, since when the classifier gets a high accuracy, the few mistakes become more important. Thus, the alpha constant can be thought of as a learning rate computed as a function of each iteration’s mistake. Additionally, this constant is employed in the final decision rule, which gives more weight to the individual classifiers with the lowest error. This process is repeated in each step for b = 1 , 2 , 3 , , B . Finally, the ensemble classifier calculates the weighted sum of each class’s votes. As a result, the class with the highest weighted vote receives the assignment. In particular [52,53],
C ( x ) = arg y j max b = 1 B α b δ C b ( x ) , y j
= arg y j max b : C b ( x ) = y j α b

3. Experimental Results and Discussions

3.1. Performance Metrics

In this study, to measure the proposed CAD diagnosis method’s effectiveness, various basic metrics, which are accuracy, precision, recall, F 1 , AUC, and MCC, are employed, and these metrics are computed as follows [54,55,56,57]:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
A U C = 1 2 T P T P + F N + T N T N + F P
M C C = T P · T N F P · F N ( T P + F P ) · ( T P + F N ) · ( T N + F P ) · ( T N + F N )
where the types of possible outcomes are TP (true positives—correctly labeled as positive tuples), TN (true negatives—correctly labeled as negative tuples), FP (false positives—negative tuples incorrectly labeled as positive), and FN (false negatives—positive tuples mislabeled as negative) in binary estimation [54,58]. The confusion matrix, a summary of the possible outcomes, is demonstrated in Figure 6.

3.2. Experiments on the Feature Extraction

The initial experiment focused on the feature extraction method employed in the proposed approach. Two distinct techniques were utilized: exclusive feature selection and PCA. The performance metrics results are presented in Table 3, utilizing various data-division methodologies, including an 80% training–20% test split, and 5-fold and 10-fold cross-validation. Notably, employing feature selection alone yielded a classification accuracy of 90.164% in the 80% training–20% test data division. Furthermore, precision, recall, F-measure, and AUC metrics exceeded 0.9, with a Matthews correlation coefficient (MCC) rate of 0.755. Subsequently, integrating PCA with the selected features led to improved accuracy, precision, recall, F-measure, and MCC metrics, resulting in an accuracy rate of 91.803%, as demonstrated in Table 3.
In the data-dividing methodology of 5-fold cross-validation, an accuracy of 86.799% was attained, and results in the range of 0.893 to 0.926 were obtained for precision, recall, F-measure, and AUC. When the PCA technique was utilized for feature extraction, some rise was observed in accuracy, precision, recall, F-measure, and MCC metrics. Finally, in 10-fold cross-validation, the use of the PCA technique increased by approximately three percent in the accuracy metric. Moreover, the results of precision, recall, F-measure and M,CC metrics rose. Considering all the results, the best classification accuracy rate of 91.803%, precision rate of 0.933, recall rate of 0.955, F-measure rate of 0.944, and MCC rate of 0.793 were achieved in 80% training–20% test splitting methodology when PCA was used. Additionally, the best AUC rate of 0.929 was achieved with the feature extraction technique of feature selection only.
In addition, the confusion matrices for each feature extraction technique and data-dividing methodologies are given in Figure 7. The confusion matrices acquired for the feature selection only are depicted in Figure 7a–c while Figure 7d–f demonstrate the confusion matrices attained for the PCA technique with the data-dividing methodologies of 80% training–20% test and 5-fold and 10-fold cross-validation, respectively.

3.3. Comparison with Traditional Methods

This subsection compares the classification results of the proposed method with basic classifiers. On the Z-Alizadeh Sani dataset, several basic techniques were tested in the 10-fold cross-validation. The aforementioned two extraction techniques of feature selection only and PCA were utilized with each basic classifier, and their performance results are shown with regard to the previously mentioned six metrics in Table 4. Along with the proposed approach, this table contains the results of eleven basic classifiers such as Naïve Bayes [59], k-NN (k = 5) [60], C4.5 decision tree [61], locally weighted learning (LWL) [62], K* [63], logistic model trees (LMT) [64], SVM [65], random forests [66], logistic regression [67], Hoeffding tree [68], and deep learning 4J [69]. As can be seen from the table, using PCA to extract the features increased the classification accuracy performance of k-NN, C4.5 decision tree, LWL, SVM, deep learning 4J and the proposed method. LMT and logistic regression with an accuracy of 88.449% are the best classifiers for feature selection only, whereas the proposed method with PCA achieves the best accuracy rate of 89.109%, a recall rate of 0.935, and an F-measure rate of 0.924, surpassing the other techniques.

3.4. Comparison with Existing Methods in the Literature

In this subsection, the proposed method was compared with the existing studies in the literature using the same dataset, the Z-Alizadeh Sani dataset. The performance comparison of the proposed method with the existing works is presented in Table 5, containing researcher names, years, used method, number of selected features, and accuracy metrics. Between 2012 and 2016, Alizadehsani et al. [19,20,21,22,23,24,25] used different numbers of features, such as 16, 20, 24, and 34, and achieved the best accuracy of 94.08% utilizing information gain and SMO. In 2017, Qin et al. [26] applied their CAD detection approach based on EA-MFS and SVM with 34 features and procured an accuracy rate of 93.70%. In the same year, Arabasadi et al. [27] proposed a genetic algorithm and MLP-ANN-based CAD prediction method selecting 22 features, while Babič et al. [28] performed various classifiers such as decision trees, Naïve Bayes, SVM, and ANN and used 27 features to feed these classifiers.
In 2018, Kılıç and Kaya Keleş [29] selected 16 features using the ABC algorithm and then classified CAD utilizing the SMO technique. As a result of their study, an accuracy rate of 89.44% was obtained. In 2019, MML-IBLMM and Var-IBLMM methods introduced by Hu et al. [30] were applied to the Z-Alizadeh Sani dataset and attained an accuracy rate of 81.84%. In the same year, Abdar et al. [31] proposed the N2Genetic-nuSVM approach, selected 29 features, and acquired an accuracy rate of 93.08%. In another work performed by Abdar et al. [32], a CAD diagnosis approach called NE-nu-SVC was presented. In this approach, 16 features were selected and an accuracy of 94.66% was achieved.
In 2020, Joloudari et al. [33] tested the classification performance of various classifiers, selected 40 features, and obtained the best accuracy rate of 91.47% with random trees. In the same year, a hybrid feature selection method called 2HFS was introduced by Nasarian et al. [34], and 38 features were selected using this method. In the sequel, SMOTE and XGBoost techniques were used together and an accuracy rate of 92.58% was reported. In another study presented by Ashish et al. [35], a random forests-, SVM-, and XGBoost-based CAD detection approach was implemented and an accuracy rate of 93.86% was achieved with 10 features. In a recent work [36], 25 features were used, and an accuracy rate of 91.78% with the MLP classifier was obtained.
Unlike these studies, the proposed method in this work utilizes five features, namely, age, hypertension, typical chest pain, t-wave inversion, and region with regional wall motion abnormality. In the dataset with these features, PCA and AdaBoostM1 techniques were used for feature extraction and classification, respectively. The best accuracy of 91.80% was achieved when using these few features on the Z-Alizadeh Sani dataset.

3.5. Limitations

This work can be considered a retrospective study because it uses a dataset based on past patient records. Researchers conduct this type of study by examining the existing records, historical data, or previous occurrences in order to determine outcomes, relationships, or correlations between variables. In contrast to prospective studies, which follow participants ahead of time, retrospective studies begin with the desired outcome or endpoint and go backwards to investigate the reasons or events that led to it. For example, in a prospective study introduced by Locuratolo et al. [70], patients were evaluated clinically and in the laboratory after 30 days, 3 months, 6 months, and 1 year following the index incident. Various endpoints related to acute coronary syndrome were evaluated. At the end of the study, the persistence of treatments and the percentage of patients who achieved therapeutic goals were evaluated.
Retrospective studies have some limitations, such as data quality, limited scope, bias, temporal ambiguity, confounding variables, validity of exposure measurement, and causality inference. In spite of these limitations, retrospective studies remain useful in epidemiological research, particularly when prospective investigations are unfeasible or unethical. The method proposed in this study can help medical doctors diagnose CAD early by using a small number of features.

4. Conclusions

This paper introduces an effective approach for diagnosing coronary artery disease (CAD) by leveraging age, hypertension, typical chest pain, T-wave inversion, and regional wall motion abnormality features. The method proposed utilizes eight distinct search techniques, including evolutionary, best first, genetic, harmony, PSO, greedy stepwise, rank, and multi-objective evolutionary search, to perform feature selection on the Z-Alizadeh Sani dataset. Principal component analysis (PCA) and the AdaBoostM1 algorithm are employed for feature extraction and CAD classification, respectively. Through extensive experiments and analyses using various performance metrics, the proposed method achieves the highest prediction performance to date with only five attributes. Notably, it achieves impressive accuracy, precision, recall, F-measure, AUC, and MCC rates of 91.8%, 93.3%, 95.5%, 94.4%, 89.5%, and 79.3%, respectively. These results demonstrate the efficiency of the proposed approach and its potential to serve as a cost-effective tool to aid medical practitioners in CAD diagnosis.

Author Contributions

Conceptualization, C.E. and O.K.; methodology, C.E.; software, C.E.; writing—original draft preparation, C.E. and O.K.; writing—review and editing, C.E. and O.K.; visualization, C.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by Cardiff University Institutional Funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This paper utilised the Z-Alizadeh Sani dataset that is freely available from the University of California—Irvine Machine Learning Repository [18].

Conflicts of Interest

The authors declare that they have no conflicts of interest. This research has not received any specific grant from public funding agencies or commercial or not-for-profit sectors.

References

  1. World Health Organization. Available online: https://www.who.int/ (accessed on 1 December 2023).
  2. International Diabetes Federation. Diabetes and Cardiovascular Disease. 2016. Available online: https://idf.org/our-activities/care-prevention/cardiovascular-disease.html (accessed on 1 December 2023).
  3. World Health Organization. Cardiovascular Diseases (CVDs). 2021. Available online: https://www.who.int/news-room/fact-sheets/detail/cardiovascular-diseases-(cvds) (accessed on 1 December 2023).
  4. World Health Organization. The Top 10 Causes of Death. 2020. Available online: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death (accessed on 1 December 2023).
  5. Eyupoglu, C. Breast cancer classification using k-nearest neighbors algorithm. Online J. Sci. Technol. 2018, 8, 29–34. [Google Scholar]
  6. Janosi, A.; Steinbrunn, W.; Pfisterer, M.; Detrano, R. Heart Disease Data Set, UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/datasets/Heart+Disease (accessed on 1 December 2023).
  7. Akgül, M.; Sönmez, Ö.E.; Özcan, T. Diagnosis of heart disease using an intelligent method: A hybrid ANN–GA approach. In Proceedings of the International Conference on Intelligent and Fuzzy Systems, Istanbul, Turkey, 21–23 July 2019; pp. 1250–1257. [Google Scholar]
  8. Rajab, W.; Rajab, S.; Sharma, V. Kernel FCM-based ANFIS approach to heart disease prediction. In Proceedings of the Emerging Trends in Expert Applications and Security, Jaipur, India, 17–19 February 2019; pp. 643–650. [Google Scholar]
  9. Uyar, K.; İlhan, A. Diagnosis of heart disease using genetic algorithm based trained recurrent fuzzy neural networks. Procedia Comput. Sci. 2017, 120, 588–593. [Google Scholar] [CrossRef]
  10. Haq, A.U.; Li, J.P.; Memon, M.H.; Nazir, S.; Sun, R. A hybrid intelligent system framework for the prediction of heart disease using machine learning algorithms. Mob. Inf. Syst. 2018, 2018, 3860146. [Google Scholar] [CrossRef]
  11. Ali, L.; Niamat, A.; Khan, J.A.; Golilarz, N.A.; Xingzhong, X.; Noor, A.; Bukhari, S.A.C. An optimized stacked support vector machines based expert system for the effective prediction of heart failure. IEEE Access 2019, 7, 54007–54014. [Google Scholar] [CrossRef]
  12. Burse, K.; Kirar, V.P.S.; Burse, A.; Burse, R. Various preprocessing methods for neural network based heart disease prediction. In Proceedings of the Smart Innovations in Communication and Computational Sciences; Springer: Singapore, 2019; pp. 55–65. [Google Scholar]
  13. Paul, A.K.; Shill, P.C.; Rabin, M.; Islam, R.; Murase, K. Adaptive weighted fuzzy rule-based system for the risk level assessment of heart disease. Appl. Intell. 2018, 48, 1739–1756. [Google Scholar] [CrossRef]
  14. Amin, M.S.; Chiam, Y.K.; Varathan, K.D. Identification of significant features and data mining techniques in predicting heart disease. Telemat. Inform. 2019, 36, 82–93. [Google Scholar] [CrossRef]
  15. Terrada, O.; Cherradi, B.; Raihani, A.; Bouattane, O. Classification and Prediction of atherosclerosis diseases using machine learning algorithms. In Proceedings of the 2019 5th International Conference on Optimization and Applications (ICOA), Kenitra, Morocco, 25–26 April 2019. [Google Scholar]
  16. Gokulnath, C.B.; Shantharajah, S.P. An optimized feature selection based on genetic approach and support vector machine for heart disease. Clust. Comput. 2019, 22, 14777–14787. [Google Scholar] [CrossRef]
  17. Karayılan, T.; Kılıç, Ö. Prediction of heart disease using neural network. In Proceedings of the 2017 International Conference on Computer Science and Engineering (UBMK), Antalya, Turkey, 5–7 October 2017; pp. 719–723. [Google Scholar]
  18. Alizadeh Sani, Z.; Alizadehsani, R.; Roshanzamir, M. Z-Alizadeh Sani Data Set, UCI Machine Learning Repository. 2017. Available online: https://archive.ics.uci.edu/ml/datasets/Z-Alizadeh+Sani (accessed on 1 December 2023).
  19. Alizadehsani, R.; Habibi, J.; Hosseini, M.J.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B.; Sani, Z.A. A data mining approach for diagnosis of coronary artery disease. Comput. Methods Programs Biomed. 2013, 111, 52–61. [Google Scholar] [CrossRef] [PubMed]
  20. Alizadehsani, R.; Habibi, J.; Sani, Z.A.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Alizadeh-Sani, F. Diagnosing coronary artery disease via data mining algorithms by considering laboratory and echocardiography features. Res. Cardiovasc. Med. 2013, 2, 133–139. [Google Scholar]
  21. Alizadehsani, R.; Hosseini, M.J.; Sani, Z.A.; Ghandeharioun, A.; Boghrati, R. Diagnosis of coronary artery disease using cost-sensitive algorithms. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining Workshops, Brussels, Belgium, 10–12 December 2012; pp. 9–16. [Google Scholar]
  22. Alizadehsani, R.; Habibi, J.; Sani, Z.A.; Mashayekhi, H.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B. Diagnosis of coronary artery disease using data mining based on lab data and echo features. J. Med. Bioeng. 2012, 1, 26–29. [Google Scholar] [CrossRef]
  23. Alizadehsani, R.; Zangooei, M.H.; Hosseini, M.J.; Habibi, J.; Khosravi, A.; Roshanzamir, M.; Khozeimeh, F.; Sarrafzadegan, N.; Nahavandi, S. Coronary artery disease detection using computational intelligence methods. Knowl. Based Syst. 2016, 109, 187–197. [Google Scholar] [CrossRef]
  24. Alizadehsani, R.; Hosseini, M.J.; Boghrati, R.; Ghandeharioun, A.; Khozeimeh, F.; Sani, Z.A. Exerting cost-sensitive and feature creation algorithms for coronary artery disease diagnosis. Int. J. Knowl. Discov. Bioinform. (IJKDB) 2012, 3, 59–79. [Google Scholar] [CrossRef]
  25. Alizadehsani, R.; Habibi, J.; Hosseini, M.J.; Boghrati, R.; Ghandeharioun, A.; Bahadorian, B.; Sani, Z.A. Diagnosis of coronary artery disease using data mining techniques based on symptoms and ecg features. Eur. J. Sci. Res. 2012, 82, 542–553. [Google Scholar]
  26. Qin, C.J.; Guan, Q.; Wang, X.P. Application of ensemble algorithm integrating multiple criteria feature selection in coronary heart disease detection. Biomed. Eng. Appl. Basis Commun. 2017, 29, 1750043. [Google Scholar] [CrossRef]
  27. Arabasadi, Z.; Alizadehsani, R.; Roshanzamir, M.; Moosaei, H.; Yarifard, A.A. Computer aided decision making for heart disease detection using hybrid neural network-Genetic algorithm. Comput. Methods Programs Biomed. 2017, 141, 19–26. [Google Scholar] [CrossRef]
  28. Babič, F.; Olejár, J.; Vantová, Z.; Paralič, J. Predictive and descriptive analysis for heart disease diagnosis. In Proceedings of the 2017 Federated Conference on Computer Science and Information Systems (fedCSIS), Prague, Czech Republic, 3–6 September 2017; pp. 155–163. [Google Scholar]
  29. Kılıc, Ü.; Kaya Keleş, M. Feature selection with artificial bee colony algorithm on Z-Alizadeh Sani dataset. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications Conference (ASYU), Adana, Turkey, 4–6 October 2018; pp. 1–3. [Google Scholar]
  30. Hu, C.; Fan, W.; Du, J.X.; Bouguila, N. A novel statistical approach for clustering positive data based on finite inverted Beta-Liouville mixture models. Neurocomputing 2019, 333, 110–123. [Google Scholar] [CrossRef]
  31. Abdar, M.; Książek, W.; Acharya, U.R.; Tan, R.S.; Makarenkov, V.; Pławiak, P. A new machine learning technique for an accurate diagnosis of coronary artery disease. Comput. Methods Programs Biomed. 2019, 179, 104992. [Google Scholar] [CrossRef] [PubMed]
  32. Abdar, M.; Acharya, U.R.; Sarrafzadegan, N.; Makarenkov, V. NE-nu-SVC: A new nested ensemble clinical decision support system for effective diagnosis of coronary artery disease. IEEE Access 2019, 7, 167605–167620. [Google Scholar] [CrossRef]
  33. Joloudari, J.H.; Hassannataj Joloudari, E.; Saadatfar, H.; Ghasemigol, M.; Razavi, S.M.; Mosavi, A.; Nadai, L. Coronary artery disease diagnosis; ranking the significant features using a random trees model. Int. J. Environ. Res. Public Health 2020, 17, 731. [Google Scholar] [CrossRef]
  34. Nasarian, E.; Abdar, M.; Fahami, M.A.; Alizadehsani, R.; Hussain, S.; Basiri, M.E.; Sarrafzadegan, N. Association between work-related features and coronary artery disease: A heterogeneous hybrid feature selection integrated with balancing approach. Pattern Recognit. Lett. 2020, 133, 33–40. [Google Scholar] [CrossRef]
  35. Ashish, L.; Kumar, S.; Yeligeti, S. Ischemic heart disease detection using support vector machine and extreme gradient boosting method. Mater. Today Proc. 2021; in press. [Google Scholar] [CrossRef]
  36. Kolukisa, B.; Bakir-Gungor, B. Ensemble feature selection and classification methods for machine learning-based coronary artery disease diagnosis. Comput. Stand. Interfaces 2023, 84, 103706. [Google Scholar] [CrossRef]
  37. Hall, M.A. Correlation-Based Feature Subset Selection for Machine Learning. Ph.D. Thesis, University of Waikato, Hamilton, New Zealand, 1998. [Google Scholar]
  38. Vikhar, P.A. Evolutionary algorithms: A critical review and its future prospects. In Proceedings of the 2016 International Conference on Global Trends in Signal Processing, Information Computing and Communication (ICGTSPICC), Jalgaon, India, 22–24 December 2016; pp. 261–265. [Google Scholar]
  39. Pearl, J. Heuristics: Intelligent Search Strategies for Computer Problem Solving; Addison-Wesley Longman Publishing Company: Boston, MA, USA, 1984. [Google Scholar]
  40. Goldberg, D.E. Genetic Algorithms in Search, Optimization and Machine Learning; Addison-Wesley Longman Publishing Company: Boston, MA, USA, 1989. [Google Scholar]
  41. Fong, S.; Biuk-Aghai, R.P.; Millham, R.C. Swarm search methods in weka for data mining. In Proceedings of the 2018 10th International Conference on Machine Learning and Computing, Macau, China, 26–28 February 2018; pp. 122–127. [Google Scholar]
  42. Moraglio, A.; Chio, C.D.; Poli, R. Geometric particle swarm optimisation. In Proceedings of the European Conference on Genetic Programming, Valencia, Spain, 11–13 April 2007; pp. 125–136. [Google Scholar]
  43. Butterworth, R.; Simovici, D.A.; Santos, G.S.; Ohno-Machado, L. A greedy algorithm for supervised discretization. J. Biomed. Inform. 2004, 37, 285–292. [Google Scholar] [CrossRef] [PubMed]
  44. Hall, M.A.; Holmes, G. Benchmarking attribute selection techniques for discrete class data mining. IEEE Trans. Knowl. Data Eng. 2003, 15, 1437–1447. [Google Scholar] [CrossRef]
  45. Jiménez, F.; Sánchez, G.; García, J.M.; Sciavicco, G.; Miralles, L. Multi-objective evolutionary feature selection for online sales forecasting. Neurocomputing 2017, 234, 75–92. [Google Scholar] [CrossRef]
  46. Statistics and Machine Learning Toolbox. 2018. Available online: https://www.mathworks.com/products/statistics.html (accessed on 1 December 2023).
  47. Salo, F.; Nassif, A.B.; Essex, A. Dimensionality reduction with IG-PCA and ensemble classifier for network intrusion detection. Comput. Netw. 2019, 148, 164–175. [Google Scholar] [CrossRef]
  48. Jackson, J.E. A User’s Guide to Principal Components; John Wiley & Sons: Hoboken, NJ, USA, 2005; Volume 587. [Google Scholar]
  49. Yavuz, E.; Eyupoglu, C. An effective approach for breast cancer diagnosis based on routine blood analysis features. Med. Biol. Eng. Comput. 2020, 58, 1583–1601. [Google Scholar] [CrossRef] [PubMed]
  50. Olson, D.L.; Delen, D. Advanced Data Mining Techniques; Springer Science & Business Media: New York, NY, USA, 2008. [Google Scholar]
  51. Eyüpoğlu, C. Büyük Veride Etkin Gizlilik Koruması Için Yazılım Tasarımı /Software Design for Efficient Privacy Preserving in Big Data. Ph.D. Thesis, İstanbul University, Istanbul, Turkey, 2018. [Google Scholar]
  52. Freund, Y.; Schapire, R.E. Experiments with a new boosting algorithm. In Proceedings of the 13th International Conference on Machine Learning, Bari Italy, 3–6 July 1996; pp. 148–156. [Google Scholar]
  53. Cortes, E.A.; Martinez, M.G.; Rubio, N.G. Multiclass corporate failure prediction by Adaboost. M1. Int. Adv. Econ. Res. 2007, 13, 301–312. [Google Scholar] [CrossRef]
  54. Eyupoglu, C.; Aydin, M.A.; Zaim, A.H.; Sertbas, A. An efficient big data anonymization algorithm based on chaos and perturbation techniques. Entropy 2018, 20, 373. [Google Scholar] [CrossRef]
  55. Eyüpoğlu, C. Kronik Böbrek Hastalığının Erken Tanısı için Yeni Bir Klinik Karar Destek Sistemi. Avrupa Bilim Teknol. Derg. 2020, 20, 448–455. [Google Scholar] [CrossRef]
  56. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  57. Eyüpoğlu, C. Korelasyon Temelli Özellik Seçimi, Genetik Arama ve Rastgele Ormanlar Tekniklerine Dayanan Yeni Bir Rahim Ağzı Kanseri Teşhis Yöntemi. Avrupa Bilim Teknol. Derg. 2020, 19, 263–271. [Google Scholar] [CrossRef]
  58. Han, J.; Kamber, M.; Pei, J. Data Mining Concepts and Techniques, 3rd ed.; Elsevier, Morgan Kaufmann Publishers: San Francisco, CA, USA, 2012. [Google Scholar]
  59. John, G.H.; Langley, P. Estimating continuous distributions in Bayesian classifiers. In Proceedings of the Eleventh conference on Uncertainty in Artificial Intelligence, Montreal, Canada, 18–20 August 1995; pp. 338–345. [Google Scholar]
  60. Aha, D.W.; Kibler, D.; Albert, M. Instance-based learning algorithms. Mach. Learn. 1991, 6, 37–66. [Google Scholar] [CrossRef]
  61. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann Publishers: San Mateo, CA, USA, 1993. [Google Scholar]
  62. Frank, E.; Hall, M.; Pfahringer, B. Locally Weighted Naive Bayes. In Proceedings of the 19th Conference in Uncertainty in Artificial Intelligence, Acapulco, Mexico, 7–10 August 2003; pp. 249–256. [Google Scholar]
  63. Cleary, J.G.; Trigg, L.E. K*: An instance-based learner using an entropic distance measure. In Proceedings of the 12th International Conference on Machine Learning, Tahoe City, CA, USA, 9–12 July 1995; pp. 108–114. [Google Scholar]
  64. Landwehr, N.; Hall, M.; Frank, E. Logistic model trees. Mach. Learn. 2005, 59, 161–205. [Google Scholar] [CrossRef]
  65. Keerthi, S.S.; Shevade, S.K.; Bhattacharyya, C.; Murthy, K.R. Improvements to Platt’s SMO algorithm for SVM classifier design. Neural Comput. 2001, 13, 637–649. [Google Scholar] [CrossRef]
  66. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  67. Le Cessie, S.; Van Houwelingen, H.C. Ridge estimators in logistic regression. J. R. Stat. Soc. Ser. Appl. Stat. 1992, 41, 191–201. [Google Scholar] [CrossRef]
  68. Hulten, G.; Spencer, L.; Domingos, P. Mining time-changing data streams. In Proceedings of the 7th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 26–29 August 2001; pp. 97–106. [Google Scholar]
  69. Deeplearning4j. Deep Learning for Java. Available online: https://deeplearning4j.org/ (accessed on 1 December 2023).
  70. Locuratolo, N.; Scicchitano, P.; Antoncecchi, E.; Basso, P.; Bonfantino, V.M.; Brescia, F.; Carrata, F.; De Martino, G.; Landriscina, R.; Lanzone, S.; et al. Follow-up of patients after an acute coronary event: The Apulia PONTE-SCA program. G. Ital. Cardiol. (2006) 2022, 23, 63–74. [Google Scholar]
Figure 1. The main types (rectangle at the centre) and risk factors (around the rectangle with arrows) of CVDs [2].
Figure 1. The main types (rectangle at the centre) and risk factors (around the rectangle with arrows) of CVDs [2].
Jcm 13 02868 g001
Figure 2. Top ten global causes of mortality in 2019 [4].
Figure 2. Top ten global causes of mortality in 2019 [4].
Jcm 13 02868 g002
Figure 3. Flowchart of the proposed CAD diagnosis method.
Figure 3. Flowchart of the proposed CAD diagnosis method.
Jcm 13 02868 g003
Figure 4. Variance values explained by each principal component.
Figure 4. Variance values explained by each principal component.
Jcm 13 02868 g004
Figure 5. Visualization of 10-fold cross-validation.
Figure 5. Visualization of 10-fold cross-validation.
Jcm 13 02868 g005
Figure 6. Confusion Matrix.
Figure 6. Confusion Matrix.
Jcm 13 02868 g006
Figure 7. Confusion matrices for each feature extraction technique and data-dividing methodology. Figures in (ac) refer to “Feature-selection-only” results for 80%-Training-20%Test, 5-fold-CV and 10-fold-CV, respectively. Similarly, (df) refer to the same results respectively but this time for the “PCA” method.
Figure 7. Confusion matrices for each feature extraction technique and data-dividing methodology. Figures in (ac) refer to “Feature-selection-only” results for 80%-Training-20%Test, 5-fold-CV and 10-fold-CV, respectively. Similarly, (df) refer to the same results respectively but this time for the “PCA” method.
Jcm 13 02868 g007
Table 1. The Overview of Z-Alizadeh Sani Dataset.
Table 1. The Overview of Z-Alizadeh Sani Dataset.
#Attribute NameAttribute CategoryAttribute Range
1AgeDemographic30–86
2WeightDemographic48–120
3LengthDemographic140–188
4SexDemographicFemale, Male
5Body Mass Index (BMI, Kg/m2)Demographic18.115–40.901
6Diabetes Mellitus (DM)DemographicYes, No
7Hypertension (HTN)DemographicYes, No
8Current SmokerDemographicYes, No
9Ex-SmokerDemographicYes, No
10Family History (FH)DemographicYes, No
11Obesity (BMI > 25)DemographicYes, No
12Chronic Renal Failure (CRF)DemographicYes, No
13Cerebrovascular Accident (CVA)DemographicYes, No
14Airway DiseaseDemographicYes, No
15Thyroid DiseaseDemographicYes, No
16Congestive Heart Failure (CHF)DemographicYes, No
17Dyslipidemia (DLP)DemographicYes, No
18Blood Pressure (BP, mmHg)Symptom and examination90–190
19Pulse Rate (PR, ppm)Symptom and examination50–110
20EdemaSymptom and examinationYes, No
21Weak Peripheral PulseSymptom and examinationYes, No
22Lung RalesSymptom and examinationYes, No
23Systolic MurmurSymptom and examinationYes, No
24Diastolic MurmurSymptom and examinationYes, No
25Typical Chest PainSymptom and examinationYes, No
26DyspneaSymptom and examinationYes, No
27Function ClassSymptom and examination0, 1, 2, 3
28AtypicalSymptom and examinationYes, No
29Nonanginal Chest PainSymptom and examinationYes, No
30Exertional Chest PainSymptom and examinationYes, No
31Low Threshold Angina (Low TH Ang)Symptom and examinationYes, No
32Q-WaveECGYes, No
33ST ElevationECGYes, No
34ST DepressionECGYes, No
35T-Wave InversionECGYes, No
36Left Ventricular Hypertrophy (LVH)ECGYes, No
37Poor R-Wave ProgressionECGYes, No
38Bundle Branch Block (BBB)ECGLeft, Right, Normal
39Fasting Blood Sugar (FBS, mg/dL)Laboratory and echo62–400
40Creatine (Cr, mg/dL)Laboratory and echo0.5–2.2
41Triglyceride (TG, mg/dL)Laboratory and echo37–1050
42Low Density Lipoprotein (LDL, mg/dl)Laboratory and echo18–232
43High Density Lipoprotein (HDL, mg/dL)Laboratory and echo15.9–111
44Blood Urea Nitrogen (BUN, mg/dL)Laboratory and echo6–52
45Erythrocyte Sedimentation Rate (ESR, mm/h)Laboratory and echo1–90
46Hemoglobin (HB, g/dL)Laboratory and echo8.9–17.6
47Potassium (K, mEq/lit)Laboratory and echo3–6.6
48Sodium (Na, mEq/lit)Laboratory and echo128–156
49White Blood Cell (WBC, cells/mL)Laboratory and echo3700–18,000
50Lymphocyte (%)Laboratory and echo7–60
51Neutrophil (%)Laboratory and echo32–89
52Platelet (PLT, 1000/mL)Laboratory and echo25–742
53Ejection Fraction (%)Laboratory and echo15–60
54Region-RWMALaboratory and echo0, 1, 2, 3, 4
55Valvular Heart Disease (VHD)Laboratory and echoMild, Severe, Moderate, Normal
Table 2. The Attributes Chosen Using Search Methods. Bold attribute numbers refer to the common features for all search methods.
Table 2. The Attributes Chosen Using Search Methods. Bold attribute numbers refer to the common features for all search methods.
Search MethodNumber of Chosen AttributesAttribute No.
Evolutionary171, 7, 9, 14, 15, 18, 24, 25, 28, 29, 31, 35, 39, 41, 45, 47, 54
Best first121, 6, 7, 18, 25, 28, 29, 35, 45, 47, 53, 54
Genetic151, 4, 6, 7, 12, 18, 25, 28, 29, 32, 34, 35, 47, 53, 54
Harmony171,7, 12, 13, 14, 15, 17, 25, 27, 28, 29, 35, 37, 45, 47, 53, 54
PSO141, 6, 7, 18, 25, 28, 29, 32, 34, 35, 45, 47, 53, 54
Greedy stepwise101, 6, 7, 14, 25,35, 45, 47, 53, 54
Rank131, 6, 7, 14, 25, 28, 29, 32, 33, 35, 45, 53, 54
Multi-objective evolutionary101, 6, 7, 14, 25, 35, 45, 47, 53, 54
Table 3. Performance Metric Results of Various Feature Extraction Techniques with AdaBoostM1.
Table 3. Performance Metric Results of Various Feature Extraction Techniques with AdaBoostM1.
Feature Extraction Technique80/20% Train/Test Split5-Fold CV10-Fold CV
Feature selection only%Acc.90.16486.79986.469
Precision0.9320.8930.900
Recall0.9320.9260.912
F10.9320.9090.906
AUC0.9290.9090.907
MCC0.7550.6700.666
PCA%Acc.91.80388.11989.109
Precision0.9330.9130.914
Recall0.9550.9210.935
F10.9440.9170.924
AUC0.8950.8880.879
MCC0.7930.7080.730
Table 4. Classification Performance Results Of Basic Classifiers. Bold-faced results refer to the best performing results for each metric.
Table 4. Classification Performance Results Of Basic Classifiers. Bold-faced results refer to the best performing results for each metric.
Feature Extraction Technique Naïve Bayes [59]k-NN [60]C4.5 DT [61]LWL [62]K* [63]LMT [64]SVM [65]RF [66]Log Reg [67]Hoeff. Tree [68]DL 4J [69]Ours
Feature selection only%Acc.88.12085.48085.15087.13083.50088.45087.79081.52088.45087.79085.48086.470
Preci.0.9050.9100.8900.9000.8770.9100.9160.8640.9100.9050.9390.900
Recall0.9310.8840.9030.9210.8940.9310.9120.8800.9310.9260.8520.912
F10.9180.8970.8970.9110.8850.9200.9140.8720.9200.9150.8930.906
AUC0.9230.8940.8300.9070.9010.9190.8530.8810.9220.9230.9220.907
MCC0.7050.6530.6340.6810.5920.7140.7030.5430.7140.6970.6760.666
PCA%Acc.80.53086.47087.79087.79081.85086.80088.12081.19088.45080.20086.80089.110
Preci.0.9200.9000.9120.9050.8710.8890.9170.8630.9100.9190.9440.914
Recall0.7960.9120.9170.9260.8750.9310.9170.8750.9310.7920.8660.935
F10.8540.9060.9150.9150.8730.9100.9170.8690.9200.8510.9030.924
AUC0.8920.8780.8460.6970.8580.9180.8550.8740.9220.8920.9210.879
MCC0.5810.6660.7010.8850.5550.6680.7100.5360.7140.5750.7030.730
Table 5. Performance Comparison of The Proposed Method with The Existing Studies using the Z-Alizadeh Sani Dataset.
Table 5. Performance Comparison of The Proposed Method with The Existing Studies using the Z-Alizadeh Sani Dataset.
PaperYearMethod# of FeaturesAccuracy (%)
[22]2012SMO1682.16
[25]2012Naïve Bayes–SMO1688.52
[21]2012SMO3492.09
[24]2012SMO 1-13492.74
[19]2013Information gain + SMO3494.08
[20]2013Bagging2079.54 (LAD)
+61.46 (LCX)
C4.568.96 (RCA)
[23]2016Average and combined2486.14 (LAD)
information gain83.17 (LCX)
+ SVM83.50 (RCA)
[26]2017EA-MFS + SVM3493.7
[27]2017GA + MLP-ANN2293.85
[28]2017SVM2786.67
[29]2018ABC + SMO1689.44
[30]2019MML-IBLMM and Var-IBLMM5581.84
[31]2019N2Genetic-nuSVM2993.08
[32]2019NE-nu-SVC1694.66
[33]2020Random trees4091.47
[34]20202HFS + SMOTE + XGBoost3892.58
[35]2021Random forests + SVM + XGBoost1093.86
[36]2023MLP2591.78
Proposed PCA + AdaBoostM1591.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Eyupoglu, C.; Karakuş, O. Novel CAD Diagnosis Method Based on Search, PCA, and AdaBoostM1 Techniques. J. Clin. Med. 2024, 13, 2868. https://doi.org/10.3390/jcm13102868

AMA Style

Eyupoglu C, Karakuş O. Novel CAD Diagnosis Method Based on Search, PCA, and AdaBoostM1 Techniques. Journal of Clinical Medicine. 2024; 13(10):2868. https://doi.org/10.3390/jcm13102868

Chicago/Turabian Style

Eyupoglu, Can, and Oktay Karakuş. 2024. "Novel CAD Diagnosis Method Based on Search, PCA, and AdaBoostM1 Techniques" Journal of Clinical Medicine 13, no. 10: 2868. https://doi.org/10.3390/jcm13102868

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop