Next Article in Journal
Neutrino Signals in Electron-Capture Storage-Ring Experiments
Next Article in Special Issue
Exact and Numerical Solutions of a Spatially-Distributed Mathematical Model for Fluid and Solute Transport in Peritoneal Dialysis
Previous Article in Journal
Coherent States of Harmonic and Reversed Harmonic Oscillator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Data Mining Approach for Cardiovascular Disease Diagnosis Using Heart Rate Variability and Images of Carotid Arteries

1
College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
2
Department of Computer Engineering, Dongguk University Kyeongju Campus, Gyeongju 38066, Korea
3
Bigsun Systems Co., Ltd., Seoul 06266, Korea
*
Author to whom correspondence should be addressed.
Symmetry 2016, 8(6), 47; https://doi.org/10.3390/sym8060047
Submission received: 20 April 2016 / Revised: 31 May 2016 / Accepted: 6 June 2016 / Published: 13 June 2016
(This article belongs to the Special Issue Symmetry in Systems Design and Analysis)

Abstract

:
In this paper, we proposed not only an extraction methodology of multiple feature vectors from ultrasound images for carotid arteries (CAs) and heart rate variability (HRV) of electrocardiogram signal, but also a suitable and reliable prediction model useful in the diagnosis of cardiovascular disease (CVD). For inventing the multiple feature vectors, we extract a candidate feature vector through image processing and measurement of the thickness of carotid intima-media (IMT). As a complementary way, the linear and/or nonlinear feature vectors are also extracted from HRV, a main index for cardiac disorder. The significance of the multiple feature vectors is tested with several machine learning methods, namely Neural Networks, Support Vector Machine (SVM), Classification based on Multiple Association Rule (CMAR), Decision tree induction and Bayesian classifier. As a result, multiple feature vectors extracted from both CAs and HRV (CA+HRV) showed higher accuracy than the separative feature vectors of CAs and HRV. Furthermore, the SVM and CMAR showed about 89.51% and 89.46%, respectively, in terms of diagnosing accuracy rate after evaluating the diagnosis or prediction methods using the finally chosen multiple feature vectors. Therefore, the multiple feature vectors devised in this paper can be effective diagnostic indicators of CVD. In addition, the feature vector analysis and prediction techniques are expected to be helpful tools in the decisions of cardiologists.

1. Introduction

According to the recent World Health Organization (WHO)’s report about the main causes of death, the top two causes are still cardiovascular diseases (CVD) [1]. In South Korea, CVD is ranked second in causes of death which turns the country into a demographical structure of high incidence of the disease [2]. Consequently, early diagnosis and the reliability of the diagnosis has been recognized as a very important social issue. The current method of diagnosis for CVD at a hospital includes echocardiography cardiac ultrasound, electrocardiogram (ECG) inspection, the magneto cardiogram (MEG) and the coronary angiography inspection. However, the majority of inspections are invasive and unreliable [3,4].
Nowadays, early diagnosis of CVD has been realized after the introduction of a method measuring carotid arterial intima-media thickness by ultrasound that can prescreen the CVD. The thickness of the common carotid artery (CCA) has been identified to be related with CVD in various studies and has become one of the typical cardiovascular risk factors together with hypertension of blood, hyperlipidemia, smoking and diabetes mellitus. It is also known as an independent predictor of CVD [5,6,7].
The correlation between the autonomic nervous system and mortality of CVD including sudden cardiac death has been proved as a significant factor during the past 30 years. The development of indicators that can evaluate quantitatively the activity of the autonomic nervous system is urgently required, and heart rate variability (HRV) has been one of the most promising indicators. The wide variety of linear and nonlinear characteristics of HRV have been studied as indicators to improve the diagnostic accuracy. Dynamic stability of the cardiovascular system is achieved by the heart rate’s quick reactions and automatically adjusting to internal or external stimuli [8,9,10].
Heart rate changes are complexly reacted to these stimuli and are stimulated intensively by the two systems: the sympathetic nervous system and the parasympathetic nervous system. The activation of the sympathetic nervous system slows the heart rate and the activation of the parasympathetic nervous system increases the speed of the heart beat with the growth of contractility. By this difference, the two systems of the autonomic nervous systems operate on different frequencies, and it allows us to know whether the variability of heart rate changes is dominantly related to the sympathetic nervous system or the parasympathetic one [11]. The level of sympathetic and parasympathetic nerve activity can be evaluated quantitatively through linear and/or nonlinear feature analysis. For instance, if we analyze the variability of the heart rates of patients with coronary artery disease (CAD), the regulatory role of the autonomic nervous system is reduced, and the risk of death in the case of acute myocardial infarction is reduced when the autonomic nervous system is actively interact. However, there is a problem with not directly introducing the developed algorithms and feature vectors standardized for western patients because it is reportedly known that feature vectors may cause different diagnosis results due to racial pathological and physiological deviations.
Therefore, the carotid artery (CA) and HRV diagnostic feature vectors need to be analyzed to ensure the reliability and early diagnosis for CVD of South Koreans. In order to analyze diagnostic feature vectors for South Korean CVD patients, we proposed an extraction methodology of multiple feature vectors from CA and HRV. The details of the proposed methods and how to perform the steps for the diagnosis of CVD are as follows:
(1)
Extracting diagnostic feature vectors: the feature vectors significant to disease diagnosis are extracted by applying image processing to the CA images taken by ultrasound;
(2)
Evaluation on feature vector and classification method for diagnosis of CVD: some diagnostic feature vectors that are significant by types of CVD through statistical analysis of the data should be selected as a preprocessing step. Classification or prediction algorithm is applied to the selected diagnostic feature vectors for CVD, and the vectors were evaluated.
For effective understanding of the paper, the paper is organized as follows. CA imaging and HRV analysis through complex diagnostic feature vector extraction process will be explained in Section 2 and Section 3, respectively. In Section 4, a feature vector selection process as pre-processing steps and experimental evaluations results using classification of forecasting techniques for disease diagnosis will be described. Finally, concluding remarks will be shown in Section 5.

2. Carotid Artery Scanning and Image Processing

The CA consists of common carotid artery (CCA), carotid bifurcation (BIF), internal carotid artery (ICA), and external carotid artery (ECA). The intima-media thickness (IMT) of the carotid can be measured at the far wall CCA region 10 mm proximal to bifurcation of carotid rather than the ICA or CA or BIF itself (see Figure 1). Intima is the high-density band-shaped and the media looks like a band with a low brightness between intima and adventitia. Adventitia generally has the brightest pixel value and it corresponds to the thick part below the intima-media having the high brightness. In addition, since the intima is thinnest among the three floors and its brightness is so similar to that of media, the endometrial thickness is difficult to detect. Thus, In general, It is sufficient to measure IMT including the intima and media.
The Common Carotid Arterial scanning using a high-resolution ultrasound system can acquire the image by scanning the right side ICA longitudinally at R-peak of the electrocardiogram. At first, we load the ultrasound image of target common carotid arterial scanning from computer hard-disc memory in order to measure the carotid artery intima-media. Next, the calibration factor of pixel length is determined using electronic range caliper in a B-mode ultrasound system. After we select at least the 10 mm–long image of the Region of Interest (ROI) picture at 10 mm proximal around the area of BIF transition to CCA, we can evaluate the quality of the selected ROI image and remove speckle noise. After obtaining the edge image by applying the edge detection algorithm, IMT is measured [12].
After the acquisition of carotid image and IMT measurement, all of the diagnostic feature vectors for CVDs are extracted. The feature vector extraction will be performed in the following eight steps [13]:
(1)
The ROI image with 64 × 100 pixels is acquired by defining the area of two ’+’ markers (from a to b ) on the image of the carotid IMT in Figure 2a;
(2)
Each pixel is expressed by a number in the range of 0– 255 ( 2 8 ) for the brightness (Figure 2b);
(3)
The trend of variation is shown in a graph in a vertical line (Figure 2c);
(4)
Thirty vertical lines are randomly selected as samples among a total of 100 vertical lines (Figure 2d);
(5)
The difference between V 1 and V 2 is calculated using the 30 random samples of vertical lines;
(6)
Only IMT ( V 1 - V 2 ) values within one sigma in Gaussian distribution are extracted;
(7)
Four basic feature vectors are extracted and an average value is calculated;
(8)
The other 18 additional feature vectors are extracted through a calculation using the four basic feature vectors in Figure 2d, and the mean value is obtained.
All the feature vectors extracted from carotid image and IMT measurement are described in Table 1.

3. Linear and Non-Linear Feature Vectors of HRV

Extracting the linear and non-linear indicators of HRV, the main diagnostic indices for CVDs such as angina pectoris or acute coronary syndrome, starts from ECG. The ECG signal is measured during five minutes using a Lead II channel. The sampling frequency obtained from ECG signals with such measurements show 500 Hz, and ectopic beats and artifacts are removed. HRV is the physiological phenomenon of variation in the time interval between heartbeats. It is measured by the variation in the beat-to-beat interval. Other terms used are RR variability. where R is a point corresponding to the peak of the QRS complex which is the name for the combination of three of the graphical deflections seen on a typical ECG. RR is the interval between successive Rs. To analyze HRV, all RR intervals of the ECG signal are calculated by Thomkin’s algorithm [14], and time-series data is generated as shown in Figure 3. RR interval times-series data are re-sampled at a rate of 4 Hz in order to extract the indicators in the frequency domain, which is one of the linear analysis methods. We extract linear feature vectors in the time and frequency domain and extract non-linear feature vectors of HRV. The literature on HRV feature vector extraction was described in detail in [15].

3.1. Linear Feature Vectors in Time Domain

Time domain statistics are generally calculated on RR intervals without re-sampling. In a continuous ECG record, each QRS complex was detected and the RR intervals or the instantaneous heart rates were determined. RR intervals is denoted by R R n , with n = 0 , . . . , N . For practical purposes, the following basic properties are true: E [ R R n ] = E [ R R n + m ] and E [ R R n 2 ] = E [ R R n + m 2 ] . The standard time domain measures of HRV are as follows [16].
The standard deviation of the RR intervals (SDRR) is often employed as a measure of overall HRV. It is defined as the square root of the variance of the RR intervals as follows:
S D R R = E [ R R n 2 ] - R R ¯ 2 ,
where the mean of RR interval is denoted by R R ¯ = E [ R R n ] . The standard deviation of the successive differences of the RR intervals (SDSD) is an important measure of short-term HRV. It is defined as the square root of the variance of the sequence Δ R R n = R R n - R R n + 1 (the delta-RR intervals):
S D S D = E [ Δ R R n 2 ] - Δ R R n ¯ 2 .
Where, Δ R R n ¯ = E [ R R n ] - E [ R R n + 1 ] = 0 for stationary intervals. This means that the root mean square (rms) of the successive differences is statistically equivalent to the standard deviation of the successive differences as follows:
S D S D = r m s S D = E [ ( R R n - R R n + 1 ) 2 ] .
Linear feature vectors in time domains include the mean of RR intervals ( R R ¯ ), the standard deviation of the RR Intervals (SDRR), and the standard deviation of successive differences of the RR intervals (SDSD).

3.2. Linear Feature Vectors in Frequency Domain

The feature vectors in frequency mode use power spectral density (PSD) analysis and extract seven types of feature vectors as follows [13,15]:
(1)
Total power ( T P ), from 0 Hz to 0.4 Hz;
(2)
Very Low Frequency ( V L F ) power, from 0 Hz to 0.04 Hz;
(3)
Low Frequency ( L F ) power, from 0.04 Hz to 0.15 Hz;
(4)
High Frequency ( H F ) power, from 0.15 Hz to 0.4 Hz;
(5)
Normalized value of L F ( n L F = ( T P - V L F ) L F × 100 );
(6)
Normalized value of H F ( n H F = ( T P - V L F ) H F × 100 );
(7)
The ratio of L F and H F ( L F / H F ).
Diagnostic feature vector employs only three vectors; n L F to reflect sympathetic activity, n H F to show parasympathetic activity and L F / H F ratio to mirror the sympathovagal balance (see Table 2).

3.3. Poincare Plot of Nonlinear Feature Vectors

The Poincare plot may be analyzed quantitatively by fitting an ellipse to the plotted shape. The center of the ellipse is determined by the average RR intervals. S D 1 refers to the standard deviation of the distances of points from the y = x axis, S D 2 stands for the standard deviation of the distances of points from y = - x + 2 R R ¯ axis, where R R ¯ is the mean of RR intervals as shown in Figure 4. We also compute the features, S D 2 / S D 1 , and S D 1 · S D 2 , describing the relationship between S D 1 and S D 2 in our study.

3.4. A Non-Linear Vector: Approximate Entropy (ApEn)

Defined as the rate of information production, entropy quantifies the chaos of motion. A p E n quantifies the regularity of time series, and is also called a "regularity statistic". It is represented as a simple index for the overall complexity and predictability of each time series. In this study, A p E n quantifies the regularity of the RR intervals. The more regular and predictable the RRI series, the lower will be the value of A p E n . First of all, we reconstructed the RRI time series in the n-dimensional phase space using Takens’ theorem [17]. Takens suggested the following time delay method for the reconstruction of the state space as follows:
D t = [ R R ( t ) , R R ( t + τ ) , . . . , R R ( t + ( n - 1 ) τ ) ] ,
where n is the embedding dimension and τ is the time delay. In this study, the optimal value of τ was 10. The mean of the fraction of patterns with length m that resembles the pattern with the same length beginnings at interval i is defined by
Φ m ( r ) = 1 N - m + 1 i = 1 N - m + 1 ln number of D m ( j ) - D m ( i ) < r N - m - 1 .
In the equation above, D m ( i ) and D m ( j ) are state vectors in the embedding dimension m. Given N data points, we can define A p E n as A p E n ( m , r , N ) = ϕ m ( r ) - Φ m + 1 ( r ) , where A p E n estimates the logarithmic likelihood that the next intervals after each of the patterns will differ. In general, the embedding dimension m, and the tolerance, r are fixed at m = 2 and r = 0 . 2 × S D in physiological time series data.

3.5. Hurst Exponent (H) Non-Linear Vector)

The Hurst Exponent H is the measure of the smoothness of a fractal time series based on the asymptotic behavior of the rescaled range of the process. H is defined as, l o g ( R / S ) l o g ( T ) , where T is the duration of the sample of data and R / S is the corresponding value of the rescaled range. If H = 0 . 5 , the behavior of the time series is similar to a random walk. If H < 0 . 5 , the time series covers less distance than a random walk. If H > 0 . 5 , the time series covers more distance than a random walk.

3.6. Exponent α of the 1 / f Spectrum ( f α ) Non-Linear Vector

Self-similarity is the most distinctive property of fractal signals. Fractal signals usually have a power spectrum of the inverse power law form, 1 / f α , where f is frequency, and the amplitude of the fluctuations is small at high frequencies and large at low frequencies. The exponent α is calculated by a first least-squares fit in a l o g - l o g spectrum, after finding the power spectrum from RR intervals. The exponent is clinically significant because it has different values for healthy and heart rate failure patients.
All feature vectors extracted from linear and non-linear analysis of HRV are described in Table 2.

4. Evaluation of Diagnostic Feature Vectors

All the data used in our experiment were provided as a sample by the Bio-signal Research Center of the Korea Research Institute of Standards and Science. In this experiment, after coronary arteriography was performed for each of the 214 cardiovascular patients, patients showing more than 50% of stenosis are categorized as Coronary Artery Disease (CAD), whereas other patients having less than 50% stenosis are designated as the control group. Furthermore, CAD patients are also re-sorted by cardiologists into two groups, Angina Pectoris (AP) and Acute Coronary Syndrome (ACS). Clinical characteristics of the studied patients are shown in Table 3.

4.1. Data Preprocessing

The extracted vectors from carotid imaging and HRV are evaluated in order to determine whether those vectors can be representative indicators of cardiovascular diseases or not by applying typical classification or prediction models of machine learning.
As a pre-processing step, feature selection method is used for eliminating the improper information to disease diagnosis. The performing steps are composed of feature ranking and feature selection steps. Selection algorithms evaluate the redundancy in feature vectors and prediction capability of each vector. Feature ranking considers one feature at a time to see how well each feature alone predicts the target class. The features are ranked according to a user-defined criterion. Available criteria depend on the measurement levels of the target class and feature. In the feature vector selection problem, a ranking criterion is used to find feature vectors that discriminate between healthy and diseased patients. The ranking value of each feature is calculated as ( 1 - p ), where p is the p-value of appropriate statistical tests of association between the candidate features and the target class. All diagnostic feature vectors are continuous-valued, and we use p-values based on F-statistics. This method is to perform a one-way ANOVA F-test [18] for each continuous feature.
Let C denote a target class with J categories, N be a total number of cases and X is the feature under consideration with I categories. The p-value based on F-statistics is calculated by;
Pr ( F ( J - 1 , N - J ) > F ) , F = j = 1 J N j ( x j ¯ - x ˜ ) 2 ( J - 1 ) j = 1 J ( N j - 1 ) s j 2 ( N - 1 ) ,
where N j is the number of cases with C = j , x j ¯ is the mean of feature X for target class C = j , s j 2 is the sample variance of feature X for class C = j , x ˜ is the grand mean of feature X and F ( J - 1 , N - J ) is a random variable follows a F-distribution with degrees of freedom J - 1 and N - J . If the denominator for a feature is zero, set the p-value = 0 for the feature. Features are ranked by p-value in ascending order. In this study, any p-value less than 0.1 significant test threshold was accepted as significant. A feature relevance score ( 1 - p ) is calculated. The features having values less than 0.1 indicate that they have low score and therefore they are removed. Afterwards, this subset of features is presented as input to the classification methods. We perform feature selection only once for each dataset and then different classification methods are evaluated. The results of feature selection and evaluation for each dataset (CA, HRV and CA+HRV) are described in Table 4.

4.2. Verification of Feature Vectors Using Classification Methods

In order to determine whether the combined 20 feature vectors extracted from both CA and HRV (indicated as CA+HRV) can be effective diagnostic indicators of CVDs than feature vectors of CA and HRV separately, the famous classification or prediction method of machine learning is used as the way of evaluation. The classification method generates and compares several models including Neural Networks (NNs), Bayesian classifiers, decision tree induction model, Support Vector Machine (SVM) and Classification Based on Multiple Association Rules (CMAR).

4.2.1. Neural Networks

The NNs method uses back propagation to classify instances [19]. NNs are composed of nodes (neurons) and their interconnections. In general, input values are converted into values ranging from zero to one. Each input node is connected to an output node through the link with weight value. We use the back-propagation multi-layer perceptron (MLP) consisting of three layers: input, hidden, and output layers. An NNs learns through changes in the weight of each node, and its goal is to determine the weight w that minimizes the sum of the squared error between target class y and predicted class y ^ , which is calculated using in the following equation below:
E ( w ) = 1 2 i = 1 N ( y i - y ^ ) 2 .

4.2.2. Bayesian Network

The Bayesian Network chooses the highest posterior probability class using the prior probability computed from the training data set. The Naïve Bayes classifier assumes that the effect of an attribute on a given class is independent of the values of the other attributes. This assumption is called class conditional independence. However, attribute values of CA and HRV data may not be entirely independent from each other. In order to address this problem, we considered a set of extended Bayesian classifiers known to work well with correlated data, including Tree Augmented Naïve Bayes (TAN) [20].

4.2.3. Decision Tree Induction (C4.5)

C4.5 [21] is one of the most popular decision trees. A decision tree can be viewed as a partitioning of the instance space. Each partition, called a leaf, represents a number of similar instances that belong to the same class. C4.5 is also a decision tree generating algorithm based on the ID3 algorithm. It contains several improvements especially needed for software implementation. These improvements contain: (1) choosing an appropriate attribute selection measure; (2) handling training data with missing attribute values; (3) handling attributes with differing costs; and (4) handling continuous attributes. In order to perform experiments, we used the j48.part method that implemented the C4.5 algorithm [19].

4.2.4. Support Vector Machine (SVM)

The SVM is basically a two-class classifier and can be extended for multi-class classification. The main reason for interest in support vector and kernel methods is their flexibility and remarkable resistance to overfitting, and their simplicity and theoretical elegance, all appealing to practitioners as well as theoreticians. In our model, each object is mapped to a point in a high dimensional space, each dimension corresponding to features. The coordinates of the point are the frequencies of the features in the corresponding dimensions. The SVM learns, in the training step, the maximum-margin hyper-planes separating each class. In the testing step, it classifies a new object by mapping it to a point in the same high-dimensional space divided by the hyper-plane learned in the training step. For our experiments, we applied the sequential minimal optimization (SMO) algorithm by using the radial basis function (RBF) kernel for training a support vector classifier [19].

4.2.5. Classification Based on Multiple Association Rules (CMAR)

CMAR classifiers [22] extend CBA by using more than one rule to classify a new case. It is a two-step process: (1) rule generation and (2) classification. In rule generation, CMAR uses an approache based on the FP-growth method [23] to find the complete set of rules satisfying the minimum confidence and minimum support thresholds. FP-growth uses a tree structure called FP-tree, which retains the item set association information contained in the given data set D. CMAR uses an enhanced FP-tree that maintains the distribution of various class labels among tuples satisfying frequent item sets. From the FP-tree, created rules can be generated immediately. Thus, CMAR allows rule generation and frequent item set mining in a single step. Once a rule is generated, it is stored in a CR-tree. The CR-tree is a prefix tree data structure. Its function is to store and retrieve rules efficiently and prune rules based on confidence, correlation and database coverage. Whenever a rule is inserted into the CR2-tree, it starts a pruning rule. Highly specialized rules with low confidence can be pruned if more generalized rules with higher confidence exist. CMAR also prunes rules based on χ 2 and database coverage [22,24].
In our experiment, four classifiers, except for the CMAR classifier, utilize the following source code provided by the Java WEKA project (University of Waikato, Hamiton, New Zealand) [19]. The CMAR classifier utilizes CMAR software provided by the LUCS-KDD group (University of Liverpool, Liverpool, England) [25].
  • weka.classifiers.bayes.BayesNet (TAN)
  • weka.classifiers.tree.j48.J48 (C4.5)
  • weka.classifiers.funtions.SMO (SVM)
  • weka.classifiers.functions.MultilayerPerceptron (MLP)
Through the statistical analysis of all the diagnostic feature vectors listed in Table 1 and Table 2, we apply each classification model to the data set that passed the feature selection step. We build the above classifiers from the preprocessed training data. Accuracy was obtained by using the methodology of stratified 10-fold cross-validation (CV-10) for three classes. To evaluate classification performance with respect to the number of instances and class labels, we used a confusion matrix. We also used P r e c i s i o n , R e c a l l , F- m e a s u r e and A c c u r a c y to evaluate the classifiers’ performance for analyzing our training sets with imbalanced class distribution. Formal definitions of these measures are given in the equations below [15,26,27]:
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N F - m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l A c c u r a c y = T P + T N T P + F P + T N + F N ,
where, T P is True Positive, T N is True Negative, F P is False Positive, and F N is False Negative in the confusion matrix. In the performance evaluation with P r e c i s i o n , R e c a l l and F - m e a s u r e , Table 5 shows the features used for each data set which are CA, HRV and CA+HRV (combining CA and HRV).
The parameters of the five classification methods were set as follows. For the CMAR algorithm, the minimum support was set to 0.4%, the minimum confidence to 70%, and the database coverage was set to 3.75 (critical threshold for a 5% “significance” level, assuming degree of freedom is equivalent to one). For the SVM, the soft margin allowed errors during training. We set 0.1 for the two-norm soft margin value. The Neural Networks (MLP), Bayesian classifier (TAN) and C4.5 parameters were default values.
In addition, summarizing classifiers’ performance with a single number would make it more convenient to compare the performance of diffenent classifiers. This can be done using a performance metric such as accuracy and Root Mean Squared Error (RMSE). The results of RMSE and the accuracy for each classifier are shown in Table 6 and Figure 5, respectively.
According to the result, shown in Figure 5, the highest accuracy for CA (16 features) and HRV (12 features) is 82.94% and 78.82%, respectively. However, both accuracies are lower than the highest accuracy rate, 89.51%, obtained from various features of CA+HRV (20 features), considering CA and HRV separately. Therefore, we anticipated that the use of the multi-features for CA+HRV will produce more effective diagnostic indicators than using features of CA and HRV. In order to address this expectation, we used the confusion matrix of classification results for each feature in Table 7. As SVM and CMAR give higher accuracy in comparison to other methods, we apply these two best classifiers in the following experiment. Table 7 records the hit rates (correctly classified instance rate) of SVM and CMAR in a confusion matrix, and the results are reasonably good. The hit rates of SVM for AP, Control and ACS classes are 85.42%, 92.55% and 66.39%. The hit rates of CMAR are also 94.51%, 84.51% and 69.23%, respectively. In particular, ACS class can not be correctly separated from AP class when we used features for each CA and HRV (see Table 7). These results indicate that using the multiple features of CA+HRV gives more effective results in discriminating between the groups.

5. Conclusions

This paper suggests multiple diagnostic feature vectors with the CA and HRV analyses for the purpose of more accurate prediction and early diagnosis of CVD, recently growing at a rapid speed. Moreover, we performed experiments and evaluations to verify the reliability of the prediction system and test the significance of diagnostic feature vectors. According to the results of experiments, 20 types of feature vectors are determined as the essential elements for disease diagnosis and SVM and CMAR show an excellent result in terms of the appropriate classification or prediction algorithm. These kind of complex diagnostic indicators would be useful for the automatic diagnosis of CVDs in Korea. The limitation of this paper is in the fact that the sample data provided by a certain organization without collecting sufficient data from the domestic hospitals. Data accumulation of various ages and genders is required to secure high reliability of the developed systems and play a reference database role in this disease area.

Acknowledgments

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT & Future Planning (No.2013R1A2A2A01068923).

Author Contributions

All authors discussed the contents of the manuscript and contributed to its preparation. Hyeongsoo Kim and Musa Ibrahim M. Ishag collected the data and analyzed the data; Minghao Piao carried out the experiments; Taeil Kwon designed and implemented the prototype for the extraction of the feature vectors; Hyeongsoo Kim wrote the paper; Keun Ho Ryu provided critical insight and discussion. All authors read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. The Top 10 Causes of Death. WHO Report. Available online: http://www.who.int/mediacentre/factsheets/fs310/en/ (accessed on 1 April 2015).
  2. Korea National Statistical Office. Statistics Report of Causes of Death. Available online: http://kosis.kr/statHtml/statHtml.do?orgId=101&tblId=DT_1B34E02&conn_path=I2 (accessed on 1 April 2015).
  3. Ryu, K.S.; Park, H.W.; Park, S.H.; Shon, H.S.; Ryu, K.H.; Lee, D.G.; Bashir, M.E.; Lee, J.H.; Kim, S.M.; Lee, S.Y.; et al. Comparison of clinical outcomes between culprit vessel only and multivessel percutaneous coronary intervention for ST-segment elevation myocardial infarction patients with multivessel coronary diseases. J. Geriatr. Cardiol. 2015, 12, 208. [Google Scholar] [PubMed]
  4. Lee, D.G.; Ryu, K.S.; Bashir, M.; Bae, J.W.; Ryu, K.H. Discovering medical knowledge using association rule mining in young adults with acute myocardial infarction. J. Med. Syst. 2013, 37, 1–10. [Google Scholar] [CrossRef] [PubMed]
  5. Bae, J.H.; Seung, K.B.; Jung, H.O.; Kim, K.Y.; Yoo, K.D.; Kim, C.M.; Cho, S.W.; Cho, S.K.; Kim, Y.K.; Rhee, M.Y.; et al. Analysis of Korean carotid intima-media thickness in Korean healthy subjects and patients with risk factors: Korea multi-center epidemiological study. Korean Circ. J. 2005, 35, 513–524. [Google Scholar] [CrossRef]
  6. Cheng, K.S.; Mikhailidis, D.P.; Hamilton, G.; Seifalian, A.M. A review of the carotid and femoral intima-media thickness as an indicator of the presence of peripheral vascular disease and cardiovascular risk factors. Cardiovasc. Res. 2002, 54, 528–538. [Google Scholar] [CrossRef]
  7. Nambi, V.; Chambless, L.; He, M.; Folsom, A.R.; Mosley, T.; Boerwinkle, E.; Ballantyne, C.M. Common carotid artery intima–media thickness is as good as carotid intima–media thickness of all carotid artery segments in improving prediction of coronary heart disease risk in the Atherosclerosis Risk in Communities (ARIC) study. Eur. Heart J. 2012, 33, 183–190. [Google Scholar] [CrossRef] [PubMed]
  8. Tarvainen, M.P.; Niskanen, J.P.; Lipponen, J.A.; Ranta-Aho, P.O.; Karjalainen, P.A. Kubios HRV—Heart Rate Variability Analysis Software. Comp. Method. Progr. Biomed. 2014, 113, 210–220. [Google Scholar] [CrossRef] [PubMed]
  9. dos Santos, L.; Barroso, J.J.; Macau, E.E.; de Godoy, M.F. Application of an automatic adaptive filter for heart rate variability analysis. Med. Eng. Phys. 2013, 35, 1778–1785. [Google Scholar] [CrossRef] [PubMed]
  10. ChuDuc, H.; NguyenPhan, K.; NguyenViet, D. A review of heart rate variability and its applications. APCBEE Proced. 2013, 7, 80–85. [Google Scholar] [CrossRef]
  11. Pumprla, J.; Howorka, K.; Groves, D.; Chester, M.; Nolan, J. Functional assessment of heart rate variability: Physiological basis and practical applications. Int. J. Cardiol. 2002, 84, 1–14. [Google Scholar] [CrossRef]
  12. Bae, J.H.; Kim, W.S.; Rihal, C.S.; Lerman, A. Individual Measurement and Significance of Carotid Intima, Media, and Intima–Media Thickness by B-Mode Ultrasonographic Image Processing. Arterioscler. Thromb. Vasc. Biol. 2006, 26, 2380–2385. [Google Scholar] [CrossRef] [PubMed]
  13. Piao, M.; Lee, H.G.; Pok, C.; Ryu, K.H. A data mining approach for dyslipidemia disease prediction using carotid arterial feature vectors. In Proceedings of the 2010 2nd International Conference on Computer Engineering and Technology (ICCET), Chengdu, China, 16–18 April 2010; Volume 2, pp. 171–175.
  14. Tompkins, W.J. Biomedical Digital Signal Processing: C Language Examples and Laboratory Experiments for the IBM PC; Prentice Hall, Inc.: Upper Saddle River, NJ, USA, 1993. [Google Scholar]
  15. Lee, H.G.; Kim, W.S.; Noh, K.Y.; Shin, J.H.; Yun, U.; Ryu, K.H. Coronary artery disease prediction method using linear and nonlinear feature of heart rate variability in three recumbent postures. Inf. Syst. Front. 2009, 11, 419–431. [Google Scholar] [CrossRef]
  16. Brennan, M.; Palaniswami, M.; Kamen, P. Do existing measures of Poincare plot geometry reflect nonlinear features of heart rate variability? IEEE Trans. Biomed. Eng. 2001, 48, 1342–1347. [Google Scholar] [CrossRef] [PubMed]
  17. Takens, F. Detecting strange attractors in turbulence. In Dynamical Systems and Turbulence, Warwick 1980; Lecture Notes in Mathematics; Springer Verlag: Berlin, Germany, 1981; Volume 898, pp. 366–381. [Google Scholar]
  18. Bhanot, G.; Alexe, G.; Venkataraghavan, B.; Levine, A.J. A robust meta-classification strategy for cancer detection from MS data. Proteomics 2006, 6, 592–604. [Google Scholar] [CrossRef] [PubMed]
  19. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed.; Morgan Kaufmann Publishers: San Francisco, CA, USA, 2005. [Google Scholar]
  20. Cheng, J.; Greiner, R. Comparing Bayesian network classifiers. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, Stockholm, Sweden, 30 July–1 August 1999; pp. 101–108.
  21. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann Publishers: San Mateo, CA, USA, 1993. [Google Scholar]
  22. Li, W.; Han, J.; Pei, J. CMAR: Accurate and efficient classification based on multiple class-association rules. In Proceedings of the IEEE International Conference on Data Mining 2001 (ICDM 2001), San Jose, CA, USA, 29 November–2 December 2001; pp. 369–376.
  23. Han, J.; Pei, J.; Yin, Y.; Mao, R. Mining frequent patterns without candidate generation: A frequent-pattern tree approach. Data Min. Knowl. Discov. 2004, 8, 53–87. [Google Scholar] [CrossRef]
  24. Lairenjam, B.; Wasan, S.K. Neural network with classification based on multiple association rule for classifying mammographic data. In Intelligent Data Engineering and Automated Learning-IDEAL 2009; Springer Berlin Heidelberg: Berlin, Heidelberg, Germany, 2009; pp. 465–476. [Google Scholar]
  25. Coenen, F. The LUCS-KDD Software Library. 2004. Available online: http://cgi.csc.liv.ac.uk/%7efrans/KDD/Software/ (accessed on 6 June 2016).
  26. Piao, Y.; Piao, M.; Jin, C.H.; Shon, H.S.; Chung, J.M.; Hwang, B.; Ryu, K.H. A New Ensemble Method with Feature Space Partitioning for High-Dimensional Data Classification. Math. Probl. Eng. 2015. [Google Scholar] [CrossRef]
  27. Lee, H.G.; Choi, Y.H.; Jung, H.; Shin, Y.H. Subspace Projection–Based Clustering and Temporal ACRs Mining on MapReduce for Direct Marketing Service. ETRI J. 2015, 37, 317–327. [Google Scholar] [CrossRef]
Figure 1. The measurement of intima-media thickness in the carotid artery using ultrasonograph. IMT: intima-media thickness, CCA: common carotid artery, BIF: bifurcation, ICA: internal carotid artery.
Figure 1. The measurement of intima-media thickness in the carotid artery using ultrasonograph. IMT: intima-media thickness, CCA: common carotid artery, BIF: bifurcation, ICA: internal carotid artery.
Symmetry 08 00047 g001
Figure 2. Carotid artery image processing and feature vector extraction step. (a) acquisition of carotid image and IMT measurement; (b) the acquired Region of Interest (ROI) image ( 64 × 100 pixels); (c) graph for the trend of variation of each vertical line; and (d) computation of four basic feature vectors (points).
Figure 2. Carotid artery image processing and feature vector extraction step. (a) acquisition of carotid image and IMT measurement; (b) the acquired Region of Interest (ROI) image ( 64 × 100 pixels); (c) graph for the trend of variation of each vertical line; and (d) computation of four basic feature vectors (points).
Symmetry 08 00047 g002aSymmetry 08 00047 g002b
Figure 3. RR interval extraction process in electrocardiogram (ECG) signals.
Figure 3. RR interval extraction process in electrocardiogram (ECG) signals.
Symmetry 08 00047 g003
Figure 4. Diagnostic indicators in a Poincare plot.
Figure 4. Diagnostic indicators in a Poincare plot.
Symmetry 08 00047 g004
Figure 5. Accuracy comparison.
Figure 5. Accuracy comparison.
Symmetry 08 00047 g005
Table 1. All the extracted vectors of carotid arteries.
Table 1. All the extracted vectors of carotid arteries.
Feature vectorIndexDescription
Carotid basic feature V 1 Starting point of intima
V 2 Starting point of adventitia
V 3 Max. point between V 1 and V 2
V 4 Min. point between V 1 and V 2
Carotid calculated feature V 5 Distance between V 1 and V 2
V 6 Area of the vector V 5
V 7 Value of the point V 3
V 8 Distance between V 1 and V 3
V 9 Area of the vector V 8
V 10 Value of the point V 4
V 11 Distance between V 1 and V 4
V 12 Area of the vector V 11
V 13 Slope between V 1 and V 3
V 14 Slope between V 3 and V 4
V 15 Slope between V 1 and V 2
V 16 V 3 - V 1
V 17 V 3 - V 4
V 18 Standard deviation between V 1 and V 4
V 19 Variance between V 1 and V 4
V 20 Skewness between V 1 and V 4
V 21 Kurtosis between V 1 and V 4
V 22 Moment between V 1 and V 4
IMT V 23 Intima-media thickness
Table 2. Diagnostic indicators from heart rate variability (HRV) analysis.
Table 2. Diagnostic indicators from heart rate variability (HRV) analysis.
Feature VectorDescription
Linear features Frequency domain n L F Normalized low frequency power ( n L F = ( T P - V L F ) L F × 100 ).
n H F Normalized high frequency power ( n H F = ( T P - V L F ) H F × 100 ).
L F / H F The ratio of low- and high-frequency power.
Time domain R R ¯ The mean of RR intervals.
SDRRStandard deviation of the RR intervals.
SDSDStandard deviation of the successive differences RR intervals.
Nonlinear features S D 1 Standard deviation of the distance of R R ( i ) from the line y = x in the Poincare
S D 2 Standard deviation of the distance of R R ( i ) from the line y = - x + 2 R R ¯ in the Poincare
S D 2 / S D 1 The ratio S D 2 to S D 1
S D 1 · S D 2 S D 1 × S D 2
A p E n Approximate Entropy
HHurst Exponent
f α 1/f scaling of Fourier spectra
Table 3. Clinical characteristics of the study subjects.
Table 3. Clinical characteristics of the study subjects.
GroupNSex (Male/Female)Age (Years)
AP10250/5259.98±8.41
Control7240/4656.70±9.23
ACS4018/2258.94±8.68
* AP: Angina Pectoris; ACS: Acute Coronary Syndrome.
Table 4. Selected feature vectors of carotid artery (CA), HRV and CA+HRV.
Table 4. Selected feature vectors of carotid artery (CA), HRV and CA+HRV.
RankCAHRVCA+HRV
FeatureRS( 1 - p )FeatureRS( 1 - p )FeatureRS( 1 - p )
1 V 3 1.000 S D 2 0.999 V 3 1.000
2 V 10 1.000 S D R R 0.998 V 10 1.000
3 V 18 0.998 f α 0.993 S D 2 0.998
4 V 2 0.997 S D 2 / S D 1 0.991 S D R R 0.997
5 V 8 0.997 S D 1 0.986 f α 0.986
6 V 21 0.995H0.984 S D 2 / S D 1 0.985
7 V 23 0.989 S D 1 × S D 2 0.975 S D 1 0.979
8 V 20 0.989 n L F 0.968 V 2 0.979
9 V 4 0.975 n H F 0.967 V 18 0.965
10 V 11 0.968 A p E n 0.961 S D 1 × S D 2 0.965
11 V 6 0.968 R R ¯ 0.958H0.965
12 V 16 0.967 L F / H F 0.955 V 8 0.963
13 V 13 0.966 V 21 0.962
14 V 19 0.962 V 23 0.962
15 V 5 0.961 n L F 0.960
16 V 17 0.953 n H F 0.958
17 A p E n 0.955
18 V 20 0.954
19 V 11 0.952
20 V 16 0.951
* CA: Carotid Artery, HRV: Heart Rate Variability, RS: Relevance score.
Table 5. A description of summary results (all features).
Table 5. A description of summary results (all features).
ClassifierCAHRVCA+HRVClass
(Using 16 Features)(Using 12 Features)(Using 20 Features)
PrecisionRecall F 1 PrecisionRecall F 1 PrecisionRecall F 1
NNs ( MLP ) 0.7010.7540.7270.6810.7490.7130.7630.9130.831AP
0.7070.7140.7110.6960.7130.7040.8350.7490.790Control
0.5190.4090.4570.4800.3360.3950.8330.5680.675ACS
BayesNet ( TAN ) 0.6270.7820.6960.5890.7490.6590.6600.8710.751AP
0.6690.5410.5980.6320.5320.5780.7680.5530.643Control
0.5950.4250.4960.5010.2970.3730.7250.4990.591ACS
C4.50.6560.7160.6850.6690.7270.6970.7340.8700.796AP
0.7220.7060.7140.7270.7110.7190.8460.7420.790Control
0.4880.3940.4360.4630.3780.4160.6450.4820.552ACS
SVM0.7560.8100.7820.6850.8040.7400.8720.8540.863AP
0.7950.7350.7640.8030.7450.7730.8640.9260.894Control
0.6210.5920.6060.3760.2580.3050.7180.6640.690ACS
CMAR0.7190.8140.7640.6170.8180.7030.8390.9450.889AP
0.6690.7690.7160.7020.7740.7360.8360.8450.840Control
0.6940.4620.5540.5420.2350.3280.8550.6920.765ACS
* F 1 : F- m e a s u r e ; NNs: Neural networks; MLP: Multi-layer perceptron; BayesNet: Bayesian network; TAN: Tree augmented naïve bayes; C4.5: Decision tree induction; SVM: Support vector machine; CMAR: Classification based on multiple association rules.
Table 6. A comparison of the classifiers’ root mean squared error (RMSE).
Table 6. A comparison of the classifiers’ root mean squared error (RMSE).
ClassifierCA (Using 16 Features)HRV (Using 12 Features)CA+HRV (Using 20 Features)
NNs (MLP)0.4050.4420.293
BayesNet (TAN)0.4430.5090.395
C4.50.4220.4260.342
SVM0.3010.4370.216
CMAR0.3550.4720.201
Table 7. Confusion matrix for SVM and CMAR (use of each feature for CA, HRV and CA+HRV).
Table 7. Confusion matrix for SVM and CMAR (use of each feature for CA, HRV and CA+HRV).
FeaturesClassifierActual ClassPredicted Class
AP (%)Control (%)ACS (%)
CA ( 16 features ) SVMAP81.016.2912.7
Control24.473.512.09
ACS22.6618.1259.22
CMARAP81.377.8410.79
Control18.0476.895.07
ACS26.9226.9246.16
HRV ( 12 features ) SVMAP80.46.0513.55
Control20.8574.524.63
ACS56.6617.525.84
CMARAP81.827.2810.9
Control18.1377.434.44
ACS53.9422.5723.49
CA + HRV ( 20 features ) SVMAP85.424.69.98
Control7.192.550.35
ACS19.0614.5566.39
CMARAP94.511.573.92
Control9.2784.516.22
ACS16.3914.3869.23

Share and Cite

MDPI and ACS Style

Kim, H.; Ishag, M.I.M.; Piao, M.; Kwon, T.; Ryu, K.H. A Data Mining Approach for Cardiovascular Disease Diagnosis Using Heart Rate Variability and Images of Carotid Arteries. Symmetry 2016, 8, 47. https://doi.org/10.3390/sym8060047

AMA Style

Kim H, Ishag MIM, Piao M, Kwon T, Ryu KH. A Data Mining Approach for Cardiovascular Disease Diagnosis Using Heart Rate Variability and Images of Carotid Arteries. Symmetry. 2016; 8(6):47. https://doi.org/10.3390/sym8060047

Chicago/Turabian Style

Kim, Hyeongsoo, Musa Ibrahim M. Ishag, Minghao Piao, Taeil Kwon, and Keun Ho Ryu. 2016. "A Data Mining Approach for Cardiovascular Disease Diagnosis Using Heart Rate Variability and Images of Carotid Arteries" Symmetry 8, no. 6: 47. https://doi.org/10.3390/sym8060047

APA Style

Kim, H., Ishag, M. I. M., Piao, M., Kwon, T., & Ryu, K. H. (2016). A Data Mining Approach for Cardiovascular Disease Diagnosis Using Heart Rate Variability and Images of Carotid Arteries. Symmetry, 8(6), 47. https://doi.org/10.3390/sym8060047

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop