Next Article in Journal
A Pre-Trained Fuzzy Reinforcement Learning Method for the Pursuing Satellite in a One-to-One Game in Space
Next Article in Special Issue
EEG Signal Analysis for Diagnosing Neurological Disorders Using Discrete Wavelet Transform and Intelligent Techniques
Previous Article in Journal
A Strong Tracking Mixed-Degree Cubature Kalman Filter Method and Its Application in a Quadruped Robot
Previous Article in Special Issue
Emphasis Learning, Features Repetition in Width Instead of Length to Improve Classification Performance: Case Study—Alzheimer’s Disease Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Arrhythmia Diagnosis by Using Level-Crossing ECG Sampling and Sub-Bands Features Extraction for Mobile Healthcare

by
Saeed Mian Qaisar
1,* and
Syed Fawad Hussain
2
1
College of Engineering, Effat University, Jeddah 22332, Saudi Arabia
2
Machine Learning and Data Science Lab, Ghulam Ishaq Khan Institute of Engineering Sciences and Technology, Topi 23460, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(8), 2252; https://doi.org/10.3390/s20082252
Submission received: 3 February 2020 / Revised: 24 March 2020 / Accepted: 1 April 2020 / Published: 16 April 2020
(This article belongs to the Special Issue Biomedical Signal Processing for Disease Diagnosis)

Abstract

:
Mobile healthcare is an emerging technique for clinical applications. It is usually based on cloud-connected biomedical implants. In this context, a novel solution is presented for the detection of arrhythmia by using electrocardiogram (ECG) signals. The aim is to achieve an effective solution by using real-time compression, efficient signal processing, and data transmission. The system utilizes level-crossing-based ECG signal sampling, adaptive-rate denoising, and wavelet-based sub-band decomposition. Statistical features are extracted from the sub-bands and used for automated arrhythmia classification. The performance of the system was studied by using five classes of arrhythmia, obtained from the MIT-BIH dataset. Experimental results showed a three-fold decrease in the number of collected samples compared to conventional counterparts. This resulted in a significant reduction of the computational cost of the post denoising, features extraction, and classification. Moreover, a seven-fold reduction was achieved in the amount of data that needed to be transmitted to the cloud. This resulted in a notable reduction in the transmitter power consumption, bandwidth usage, and cloud application processing load. Finally, the performance of the system was also assessed in terms of the arrhythmia classification, achieving an accuracy of 97%.

1. Introduction

An electrocardiogram (ECG) signal possesses critical information about cardiac functionality [1]. An abnormality of cardiac rhythm is a sign of certain diseases that can be diagnosed by an effective analysis of an ECG [2]. Heart diseases are one of the major threats to human life [3], and a timely diagnosis can lead to better measures. To do this, individual heartbeats are analyzed by exploiting their frequency content and morphological patterns for the automatic diagnosis of ECG arrhythmia [4,5,6,7,8]. Interferences and physiological artifacts can modify the ECG signal, thus decreasing the effectiveness of the automatic diagnosis mechanism. To overcome these deficiencies, numerous signal processing techniques have been used including eigenvalue composition [9], extended Kalman filtering [10], and Fourier decomposition [11]. The denoised ECG signals are processed and analyzed to extract classifiable features that can assist in automatic diagnosis. Some broadly used features extraction techniques are wavelet packet decomposition (WPD) [4], wavelet-based kernel principle component analysis (wkPCA) [4], wavelet packet entropy (WPE) [5], discrete wavelet transform (DWT) [6,12,13], temporal and morphological features [12], principle component analysis (PCA) [13,14], bispectrum [14], and Hermite function coefficient and temporal features [15].
The extracted features are employed for the automatic diagnosis of cardiac diseases. Supervised machine learning techniques are quite popular in these applications. In this regard, numerous approaches have been presented for ECG signal classification including those of the back propagation neural network (BNN) [4], random forest (RF) [5], probabilistic neural network (PNN) [6], support vector machine (SVM) [12], support vector machine (SVM)-radial basis function (RBF) [13,14], and optimized block-based neural network (OBNN) [15].
A timely detection of arrhythmia conditions allows for effective cardiac failure treatment. Therefore, patients with cardiac problems require continuous monitoring. In this context, wearable ECG sensors have been used [16,17]. These are usually linked to the cloud, and their data are logged and further processed via cloud-based applications. The decision support outcome is then shared with a monitoring center [17,18,19,20]. Such an approach allows healthcare specialists to make timely decisions and offer emergency care for patients with chronic disorders.
The design of ECG wearable devices is challenging because of strict constraints on size, weight, and power consumption. Self-powered wireless sensors are preferred because they enable the acquisition of ECG signals without causing too much uneasiness to the patient. A precise diagnosis requires the fine-grained recording and analysis of multichannel ECG signals. Consequently, the data dimension is exponentially augmented [21]. The transmission, analysis, and storage of such a big amount of data are not efficient. The substantial saving of power in a wireless ECG device is only possible by minimizing the activity of acquisition, processing, and wireless data transmission. In this context, studies have been carried out on ECG signal compression [20,21,22], non-uniform sampling [16,23,24], and adaptive-rate processing [25].
Conventional ECG processing solutions are based on the Nyquist sampling theory. Systems are time-invariant, which results in a worst-case parameterization [25,26,27,28,29]. A system’s computational load and processing activity remain fixed regardless of the incoming ECG signal time-varying nature. As a result, the system is constrained. It captures and processes redundant samples that increase the system’s overall computational load, power consumption, processing, and transmission activities [27,29]. These shortfalls can be compensated for to a certain extent by using level-crossing analog-to-digital converters (LCADCs) [16,24,30,31,32,33,34,35]. These converters adapt the system acquisition and processing rates according to the incoming signal’s temporal variations, a process that renders a significant computational efficiency of the suggested approach compared to traditional ones. In [27,29], Qaisar et al. reported more than one order of magnitude gain in terms of computational outperformance compared to the classical counterparts.
This work contributes to the development of efficient automatic arrhythmia diagnosis in the mobile healthcare framework. The aim is to realize an effective solution by achieving real-time compression and efficient ECG signal processing and transmission. In a continuation of previous works, such as [16,24,28,30,31], level-crossing sampling and adaptive-rate processing techniques are proposed in this paper. The realization of such is achieved by using an efficient assembly of the LCADC, robust classifiers, adaptive-rate signal processing, and features extraction modules. The use of level-crossing sampling and adaptive-rate processing for arrhythmia detection is an original concept thatis explored in this paper.

2. Materials and Methods

The proposed framework is shown in Figure 1, where the dotted blue-colored border (‘….’) encloses blocks embedded in the ECG real-time wearable processing device. The cloud processing module is enclosed in a green-colored dashed line (‘---’).

2.1. Dataset

The performance of the designed solution was studied by using the MIT-BIH arrhythmia database (ECG arrhythmia time series are available under (https://physionet.org/content/mitdb/1.0.0/)) [36]. A set of twelve 30-min ECG recordings containing clinically important arrhythmias was employed. Each channel was band-limited between [0.5, 60] Hz by using an analog antialiasing filter and was then acquired with an 11-bit resolution analog-to-digital converter (ADC). The sampling rate was 360 Hz. Experienced cardiologists labeled all acquired heartbeats. To extract individual heartbeats, each considered ECG signal wassegmented for a time length of 0.9 s. Five different arrhythmias classes were considered: normal signals (N), right bundle branch block (RBBB), left bundle branch block (LBBB), atrial premature contraction (APC) and premature ventricular contraction (PVC). To achieve equal representation, 300 instances were considered for each class. Therefore, in total, there were 1500 instances, each belonging to one of the five considered classes. To avoid any bias, heartbeats related to each class were collected from various records (clear from Table 1).

2.2. Level-Crossing A/D Converter (LCADC)

To evaluate the LCADC, the sampled version of a band-limited signal, given by y ( t n ) and obtained from the MIT-BIH dataset, was reconstructed. If x ˜ ( t ) is the reconstructed quasi-analog version of y ( t n ) , then the relationship between x ˜ ( t ) and y ( t n ) can be mathematically presented by using Equation (1), where U is the up-sampling factor. The up-sampling is realized by using the standard cascaded cubic spline interpolators and anti-imaging filters [37].
x ˜ ( t ) = y ( t n U ) .
The band-limited signal x ˜ ( t ) is digitized with a LCADC. The frequency content of x ˜ ( t ) is limited to [0.5, 60] Hz [9,36]. The LCADC is designed on the basis of level-crossing sampling (LCS) [28,38]. In this case, a sample is only acquired when x ˜ ( t ) crosses one of the predefined thresholds. For a given LCADC amplitude dynamic, Δ V , and resolution, M , the sampling frequency is piloted by the signal. Samples are irregularly spaced in time, and the count of samples is proportional to the slope of x ˜ ( t ) [31]. Sample amplitudes are equal to the predefined thresholds. The sampling instants are defined by Equation (2), and the process is shown in Figure 2 [28,38], where t n is the present sampling instant, t n 1 , is the preceding one and the time step between the present and the preceding sampling instants is d t n . LCADC parameters are selected according to the approach described in [35]. This approach is based on the uniform-quantization scheme. Therefore, its quantum, q , can be calculated as: q = Δ V 2 M 1 [38].
t n = t n 1 + d t n .
The phenomenon of hysteresis is also embedded in the LCADC [29]. A new sample is only acquired when there is a difference of q with respect to the preceding sample amplitude. The process is shown in Figure 2 and can be mathematically expressed as: x n = x n 1 ± q . It improves LCADC efficiency in terms of real-time compression [29]. The QRS complexes of heartbeats contain the most significant arrhythmia-related information [16,24]. The LCADC acquires the relevant ECG information, QRS complexes, at adaptive-rates while avoiding the remaining low amplitude components [24]. Therefore, it collects a reduced number of samples while comparing them with the classical counterparts [24].
The working principle of an LCADC is different from conventional ADCs [38]. While considering the ideal case, for conventional ADCs, the sampling instants are accurately known. However, the amplitudes of samples are quantized [38]. Quantization is the only source of error, and it depends on the selected Δ V and M [35]. It is assessed in terms of the signal-to-noise ratio (SNR) [38]. The SNR is computable as: S N R d B = 6.02 M + 1.76 . It presents an ideal ADC SNR for a full-scale monotonous sinusoidal input and describes its dependency on M . On the contrary, the amplitudes of samples are ideally known for an LCADC. However, the instants of these samples are quantized according to the operating frequency, F T i m e r , of the timer circuit thatis used to record these instants [28]. The SNR of an ideal LCADC can be calculated by using Equation (3) [38], where f s i g is the frequency of the full-scale sinusoid used to evaluate the LCADC. Equation (3) shows that the SNR of an ideal LCADC is independent of M and is a function of f s i g and T T i m e r = 1 F T i m e r . A 6.02 dB improvement in the value of the SNR is achievable by halving Ttimer [28]. In this study, a 21-bit resolution timer wasused with F T i m e r = 1 MHz. These parameters allowed us to properly record a heartbeat without timer overflow and resulted in an ideal LCADC SNR of 73.25 dB. According to [4,5,6,12,13,14,15], an11-bit ADC resolution is appropriate and results in a precise, computer-based arrhythmia diagnosis. For the selected timer parameters, the obtained LCADC SNR wasequal to the theoretical SNR of an 11.9-bit classical ADC. This justified the selected system parameters for the considered application.
S N R d B = 11.19 20 l o g ( f s i g · T t i m e r )

2.3. Activity Selection Algorithm (ASA)

The activity selection algorithm (ASA) segments LCADC output [27,28]. Thisalgorithmemploys sampling process non-uniformity for selecting the active parts of the signal while avoiding the redundant baseline [16,24,28]. The principle is clear from the algorithmic state machine (ASM) chart shown in Figure 3. In Figure 3, T 0 = 1 f m i n is the fundamental period of x ˜ ( t ) , and f m i n is the lowest frequency component whose value is equal to 0.5 Hz [9,36]. The level-crossing concept based sampled signal active parts are identified by using values of T 0 and d t n . The condition d t n T 0 is selected to respect the Nyquist criterion for f m i n . L i is the length of the ith selected segment W i . N i is the number of samples for W i . L r e f is the superior bound on L i , and its choice depends on the system parameters and the characteristics of the intended signal [28,38]. For this study L r e f = 1-swas selected. At the beginning of each iteration, ‘i’ is incremented, and N i and L i are initialized to zero.
The traditional windowing functions do not provide interesting features of the ASA [27,29]. The ASA allows for the selection of the signal active portions while avoiding the redundant, unwanted ones [16,24,28]. In addition, the length of the window-function is automatically adjusted according to the temporal variations of the signal. This process avoids signal truncation, and, therefore, segmentation can be performed by using adaptive length rectangular windows. This avoids the use of arithmetically complex smoothening-window functions and creates an effective solution of the spectral leakage phenomenon [27].

2.4. Adaptive-Rate Resampling

For a given resolution M, the LCADC sampling frequency organizes by following the temporal variations of x ˜ ( t ) . The maximum sampling frequency, F s m a x , of a uniform quantization-based LCADC is defined by Equation (4) [28,38], where fmax is the x ˜ ( t ) bandwidth. Ain is the input signal amplitude.
F s m a x = 2 · f m a x ( 2 M 2 ) · A i n Δ V
The sampling frequency for W i can be calculated as: F s i = N i L i . To benefit from the established signal processing techniques, W i is uniformly resampled by using simplified linear interpolation (SLI) [29], which modifies the resampled signal compared to the original; this variation depends on M , q , and the employed interpolator [39]. For SLI, the superior bound of error per resampled observation is q 2 [39].
The LCADC focuses on the active signal parts. Nevertheless, one LCADC shortfall is that the active signal parts can be digitized at superior rates compared to conventional digitization approaches [27,29]. The ASA overcomes this shortfall by examining features of W i and thenadjusting the system parameters accordingly [29]. In this way, the resampling frequency, F r s i , and the arithmetic complexity of the post-processing modules are adjusted by following the x ˜ ( t ) temporal disparities [29].

2.5. Adaptive-Rate Denoising

A band-pass finite impulse response (FIR) filters bank is designed for the effective online diminishing of unwanted noise from ECG signals [9]. The FIR filtering process can be mathematically described by using Equation (5), where x n is the incoming signal, x f n is the filtered signal, and h k are the coefficients of the K t h order FIR filter.
x f n = k = 0 K 1 h k × x n k .
Here, the filters bank was designed for the cut-off frequencies of [ Fc min , Fc max ] Hz. Each filter was implemented for a different sampling frequency that was chosen from the set F r e f (cf. Equation (6)). The upper bound on Fref was selected as Fr and, to assure a proper digital filtering operation, the lower bound on Fref was chosen as Fsmin ≥ 2·FCmax [29]. Fr was the sampling frequency such that its value remained greater than and closer to FNyq = 2·FCmax. Q is the length of Fref, and its value is always chosen as a binary-weighted. In Equation (6), Δ is a unique offset and can be computed as: Δ= F r F s m i n Q 1 .
F r e f = { F s min , F s min + Δ , , F s min + ( Q 1 ) Δ = F r }
The ASA examines the properties of W i and uses them for adjusting the denoising parameters such as the resampling frequency, F r s i , and the filter order, K i . The method of choosing F r s i and keeping it aligned with F r e f C is shown in Figure 4, which shows that a suitable filter, from the reference bank, is selected for each W i . Let h c k be the selected filter for W i that is sampled at F r e f C . Then, this selection can be made on the basis of F r e f and F s i . For proper denoising, F r s i = F r e f C must bechosen [29].

2.6. Features Extraction

2.6.1. Adaptive-Rate Discrete Wavelet Transform (ARDWT)

The wavelet transform (WT) is frequently used for the multi-resolution time-frequency analysis of non-stationary ECG like signals [4,40]. This transform can be mathematically expressed by Equation (7), where s and u, respectively, represent the dilation and the translation parameters.
W x ψ ( u , s ) = 1 S + x ( t ) ψ ( ( t u ) s ) d t
A discrete-time wavelet transform (DWT) is used for analyzing digital signals. A translation-dilation representation is attained by employing digital filters. In this case, a denoised segment, W i , is passed through the Daubechies algorithm-based wavelet decomposition, which consists of half-band high-pass and low-pass filters. This allows for the computation of approximation, a m i , and detail, d m i , coefficients at each level of decomposition. The mathematical processes of computing a m i and d m i are, respectively, expressed by Equations (8) and (9), where m represents the level of decomposition. In this study, a fourth level of decomposition was performed, i.e., m { 1 , 2 , 3 , 4 } . g 2 n k and h 2 n k are, respectively, the low-pass and high-pass FIR filters with a subsampling factor of 2. The process is further illustrated in Figure 5.
According to [40], sub-bands extracted by wavelet decomposition are functions of the incoming signal sampling frequency. In the proposed solution, the W i can be resampled at a specific frequency F r s i , resulting in an adaptive-rate decomposition for each W i , potentially achievingsub-bands with a lesser computational cost. This happens because the system has to process fewer number of samples compared to fix-rate decomposition concepts [4,5,6,12,13,14,15]. Furthermore, the adjustment of F r s i allows for a better focus on the incoming signal band of interest compared to the fix-rate counterparts [4,5,6,12,13,14,15].
a m i = k = 1 K g i x f n i · g 2 n k
d m i = k = 1 K g i x f n i · h 2 n k

2.6.2. Features

The wavelet coefficients obtained for each intended sub-band— d 1 i , d 2 i , d 3 i , d 4 i , and a 4 i —were used for extracting classifiable signal features. Nine statistical features were extracted for each considered sub-band. Therefore, in total, 45 features were employed to represent each selected segment.
The same features were extracted from each considered sub-band, as listed below.
Power Spectrum of the Signal (PS)
Power spectrum (PS) is computed as the average absolute value of the spectral means.
Mean Absolute Value of the Signal (MAV)
The mean absolute value (MAV) is calculated by adding all the absolute values of coefficients and then normalizing the sum.
Standard Deviation (STD)
Standard deviation (STD) is the measure of intended coefficients dispersion from the mean value.
Skewness of the Signal (SK)
Skewness (SK) is a measure of the asymmetry of the frequency distribution around its mean.
Kurtosis of the Signal (K)
Kurtosis (K) is a measure of the curvature of the considered coefficients.
Mean Ratio (R)
The mean ratio (R) is the ratio of the mean value of the detailed signal to the mean value of the approximate signal.
Peak Positive Value (PV)
The peak positive value (PV) is the maximum positive amplitude of the considered coefficients.
Peak Negative Value (NV)
The peak negative value (NV) is the maximum negative amplitude of the considered coefficients.
Second Peak Negative Value (NV2)
The second peak negative value (NV2) is the second maximum negative amplitude of the considered coefficients.

2.7. Classification Methods

Here, once relevant features were extracted, the data were represented in the form of a reduced data matrix, composed of 45 features, with 9 from each of the 5 selected sub-bands for each intended instance. Since the employed dataset had 5 ECG classes with300 instances per class, the resulting data matrix had a size of 1500 × 45. To classify this data matrix, the following classification techniques were employed.
Several of the used classification techniques require parameters tuning. To do this, we used the standard method of validation to tune parameters during the training phase to get the appropriate average result. These values were then fixed for the testing phase and are mentioned below.

2.7.1. k-Nearest Neighbors (k-NN)

The k-nearest neighbors (k-NN) algorithm [41,42] uses the k nearest neighbors of a test sample from the training dataset. Let vj represent a sample and <vj, lj> denote a tuple of a training sample and its label, lj ∈ [1,C] where C is the number of classes. Given a test sample, z, the mathematical process of computing the nearest neighbor, j, is presented by Equation (10):
argmin   d i s t j ( d j , t ) j = 1 . . N
In the designed solution, k = 3 and the Euclidean distance metric was used. The final label of z was selected as the most frequent label of the k chosen neighbors.

2.7.2. Artificial Neural Network (ANN)

The artificial neural network (ANN) is a popular class of algorithms and is based on the perceived mechanics of the human brain [43]. Standard multi-layered perceptron (MLP) was used here. The input and output layers, respectively, corresponded to the 45 features and the 5 possible classes. Hidden layers are important to model complex data, but care must be taken to avoid over-fitting. The number of the hidden layers was chosen equal to 5. The training function property ‘traingdx,’ in MATLAB [44] was used. This set the learning to gradient descent momentum and set a variable learning rate. The activation function used was the radial basis, and the maximum epochs were set to 1000.

2.7.3. Support Vector Machines (SVMs)

The support vector machine (SVM) was developed by Cortes and Vapnik [45]. It searches for the optimal separating hyperplane between support vectors of two classes. The process of separating the hyperplane is described mathematically in Equation (11), where x is the sample vector x = [x1, x2xp] with p attributes, w = [w1, w2wp] is the weights vector, and b is a scalar bias.
h ( x ) = s i g n ( w · x T + b )
For categorizing multiple classes, different approaches can be used [46,47], such as the one-vs-all approach or the one-vs-one approach. In this solution, the one-vs-all strategy was used with sequential minimal optimization (SMO) to train the classifier weights. The kernel function used was the polynomial of order 3, and the regularization parameter was set to 50.

2.7.4. Random Forest (RF)

The random forest (RF) was developed by using ideas put forward by Ho Tim Kam [48]. It is a technique that takes advantages of multiple classifiers. It constructs a multitude of decision trees at the training time and use the output from these trees to form a consensus. In contrast with bagging, RF may employ different decision tree techniques for the multiple subsets created. In the designed solution, the number of trees was set equal to 60, and out-of-bag predictions were retained for each tree. For the split at the nodes, the interaction-curvature method was selected, as it minimized the p-value of the chi-square tests of independence between each predictor and the response. The number of splits on each branch was limited to 10.

2.7.5. Bagging (BG)

Bagging (BG) is a bootstrap aggregation of classification trees. Multiple trees are allowed to fit the training data so that any bias, such as over-fitting, can be dealt with by using the ensemble of trees. Hence, bagging can be a powerful tool in classification. In this study, the number of bagged trees was set to 50, and ensemble predictions were used for out-of-bag observations.

2.8. Performance Evaluation Metrics

2.8.1. Compression Ratio

The compression ratio compares the proposed system performance in terms of reduction in the amount of information to be transmitted and classified compared to the conventional approach. In the classical case, acquired ECG data points are transmitted to the cloud without performing any features selection [24]. If N r and P are, respectively, the count of data points to be classified in the conventional and the proposed approach, then the compression ratio, R C O M P , can be calculated by using Equation (12):
R C O M P = N r P

2.8.2. Computational Complexity

The computational complexity of the embedded processing chain till the denoising module was studied in detail. The complexity of wavelet decomposition, features extraction, and cloud processing modules was analyzed at an abstract level by considering the reduction in amount of information to be processed by these modules.
The resampled signal was denoised by using an enhanced adaptive-rate FIR (ARFIR) filtering concept [27]. The arithmetic complexity of a classical K order FIR filter is clear from Equation (5). It executes (K − 1) additions and K multiplications while calculating an output sample. Therefore, for N samples, the entire computational complexity CFIR can be calculated by using Equation (13):
C F I R = ( K 1 ) · N A d d i t i o n s + K · N M u l t i p l i c a t i o n s
For the suggested solution, the online filter selection and the selected segment resampling processes necessitated additional operations.
  • Filter selection for W i wasresolved by using the successive approximation algorithm. Therefore, resolving the value of F r s i for Q reference filters, in the worst case, requires l o g 2 ( Q ) comparisons [29].
  • Resampling wasrealized by using the SLI. For W i , the complexity of SLI was N r i additions and N r i binary weighted right shifts. The complexity of binary weighted right shift wasnegligible compared to the addition and multiplication processes [49]. Therefore, it wasignored.
  • The complexity of the K i order FIR filtering for N r i samples could be calculated as: ( K i 1 ) · N r i additions and K i · N r i multiplications.
Therefore, the overall complexity of the proposed ARFIR method for W i can be calculated by using Equation (14):
C A R F I R = ( K i 1 ) · N r i + N r i + log 2 ( Q ) A d d i t i o n s + K i · N r i M u l t i p l i c a t i o n s

2.8.3. Classification Precision

The proposed solution seems appealing in terms of hardware complexity, compression, processing, and transmission efficiencies. However, it can lag in terms of precision. Therefore, the performance of the whole system wasstudied in terms of its classification accuracy. To avoid any bias in estimating the classification performance due to the limited dataset, cross-validation schema has been popularly used in the literature [47]. Therefore, 10-fold cross-validation was used in this study. All tested algorithms are provided with the same dataset, both training and test, for each fold. Similarly, to avoid any biasness in findings from any one measure, the following evaluation measures were used.

Accuracy (Acc)

Accuracy (Acc) is the percentage of labels that have been correctly classified. Let TP, TN, FP, and FN, respectively, denote true positives, true negatives, false positives, and false negatives in the predicted labels. Then, the mathematical formulation for accuracy is given by Equation (15). The accuracy measure results in a value between 0 and 1, with a higher value signifying a better performance.
A c c u r a c y = TP + TN TP + TN + FP + FN

Normalized Mutual Information (NMI)

Normalized Mutual Information (NMI) is an information theoretic score calculated as the mutual information between two distributions-labels predicted by a classifier and the real labels of the data. The value of NMI scores ranges between 0 and 1, with a higher score signifying a better classification. It is mathematically computable by using Equation (16), where X is the predicted clustering labels via the algorithm, Y is the real labels from the data, k(X) denotes the number of clusters in X, r j X is the number of elements in cluster j according to X, r u j is the number of elements predicted in u but actually belonging to j, and n is the total number of elements.
M I = 1 k ( X ) 1 k ( Y ) r u j l o g k ( X ) · k ( Y ) ( n · r u j r u X r j Y ) ( 1 k ( X ) r u X l o g ( r u X n ) ) ( 1 k ( Y ) r u Y l o g ( r u Y n ) )

F-Measure (F1)

The F-measure (F1) balances the values of precision and recall. We usually talk about a macro (without taking class sized into account) or micro (taking class sizes into account) F-measure. However, as all classes had the same data size in the studied case, we simply employ the F-measure. Mathematically, the F-measure is expressed by Equation (17), where precision = T P ( T P + F P ) and recall = T P ( T P + F N ) .
F = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l

Kappa Index (Kappa)

The kappa index (kappa) is a widely used statistics to judge the agreement of two clustering results. It is usually considered more robust than simple accuracy because it takes the possibility of agreement by chance into account. The most popular version is the Cohen’s kappa measure, which is mathematically expressed by Equation (18), where p0 is the percentage of agreement between the predicted and actual labels, similar to accuracy, and pe is the hypothetical probabilistic chance of such an agreement occurring randomly. It is given as p e = ( T P + T N ) ( T P + F N ) + ( F P + T N ) ( F P + F N ) ( T P + T N + F P + F N ) 2 . A perfect classification gives kappa = 1, and if the classification is merely due to chance, kappa = 0.
k a p p a = 1 1 p 0 1 p e

Specificity (Sp)

Specificity (Sp) measures a test’s ability to correctly generate a negative result for instances thatdo not belong to a given class. It is also known as the “true negative” rate. It is expressed by Equation (19):
S p = T N T N + F P

3. Results

The performance of the suggested solution wasstudied for five arrhythmia classes [36]. All system modules were implemented and validated by using the MATLAB® [44]. Examples of the pre-segmented heartbeats from all considered classes are shown in Figure 6.
In the classical case, the signal is band-limited to 60 Hz, and it is acquired with traditional ADC of 11-bit resolution at a sampling rate of 360 Hz [38]. The signal is divided into segments of 0.9 s, and each segment consists of 320 samples.
In the proposed case, to test the LCADC, the considered ECG signals were reconstructed by using U = 400. The reconstructed signals were acquired with an LCADC of M = 5-bit. The ECG signals were band-limited up to fmax = 60 Hz. Therefore, the maximum LCADC sampling frequency was F s m a x = 3.6 kHz. Examples of the pre-segmentation ECG signal, from the PVC class, acquired with an 11-bit classical ADC and a 5-bit LCADC, are shown in Figure 7.
The LCADC output was segmented by using the ASA. The ASA adapted L i and F r s i according to the x ˜ ( t ) temporal variations. It contributed to the enhancement of the system’s computational efficiency. The average compression ratios were computed for all 300 instances of each considered class. These were, respectively, 3.13-, 2.86-, 3.01-, 3.02-, and 3.05-fold for classes N, RBBB, LBBB, APC, and PVC. The attained overall average reduction in the number of collected samples for all five classes was 3.01-fold.
The resampled signal wasdenoised by using the ARFIR filtering technique [29]. A band-pass filter bank was designed for the cut-off frequencies of [FCmin = 0.7; FCmax = 35] Hz. The Filter bank was implemented for a set of sampling frequencies, Fref, between F s m i n = 75 Hz ˃ 2·FCmax to F r = 360 Hz. In this case, Δ = 19 Hz was chosen. It realized a bank of Q = 16 band-pass filters. A summary of the designed filter bank is shown in Table 2.
Denoising enhanced the intended signal SNR and resulted in a better classification [25]. The online filter order adaption resulted in signal denoising with a reduced computational cost compared to the time-invariant traditional ones [29]. The gains of the proposed ARFIR were computed over the classical one by using Equations (13) and (14). A summary of results for all 300 instances for each considered ECG class is presented in Table 3 and Table 4.
Table 3 and Table 4 illustrate that the proposed ARFIR technique achieved a noticeable computational gain over the traditional counterpart. It was attained by intelligently adapting the system parameters like Frsi, Nri, and hck for each selected segment (Wi) by following the x ˜ ( t ) temporal variations.
Each classifier was adjusted during the training phase with parameters set as mentioned in Section 2.7. Five classifiers, namely ANN, k-NN, SVM, RF, and bagging, were used. Their parameters were adjusted according to methods described in Section 2.7. They classified the intended dataset, and we measured the different evaluation metrics. The results are summarized in Table 5.
A summary of detail classifications results, in terms of TP, FP, FN, and TN for the five considered classes, is presented in Figure 8. These results showed that the integration of adaptive-rate signal acquisition and processing chain with RF and bagging resulted in superior classification performances as compared to k-NN, ANN, and SVM. Since RF uses multiple classifiers, it is less likely to be biased. This was reflected in the evaluation where RF resulted in the best NMI score and kappa statistics. On the other hand, k-NN can be easily biased by the chosen neighbors, particularly if there are outliers in the data. This was reflected in its low score across all indices. Overall, the highest accuracy score of 0.97 was attained by the designed solution.
For further analysis, we considered the detail classification results for each intended class, as shown in Figure 8. The overall false positive and false negative counts by using RF were the lowest compared to bagging, k-NN, ANN, and SVM, thus confirming the outperformance of RF compared to other considered algorithms. The most difficult class to discriminate was the RBBB, which was confused with LBBB, while the easiest class to discriminate was normal compared to the other four classes.

4. Discussion

The appealing features of the suggested solution are clear from results, presented in Section 3. This solution was attained by intelligently exploiting level-crossing sampling, adaptive-rate processing, and robust classifiers. The values of N r i were adapted by the ASA. We showed how F s i , L i , and F r s i were adjusted as a function of the temporal variations of x ˜ ( t ) . The adjustment of F r s i avoided unnecessary interpolations during the resampling process [29]. This signal-driven, adaptive-rate sampling resulted in a three-fold decrease in the number of collected samples compared to classical counterparts. It guaranteed a noticeable reduction in the processing load of the post denoising and features extraction.
K i represents how the adjustment of hck for Wi avoids the unnecessary operations while conditioning the selected segments. It yielded a noteworthy computational gain of the proposed denoising method over the conventional counterparts. The average gains in additions and multiplications, for the considered 1500 instants, of the employed ARFIR over the conventional one were 7.81- and 7.99-fold, respectively. Additionally, the adaptation of F r s i resulted in an adaptive-rate sub-band decomposition. F r s i reflected the x ˜ ( t ) frequency content [35]. Therefore, it was able to result in an improved focus on the band of interest, as the decomposition process is a function of F r s i [40]. The decomposition was attained by using the half-band FIR filters. A three-fold decrease in the number of collected samples confirmed computationally efficient sub-band decomposition compared to the counter fix-rate solutions [4,6,12,13,29,40].
Originally, each intended instance was composed of 320 samples. After features extraction, their dimensionality was reduced to 45 features, resulting in a 7.1-fold dimensionality reduction and guaranteeing the same factor of reduction in the data transmission activity, bandwidth, and power consumption. Additionally, on the cloud side, the processing of 7.1-fold less amount of information assured a similar factor of processing and resource utilization gain during the classification process.
The adoption of performing most of the signal processing tasks via the front-end processor and transmitting only the extracted features to the cloud was also beneficial in realizing an optimized front-end ECG wearable device while keeping the system configurable. Moreover, in the designed framework, the signal was digitized with a M = 5-bit resolution LCADC. In the classical case, the digitization is performed with an M = 11-bit resolution ADC [4,5,6,12,13,14,15]. Our study confirmed a simpler circuit level realization compared to the traditions ones. A low resolution ADC can also be used in the classical case, but it can reduce the system SNR [38] and can therefore degrade classification performance. On other hand, the effective resolution of LCADC is fairly independent of M [38]. Therefore, an accuracy of 97% was achieved by the proposed solution, even when using a 5-bit resolution LCADC.
The idea of embedding the level-crossing sampling and adaptive-rate processing in ECG signal-based automatic arrhythmia diagnosis is a novel concept. In [16,24,30,31,32], new approaches were proposed for effective ECG acquisition. These mainly focused on the analog-to-digital conversion of ECG signals. In [30,31], the authors demonstrated that how the use of LCS-based architectures led towards a simplified and power efficient analog/digital (A/D) conversion. In [32], the authors showed how the proposed ADC could adjust its effective resolution as a function of the activity of the input signal. In [16,24], efficient and real-time QRS detection mechanisms were proposed.
In contrast to [16,24,30,31,32], the proposed solution did not only focus on the design of an effective LCADC for ECG signal acquisition or QRS complexes detection; it additionally presented an application of the ASA and the adaptive-rate resampling, denoising, and sub-band decomposition approaches on the LCADC output. It was performed to achieve efficient segmentation, noise reduction, and features extraction of digitized ECG pulses to well prepare them for post cloud-based automatic arrhythmia classification.
The proposed technique is original, and comparing it with existing state-of-the-art methods was not straight forward, since they are based on classical sampling and processing. Additionally, each study uses a different number of subjects and different types of arrhythmia classes. It is also delicate to compare the diversity of classification methods and techniques for ECG signal processing. However, a comparison was performed with previous studies that use the same ECG dataset and DWT-based features extraction. The highest accuracies of classification for all considered studies are presented in Table 6, which shows that the proposed solution attained an analogous or better classification accuracy as compared to the fix-rate counterparts [4,5,6,12,13,14,15] while assuring noticeable processing, transmission, and hardware simplicity gains.
The main advantage of the proposed solution over the previous ones is the elimination of unnecessary samples to process and to introduce a real-time compression gain in the system. This was achieved by tactfully embedding the level-crossing sampling and ASA into the system. Similar gains could be achieved by embedding these concepts in counterparts [4,5,6,12,13,14,15].

5. Conclusions

An original level-crossing ECG signal sampling, adaptive-rate denoising, and sub-band decomposition and features extraction approach was proposed. It was shown that the proposed framework achieved a three-fold reduction in the number of acquired samples, thus leading to a significant reduction in the computational complexity of the designed system compared to the conventional counterparts. The overall average gains in additions and multiplications for the designed adaptive-rate denoising module were computed as 7.81- and 7.99-fold, respectively. The adaptive-rate processing promised a similar factor of processing gain during the adaptive-rate sub-band decomposition. The proposed features extraction concept reduced the incoming data dimensionality with a factor of 7.1 and assured a remarkable gain in terms of the power consumption and transmission activity between the ECG wearable device and cloud-based classification. Moreover, a similar magnitude of processing gain was also evident in the post cloud-based classifier because it had to deal with a 7.1 times lesser amount of information. The highest classification accuracy of 97% was attained by the random forest, which is comparable, and in some cases better, to existing state-of-the-art solutions. The plus side is our gain in compression, processing, and transmission bandwidth, which show the potential of using the suggested solution for the design and development of low power and efficient ECG wearable devices in the mobile healthcare framework.

Author Contributions

S.M.Q. proposed the concept and acquired funding. S.M.Q. and S.F.H. designed, implemented, and investigated the proposed solution. S.M.Q. and S.F.H. prepared original draft of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

The project was funded by the EFfat University, under the grant number UC#7/28Feb 2018/10.2-44g.

Acknowledgments

Authors are thankful to anonymous reviewers for their valuable feedback.

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Approval

This article does not contain any studies with human participants or animals performed by any of the authors.

References

  1. Brenyo, A.; Aktas, M.K. Review of complementary and alternative medical treatment of arrhythmias. Am. J. Cardiol. 2014, 113, 897–903. [Google Scholar] [CrossRef] [PubMed]
  2. Banerjee, S.; Mitra, M. Application of cross wavelet transform for ECG pattern analysis and classification. IEEE Trans. Instrum. Meas. 2013, 63, 326–333. [Google Scholar] [CrossRef]
  3. Mozaffarian, D.; Benjamin, E.J.; Go, A.S.; Arnett, D.K.; Blaha, M.J.; Cushman, M.; Howard, V.J. Heart disease and stroke statistics-2016 update a report from the American Heart Association. Circulation 2016, 133, e38–e48. [Google Scholar] [CrossRef] [PubMed]
  4. Hai, N.T.; Nguyen, N.T.; Nguyen, M.H.; Livatino, S. Wavelet-based kernel construction for heart disease classification. Adv. Electr. Electr. Eng. 2019, 17, 306–319. [Google Scholar] [CrossRef]
  5. Li, T.; Zhou, M. ECG classification using wavelet packet entropy and random forests. Entropy 2016, 18, 285. [Google Scholar] [CrossRef]
  6. Gutiérrez-Gnecchi, J.A.; Morfin-Magana, R.; Lorias-Espinoza, D.; del Carmen Tellez-Anguiano, A.; Reyes-Archundia, E.; Méndez-Patiño, A.; Castañeda-Miranda, R. DSP-based arrhythmia classification using wavelet transform and probabilistic neural network. Biomed. Signal Process. Control 2017, 32, 44–56. [Google Scholar] [CrossRef] [Green Version]
  7. Sharma, R.R.; Pachori, R.B. Baseline wander and power line interference removal from ECG signals using eigenvalue decomposition. Biomed. Signal Process. Control 2018, 45, 33–49. [Google Scholar] [CrossRef]
  8. Hesar, H.D.; Mohebbi, M. ECG denoising using marginalized particle extended kalman filter with an automatic particle weighting strategy. IEEE J. Biomed. Health Inform. 2016, 21, 635–644. [Google Scholar] [CrossRef]
  9. Singhal, A.; Singh, P.; Fatimah, B.; Pachori, R.B. An efficient removal of power-line interference and baseline wander from ECG signals by employing Fourier decomposition technique. Biomed. Signal Process. Control 2020, 57, 101741. [Google Scholar] [CrossRef]
  10. Sahoo, S.; Kanungo, B.; Behera, S.; Sabut, S. Multiresolution wavelet transform based feature extraction and ECG classification to detect cardiac abnormalities. Measurement 2017, 108, 55–66. [Google Scholar] [CrossRef]
  11. Martis, R.J.; Acharya, U.R.; Min, L.C. ECG beat classification using PCA, LDA, ICA and discrete wavelet transform. Biomed. Signal Process. Control 2013, 8, 437–448. [Google Scholar] [CrossRef]
  12. Martis, R.J.; Acharya, U.R.; Mandana, K.; Ray, A.; Chakraborty, C. Cardiac decision making using higher order spectra. Biomed. Signal Process. Control 2013, 8, 193–203. [Google Scholar] [CrossRef]
  13. Shadmand, S.; Mashouf, B. A new personalized ECG signal classification algorithm using block-based neural network and particle swarm optimization. Biomed. Signal Process. Control 2016, 25, 12–23. [Google Scholar] [CrossRef]
  14. Zhang, X.; Lian, Y. A 300-mV 220-nW event-driven ADC with real-time QRS detection for wearable ECG sensors. IEEE Trans. Biomed. Circ. Syst. 2014, 8, 834–843. [Google Scholar] [CrossRef]
  15. de Ruvo, E.; Sciarra, L.; Martino, A.M.; Rebecchi, M.; Iulianella, R.V.; Sebastiani, F.; Fagagnini, A.; Borrelli, A.; Scarà, A.; Grieco, D.; et al. A prospective comparison of remote monitoring systems in implantable cardiac defibrillators: Potential effects of frequency of transmissions. J. Interv. Card. Electrophysiol. 2016, 45, 81–90. [Google Scholar] [CrossRef]
  16. Guzik, P.; Malik, M. ECG by mobile technologies. J. Electrocardiol. 2016, 49, 894–901. [Google Scholar] [CrossRef]
  17. Miah, S.J.; Gammack, J.; Hasan, N. Extending the framework for mobile health information systems Research: A content analysis. Inf. Syst. 2017, 69, 1–24. [Google Scholar] [CrossRef] [Green Version]
  18. Rezaii, T.Y.; Beheshti, S.; Shamsi, M.; Eftekharifar, S. ECG signal compression and denoising via optimum sparsity order selection in compressed sensing framework. Biomed. Signal Process. Control 2018, 41, 161–171. [Google Scholar] [CrossRef]
  19. Singh, A.; Sharma, L.N.; Dandapat, S. Multi-channel ECG data compression using compressed sensing in eigenspace. Comput. Biol. Med. 2016, 73, 24–37. [Google Scholar] [CrossRef]
  20. Chen, S.; Hua, W.; Li, Z.; Li, J.; Gao, X. Heartbeat classification using projected and dynamic features of ECG signal. Biomed. Signal Process. Control 2017, 31, 165–173. [Google Scholar] [CrossRef]
  21. Niederhauser, T.; Haeberlin, A.; Jesacher, B.; Fischer, A.; Tanner, H. Model-based delineation of non-uniformly sampled ECG signals. In Proceedings of the 2017 Computing in Cardiology (CinC), Rennes, France, 24–27 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  22. Tekeste, T.; Saleh, H.; Mohammad, B.; Ismail, M. Ultra-Low Power QRS Detection and ECG Compression Architecture for IoT Healthcare Devices. IEEE Trans. Circ. Syst. Regul. Pap. 2018, 66, 669–679. [Google Scholar] [CrossRef]
  23. Qaisar, S.M.; Subasi, A. An adaptive rate ECG acquisition and analysis for efficient diagnosis of the cardiovascular diseases. In Proceedings of the 2018 IEEE 3rd International Conference on Signal and Image Processing (ICSIP), Shenzhen, China, 13–15 July 2018; pp. 177–181. [Google Scholar] [CrossRef]
  24. Budiman, E.S. Multi-Rate Analyte Sensor Data Collection with Sample Rate Configurable Signal Processing. U.S. Patent 20170027483A1, 2017. [Google Scholar]
  25. Qaisar, S.M.; Fesquet, L.; Renaudin, M. Adaptive rate filtering a computationally efficient signal processing approach. Signal Process. 2014, 94, 620–630. [Google Scholar] [CrossRef]
  26. Qaisar, S.M.; Ben-Romdhane, M.; Anwar, O.; Tlili, M.; Maalej, A.; Rivet, F.; Rebai, C.; Dallet, D. Time-domain characterization of a wireless ECG system event driven A/D converter. In Proceedings of the 2017 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Torino, Italy, 22–25 May 2017; pp. 1–6. [Google Scholar] [CrossRef]
  27. Qaisar, S.M. Efficient mobile systems based on adaptive rate signal processing. Comput. Electr. Eng. 2019, 79, 106462. [Google Scholar] [CrossRef]
  28. Hou, Y.; Qu, J.; Tian, Z.; Atef, M.; Yousef, K.; Lian, Y.; Wang, G. A 61-nW level-crossing ADC with adaptive sampling for biomedical applications. IEEE Trans. Circ. Syst. II Express Briefs 2018, 66, 56–60. [Google Scholar] [CrossRef]
  29. Marisa, T.; Niederhauser, T.; Haeberlin, A.; Wildhaber, R.A.; Vogel, R.; Goette, J.; Jacomet, M. Pseudo asynchronous level crossing ADC for ECG signal acquisition. IEEE Trans. Biomed. Circuits Syst. 2017, 11, 267–278. [Google Scholar] [CrossRef] [PubMed]
  30. Nasserian, M.; Peiravi, A.; Moradi, F. An adaptive-resolution signal-specific ADC for sensor-interface applications. Analog Integr. Circ. Signal Process. 2019, 98, 125–135. [Google Scholar] [CrossRef]
  31. Miskowicz, M. Event-Based Control and signal Processing; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar] [CrossRef]
  32. Mashhadi, M.B.; Salarieh, N.; Farahani, E.S.; Marvasti, F. Level crossing speech sampling and its sparsity promoting reconstruction using an iterative method with adaptive thresholding. IET Signal Process. 2017, 11, 721–726. [Google Scholar] [CrossRef]
  33. Moser, B.A.; Lunglmayr, M. On quasi-isometry of threshold-based sampling. IEEE Trans. Signal Process. 2019, 67, 3832–3841. [Google Scholar] [CrossRef] [Green Version]
  34. Moody, G.B.; Mark, R.G. The impact of the MIT-BIH arrhythmia database. IEEE Eng. Med. Biol. Mag. 2001, 20, 45–50. [Google Scholar] [CrossRef]
  35. Qaisar, S.M.; Yahiaoui, R.; Gharbi, T. An efficient signal acquisition with an adaptive rate A/D conversion. In Proceedings of the 2013 IEEE International Conference on Circuits and Systems (ICCAS), Kuala lumpur, Malaysia, 18–19 September 2013; pp. 124–129. [Google Scholar] [CrossRef]
  36. Qaisar, S.M.; Akbar, M.; Beyrouthy, T.; Al-Habib, W.; Asmatulah, M. An error measurement for resampled level crossing signal. In Proceedings of the 2016 Second International Conference on Event-based Control, Communication, and Signal Processing (EBCCSP), Krakow, Poland, 13–15 June 2016; pp. 1–4. [Google Scholar] [CrossRef]
  37. Massopust, P. Interpolation and Approximation with Splines and Fractals; Oxford University Press, Inc.: New York, NY, USA, 2010; ISBN 978-0-19-533654-2. [Google Scholar]
  38. Zhang, Z.; Telesford, Q.K.; Giusti, C.; Lim, K.O.; Bassett, D.S. Choosing wavelet methods, filters, and lengths for functional brain network construction. PLoS ONE 2016, 11, e0157243. [Google Scholar] [CrossRef] [PubMed]
  39. Hussain, S.F.; Bisson, G. Text categorization using word similarities based on higher order co-occurrences. In Proceedings of the 2010 SIAM International Conference on Data Mining (SDM), Columbus, OH, USA, 29 April–1 May 2010; pp. 1–12. [Google Scholar] [CrossRef] [Green Version]
  40. Dudani, S.A. The distance-weighted k-nearest-neighbor rule. IEEE Trans. Syst. Man Cybern. 1976, 325–327. [Google Scholar] [CrossRef]
  41. Parisi, G.I.; Kemker, R.; Part, J.L.; Kanan, C.; Wermter, S. Continual lifelong learning with neural networks: A review. Neural Netw. 2019, 113. [Google Scholar] [CrossRef] [PubMed]
  42. Guo, H.; Wang, W. Granular support vector machine: A review. Artif. Intell. Rev. 2019, 51, 19–32. [Google Scholar] [CrossRef]
  43. Hussain, S.F.; Suryani, A. On retrieving intelligently plagiarized documents using semantic similarity. Eng. Appl. Artif. Intell. 2015, 45, 246–258. [Google Scholar] [CrossRef]
  44. Moore, H. MATLAB for Engineers; Pearson: London, UK, 2017; ISBN 978-0-13-458964-0. [Google Scholar]
  45. Hussain, S.F. A novel robust kernel for classifying high-dimensional data using Support Vector Machines. Expert Syst. Appl. 2019, 131, 116–131. [Google Scholar] [CrossRef]
  46. Rokach, L. Decision forest: Twenty years of research. Inf. Fusion 2016, 27, 111–125. [Google Scholar] [CrossRef]
  47. Cavanagh, J. Computer Arithmetic and Verilog HDL Fundamentals; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar] [CrossRef]
  48. de Chazal, P.; O’Dwyer, M.; Reilly, R.B. Automatic classification of heartbeats using ECG morphology and heartbeat interval features. IEEE Trans. Biomed. Eng. 2004, 51, 1196–1206. [Google Scholar] [CrossRef] [Green Version]
  49. Christov, I.; Gómez-Herrero, V.; Krasteva, G.; Jekova, I.; Gotchev, A.; Egiazarian, K. Comparative study of morphological and time-frequency ECG descriptors for heartbeat classification. Med. Eng. Phys. 2006, 28, 876–887. [Google Scholar] [CrossRef]
Figure 1. The proposed system block diagram. LCADC: level-crossing analog-to-digital converters; ASA: activity selection algorithm.
Figure 1. The proposed system block diagram. LCADC: level-crossing analog-to-digital converters; ASA: activity selection algorithm.
Sensors 20 02252 g001
Figure 2. The principle of a level-crossing ADC with hysteresis.
Figure 2. The principle of a level-crossing ADC with hysteresis.
Sensors 20 02252 g002
Figure 3. The algorithmic state machine (ASM) chart of activity selection algorithm.
Figure 3. The algorithmic state machine (ASM) chart of activity selection algorithm.
Sensors 20 02252 g003
Figure 4. The process of selecting the resampling frequency and the filter coefficients from the reference filter bank for the ith selected segment.
Figure 4. The process of selecting the resampling frequency and the filter coefficients from the reference filter bank for the ith selected segment.
Sensors 20 02252 g004
Figure 5. The adaptive-rate discrete wavelet transform (ARDWT) scheme with four levels of decomposition.
Figure 5. The adaptive-rate discrete wavelet transform (ARDWT) scheme with four levels of decomposition.
Sensors 20 02252 g005
Figure 6. Examples of the pre-segmented electrocardiogram (ECG) signals.
Figure 6. Examples of the pre-segmented electrocardiogram (ECG) signals.
Sensors 20 02252 g006
Figure 7. The pre-segmented ECG signal, acquired with an 11-bit resolution conventional ADC (a) and acquired with a 5-bit resolution LCADC (b).
Figure 7. The pre-segmented ECG signal, acquired with an 11-bit resolution conventional ADC (a) and acquired with a 5-bit resolution LCADC (b).
Sensors 20 02252 g007
Figure 8. True positive (TP), false positive (FP), false negative (FN), and true negative (TN) for different ECG classes.
Figure 8. True positive (TP), false positive (FP), false negative (FN), and true negative (TN) for different ECG classes.
Sensors 20 02252 g008
Table 1. MIT-BIH arrhythmia records used in this study. RBBB: right bundle branch block; LBBB: left bundle branch block; APC: atrial premature contraction; PVC: premature ventricular contraction.
Table 1. MIT-BIH arrhythmia records used in this study. RBBB: right bundle branch block; LBBB: left bundle branch block; APC: atrial premature contraction; PVC: premature ventricular contraction.
CLASSRecords Number of Beats Used
Normal116, 119, 209300
RBBB118, 124, 212300
LBBB109, 111, 214300
APC118, 200, 209300
PVC119, 200, 233300
Table 2. Summary of the reference filters bank parameters.
Table 2. Summary of the reference filters bank parameters.
hckh1kh2kh3kh4kh5kh6kh7kh8kh9kh10kh11kh12kh13kh14kh15kh16k
Frefc (Hz)7594113132151170189208227246265284303322341360
Kc23303643495561687480869299105111117
Table 3. Summary of the adaptive-rate finite impulse response (ARFIR) computational gains, in additions, over the classical FIR filtering.
Table 3. Summary of the adaptive-rate finite impulse response (ARFIR) computational gains, in additions, over the classical FIR filtering.
CLASSMax Gain
in Additions
Min Gain
in Additions
Mean Gain
in Additions
Median Gain
in Additions
Normal15.976.678.127.90
RBBB12.176.667.327.27
LBBB12.466.707.867.66
APC10.756.757.717.56
PVC15.066.718.037.77
Table 4. Summary of the ARFIR computational gains, in multiplications, over the classical FIR filtering.
Table 4. Summary of the ARFIR computational gains, in multiplications, over the classical FIR filtering.
CLASSMax Gain
in Multiplications
Min Gain
in Multiplications
Mean Gain
in Multiplications
Median Gain
in Multiplications
Normal16.146.748.318.09
RBBB12.306.737.487.43
LBBB12.586.768.047.84
APC10.866.817.897.74
PVC15.226.768.227.95
Table 5. Performance of different classifiers using five evaluation metrics on the test dataset. ANN: artificial neural network; k-NN: k-nearest neighbors; SVM: support vector machine; and RF: random forest.
Table 5. Performance of different classifiers using five evaluation metrics on the test dataset. ANN: artificial neural network; k-NN: k-nearest neighbors; SVM: support vector machine; and RF: random forest.
ClassifierAccNMIF1KappaSp
ANN0.930.820.930.770.98
k-NN0.830.710.810.460.96
SVM0.930.840.930.750.98
RF0.970.940.970.900.99
Bagging0.960.890.960.790.99
Table 6. Comparison with state-of-the-art methods.
Table 6. Comparison with state-of-the-art methods.
StudyFeatures ExtractionClassification MethodNo. of ClassesAccuracy (%)
[4]Wavelet Packet Decomposition (WPD) and Wavelet-Based kernel Principle Component Analysis (wkPCA)Backpropagation Neural Network (BNN)598.03
[5]Wavelet Packet Entropy (WPE)Random Forests (RF)594.61
[6]Discrete Wavelet Transform (DWT)Probabilistic Neural Network (PNN)892.75
[12]Discrete Wavelet Transform (DWT), Temporal, and MorphologicalSupport Vector Machine (SVM)498.39
[13]Discrete Wavelet Transform (DWT) and Principle Component Analysis (PCA)Support Vector Machine (SVM)-Radial Basis Function (RBF)596.92
[14]Bispectrumand Principle Component Analysis (PCA)Support Vector Machine (SVM)-Radial Basis Function (RBF)593.48
[15]HermiteFunction Coefficient and Temporal FeaturesOptimized block-based Neural Network (OBNN)597.00
This StudyWavelet Decomposition and Sub-band statistical featuresRandom Forests (RF)597.00

Share and Cite

MDPI and ACS Style

Mian Qaisar, S.; Fawad Hussain, S. Arrhythmia Diagnosis by Using Level-Crossing ECG Sampling and Sub-Bands Features Extraction for Mobile Healthcare. Sensors 2020, 20, 2252. https://doi.org/10.3390/s20082252

AMA Style

Mian Qaisar S, Fawad Hussain S. Arrhythmia Diagnosis by Using Level-Crossing ECG Sampling and Sub-Bands Features Extraction for Mobile Healthcare. Sensors. 2020; 20(8):2252. https://doi.org/10.3390/s20082252

Chicago/Turabian Style

Mian Qaisar, Saeed, and Syed Fawad Hussain. 2020. "Arrhythmia Diagnosis by Using Level-Crossing ECG Sampling and Sub-Bands Features Extraction for Mobile Healthcare" Sensors 20, no. 8: 2252. https://doi.org/10.3390/s20082252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop