Next Article in Journal
Microscale Sensor Arrays for the Detection of Dopamine Using PEDOT:PSS Organic Electrochemical Transistors
Previous Article in Journal
Walking on Virtual Surface Patterns Leads to Changed Control Strategies
Previous Article in Special Issue
An Optical Signal Simulator for the Characterization of Photoplethysmographic Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Detection Method of Atrial Fibrillation by CEPNCC-BiLSTM Based on Long-Term Photoplethysmography Data

1
School of Mechatronics Engineering and Automation, Foshan University, Foshan 528000, China
2
Guangdong Provincial Key Laboratory of Industrial Intelligent Inspection Technology, Foshan University, Foshan 528000, China
3
School of Education, City University of Macau, Macau 999078, China
4
Cosmetic Dermatology Department, Dermatology Hospital of Southern Medical University, Guangzhou 510091, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(16), 5243; https://doi.org/10.3390/s24165243
Submission received: 18 June 2024 / Revised: 14 July 2024 / Accepted: 25 July 2024 / Published: 14 August 2024

Abstract

:
Atrial fibrillation (AF) is the most prevalent arrhythmia characterized by intermittent and asymptomatic episodes. However, traditional detection methods often fail to capture the sporadic and intricate nature of AF, resulting in an increased risk of false-positive diagnoses. To address these challenges, this study proposes an intelligent AF detection and diagnosis method that integrates Complementary Ensemble Empirical Mode Decomposition, Power-Normalized Cepstral Coefficients, Bi-directional Long Short-term Memory (CEPNCC-BiLSTM), and photoelectric volumetric pulse wave technology to enhance accuracy in detecting AF. Compared to other approaches, the proposed method demonstrates faster preprocessing efficiency and higher sensitivity in detecting AF while effectively filtering out false alarms from photoplethysmography (PPG) recordings of non-AF patients. Considering the limitations of conventional AF detection evaluation systems that lack a comprehensive assessment of efficiency and accuracy, this study proposes the ET-score evaluation system based on F-measurement, which incorporates both computational speed and accuracy to provide a holistic assessment of overall performance. Evaluated with the ET-score, the CEPNCC-BiLSTM method outperforms EEMD-based improved Power-Normalized Cepstral Coefficients and Bi-directional Long Short-term Memory (EPNCC-BiLSTM), Support Vector Machine (SVM), EPNCC-SVM, and CEPNCC-SVM methods. Notably, this approach achieves an outstanding accuracy rate of up to 99.2% while processing PPG recordings within 5 s, highlighting its potential for long-term AF monitoring.

1. Introduction

Stroke is a leading cause of death and disability worldwide [1,2]. Reliable research has established a strong correlation between AF and ischemic stroke [3]. Furthermore, strokes associated with AF carry higher risks [4,5,6,7,8,9]. Although there are numerous methods for detecting AF based on electrocardiogram (ECG) and ECG-like signals, these methods have inherent limitations [10]. While ECG serves as the gold standard for AF detection and accurately analyzes heart rate and rhythm [11], identifying paroxysmal AF with short-period ECG recordings remains challenging due to its paroxysmal and complex nature [12]. Additionally, long-term ECG monitoring requires medical professionals, disrupts daily activities, and increases monitoring costs. Therefore, there is an urgent need to develop a novel, accurate, and efficient method for processing long-term PPG recordings that overcomes the limitations of traditional methods while enhancing AF detection rates.
Studies suggest that PPG recordings hold potential as an alternative to ECG for AF detection. Chong et al. [13] proposed that PPG waveforms may contain features beyond heart rate. However, further validation is required regarding the accuracy of PPG devices and the ability of physicians to detect AF based on PPG recordings [14]. To address these challenges, researchers have explored the application of image processing techniques and artificial intelligence methods in analyzing PPG recordings [15,16,17]. With advancements in computing technology, deep learning methods have gained significant attention [18]. For instance, Wu et al. [19] utilized a smartwatch device to obtain 30 s PPG recordings and achieved an overall accuracy of 92.38% using a hybrid deep learning model called Res-BiANet across 28,440 signal segments from 102 patients. Although studies have demonstrated the effective detection of AF using shorter periods of PPG recording in some cases, the paroxysmal nature of AF presents the risk of missed detection with short-duration signal segments [20].
PPG devices have attracted significant attention due to their potential for long-term monitoring and affordability [21]. However, the application of long-term PPG recordings is limited as it increases the computation time and poses challenges for rapid AF detection. Chen et al. [22] developed a novel smart bracelet that utilizes long-term PPG recordings to detect AF, demonstrating the integration potential of PPG with smart devices. For instance, Saarinen et al. [23] employed 5 min PPG and ECG recordings to analyze AF through a random forest algorithm, achieving an accuracy of 97.4%. Similarly, Kotlyarov et al. [24] analyzed over 2 min PPG recordings using a support vector machine (SVM) with a processing time of approximately 3 min, achieving sensitivity of 92.3%, specificity of 94.7%, and accuracy of 93.5%. While utilizing long-term PPG recordings can indeed enhance detection performance, the extended processing time is unfavorable for rapid AF diagnosis [25]. Moreover, there is currently no established standard in the field of AF diagnosis regarding the selection criteria for term length and processing time when utilizing PPG recordings. Evaluating the efficiency of an AF diagnostic method requires considering the appropriate recording duration, along with corresponding signal processing and classification techniques.
In the field of AF detection, convolutional neural networks (CNNs) are widely adopted due to their excellent spatial feature extraction capabilities [26]. However, they face challenges in capturing temporal dependencies within electrophysiological signals. To address this issue, recurrent neural networks (RNNs) have been introduced. For example, Senturk et al. [27] integrated ECG and PPG recordings as inputs into an RNN with the objective of estimating continuous blood pressure. Xu-K et al. [28] employed RNNs to achieve precise PPG cardiogram segmentation, resulting in three key indicators for heart rate variability estimation. However, traditional RNNs often face the issue of gradient vanishing when processing long-sequence data. To address this limitation, the bidirectional long short-term memory (BiLSTM) model has been employed [29], with the aim of enhancing the model’s capacity to capture contextual information in time series analysis and thereby improving AF event recognition.
In conclusion, this paper proposes an innovative method named CEPNCC-BiLSTM for the precise classification of AF characteristics based on long-term PPG recordings. To address the challenges in AF detection from PPG recordings and improve the processing speed of long-term PPG data, the proposed method enhances power normalized cepstral coefficients (PNCC) by incorporating complementary ensemble empirical mode decomposition (CEEMD) to extract frequency domain features. These features are then combined with time domain features to form a feature matrix. Subsequently, the feature matrix is fed into a BiLSTM neural network for accurate and effective identification and classification of AF characteristics. Additionally, to address the limitations of traditional evaluation systems that assess efficiency and accuracy comprehensively, this paper introduces a novel evaluation method called the ET-score, which incorporates time factors into the F-measure. This enables a comprehensive evaluation of various methods’ efficiency and accuracy.

2. Materials and Methods

2.1. Data

In this study, the effectiveness of the proposed algorithm is validated through experimental analysis. The PPG recordings are sourced from the MIT-BIH-MIMIC-III database at the Massachusetts Institute of Technology (MIT). MIMIC-III (‘Medical Information Mart for Intensive Care’) is an internationally recognized ECG database that accurately represents differences in pulse classification algorithms. The MIMIC-III Waveform Database is a large single-center database comprising vital signs, medications, laboratory measurements, observations, etc. [30]. In this research, PPG recordings were collected from both patients with AF and individuals exhibiting normal sinus rhythm (NSR) in the Synthetic Dataset [31] (1500 records in total). Among them, the signals of AF patients were confirmed by the diagnostic opinions provided by MIMIC-III. Furthermore, to validate the ability of the proposed method to detect AF patients with additional diseases, a small number of AF patients with other diseases were also included in the dataset.
To develop the classification models, the dataset was divided into two parts: a training set and a test set. The former was used for model training and the latter was kept as unseen data to test the classification performance. To ensure similar rhythm distributions in both datasets, 66.6% of the patients were assigned to the training set and 33.3% were assigned to the test set based on the rhythm. The details of the rhythm distributions are presented in Table 1, including the percentage of beats in each class and the number of patients having that rhythm.

2.2. Evaluation Indicators

To evaluate the sensitivity and time complexity of different methods simultaneously and provide a more intuitive assessment of their overall performance, this paper introduces the ET-score to evaluate the performance of AF detection methods using the following evaluation formula:
E T s c o r e = β 2 + 1 × ( F A F × F N S R ) β 2 · ( F A F + F N S R ) × t i m e ,
where β represents the estimated parameter, and in this paper, β is set to 1. F A F and F N S R are the F-measures corresponding to AF and NSR, respectively. The calculation speed factor of the AF detection method is t i m e , whose maximum value is 100 and decreases by 2 in order of calculation speed ranking.
Several evaluation methods have been investigated to assess the accuracy and computation time of AF detection algorithms, aiming to demonstrate their performance in different aspects. However, none of these methods can comprehensively evaluate the overall performance of the algorithm. When comparing multiple methods, the ET-score incorporates a time factor related to the computational speed in addition to the accuracy rate. The ET-score considers both efficiency and accuracy, providing an assessment of the AF detection methods’ effectiveness. A high ET-score indicates that a method is both accurate and computationally efficient.

2.3. Preprocessing and Signal Analysis

The preprocessed PPG recordings commence from a peak and have a duration of 2 min. Examples of 10 s segments of the PPG waveforms for different rhythm types are presented in Figure 1. Compared to NSR, AF patients exhibit intermittent fluctuations in PPG recordings, which are significant factors influencing AF detection [32]. To determine the optimal PPG feature extraction times for AF identification, PPG recordings were divided into 1 min, 2 min, and 5 min for pre-experiments. The feature extraction time, network training time, classification accuracy, and ET-score were calculated, respectively. Our study reveals that 2 min PPG recordings achieve the highest ET-score; in other words, 2 min PPG recordings strike a balance between efficiency and convenience, as shown in Table 2. Tang et al. [33] also demonstrated that longer signals result in better classification performance, but excessively long signal records will prolong the computation time. Based on the results of the pre-experiments, PPG recordings with a length of 2 min containing 15,000 data points were utilized in this study. The selection of different signals was performed by detecting local maxima, and all the signals were normalized to ensure the accuracy of the subsequent algorithm.

2.4. CEPNCC Algorithm Details

PNCC is a speech-processing method that demonstrates strong efficiency in dealing with physiological signals [34]. Figure 1 shows the statistical characteristics of the PPG recordings varying with time, indicating that it is a non-smooth signal. However, the application of the fast Fourier transform (FFT) for signal preprocessing proves to be ineffective for non-smooth signals. Chen et al. [35] proposed an EEMD-based improved Power-Normalized Cepstral Coefficients (EPNCC) approach by employing Ensemble Empirical Mode Decomposition (EEMD) for signal preprocessing instead of FFT. However, EEMD requires hundreds of calculations to reduce the residual error to a low level. In order to enhance efficiency while reducing error, the CEEMD [36] and EEMD were utilized to preprocess PPG recordings, with their respective processing times detailed in Table 3. Notably, CEEMD exhibits a significant advantage in terms of processing time.
CEEMD is used to preprocess the PPG recordings with the following steps.
Step 1: Let N be the overall number of processing of the original signal, indicating the total number of white noise signals added.
Step 2: A positive white noise signal sequence n i and a negative white noise signal sequence n i are added to the original signal s ( t ) , forming two new sets of signals, r i + ( t ) and r i ( t ) , respectively, where i = 1 ,   2 ,   ,   N .
Step 3: Using EMD to decompose r i + ( t ) and r i ( t ) separately, a series of intrinsic modal functions (IMFs) are obtained. Note that I i j + ( t ) and I i j ( t ) are the j t h intrinsic modal functions of r i + ( t ) and r i ( t ) , respectively.
Step 4: Let I j ( t ) be the jth intrinsic modal function of the reconstructed signal where I j ( t ) is the average of the components I i j + ( t ) and I i j ( t ) , as shown in the following equation:
I j t = 1 2 N i = 1 N ( I i j + t + I i j ( t ) ) .
Step 5: Let p be the total number of all IMFs in the reconstructed signal and let r ( t ) be the residual of the reconstructed signal. The corresponding IMFs are averaged as the result of the decomposition, as shown in the following equation:
s t = j = 1 p I j + t + r t .
After decomposing the PPG recordings using CEEMD, each IMF component and the residuals were obtained. Due to the time-varying and random nature of the noise, the IMF components after EEMD decomposition will vary. The large differences in these IMF components are meaningless for characterizing the impulse signal properties. Therefore, before extracting PNCC frequency domain features, a correlation analysis was conducted on the EEMD-decomposed IMF components to filter out those that characterize the impulse signal. Correlation analysis can determine the coherence between each IMF component and the original PPG recordings, and the expression for the coherence coefficient is
μ i = c o v ( i m f i ,   S i g ) V a r i m f i   V a r [ S i g ] ,
where n is the total number of the obtained IMFs, S i g is the original PPG record, and the desired IMF is selected by the difference from the set threshold value.
After the PPG record is processed by EEMD, the components are arranged and expressed as a feature matrix composed of IMFs. Here are the CEPNCC feature extraction steps:
Step 1: The estimated power spectrum P i ω for each IMF is calculated as follows:
P i ω = lim T I M F i 2 π T
Step 2: The power estimate   P i ω is input into the Gammatone filter for filtering, and the time domain impulse response of the Gammatone filter is formulated as follows:
G t = a t n 1 e 2 π w t cos 2 π f 0 t + φ , t > 0
where w is the filter bandwidth and n is the filter order.
Step 3: After filtering, the filter is normalized to its power spectrum (PN). The power-normalization expression is as follows:
U i = P o w n ( ω g ) μ [ ω ]
where P O W n is the value obtained from the Gammatone filter and μ [ ω ] is the average power.
Step 4: Power function nonlinear processing is expressed as follows:
P O W n = U n ϑ
where ϑ is the exponential factor, while in general 0 < ϑ < 1 .
Step 5: The CEPNCC eigencoefficients are the obtained eigenmatrices.
After obtaining the CEPNCC feature parameters through a series of processing steps, the CEPNCC feature parameters are mixed with the time domain features and input into the neural network for training.

2.5. Net Modeling

BiLSTM was utilized for processing PPG recordings in the development of AF classification models in this study. BiLSTM, comprising a forward LSTM and a backward LSTM, is well-suited for modeling time series data and is commonly employed in natural language processing tasks to capture bidirectional temporal dependencies and learn contextual information.
In this study, a neural network model based on BiLSTM was constructed for AF classification. Figure 2a illustrates the basic unit of the BiLSTM, while Figure 2b depicts a block diagram of the overall model. The CEPNCC feature matrix obtained from PPG recordings serves as the input for calculations in this model.
The BiLSTM model (Figure 2) was utilized for binary classification. The PPG recordings diagnosed as AF were grouped into one category, while non-AF rhythms were categorized separately.
To comprehensively characterize the PPG record features, the periodic time domain and frequency domain features were first normalized and activated before being input into the BiLSTM network for training. Subsequently, the classification of PPG recordings was achieved using a SoftMax classifier, which is well-suited to the specific classification requirements of this paper. The SoftMax classifier transforms the logits from the BiLSTM model into a probability distribution by applying the SoftMax function. This allows the score for each category to be interpreted as the probability of that category. Notably, the SoftMax classifier is characterized by its computational simplicity and high efficiency. The number of categories corresponds to clinical diagnostic categories associated with the PPG recordings. Specifically, the presence or absence of AF is indicated by [1] and [0], respectively.

3. Experimental Results

After conducting a comprehensive analysis of data preprocessing techniques, the subsequent crucial step involves extracting relevant features from the preprocessed signals to facilitate the accurate classification of AF.

3.1. PPG Recordings

In this study, we identified peak points in the database of PPG recordings and subsequently grouped and segmented them. The sampling frequency of the PPG recordings from the MIMIC-III database was 125 Hz. Therefore, 15,000 data points starting from the peak were selected as the data sample for this experiment. We obtained the feature matrix from the CEPNCC decomposition, as shown in Figure 3. Due to the intermittent fluctuations in the PPG recordings of AF patients, their feature matrix exhibited confounding patterns in the low-frequency domain, with a more concentrated energy distribution in the image. As illustrated in Figure 3, the NSR feature matrix and the AF feature matrix displayed significant differences in amplitude and distribution. The AF feature matrix showed higher energy and uneven distribution at the initial points, with more low-value regions, indicating greater energy variations in AF patients. In contrast, the NSR feature matrix had a more uniform energy distribution and exhibited smaller fluctuations, reflecting the more stable energy variations in patients with normal sinus rhythm. The BiLSTM network captures these features effectively. After combining the frequency domain feature matrix with the time domain features, we employed the BiLSTM network for training.

3.2. Net Training and Analysis

To ensure optimal performance of the neural network, it is crucial to carefully set and fine-tune the network parameters. This includes selecting an appropriate optimizer and determining the optimal learning rate, as these factors significantly influence the accuracy and efficiency of the AF detection model.

3.2.1. Network Settings

We set the network structure parameters according to the description in the previous Section 2.5. Following the testing of different network parameters, the Adam optimizer was chosen to regulate the learning rate, which could significantly impact the final AF detection results in machine learning. To select an appropriate learning rate, we designed an optimal learning rate estimation method. A very small initial learning rate for the AF detection network was chosen in this work. The learning rate was increased after each batch, with the retention of the loss obtained from each batch. Subsequently, we created a plot of the learning rate versus loss curve, as shown in Figure 4.
As illustrated in Figure 4, the learning rate achieving the minimum loss was located roughly within the interval [ 0.6 × 10 2 , 0.2 × 10 1 ] . To accurately determine the optimal learning rate, the accuracy rate α should satisfy the following equation:
α = f φ , ω
where φ is the network loss rate and ω is the initial learning rate. The optimal estimate of the network’s initial learning rate is determined by the following equation based on the obtained accuracy parameters ( φ i , α i ) :
L α , f φ , ω = i = 1 N [ α i , f ( φ i , α i ) ] 2
The minimum value can be obtained from the sequence [ ω 1 , ω 2 , , ω i ] . The relationship between various learning rates, learning rate decay factors (LRDFs), and the network’s classification accuracy was analyzed to select the appropriate LRDF. The results are presented in Figure 5.
It should be noted that both different learning rates and LRDFs can affect the accuracy of the network. As the LRDF increases, there is generally a decrease in the classification accuracy for different learning rates; notably, when the LRDF is 0.5, the classification accuracy reaches its peak. The moderate learning rate decay allows the model to gradually converge during training while maintaining adequate exploration capabilities. This ultimately enables the model to effectively find the optimal solution. When the LRDF falls below 0.5, the AF classification accuracy drops below 80%. This slow decay leads to inadequate parameter adjustment during training and suboptimal performance with decreased classification accuracy. In order to ensure the clarity of the image, the accuracy of the network with an LRDF less than 0.5 is not listed in Figure 5. Therefore, the initial learning rate used for our network was 0.01525, the LRDF was 0.5, and the maximum iteration period was 100.

3.2.2. PPG Classification Recognition Rate of CEPNCC-BiLSTM

We utilized the CEPNCC feature matrix, extracted from the PPG recordings, for training, classifying, and validating the network. The categorical data in Table 4 were derived from the confusion matrix presented in Figure 6. The model acquired through network training was cross-validated on the test dataset, yielding a final classification accuracy of 99.20%. As depicted in Table 4, the proposed method demonstrates the precise detection of PPG recordings in AF patients with a recall rate of 100%. For PPG recordings from NSR, the false alarm rate was merely 2%. Across multiple experiments, CEPNCC-BiLSTM consistently exhibited the accurate screening of AF patients.

3.2.3. Compared with Existing Methods

To demonstrate the superiority of the proposed method in AF classification, we compared it with other methods, including EPNCC-BiLSTM, CEPNCC-BiLSTM, SVM, EPNCC-SVM, and CEPNCC-SVM. All the models were trained on the same hardware environment using the same training set and test set.
As illustrated in Table 5 and Table 6, the SVM classifier was capable of producing results in a relatively short time; however, it exhibited a significant loss of accuracy and an AF false alarm rate of 26%, which was insufficient to meet the requirements for AF detection. In contrast, AF detection could be classified rapidly and with minimal loss of accuracy by employing EPNCC or CEPNCC to transform PPG recordings into feature matrices. Nevertheless, it should be noted that there were notable differences between the aforementioned two methods for AF detection. EPNCC required a considerable number of calculations to fulfill the accuracy requirements due to residual errors, which consequently prolonged its preprocessing time.
The application of CEPNCC for the preprocessing of PPG recordings had the effect of reducing both the necessary preprocessing and training times while simultaneously achieving a high degree of accuracy. As illustrated in Table 5, the CEPNCC-BiLSTM classification method exhibited high sensitivity in AF detection and was capable of accurately classifying PPG recordings with AF features. In multiple experiments, the proposed method demonstrated the capacity to not only detect AF features in PPG recordings with precision but also achieve a markedly lower false-positive rate than other methods under comparison.

3.2.4. Proposed Evaluation Methodology

Table 7 demonstrates that the EPNCC-BiLSTM achieved a classification accuracy of 91.46% and a recognition accuracy of 97.79%. However, it exhibited the slowest computation speed, resulting in the lowest ET-score. Conversely, using only SVM as a classifier yields faster computation but lower ET-score due to decreased accuracy. When EPNCC-SVM was employed for AF classification, it attained an accurate AF recognition rate of 93.62% with rapid computational speed. In contrast, CEPNCC-SVM proved relatively more efficient with an AF detection accuracy of 96.18%. Notably, CEPNCC-BiLSTM outperformed all the methods by achieving the highest ET-score through preprocessing PPG recordings in approximately 5.8 s and achieving exceptional AF detection accuracy at 99.2%. Furthermore, CEPNCC-BiLSTM consistently demonstrated high sensitivity to AF detection across multiple experiments, with most false positives originating from PPG recordings of non-AF patients.

4. Discussion

This study proposes an intelligent method for AF detection based on long-term photoplethysmography (PPG) data and the CEPNCC-BiLSTM model. The experimental results demonstrate that this method significantly outperforms existing approaches in terms of accuracy and efficiency, addressing the issue of under-detection caused by the AF’s paroxysmal nature and meeting AF detection requirements.
The CEPNCC-BiLSTM model enhances AF detection sensitivity and specificity by combining CEEMD and PNCC techniques to extract both time domain and frequency domain features. CEEMD efficiently processes non-smooth signals, contributing to reduced computation time and improved accuracy. Additionally, the bidirectional characteristic of the BiLSTM model enables better capture of contextual information in time series data, enhancing its ability to recognize AF features.
Compared to traditional methods, the CEPNCC-BiLSTM model excels in various performance metrics. When compared to EPNCC-BiLSTM, our proposed method not only improves the detection accuracy (99.2% vs. 93.62%) but also significantly reduces the preprocessing time (5.8291 s vs. 36.1267 s). In comparison with the SVM method, the CEPNCC-BiLSTM model demonstrates superior accuracy (99.2% vs. 96.18%) and computational efficiency. Moreover, these improvements are attributed to the efficient preprocessing by CEEMD and the robust time series processing capability of the BiLSTM model, which reduces the computational overhead while capturing signal features effectively. Furthermore, the CEPNCC-BiLSTM model exhibits high efficiency in processing non-smooth signals, making it more stable and reliable in long-term monitoring: a crucial characteristic for practical applications.
The proposed method holds great potential for practical applications, particularly in long-term AF monitoring. Its efficient processing capabilities make the CEPNCC-BiLSTM more suitable for wearable devices and real-time monitoring systems, which can help improve early detection rates of AF, thereby reducing the risk of stroke and other complications without enabling continuous cardiac health monitoring.
While the CEPNCC-BiLSTM method demonstrated strong performance, it also exhibited certain limitations. Firstly, the diversity and size of the dataset may impact the model’s ability to generalize. This study utilized publicly available data, consisting of 400 healthy human PPG recordings from the Synthetic Dataset and 1100 PPG recordings of AF patients from the MIMIC-III Waveform Database. To improve the model’s generalization performance, additional AF data were appropriately selected. The MIMIC-III Waveform Database was obtained from bedside monitors in an intensive care unit (ICU) within an inpatient unit. Among the AF PPG recordings, 89.09% of patients had a sole diagnosis of AF, while another 10.91% had a concurrent diagnosis of AF and other disorders; however, detailed information regarding these disorders was not disclosed by the data organization. ICU patients typically exhibit more severe physiological states due to high stress levels and various comorbidities, which can introduce complexity into PPG recordings. Consequently, these data might not accurately represent outpatient or community patient populations, thereby limiting their applicability to other scenarios such as general wards or outpatient clinics.
The experimental results in Figure 6 revealed four false positive samples, which can be attributed to (1) significant individual differences among samples and (2) interference from other diseases affecting the model’s learning of atrial fibrillation-related features. Consequently, enhancing the model’s generalization ability will be a primary focus in our team’s future research endeavors. Furthermore, incorporating more diverse population data and exploring atrial fibrillation detection effectiveness across different contexts are essential considerations. Additionally, future research should encompass classifying various arrhythmias (e.g., ventricular premature beats and tachycardia) to offer a comprehensive cardiovascular health monitoring solution while identifying other sources of interference in arrhythmia signals.

5. Conclusions

In this study, an intelligent method named CEPNCC-BiLSTM has been proposed for the detection and diagnosis of AF with long-term PPG recordings. This method significantly improves the accuracy of AF detection and overcomes the limitations of intermittent monitoring in traditional methods. Additionally, it achieves superior classification accuracy while demonstrating improved computational efficiency compared to other long-term monitoring methods. Specifically, our approach accurately identifies PPG signals with AF characteristics, achieving an impressive accuracy rate of 99.2%. Moreover, it processes 2 min PPG recordings in just 5.8 s, making it the fastest among the five methods considered in this study. Furthermore, we introduce a novel performance metric based on the ET-score that incorporates a temporal dimension beyond F-measure evaluation. The ET-score for CEPNCC-BiLSTM is evaluated to be the highest among all five methods, further confirming its effectiveness. Importantly, our research findings demonstrate that PPG signals contain sufficient information for precise and rapid differentiation between AF and normal NSR using CEPNCC-BiLSTM. Therefore, long-term PPG recordings can be effectively used for accurate AF detection.

Author Contributions

Conceptualization, Z.W. (Zhifeng Wang) and. J.F.; methodology, Z.W. (Zhifeng Wang), J.F. and Y.D.; software, J.F. and H.Z.; writing—original draft preparation, Z.W. (Zhifeng Wang) and J.F.; validation, J.F. and P.W.; writing—review and editing, J.F., H.C. and Z.W. (Zetao Wu); supervision, H.C.; funding acquisition, Z.W. (Zhifeng Wang) and Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Macao funding schemes granted by the Science and Technology Development Fund (FDCT) “Research on intelligent learning system & telephoto learning machine and their key technologies in Macao for K-12 students” (0071/2023/RIB3), the Science and Technology Program of Sichuan Province (2022JDR0039), the Basic and Applied Basic Research Foundation of Guangdong Province (2019A1515110304, 2020A1515110461), the Special Project in Key Fields of Universities in Guangdong Province (2022ZDZX3093), and the 2021 General Project of the 14th Five-Year Plan of Philosophy and Social Sciences of Guangdong Province of China (GD21CJY08).

Institutional Review Board Statement

MIMIC-III Waveform Database: The project was approved by the Institutional Review Boards of Beth Israel Deaconess Medical Center (Boston, MA) and the Massachusetts Institute of Technology (Cambridge, MA, USA); Synthetic Dataset: Ethical approval was obtained from the London Westminster Research Ethics Committee (approval number 11/LO/1667).

Informed Consent Statement

Not applicable.

Data Availability Statement

The PPG dataset adopted in this research is openly available in [Physionet] at the MIMIC-III Waveform Database v1.0 (physionet.org) and https://archive.physionet.org/cgi-bin/atm/ATM (accessed on 7 June 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Feigin, V.L.; Brainin, M.; Norrving, B.; Martins, S.; Sacco, R.L.; Hacke, W.; Fisher, M.; Pandian, J.; Lindsay, P. World Stroke Organization (WSO): Global stroke fact sheet 2022. Int. J. Stroke 2022, 17, 18–29. [Google Scholar] [CrossRef] [PubMed]
  2. Chen, G.; Li, L.; Tao, H. Bioinformatics identification of ferroptosis-related biomarkers and therapeutic compounds in ischemic stroke. Front. Neurol. 2021, 12, 745240. [Google Scholar] [CrossRef] [PubMed]
  3. Ding, C.; Guo, Z.; Rudin, C.; Xiao, R.; Shah, A.; Do, D.H.; Lee, R.J.; Clifford, G.; Nahab, F.B.; Hu, X. Learning From Alarms: A Robust Learning Approach for Accurate Photoplethysmography-Based Atrial Fibrillation Detection Using Eight Million Samples Labeled With Imprecise Arrhythmia Alarms. IEEE J. Biomed. Health Inf. 2024, 28, 2650–2661. [Google Scholar] [CrossRef] [PubMed]
  4. Alvarado-Bolaños, A.; Ayan, D.; Khaw, A.V.; Mai, L.M.; Mandzia, J.L.; Bogiatzi, C.; Mrkobrada, M.; Bres-Bullrich, M.; Fleming, L.A.; Lippert, C. Differences in stroke recurrence risk between atrial fibrillation detected on ECG and 14-day cardiac monitoring. Stroke 2023, 54, 2022–2030. [Google Scholar] [CrossRef] [PubMed]
  5. Yasmin, F.; Moeed, A.; Adil, M.; Rahman, S.U.; Wilson, F.P. Accuracy of Apple Watch for Detection of Atrial Fibrillation: A Systematic Review and Meta-Analysis. Circulation 2023, 148, e20362. [Google Scholar] [CrossRef]
  6. Hermans, A.N.L.; Isaksen, J.L.; Gawalko, M.; Pluymaekers, N.A.H.A.; van der Velden, R.M.J.; Snippe, H.; Evens, S.; De Witte, G.; Luermans, J.G.L.M.; Manninger, M.; et al. Accuracy of continuous photoplethysmography-based 1 min mean heart rate assessment during atrial fibrillation. Europace 2023, 25, 835–844. [Google Scholar] [CrossRef]
  7. Sun, Z.; Junttila, J.; Tulppo, M.; Seppanen, T.; Li, X. Non-Contact Atrial Fibrillation Detection From Face Videos by Learning Systolic Peaks. IEEE J. Biomed. Health Inf. 2022, 26, 4587–4598. [Google Scholar] [CrossRef] [PubMed]
  8. Lueken, M.; Gramlich, M.; Leonhardt, S.; Marx, N.; Zink, M.D. Automated Signal Quality Assessment of Single-Lead ECG Recordings for Early Detection of Silent Atrial Fibrillation. Sensors 2023, 23, 5618. [Google Scholar] [CrossRef]
  9. Velraeds, A.; Strik, M.; van der Zande, J.; Fontagne, L.; Haissaguerre, M.; Ploux, S.; Wang, Y.; Bordachar, P. Improving Automatic Smartwatch Electrocardiogram Diagnosis of Atrial Fibrillation by Identifying Regularity within Irregularity. Sensors 2023, 23, 9283. [Google Scholar] [CrossRef]
  10. Sejr, M.H.; May, O.; Damgaard, D.; Sandal, B.F.; Nielsen, J.C. External continuous ECG versus loop recording for atrial fibrillation detection in patients who had a stroke. Heart 2019, 105, 848–854. [Google Scholar] [CrossRef]
  11. Fuster-Barcelo, C.; Guerrero-Lopez, A.; Camara, C.; Peris-Lopez, P. Exploring the power of photoplethysmogram matrix for atrial fibrillation detection with integrated explainability. Eng. Appl. Artif. Intell. 2024, 133, 108325. [Google Scholar] [CrossRef]
  12. Tonko, J.B.; Wright, M.J. Review of the 2020 ESC Guidelines for the Diagnosis and Management of Atrial Fibrillation—What Has Changed and How Does This Affect Daily Practice. J. Clin. Med. 2021, 10, 3922. [Google Scholar] [CrossRef] [PubMed]
  13. Chong, J.W.; Esa, N.; McManus, D.D.; Chon, K.H. Arrhythmia discrimination using a smart phone. IEEE J. Biomed. Health Inf. 2015, 19, 815–824. [Google Scholar] [CrossRef] [PubMed]
  14. El-Hajj, C.; Kyriacou, P.A. A review of machine learning techniques in photoplethysmography for the non-invasive cuff-less measurement of blood pressure. Biomed. Signal Process. Control 2020, 58, 101870. [Google Scholar] [CrossRef]
  15. Jiang, F.; Zhou, Y.; Ling, T.; Zhang, Y.; Zhu, Z. Recent research for unobtrusive atrial fibrillation detection methods based on cardiac dynamics signals: A survey. Sensors 2021, 21, 3814. [Google Scholar] [CrossRef] [PubMed]
  16. Perpetuini, D.; Chiarelli, A.M.; Cardone, D.; Filippini, C.; Rinella, S.; Massimino, S.; Bianco, F.; Bucciarelli, V.; Vinciguerra, V.; Fallica, P. Prediction of state anxiety by machine learning applied to photoplethysmography data. PeerJ 2021, 9, e10448. [Google Scholar] [CrossRef]
  17. Ouzar, Y.; Djeldjli, D.; Bousefsaf, F.; Maaoui, C. X-iPPGNet: A novel one stage deep learning architecture based on depthwise separable convolutions for video-based pulse rate estimation. Comput. Biol. Med. 2023, 154, 106592. [Google Scholar] [CrossRef] [PubMed]
  18. Aldughayfiq, B.; Ashfaq, F.; Jhanjhi, N.Z.; Humayun, M. A Deep Learning Approach for Atrial Fibrillation Classification Using Multi-Feature Time Series Data from ECG and PPG. Diagnostics 2023, 13, 2442. [Google Scholar] [CrossRef]
  19. Wu, Y.; Tang, Q.; Zhan, W.; Li, S.; Chen, Z. Res-BiANet: A Hybrid Deep Learning Model for Arrhythmia Detection Based on PPG Signal. Electronics 2024, 13, 665. [Google Scholar] [CrossRef]
  20. Pereira, T.; Tran, N.; Gadhoumi, K.; Pelter, M.M.; Do, D.H.; Lee, R.J.; Colorado, R.; Meisel, K.; Hu, X. Photoplethysmography based atrial fibrillation detection: A review. NPJ Digit. Med. 2020, 3, 3. [Google Scholar] [CrossRef]
  21. Alam, J.; Khan, M.F.; Khan, M.A.; Singh, R.; Mundazeer, M.; Kumar, P. A Systematic Approach Focused on Machine Learning Models for Exploring the Landscape of Physiological Measurement and Estimation Using Photoplethysmography (PPG). J. Cardiovasc. Transl. Res. 2023, 17, 669–684. [Google Scholar] [CrossRef]
  22. Chen, E.; Jiang, J.; Su, R.; Gao, M.; Zhu, S.; Zhou, J.; Huo, Y. A new smart wristband equipped with an artificial intelligence algorithm to detect atrial fibrillation. Heart Rhythm. 2020, 17, 847–853. [Google Scholar] [CrossRef] [PubMed]
  23. Saarinen, H.J.; Joutsen, A.; Korpi, K.; Halkola, T.; Nurmi, M.; Hernesniemi, J.; Vehkaoja, A. Wrist-worn device combining PPG and ECG can be reliably used for atrial fibrillation detection in an outpatient setting. Front. Cardiovasc. Med. 2023, 10, 1100127. [Google Scholar] [CrossRef] [PubMed]
  24. Kotlyarov, S.; Lyubavin, A. Early Detection of Atrial Fibrillation in Chronic Obstructive Pulmonary Disease Patients. Medicina 2024, 60, 352. [Google Scholar] [CrossRef] [PubMed]
  25. Wu, B.-F.; Wu, B.-J.; Cheng, S.-E.; Sun, Y.; Chung, M.-L. Motion-Robust Atrial Fibrillation Detection Based on Remote-Photoplethysmography. IEEE J. Biomed. Health Inf. 2023, 27, 2705–2716. [Google Scholar] [CrossRef]
  26. Tutuko, B.; Nurmaini, S.; Tondas, A.E.; Rachmatullah, M.N.; Darmawahyuni, A.; Esafri, R.; Firdaus, F.; Sapitri, A.I. AFibNet: An implementation of atrial fibrillation detection with convolutional neural network. BMC Med. Inf. Decis. Mak. 2021, 21, 216. [Google Scholar] [CrossRef] [PubMed]
  27. Şentürk, Ü.; Yücedağ, İ.; Polat, K. Repetitive Neural Network (RNN) Based Blood Pressure Estimation Using PPG and ECG Signals. In Proceedings of the 2018 2nd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT), Ankara, Turkey, 19–21 October 2018; pp. 1–4. [Google Scholar]
  28. Xu, K.; Jiang, X.; Ren, H.; Liu, X.; Chen, W. Deep Recurrent Neural Network for Extracting Pulse Rate Variability from Photoplethysmography During Strenuous Physical Exercise. In Proceedings of the 2019 IEEE Biomedical Circuits and Systems Conference (BioCAS), Nara, Japan, 17–19 October 2019; pp. 1–4. [Google Scholar]
  29. Kumar, A.K.; Ritam, M.; Han, L.; Guo, S.; Chandra, R. Deep learning for predicting respiratory rate from biosignals. Comput. Biol. Med. 2022, 144, 105338. [Google Scholar] [CrossRef]
  30. Lancet Respiratory Medicine, T. Opening the black box of machine learning. Lancet Respir. Med. 2018, 6, 801. [Google Scholar] [CrossRef]
  31. Moody, B.; Moody, G.; Villarroel, M.; Clifford, G.; Silva, I. Mimic-iii Waveform Database, Version 1.0; PhysioNet: Bethesda, MD, USA, 2020; Available online: https://physionet.org (accessed on 7 June 2024).
  32. Gruwez, H.; Ezzat, D.; Van Puyvelde, T.; Dhont, S.; Meekers, E.; Bruckers, L.; Wouters, F.; Kellens, M.; Van Herendael, H.; Rivero-Ayerza, M.; et al. Real-world validation of smartphone-based photoplethysmography for rate and rhythm monitoring in atrial fibrillation. Europace 2024, 26, euae065. [Google Scholar] [CrossRef]
  33. Tang, S.-C.; Huang, P.-W.; Hung, C.-S.; Shan, S.-M.; Lin, Y.-H.; Shieh, J.-S.; Lai, D.-M.; Wu, A.-Y.; Jeng, J.-S. Identification of Atrial Fibrillation by Quantitative Analyses of Fingertip Photoplethysmogram. Sci. Rep. 2017, 7, 45644. [Google Scholar] [CrossRef]
  34. Kim, C.; Stern, R.M. Power-Normalized Cepstral Coefficients (PNCC) for Robust Speech Recognition. IEEE-ACM Trans. Audio Speech Lang. Process. 2016, 24, 1315–1329. [Google Scholar] [CrossRef]
  35. Chen, H.; Guo, C.; Wang, Z.; Wang, J. Research on recognition and classification of pulse signal features based on EPNCC. Sci. Rep. 2022, 12, 6731. [Google Scholar] [CrossRef]
  36. Zheng, Q.; Tian, X.; Yu, Z.; Jin, B.; Jiang, N.; Ding, Y.; Yang, M.; Elhanashi, A.; Saponara, S.; Kpalma, K. Application of complete ensemble empirical mode decomposition based multi-stream informer (CEEMD-MsI) in PM2.5 concentration long-term prediction. Expert Syst. Appl. 2024, 245, 123008. [Google Scholar] [CrossRef]
Figure 1. 10 s PPG recordings after Preprocessing (a) AF Patient (b) Sinus rhythm.
Figure 1. 10 s PPG recordings after Preprocessing (a) AF Patient (b) Sinus rhythm.
Sensors 24 05243 g001
Figure 2. BiLSTM and classification network. (a) Basic unit of BiLSTM; (b) Block diagram of classification network: The input sequence is activated by ReLU and passed into BiLSTM to generate sequence features (X1, X2, …, Xn). These features are extracted through the fully connected layer (FCL) to produce high-level features (a1, a2), which are then converted to probability distributions by the SoftMax layer, and the final output is the classification result (R1, R2).
Figure 2. BiLSTM and classification network. (a) Basic unit of BiLSTM; (b) Block diagram of classification network: The input sequence is activated by ReLU and passed into BiLSTM to generate sequence features (X1, X2, …, Xn). These features are extracted through the fully connected layer (FCL) to produce high-level features (a1, a2), which are then converted to probability distributions by the SoftMax layer, and the final output is the classification result (R1, R2).
Sensors 24 05243 g002
Figure 3. Characteristic Matrix Obtained by CEPNCC. (A) AF patient (B) Sinus rhythm.
Figure 3. Characteristic Matrix Obtained by CEPNCC. (A) AF patient (B) Sinus rhythm.
Sensors 24 05243 g003
Figure 4. Relationship between Different Learning Rates and Loss.
Figure 4. Relationship between Different Learning Rates and Loss.
Sensors 24 05243 g004
Figure 5. Relationship between Different Learning Rates and Learning Rate Drop Factors and Accuracy.
Figure 5. Relationship between Different Learning Rates and Learning Rate Drop Factors and Accuracy.
Sensors 24 05243 g005
Figure 6. Multi-term, CEPNCC AF Classification Confusion Matrix.
Figure 6. Multi-term, CEPNCC AF Classification Confusion Matrix.
Sensors 24 05243 g006
Table 1. Diagnosis Distributions in the Training and Test Sets.
Table 1. Diagnosis Distributions in the Training and Test Sets.
Signal CategoryTraining SetTest Set
PPG from AF patients75%46%
PPG from AF patients with other diseases5%14%
Sinus rhythm PPG20%40%
Table 2. Feature Extraction Time and Network Training Time.
Table 2. Feature Extraction Time and Network Training Time.
DataFeature Extraction Time
(Per Data)
Network Training TimeAverage Accuracy (%)ET-Score
1 min PPGAbout 3 s1 min 16 s
(Only 400 data items)
96%96.25
2 min PPGAbout 5 s98%96.80
5 min PPGAbout 10 s98%92.50
Table 3. Processing of Two Methods.
Table 3. Processing of Two Methods.
MethodRequired Processing Times (Times)Average Processing Time (s)
CEEMD205.8291
EEMD20036.1267
Table 4. CEPNCC-BiLSTM Classification Recognition Rate.
Table 4. CEPNCC-BiLSTM Classification Recognition Rate.
CategoryRecall Rate (%)Precision (%)
AF Patient10098.7
Sinus rhythm98100
Table 5. Recall Rate of Each Training Category.
Table 5. Recall Rate of Each Training Category.
CategoryEPNCC-BiLSTM (%)CEPNCC-BiLSTM (%)SVM (%)EPNCC-SVM (%)CEPNCC-SVM (%)
PPG recordingsAFNSRAFNSRAFNSRAFNSRAFNSR
Recall rate93.510098.71007410088.810092.999.4
Precision89.595.671009810071.91008199.788.5
Table 6. Training Time and Preprocessing Time of Each Training Category.
Table 6. Training Time and Preprocessing Time of Each Training Category.
Average Time (1500-Data)EPNCC-BiLSTMCEPNCC-BiLSTMInput PPG Recordings to SVMEPNCC-SVMCEPNCC-SVM
Preprocessing15 h2.25 h015 h2.25 h
Training3 min 16 s3 min 10 s1 min 7 sAbout 5 sAbout 5 s
Table 7. F-measure and ET-score of each training category.
Table 7. F-measure and ET-score of each training category.
CategoryEPNCC-BiLSTMCEPNCC-BiLSTMSVMEPNCC-SVMCEPNCC-SVM
F-measureAF Patient91.46%100%85.06%93.62%96.18%
Sinus rhythm97.79%98.34%83.65%89.50%93.63%
ET-score82.2897.2784.3586.1092.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Fan, J.; Dai, Y.; Zheng, H.; Wang, P.; Chen, H.; Wu, Z. Intelligent Detection Method of Atrial Fibrillation by CEPNCC-BiLSTM Based on Long-Term Photoplethysmography Data. Sensors 2024, 24, 5243. https://doi.org/10.3390/s24165243

AMA Style

Wang Z, Fan J, Dai Y, Zheng H, Wang P, Chen H, Wu Z. Intelligent Detection Method of Atrial Fibrillation by CEPNCC-BiLSTM Based on Long-Term Photoplethysmography Data. Sensors. 2024; 24(16):5243. https://doi.org/10.3390/s24165243

Chicago/Turabian Style

Wang, Zhifeng, Jinwei Fan, Yi Dai, Huannan Zheng, Peizhou Wang, Haichu Chen, and Zetao Wu. 2024. "Intelligent Detection Method of Atrial Fibrillation by CEPNCC-BiLSTM Based on Long-Term Photoplethysmography Data" Sensors 24, no. 16: 5243. https://doi.org/10.3390/s24165243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop