Next Article in Journal
Towards Automated Model Selection for Wind Speed and Solar Irradiance Forecasting
Previous Article in Journal
High-Resolution Rainfall Estimation Using Ensemble Learning Techniques and Multisensor Data Integration
Previous Article in Special Issue
An Intelligent Hand-Assisted Diagnosis System Based on Information Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HRP-OG: Online Learning with Generative Feature Replay for Hypertension Risk Prediction in a Nonstationary Environment

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
Beijing International Collaboration Base on Brain Informatics and Wisdom Services, Beijing University of Technology, Beijing 100124, China
3
Beijing Key Laboratory of MRI and Brain Informatics, Beijing University of Technology, Beijing 100124, China
4
Engineering Research Center of Intelligent Perception and Autonomous Control, Beijing 100124, China
5
Engineering Research Center of Digital Community, Ministry of Education, Beijing 100124, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(15), 5033; https://doi.org/10.3390/s24155033
Submission received: 4 July 2024 / Revised: 27 July 2024 / Accepted: 1 August 2024 / Published: 3 August 2024
(This article belongs to the Special Issue Artificial Intelligence for Medical Sensing)

Abstract

:
Hypertension is a major risk factor for many serious diseases. With the aging population and lifestyle changes, the incidence of hypertension continues to rise, imposing a significant medical cost burden on patients and severely affecting their quality of life. Early intervention can greatly reduce the prevalence of hypertension. Research on hypertension early warning models based on electronic health records (EHRs) is an important and effective method for achieving early hypertension warning. However, limited by the scarcity and imbalance of multivisit records, and the nonstationary characteristics of hypertension features, it is difficult to predict the probability of hypertension prevalence in a patient effectively. Therefore, this study proposes an online hypertension monitoring model (HRP-OG) based on reinforcement learning and generative feature replay. It transforms the hypertension prediction problem into a sequential decision problem, achieving risk prediction of hypertension for patients using multivisit records. Sensors embedded in medical devices and wearables continuously capture real-time physiological data such as blood pressure, heart rate, and activity levels, which are integrated into the EHR. The fit between the samples generated by the generator and the real visit data is evaluated using maximum likelihood estimation, which can reduce the adversarial discrepancy between the feature space of hypertension and incoming incremental data, and the model is updated online based on real-time data using generative feature replay. The incorporation of sensor data ensures that the model adapts dynamically to changes in the condition of patients, facilitating timely interventions. In this study, the publicly available MIMIC-III data are used for validation, and the experimental results demonstrate that compared to existing advanced methods, HRP-OG can effectively improve the accuracy of hypertension risk prediction for few-shot multivisit record in nonstationary environments.

1. Introduction

Hypertension is a leading cause of early demise and debility globally [1]. It is responsible for around 7.5 million deaths per year, accounting for 12.8% of all deaths worldwide and about 18% of global deaths from cardiovascular disease [2]. Timely detection of the preclinical or high-risk stage of hypertension, namely, hypertension risk prediction, is very important. Clinical studies have shown that for individuals in the preclinical or at-risk stage, significant reductions in the progression of hypertension can be achieved through lifestyle changes or effective drug treatments [3]. However, due to limited medical resources, relying solely on healthcare workers to achieve hypertension risk prediction is not an easy task. Related research shows that about 1.13 billion people in the world are troubled by elevated blood pressure, but only about 50% are aware that they are hypertensive [4]. Hypertension is a gradual process, and specific symptoms may manifest before the onset of the disease. In this case, by analyzing EHRs, medical AI is able to understand the physical condition of patients to determine their future risk of developing hypertension [5]. Therefore, hypertension risk prediction assisted by artificial intelligence (AI) has become an urgent need and is one of the current research hotspots in medical AI.
Electronic health records (EHRs) are digital records that contain complete personal and health information of patients along with their medical history [6]. With the rapid development of EHRs, various medical AI applications based on EHRs have received widespread attention, including disease risk prediction systems [7,8,9], clinical knowledge question answering systems [10,11,12], drug recommendation systems [13,14], etc. Chronic disease risk prediction based on EHRs is an important way to achieve AI-assisted chronic disease warning, including hypertension risk prediction. Existing studies focus on developing complex models [15,16], especially deep neural networks, to extract feature information from multivisit EHR data for accurately predicting chronic disease risk. They depend on large-scale and high-quality multivisit EHRs.
However, although hospitals worldwide have established sophisticated hospital in-formation systems, there are still differences in the storage methods and the data quality of EHRs between different hospitals [17]. At the same time, there are also challenges such as data ownership and privacy protection [18]. Therefore, it is difficult to integrate long-term data from multiple hospitals at once, to build a large-scale and high-quality multivisit EHRs dataset. In practical settings, chronic disease risk prediction modeling based on EHRs is often conducted in a few-shot and nonstationary environment. Few-shot EHR data are gradually generated and collected in batches. There are also significant differences in the distribution of data attributes for each batch. Figure 1 shows the statistics of EHR data of hypertensive patients from two hospitals in Hainan Province, China, demonstrating the distribution of the age of patients examined in each of the 5 years. It can be seen that the age characteristics of patients continuously change over five years. In subplot a of Figure 1, the age of patients ranged from 67.5 to 72.5 in 2014, increased in 2015, and then dropped significantly to 56–67 in 2016, before gradually rising again to 66–73. In subplot b of Figure 1, the age of patients consistently decreased from 70.6–75.8 to 66.5–73. In both datasets, the age distribution of patients is different in each year, and such changes lead to a fixed hypertension prediction model being unable to meet the actual needs of the patients.
Such a few-shot and nonstationary environment leads to adversarial differences between the feature space of disease prediction models and the continuously inputted incremental EHR data, which brings about the need for online learning of the model. However, the factors affecting hypertension risk are often multifaceted, requiring high-dimensional data for prediction [19]. Existing online learning methods struggle with learning high-dimensional data features. For examples, Ning et al. [20] proposed a robust algorithm using polar decomposition for derivation, transforming the learning problems into optimal feedback control problems for a group of low-dimensional linear systems and using optimal control to achieve learning from high-dimensional online data. However, such complex deep learning models tend to overfit to noisy data when the data are small [21]. When data distribution changes significantly, excessive adaptation to new data will also lead to a decrease in the generalization ability of the model.
Based on the above observations, this paper proposes a new hypertension risk prediction model, called HRP-OG, which combines reinforcement learning and generative feature replay techniques to predict hypertension risk based on multivisit EHR data. The model can be updated with incoming incremental data in a few-shot and nonstationary environment, achieving more accurate prediction of hypertension risk. The main contributions are as follows:
  • This paper transforms the hypertension risk prediction problem into a Markov decision process problem and proposes a hypertension monitoring model based on reinforcement learning. A new dual experience replay strategy is proposed to further distinguish the high and low risk experiences for reducing the bias caused by the imbalance distribution of multivisit EHRs and improving the few-shot learning capability of the model.
  • A new generator updating strategy is proposed. By comparing the fitting degree between the incremental data and historical data with the hypertension feature space, the generator adopts different updating modes for reducing the adversarial differences between the incremental EHR data and the feature space.
  • This paper constructs a 12-year continuous hypertension risk prediction dataset from the public MIMIC-III dataset to simulate a few-shot, incremental, and nonstationary EHR generation scenario. Experiments are conducted on this dataset and the results demonstrate that the proposed HRP-OG can provide effective and continuously updating hypertension risk prediction in a few-shot and nonstationary environment.
The rest of this paper is organized as follows. Section 2 reviews machine learning methods for chronic disease prediction. Section 3 introduces the proposed HRP-OG model. Section 4 describes the datasets and compares the performance of this model with the baselines. In Section 5, the performance of this model is compared and analyzed with the variants. Finally, Section 6 gives the conclusions and future work.

2. Related Work

Hypertension risk prediction is very important and effective for reducing the prevalence of hypertension. Traditional hypertension risk prediction is mostly achieved using regression models [22]. These methods depend on high-quality medical resources and have difficulty meeting the needs of a large population at high risk of hypertension. Medical AI provides an effective way to alleviate the pressure on medical resources for chronic disease (e.g., hypertension) risk prediction. Its typical application is EHR-based chronic disease risk prediction.
EHR-based chronic disease risk prediction refers to predicting future disease risk by analyzing previously observed EHR information [23,24]. Early researchers focused only on the current health status of patients and tended to model single-visit EHR data for predicting chronic disease risks. For example, Syed et al. [25] integrated neural network and genetic algorithm to predict the probability of heart disease based on 12 factors, including family history, diabetes, hypertension, smoking status, obesity, etc. They used the genetic algorithm for a global search to optimize network initialization weights. Sayali et al. [26] combined Bayesian and K-nearest neighbor (KNN) algorithms to predict whether a patient has heart disease based on symptoms. Based on this, they proposed a Convolutional Neural Networks-based Uni-Modal Disease Risk Prediction (CNN-UDRP) algorithm that uses a convolutional neural network to further predict the disease risk of the patient.
However, the development of chronic diseases is a long-term process, and it is difficult to obtain more accurate prediction results by only focusing on the current health status of patients. Therefore, more and more researchers no longer focus solely on the horizontal features of EHRs but begin to combine the current and historical EHRs of patients to extract longitudinal EHR features [27,28]. Multivisit EHRs are used to predict chronic disease risks. For example, Gao et al. [29] used a phase-aware neural network to extract disease stage information from the time series of EHR data for health risk prediction. Zhao et al. [30] developed a model to extract and utilize the uncertainty between multiple visit intervals of EHRs for improving disease prediction. Nancy et al. [31] integrated recursive neural network and bidirectional long short-term memory (Bi-LSTM) to manage continuous EHR time series and build a model for predicting heart disease risks. Most of these studies adopted complex deep learning models, which depended on large-scale and high-quality multivisit EHR data.
In practical settings, it is difficult to obtain large-scale and high-quality multivisit EHR data at once. Therefore, online learning of models is gradually gaining attention for EHR-based chronic disease risk prediction. Online learning is a particular incremental learning method [32] that refers to the paradigm where a model is updated iteratively and adaptively over time as new data become available in a sequential manner [33]. For chronic disease risk prediction, the main methods of online learning can be broadly classified into three categories: regularization-based methods, dynamic network structure-based methods, and replay-based methods [34]. Regularization-based methods [35] mitigate catastrophic forgetting by imposing constraints to enable the updating of network parameters. However, because of penalizing changes to important parameters, the regularization term may lead to poor adaptability of the model, making it difficult to adapt to new tasks with different data distributions and features. Dynamic network structure-based methods [36] dynamically adapt the network to new tasks. However, due to the continuous complexity of the model structure, this kind of method is unstable during the model update process and is prone to overfitting on few-shot data. Therefore, these two kinds of online learning methods are not suitable for hypertension risk prediction in a few-shot and nonstationary environment.
Replay-based methods rely on memory replay, meaning that samples from previous tasks are replayed when the samples from the current task are used for learning [37]. It can effectively retain historical samples while not requiring adjustments to the model structure. Therefore, it does not have the aforementioned drawbacks and is currently a research hotspot in online learning for chronic disease prediction models. For example, Wang et al. [38] proposed a novel method named embedding episodic memory and consolidation to prevent catastrophic forgetting on disease diagnosis tasks. This method retrieved a replay batch from the memory set and leveraged medical domain knowledge, incorporating both context and medical entity features to transfer knowledge to new stages. Gao et al. [39] proposed a brain disease prediction method that combines a multiloop learning algorithm with a generative adversarial network. The method combines evidence with better sample contribution ranking during training processes, selecting 1/2 of the samples each time for learning and generating samples to input into the database. In chronic disease risk prediction, EHR data are characterized by continuous changes and interconnections. Combining historical data with current data can achieve better predictions, making the replay-based methods highly suitable.
However, when there is a significant difference in the feature distribution between new data and historical data, the accuracy of replay-based methods will be poor. The physiological characteristics of high-risk individuals for hypertension are influenced by various factors such as climate and psychology. Therefore, there are significant differences in the distribution of attributes between different batches of data in few-shot and nonstationary environments. In this case, using existing replay-based methods for online learning of the model may cause differences between the generated samples and the feature space of hypertension, thereby affecting the accuracy of the model. Therefore, it is necessary to innovate existing replay-based methods to achieve online learning of the hypertension risk prediction model in the few-shot and nonstationary environment.

3. Methods

3.1. The Model Framework of HRP-OG

This study proposes an online hypertension risk monitoring model based on reinforcement learning and generative feature replay. The overall framework consists of three parts: the input representation module, the hypertension risk prediction module, and the model online updating module. The specific model framework is illustrated in Figure 2.

3.2. The Input Representation Module

The input representation module converts each visit record into a multidimensional input embedding that serves as the input to subsequent modules. In EHR data, each visit record is represented as P n = { A n , I S H Y P E R n } , where n represents the n-th patient and IS-HYPER represents the sample tag. A represents the input embedding of record containing the above 20 attributes, which can be expressed as follows:
A = | | i M i t e m i
where i t e m [ i ] is the input embedding corresponding to the i-th attribute of the visit record, | M | represents the number of attributes, and | | represents the concatenation of i t e m 1 , i t e m 2 i t e m | M | .

3.3. The Risk Prediction Module

The hypertension risk prediction module pools all attribute embedding sequences transformed from longitudinal EHR data, which are based on the reinforcement learning framework, to perform the hypertension risk warning task and record the changes in prediction states for patients. In addition, the unique reward, penalty functions, and sample selection strategies can further improve prediction accuracy.
This study abstracts the hypertension risk prediction module as an agent, in which the visit record of each patient serves as a state and all EHR data as the environment. The choice of predicting hypertension risk for patients is considered as an action, and the accuracy of hypertension risk prediction is the reward. Ultimately, the agent provides hypertension risk predictions based on the current environment.
At the beginning, we set the state of the patient as 0 vector. At time t, we update the state vector of the patient by connecting the historical visit records with the current visit record of the patient; the state can be calculated as follows:
s t a t e t n = i = 0 M i t e m i i n | | 1 t 1 i t 1 P i n
The agent will predict the hypertension risk of the patient. The actions of the agent are divided into two categories, which can be expressed as follows:
a t = 1       i s _ h y p e r 0       n o _ h y p e r
After the agent performs the action, the environment evaluates whether the hypertension risk prediction for the patient is correct and provides a reward signal to the agent. The reward function consists of two parts, which can be expressed as follows:
r e w a r d = r e w a r d 1 + r e w a r d 2
where reward1 represents the judgement of whether the hypertension risk prediction is correct, and it can be calculated as follows:
r e w a r d 1 = 1   i f   a ! = l a b l e 1   i f   a = = l a b l e
where label represents the sample tag of visit record. reward2 represents the reciprocal of the distance between the predicted and true results, and it can be calculated as follows:
r e w a r d 2 = S i g m o i d 1 l e f t i s _ h y p e r 2 ( r i g h t n o _ h y p e r ) 2 2
where left and right values are obtained through the DQN network and represent the probabilities of the patient having hypertension and not having hypertension, respectively. The agent predicts the risk of hypertension by comparing these probabilities and selecting the larger one.
This study uses a DQN network to receive the current visit data status of the patient transmitted by the agent. The DQN stores the current health status of the patient, predicted disease condition, reward value, and the last health status in a dual replay buffer. The dual replay buffer stores data in the corresponding buffer based on whether the current versus the last visit has changed. The main network is trained by sampling data samples from the buffer according to the sampling strategy, aiming to fit the real hypertension condition as closely as possible after training. Finally, the loss function is calculated based on the Q-values of the main network and the target network, and the network parameters are updated using historical data.
To expedite the model training process, we design a time-based priority allocation strategy and a priority-based sampling strategy. In the allocation strategy, the DQN network calculates the difference between the estimated value of the action and the current function output value. The priority increases as the error grows larger. The gap value can be calculated as follows:
g a p s t , a t , r t , s t + 1 = r t + y m a x a t Q s t + 1 , a t + 1 Q ( s t , a t )
where y represents the discount factor, y m a x a t Q s t + 1 , a t + 1 represents the target value, and Q ( s t , a t ) represents the current value.
After calculating the priority scores of the samples, the model uses the softmax function to map the gap values, converting the priority scores of the data samples into probabilities of being selected from the buffer. The value can be calculated as follows:
P ( n ) = e g a p n i = 1 n e g a p i
where P ( n ) represents the probability of the n-th data sample being selected from the buffer. After assigning sampling probabilities to each sample in the buffer, the DQN network selects a batch of data samples for learning based on these probabilities. To better extract sample data, we propose a dual experience replay strategy, establishing two buffers to store the samples predicted correctly and incorrectly. We define the loss function as follows:
L = φ 1 t P a = = l a b l e g a p s t , a t , r t , s t + 1 2 + φ 2 t P a ! = l a b l e g a p s t , a t , r t , s t + 1 2
where φ 1 and φ 2 represent the weights used to select data samples from the double experience buffer. When selecting data samples from the first buffer, we set φ 1 = 1 and φ 2 = 0 . Conversely, when selecting data samples from the second buffer, we set φ 1 = 0 and φ 2 = 1 . t P a = = l a b l e represents the buffer where data samples have the same action values and label values, while t P a ! = l a b l e represents the buffer where data samples have different action values and label values.

3.4. The Model Updating Module

The model online updating module uses maximum likelihood estimation to evaluate the distance between old data features and new data features. By utilizing a generator, discriminator, and replay alignment strategy, the generalization ability of the model is enhanced.
When the t-th batch of data arrives, the model does not directly feed the data into the hypertension risk warning module for learning. Instead, it generates a new generator G t to learn the feature distribution of the new data. At the same time, it calls the old generator that has been updated t − 1 times. Both generators generate sample data to be fitted to the real data. In order to evaluate the fitting degree of the generated samples to real data, we propose a maximum likelihood estimation and the negative-log-likelihood (NLL) function, as follows:
N L L = 1 N t i = 1 N l o g G θ ( P f a t e | P i )
where N t represents the number of samples arriving at the t-th time, P f a t e represents the generated data samples by the generator, and P i represents the medical examination data of the i-th patient in the data arriving at the t-th time. If the condition N L L n e w < N L L t 1 is met, the data will enter the online updating framework of the model; otherwise, the model will directly learn from the new data.
This study designs an experience replay strategy similar to generative adversarial networks. It integrates generated data samples with new visit data to form a fusion dataset. In the generative adversarial networks, the generator attempts to capture the distribution of EHR data from the patients, while the discriminator tries to assess the probability that input samples come from real data rather than the generator.
In order to enable the model to use the features learned from previous tasks to train the new tasks, we propose a feature distillation strategy to extract features generated from previous tasks. Before starting a new task, we freeze the feature extractor and create a new identical feature extractor. After the two feature extractors perform feature extraction separately, the loss function is computed between them to prevent forgetting. The loss function can be calculated as follows:
L F t = 1 N r e c R E C t | | F t ( r e c ) F t 1 ( r e c ) | | 2
where F t represents the t-th feature extractor, R E C t represents the collection of multivisit EHR for patients in task t, and rec represents the multivisit EHR of the patient.
In this study, the generator consists of three linear layers, and each is followed by an activation function. It concatenates random noise vectors and is-hyper vectors to form the input vectors, which are then passed through the network composed of linear layers and activation functions to generate normalized output features. The loss function of the generator can be calculated as follows:
L G t = E z ~ q z , i s h y p e r l a b e l [ D t ( i s h y p e r , G t i s h y p e r , z ) ]
where G t represents the generator incrementally generated at the t-th time, and q z is a randomly generated noise variable, which is input together with the sample label into the generator to obtain the generated data sample. D t represents the discriminator incrementally generated at the t-th time. The discriminator calculates the probability of the sample coming from real data or the generator by, respectively, receiving the generated data from G t and the visit record data from the patient. It consists of two linear layers, and each is followed by an activation function. It receives real data and generated data, extracts hidden features from the data, and calculates the real probability and class classification probability of the generated data through validity scores. Through adversarial learning, both the generator and discriminator are alternately updated to continuously improve the quality of the generated data. The loss function of the discriminator can be calculated as follows:
L D t = E z ~ q z , i s h y p e r l a b e l [ D t ( i s h y p e r , G t i s h y p e r , z ) ] E [ D t ( i s h y p e r , F t r e c ) ]
By training the discriminator G t to maximize the probability of discriminating between samples from real data and training samples, we achieve alternating updates of the generator and discriminator in the t-th incremental process in an adversarial learning manner.
In order to avoid model overfitting on new data when using old knowledge to predict new data, we design a replay alignment mechanism. This mechanism generates a replay generator G s with the exact same network structure as the current generator G t . Before the start of the t-th incremental task, the noise variables, parameter space, and conditional space of the two generators are aligned.
During the training of the generator, EHR data features are generated by inputting the same hypertension category and noise variables as the generator. The loss function is minimized to align the medical examination record features of the current generator as closely as possible with those of the replay generator, thereby preventing the hypertension monitoring model from overfitting to the training data completely. The loss function of the replay alignment mechanism can be calculated as follows:
L R A = E z ~ q z , i s h y p e r l a b e l [ | | G t i s h y p e r G s i s h y p e r , z | | 2 2 ]

3.5. The Pseudo-Code of Model

Algorithm 1 presents the core algorithm for the proposed model. At the beginning of the algorithm, the feature extractor and generator are initialized. As the segmented dataset is continuously fed into the algorithm, the generator, discriminator, and feature extractor are continuously updated. Ultimately, the agent will predict the hypertension risk of the patient and produce the output.
Algorithm 1: Online Learning with Generative Feature Replay for Hypertension Risk Prediction.
 1:Input: Sequence D 1 , . . , D T
 2:Require: Feature extractor F 0 , Generator G 0
 3:for t = 1, …, T
 4:  if t == 1 then
 5:    train F 1 F 0   w i t h   D 1
 6:    train G 1 G 0   w i t h   F 1
 7:  else
 8:    If  N L L t N L L t 1 then
 9:       t r a i n   F t F t 1     w i t h     D t
10:       t r a i n   G t G t 1   w i t h     F t
11:    else
12:      set G t = G t 1
13:  Repeat
14:     s t a t e t n i t e m i i n   a n d   P i n
15:    choose action a t and provide reword
16:     g a p r e c o r d + Q s t + 1 , a t + 1 Q ( s t , a t )  
17:    put sample into replay buffer
18:     L R A G t G s
19:  end Repeat
20:end for

4. Experiments

4.1. Dataset

This study uses personal physical examination records obtained from the public dataset MIMIC-III as the dataset for our experiments. The MIMIC-III dataset is widely used and contains clinical data from tens of thousands of patients spanning from 2001 to 2012. According to the 2023 edition of the Chinese Guidelines for Prevention and Treatment of Hypertension, hypertension is associated with various risk factors such as “daily blood pressure fluctuations”, “abnormal heart rate”, “obesity”, “irregular lifestyle”, and “age”. Based on these risk factors, 20 attributes were chosen from the MIMIC-III dataset and all attributes are listed in Table 1. The attribute “age” corresponds to the risk factor “age”. The attributes “heart rate”, “respiratory rate”, “heart rhythm”, “low heart rate (LHR)”, “high heart rate (HHR)”, “low respiratory rate (LRR)”, “high respiratory rate (HRR)”, and “respiratory pattern” correspond to the risk factor “abnormal heart rate”. The attributes “low blood oxygen saturation (LBOS)”, “high blood oxygen saturation (HBOS)”, “systolic blood pressure (SBP)”, “diastolic blood pressure (DBP)”, “mean arterial blood pressure (MBP)”, “central venous blood pressure (CVBP)”, “low arterial blood pressure (LABP)”, and “high arterial blood pressure (HABP)” correspond to the risk factor “daily blood pressure fluctuations”. The attribute “life pattern” corresponds to the risk factor “irregular lifestyle” and the attribute “weight” corresponds to the risk factor “obesity”. In addition, these attributes and “temperature” reflect the basic lifestyle and physical state of the patient.
This study uses a multivisit dataset based on the number of personal physical examination records. Table 2 shows the statistics of the multivisit samples; as the number of visits increases, the number of patients decreases. Clearly, the availability of multivisit samples for patients each year remains scarce.
In order to simulate a few-shot, incremental, and nonstationary EHR generation scenario, this study divides the MIMIC-III dataset into 12 batches of continuous dataset based on the year of data generation. Table 3 shows the number of patients and the number of records for each of the 12 years. It can be seen that the number of EHRs available per year does not exceed 610, and records of multivisit are scarce. Taking age, diastolic blood pressure, systolic blood pressure, and weight as examples, the distribution of the characteristics of patients examined in each of the 12 years is shown in Figure 3. It can be observed that regardless of the examination value, there is a consistent range of distribution. However, the examination values of most patients still undergo changes, which are irregular and present in each batch of data. Therefore, a fixed hypertension prediction model is insufficient to meet the needs of patients in real-world scenarios. And an adaptive updating model that can adapt to continuous data streams is still of significant research value.

4.2. Baselines

In this study, the proposed model H R P O G is compared with the following baselines. All methods are implemented in Pytorch and trained on an NVIDIA GeForce RTX 3070Ti. Their source codes are available in original papers:
(1)
LSTM [40] is a predictive model based on sequential data that can handle irregular time intervals in patient examination records. In the experiment, the learning rate is set to 1 × 10−3, the hidden layer size is set to 100, and the number of hidden layers is set to 2.
(2)
KNN-SVM [41] is a model that utilizes grid search technique for hyperparameter tuning and can handle imbalanced data for prediction. In the experiment, the regularization parameter is set to 1, and the kernel function is set to radial basis function.
(3)
ConCare [42] is a multichannel longitudinal disease risk prediction model. It adopts improved multihead self-attention to extract the interdependencies among time series of dynamic features as well as static baseline information for disease prediction. In the experiment, the learning rate was set to 1 × 10−3, the input feature dimension was 76, and the number of heads in the multihead attention mechanism was set to 4.
(4)
LightGBM [43] is a gradient boosting framework that builds multiple decision trees and learns from the gradients to handle few-shot learning. In the experiment, the learning rate is set to 0.1, the boosting type is set to gradient boosting decision tree, the number of trees built is set to 10, and the maximum number of leaf nodes is set to 31.
(5)
KNN-LightGBM [44] is a disease risk prediction model based on deep learning. It predicts hypertension using a hybrid deep neural network combining KNN and LightGBM based on single-instance EHR. In this experiment, the number of nearest neighbors is set to 5, the learning rate is set to 1 × 10−3, and the distance metric used Minkowski.
(6)
SACNN-SOA [45] is a chronic disease diagnosis based on the Self-Attention Convolutional Neural Network optimized with Season Optimization Algorithm, which uses IoT and cloud computing in the Smart Medical Big Data health care system. In this experiment, the learning rate is set to 1 × 10−5, batch size is set to 20, and the number of convolutional layers is 9.

4.3. Metrics

This paper uses the accuracy, precision recall AUC (PR-AUC), and ROC-AUC to measure experimental results.
Accuracy quantifies the proportion of correct predictions in the overall test set, which can demonstrate the reliability of the model’s prediction results. It can be calculated as follows:
A c c u r a c y = T N + T P T N + T P + F N + F P
where TP represents the true positive, FP represents the false positive, TN represents true negative, and FN represents the false negative.
Precision, recall, and FPR can be calculated as follows:
P r e c i s i o n = T P T P + F P ,   R e c a l l = T P T P + F N ,   F P R = F P T N + F P
PR-AUC is the area under the precision–recall curve, which can provide a comprehensive model evaluation that combines precision rate and recall rate. In the case of imbalanced data, PR-AUC can effectively reflect the performance of the model. It can be calculated as follows:
P R A U C = 0 1 P r e c i s i o n ( R e c a l l ) d R e c a l l
ROC-AUC is the area under the ROC (receiver operating characteristic) curve, which can also provide a comprehensive model evaluation. Unlike PR-AUC, which focuses on the accuracy of the model in predicting positive samples, ROC-AUC is suitable for scenarios where the model is expected to perform well on both positive and negative samples (e.g., reducing the misdiagnosis rate). Therefore, it can provide a relatively robust evaluation for imbalanced datasets. ROC-AUC can be calculated as follows:
R O C A U C = 0 1 R e c a l l ( F P R ) d F P R

4.4. Results

In the experiments, the multivisit records are divided into 12 batches by year to simulate the scenario of incremental data in a nonstationary environment. Each batch of data is randomly split into training set, test set, and validation set in a ratio of 3:1:1. The results of 5-fold cross-validation are shown in Table 4 and Table 5, which present the accuracy and PR-AUC values of different models, respectively. We recorded the performance changes of the models over 12 stages and calculated the average performance over these 12 stages. Despite fluctuations in the data, all models showed performance variations within a certain range, indicating differences in their ability to learn hypertension features.
In all the experiments, SACNN-SOA showed the worst performance; both the accuracy value and the PR-AUC value were the lowest among all the models. The reason for this is related to the characteristics of the SACNN-SOA model; as described in its paper, the model is based on big data and cloud computing platforms, and requires a large number of data samples for training. However, due to the limited number of samples in our experiments, this model could not accurately capture the feature distribution of hypertensive patients.
In contrast, LightGBM performed excellently in several comparative experiments, achieving accuracy and PR-AUC values of 0.6086 and 0.8043, respectively, indicating its ability to effectively learn hypertension features from multivisit records.
The KNN-LightGBM hybrid prediction model, which combines LightGBM and KNN, is expected to perform better than LightGBM alone according to its paper. However, in our experiment, KNN-LightGBM and LightGBM have similar predictive performance. Comparing the accuracy and PR-AUC indicators across several experiments, it can be seen that KNN-LightGBM achieved better performance in the seventh, eighth, and tenth stages, while LightGBM performed better in the second, fourth, and sixth stages. Overall, the accuracy and PR-AUC of LightGBM were slightly higher than those of KNN-LightGBM. The reason why each of the two models can outperform the other in certain stages is that KNN and LightGBM are trained simultaneously, and their final results are weighted averages. In some data streams, the weights of the two models cannot be effectively allocated, preventing the model from achieving better performance.
KNN-SVM is also a predictive model that combines KNN. In our experiment, the overall performance of the model is slightly better than KNN-LightGBM, and the model achieves better predictive results in the first, ninth, eleventh, and twelfth stages. The reason is that this model uses a more complex ensemble learning method, which can better integrate multiple modules.
LSTM does not perform well on either indicator, possibly because the model is relatively simple and cannot effectively capture long sequence features from few-shot EHR samples.
The ConCare model performs well on the PR-AUC indicator but poorly on the accuracy indicator, suggesting that while the model does not predict overall physical examination data well, it is better at identifying hypertensive patients, failing to effectively recognize nonhypertensive samples.
Finally, the model we proposed achieves the best results across continuous testing of 12 data batches. In terms of the accuracy indicator, it outperformed other models by more than 20% at each stage. In terms of the PR-AUC indicator, it outperformed other models by more than 10% at each stage, and the average performance on both indicators is far superior to that of other models, demonstrating the effectiveness in learning data features from a few-shot and nonstationary environment.

5. Discussion

This study conducts ablation experiments on our model H R M A R , and three H R M A R variants models are designed as follows:
  • H R P O G r e t r a i n . Based on H R P O G , this variant directly abandons the knowledge learned from historical data, instead learning from new data and making predictions. This variant is used to justify the need for model updates.
  • H R P O G G . Based on H R P O G , the online updating module is removed. The reinforcement learning module is directly used for training and testing. This variant is used to demonstrate the necessity and effectiveness of the generative feature replay.
  • H R P O G N . Based on H R P O G , the data evaluation module is removed. The reinforcement learning module and generative feature replay module are directly used for training and testing. This variant is used to demonstrate the necessity of the sample evaluation module.
Figure 4 and Figure 5 show the results of the ablation study, displaying the performance of several models evaluated based on the accuracy and PR-AUC metrics, respectively. The performance of the models demonstrated by both metrics is similar. The result of H R P O G G is the worst, even worse than most of the instance-based baseline methods shown in Table 4 and Table 5. This indicates that using past models directly cannot make good predictions on new data, as the old models cannot effectively capture the features of hypertension patients on few-shot datasets. The result of H R P O G r e t r a i n is close to most of the instance-based baseline methods shown in Table 4 and Table 5, indicating that the ability of the model to capture features on few-shot datasets is relatively close. They can predict the hypertension risk, but the accuracy of the predictions could not be compared to the models incorporating historical visit records. This underscores the necessity of online updates to the model in response to changes in the data. The result of H R P O G N are comparable to H R P O G , and even exceed it in some stages. This indicates that the model can effectively extract feature distributions from the data. However, in few-shot datasets, there are biases in the data features. Thus, continuous effective model updates are impossible without assessing data quality. This highlights the necessity of sample selection based on maximum likelihood estimation.
Overall, in terms of the accuracy indicator, H R P O G improves by 7% compared to H R P O G r e t r a i n , by 41% compared to H R P O G G , and by 4% compared to H R P O G N . In terms of the PR-AUC indicator, H R P O G improves by 14% compared to H R P O G r e t r a i n , by 25% compared to H R P O G G , and by 2% compared to H R P O G N .
Figure 6 shows the ROC curves of the models, illustrating the overall ROC curves of the H R P O G model and its variants across 12 batches of data. The value for H R P O G is the highest, reaching 0.97, indicating that our model can effectively capture the characteristics of hypertensive patients in a few-shot and nonstationary environment, achieving accurate predictions for hypertensive patients. The ROC value for H R P O G N is 0.92. Although it is not as high as H R P O G , it still demonstrates effective prediction capabilities. However, due to the lack of evaluation for new data, its overall performance is weaker. The ROC value for H R P O G r e t r a i n is only 0.82, indicating that it can achieve limited prediction effectiveness, but its accuracy is still constrained by the few-shot size. The ROC value for H R P O G G is the lowest, at only 0.79, which is even lower than that of the retrained model. This also indicates that directly using an old model to train new data is not feasible, as it can lead to overfitting under few-shot conditions and affect prediction accuracy.

6. Conclusions

This study designed and implemented an online hypertension risk monitoring model, HRP-OG, based on few-shot multivisit data. Based on a reinforcement learning framework, we constructed the hypertension risk prediction model, enabling it to extract hypertension features and perform risk prediction. A generative feature replay module was designed to apply knowledge learned from historical data to new data domains, facilitating model updates. Furthermore, a data quality assessment method was proposed to avoid bias due to large gaps between the feature distribution of historical and new data, further enhancing the effectiveness of model updates. Experimental results on the MIMIC-III dataset validated our model, and the average accuracy of each stage reached 0.96, which can effectively identify hypertensive patients. In the future, we plan to incorporate a wider range of diseases into our model to improve its practicality. Meanwhile, more advanced deep learning methods [46] will be considered to further improve the performance of the model.

Author Contributions

Conceptualization, H.Y. and J.C.; methodology, H.Y. and J.C.; software, H.Y. and S.Z.; validation, H.Y. and Z.Q.; formal analysis, H.Y.; investigation, H.Y.; resources, S.Z.; data curation, H.Y.; writing—original draft preparation, H.Y.; writing—review and editing, H.Y., J.C. and S.L.; visualization, H.Y.; supervision, S.L.; project administration, J.C.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Beijing Natural Science Foundation (No. 4222022) and National Key Research and Development Program of China (Grant No. 2020YFB2104402).

Institutional Review Board Statement

The Massachusetts Institute of Technology (Cambridge, MA, USA) and the Institutional Review Boards of Beth Israel Deaconess Medical Center (Boston, MA, USA) have approved the data collection and the use of MIMIC-III for research purposes and granted waiver of informed consent. The author Haokang Yan obtained the certificate (no: 63730510) and the permission to use the MIMIC-III database after completing the web-based course. https://www.citiprogram.org/members/index.cfm?pageID=50 (accessed on 3 July 2024).

Informed Consent Statement

All patient information in the database is de-identified, so this study was exempt from the approval of the institutional review board and informed consent from individual patients.

Data Availability Statement

Dataset available on request from the corresponding authors J.C. ([email protected]) and H.Y. ([email protected]).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Elendu, C.; Amaechi, D.C.; Elendu, T.C.; Amaechi, E.C.; Elendu, I.D. Dependable approaches to hypertension management: A review. Medicine 2024, 24, 103. [Google Scholar] [CrossRef] [PubMed]
  2. Ojo, A.E.; Ojji, D.B.; Grobbee, D.E.; Huffman, M.D.; Peters, S.A. The Burden of Cardiovascular Disease Attributable to Hypertension in Nigeria: A Modelling Study Using Summary-Level Data. Glob. Heart 2024, 19, 50. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, W.; Kozlova, E.; Chen, L. The importance of medical and sports preventive and therapeutic measures for chronic diseases. Med. Dello Sport 2023, 76, 126–135. [Google Scholar] [CrossRef]
  4. Kaur, S.; Bansal, K.; Kumar, Y.; Changela, A. A comprehensive analysis of hypertension disease risk-factors, diagnostics, and detections using deep learning-based approaches. Arch. Comput. Methods Eng. 2024, 31, 1939–1958. [Google Scholar] [CrossRef]
  5. Visco, V.; Izzo, C.; Mancusi, C.; Rispoli, A.; Tedeschi, M.; Virtuoso, N.; Giano, A.; Gioia, R.; Melfi, A.; Serio, B.; et al. Artificial Intelligence in Hypertension Management: An Ace up Your Sleeve. J. Cardiovasc. Dev. Dis. 2023, 10, 74. [Google Scholar] [CrossRef] [PubMed]
  6. Spiranovic, C.; Matthews, A.; Scanlan, J.; Kirkby, K.C. Increasing knowledge of mental illness through secondary research of electronic health records: Opportunities and challenges. Adv. Ment. Health 2016, 14, 14–25. [Google Scholar] [CrossRef]
  7. Awari, S.V. Diseases prediction model using machine learning technique. Int. J. Sci. Res. Sci. Technol. 2021, 8, 461–467. [Google Scholar]
  8. Ghatage, M.M.; Ranjan, D.N. Survey paper on A Novel approach for Disease Prediction Scheme for Healthcare Management. J. Emerg. Technol. Innov. Res. 2020, 7, 1327–1332. [Google Scholar]
  9. Lai, K.; Wang, X.; Cao, C. A Continuous Non-Invasive Blood Pressure Prediction Method Based on Deep Sparse Residual U-Net Combined with Improved Squeeze and Excitation Skip Connections. Sensors 2024, 24, 2721. [Google Scholar] [CrossRef]
  10. Raghavan, P.; Liang, J.J.; Mahajan, D.; Chandra, R.; Szolovits, P. emrKBQA: A Clinical Knowledge-Base Question Answering Dataset; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 64–73. [Google Scholar]
  11. Park, J.; Cho, Y.; Lee, H.; Choo, J.; Choi, E. Knowledge graph-based question answering with electronic health records. Mach. Learn. Healthc. Conf. 2021, 149, 36–53. [Google Scholar]
  12. Socrates, V. Extraction of Diagnostic Reasoning Relations for Clinical Knowledge Graphs. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, Dublin, Ireland, 22–27 May 2022; pp. 413–421. [Google Scholar]
  13. Tran, T.N.T.; Felfernig, A.; Trattner, C.; Holzinger, A. Recommender systems in the healthcare domain: State-of-the-art and research issues. J. Intell. Inf. Syst. 2021, 57, 171–201. [Google Scholar] [CrossRef]
  14. Ghasemi, S.H.; Etminani, K.; Dehghan, H.; Eslami, S.; Hasibian, M.R.; Vakili Arki, H.; Saberi, M.R.; Aghabagheri, M.; Namayandeh, S.M. Design and evaluation of a smart medication recommendation system for the electronic prescription. In dHealth 2019—From eHealth to dHealth; IOS Press: Amsterdam, The Netherlands, 2019; pp. 128–135. [Google Scholar]
  15. Du, Z.; Yang, Y.; Zheng, J.; Li, Q.; Lin, D.; Li, Y.; Fang, J.P.; Cheng, W.; Chen, X.H.; Cai, Y. Accurate prediction of coronary heart disease for patients with hypertension from electronic health records with big data and machine-learning methods: Model development and performance evaluation. JMIR Med. Inform. 2020, 8, 17257. [Google Scholar] [CrossRef] [PubMed]
  16. Rasmy, L.; Xiang, Y.; Xie, Z.; Tao, C.; Zhi, D. Med-BERT: Pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction. NPJ Digit. Med. 2021, 4, 86. [Google Scholar] [CrossRef] [PubMed]
  17. Nowrozy, R. A Security and Privacy Compliant Data Sharing Solution for Healthcare Data Ecosystems. Ph.D. Thesis, Victoria University, Victoria, BC, Canada, 2024. [Google Scholar]
  18. Yoo, J.H.; Jeong, H.; An, J.H.; Chung, T.-M. Mood Disorder Severity and Subtype Classification Using Multimodal Deep Neural Network Models. Sensors 2024, 24, 715. [Google Scholar] [CrossRef] [PubMed]
  19. Deepa, D.R.; Sadu, V.B.; Sivasamy, D.A. Early prediction of cardiovascular disease using machine learning: Unveiling risk factors from health records. AIP Adv. 2024, 14, 035049. [Google Scholar] [CrossRef]
  20. Ning, H.; Zhang, J.; Feng, T.T.; Chu, E.K.W.; Tian, T. Control-based algorithms for high dimensional online learning. J. Frankl. Inst. 2020, 357, 1909–1942. [Google Scholar] [CrossRef]
  21. Shahraki, A.; Abbasi, M.; Taherkordi, A.; Jurcut, A.D. A comparative study on online machine learning techniques for network traffic streams analysis. Comput. Netw. 2022, 207, 108836. [Google Scholar] [CrossRef]
  22. Chowdhury, M.Z.I.; Leung, A.A.; Walker, R.L.; Sikdar, K.C.; O’Beirne, M.; Quan, H.; Turin, T.C. A comparison of machine learning algorithms and traditional regression-based statistical modeling for predicting hypertension incidence in a Canadian population. Sci. Rep. 2023, 13, 13. [Google Scholar] [CrossRef]
  23. Lee, S.; Yin, C.; Zhang, P. Stable clinical risk prediction against distribution shift in electronic health records. Patterns 2023, 4, 100828. [Google Scholar] [CrossRef]
  24. Guralnik, E. Utilization of Electronic Health Records for Chronic Disease Surveillance: A Systematic Literature Review. Cureus 2023, 15, e37975. [Google Scholar] [CrossRef]
  25. Amin, S.U.; Agarwal, K.; Beg, R. Genetic neural network based data mining in prediction of heart disease using risk factors. In Proceedings of the 2013 IEEE Conference on Information & Communication Technologies, Thuckalay, India, 11–12 April 2013. [Google Scholar]
  26. Ambekar, S.; Phalnikar, R. Disease risk prediction by using convolutional neural network. In Proceedings of the 2018 Fourth International Conference on Computing Communication Control and Automation, Pune, India, 16–18 August 2018. [Google Scholar]
  27. Liang, Y.; Guo, C. Heart failure disease prediction and stratification with temporal electronic health records data using patient representation. Biocybern. Biomed. Eng. 2023, 43, 124–141. [Google Scholar] [CrossRef]
  28. Javidi, H. Deep Neural Networks for Complex Disease Prediction Using Electronic Health Records and Genomic Data. Ph.D. Thesis, Cleveland State University, Cleveland, OH, USA, 2024. [Google Scholar]
  29. Gao, J.; Xiao, C.; Wang, Y.; Tang, W.; Glass, L.M.; Sun, J. Stagenet: Stage-aware neural networks for health risk prediction. In Proceedings of the Web Conference, Taipei, Taiwan, 20–24 April 2020. [Google Scholar]
  30. Zhao, D.; Shi, Y.; Cheng, L.; Li, H.; Zhang, L.; Guo, H. Time interval uncertainty-aware and text-enhanced based disease prediction. J. Biomed. Inform. 2023, 139, 104239. [Google Scholar] [CrossRef] [PubMed]
  31. Nancy, A.A.; Ravindran, D.; Raj Vincent, P.D.; Srinivasan, K.; Gutierrez Reina, D. Iot-cloud-based smart healthcare monitoring system for heart disease prediction via deep learning. Electronics 2022, 11, 2292. [Google Scholar] [CrossRef]
  32. Shan, G.; Xu, S.; Yang, L.; Jia, S.; Xiang, Y. Learn#: A Novel incremental learning method for text classification. Expert Syst. Appl. 2020, 147, 113198. [Google Scholar]
  33. Khennour, M.E.; Bouchachia, A.; Kherfi, M.L.; Bouanane, K.; Aiadi, O. Adapting Random Simple Recurrent Network for Online Forecasting Problems. In Proceedings of the 2024 IEEE International Conference on Evolving and Adaptive Intelligent Systems, Madrid, Spain, 23–24 May 2024. [Google Scholar]
  34. Zhang, H.; Wu, D.; Zhang, R.; Wang, R. Incremental Learning System for Disease Prediction: Epileptic Seizure Case. In Proceedings of the 2018 IEEE Global Communications Conference, Abu Dhabi, United Arab Emirates, 9–13 December 2018. [Google Scholar]
  35. De Lange, M.; Aljundi, R.; Masana, M.; Parisot, S.; Jia, X.; Leonardis, A.; Slabaugh, G.; Tuytelaars, T. A continual learning survey: Defying forgetting in classification tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3366–3385. [Google Scholar]
  36. Yang, B.; Xiang, L.; Chen, X.; Jia, W. An online chronic disease prediction system based on incremental deep neural network. Comput. Mater. Contin. 2021, 67, 951–964. [Google Scholar] [CrossRef]
  37. Van de Ven, G.M.; Tuytelaars, T.; Tolias, A.S. Three types of incremental learning. Nat. Mach. Intell. 2022, 4, 1185–1197. [Google Scholar] [CrossRef] [PubMed]
  38. Wang, Z.; Yang, Y.; Wen, R.; Chen, X.; Huang, S.L.; Zheng, Y. Lifelong learning based disease diagnosis on clinical notes. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer International Publishing: Cham, Switzerland, 2021; p. 12712. [Google Scholar]
  39. Cao, Y.; Kuai, H.; Liang, P.; Pan, J.S.; Yan, J.; Zhong, N. BNLoop-GAN: A multi-loop generative adversarial model on brain network learning to classify Alzheimer’s disease. Front. Neurosci. 2023, 17, 1202382. [Google Scholar] [CrossRef]
  40. Baytas, I.M.; Xiao, C.; Zhang, X.; Wang, F.; Jain, A.K.; Zhou, J. Patient subtyping via time-aware LSTM networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 65–74. [Google Scholar]
  41. Jijji, S.A.; Sandra, A.A.; Musa, M.Y. Ensemble Model for the Prediction of Hypertension using KNN and SVM Algorithms. Int. J. Comput. Appl. 2021, 183, 10-5120. [Google Scholar] [CrossRef]
  42. Ma, L.; Zhang, C.; Wang, Y.; Ruan, W.; Wang, J.; Tang, W.; Ma, X.; Gao, X.; Gao, J. Concare: Personalized clinical feature embedding via capturing the healthcare context. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; p. 34. [Google Scholar]
  43. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.W.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3149–3157. [Google Scholar]
  44. Fang, M.; Chen, Y.; Xue, R.; Wang, H.; Chakraborty, N.; Su, T.; Dai, Y. A hybrid machine learning approach for hypertension risk prediction. Neural Comput. Appl. 2023, 35, 14487–14497. [Google Scholar] [CrossRef]
  45. Alikhan, J.S.; Alageswaran, R.; Amali, S.M.J. Self-attention convolutional neural network optimized with season optimization algorithm Espoused Chronic Kidney Diseases Diagnosis in Big Data System. Biomed. Signal Process. Control 2023, 85, 105011. [Google Scholar]
  46. Yang, Y.; Su, X.; Zhao, B.; Li, G.; Hu, P.; Zhang, J.; Hu, L. Fuzzy-based deep attributed graph clustering. IEEE Trans. Fuzzy Syst. 2023, 32, 1951–1964. [Google Scholar] [CrossRef]
Figure 1. Distribution of age over 5 years. (a) Distribution of age in hospital 1; (b) distribution of age in hospital 2.
Figure 1. Distribution of age over 5 years. (a) Distribution of age in hospital 1; (b) distribution of age in hospital 2.
Sensors 24 05033 g001
Figure 2. The whole framework of HRP-OG.
Figure 2. The whole framework of HRP-OG.
Sensors 24 05033 g002
Figure 3. Distribution of characteristics over 12 years. (a) Distribution of age over 12 years; (b) distribution of diastolic blood pressure over 12 years; (c) distribution of systolic blood pressure over 12 years; (d) distribution of weight over 12 years.
Figure 3. Distribution of characteristics over 12 years. (a) Distribution of age over 12 years; (b) distribution of diastolic blood pressure over 12 years; (c) distribution of systolic blood pressure over 12 years; (d) distribution of weight over 12 years.
Sensors 24 05033 g003
Figure 4. Accuracy value of ablation experiments.
Figure 4. Accuracy value of ablation experiments.
Sensors 24 05033 g004
Figure 5. PR-AUC value of ablation experiments.
Figure 5. PR-AUC value of ablation experiments.
Sensors 24 05033 g005
Figure 6. ROC curve of ablation experiments.
Figure 6. ROC curve of ablation experiments.
Sensors 24 05033 g006
Table 1. Statistical information of hypertension dataset.
Table 1. Statistical information of hypertension dataset.
FeaturesIs-Hyper = 0Is-Hyper = 1
mean # of age67.970.1
mean # of heart rate88.984.7
mean # of respiratory rate19.018.9
mean # of heart rhythmNoneNone
mean # of LHR54.654.4
mean # of HHR121.8121.4
mean # of LBOS89.789.7
mean # of HBOS99.099.2
mean # of LRR9.29.3
mean # of HRR33.433.0
mean # of SBP118.2123.4
mean # of DBP62.264.6
mean # of MBP79.182.3
mean # of life patternNoneNone
mean # of weight93.195.5
mean # of CVBP10.911.0
mean # of LABP83.183.3
mean # of HABP147.5149.6
mean # of breathing patternNoneNone
mean # of temperature37.936.0
Table 2. Statistics of the multivisit samples.
Table 2. Statistics of the multivisit samples.
Number of VisitsNumber of PatientsNumber of Records
244798958
311243372
4–66722915
6–8137871
8–1047397
10–1541457
>1515350
all651517,320
Table 3. Distribution of the multivisit samples over 12 years.
Table 3. Distribution of the multivisit samples over 12 years.
YearNumber of PatientsNumber of Records
20015381247
20026101585
20036011639
20046121638
20056071652
20065761551
20075621493
20085691509
20096001596
20105881518
20115721470
201279421
Table 4. Accuracy value of comparison experiments.
Table 4. Accuracy value of comparison experiments.
MethodLSTMKNN-SVMConCareLightGBMKNN-LightGBMSACNN-SOAHRP-OG
10.52860.59380.59380.57810.56250.50000.8594
20.48180.50000.49220.59900.59380.48440.9010
30.53530.55880.60590.55880.57060.49710.9118
40.56730.55770.55770.58170.56250.49760.8990
50.65000.70430.70430.67830.62170.50000.9196
60.58070.63660.62110.72670.67080.47830.9379
70.51190.49080.48810.53810.58570.50000.8595
80.56010.59020.58470.57380.60660.49180.9344
90.60830.61780.48730.54140.53500.48730.9236
100.54970.57890.58190.56140.60820.49120.8830
110.68240.70440.67610.68550.63520.49370.9528
120.66750.68060.67800.68060.60730.48430.9424
average0.57700.60120.58930.60860.59670.49210.9104
Table 5. PR-AUC value of comparison experiments.
Table 5. PR-AUC value of comparison experiments.
MethodLSTMKNN-SVMConCareLightGBMKNN-LightGBMSACNN-SOAHRP-OG
10.79170.79700.79690.78910.78130.70570.9349
20.73700.75000.74480.79950.79690.74480.9453
30.80290.77650.80290.77940.78530.71760.9618
40.78130.77870.77880.79090.78130.73080.9567
50.85320.85210.85220.83910.81090.66300.9609
60.81060.81680.81060.86340.83540.69880.9752
70.74520.74530.74520.76900.79290.74760.9476
80.79510.79500.79510.78680.80330.70770.9699
90.80890.80890.80890.77070.76750.69750.9650
100.78950.79960.78950.78070.80410.70180.9474
110.84910.85220.84910.84270.81760.67300.9811
120.84030.84130.84030.84030.80360.69900.9738
average0.80030.80110.80120.80430.79830.70730.9610
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, S.; Yan, H.; Zhou, S.; Qiao, Z.; Chen, J. HRP-OG: Online Learning with Generative Feature Replay for Hypertension Risk Prediction in a Nonstationary Environment. Sensors 2024, 24, 5033. https://doi.org/10.3390/s24155033

AMA Style

Lin S, Yan H, Zhou S, Qiao Z, Chen J. HRP-OG: Online Learning with Generative Feature Replay for Hypertension Risk Prediction in a Nonstationary Environment. Sensors. 2024; 24(15):5033. https://doi.org/10.3390/s24155033

Chicago/Turabian Style

Lin, Shaofu, Haokang Yan, Shiwei Zhou, Ziqian Qiao, and Jianhui Chen. 2024. "HRP-OG: Online Learning with Generative Feature Replay for Hypertension Risk Prediction in a Nonstationary Environment" Sensors 24, no. 15: 5033. https://doi.org/10.3390/s24155033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop