Next Article in Journal
Anti-Swing Control for Quadrotor-Slung Load Transportation System with Underactuated State Constraints
Previous Article in Journal
Analysis of the Impact of Atmospheric Models on the Orbit Prediction of Space Debris
Previous Article in Special Issue
Enhancing Electroretinogram Classification with Multi-Wavelet Analysis and Visual Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework for Anesthesia Depth Prediction from Drug Infusion History

School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8994; https://doi.org/10.3390/s23218994
Submission received: 7 October 2023 / Revised: 28 October 2023 / Accepted: 31 October 2023 / Published: 6 November 2023
(This article belongs to the Special Issue Biosignal Sensing and Analysis for Healthcare Monitoring)

Abstract

:
In the target-controlled infusion (TCI) of propofol and remifentanil intravenous anesthesia, accurate prediction of the depth of anesthesia (DOA) is very challenging. Patients with different physiological characteristics have inconsistent pharmacodynamic responses during different stages of anesthesia. For example, in TCI, older adults transition smoothly from the induction period to the maintenance period, while younger adults are more prone to anesthetic awareness, resulting in different DOA data distributions among patients. To address these problems, a deep learning framework that incorporates domain adaptation and knowledge distillation and uses propofol and remifentanil doses at historical moments to continuously predict the bispectral index (BIS) is proposed in this paper. Specifically, a modified adaptive recurrent neural network (AdaRNN) is adopted to address data distribution differences among patients. Moreover, a knowledge distillation pipeline is developed to train the prediction network by enabling it to learn intermediate feature representations of the teacher network. The experimental results show that our method exhibits better performance than existing approaches during all anesthetic phases in the TCI of propofol and remifentanil intravenous anesthesia. In particular, our method outperforms some state-of-the-art methods in terms of root mean square error and mean absolute error by 1 and 0.8, respectively, in the internal dataset as well as in the publicly available dataset.

1. Introduction

In recent works, clinical pharmacology research has focused on identifying best practices for ensuring patient safety by maximizing anticipated drug effects while reducing drug-induced side effects. This tradeoff is particularly important when developing anesthetic delivery patterns [1]. Specifically, the main targets are concentrated on the following three variables: analgesia (pain relief), hypnosis (based on loss of memory and consciousness), and muscle relaxation (quiescence). Therefore, it is crucial to provide appropriate doses of different drugs during this process to ensure optimal intervention conditions while preventing unnecessary and dangerous consequences.
In clinical practice, total intravenous anesthesia (TIVA) is primarily used for drug delivery. To control the depth of anesthesia (DOA), target-controlled infusion (TCI) of propofol for TIVA is one of the most widely used techniques in anesthesia and sedation. To accurately administer medication, the DOA must be continuously measured to assess the patient’s level of anesthesia and reduce drug-induced side effects. In anesthesia operations, the bispectral index (BIS) is the preferred metric for clinicians to monitor the DOA of patients. The BIS is a noninvasive system that calculates electroencephalography values ranging from 0 (no brain activity) to 100 (fully awake patients) as dimensionless numbers [2,3].
To achieve precise drug administration, various models that relate the rate of drug infusion to the DOA in humans have been proposed in recent decades. The traditional approach to target-controlled infusion is based on a pharmacokinetic and pharmacodynamic model (PK-PD) that uses propofol doses to predict propofol effect-site concentration (Ce) as an alternative to the DOA [4,5]. Although the traditional model is simple to implement, the PK-PD model fails to use the BIS as the measured metric, which is the most commonly used metric in clinical practice. More importantly, the BIS does not always agree with the model-driven Ce of propofol, especially during the anesthesia induction and recovery period [6]. With the development of machine learning, some studies [1,7,8,9,10,11] have attempted to model the DOA and the dose of propofol by using different machine learning methods. However, these models only consider the history of propofol infusion, while entirely ignoring the synergistic effect of the concomitant infusion of remifentanil on anesthesia.
In recent works, Lee et al. proposed a deep learning model for investigating the influence of the infusion histories of propofol and remifentanil and patient physiological characteristics on the DOA [12]. In addition, Sara Afshar et al. developed a combinatorial deep learning structure that predicts the depth of anesthesia according to electroencephalography (EEG) signals, demonstrating good prediction performance [13,14,15]. However, EEG signals are susceptible to inotropic and cardiac interference and sensitive to electromagnetic interference [16]. In this situation, the infusion histories of propofol and remifentanil are widely adopted to predict the DOA in clinical practice. It is well known that the statistical properties of time series change over time, thereby causing distributions to change over time. This phenomenon is known as the distribution shift problem in the field of machine learning. Our task is a time series prediction task; thus, the distribution shift problem caused by differences in the physiological characteristics of patients cannot be ignored. But, previous works fail to consider this issue. Moreover, the existing model proposed in [12] has poor performance in the induction and recovery periods, as shown in Figure 1.
To address the above two limitations, this paper proposes a new deep learning framework for DOA prediction according to drug infusion history. The key idea of the proposed approach is to use domain adaptation to address the issue that the data distribution varies across patients. Furthermore, knowledge distillation is adopted to learn the intermediate feature map of the teacher model by using the prediction model, which is obtained by the historical moment BIS as the extra feature. Specifically, an adaptive recurrent neural network (AdaRNN) [17] is applied to our task for solving the problem of data distribution differences by characterizing and matching the temporal distributions. In contrast to the traditional AdaRNN model, which uses boosting-based importance evaluations to obtain the importance vector, a neural-network-based method is used to generate the importance vector. This method can learn much useful information about the output RNN hidden states, thereby generating an enhanced importance vector. In addition, a new knowledge distillation framework is developed in our work, which uses feature-based distillation in the recurrent neural network and bottleneck networks to transfer rich temporal information and accurate efficacy information to the prediction network, respectively.
The main contributions of this work can be summarized as follows:
  • A new deep learning framework for DOA prediction according to drug infusion history is proposed to overcome the shortcomings of the existing DOA prediction methods.
  • A modified AdaRNN algorithm is developed in our framework for DOA prediction to address the issue of distribution shifts in the physiological characteristics of different patients.
  • A feature-based knowledge distillation framework is proposed for time series prediction, which allows the prediction model to obtain more useful information. This framework enables the intermediate features of the model to accurately represent the DOA, thereby ensuring reliable and stable output results. To the best of our knowledge, this is the first time that knowledge distillation has been implemented to predict the DOA.
  • Extensive experimental results show that our method has better performance than the existing DOA prediction methods on a publicly available dataset during all periods, including the induction, maintenance, and recovery periods.
The remainder of this paper is organized as follows. In Section 2, we review some related works. The proposed method is introduced in Section 3. The experimental results are discussed in Section 4. Finally, Section 5 presents the conclusions.

2. Related Works

2.1. Total Intravenous Anesthesia

The traditional TIVA method is the PK-PD model [4,5], which uses propofol doses to predict the effect-site concentration of propofol. Gonzalez-Cava et al. [18] proposed a PK-PD model that considers only propofol kinetics and is thus insufficient for accurately performing the DOA. Jose M. Gonzalez-Cava et al. [19] developed an identification algorithm based on optimization techniques and the traditional PK-PD model to determine the hypnotic level of patients under propofol–remifentanil anesthesia. In recent years, rapid advancements in machine learning and artificial intelligence have led to the development of new anesthesia prediction models. Esteban Jove et al. [7] developed a hybrid model with clustering and regression techniques that use EMG signals and propofol infusion rates to predict BIS signals. Regina Padmanabhan et al. [1] applied reinforcement learning to develop a closed-loop anesthesia controller that uses the bispectral index as the control variable while considering the mean arterial pressure (MAP). These two parameters were used to control propofol infusion rates to regulate the BIS and MAP values to the desired range. Juan A. Méndez et al. [8] proposed an adaptive model based on fuzzy logic and a genetic algorithm that aimed to provide more accurate anesthesia prediction. Sahar Javaher Haghighi et al. [9] defined a DOA index and proposed a denoising method for extracting the 40 Hz auditory steady-state response period, using adaptive multilevel wavelet denoising to calculate the proposed DOA index. It is worth noting that this algorithm does not use any medical or physiological information to define the DOA index. Therefore, this DOA index may not be well suited to patients with different physiological characteristics. Ahmad Shalbaf et al. [10] determined the best EEG features and fed these features into an adaptive neuro-fuzzy inference system with linguistic hedges to classify the patient DOA stage. However, previous work was performed in limited experimental settings using small numbers of study participants; thus, these works cannot realistically reflect the effects of drugs on the DOA or the variety of physiological characteristics of patient conditions. Lee et al. [12] proposed a deep learning model that used the infusion histories of propofol and remifentanil and patient characteristics. This deep learning model was trained on 231 subjects who received TIVA during surgery. In contrast to previous work, Sara Afshar et al. [13] combined convolutional neural networks (CNNs), long short-term memory (LSTM) networks, and an attention mechanism to develop a new framework for predicting the DOA according to EEG signals. However, in clinical surgery, anesthesiologists are more likely to control the DOA based on the effects of the drug.

2.2. Domain Adaptation

When the distributions of the training, validation, and test sets are markedly different, domain adaptation (DA) or domain generalization (DG) is usually used to reduce these distribution differences. Yu et al. [20] proposed a method that used a generative adversarial network (GAN) to learn transferable features and reduce distribution differences between two domains. Wang et al. [21] proposed that dynamic distribution adaptation can quantitatively estimate the relative importance of each distribution and prevent feature distortion with manifold feature learning. Li et al. [22] used maximum mean discrepancy (MMD) regularization, which reduces the distributions of different domains and matches the aligned distributions to arbitrary prior distributions through adversarial feature learning. Most DA and DG techniques apply CNNs rather than recurrent neural networks (RNNs) to classification tasks. Thus, when there are distribution differences or temporal covariate shifts in time series data, existing DA and DG methods may not be applicable. Recently, Chaehwa Yoo et al. proposed a strategy for the unsupervised DA training of sleep phase networks to overcome the domain differences between the source and target domains in more realistic situations by generating domain invariant features, and Du et al. [17] applied a novel domain adaptation approach to time series forecasting, applying temporal distribution characterization (TDC) to measure the distribution information in the time series and temporal distribution matching (TDM) to match the distributions between different periods.

2.3. Knowledge Distillation

Deep neural networks are adept at learning multilayer feature representations with increasing levels of abstraction. Bengio et al. [23] proposed representation learning for multilayer features. Romero et al. [24] first introduced intermediate representations in Fitnets, which aimed to provide the output of intermediate teacher networks to improve the training of the student model. Inspired by this approach, Gao et al. [25] distilled the feature maps of different levels of the teacher and student networks. Hong et al. [26] trained teacher networks using heterogeneous tasks and applied knowledge distillation to the intermediate layer representations to ensure consistency between the intermediate feature representations of the reconstruction processes of the student and teacher networks. Zhang et al. [27] distilled features in each level of the teacher and student networks, thereby allowing the student network to learn privileged information from the teacher network. Zhang et al.’s model [27] was applied to time series prediction, demonstrating the effectiveness of knowledge distillation in time series prediction tasks.

3. Methodology

Different from the previous DOA prediction technique, our proposed method uses domain adaptation and knowledge distillation techniques to further improve the DOA prediction accuracy. Specifically, for domain adaptation, a modified AdaRNN is utilized to enhance the generalizability of the proposed model. For the knowledge distillation, the teacher network aims to provide intermediate feature representations of the BIS, while the student network aims to predict the BIS more precisely by transforming the intermediate features in the teacher’s intermediate feature domain. Figure 2 shows an overview of our proposed framework, which consists of a teacher model and a student model. The teacher model has an additional input to learn an excellent representation of the BIS and transfer the BIS representation via knowledge distillation. Moreover, the distribution shift problem caused by the various physiological characteristics of different patients is also solved by modifying AdaRNN to improve the robustness and generalizability of the proposed DOA prediction model.

3.1. AdaRNN

AdaRNN is mainly adopted to solve the temporal covariate shift (TCS) problem. AdaRNN can characterize the temporal distributions and capture the long-term time dependencies in time series. In general, AdaRNN includes temporal distribution characterization (TDC) and temporal distribution matching (TDM). In order to take advantage of AdaRNN for our task, the TDM approach in AdaRNN is modified to efficiently adapt the proposed network. Figure 3 illustrates an overview of the TDC approach, which divides the data into regions with the largest distribution differences, thereby ensuring that the model is trained starting from the worst case. Figure 4 depicts an overview of the TDM method, which reduces cross-domain distribution differences in the model to enhance the generalizability of the model.

3.1.1. Temporal Distribution Characterization

According to the maximum entropy (ME) principle [28], it is reasonable to have as diverse a distribution as possible for each period to maximize the entropy of the total distribution, without any prior assumptions about how to divide the time series data. This approach allows for more general and flexible modeling of future data. In addition, the use of shared knowledge in the time series data can be maximized by identifying periods that are least similar to each other. Thus, the problem can be expressed as follows:
max 0 < K K 0 max n 1 , , n k 1 K 1 i j K d D i , D j s . t . i , Δ 1 < D i < Δ 2 ; i D i = n
where d ( · , · ) is a distance metric, Δ 1 and Δ 2 are predefined parameters, K 0 is a hyperparameter, and D i denotes the i-th time series segment.
The goal of the optimization problem (1) is to maximize the difference in the distribution in each period by determining K. If the model learns from the worst-case scenario, it has better generalizability to the unseen test data. This assumption has been verified in theoretical analyses of the time series models [29,30], showing that diversity is crucial in time series modeling. The time series partitioning of the optimization problem in (1) is a computationally intractable problem that may not have a closed solution. However, the optimization problem in (1) can be solved by using dynamic programming (DP) [31] with a suitable distance metric. Considering the issues of scalability and efficiency for the large-scale data, similar to previous methods, a greedy algorithm is adopted for the AdaRNN in our work to solve the optimization problem in (1). To show the flow of TDC more clearly, the pseudo-code is summarized in Algorithm 1.
Algorithm 1: Temporal distribution characterization
Sensors 23 08994 i001

3.1.2. Temporal Distribution Matching

Once the TDC approach determines the time periods to be learned, the temporal distribution matching approach is required, which learns the public knowledge shared across each time period by matching the distributions of different time periods. Thus, the learned model M has better generalizability on the test data than other methods that only rely on the local or statistical information.
The TDM prediction loss L p r e d can be calculated as follows:
L pred ( θ ) = 1 K j = 1 K 1 D j i = 1 D j l y i j , M t x i j ; θ
where x i j , y i j denotes the i-th labeled segment in period D i , l ( , ) is a loss function, e.g., the mean square error (MSE) loss, θ denotes the learnable model parameters, and M t denotes the teacher model.
In contrast, the minimization of Formula (2) can learn only the predictive knowledge for each period and cannot reduce distribution differences between each period to exploit common knowledge. A primitive approach is to use some distribution measure of distance as a regularization term for each period pair D i and D j . Based on previous domain adaptation research [32,33] in which distribution matching is typically performed on high-level representations, Du et al. [17] applied a distribution matching regularization term to the final output of the RNN model.
Since each hidden state only contains the part of the distribution information of an input sequence, Du et al. [17] adopted the distribution regularization term in each hidden state of the RNN model to capture the temporal dependency of each hidden state. Then, they introduced the importance vector α , which aims to learn the relative importance of each hidden state, with each hidden state weighted with a normalized α .
Given a period pair ( D i , D j ) , the temporal distribution matching loss can be formulated as:
L t d m D i , D j ; θ = t = 1 V α i , j t d h i t , h j t ; θ
where α i , j t denotes the distribution importance between periods D i and D j in state t, and V denotes the V hidden states of the RNN.
By integrating (2) and (3), the final temporal distribution matching loss can be formulated as
L ( θ , α ) = L p r e d + λ 2 K ( K 1 ) i , j i j L t d m D i , D j ; θ , α
where λ is a balance hyperparameter.

3.1.3. Neural-Network-Based Importance Evaluation

Du et al. [17] used a boosting-based importance evaluation algorithm to learn α . However, in our experiments, we observe that the boosting-based method is difficult to improve the experimental results. In contrast, a naive method that obtains α through a fully connected network usually results in good results. The boosting-based method is not suitable for our task since α and θ are highly relevant in the early training phase and the hidden state representations are learned according to the model parameter θ . Thus, α must be determined according to θ . Therefore, a neural network is used in our proposed framework to comprehensively explore the importance of each hidden state. Specifically, the importance of the RNN output hidden states is initialized as the mean value of the RNN output hidden states. The initialization can be expressed as α i , j = { 1 / V } V . Then, we use a fully connected network with weight W i , j that takes ( H i , H j ) as input and outputs α . This process can be formulated as α i , j = g W i , j H i , H j T , where g denotes an activation and normalization function and ⊙ stands for the elementwise product.

3.2. Knowledge Distillation

Since response-based knowledge is mainly used in classification tasks and our task is a regression task, we use a knowledge distillation framework based on feature representations to enable the student network to learn better intermediate feature representations from the more informative teacher network. Previous work on feature-based knowledge distillation [25,26,27] has shown that teacher knowledge is typically propagated through high and low levels. Thus, in our method, the knowledge is applied to propagate through the final outputs of the RNN and bottleneck network to learn the temporal relationships in the teacher RNN and the feature map representations of the teacher bottleneck network, respectively.
Firstly, we introduce the historical moment BIS data (i.e., BIS data before the observation moment) in the teacher network to learn the DOA representation. The student network uses the propofol and remifentanil infusion histories as the inputs. Secondly, to propagate the knowledge from the teacher network to the student network more effectively, the hidden states of the output of the student RNN are used to mimic the hidden states of the output of the teacher RNN in order to improve the prediction performance of the student network. Moreover, the feature map of the bottleneck of the student network is adopted in our framework to mimic the feature map of the bottleneck of the teacher network. In this situation, the similarity between the predictions of the student and teacher networks will be increased.

3.2.1. Knowledge Distillation via RNN

Since the teacher network introduces the historical moment BIS data, the output hidden states of the teacher RNN contain more richer temporal information about the DOA, thus allowing the model to predict the BIS change trends more accurately. Therefore, the output hidden states of the student RNN is utilized in our proposed framework to imitate the output hidden states of the teacher RNN. In contrast to the TDM regularization term, we imitate the feature representations of the output hidden states of the teacher RNN to learn the temporal relationships of the teacher RNN for improving the prediction performance of the student network. Therefore, the feature imitation loss is introduced. Specifically, we use a gated recurrent unit (GRU) as the basic unit of the RNN module. Assume that T g r u ( z ) is the feature representation of the teacher GRU trained on z, which denotes the input data to the teacher network and includes historical BIS data, and S g r u ( x ) denotes the feature representation of the student GRU trained on x, which includes the propofol and remifentanil infusion histories. L g ( θ ) can be formulated as follows:
L g ( θ ) = | | T g r u ( z ; θ ) S g r u ( x ; θ ) | | 2

3.2.2. Knowledge Distillation via Bottleneck

For the RNN output feature, only the effects of the propofol and remifentanil on the BIS are considered. However, according to the experimental results in [34], the drug elimination rates of the remifentanil and propofol were different among people with different physiological characteristics (e.g., older people with the same duration of surgery had higher drug elimination rates). In contrast to the temporal information contained in RNNs, the bottlenecks add the static physiological information (e.g., height, weight, sex, and age) into the RNN output to fully learn the anesthetic effects of both propofol and remifentanil drugs on people with different physiological characteristics. Moreover, the teacher network uses the historical BIS data features, which can accurately represent the effects of static physiological information and dynamic drug information on the BIS. Thus, the feature representations of the student bottleneck output is utilized to mimic the feature representations of the teacher bottleneck output. Specifically, supposing that T b ( z ) denotes the feature representations of the teacher bottleneck trained on z and S b ( x ) denotes the feature representations of the student bottleneck trained on x, L b ( θ ) can be formulated as follows:
L b ( θ ) = | | T b ( z ; θ ) S b ( x ; θ ) | | 2

3.3. Loss Function

3.3.1. Loss Function of the Teacher Network

To obtain more effective feature representations for the teacher network, combining (2) and (3), the teacher loss can be formulated as follows:
L T ( θ , α ) = L p r e d + λ 2 K ( K 1 ) i , j i j L t d m D i , D j ; θ , α

3.3.2. Loss Function of the Student/Prediction Networks

The loss of the prediction network has three components: the predicted loss, the TDM regulation loss, and the distillation loss (i.e., L g ( θ ) and L b ( θ ) ), which can be formulated as follows:
L S ( θ , α ) = L pred + λ 2 K ( K 1 ) i , j i j L t d m D i , D j ; θ , α + L g ( θ ) + L b ( θ )
In our method, the predicted loss is used to learn the DOA during each period, and the TDM loss is adopted to reduce the distribution differences between various periods. Then, the effective DOA feature representation in the teacher model is transferred to the student model by the distillation loss.

4. Experiments

4.1. DOA Type and Dataset

In order to calculate the DOA, some typical measurements have been proposed in clinical anesthesia, which are shown in Table 1. Compared with other measurements of DOA type (e.g., Narcotrend index and Patient State index), the BIS has two main advantages. On the one hand, the BIS is the most widely used anesthesia depth indicator and has been approved by the FDA to be marketed as a monitor of anesthetic effects on the brain. The BIS correlates well with the sedative effects of many anesthetic drugs and can accurately reflect their depth of sedation, especially for commonly used anesthetic drugs such as propofol and sevoflurane. On the other hand, the BIS has been widely adopted in clinical validation. Particularly, the BIS is collected as the only indicator of anesthesia depth in the publicly available dataset (i.e., VitalDB). Therefore, the BIS is adopted as the measurement of DOA type in our work.
The dataset used in the experiments is the VitalDB (https://vitaldb.net) (accessed on 1 January 2022). database, which is collected and registered by the Department of Anesthesiology and Pain Medicine, Seoul National University Hospital, Seoul Metropolitan Medical College, Seoul, Republic of Korea [40]. The VitalDB database is an open-access dataset, which can freely download from the website, https://vitaldb.net (accessed on 1 January 2022). In addition, the VitalDB database is a comprehensive dataset that includes the intraoperative biosignals and clinical information of 6388 surgical patients. It not only contains the demographic data (height, weight, sex, and age), but also contains more than 60 procedure-related clinical indicators of basic equipment used in operating rooms, such as patient monitors, anesthesia machines, and BIS monitors, as well as target-controlled infusion pumps, cardiac output monitors, and local oximeters. The BIS data include the BIS values and the signal quality indices collected by BIS VISTA at 1 s intervals. The propofol and remifentanil data include the cumulative infusion volumes, Ce values, and plasma concentration (Cp) values of the two drugs collected by target-controlled infusion pumps at 1 s intervals.
In our experiments, the data of 1000 patients are randomly selected from the VitalDB. After data processing, only 332 data are retained and the rest are discarded due to the excessive missing BIS or the drug administration records. The detailed data processing is given in Section 4.2. The retained 332 patients are randomly divided into training, validation, and testing datasets, in which the training, validation, and testing dataset contain 180, 76, and 76 cases, respectively. The characteristics of these three datasets are shown in Table 2.

4.2. Data Processing

From the start of the propofol or remifentanil infusion to the end of the BIS measurement, the data meeting the following conditions are discarded:
  • If the BIS value at the start of the drug infusion is less than 80;
  • If the data are missing for more than 300 s;
  • If the first BIS is recorded when the cumulative amount of the infused drug is not 0.
After the above processing steps, the patient sample data may still contain a small number of missing or erroneous values. To address the issue of missing values in data, according to the linear relationship between propofol and remifentanil over time, the missing values in the patient data are determined by linear interpolation. Since the total dosage of propofol and remifentanil is monotonically nondecreasing, for cases in which there is a partial decrease in the cumulative infusion volumes in the patient data, the erroneous values may be caused by instrument recordings or how the data are exported. To solve this problem, the mean values of the moments before and after the erroneous moment are used. The above processing techniques were applied to the training set.
The cumulative injection volume is recorded by target-controlled infusion pumps, and the injection history data collected by the target-controlled infusion pumps are updated every 10 s [41]. Therefore, the cumulative medication usage of propofol and remifentanil over 10 s is adopted as a feature to create a time window with a sequence length of 120. Since the data variations among different patients are useless for training the network, we downsample the training data, i.e., one sampling point every 10 s is selected. Since the BIS values of different patients have diverse trends and the BIS values of each patient change frequently, we smoothed the BIS values in the training set by using locally weighted scatter plot smoothing (LOWESS) with a smoothing parameter of 0.03 to reduce computational errors during training. The unprocessed BIS values are used in the validation and test datasets.

4.3. Experiment Settings

4.3.1. Teacher and Student Network

The teacher and student networks have the same structure, except that the input to the teacher network has an additional feature (the BIS value of the previous moment) that is not available to the student network. Our network includes an RNN module, a bottleneck module, and an output module. We use a GRU as the basic unit in the RNN module because of its high performance in practical applications. The bottleneck module focuses on learning the anesthetic effects of propofol and remifentanil on patients with different physiological characteristics. In our work, the bottleneck module consists of two fully connected networks. After the bottleneck network, our proposed model learns the combined dynamic (propofol and remifentanil drug dosage information) and static (patient physiological characteristics) information. Finally, the predicted BIS values with a fully connected network are obtained.

4.3.2. Implementation Details

In the experiments, PyTorch 1.7.1 is used to implement our proposed model with a 24 GB NVIDIA TITAN RTX GPU and an Intel(R) Xeon(R) CPU E5-2680 v3 @ 2.50 GHz with 32 GB RAM and 128 GB SSD, and our teacher and student networks are trained simultaneously. The network is trained by an Adam optimizer with an initial learning rate of 0.005. The learning rate is decayed every 10 epochs by a factor of 0.1, and the weight clipping is utilized to prevent gradient explosion. We use a batch size of 256 in the training phase and a batch size of 128 in the validation and testing phases. The balance parameter λ is set to 0.05 and the number of difference distribution periods K is set to 5. The cosine similarity is adopted to measure the distribution differences across each period. The entire training process takes no more than 2 h, where the training batch size is set to 256. Our code is available at https://github.com/chanwendy/Domain-adaptation-DOA-prediction (accessed on 25 June 2023).

4.4. Evaluation Metrics and Results

The experimental results are expressed as the regression errors. To evaluate the regression performance of our proposed framework, the predicted values are assessed according to the following commonly used evaluation metrics [42,43]: the mean absolute error (MAE) [44] and the root mean square error (RMSE) [44]. Based on [28], the performance error (PE), median performance error (MDPE) [28], and median absolute performance error (MDAPE) [28] are also used to measure the performance for comparison. Particularly, the PE is calculated as the measured BIS-predicted BIS/predicted BIS. Since the PE is based on performance error as part of the predicted drug concentration, it is particularly useful to clinicians. The MDPE is a signed value and therefore indicates the direction of the performance error (overprediction or underprediction) rather than the size of the error. The MDAPE measures the computer-controlled infusion pump (CCIP) performance and reflects the inaccuracy of the CCIP. To further reveal the performance of the model, we evaluated the performance of the three periods (induction, maintenance, recovery) in TIVA. The induction period in TIVA represents the 10 min period starting with the surgical injection of propofol, and the recovery period represents the period between stopping the surgical injection of propofol to the end of anesthesia; the remaining time represents the maintenance period.

4.4.1. Open-Access Dataset

In the experiment, we conduct a comprehensive analysis of mean comparisons across all periods, including the induction period, the maintenance period, and the recovery period. The compared methods are, respectively, the baseline method [12], FEDformer [45], and Crossformer [46]. The experiment results are shown in Table 3, and are depicted as the mean ± standard deviation. According to Table 3, one can observe that, in most situations, our proposed method has better performance than other compared methods during each period of anesthesia. This indicates that our proposed method can efficiently solve these issues of the baseline method, i.e., the weak generalizability of the model and the poor performance in the induction and recovery periods. In addition, we use the vanilla transformer [47] as the transformer backbone of our pipeline; the results are shown in Table 3. Benefiting from the long-term modeling capability of the transformer model, it can effectively capture the global features of the time series data, which can effectively learn the continuous effect of anesthesia drugs on the human body; therefore, the transformer encoder can obtain more accurate features of the depth of anesthesia. In addition, there is only one case in which FEDformer outperforms the proposed model in the recovery period. The main reason may be that FEDformer is a transformer-based method that can better capture the global features of the time series and thus is more able to synthesize information before the recovery period. However, since FEDformer fails to take into account the excessive differences in the distribution of each individual, the experimental results in the final evaluation metrics (e.g., RMSE and MAE) of anesthesia prediction are worse than those of ours.
Figure 5 shows the results of the visual comparison between the proposed method and the compared methods. From this figure, it can be seen that the proposed method can predict the DOA of unknown patients more accurately than these compared methods. The main reason that our model outperforms the compared models is that the proposed framework considers the distribution shift problem caused by different patient data and solves this problem with an appropriate method. Moreover, the introduction of historical BIS values to the teacher network in the distillation model allows the prediction model to obtain much useful information about the intermediate feature representations of the BIS values, thereby allowing the model to predict partial mutations in the induction and recovery periods.

4.4.2. In-House Dataset

To further validate the generalization ability of our model and the applicability of our model under different measurements of DOA type, we perform the training and testing on our in-house dataset. This dataset uses the Narcotrend index (NI) as the measurement for the depth of anesthesia. Compared with the BIS, the NI has a larger response gap for patients of different age groups and different disease states, creating a larger domain shift. The main difference between the VitalDB dataset and our in-house dataset is shown in Table 4.
In the experiment, we randomly select 24 cases as the training set, and others as the test cases. The experiment results are shown in Table 5. Obviously, our proposed method still performs better than these compared methods during each period of anesthesia. This demonstrates that our proposed method can achieve good results when there is a large domain shift in the dataset, while other compared methods suffer from significant performance degradation due to the domain shift. In addition, the results without the preprocessed data of our in-house dataset show that noisy data will lead to performance degradation, but our method can still achieve better performance than the SOTA models using the preprocessed data for training.
Figure 6 shows the results of the visual comparison between the proposed method and these compared methods in our in-house dataset. From this figure, it can be seen that, compared with other compared methods, our proposed method has better trend and prediction results in all periods. This is because our proposed method is not only trained from the interval with the largest distribution differences but also utilizes the NI data of historical moments to the teacher network.

4.4.3. Statics Test and Application in the Real Word

In addition, we perform some statistical tests, such as pair-t test and F-test, with these compared methods in both public and in-house datasets, and the experimental results are shown in Table 6. According to this table, one can see that, in most situations, the p-values for the pair-t test and F-test are less than 0.05 in both datasets, which indicates that our proposed model has better performance than other compared methods from the viewpoint of statistical test. There is only one case where the p-values for the Crossformer in the F-test are greater than 0.05. The main reason may be that the Crossformer takes into account the information interaction between multiple variables and achieves relatively good stability. However, from the results in Table 3, one can see that, in most cases, the standard deviations of our proposed model are still better than Crossformer.
In a real clinical environment, the closed-loop target-controlled infusion system is divided into two main parts, which are the prediction of DOA and the control of drug dosage according to the DOA, respectively. Our work mainly addresses the first part of predicting the DOA based on drug efficacy, and the part of controlling the dose based on the DOA is our future work. The prediction of DOA is commonly performed at a frequency of one prediction per second, aiming to estimate the BIS value. To simulate this inference process and evaluate the real-time capabilities of our model, we set the batch size to one in the experiments. Remarkably, our findings reveal that our model achieves a remarkable inference time of only 1 × 10−5 s per prediction of the BIS value. In addition, we compute model complexity between our model and the compared methods; the results are shown in Table 7. One can see that our model has a smaller parameter and FLOPS compared with the SOTA models. Although the baseline model parameter and FLOPS are the smallest, its ability is also the worst. In contrast, although our model increases the number of parameters with respect to the baseline, at the same time, it obtains a better performance improvement that outperforms the SOTA models, and our proposed model predicts the depth of anesthesia in only 1 × 10−5 per second, which can fully meet the clinical requirements. Therefore, our proposed model can be suited for real-time monitoring applications.

4.5. Ablation Study

To verify the effectiveness of the neural-network-based importance evaluation method, a comparative experiment with our TDM method and the boosting-based TDM method is derived to evaluate the respective performance. Table 8 shows the experimental results of our TDM method and the boosting-based method. Compared with the baseline method in Table 3, our TDM method and the boosting-based TDM method both have better experimental results than the baseline. From Table 8, one can see that our TDM method outperforms the boosting-based method in all periods, including the induction and maintenance periods. In addition, the performance between our TDM method and the boosting-based TDM method is slightly different in the recovery period. This is because the boosting-based method focuses more on the distribution at the end of the time series, so it has better results in the recovery phase, while our neural-based method focuses more on the early part of the time series, i.e., the induction period and maintenance period. This only causes our method to be a little bit worse than the boosting-based method in the recovery period. However, our method achieves significantly better results in the induction period, the maintenance period, and all three periods combined. Therefore, the proposed TDM method is adopted in our task.
The experimental results of the ablation study are illustrated in Figure 7. As can be seen from Figure 7, when the baseline is introduced with the AdaRNN, the proposed model usually works better than the original model in all stages of anesthesia. This is due to the fact that the proposed model trains in the worst case, which can alleviate the effects of excessive differences in the distribution of different patient data to a certain extent, making the model more robust and generalized. Thus, the model is able to predict better in the face of unknown patient data.
According to Figure 7a, when a knowledge distillation approach is applied in the baseline model, the new model has better performance during each period. In Figure 7c,d, the MDPE achieves better performance in the induction and maintenance periods than the AdaRNN model, which indicates that the predicted values for these two periods are more accurate and display fewer fluctuations. Therefore, knowledge distillation is useful for solving the overprediction problem faced by the AdaRNN model in the induction and maintenance periods. Based on Figure 7a,b, although incorporating the knowledge distillation into the baseline model has good stability, the difference between the AdaRNN and the knowledge distillation approach is very small in terms of the RMSE and MAE. However, when the AdaRNN and knowledge distillation are used in the same model, the prediction performance can improve, and the combined model still brings good stability. Thus, incorporating knowledge distillation into the AdaRNN model results in a great improvement in our proposed work.

5. Conclusions

Accurate anesthesia depth prediction is crucial in clinical surgeries. In this paper, a new deep learning framework is proposed that combines the domain adaptation technique and the knowledge distillation technique to improve the generalizability and prediction accuracy of anesthesia depth prediction models. Specifically, different from previous works, the proposed framework first considers the problem of distribution shifts among patients with different physiological characteristics data and trains the model from cases with the largest distribution differences in order to enhance the generalizability of the proposed model. Moreover, to improve the accuracy of anesthesia depth prediction, the BIS values of the historical moments are introduced in the teacher network of the proposed method to extract the intermediate features, representing the anesthesia depth accurately. After that, the prediction network is utilized to mimic the feature representations of the teacher network. Experimental results have shown the effectiveness of our proposed framework on a publicly available dataset compared with the traditional response model and the baseline model.
In future studies, more information on historical drug doses or patient physiological indicators related to clinical anesthesia, such as plasma concentration, effect-site concentration, target concentration, and infusion rate, can be added to our model to further improve the prediction accuracy. Furthermore, an attention mechanism can be incorporated to allow the model to focus on anesthesia depth features that are effective for predicting anesthesia depth. Finally, due to the imbalance in the quantity of data (i.e., most of the BIS is in the range of 35–45, while the other quantities are sparse). Thus, addressing data imbalance in our task also can improve the performance.

Author Contributions

Methodology, M.C.; Formal analysis, M.C.; Data curation, Y.H.; Writing—original draft, M.C.; Writing—review & editing, Z.Y.; Visualization, Y.H.; Supervision, Z.Y.; Project administration, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science and Technology Project of Guangdong Province (no. 2021A1515011341), Guangzhou Science and Technology Plan Project (no. 202002030386), and Guangdong Provincial Key Laboratory of Human Digital Twin (no. 2022B1212010004).

Institutional Review Board Statement

In this work our experiment not involving humans. In a real clinical environment, the closed-loop target-controlled infusion system is divided into two main parts, which are the prediction of DOA and the control of drug dosage according to the DOA. Our work mainly addresses the first part of predicting the DOA based on drug efficacy, and the part of controlling the dose based on the DOA is our future work.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The main data we use comes from VitalDB, available from https:vitaldb.net.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Padmanabhan, R.; Meskin, N.; Haddad, W.M. Closed-loop control of anesthesia and mean arterial pressure using reinforcement learning. Biomed. Signal Process. Control 2015, 22, 54–64. [Google Scholar] [CrossRef]
  2. Glass, P.S.; Bloom, M.; Kearse, L.; Rosow, C.; Sebel, P.; Manberg, P. Bispectral analysis measures sedation and memory effects of propofol, midazolam, isoflurane, and alfentanil in healthy volunteers. J. Am. Soc. Anesthesiol. 1997, 86, 836–847. [Google Scholar] [CrossRef]
  3. Sigl, J.C.; Chamoun, N.G. An introduction to bispectral analysis for the electroencephalogram. J. Clin. Monit. 1994, 10, 392–404. [Google Scholar] [CrossRef]
  4. Marsh, B.; White, M.; Morton, N.; Kenny, G. Pharmacokinetic model driven infusion of propofol in children. BJA Br. J. Anaesth. 1991, 67, 41–48. [Google Scholar] [CrossRef]
  5. Schnider, T.W.; Minto, C.F.; Gambus, P.L.; Andresen, C.; Goodale, D.B.; Shafer, S.L.; Youngs, E.J. The influence of method of administration and covariates on the pharmacokinetics of propofol in adult volunteers. J. Am. Soc. Anesthesiol. 1998, 88, 1170–1182. [Google Scholar] [CrossRef]
  6. Coppens, M.; Van Limmen, J.; Schnider, T.; Wyler, B.; Bonte, S.; Dewaele, F.; Struys, M.; Vereecke, H. Study of the time course of the clinical effect of propofol compared with the time course of the predicted effect-site concentration: Performance of three pharmacokinetic–dynamic models. Br. J. Anaesth. 2010, 104, 452–458. [Google Scholar] [CrossRef]
  7. Jove, E.; Gonzalez-Cava, J.M.; Casteleiro-Roca, J.L.; Méndez-Pérez, J.A.; Antonio Reboso-Morales, J.; Javier Pérez-Castelo, F.; Javier de Cos Juez, F.; Luis Calvo-Rolle, J. Modelling the hypnotic patient response in general anaesthesia using intelligent models. Log. J. IGPL 2019, 27, 189–201. [Google Scholar] [CrossRef]
  8. Méndez, J.A.; Marrero, A.; Reboso, J.A.; León, A. Adaptive fuzzy predictive controller for anesthesia delivery. Control. Eng. Pract. 2016, 46, 1–9. [Google Scholar] [CrossRef]
  9. Haghighi, S.J.; Komeili, M.; Hatzinakos, D.; El Beheiry, H. 40-Hz ASSR for measuring depth of anaesthesia during induction phase. IEEE J. Biomed. Health Inform. 2017, 22, 1871–1882. [Google Scholar] [CrossRef]
  10. Shalbaf, A.; Saffar, M.; Sleigh, J.W.; Shalbaf, R. Monitoring the depth of anesthesia using a new adaptive neurofuzzy system. IEEE J. Biomed. Health Inform. 2017, 22, 671–677. [Google Scholar] [CrossRef]
  11. Raju, D.N.; Shanmugasundaram, H.; Sasikumar, R. Fuzzy segmentation and black widow–based optimal SVM for skin disease classification. Med. Biol. Eng. Comput. 2021, 59, 2019–2035. [Google Scholar] [CrossRef]
  12. Lee, H.C.; Ryu, H.G.; Chung, E.J.; Jung, C.W. Prediction of bispectral index during target-controlled infusion of propofol and remifentanil: A deep learning approach. Anesthesiology 2018, 128, 492–501. [Google Scholar] [CrossRef]
  13. Afshar, S.; Boostani, R.; Sanei, S. A combinatorial deep learning structure for precise depth of anesthesia estimation from EEG signals. IEEE J. Biomed. Health Inform. 2021, 25, 3408–3415. [Google Scholar] [CrossRef]
  14. Chen, Y.F.; Fan, S.Z.; Abbod, M.F.; Shieh, J.S.; Zhang, M. Nonlinear Analysis of Electroencephalogram Variability as a Measure of the Depth of Anesthesia. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  15. Dutt, M.I.; Saadeh, W. A Multilayer Perceptron (MLP) Regressor Network for Monitoring the Depth of Anesthesia. In Proceedings of the 2022 20th IEEE Interregional NEWCAS Conference (NEWCAS), Quebec City, QC, Canada, 19–22 June 2022; pp. 251–255. [Google Scholar]
  16. Plourde, G.; Boylan, J. The auditory steady state response during sufentanil anaesthesia. BJA Br. J. Anaesth. 1991, 66, 683–691. [Google Scholar] [CrossRef]
  17. Du, Y.; Wang, J.; Feng, W.; Pan, S.; Qin, T.; Xu, R.; Wang, C. Adarnn: Adaptive learning and forecasting of time series. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Gold Coast, QSD, Australia, 1–5 November 2021; pp. 402–411. [Google Scholar]
  18. Björnsson, M.A.; Norberg, Å.; Kalman, S.; Karlsson, M.O.; Simonsson, U.S. A two-compartment effect site model describes the bispectral index after different rates of propofol infusion. J. Pharmacokinet. Pharmacodyn. 2010, 37, 243–255. [Google Scholar] [CrossRef]
  19. Gonzalez-Cava, J.M.; Reboso, J.A.; Calvo-Rolle, J.L.; Mendez-Perez, J.A. Adaptive drug interaction model to predict depth of anesthesia in the operating room. Biomed. Signal Process. Control 2020, 59, 101931. [Google Scholar] [CrossRef]
  20. Yu, C.; Wang, J.; Chen, Y.; Huang, M. Transfer learning with dynamic adversarial adaptation network. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 778–786. [Google Scholar]
  21. Wang, J.; Chen, Y.; Feng, W.; Yu, H.; Huang, M.; Yang, Q. Transfer learning with dynamic distribution adaptation. ACM Trans. Intell. Syst. Technol. (TIST) 2020, 11, 1–25. [Google Scholar] [CrossRef]
  22. Li, H.; Pan, S.J.; Wang, S.; Kot, A.C. Domain generalization with adversarial feature learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5400–5409. [Google Scholar]
  23. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
  24. Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. Fitnets: Hints for thin deep nets. arXiv 2014, arXiv:1412.6550. [Google Scholar]
  25. Gao, Q.; Zhao, Y.; Li, G.; Tong, T. Image super-resolution using knowledge distillation. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 527–541. [Google Scholar]
  26. Hong, M.; Xie, Y.; Li, C.; Qu, Y. Distilling image dehazing with heterogeneous task imitation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3462–3471. [Google Scholar]
  27. Zhang, G.; Etemad, A. Distilling eeg representations via capsules for affective computing. arXiv 2021, arXiv:2105.00104. [Google Scholar] [CrossRef]
  28. Varvel, J.R.; Donoho, D.L.; Shafer, S.L. Measuring the predictive performance of computer-controlled infusion pumps. J. Pharmacokinet. Biopharm. 1992, 20, 63–94. [Google Scholar] [CrossRef] [PubMed]
  29. Kuznetsov, V.; Mohri, M. Generalization bounds for time series prediction with non-stationary processes. In Proceedings of the International Conference on Algorithmic Learning Theory, Bled, Slovenia, 8–10 October 2014; pp. 260–274. [Google Scholar]
  30. Oreshkin, B.N.; Carpov, D.; Chapados, N.; Bengio, Y. Meta-learning framework with applications to zero-shot time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 9242–9250. [Google Scholar]
  31. Ross, S.M. Introduction to Stochastic Dynamic Programming; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  32. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2096–3030. [Google Scholar]
  33. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 443–450. [Google Scholar]
  34. Schüttler, J.; Ihmsen, H. Population pharmacokinetics of propofol: A multicenter study. J. Am. Soc. Anesthesiol. 2000, 92, 727–738. [Google Scholar] [CrossRef]
  35. Lan, J.Y.; Abbod, M.F.; Yeh, R.G.; Fan, S.Z.; Shieh, J.S. Intelligent modeling and control in anesthesia. J. Med. Biol. Eng. 2012, 32, 293–307. [Google Scholar] [CrossRef]
  36. Kreuer, S.; Wilhelm, W.; Grundmann, U.; Larsen, R.; Bruhn, J. Narcotrend index versus bispectral index as electroencephalogram measures of anesthetic drug effect during propofol anesthesia. Anesth. Analg. 2004, 98, 692–697. [Google Scholar] [CrossRef]
  37. Jun, M.R.; Yoo, J.H.; Park, S.Y.; Na, S.; Kwon, H.; Nho, J.H. Assessment of phase-lag entropy, a new measure of electroencephalographic signals, for propofol-induced sedation. Korean J. Anesthesiol. 2019, 72, 351–356. [Google Scholar] [CrossRef]
  38. Jones, J.H.; Nittur, V.R.; Fleming, N.; Applegate, R.L. Simultaneous comparison of depth of sedation performance between SedLine and BIS during general anesthesia using custom passive interface hardware: Study protocol for a prospective, non-blinded, non-randomized trial. BMC Anesthesiol. 2021, 21, 105. [Google Scholar] [CrossRef]
  39. Kim, D.; Ahn, J.H.; Heo, G.; Jeong, J.S. Comparison of Bispectral Index and Patient State Index values according to recovery from moderate neuromuscular block under steady-state total intravenous anesthesia. Sci. Rep. 2021, 11, 5908. [Google Scholar] [CrossRef]
  40. Lee, H.C.; Jung, C.W. Vital Recorder—A free research tool for automatic recording of high-resolution time-synchronised physiological data from multiple anaesthesia devices. Sci. Rep. 2018, 8, 1527. [Google Scholar] [CrossRef]
  41. Shafer, S.L.; Gregg, K.M. Algorithms to rapidly achieve and maintain stable drug concentrations at the site of drug effect with a computer-controlled infusion pump. J. Pharmacokinet. Biopharm. 1992, 20, 147–169. [Google Scholar] [CrossRef] [PubMed]
  42. Lewis, S.R.; Pritchard, M.W.; Fawcett, L.J.; Punjasawadwong, Y. Bispectral index for improving intraoperative awareness and early postoperative recovery in adults. Cochrane Database Syst. Rev. 2019, 9, CD003843. [Google Scholar] [PubMed]
  43. Liu, Q.; Ma, L.; Fan, S.Z.; Abbod, M.F.; Shieh, J.S. Sample entropy analysis for the estimating depth of anaesthesia through human EEG signal at different levels of unconsciousness during surgeries. PeerJ 2018, 6, e4817. [Google Scholar] [CrossRef]
  44. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE). Geosci. Model Dev. Discuss. 2014, 7, 1525–1534. [Google Scholar]
  45. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  46. Zhang, Y.; Yan, J. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In Proceedings of the The Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  47. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
Figure 1. Visualization of the results of the baseline method [12]. The baseline method performs poorly during the induction and recovery periods. In addition, the baseline model predicts that the patient’s BIS is approximately 40 during the maintenance period. However, some patients have BIS values of approximately 50 or 30 during the maintenance period.
Figure 1. Visualization of the results of the baseline method [12]. The baseline method performs poorly during the induction and recovery periods. In addition, the baseline model predicts that the patient’s BIS is approximately 40 during the maintenance period. However, some patients have BIS values of approximately 50 or 30 during the maintenance period.
Sensors 23 08994 g001
Figure 2. An overview of our deep learning framework for DOA prediction. The teacher model has an extra input (i.e., the BIS history), which allows the teacher model to learn a more accurate representation of the DOA and transfer the DOA representation to the student model through various kinds of layers by knowledge distillation. The AdaRNN model includes temporal distribution characterization (TDC) and temporal distribution matching (TDM), which are used to determine the K-segment intervals with the largest distribution differences and reduce cross-domain distribution differences according to the drug infusion history, thereby improving the generalizability of the model. GRU denotes the gated recurrent unit, and FC denotes the fully connected network.
Figure 2. An overview of our deep learning framework for DOA prediction. The teacher model has an extra input (i.e., the BIS history), which allows the teacher model to learn a more accurate representation of the DOA and transfer the DOA representation to the student model through various kinds of layers by knowledge distillation. The AdaRNN model includes temporal distribution characterization (TDC) and temporal distribution matching (TDM), which are used to determine the K-segment intervals with the largest distribution differences and reduce cross-domain distribution differences according to the drug infusion history, thereby improving the generalizability of the model. GRU denotes the gated recurrent unit, and FC denotes the fully connected network.
Sensors 23 08994 g002
Figure 3. An overview of the temporal distribution characterization (TDC) approach, which divides the drug infusion history data into K intervals and obtains the largest distribution difference between every two intervals.
Figure 3. An overview of the temporal distribution characterization (TDC) approach, which divides the drug infusion history data into K intervals and obtains the largest distribution difference between every two intervals.
Sensors 23 08994 g003
Figure 4. Overview of the temporal distribution matching, which reduces cross-domain shifts in the K intervals according to the GRU output.
Figure 4. Overview of the temporal distribution matching, which reduces cross-domain shifts in the K intervals according to the GRU output.
Sensors 23 08994 g004
Figure 5. Visualization of the test cases with other compared methods in the VitalDB dataset.
Figure 5. Visualization of the test cases with other compared methods in the VitalDB dataset.
Sensors 23 08994 g005
Figure 6. Visualization of the test cases with other compared methods in our in-house dataset.
Figure 6. Visualization of the test cases with other compared methods in our in-house dataset.
Sensors 23 08994 g006
Figure 7. Ablation analysis of the various methods in all periods for different evaluation metrics. (a) RMSE; (b) MAE; (c) MDPE, where * denotes a negative value; (d) MDAPE.
Figure 7. Ablation analysis of the various methods in all periods for different evaluation metrics. (a) RMSE; (b) MAE; (c) MDPE, where * denotes a negative value; (d) MDAPE.
Sensors 23 08994 g007
Table 1. Different measurements of DOA type.
Table 1. Different measurements of DOA type.
DOA Type
Bispectral Index [35]Narcotrend Index [36]
Phase Lag Entropy [37]Entropy [35]
SedLine [38]Patient State Index [39]
Auditory Evoked Potential [35]Surgical Stress Index [35]
Table 2. Patient characteristics, mean ± standard deviation (min-max).
Table 2. Patient characteristics, mean ± standard deviation (min-max).
Training Data SetValidation Data SetTesting Data Set
N1807676
Age (yr)56.1 ± 14.0 (17–82)56.3 ± 15.0 (17–79)56.2 ± 15.1 (17–79)
Sex (male/female)113/6747/2940/36
Weight (kg)61.5 ± 10.2 (37.9–98.1)60.7 ± 10.3 (37.9–98.1)60.0 ± 9.8 (37.9–81.6)
Height (cm)163.2 ± 8.2 (138.8–186.6)162.3 ± 7.9 (138.8–182.0)161.2 ± 7.5 (138.8–182.0)
Median BIS41.1 ± 5.4 (25.9–59.5)43.1 ± 6.1 (23.1–57.2)42.5 ± 5.8 (30.6–55.9)
Propofol total dose (g)1.19 ± 0.63 (0.28–3.41)1.27 ± 0.71 (0.32–3.31)1.32 ± 0.71 (0.30–4.24)
Propofol median Ce (μg/mL)3.02 ± 0.47 (1.91–4.30)3.06 ± 0.49 (2.00–4.00)3.05 ± 0.50 (1.60–4.00)
Remifentanil total dose (g)1.46 ± 1.01 (0.29–6.29)1.43 ± 0.84 (0.25–3.70)1.46 ± 0.91 (0.34–5.16)
Remifentanil median Ce (μg/mL)3.73 ± 1.08 (1.50–6.01)3.67 ± 0.95 (2.00–6.97)3.70 ± 0.87 (2.00–6.00)
Table 3. Comparison of the metrics between our model and the compared model during three anesthesia periods in vitalDB dataset with BIS as the DOA.
Table 3. Comparison of the metrics between our model and the compared model during three anesthesia periods in vitalDB dataset with BIS as the DOA.
RMSEMAE
Anesthesia PeriodBaseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Baseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)
All10.20 ± 2.4510.08 ± 2.409.66 ± 2.379.40 ± 2.299.07 ± 1.768.07 ± 2.567.75 ± 2.397.61 ± 2.347.45 ± 2.307.29 ± 1.40
Induction14.57 ± 3.3913.75 ± 5.0013.64 ± 4.6613.09 ± 3.9010.91 ± 3.9312.80 ± 3.7911.45 ± 4.6311.63 ± 4.3911.14 ± 3.699.00 ± 3.54
Maintenance8.72 ± 2.788.56 ± 2.558.37 ± 2.628.16 ± 2.518.63 ± 2.067.12 ± 2.796.95 ± 2.586.80 ± 2.616.66 ± 2.527.00 ± 1.60
Recovery15.18 ± 6.4313.30 ± 5.4313.16 ± 5.9012.96 ± 5.4510.46 ± 2.9413.43 ± 6.4311.30 ± 5.1311.43 ± 5.6711.26 ± 5.228.74 ± 2.45
MDPE (%)MDAPE (%)
Anesthesia PeriodBaseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Baseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)
All−3.64 ± 14.961.76 ± 13.84−0.59 ± 14.050.57 ± 13.77−13.89 ± 4.9815.97 ± 7.9115.35 ± 7.6114.83 ± 6.6614.62 ± 6.4016.76 ± 3.53
Induction−6.35 ± 20.507.72 ± 16.846.79 ± 17.730.89 ± 18.54−3.15 ± 14.3022.75 ± 9.3921.90 ± 19.5018.88 ± 8.6918.87 ± 8.1415.64 ± 9.40
Maintenance−2.99 ± 15.241.17 ± 14.31−0.97 ± 14.480.56 ± 14.10−15.20 ± 5.5115.08 ± 8.3414.96 ± 8.0114.44 ± 8.3714.12 ± 8.9717.27 ± 4.14
Recovery−13.92 ± 25.441.94 ± 19.35−2.33 ± 21.69−3.98 ± 21.38−7.44 ± 11.5624.97 ± 16.0319.75 ± 12.9019.81 ± 11.1019.52 ± 11.1815.34 ± 6.22
Table 4. Difference in VitalDB and our dataset.
Table 4. Difference in VitalDB and our dataset.
Our DatasetVitalDB (Training Set)
N44180
Age (yr)39.9 ± 13.4 (19–69)56.1 ± 14.0 (17–82)
Sex (male/female)22/22113/67
Weight (kg)62.6 ± 10.5 (43–105)61.5 ± 10.2 (37.9–98.1)
Height (cm)166 ± 8.1 (147–183)163.2 ± 8.2 (138.8–186.6)
Median NI/BIS43.5 ± 10.1 (23.0–68.2)41.1 ± 5.4 (25.9–59.9)
Table 5. Comparison of metrics between our proposed model and other methods during three anesthesia periods in our in-house dataset with NI as the DOA.
Table 5. Comparison of metrics between our proposed model and other methods during three anesthesia periods in our in-house dataset with NI as the DOA.
RMSEMAE
Anesthesia PeriodBaseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Ours (without data preprocessing)Baseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Ours (without data preprocessing)
All13.30 ± 5.9812.19 ± 3.6711.43 ± 3.4310.44 ± 3.2310.10 ± 1.6110.64 ± 3.2411.40 ± 6.299.96 ± 3.569.80 ± 3.568.76 ± 2.987.52 ± 1.478.93 ± 3.12
Induction15.85 ± 5.7115.42 ± 3.8710.95 ± 2.588.01 ± 2.9221.30 ± 4.758.28 ± 3.3214.64 ± 5.9713.47 ± 4.439.35 ± 2.786.79 ± 2.6818.69 ± 5.706.72 ± 2.78
Maintenance12.22 ± 6.8010.56 ± 4.2410.66 ± 4.279.77 ± 3.897.18 ± 2.0310.15 ± 3.9710.72 ± 6.979.15 ± 4.179.37 ± 4.448.43 ± 3.665.87 ± 1.818.75 ± 3.86
Recovery15.12 ± 6.7214.60 ± 6.7114.75 ± 6.3213.88 ± 6.6910.12 ± 2.9613.97 ± 6.1313.80 ± 6.7613.54 ± 6.8913.72 ± 6.4212.80 ± 6.868.44 ± 2.7212.93 ± 6.31
MDPE (%)MDAPE (%)
Anesthesia PeriodBaseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Ours (without data preprocessing)Baseline [12]FEDformer [45]Crossformer [46]OursOurs (Transformer)Ours (without data preprocessing)
All−6.02 ± 22.622.71 ± 17.136.08 ± 16.851.34 ± 15.650.64 ± 8.024.67 ± 16.1323.23 ± 12.2119.66 ± 9.1719.39 ± 9.0918.29 ± 7.7513.08 ± 3.7118.05 ± 8.48
Induction15.16 ± 11.41−14.31 ± 11.23−6.13 ± 9.11−1.78 ± 7.99−27.21 ± 13.28−2.11 ± 8.3417.78 ± 8.1817.18 ± 7.7911.33 ± 4.987.74 ± 3.8728.00 ± 12.308.04 ± 4.47
Maintenance−8.77 ± 24.835.43 ± 18.508.87 ± 18.161.53 ± 18.013.64 ± 9.085.63 ± 18.1125.00 ± 13.9819.86 ± 10.2219.78 ± 10.8019.36 ± 9.4612.10 ± 4.9319.17 ± 10.10
Recovery−6.04 ± 32.20−9.73 ± 29.35−11.95 ± 28.23−4.07 ± 29.83−3.98 ± 9.76−2.59 ± 27.4425.49 ± 14.0827.27 ± 17.1928.33 ± 15.2825.16 ± 17.4614.29 ± 5.9723.46 ± 15.48
Table 6. Comparison of the statistic test between our proposed model and the compared methods.
Table 6. Comparison of the statistic test between our proposed model and the compared methods.
VitalDBIn-House
Statics TestPairt-tFPair-tF
Ours&Baseline0.0140.0270.0130.004
Ours&Fedformer0.0070.0360.0030.044
Ours&Crossformer0.0230.1630.0030.003
Table 7. Comparison of the model complexity between our proposed model and the compared methods.
Table 7. Comparison of the model complexity between our proposed model and the compared methods.
Paramters (M)FLOPS (G)
Baseline [12]0.10.0004
FEDformer [45]16.5139
Crossformer [46]11.480
Ours9.40.1
Ours (Transformer)6.251
Table 8. Comparison of the evaluation metrics between our TDM method and the boosting-based TDM method in three anesthesia periods.
Table 8. Comparison of the evaluation metrics between our TDM method and the boosting-based TDM method in three anesthesia periods.
MDPE (%)MDAPE (%)RMSEMAE
Anesthesia PeriodBoosting-based
TDM
OursBoosting-based
 TDM
OursBoosting-based
 TDM
OursBoosting-based
TDM
Ours
All5.22 ± 13.251.93 ± 13.3614.91 ± 6.1214.53 ± 6.189.93 ± 2.499.71 ± 2.297.95 ± 2.387.51 ± 2.29
Induction10.75 ± 17.545.45 ± 17.8319.01 ± 10.5618.64 ± 8.7314.24 ± 5.9313.09 ± 4.6012.17 ± 5.6511.31 ± 4.41
Maintenance5.06 ± 13.651.87 ± 13.7014.63 ± 6.6214.06 ± 6.738.69 ± 2.448.27 ± 2.427.19 ± 2.506.75 ± 2.44
Recovery0.84 ± 20.88−2.40 ± 21.0218.64 ± 11.7419.44 ± 10.8612.75 ± 6.2812.80 ± 5.2811.07 ± 5.9311.14 ± 5.15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, M.; He, Y.; Yang, Z. A Deep Learning Framework for Anesthesia Depth Prediction from Drug Infusion History. Sensors 2023, 23, 8994. https://doi.org/10.3390/s23218994

AMA Style

Chen M, He Y, Yang Z. A Deep Learning Framework for Anesthesia Depth Prediction from Drug Infusion History. Sensors. 2023; 23(21):8994. https://doi.org/10.3390/s23218994

Chicago/Turabian Style

Chen, Mingjin, Yongkang He, and Zhijing Yang. 2023. "A Deep Learning Framework for Anesthesia Depth Prediction from Drug Infusion History" Sensors 23, no. 21: 8994. https://doi.org/10.3390/s23218994

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop