Next Article in Journal
Characterization and Efficient Management of Big Data in IoT-Driven Smart City Development
Next Article in Special Issue
Classification of Motor Functions from Electroencephalogram (EEG) Signals Based on an Integrated Method Comprised of Common Spatial Pattern and Wavelet Transform Framework
Previous Article in Journal
Intrinsic Physical Unclonable Function (PUF) Sensors in Commodity Devices
Previous Article in Special Issue
A Novel Eye Movement Data Transformation Technique that Preserves Temporal Information: A Demonstration in a Face Processing Task
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FusionAtt: Deep Fusional Attention Networks for Multi-Channel Biomedical Signals

1
College of Information and Communication Engineering, Beijing University of Technology, Beijing 100124, China
2
Beijing Key Laboratory of Computational Intelligence and Intelligent System, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(11), 2429; https://doi.org/10.3390/s19112429
Submission received: 10 April 2019 / Revised: 24 May 2019 / Accepted: 25 May 2019 / Published: 28 May 2019

Abstract

:
Recently, pervasive sensing technologies have been widely applied to comprehensive patient monitoring in order to improve clinical treatment. Various types of biomedical signals collected by different sensing channels provide different aspects of patient health information. However, due to the uncertainty and variability in clinical observation, not all the channels are relevant and important to the target task. Thus, in order to extract informative representations from multi-channel biosignals, channel awareness has become a key enabler for deep learning in biosignal processing and has attracted increasing research interest in health informatics. Towards this end, we propose FusionAtt—a deep fusional attention network that can learn channel-aware representations of multi-channel biosignals, while preserving complex correlations among all the channels. FusionAtt is able to dynamically quantify the importance of each biomedical channel, and relies on more informative ones to enhance feature representation in an end-to-end manner. We empirically evaluated FusionAtt in two clinical tasks: multi-channel seizure detection and multivariate sleep stage classification. Experimental results showed that FusionAtt consistently outperformed the state-of-the-art models in four different evaluation measurements, demonstrating the effectiveness of the proposed fusional attention mechanism.

1. Introduction

With the broad adoption of pervasive sensors, comprehensive patient monitoring becomes feasible for a wide range of medical applications [1]. Various types of biomedical signals collected by different sensing channels (i.e., multi-channel biosignals) provide abundant physiological information and reflect different aspects of patients’ health condition. For instance, multi-channel scalp electroencephalogram (EEG) measures electrical activity in different brain areas, and each of these channels can be regarded as a sensing source providing unique brain information. It is applicable to several clinical tasks, including schizophrenia diagnosis [2], emotion recognition [3], and epileptic seizure detection [4]. In order to learn meaningful representations for the multi-channel biosignals, significant efforts were recently made to explore feature extraction techniques using deep neural networks [5,6]. These deep learning approaches are designed to capture abstract characteristics of signal patterns among all the biomedical channels, referred to as multi-view learning. However, not all the channels (or views) are relevant and important to the target task due to the uncertainty and variability in clinical observation (e.g., different brain regions are often involved in different types of brain activity) [4]. The irrelevance and redundant raw features may influence the model performance. Intuitively, an ideal deep learning approach should be able to distinguish the task-related importance among different channels and rely on more informative ones to enhance feature representation. Thus, channel awareness has become a key enabler for deep learning in biosignal processing, and has recently attracted increasing research interest in health informatics.
To address the channel awareness issue, several researchers have attempted to combine deep learning with a channel selection module, regarded as a multi-stage model, to mitigate the influence of irrelevance and redundancy in the raw feature space [4,7,8]. It has been validated that such deep-learning-based channel selection methods can improve the performance in modeling multi-channel biomedical data. However, these methods adopt the same selection strategy: only the critical channels are determined as the input to train the model, while the rest of the channels are one-off eliminated. Utilizing such a hard channel selection procedure ignores the detailed task-related information among channels over different situations. Take EEG seizure detection as an example: subjects have epileptic seizures to different extents (i.e., different brain regions), and hence the importance of each channel varies significantly across individuals. Moreover, since the existing deep-learning-based models extract features and select channels separately (i.e., multi-stage training), they do not consistently make all the functional modules work together, rendering it a challenging task to develop a unified channel-aware deep learning model.
To this end, we propose FusionAtt, a deep fusional attention network, to extract channel-aware representative features from multi-channel biosignals. We developed a new fusional attention layer that adopts a fusion gate to fully incorporate multi-view information, in order to dynamically measure the contribution of each biomedical channel. A multi-view convolutional encoding layer combined with convolutional neural networks (CNNs) [9] is also adopted to train a unified deep learning model. Experimental results conducted on different clinical datasets demonstrate that FusionAtt consistently obtained better performance than nine biosignal feature learning baselines in terms of four evaluation metrics. We summarize our main contributions as follows:
  • We propose FusionAtt, a unified fusional attention neural network combined with multi-view convolutional encoder, designed to learn channel-aware representations of multi-channel biosignals.
  • FusionAtt dynamically quantifies the importance of each biomedical channel by gated fusion, and relies on more informative ones to enhance feature representation, without prior expert knowledge.
  • We empirically show that FusionAtt consistently achieved the best performance compared with nine biosignal feature learning baselines on two clinical datasets, demonstrating the effectiveness of the proposed fusional attention mechanism.
The rest of the paper is organized as follows: We first review the related work in the next section. In Section 3, we present the details of the proposed FusionAtt model. The experimental results are then discussed in Section 4. Finally, we conclude this work in Section 5.

2. Related Work

In this section, we summarize the literature related to our work in the following two categories: deep learning for multi-channel biosignals, and attention-based neural networks in clinical diagnosis.

2.1. Deep Learning for Multi-Channel Biosignals

Learning deep representations of biomedical data is crucial in the healthcare domain. The existing studies on deep learning for multi-channel biosignals are diverse due to the wide range of medical applications. Deep belief networks (DBNs) have been widely adopted to learn inherent representations of polysomnography (PSG) signals to classify sleep stages [10,11]. To capture temporal patterns from multi-channel waveform data, context learning was employed for emotion recognition [3] and seizure detection [12,13]. CNN-based approaches were also proposed to fuse information from different sensing sources into a high-order feature space [14,15]. Compared with these models that utilize parameter sharing architecture to extract joint features among channels, we integrate features explicitly according to the relative significance of each biomedical channel.
More recently, several researchers have attempted to combine deep neural networks with a channel selection module in order to mitigate the influence of irrelevance and redundancy in the raw feature space. Yuan et al. [4] considered the response energy of stacked autoencoders (SAEs) to jointly determine critical EEG channels. Li et al. [7] and Jia et al. [8] proposed different DBN-based channel selection modules to recognize affective state from multi-channel biosignal data. However, these advanced methods, though yielding reasonably good performance, adopt hard channel selection strategies and are not unified models. In contrast, we propose the fusional attention mechanism to dynamically select and fuse channel information in an end-to-end manner.

2.2. Attention-Based Neural Networks in Clinical Diagnosis

Recently, attention mechanisms have attracted increasing research interest in clinical diagnosis, due to their strong ability of feature extraction and model interpretability. Choi et al. [16] and Ma et al. [17] adopted attention mechanisms to explain medical codes (e.g., procedure, diagnosis, and medication codes) from electronic health records (EHRs). Yuan et al. [18] first exploited an attention mechanism based on multi-view learning (i.e., ChannelAtt) to achieve soft channel selection for EEG seizure detection. In general, there are two differences between ChannelAtt and our proposed FusionAtt model. First, ChannelAtt uses local information and concatenated information to assign attention energy, referred to as local attention and global attention, respectively. In our FusionAtt model, we adopt a gated function to better fuse the multi-view information for the attention energy assignment. Second, the multi-view representation module in ChannelAtt is based on a SAE, a fully-connected neural network, while ours is based on a CNN, a locally-connected neural network, which can better extract deep features from multi-channel biosignals.

3. Methodology

In this section we present the technical details of our FusionAtt model with multi-channel biosignal inputs. The architecture of FusionAtt is illustrated in Figure 1. In the following subsections, we detail the main components of our FusionAtt model.

3.1. Multi-View Convolutional Encoder

In order to learn deep representations of multi-channel waveform data, simply concatenating the raw inputs of all the channels may not be enough to preserve the unique characteristics of each channel. Inspired by the rapid development of multi-view deep learning techniques [4,15,19], we propose the utilization of two convolutional feature encoders (i.e., channel-encoder and global-encoder) to extract abstract features from channel-specific and global views, respectively.
Formally, we assume that the input biosignal fragments consist of C channels, denoted as X = { x 1 , , x i , , x C } . Given the input vector of the i-th channel, denoted as x i R n , a channel-view representation h i R p can be obtained using the channel-encoder (i.e., E n c o d e r c ), as follows:
h i = E n c o d e r c ( x i ; θ c ) ,
where θ c is the learnable parameter set in E n c o d e r c . Similarly, we can calculate a global-view representation h g R p through the global-encoder (i.e., E n c o d e r g ), as follows:
h g = E n c o d e r g ( x 1 : C ; θ g ) ,
where θ g is the learnable parameter set E n c o d e r g . Generally speaking, both E n c o d e r c and E n c o d e r g can be parameterized by different deep learning methods designed for feature extraction.
In our model, we construct the multi-view convolutional encoder by stacking several multi-kernel CNN cells consisting of convolutional, nonlinear, and pooling layers. Specifically, regarding our two feature encoders, the latent multi-view representations of the k-th feature map, denoted as h ˜ i ( k ) and h ˜ g ( k ) , can be obtained as follows:
h ˜ i ( k ) = f ( x i * W c ( k ) + b c ( k ) ) ,
h ˜ g ( k ) = f ( i = 1 C x i * W g ( k ) + b g ( k ) ) ,
where θ g ( k ) = { W c ( k ) , b c ( k ) } and θ c ( k ) = { W g ( k ) , b g ( k ) } are the learnable parameters. Subsequently, we flatten all the features extracted by different kernels and derive the global-view and channel-view representations, that is, h i and h g , respectively. Note that the dimension of h ˜ i ( k ) and h ˜ g ( k ) relies on the structure configuration of the multi-view convolutional encoder, which is given in Section 4.3.

3.2. Fusional Attention Mechanism

In order to dynamically qualify the contribution of each biomedical channel, we propose the fusional attention mechanism, which incorporates a gated function for the final task. Specifically, the fusion gate r i R can be calculated considering both the global-view and channel-view representations, defined as:
r i = σ ( W r g h g + W r c h i + b r c ) ,
where θ a ( 1 ) = { W r g , W r c R p , b r c R } denotes the learnable parameter set. Here in Equation (5), we rescale r i into the range of [ 0 , 1 ] by adopting the sigmoid function σ ( · ) , in order to control the flow of multi-view information through the neural networks. We then integrate the information of the global-view and its own channel-view representations according to the fusion gate r i , defined as:
h ¯ i = ( 1 r i ) h g + r i h i ,
where ⊙ is the element-wise multiplication operator. According to Equation (6), the fusion gate r i is able to learn how much information carried by each encoder is relevant to keep or forget during end-to-end training. If r i = 1 , then h ¯ i = h i . This means that only channel-view information is passed. if r i = 0 , then h ¯ i = h g . This means that only global information is passed. We use the gated unit to derive a more representative integrated feature vector, that is, h ¯ i R p , as the input of the attention energy assignment function.
The attention energy e g , i of the i-th channel can be further assigned based on the integrated feature vector h ¯ i , as follows:
e g , i = W e h ¯ i + b e ,
where θ a ( 2 ) = { W e R p , b e R } is the learnable parameter set. Given all the attention energy values, the contribution score vector α g R C can be normalized using softmax function, as follows:
α g = S o f t m a x ( [ e g , 1 , , e g , i , , e g , C ] ) .
Intuitively, if the contribution score α g , i of the i-th channel is large, the information of the i-th channel is high related to the corresponding task label. Subsequently, we use weighted aggregation to compute a context vector c g R p based on the integrated features h ¯ i ( 1 i C ) and the contribution score vector α g , as follows:
c g = i = 1 C α g , i h ¯ i .
In this way, our model is able to effectively incorporate the multi-view information carried by both feature views, and hence fuse representative features from multi-channel biosignals.

3.3. Unified Training Procedure

To train our FusionAtt model in an end-to-end manner, we combine the context vector with the global-view vector to derive an attentional representation h α R r , defined as:
h α = f ( W h [ c g h g ] + b h ) ,
where ⊕ is the concatenation operator, and W h R r × 2 p and b h R r are the learnable parameters. Finally, a softmax layer is applied to produce the classification task, as follows:
y ^ = S o f t m a x ( W s h α + b s ) ,
where W s R | c l a s s | × r and b s R | c l a s s | denote the learnable parameters. Here we employ cross-entropy as the classification loss. Given M training samples { ( X ( m ) , y ( m ) ) } m = 1 M , the cost function of our unified FusionAtt network in terms of the learnable parameter set Θ = { θ c , θ g , θ a , W h , s , b h , s } , is defined as:
J F u s i o n A t t ( X ( 1 ) , , X ( m ) , , X ( M ) ; Θ ) = 1 M m = 1 M y ( m ) log y ^ ( m ) + ( 1 y ( m ) ) log ( 1 y ^ ( m ) ) .

4. Experiments

In this section, we evaluate FusionAtt using two benchmark clinical tasks: multi-channel EEG seizure detection and multivariate PSG sleep stage classification. We first introduce the two datasets, then describe the baselines and implementation details. We finally present the quantitative results and analyze the learned contribution scores through a clinical case study.

4.1. Datasets

CHB-MIT. We performed the task of multi-channel EEG seizure detection using the public CHB-MIT dataset provided by the Children’s Hospital Boston [20]. This dataset contains 23-channel 256 Hz EEG signals. All the seizures are manually labeled by medical experts. Following the segmentation experience [21], we set both window length and step length as 1 second, and finally generated 252 , 862 input vectors from all 23 subjects.
UCD. We conducted experiments for the multivariate PSG sleep stage classification task based on the UCD dataset collected from St. Vincent’s University Hospital and University College Dublin [22]. This dataset contains 14-channel overnight PSG data, consisting of 128 Hz EEG, 64 Hz electromyography (EMG), and other types of biosignals. We generated 287,840 input vectors from all 25 subjects, and each 30-second fragment is labeled as being in one of the five sleep stages.

4.2. Baseline Approaches

We compared FusionAtt with the following nine existing biosignal feature learning baselines:
Support vector machine (SVM) [23]. SVM is a classic machine learning method. Here we adopted a one-vs-all SVM for the multi-class classification. To avoid the curse of dimensionality, we used principal component analysis (PCA) [24] to select the top-r related components from all channels as features before training the SVM, namely, PSVM.
SAE [25]. SAE is a widely adopted deep learning method for biosignal feature learning. For the sake of fairness, we incorporated global and multi-view strategies to extend the SAE model, referred to as GSAE and MSAE, respectively.
CNN [9]. The CNN is another commonly used deep learning method in biosignal processing. Similarity, we performed the same process as with SAE (i.e., GCNN and MCNN).
CtxFusionEEG [13]. CtxFusionEEG is a multi-stage feature learning model that focuses on using context learning to detect temporal patterns from epilepsy EEG signals. This method derives fusional representations of multi-channel biosignals by incorporating both deep learning and handcrafted engineering.
mSSDA [4]. mSSDA is an SAE-based multi-view deep-learning variant for multi-channel biosignals. This method determines channels by adopting a hard selection procedure after the deep feature extraction.
ChannelAtt [18]. ChannelAtt adopts fully connected multi-view learning to soft-select critical channels from multi-channel biosignals. According to the original model, both the local and global attention mechanisms are included, referred to as ChannelAtt l o c and ChannelAtt g l o , respectively.

4.3. Implementation Details

We implemented our proposed model with PyTorch. Regarding the evaluation procedure, 5-fold subject-independent cross-validation was adopted, and we report the average test results for performance comparisons. The ratio of training, validation, and test sets was 0.7:0.1:0.2. Furthermore, we utilized short-time Fourier transform (STFT) for data preprocessing. Adadelta was adopted for the training process to optimize the cost function in terms of the learnable parameters. We also used weight decay with a 0 . 001 L 2 penalty coefficient, 0 . 95 momentum, and 0 . 5 dropout rate to train all the deep learning methods. Table 1 lists the structure configuration of the multi-view convolutional encoder in FusionAtt, and we set p = 128 and r = 128 for our model and baselines.
To quantify the performance, both Accuracy and F1-score were adopted for evaluation. The area-under-the-curve of precision–recall (AUC-PR) and receiver operator characteristic (AUC-ROC) were also utilized to evaluate each approach. Note that we used the Macro metric for the multi-class classification task.

4.4. Performance on Clinical Tasks

We investigated the effectiveness of our proposed FusionAtt model compared to the baseline methods in different biomedical tasks. Table 2 reports the experimental results of all the aforementioned approaches on all the benchmark datasets. We can see that our FusionAtt model achieved the best performance compared with the corresponding baselines utilizing different encoder methods on both datasets.
Given the results of the baselines, the traditional classification method PSVM outperformed the SAE-based models in both tasks. This demonstrates the effectiveness of the handcrafted PCA features where SVM could find a relatively clear hyper-lane to separate classes in the vector space. Among all the deep learning baselines, we observed that the multi-view-based methods obtained better results than the globally based methods. This illustrates the advantage of multi-view representation, in which the meaningful features are effectively extracted from multi-channel biosignal data. The improvement of the CNN-based methods compared with the SAE-based methods demonstrates that all the deep learning models take advantage of the CNN feature encoder. The reason is that the spatial information of channels is retained and the classifier can hence capture more detailed information from multi-channel biosignals. Since mSSDA employs the hard channel selection procedure, all the evaluation measurements increase compared with the SAE-based methods. This suggests the benefit of channel selection on multi-channel biosignals to capture critical information. CtxFusionEEG performed on par with mSSDA in terms of accuracy, but achieved a much worse F1-score than mSSDA, even the other deep learning baselines. This means that adopting multi-stage training would not consistently make all the functional modules work together to yield good results across different classes.
From the results of the attention-based models, the comparisons between ChannelAtt l o c and ChannelAtt g l o indicate that the concatenated information captured by the global attention mechanism is helpful for feature extraction. Our proposed FusionAtt model performed better than all the baselines, due to the effective integration of the multi-view representation and fusional attention mechanism. Moreover, Figure 2 and Figure 3 illustrate the ROC and PR curves of all the test folds on the CHB-MIT and UCD datasets, respectively. We can see that FusionAtt achieved the best performance in terms of the AUC-ROC and AUC-PR. This means that FusionAtt was able to focus on providing complementary information toward each view representations, and hence could help to learn more informative features from multi-channel biosignal data. According to the overall analysis on both datasets, we conclude that the fusional attention mechanism is key to identifying critical patterns from biomedical signals, and the attentional representation learned by FusionAtt can improve the performance for different biomedical tasks in healthcare.

4.5. Sensitivity Analysis

In this subsection, we discuss the effects of various hyper-parameter choices in FusionAtt, including the dimensionality of multi-view representation p and the size of attentional representation r. Specifically, we plot the accuracy results of our proposed FusionAtt model and the two ChannelAtt variants under different hyper-parameter settings on both datasets. Here we use the hyper-parameter setting mentioned in Section 4.3 as the basic configuration, and vary one from 32 to 512 while keeping others fixed to the basic configuration.
Figure 4 illustrates the variation of accuracy under different sizes of multi-view representation p using different feature encoders on the two datasets. From the figures, we observe that when the dimension was small, all the models resulted in limited accuracy performance, demonstrating that a smaller size of p led to worse capture of the multi-channel information. When we increased p, the model performance increased rapidly. However, when p was too large, the performance decreased slightly. This shows that properly setting the size of multi-view representation can help to generate robust features to improve the performance, while excessive hidden neurons negatively impact the model performance. Similarity, as shown in Figure 5, we can see the same trends of the influence under different sizes of attentional representation r. Among all the figures, regarding the feature encoders, the proposed FusionAtt model consistently beat the two ChannelAtt models. This resulted from the fact that the multi-view convolutional encoder has strong model generalization capability. Moreover, the FusionAtt model was less sensitive to different choices of hyper-parameters than the other models. This demonstrates that the fusion gate in Equation (5) enables FusionAtt to dynamically fuse features from multiple views, deriving more comprehensive features from multi-channel biosignals.

4.6. Case Study

In this subsection, we discuss the learned contribution scores in FusionAtt to justify the benefit of adopting the fusional attention mechanism in clinical settings. Figure 6 presents a clinical case study of multi-channel EEG seizure detection on the CHB-MIT dataset where a subject suffered epileptic seizure from the 6-th to 13-th timestamps. We display the mean scores of every 5 fragments and mark the max value among all the channels within each timestamp for clear visualization. Note that according to the EEG seizure detection results discussed in Section 4.4, the contribution scores learned from all of the channel-aware attention variants were similar, due to the similar attention learning scenarios.
From Figure 6, we can observe that the learned contribution scores of all the channels were different within different timestamps. Analyzing the timestamps that belong to the non-seizure stage, the distribution of contribution scores was relatively uniform. Therefore, no seizure pattern was recognized, and hence our model paid equal attention to all the EEG channels in detecting seizure onset. Regarding the timestamps in the seizure stage, we can observe that the 17-th and 18-th channels were the two most active ones that significantly contributed to the final task of seizure detection. This indicates that the seizure onset was highly related to the brain regions measured by these two channels. According to the International 10–20 system employed in this dataset [20], the seizure onset in this case was located in the central region of the brain, which conforms to the channels selected by our model. To sum up, the case study justifies the ability of our proposed FusionAtt model to learn interpretable contribution scores with clinical meanings. According to Table 2, the meaningful representations support the effectiveness of FusionAtt for good performance in seizure detection.

5. Conclusions

In this paper, we present deep fusional attention networks, namely FusionAtt, to learn deep representations of multi-channel biosignals for clinical tasks. The proposed FusionAtt is a unified deep learning framework that fuses gated attentional features by capturing dependencies among biomedical channels based on multi-view convolutional encoding. Experiments on two benchmark clinical tasks showed that FusionAtt was able to efficiently fuse channel-aware information from multi-channel biosignals. The case study intuitively showed how FusionAtt was able to interpret influential clinical observations by analyzing the learned contribution scores of biomedical channels.
Our future work will include the incorporation of different deep-learning architectures for feature encoding, in addition to investigating the performance of FusionAtt on larger datasets. In addition, FusionAtt can be extended to other task-oriented applications with similar data structure where the channel awareness is still a major challenge.

Author Contributions

Y.Y. and K.J. conceived the idea and designed the experiments; Y.Y. programmed the model, carried out the experiments, and wrote most of the manuscript; All the authors revised and approved the final manuscript.

Funding

This research was funded by the National Science Foundation of China under Grant 81871394 and 61672064, and the Beijing Laboratory of Advanced Information Networks under Grant 040000546618017.

Acknowledgments

The authors would like to thank the anonymous reviewers, and NVIDIA Corporation for the donation of the Titan Xp GPU.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Johnson, A.E.; Ghassemi, M.M.; Nemati, S.; Niehaus, K.E.; Clifton, D.A.; Clifford, G.D. Machine learning and decision support in critical care. Proc. IEEE Inst. Electr. Electr. Eng. 2016, 104, 444. [Google Scholar] [CrossRef] [PubMed]
  2. Acar, E.; Levin-Schwartz, Y.; Calhoun, V.D.; Adali, T. Tensor-based fusion of EEG and FMRI to understand neurological changes in schizophrenia. In Proceedings of the 2017 IEEE International Symposium on Circuits and Systems (ISCAS), Baltimore, MD, USA, 28–31 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–4. [Google Scholar] [Green Version]
  3. Li, X.; Jia, X.; Xun, G.; Zhang, A. Improving eeg feature learning via synchronized facial video. In Proceedings of the 2015 IEEE International Conference on Big Data (Big Data), Santa Clara, CA, USA, 29 Octorber–1 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 843–848. [Google Scholar]
  4. Yuan, Y.; Xun, G.; Jia, K.; Zhang, A. A multi-view deep learning method for epileptic seizure detection using short-time fourier transform. In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Boston, MA, USA, 20–23 August 2017; ACM: New York, NY, USA, 2017; pp. 213–222. [Google Scholar]
  5. Längkvist, M.; Karlsson, L.; Loutfi, A. A review of unsupervised feature learning and deep learning for time-series modeling. Pattern Recognit. Lett. 2014, 42, 11–24. [Google Scholar] [CrossRef] [Green Version]
  6. Supratak, A.; Wu, C.; Dong, H.; Sun, K.; Guo, Y. Survey on feature extraction and applications of biosignals. In Machine Learning for Health Informatics; Springer: Berline, Germany, 2016; pp. 161–182. [Google Scholar]
  7. Li, K.; Li, X.; Zhang, Y.; Zhang, A. Affective state recognition from EEG with deep belief networks. In Proceedings of the 2013 IEEE International Conference on Bioinformatics and Biomedicine, Shanghai, China, 18–21 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 305–310. [Google Scholar]
  8. Jia, X.; Li, K.; Li, X.; Zhang, A. A novel semi-supervised deep learning framework for affective state recognition on eeg signals. In Proceedings of the 2014 IEEE International Conference on Bioinformatics And Bioengineering, Boca Raton, FL, USA, 10–12 November 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 30–37. [Google Scholar]
  9. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, J.; Wu, Y.; Bai, J.; Chen, F. Automatic sleep stage classification based on sparse deep belief net and combination of multiple classifiers. Trans. Inst. Meas. Control 2016, 38, 435–451. [Google Scholar] [CrossRef]
  11. Längkvist, M.; Karlsson, L.; Loutfi, A. Sleep stage classification using unsupervised feature learning. Adv. Artif. Neural Syst. 2012, 2012, 5. [Google Scholar] [CrossRef]
  12. Xun, G.; Jia, X.; Zhang, A. Detecting epileptic seizures with electroencephalogram via a context-learning model. BMC Med. Inf. Dec. Mak. 2016, 16, 70. [Google Scholar] [CrossRef] [PubMed]
  13. Yuan, Y.; Xun, G.; Jia, K.; Zhang, A. A novel wavelet-based model for eeg epileptic seizure detection using multi-context learning. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 694–699. [Google Scholar]
  14. Ha, S.; Choi, S. Convolutional neural networks for human activity recognition using multiple accelerometer and gyroscope sensors. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 381–388. [Google Scholar]
  15. Yao, S.; Hu, S.; Zhao, Y.; Zhang, A.; Abdelzaher, T. Deepsense: A unified deep learning framework for time-series mobile sensing data processing. In Proceedings of the 26th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, Perth, Australia, 3–7 April 2017; pp. 351–360. [Google Scholar]
  16. Choi, E.; Bahadori, M.T.; Song, L.; Stewart, W.F.; Sun, J. GRAM: graph-based attention model for healthcare representation learning. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; ACM: New York, NY, USA, 2017; pp. 787–795. [Google Scholar]
  17. Ma, F.; Chitta, R.; Zhou, J.; You, Q.; Sun, T.; Gao, J. Dipole: Diagnosis prediction in healthcare via attention-based bidirectional recurrent neural networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; ACM: New York, NY, USA, 2017; pp. 1903–1911. [Google Scholar]
  18. Yuan, Y.; Xun, G.; Ma, F.; Suo, Q.; Xue, H.; Jia, K.; Zhang, A. A novel channel-aware attention framework for multi-channel eeg seizure detection via multi-view deep learning. In Proceedings of the 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 206–209. [Google Scholar]
  19. Zhao, J.; Xie, X.; Xu, X.; Sun, S. Multi-view learning overview: Recent progress and new challenges. Inf. Fus. 2017, 38, 43–54. [Google Scholar] [CrossRef]
  20. Shoeb, A.H. Application of Machine Learning to Epileptic Seizure Onset Detection and Treatment. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2009. [Google Scholar]
  21. Yuan, Y.; Xun, G.; Suo, Q.; Jia, K.; Zhang, A. Wave2vec: Learning deep representations for biosignals. In Proceedings of the 2017 IEEE International Conference on Data Mining (ICDM), New Orleans, LA, USA, 18–21 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1159–1164. [Google Scholar]
  22. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed]
  23. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  24. Jolliffe, I. Principal Component Analysis; Springer: Berline, Germany, 2011. [Google Scholar]
  25. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Architecture of the proposed FusionAtt model.
Figure 1. Architecture of the proposed FusionAtt model.
Sensors 19 02429 g001
Figure 2. Receiver operating characteristic (ROC) and precision–recall (PR) curves of the proposed method and the baselines on the CHB-MIT dataset. (a) ROC curves; (b) PR curves.
Figure 2. Receiver operating characteristic (ROC) and precision–recall (PR) curves of the proposed method and the baselines on the CHB-MIT dataset. (a) ROC curves; (b) PR curves.
Sensors 19 02429 g002
Figure 3. ROC and PR curves of the proposed method and the baselines on the UCD dataset. (a) ROC curves; (b) PR curves.
Figure 3. ROC and PR curves of the proposed method and the baselines on the UCD dataset. (a) ROC curves; (b) PR curves.
Sensors 19 02429 g003
Figure 4. Performance variations with different dimensions of multi-view representation. (a) Sensitivity analysis on the CHB-MIT dataset; and (b) Sensitivity analysis on the UCD dataset.
Figure 4. Performance variations with different dimensions of multi-view representation. (a) Sensitivity analysis on the CHB-MIT dataset; and (b) Sensitivity analysis on the UCD dataset.
Sensors 19 02429 g004
Figure 5. Performance variations with different dimensions of attentional representation. (a) Sensitivity analysis on the CHB-MIT dataset; (b) Sensitivity analysis on the UCD dataset.
Figure 5. Performance variations with different dimensions of attentional representation. (a) Sensitivity analysis on the CHB-MIT dataset; (b) Sensitivity analysis on the UCD dataset.
Sensors 19 02429 g005
Figure 6. Multi-channel electroencephalogram (EEG) seizure detection of a patient in the case study.
Figure 6. Multi-channel electroencephalogram (EEG) seizure detection of a patient in the case study.
Sensors 19 02429 g006
Table 1. Structure configuration of the multi-view convolutional encoder. ReLU: rectified linear unit.
Table 1. Structure configuration of the multi-view convolutional encoder. ReLU: rectified linear unit.
Cell No.ConvNon-linearPooling
E n c o d e r c 1 ( 8 , 16 , 32 , 64 ) × 8 ReLU6
E n c o d e r c 2 ( 3 , 5 ) × 16 ReLU3
E n c o d e r c 3 ( 3 , 5 ) × 16 ReLU3
E n c o d e r g 1 ( 1 × 8 , 1 × 16 , 1 × 32 , 1 × 64 ) × 8 ReLU 1 × 6
E n c o d e r g 2 ( 3 × 3 , 5 × 5 ) × 16 ReLU 1 × 3
E n c o d e r g 3 ( 3 × 3 , 5 × 5 ) × 16 ReLU C × 3
Table 2. Performance comparisons on two clinical biosignal datasets. GCNN: global extension of convolutional neural network (CNN); MCNN: multi-view extension of CNN; GSAE: global extension of stacked autoencoder (SAE); MSAE: multi-view extension of SAE.
Table 2. Performance comparisons on two clinical biosignal datasets. GCNN: global extension of convolutional neural network (CNN); MCNN: multi-view extension of CNN; GSAE: global extension of stacked autoencoder (SAE); MSAE: multi-view extension of SAE.
MethodSeizure Detection (CHB-MIT)Sleep Stage Classification (UCD)
AUC-ROCAUC-PRF1-ScoreAccuracyAUC-ROCAUC-PRF1-ScoreAccuracy
PSVM0.82910.70210.64210.87680.81770.57640.52040.6193
GSAE0.59340.41800.06680.79870.70680.49650.27600.4917
MSAE0.75290.49370.14790.80130.72130.52240.35420.5262
GCNN0.92550.80540.75060.88490.86550.65890.50420.6270
MCNN0.92630.87020.79590.90880.87320.67250.59250.6590
CtxFusionEEG0.92870.78330.72020.90250.84830.53300.46800.6688
mSSDA0.94500.88010.81860.93640.85440.62080.59690.6741
ChannelAtt l o c 0.95540.91340.86250.94770.86990.63700.58900.6855
ChannelAtt g l o 0.95560.91190.86750.95060.86620.64580.61370.6859
FusionAtt0.97010.91450.89530.96220.88940.70210.66370.7257

Share and Cite

MDPI and ACS Style

Yuan, Y.; Jia, K. FusionAtt: Deep Fusional Attention Networks for Multi-Channel Biomedical Signals. Sensors 2019, 19, 2429. https://doi.org/10.3390/s19112429

AMA Style

Yuan Y, Jia K. FusionAtt: Deep Fusional Attention Networks for Multi-Channel Biomedical Signals. Sensors. 2019; 19(11):2429. https://doi.org/10.3390/s19112429

Chicago/Turabian Style

Yuan, Ye, and Kebin Jia. 2019. "FusionAtt: Deep Fusional Attention Networks for Multi-Channel Biomedical Signals" Sensors 19, no. 11: 2429. https://doi.org/10.3390/s19112429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop