Next Article in Journal
The Use of Pulse Oximetry in the Assessment of Acclimatization to High Altitude
Next Article in Special Issue
SEED-G: Simulated EEG Data Generator for Testing Connectivity Algorithms
Previous Article in Journal
Characterization of Tiled Architecture for C-Band 1-Bit Beam-Steering Transmitarray
Previous Article in Special Issue
An Open Dataset for Wearable SSVEP-Based Brain-Computer Interfaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition

1
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
2
MoE Key Laboratory of Advanced Perception and Intelligent Control of High-End Equipment, Anhui Polytechnic University, Wuhu 241000, China
3
Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1262; https://doi.org/10.3390/s21041262
Submission received: 26 January 2021 / Revised: 5 February 2021 / Accepted: 6 February 2021 / Published: 10 February 2021

Abstract

:
Emotion recognition has a wide range of potential applications in the real world. Among the emotion recognition data sources, electroencephalography (EEG) signals can record the neural activities across the human brain, providing us a reliable way to recognize the emotional states. Most of existing EEG-based emotion recognition studies directly concatenated features extracted from all EEG frequency bands for emotion classification. This way assumes that all frequency bands share the same importance by default; however, it cannot always obtain the optimal performance. In this paper, we present a novel multi-scale frequency bands ensemble learning (MSFBEL) method to perform emotion recognition from EEG signals. Concretely, we first re-organize all frequency bands into several local scales and one global scale. Then we train a base classifier on each scale. Finally we fuse the results of all scales by designing an adaptive weight learning method which automatically assigns larger weights to more important scales to further improve the performance. The proposed method is validated on two public data sets. For the “SEED IV” data set, MSFBEL achieves average accuracies of 82.75%, 87.87%, and 78.27% on the three sessions under the within-session experimental paradigm. For the “DEAP” data set, it obtains average accuracy of 74.22% for four-category classification under 5-fold cross validation. The experimental results demonstrate that the scale of frequency bands influences the emotion recognition rate, while the global scale that directly concatenating all frequency bands cannot always guarantee to obtain the best emotion recognition performance. Different scales provide complementary information to each other, and the proposed adaptive weight learning method can effectively fuse them to further enhance the performance.

1. Introduction

Developing automatic and accurate emotion recognition technologies has gained more and more attention due to its wide range of potential applications. In engineering, it facilitates the human–machine interaction more friendly, where machines might understand emotions and interact with us according to our emotions [1,2]. In the medical research, it is beneficial for diagnosing and treating various mental diseases, such as depression and autism spectrum disorders [3,4]. In the education field, it helps to track and improve the learning efficiency of students [5,6]. EEG signals record the neural activities of human cerebral cortex and reflect emotion states, providing an objective and reliable way to perform emotion recognition [7,8,9]. Besides, the advantages of EEG such as noninvasive, fast, and inexpensive in data acquisition make it become a preferred media in emotion recognition [10]. A popular video evoked EEG-based emotion recognition system is shown in Figure 1, which generally consists of the following stages. First, emotional video clips should be collected and subjects should be recruited before the experiments, and then EEG signals could be recorded from subjects who generate corresponding emotion states during watching emotional clips. Second, the raw EEG signals will be preprocessed including removing noise and filtering. Third, related features will be extracted and fed into a classifier to perform emotion classification. In this paper, we mainly focus on the last stage.
In the past decade, many feature extraction methods and classifiers were proposed for EEG-based emotion recognition [10]. Basically, EEG features can be divided into two types, the time-domain features and the frequency-domain features. The time-domain features aim to extract the temporal information from EEG, e.g., the fractal dimension feature [11], the Hjorth feature [12] and the higher order crossing feature [13]. For the frequency-domain features, researchers usually first filter EEG signals into several frequency bands, and then extract EEG features from each frequency band. The frequency interval of interest is 1–50 Hz which is usually partitioned into five frequency bands, Delta (1–4 Hz), Theta (4–8 Hz), Alpha (8–14 Hz), Beta (14–31 Hz), and Gamma (31–50 Hz). The frequency-domain features mainly include the differential entropy (DE) feature [14], the power spectral density (PSD) feature [15], the differential asymmetry (DASM) feature [11], the rational asymmetry (RASM) feature [16] and so on. Lu et al. made a detailed comparison among these features, and found that DE was the most stable and accurate feature for emotion recognition than the others [14,17]. Therefore, the DE feature is adopted in this paper. On the classifiers, many machine learning methods were proposed for EEG-based emotion recognition [18,19]. Peng et al. designed a discriminative manifold extreme learning machine (DMELM) method for emotion recognition, and found that Beta and Gamma frequency bands were more relevant to emotional states transition than the others [20]. Li et al. proposed a hierarchical convolutional neural network (HCNN) to classify emotion states, and their experimental results also indicated that the high frequency bands (Beta and Gamma) performed better than the low frequency bands [21]. Moreover, Zheng et al. introduced deep belief networks (DBNs) to construct EEG-based emotion recognition models, and their results demonstrated that combining all frequency bands together performed better than individual bands [22]. Yang et al. drew similar conclusion by designing a Continuous Convolutional Neural Network [23]. The potential reason was that the multiple bands could provide complementary information to each other.
Although the methods mentioned above have achieved improvement in EEG-based emotion recognition, there still exists a problem. The way of frequency bands combination of them is directly concatenating all frequency bands together, which termed as the global scale in this paper and depicted in Figure 2a. However, such way cannot always achieve the best results since it essentially assumes that all frequency bands share the same importance. In this paper, we make extension on the way of combining frequency bands, which is termed as local scales and shown in Figure 2b. Here, by taking the face recognition task as an example, we illustrate the rationality of such multi-scale setting. Human faces manifest distinct characteristics and structures when observe in different scales, and different scales provide complementary information to each other [24]. Similarly, we assume that different scales of frequency bands hold different characteristics of emotion, as well as complement to each other. In each scale, we combine adjacent frequency bands into patches. For example, when the scale is 2, patches are formed by combining 2 adjacent frequency bands, and there are 4 patches in total. It should be noted that we only combine adjacent frequency bands into a patch because the frequency bands changing from Delta to Gamma reflects the conscious mind going from weak to active, which is a continuous process [10]. Therefore, it is reasonable to combine adjacent frequency bands.
For each scale, we train a base classifier to obtain the single-scale classification result. After that, the critical step is how to fuse the results of all scales to enhance the overall performance. It essentially belongs to an ensemble learning task [25,26,27], which combines the results of a set of base classifiers to perform better. In this paper, we design an adaptive weight learning method to combine all scales, which considers the classifier on each scale as a base classifier and learns the weight of each scale to fuse multi-scale results.
From the above, we propose a novel multi-scale frequency bands ensemble learning (MSFBEL) for EEG-based emotion recognition. Generally, the main contributions of this work are summarized as follows.
  • We extended the way of combining different frequency bands into four local scales and one global scale, and then performed emotion recognition on every scale with a single-scale classifier.
  • We proposed an effective adaptive weight learning method to ensemble multi-scale results, which can adaptively learn the respective weights of different scales according to the maximal margin criterion, whose objective can be formulated as a quadratic programming problem with the simplex constraint.
  • We conducted extensive experiments on benchmark emotional EEG data sets, and the results demonstrated that the global scale that directly concatenating all frequency bands cannot always guarantee to obtain the best emotion recognition performance. Different scales provide complementary information to each other, and the proposed method can effectively combine these information to further improve the performance.
The rest of this paper is organized as follows. Section 2 presents the proposed method in detail. Section 3 displays the emotional EEG data sets, experiments, and results of the proposed method. Section 4 concludes the whole paper and presents the future work.
In this paper, vectors are written as boldface lowercase letters, and matrices are written as boldface uppercase letters. The trace of matrix A is represented by Tr ( A ) . For a vector a , the  2 -norm of it is denoted by a 2 = ( a T a ) 1 2 , where a T is the transpose of a . a 0 represents that every element of vector a is larger than or equal to zero. 1 and I represent a column vector that all elements are “1” and an identity matrix, respectively. I denotes an indicator function which takes the value of 1 when the condition is true, and 0 otherwise.

2. Method

In this section, we present the proposed method MSFBEL (Figure 3) in detail, which mainly contains two stages. First, we re-arrange every EEG sample into different scales as shown in Figure 2, and then perform emotion classification on each scale by a single-scale classifier, called single-scale frequency band ensemble learning (SSFBEL). Second, the results of all scales are fused by the adaptive weight learning method to further improve the performance.

2.1. Single-Scale Frequency Band Ensemble Learning

In this paper, we use DE feature to model emotion information from EEG signals. Without loss of generality, supposing that DE features are extracted from s frequency bands, we divide it into s scales. In each scale j ( j = 1 , 2 , , s ), we combine j adjacent frequency bands into patches and then obtain p j = s j + 1 patches of this scale. Figure 2 displays an example when s = 5 , which represents that DE features are extracted from 5 frequency bands (Delta, Theta, Alpha, Beta, and Gamma).
SSFBEL is proposed to perform emotion classification on each scale of frequency bands, whose architecture is displayed in Figure 4, which takes an example with s = 5 and j = 2 . Below we give some explanations of it.
  • First, given an unlabeled DE feature-based sample y R d , we divide it into a set of patches y i R d j ( i = 1 , 2 , , p j ) . Here d is the feature dimension of EEG samples. For example, if we use the DE-based EEG feature representation, d is equal to the product of the numbers of channels and frequency bands. Similarly, d j denotes the feature dimension of DE patches under scale j, that is, d j equals the product of the numbers of channels and frequency bands in a patch.
  • Second, these patches are, respectively fed into base classifiers and then the corresponding predicted labels { z 1 , z 2 , , z p j } can be obtained.
  • Finally, the predicted labels of all patches are combined by simple majority voting [28] to generate the final label r j for the sample y under scale j.
In SSFBEL, the collaborative representation based classification (CRC) [29] is used as the base classifier. CRC usually represents a test sample with an over-complete dictionary formed by training samples, whose representation coefficient vector is regularized with an 2 -norm to improve its computational efficiency. Once the representation coefficient vector is obtained, the test sample can be categorized into the class which yields the minimum reconstruction error. In current study, for each patch, we construct the corresponding dictionary according to the principle that the combination and order of frequency bands of them are consistent.
Suppose that we have a patch y i and the corresponding dictionary formed by training samples X i = [ X i 1 , X i 2 , , X i c ] R d j × n , where c, d j , and n are the number of classes, the feature dimension of DE patches under scale j, and the number of training samples, respectively. X i k R d j × n k ( k = 1 , 2 , , c ) is the collection of samples from the k-th class in which each column is a sample, where n k denotes the number of samples in the k-th class. For the patch y i , its representation coefficient α i R n can be obtained by solving the following objective
min α i X i α i y i 2 2 + λ α i 2 2 ,
where λ is a regularization parameter. Obviously, the optimal representation coefficient to (1) is
α i * = ( X i T X i + λ I ) 1 X i T y i .
Let α i k * R n k represent the vector whose only nonzero entries are the entries of α i * associated with class k. The sample y i can be reconstructed by the training samples of class k as y i k = X i k α i k * . The label of y i is determined as the class which yields the minimum reconstruction error
z i = arg min k   X i k α i k * y i 2 .
In our experiments, to further improve the classification accuracy, we divided the above reconstruction error by α i k * 2 , because it can bring some discrimination information for classification [29]. Finally, the single-scale result r j is obtained, which combines all patches’ results z i ( i = 1 , 2 , , p j ) by using majority voting r j = arg max k [ c ] { i = 1 p j I ( z i = k ) } .

2.2. Adaptive Weight Learning

Assuming that different scales of frequency bands might have complementary information to each other, we combine the classification results of all scales obtained by SSFBEL to enhance the emotion recognition performance. The whole architecture of the proposed MSFBEL is shown in Figure 3.
Obviously, the current task is how to determine the optimal weights for different scales in the ensemble stage. In this work, we propose an adaptive weight learning method by maximizing the ensemble margin. Suppose the ensemble weight vector is w = [ w 1 ; w 2 ; ; w s ] R s , where j = 1 s w j = 1 , and we learn it from data. Specifically, we select m samples from the total n training samples, whose data and labels can be, respectively represented as X s u b = [ x 1 , x 2 , , x m ] R d × m and O s u b = [ o 1 , o 2 , , o m ] R c × m , where x p R d ( p = 1 , 2 , , m ) represents the p-th DE-based EEG sample, o p is the corresponding ground-truth label. We divide a sample x p into s scales and feed them into SSFBEL to obtain the classification result of each scale r p j ( j = 1 , 2 , , s ) . We define the decision matrix D = [ d p j ] R m × s as
d p j = + 1 , if o p = r p j 1 , if o p r p j .
Then, the ensemble margin of sample x p is defined as
ε ( x p ) = j = 1 s w j d p j .
To get the optimal weight vector w , we should make the ensemble margin in (5) as large as possible. Based on the studies of [30,31], margin maximization can be transformed into a loss minimization problem. To be specific, the ensemble loss function can be formulated as
l o s s = p = 1 m [ 1 ε ( x p ) ] = p = 1 m [ 1 j = 1 s w j d p j ] = 1 m Dw 2 2 ,
where 1 m = [ 1 ; 1 ; ; 1 ] R m is a column vector. Therefore, objective (6) is equivalent to optimize
min w 0 , w T 1 s = 1 1 m Dw 2 2 ,
where 1 s = [ 1 ; 1 ; ; 1 ] R s is a column vector. By denoting M = D T D and b = 2 D T 1 m , we can rewrite the objective (7) as
min w 0 , w T 1 s = 1 Tr ( ( 1 m Dw ) T ( 1 m Dw ) ) min w 0 , w T 1 s = 1 Tr ( 1 m T 1 m 2 w T D T 1 m + w T D T Dw ) min w 0 , w T 1 s = 1 Tr ( w T D T Dw 2 w T D T 1 m ) min w 0 , w T 1 s = 1 Tr ( w T M w w T b ) .
Since w is a column vector, objective (8) is equivalent to optimize
min w 0 , w T 1 s = 1 w T M w w T b .
To make the above objective separable, we introduce an auxiliary variable v with respect to w , then we get
min w 0 , w T 1 s = 1 v T M w v T b .
According to the augmented Lagrangian multiplier (ALM) method [32,33], objective (10) can be rewritten as
min w 0 , w T 1 s = 1 , v v T M w v T b + μ 2 w v + β μ 2 2 ,
where μ is a quadratic penalty parameter and β R s is the Lagrangian multiplier.
Accordingly, an alternative optimization method is applied to solving problem (11). The details are given in Appendix A. Finally, we get the optimum solution w * to problem (11), based on which we can make prediction on the test sample y as l = arg max k [ c ] { j = 1 s w j * I ( r j = k ) } . The procedure of MSFBEL framework is outlined in Algorithm 1.
Algorithm 1 The procedure for MSFBEL framework.
Input: Number of scales s, number of classes c, training data X R d × n , training data label
   O R c × n , a subset of training data X s u b R d × m , the labels of subset O s u b R c × m ,
  testing data y R d ;
Output: The label of testing data: l.
 1: for j = 1 : s do
 2:  Compute the label of testing data under scale j via Algorithm 2;
 3: end for
 4: Compute decision matrix D by Equation (4) with X s u b and O s u b ;
 5: Compute M = D T D and b = 2 D T 1 m ;
 6: Compute the adaptive weight w * via Algorithm 3;
 7: Compute l = arg max k [ c ] { j = 1 s w j * I ( r j = k ) } .

3. Experiments and Results

In this section, we first describe two emotional EEG data sets used in the experiments, including their data collection and feature extraction. Then, the experimental settings are given based on which we perform EEG-based emotion recognition to evaluate the effectiveness of MSFBEL.
Algorithm 2 The procedure for SSFBEL framework.
Input:s, c, λ , X i = [ X i 1 , X i 2 , , X i c ] R d j × n , y i R d j , where i = 1 , 2 , , p j and
    p j = s j + 1 ;
Output: The label of testing data under scale j: r j .
  1: Compute α i * by Equation (2);
  2: Compute z i by Equation (3);
  3: Compute r j = arg max k [ c ] { i = 1 p j I ( z i = k ) } .
Algorithm 3 The algorithm to solve problem (11).
Input: M and b ;
Output: The weight vector w R s .
  1: Initialize v , β , μ , and  ρ ;
  2: while not converged do
  3:  Update v by Equation (A1);
  4:  Update w by solving problem (A3) via Algorithm 4;
  5:  Update β = β + μ ( w v ) ;
  6:  Update μ = ρ μ ;
  7: end while
Algorithm 4 The algorithm to solve problem (A3).
Input: M , v , β , μ and s;
Output: The weight vector w R s .
  1: Compute g = v β + M T v μ R s ;
  2: Compute h = g 1 s T g s 1 s + 1 s 1 s ;
  3: Use Newton’s method to obtain the root η ¯ * of Equation (A12);
  4: The optimal solution w j * = ( h j η ¯ * ) + , where j = 1 , 2 , , s .

3.1. Data Set

3.1.1. SEED IV

The “SEED IV” is a publicly available emotional EEG data set [34]. The EEG signals were collected from 15 healthy subjects when they watched emotion-eliciting videos. In the EEG data collection experiment, 24 two-minute video clips were played to each subject. There are four types of emotional video clips, sadness, fear, happiness, and neutral, and each emotion has 6 video clips. While watching video clips, EEG signals of subjects were recorded at a 1000 Hz sampling rate by the 62-channel ESI NeuroScan system (https://compumedicsneuroscan.com/ (accessed on 6 August 2020)). Every subject was asked to complete the experiment three times on different days, and therefore we obtained three sessions of EEG signals for each subject.
In our experiments, we used the “EEG_feature_smooth” EEG data recordings downloaded from the “SEED IV” web site (http://bcmi.sjtu.edu.cn/home/seed/seed-iv.html (accessed on 20 December 2019)). Preprocessing and feature extraction of EEG data had already been conducted, including downsampling all data to 200 Hz, filtering out noise and artifacts by linear dynamic system (LDS) [35], and extracting DE features from 5 frequency bands: Delta, Theta, Alpha, Beta, and Gamma, with a four-second time window without overlapping. The dimensions of DE feature were 62 × W 1 × 5 (format: # channel × # samples × # frequency bands), where W 1 was the number of samples of one subject in each trial. We reshaped the 62 points of each of the 5 frequency bands, and then obtained DE features with the shape of 310 × W 1 . Since the time durations of different video clips in each session were slightly different, the total sample numbers for each subject in each session were approximately 830.

3.1.2. DEAP

Another emotional EEG data set used to validate our proposed method is “DEAP” [36]. It is a music video evoked EEG data set. There were 32 subjects invited to watch 40 one-minute music video clips. At the end of each video, subjects were asked to make a self-assessment of their level in terms of arousal, valence, liking, and dominance. During the experiment, the EEG signals were recorded by Biosemi ActiveTwo system (http://www.biosemi.com (accessed on 6 August 2020) with 32-channel electrode according to the international 10–20 system placement. The sampling rate is 512 Hz.
In our experiments, we utilized the “Data_preprocessed_matlab” EEG data recordings downloaded from the “DEAP” web site (http://www.eecs.qmul.ac.uk/mmv/datasets/deap/index.html (accessed on 20 December 2019)). EEG signals were down-sampled to 128 Hz. EOG artefacts were removed by using a blind source separation technique (http://www.cs.tut.fi/gomezher/projects/eeg/aar.htm (accessed on 6 August 2020)). A bandpass frequency filter from 4.0 to 45.0 Hz was applied. Then, we extracted DE features from 4 frequency bands: Theta, Alpha, Beta, and Gamma, with a one-second window size without overlapping. The dimensions of DE feature were 32 × W 2 × 4 , where W 2 was the number of samples in each trial. We concatenated the 32 values of the 4 frequency bands and then obtained DE features with the shape of 128 × W 2 . We got 63 samples for each trial in which the first 3 samples were baseline signals, and the last 60 sample were trial signals. According to the study of Yang et al. [37] that the baseline signals were useful for emotion recognition, we further processed the data as they did: calculating the deviation between every trial sample and the average of 3 baseline samples as the final input. Therefore, we obtained 60 samples in each trial, and there were totally 2400 samples for each subject.

3.2. Experimental Settings

We evaluated methods on every subject of the two data sets. For the “SEED IV” data set, we evaluated methods under the within-session experimental paradigm as [34]. Specifically, for every subject in each session, we utilized the last 8 trials as the test data, which not only contained all emotional states but also guaranteed that each emotional state has exact 2 trials, and the rest 16 trials as the training data. On the “DEAP” data set, we split every subject’s data as [36], which chose 5 as a threshold to divide all samples into four categories according to the different levels of valence and arousal: high valence and high arousal (HVHA), high valence and low arousal (HVLA), low valence and high arousal (LVHA), low valence and low arousal (LVLA). Then we performed 5-fold cross validation on each subject’s data. The recognition accuracy and standard deviation were used as evaluation metrics. The average accuracy and standard deviation of all subjects represent the final performance of a method.
We compare MSFBEL with support vector machine (SVM), K Nearest Neighbors (KNN) and SSFBEL. Linear kernel was used in SVM and the regularization parameter C was determined by grid search from { 10 5 , 10 4 , , 10 5 } . For KNN, Euclidean distance measure was used and the number of neighboring samples K was searched from { 10 , 20 , 30 , , 150 } . The regularization parameter λ in objective (2) of SSFBEL was fine-tuned with grid search from { 10 5 , 10 4 , , 10 1 } . The parameters μ in objective (11) of MSFBEL was fine-tuned with grid search from { 10 5 , 10 4 , , 10 1 } , β and ρ were initialized with [ 1 ; 1 ; ; 1 ] R s and 1.1 , respectively. To learn the optimal weights corresponding to different scales, we need to get the decision matrix D defined in Equation (4). Therefore, we divided the training data into two parts; one was used for training and the other was used for validation. For the “SEED IV” data set, each part contained 2 trials of each emotional state. For the “DEAP” data set, each part contained half samples of each emotional category. Then we calculated the decision matrix D based on the ground truth labels and the estimated labels of samples in the validation set.

3.3. Experimental Results and Analysis

3.3.1. The Effect of Different Scales

In this section, we compare the performance of global scale and local scales of frequency bands for emotion recognition by SSFBEL. First, we perform classification on different scales for every subject’ data from the “SEED IV” and “DEAP” data sets, and the results are shown in Table 1, Table 2, Table 3 and Table 4, respectively. The best results are highlighted in boldface. From these tables, we can observe that not every subject gets the best results on the global scale (scale = 5 for “SEED IV”, and scale = 4 for “DEAP”). For example, for subject #1 in Table 1, SSFBEL gets the best accuracy 84.85% when scale = 1, which exceeds the accuracy of scale = 5 by 1.17%. For subject #6, it obtains the optimal result 79.25% when scale = 2, which outperforms the result of scale = 5 by 12.58%. We can find similar results from the other three tables. Second, we calculate the average accuracy of all subjects in each scale of frequency bands. The average accuracies of each scale on the three sessions of “SEED IV” data set are shown in Figure 5. From this figure, we can find that the global scale cannot always achieve the optimal accuracy on every session. For example, in session 1, SSFBEL achieves the highest average accuracy 81.65% on scale = 1, which is higher than that of scale = 5 by 1.46%. These results are consistent with our proposed idea that the global scale of frequency bands which direct concatenating all frequency bands together cannot always achieve the best performance. That is, sometimes local scales can achieve higher emotion recognition accuracy than the global scale. Therefore, it is reasonable for us to re-organize all frequency bands into different scales, which can provide more potential for improving the performance of emotion recognition.
However, there still exists a problem that the optimal scale of frequency bands varies in terms of different subjects. For example, in Table 1, some subjects (#1, #3, #4, #7, #9, #10, #12, and #15) get the best result on scale = 1, and some subjects (#2, #6, #9, and #14) obtain the optimal result on scale = 2, and so on. Besides, some subjects (#2 and #9) achieve the best result on several scales at the same time. These problems can be found in the other three tables. The uncertainty of the optimal scale may be influenced by the characteristics of the EEG signals, which not only have low signal-to-noise ratio (SNR) but also exhibit significant differences across subjects [38]. Therefore, it is necessary to fuse all scales to reduce the impact of these factors. In this paper, MSFBEL is proposed to fuse different scales, in which the adaptive weight learning can fuse all scales’ results through automatically assigning larger weights to more important scales to further enhance the performance. The effectiveness of MSFBEL will be evaluated in Section 3.3.2.

3.3.2. The Performance of MSFBEL

In this part, we compare MSFBEL with SVM, KNN and SSFBEL to show the effectiveness of it. In this experiment, SVM and KNN take global scale frequency bands as input. As for SSFBEL, we choose the best frequency band scale for every data set as input. Specifically, according to the average accuracies shown in Table 1, Table 2, Table 3 and Table 4, we choose the results of 1, 3, and 5 for session 1, 2, and 3 of the “SEED IV” data set, and we select the results of scale = 4 for the “DEAP” data set. MSFBEL method takes all frequency band scales as input.
For the “SEED IV” data set, emotion recognition accuracies of the four methods in the three sessions are shown in Table 5. From the experimental results, we observe that MSFBEL achieves the best average recognition rates of 82.75%, 87.87%, and 78.27% in the three sessions, respectively. When compared with SVM, MSFBEL respectively achieves 11.53%, 10.27% and 6.51% improvements in the three sessions. As for KNN, MSFBEL respectively obtains 11.78%, 7.69% and 9.33% improvements than it in the three sessions. Moreover, MSFBEL exceeds SSFBEL by 1.10%, 2.12%, and 1.57% corresponding to the three sessions. For the “DEAP” data set, accuracies of the four methods are displayed in Table 6. From this table, we can find that MSFBEL achieves the average accuracy of 74.23%, which obtains 20.63% and 11.5% improvements in comparison with SVM and KNN. In addition, the performance of MSFBEL is better than SSFBEL by 2.03% in terms of average accuracy. Besides, Figure 6 presents the overall performance of the four methods. We observe that MSFBEL achieves better performance than the other three methods on both data sets. The underlying reason may be the combination of different scales of frequency bands by assigning larger weights to important scales by MSFBEL, while the other three methods only conduct classification on one scale. Therefore, we declare that there are complementary information among different scales, and the proposed adaptive weight learning method effectively fuses these information to enhance classification performance.
Besides the comparison of average accuracy of these four methods, we perform the Friedman test [39] to illustrate the statistical significance among them. The Friedman test is a non-parametric statistical test, which is used to detect differences of multiple methods across multiple test results. The null-hypothesis is “all the methods have the same performance”. If the null-hypothesis is rejected, the Nemenyi test is utilized to further distinguish whether the performances of the two among all methods are significantly different. We analyze the difference in performance among the four methods, and the results are shown in Figure 7. In the figure, the solid circle represents the average rank of each method, and the vertical line represents the critical distance (CD) of Nemenyi test, which is defined as follows
C D = q α k ( k + 1 ) 6 N .
where k denotes the number of methods, N denotes the number of result groups, and q α is the critical value which is defaulted as 0.05 [40]. In this paper, we set k = 4 because there are four methods in total. For “SEED IV” data set, N = 45 because there are 15 subjects and each subject has 3 sessions. For “DEAP” data set, N = 32 since there are 32 subjects. If two vertical lines do not have overlap, it means that the corresponding methods have statistically different performance. As shown in Figure 7, both on “SEED IV” and “DEAP” data set, SSFBEL and MSFBEL do not have overlap with SVM and KNN, which represents that our proposed methods are significantly different in performance with these two compared methods.
Figure 8 and Figure 9 present the confusion matrices for the emotion recognition results of SVM, KNN, SSFBEL, and MSFBEL on the two data sets, respectively. From them, we first can obtain the average recognition accuracies of every emotional state. For example, in Figure 8, the average accuracy of the “neutral” emotional state classified by SVM, KNN, SSFBEL, and MSFBEL are 80.48%, 79.31%, 82.36%, and 87.73%, respectively. In Figure 9, the average accuracies of the “HVHA” emotional state classified by these four methods are 72.45%, 76.7%, 85.74%, and 86.89%, respectively. Second, we can get the misclassification rate of each emotion state. For example, from the confusion matrix of MSFBEL on the “SEED IV” data set (Figure 8d), 87.73% of the EEG samples are correctly recognized as “neutral” state while 4.59%, 1.19%, 6.5% of them are incorrectly classified as “sad”, “fear”, and “happy” states, respectively. From the confusion matrix of MSFBEL on the “DEAP” data set (Figure 9d), 86.89% of the EEG samples are correctly classified as “HVHA” state while 3.13%, 6.26%, 3.36% of them are misclassified as “HVLA”, “LVHA”, and “LVLA” states, respectively. Third, comparing with the other three methods, MSFBEL shows improvement on each of the four emotional state. For instance, as shown in Figure 8, MSFBEL exceeds the accuracies of SVM, KNN, and SSFBEL on the “neutral” state recognition by 7.25%, 8.42%, and 5.37%, respectively. Further, we notice that “neutral” state always gets the highest accuracies on the four methods, which can deduce that it is the easiest emotional state to be identified. As displayed in Figure 9, MSFBEL exceeds the accuracies of other three methods on the “HVHA” state recognition by 14.44%, 10.19%, and 1.15%, respectively.

4. Conclusions and Future Work

In this paper, a new frequency bands ensemble method (MSFBEL) was proposed to recognize the emotional states from EEG data. The main advantages of MSFBEL are that (1) It re-organizes all frequency bands into different scales (several local scales and one global scale) to extract features and classify. (2) It combines the results of different scales by learning an adaptive weight to further improve emotion classification performance. Extensive experiments were conducted on the “SEED IV” and “DEAP” data set to evaluate the performance of MSFBEL. The results demonstrate that the scale of frequency bands influences the emotion recognition rate, while the global scale that directly concatenating all frequency bands cannot always guarantee to obtain the best emotion recognition performance. Moreover, the results also illustrate that different scales of frequency bands provide complementary information to each other, and the proposed adaptive weight learning method can effectively fuse these information. The results indicate the effectiveness of our proposed MSFBEL model in emotion recognition task.
In the future, we will focus on the following problem not covered in this paper. We will explore the cross subject or cross session domain adaptation problem, so that we may try to find the optimal scale of frequency bands for EEG-based emotion recognition.

Author Contributions

Conceptualization, Y.P., W.K. and G.D.; Data curation, F.S.; Methodology, F.S., Y.P.; Software, F.S., Y.P.; Validation, W.K., G.D.; Writing—original draft preparation, F.S., Y.P.; Writing—review and editing, W.K. and G.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (2017YFE0118200, 2017YFE0116800), National Natural Science Foundation of China (61971173, 61671193, U1909202), Science and Technology Program of Zhejiang Province (2018C04012), Zhejiang Provincial Natural Science Foundation of China (LY21F030005), Fundamental Research Funds for the Provincial Universities of Zhejiang (GK209907299001-008), Key Laboratory of Advanced Perception and Intelligent Control of High-end Equipment of Ministry of Education, Anhui Polytechnic University (GDSC202015). The authors also would like to thank the anonymous reviewers for their comments on this work.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Below we give the solution of the problem (11). We alternately optimize w and v , which updates one variable with the other fixed.
(1) Update v with w fixed. By taking the derivative of objective (11) with respect to v and setting it to zero, we get
v = w + b + β M w μ .
(2) Update w with v fixed. When v is fixed, objective (11) becomes
min w 0 , w T 1 s = 1 v T M w + μ 2 w v + β μ 2 2 .
By completing the square form with respect to w , we rewrite the above objective as
min w 0 , w T 1 s = 1 μ 2 w v + β + M T v μ 2 2 .
Problem (A3) is a typical Euclidean projection on the simplex [33,41], and we solve it as below. By denoting g = v β + M T v μ , the Lagrangian function of problem (A3) is
L ( w , γ , η ) = 1 2 w g 2 2 γ ( w T 1 s 1 ) η T w ,
where γ and η R s are the Lagrangian multipliers. Suppose that w * is the optimal solution, γ * and η * are the corresponding Lagrangian multipliers. According to KKT condition [42], we get the equations for every j ( j = 1 , 2 , , s ) as follows
w j * g j γ * η j * = 0 w j * 0 η j * 0 w j * η j * = 0 .
The first equation of Equation (A5) can be reformulated into the following form
w * g γ * 1 s η * = 0 .
With the constraint w T 1 s = 1 , the above equation becomes
γ * = 1 1 s T g 1 s T η * s .
By substituting Equation (A7) into Equation (A6), we obtain
w * = g 1 s T g s 1 s + 1 s 1 s 1 s T η * s 1 s + η * .
Suppose η ¯ * = 1 s T η * s and h = g 1 s T g s 1 s + 1 s 1 s , Equation (A8) becomes
w * = h + η * η ¯ * 1 s .
Hence, for every j, we obtain
w j * = h j + η j * η ¯ * .
According to Equations (A5) and (A10), we obtain h j + η j * η ¯ * = ( h j η ¯ * ) + , where ( f ( · ) ) + = m a x ( f ( · ) , 0 ) . Hence, we have
w j * = ( h j η ¯ * ) + .
Once the optimal η ¯ * is determined, the optimal solution w * will be obtained from Equation (A11). Similarly, Equation (A10) can also be reformulated into η j * = w j * + η ¯ * h j such that η j * = ( η ¯ * h j ) + . Thus, η ¯ * can be computed by
η ¯ * = 1 s t = 1 s ( η ¯ * h j ) + .
Defining a function as
f ( η ¯ ) = 1 s t = 1 s ( η ¯ h j ) + η ¯ .
When Equation (A13) equals to zero, the optimal η ¯ * can be obtained by Newton method [43] to find the root of Equation (A13) as following
η ¯ p + 1 = η ¯ p f ( η ¯ p ) f ( η ¯ p ) .

References

  1. Cowie, R.; Douglas-Cowie, E.; Tsapatsoulis, N.; Votsis, G.; Kollias, S.; Fellenz, W.; Taylor, J.G. Emotion recognition in human-computer interaction. IEEE Signal Process. Mag. 2001, 18, 32–80. [Google Scholar] [CrossRef]
  2. Swangnetr, M.; Kaber, D.B. Emotional state classification in patient–robot interaction using wavelet analysis and statistics-based feature selection. IEEE Trans. Hum. Mach. Syst. 2012, 43, 63–75. [Google Scholar] [CrossRef]
  3. Qureshi, S.A.; Dias, G.; Hasanuzzaman, M.; Saha, S. Improving depression level estimation by concurrently learning emotion intensity. IEEE Comput. Intell. Mag. 2020, 15, 47–59. [Google Scholar] [CrossRef]
  4. Hu, B.; Rao, J.; Li, X.; Cao, T.; Li, J.; Majoe, D.; Gutknecht, J. Emotion regulating attentional control abnormalities in major depressive disorder: An event-related potential study. Sci. Rep. 2017, 7, 1–21. [Google Scholar] [CrossRef] [Green Version]
  5. Yang, D.; Alsadoon, A.; Prasad, P.C.; Singh, A.K.; Elchouemi, A. An emotion recognition model based on facial recognition in virtual learning environment. Procedia Comput. Sci. 2018, 125, 2–10. [Google Scholar] [CrossRef]
  6. Li, T.M.; Shen, W.X.; Chao, H.C.; Zeadally, S. Analysis of Students’ Learning Emotions Using EEG. In International Conference on Innovative Technologies and Learning; Springer: Cham, Switzerland, 2019; pp. 498–504. [Google Scholar]
  7. Zhang, J.; Zhao, S.; Yang, G.; Tang, J.; Zhang, T.; Peng, Y.; Kong, W. Emotional-state brain network analysis revealed by minimum spanning tree using EEG signals. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 1045–1048. [Google Scholar]
  8. Peng, Y.; Li, Q.; Kong, W.; Qin, F.; Zhang, J.; Cichocki, A. A joint optimization framework to semi-supervised RVFL and ELM networks for efficient data classification. Appl. Soft Comput. 2020, 97, 106756. [Google Scholar] [CrossRef]
  9. Wu, S.; Xu, X.; Shu, L.; Hu, B. Estimation of valence of emotion using two frontal EEG channels. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017; pp. 1127–1130. [Google Scholar]
  10. Alarcao, S.M.; Fonseca, M.J. Emotions recognition using EEG signals: A survey. IEEE Trans. Affect. Comput. 2017, 10, 374–393. [Google Scholar] [CrossRef]
  11. Liu, Y.; Sourina, O. Real-time fractal-based valence level recognition from EEG. In Transactions on Computational Science XVIII; Springer: Berlin/Heidelberg, Germany, 2013; pp. 101–120. [Google Scholar]
  12. Hjorth, B. EEG analysis based on time domain properties. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 306–310. [Google Scholar] [CrossRef]
  13. Petrantonakis, P.C.; Hadjileontiadis, L.J. Emotion recognition from EEG using higher order crossings. IEEE Trans. Inf. Technol. Biomed. 2009, 14, 186–197. [Google Scholar] [CrossRef]
  14. Shi, L.C.; Jiao, Y.Y.; Lu, B.L. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 6627–6630. [Google Scholar]
  15. Frantzidis, C.A.; Bratsas, C.; Papadelis, C.L.; Konstantinidis, E.; Pappas, C.; Bamidis, P.D. Toward emotion aware computing: An integrated approach using multichannel neurophysiological recordings and affective visual stimuli. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 589–597. [Google Scholar] [CrossRef]
  16. Lin, Y.P.; Wang, C.H.; Jung, T.P.; Wu, T.L.; Jeng, S.K.; Duann, J.R.; Chen, J.H. EEG-based emotion recognition in music listening. IEEE Trans. Biomed. Eng. 2010, 57, 1798–1806. [Google Scholar]
  17. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar]
  18. Paszkiel, S. Using neural networks for classification of the changes in the EEG signal based on facial expressions. In Analysis and Classification of EEG Signals for Brain–Computer Interfaces; Springer: Cham, Switzerland, 2020; pp. 41–69. [Google Scholar]
  19. Chen, X.; Xu, L.; Cao, M.; Zhang, T.; Shang, Z.; Zhang, L. Design and Implementation of Human-Computer Interaction Systems Based on Transfer Support Vector Machine and EEG Signal for Depression Patients’ Emotion Recognition. J. Med. Imaging Health Inform. 2021, 11, 948–954. [Google Scholar] [CrossRef]
  20. Peng, Y.; Lu, B.L. Discriminative manifold extreme learning machine and applications to image and EEG signal classification. Neurocomputing 2016, 174, 265–277. [Google Scholar] [CrossRef]
  21. Li, J.; Zhang, Z.; He, H. Hierarchical convolutional neural networks for EEG-based emotion recognition. Cogn. Comput. 2018, 10, 368–380. [Google Scholar] [CrossRef]
  22. Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  23. Yang, Y.; Wu, Q.; Fu, Y.; Chen, X. Continuous convolutional neural network with 3d input for eeg-based emotion recognition. In International Conference on Neural Information Processing; Springer: Cham, Switzerland, 2018; pp. 433–443. [Google Scholar]
  24. Lin, D.; Tang, X. Recognize high resolution faces: From macrocosm to microcosm. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1355–1362. [Google Scholar]
  25. Krawczyk, B.; Minku, L.L.; Gama, J.; Stefanowski, J.; Woźniak, M. Ensemble learning for data stream analysis: A survey. Inf. Fusion 2017, 37, 132–156. [Google Scholar] [CrossRef] [Green Version]
  26. Dietterich, T.G. Ensemble methods in machine learning. In International Workshop on Multiple Classifier Systems; Springer: Berlin/Heidelberg, Germany, 2000; pp. 1–15. [Google Scholar]
  27. Yang, P.; Hwa Yang, Y.; B Zhou, B.; Y Zomaya, A. A review of ensemble methods in bioinformatics. Curr. Bioinform. 2010, 5, 296–308. [Google Scholar] [CrossRef] [Green Version]
  28. Kumar, R.; Banerjee, A.; Vemuri, B.C. Volterrafaces: Discriminant analysis using volterra kernels. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 150–155. [Google Scholar]
  29. Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 471–478. [Google Scholar]
  30. Rosset, S.; Zhu, J.; Hastie, T. Boosting as a regularized path to a maximum margin classifier. J. Mach. Learn. Res. 2004, 5, 941–973. [Google Scholar]
  31. Shen, C.; Li, H. Boosting through optimization of margin distributions. IEEE Trans. Neural Netw. 2010, 21, 659–666. [Google Scholar] [CrossRef] [PubMed]
  32. Bonnans, J.F.; Gilbert, J.C.; Lemaréchal, C.; Sagastizábal, C.A. Numerical Optimization: Theoretical and Practical Aspects; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  33. Nie, F.; Yang, S.; Zhang, R.; Li, X. A general framework for auto-weighted feature selection via global redundancy minimization. IEEE Trans. Image Process. 2018, 28, 2428–2438. [Google Scholar] [CrossRef] [PubMed]
  34. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. Emotionmeter: A multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 2018, 49, 1110–1122. [Google Scholar] [CrossRef]
  35. Shi, L.C.; Lu, B.L. Off-line and on-line vigilance estimation based on linear dynamical system and manifold learning. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, Buenos Aires, Argentina, 31 August–4 September 2010; pp. 6587–6590. [Google Scholar]
  36. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. Deap: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  37. Yang, Y.; Wu, Q.; Qiu, M.; Wang, Y.; Chen, X. Emotion recognition from multi-channel EEG through parallel convolutional recurrent neural network. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  38. Li, J.; Qiu, S.; Shen, Y.Y.; Liu, C.L.; He, H. Multisource transfer learning for cross-subject EEG emotion recognition. IEEE Trans. Cybernet. 2019, 50, 3281–3293. [Google Scholar] [CrossRef] [PubMed]
  39. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  40. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  41. Kyrillidis, A.; Becker, S.; Cevher, V.; Koch, C. Sparse projections onto the simplex. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 235–243. [Google Scholar]
  42. Boyd, S.; Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  43. Sherman, A.H. On Newton-iterative methods for the solution of systems of nonlinear equations. SIAM J. Numer. Anal. 1978, 15, 755–771. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The flow chart of EEG-based emotion recognition system.
Figure 1. The flow chart of EEG-based emotion recognition system.
Sensors 21 01262 g001
Figure 2. Different scales of frequency bands: (a) global scale, (b) local scales.
Figure 2. Different scales of frequency bands: (a) global scale, (b) local scales.
Sensors 21 01262 g002
Figure 3. The framework of MSFBEL.
Figure 3. The framework of MSFBEL.
Sensors 21 01262 g003
Figure 4. The architecture of SSFBEL.
Figure 4. The architecture of SSFBEL.
Sensors 21 01262 g004
Figure 5. The average accuracies of different scales of frequency bands on the three sessions of “SEED IV” by using SSFBEL.
Figure 5. The average accuracies of different scales of frequency bands on the three sessions of “SEED IV” by using SSFBEL.
Sensors 21 01262 g005
Figure 6. The average accuracies of different methods on the two data sets for EEG-based emotion recognition.
Figure 6. The average accuracies of different methods on the two data sets for EEG-based emotion recognition.
Sensors 21 01262 g006
Figure 7. Friedman test of different methods on the two data sets.
Figure 7. Friedman test of different methods on the two data sets.
Sensors 21 01262 g007
Figure 8. Confusion matrices of different methods on the “SEED IV” data set.
Figure 8. Confusion matrices of different methods on the “SEED IV” data set.
Sensors 21 01262 g008
Figure 9. Confusion matrices of different methods on the “DEAP” data set.
Figure 9. Confusion matrices of different methods on the “DEAP” data set.
Sensors 21 01262 g009
Table 1. The accuracies (%) of different scales of frequency bands on session 1 of “SEED IV” by using SSFBEL.
Table 1. The accuracies (%) of different scales of frequency bands on session 1 of “SEED IV” by using SSFBEL.
SubjectLocal ScaleGlobal Scale
Scale = 1Scale = 2Scale = 3Scale = 4Scale = 5
184.8573.4380.1976.6983.68
296.0410010090.68100
382.7572.2681.8267.6082.05
410092.0789.7489.7489.74
568.3073.6669.0075.2979.72
665.7379.2578.5572.7366.67
797.2096.0469.9376.9279.72
865.0379.4983.2270.6377.86
993.4793.4793.4793.4793.47
1083.4574.1363.6465.2766.20
1180.1969.0081.8275.2976.46
1279.9577.6277.6274.1373.66
1366.6775.7680.6576.9276.69
1472.4984.1577.3970.8670.86
1588.5882.2881.5985.5583.92
Average81.6581.5180.5777.4580.05
Table 2. The accuracies (%) of different scales of frequency bands on session 2 of “SEED IV” by using SSFBEL.
Table 2. The accuracies (%) of different scales of frequency bands on session 2 of “SEED IV” by using SSFBEL.
SubjectLocal ScaleGlobal Scale
Scale = 1Scale = 2Scale = 3Scale = 4Scale = 5
164.0667.9776.2872.1376.28
297.5691.6997.5696.3388.51
391.2085.8288.7591.2091.20
482.8988.7584.1188.7588.75
580.9382.4088.5188.5188.51
688.7592.1894.1385.0989.49
797.3197.5692.9197.5697.56
871.1578.2487.0478.2480.93
988.7575.7973.1173.3573.11
1088.7594.1392.9194.1394.13
1169.6882.6480.6878.0080.68
1277.2679.9585.5771.3979.95
1378.9773.1172.3779.2275.79
1474.8273.1180.6886.5582.40
1510094.1391.6991.6994.13
Average83.4783.8385.7584.8185.43
Table 3. The accuracies (%) of different scales of frequency bands on session 3 of “SEED IV” by using SSFBEL.
Table 3. The accuracies (%) of different scales of frequency bands on session 3 of “SEED IV” by using SSFBEL.
SubjectLocal ScaleGlobal Scale
Scale = 1Scale = 2Scale = 3Scale = 4Scale = 5
160.3278.8280.1680.4369.97
289.0189.0189.0189.0189.01
356.5760.5961.1356.5778.55
489.0180.1689.0191.9689.01
564.3472.1272.1285.7972.12
670.2480.9788.4789.0189.01
789.0194.9184.4570.2484.99
863.0071.5878.5575.8786.33
954.9662.7362.7362.7366.49
1064.6164.6169.1764.6174.80
1167.2956.5761.1361.1361.13
1254.9649.0649.0653.6253.62
1361.1364.3461.1361.1361.13
1479.3670.2495.4489.0195.44
1578.5578.5578.5578.5578.82
Average69.4971.6274.6773.9876.79
Table 4. The accuracies of different scales of frequency bands on the “DEAP” by using SSFBEL (%).
Table 4. The accuracies of different scales of frequency bands on the “DEAP” by using SSFBEL (%).
SubjectLocal ScaleGlobal Scale
Scale = 1Scale = 2Scale = 3Scale = 4
160.2177.9280.4286.88
241.0450.4251.6757.92
353.1360.6362.5073.75
455.4260.2167.2963.54
548.5458.3358.5468.33
658.5467.5072.7177.71
745.6360.0060.8371.88
845.6365.6363.5475.00
947.9260.8363.7567.08
1065.4278.0472.9275.92
1142.9256.8858.3364.38
1251.2564.7962.7164.21
1373.1374.1780.1777.58
1455.0063.9676.0472.50
1571.0479.7980.4284.58
1667.0882.2988.7587.71
1748.3365.6366.6777.08
1849.5869.5870.6381.67
1947.9257.0857.5064.17
2058.3369.7973.5477.50
2148.5460.2161.2567.71
2242.7154.0450.2151.75
2364.3878.7583.7587.29
2454.1759.7963.5461.67
2558.9666.6768.7570.83
2651.8864.7966.2570.21
2762.5063.5464.1767.50
2849.7973.3372.9280.21
2947.2962.2964.1772.71
3044.1762.5064.3872.50
3146.6762.9266.8873.75
3249.3856.6768.5465.00
Average53.3365.2867.6272.20
Table 5. The comparison of different methods on the “SEED IV” (%).
Table 5. The comparison of different methods on the “SEED IV” (%).
SubjectSession 1Session 2Session 3
SVMKNNSSFBELMSFBELSVMKNNSSFBELMSFBELSVMKNNSSFBELMSFBEL
173.4365.9784.8580.1976.2875.0676.2876.2863.2757.9169.9780.16
286.0110096.0410097.5694.6297.5697.5681.7770.2489.0189.01
375.9975.7682.7573.6675.0681.9188.7591.2056.5749.0678.5578.55
483.2279.2510010071.1584.1184.1188.7594.9189.0189.0191.42
552.6859.2168.3075.2982.6477.2688.5185.5778.0276.1472.1275.87
659.4454.5565.7366.6764.5580.4494.1393.4069.7189.0189.0189.01
752.4571.3397.2097.2091.6997.5692.9199.2785.2580.1684.9994.91
877.1677.1665.0379.4966.9988.0287.0487.0489.5481.2386.3386.33
983.2287.4193.4710073.1167.2473.1174.8262.7362.7366.4966.49
1045.2258.2883.4573.6685.3391.2092.9191.6968.3675.0774.8064.88
1170.1672.4980.1979.7280.9366.9980.6880.6861.1349.0661.1367.29
1275.7671.3379.9577.6257.2161.3785.5783.8649.0659.5253.6254.96
1383.6860.1466.6777.6278.4862.5972.3787.2961.1361.1361.1361.13
1464.3466.9072.4972.4977.7574.3380.6886.5586.6054.9695.4495.44
1585.5564.8088.5887.6585.3310091.6994.1368.3678.8278.8278.55
Average71.2270.9781.6582.7577.6080.1885.7587.8771.7668.9476.6978.27
Std12.9411.5811.5710.9210.0712.057.526.8313.0313.0812.0212.45
Table 6. The comparison of different methods on the “SEED IV” (%).
Table 6. The comparison of different methods on the “SEED IV” (%).
SubjectSVMKNNSSFBELMSFBELSubjectSVMKNNSSFBELMSFBEL
159.7975.8386.8887.711756.0459.7977.0875.63
243.7557.5057.9259.791855.4272.2981.6782.92
352.7166.0473.7575.631961.2564.5864.1765.63
454.1762.5063.5470.212061.0462.9277.5079.38
542.7155.6368.3368.332154.3862.5067.7169.58
649.5861.8877.7180.832236.8850.2151.7563.96
763.1371.8871.8873.542362.9276.2587.2985.00
851.2560.2175.0076.042456.6760.4261.6764.58
952.5056.8867.0867.292552.2963.7570.8374.38
1055.8360.2175.9277.922652.5055.4270.2169.79
1134.3851.6764.3870.212762.0868.7567.5068.75
1256.4661.2564.2166.042848.3361.0480.2182.92
1368.1369.3877.5881.882953.3360.4272.7173.54
1454.7959.3872.5073.543046.4661.6772.5073.96
1559.7970.4284.5884.173152.7164.1773.7573.75
1657.0868.3387.7188.963246.8854.1765.0069.38
MethodSVMKNNSSFBELMSFBEL
Average53.6062.7372.2074.23
Std7.326.348.427.23
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shen, F.; Peng, Y.; Kong, W.; Dai, G. Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition. Sensors 2021, 21, 1262. https://doi.org/10.3390/s21041262

AMA Style

Shen F, Peng Y, Kong W, Dai G. Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition. Sensors. 2021; 21(4):1262. https://doi.org/10.3390/s21041262

Chicago/Turabian Style

Shen, Fangyao, Yong Peng, Wanzeng Kong, and Guojun Dai. 2021. "Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition" Sensors 21, no. 4: 1262. https://doi.org/10.3390/s21041262

APA Style

Shen, F., Peng, Y., Kong, W., & Dai, G. (2021). Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition. Sensors, 21(4), 1262. https://doi.org/10.3390/s21041262

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop