Next Article in Journal
Siam Deep Feature KCF Method and Experimental Study for Pedestrian Tracking
Previous Article in Journal
Improved Face Detection Method via Learning Small Faces on Hard Images Based on a Deep Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status

1
Department of Information systems, Faculty of Computing and Information Technology (FCIT), King Abdulaziz University, Jeddah 22254, Saudi Arabia
2
College of Dentistry, Chosun University, Gwangju 61452, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(1), 498; https://doi.org/10.3390/s23010498
Submission received: 27 November 2022 / Revised: 23 December 2022 / Accepted: 27 December 2022 / Published: 2 January 2023
(This article belongs to the Section Biomedical Sensors)

Abstract

:
The accurate identification of the human emotional status is crucial for an efficient human–robot interaction (HRI). As such, we have witnessed extensive research efforts made in developing robust and accurate brain–computer interfacing models based on diverse biosignals. In particular, previous research has shown that an Electroencephalogram (EEG) can provide deep insight into the state of emotion. Recently, various handcrafted and deep neural network (DNN) models were proposed by researchers for extracting emotion-relevant features, which offer limited robustness to noise that leads to reduced precision and increased computational complexity. The DNN models developed to date were shown to be efficient in extracting robust features relevant to emotion classification; however, their massive feature dimensionality problem leads to a high computational load. In this paper, we propose a bag-of-hybrid-deep-features (BoHDF) extraction model for classifying EEG signals into their respective emotion class. The invariance and robustness of the BoHDF is further enhanced by transforming EEG signals into 2D spectrograms before the feature extraction stage. Such a time-frequency representation fits well with the time-varying behavior of EEG patterns. Here, we propose to combine the deep features from the GoogLeNet fully connected layer (one of the simplest DNN models) together with the OMTLBP_SMC texture-based features, which we recently developed, followed by a K-nearest neighbor (KNN) clustering algorithm. The proposed model, when evaluated on the DEAP and SEED databases, achieves a 93.83 and 96.95% recognition accuracy, respectively. The experimental results using the proposed BoHDF-based algorithm show an improved performance in comparison to previously reported works with similar setups.

1. Introduction

In the last decade, research efforts, from both academia and industry, have shifted toward the development of human interactive social robots that can build socio-emotional relationships with humans. The human emotional status has a strong impact on the daily life decision-making process [1]. Social robots require a precise understanding of human emotion for effective assistance during daily routine tasks [2]. Social robots require a highly robust brain–computer interfacing (BCI) framework for the precise identification of the human emotional status [3]. To bring noise robustness and invariance in the identification of human emotional states, both physiological and non-physiological signals are employed either simultaneously or separately.
The non-physiological signals include facial expression [4], vocal pattern, and text data [5,6,7]. Emotional status identification via facial expressions requires uninterrupted facial cues within each frame of the video. The overall framework includes face detection, localization, feature extraction, and facial landmark tracking using a certain machine learning mechanism with adequate training data and a stable evaluation strategy [8]. Face expression-based emotion identification is a contactless method and can track multiple subjects simultaneously; however, the photometric variations in the personal appearance of the subject degrade the accuracy of the emotion identification model. The vocal pattern also varies with the emotional conditions of the well-being. Vocal pattern-dependent frameworks require stable and distinctive speech temporal signal representation and classification models. Vocal pattern-based methods are very prone to environmental noise. Text mining and natural language processing applications require identifying the subject’s emotions from the text presented [9]. The frequency of words is employed in the feature extraction mechanism for text-based emotion identification. Non-physiological signal-based emotion recognition techniques are less complex; however, they can easily be deceived if the subject hides their true facial expressions, vocal tone, and textual words selection. Moreover, the non-physiological signal-dependent emotion recognition techniques fail when the subject is suffering from physical impairments [10].
The physiological signals include skin impedance [11], respiration [12], and blood volume pulsation [13], among others. It is hard for the subject to conceal their emotions in physiological signals. In comparison to other biosignals, the EEG patterns collected from various scalp locations on the skull can describe the emotional pattern more effectively in comparison to other biosignals, such as an ECG, and EMG signals [14]. In the recent decade, various frameworks were developed for emotion identification based on EEG biosignals. However, the development of a robust and invariant features extraction system continues to be a challenging task. Channel selection and signal pre-processing always play a crucial role in the robustness of emotion identification systems. The unwanted noise such as moment artifacts, respiration, cardiac beats, and blood flow need to be removed to improve the classification accuracy of the human emotion identification frameworks [15,16]. A human–robot interaction requires a hardware device to sense the EEG signals, which spatially covers the scalp and senses the electrical activity of the neurons in the brain cortex. The multi-channel EEG devices consisting of numerous mini electrodes record the voltage amplitude values at a specified sampling frequency. The data collected by a multi-channel EEG device are pre-processed before the feature extraction and classification stage. The feature extraction model for EEG signals can either be based on handcrafted features or deep features [17]. The handcrafted features extraction stage is independent of the training procedure of the classifier. On the other hand, the deep neural network model requires a large amount of training data to adjust the weights of the neurons through an iterative optimization algorithm and combines both the feature extraction and the classification stages together. In many cases, existing neural network models are employed for the extraction of features through transfer learning to reduce the complexity and training stage with large training datasets.
In most existing applications, the pre-existing neural network models are employed for the extraction of features through transfer learning methods. The features are not only extracted directly from the raw one-dimensional EEG signals but also from the two-dimensional spectrogram of the EEG data. The pre-existing DNN models for feature extraction include ResNet, GoogLeNet, VGG-19, and InceptionV3, which can be combined with classification models, such as the multi-class SVM, KNN, and decision trees. Such hybrid models can work effectively on time patterns as well as spectrograms of EEG data. The emotion identification framework precision is hugely dependent upon the pre-processing methods, the robust feature extraction models, and the final classification models. However, the non-linear nature of the EEG results in large numbers of false positives and false negative cases which degrade the performance of the existing emotion identification frameworks. The PhysioNet repository publicly provides complex physiologic EEG signals for a happy, sad, and normal emotional status for subjects with varying gender and age. In this work, we approach the problem of emotion recognition from EEG data using a new hybrid framework. This contribution involves the fusion of both handcrafted and robust deep features using a novel bag-of-features (BoF) model for solving the emotion recognition problem from EEG biosignals. More specifically, the main contributions of this paper are as follows:
  • A new hybrid feature extraction framework is developed for emotional status identification using a robust fusion approach of handcrafted features together with DNN-based features.
  • The Short-Time Fourier Transform is used to map the 1D EEG patterns into 2D spectrograms before the final feature extraction stage.
  • A bag-of-features framework is used to reduce the dimension of the hybrid raw feature vector.
  • A KNN-based clustering approach is introduced to combine the feature vectors from multiple EEG channels.
The remainder of this paper is arranged as follows: Section 2 describes the existing literature related to the problems addressed in this research. Section 3 discusses the proposed methodology, Section 4 presents the different experimental results, and finally, Section 5 provides a conclusion and some future research directions.

2. Related Work

The traditional EEG-based emotion identification schemes were developed along three main directions: pre-processing, feature extraction, and classification. Moreover, feature selection and reduction are some of the post-processing methods used to lower the computational complexity and storage requirements of diverse systems. In the pre-processing stage, the EEG signals are conditioned using various filtering approaches to remove the unwanted noise artifacts. To induce the emotional status of subjects, various data types are used, such as images [18], textual data [19], music clips [20], and videos [21], with different degrees of success. In what follows, we briefly describe the main stages involved in the development of emotion identification frameworks from EEG data. We then present our proposed framework.

2.1. Data Pre-Processing

The pre-processing stage includes effective channel selection, the filtering of the raw EEGs, and 2D spectral representation. The channel selection criterion aims at reducing raw data by excluding the EEG channels with fewer interclass variations [22]. Researchers adopted various methods to pick the most effective EEGs out of 64 channels on the basis of statistical information. In [23], the statistical dependencies of the data collected by an electrode positioned at a different area of the scalp were used for the selection of channels. Figure 1 presents the configuration and spatial distribution of 64-channel EEG electrodes on the human scalp for recording the electrical activity of brain neurons. Optimal channels subset identification methods were broadly classified into two classes based on their dependency on the classification model. The classifier-independent method [24], relying on a threshold value of 1.725 on the differential entropy, was applied for the selection of effective channels in the SEED [25] and DEAP [26] datasets. A total of 12 channels in SEED and 32 channels in the DEAP database fulfilled the differential thresholding criterion. In [27], 10 optimal EEG channels out of 64 were identified by analyzing the statistical parameters, such as the Shannon entropy, differential entropy, and first difference method. In [28], the normalized mutual information was estimated based on inner channel connection matrix thresholding to identify an optimal EEG channel subset. In [29], mutual information maximization based on joint entropy was used to detect an optimal subset of the EEG channels. Both entropy and mutual information-based methods do not rely on the classification model; however, it degrades the accuracy of the overall model due to the absence of correlation details among the EEG channels [30]. The classifier-dependent methods developed in [31,32] select those EEG channels for which the classification performance is high. Classifier-dependent methods do not affect the accuracy of the overall model; however, they are sensitive to feature extraction methods [32].
In general, the data acquisition devices record diverse noise artifacts due to joint movements, blood flow, and inconsistent environmental noise. Both personal and environmental artifacts removal require carefully designed filters at the pre-processing stage of the proposed pattern identification framework. In [33], raw EEG signals were passed through a low-pass filter to exclude the unwanted artifacts. The EEG signals were de-noised and the continuous wavelet transform (CWT) was applied to further reduce the noise interference. The discrete wavelet transform was employed in [34] to identify and filter out unwanted frequency components from the raw EEGs. Learnable temporal filters were introduced in [35] that learn and optimize the filter weights and parameters for EEG-based brain activity pattern identification. Instead of raw EEGs, the spectrogram of an EEG using Short-Time Fourier Transform (STFT) [36] was employed as a pre-processing method to enhance the classification accuracy of the model. The hybrid deep lightweight features were extracted from both a 1D raw EEG and their 2D spectrogram using a 1D-LBP and pre-trained DNN, respectively [37]. The various pre-processing methods are summarized in Table 1.

2.2. Feature Extraction

The pre-processing stage is followed by the feature extraction stage. Both the 1D and 2D signal-based feature extraction models developed so far can broadly be classified into two classes, namely handcrafted and deep learning-based methods. The handcrafted features based on statistical parameters are manually engineered by data scientists, while the deep learning-based methods are automatically learned from the labeled data provided for neural network weights optimization. In [38], the power spectrum of a frequency-based segmented EEG denotes the elements of the feature vector. The differential asymmetry (DASM) [39] and asymmetry rational (RASM) [40] were computed from the difference in absolute power and the ratio of absolute power between symmetrical channel pairs. The power-based method is less complex and low dimensional; however, it provides a lower accuracy for emotion identification. Entropy-based features descriptors, including simple entropy [41], permutation entropy (PE) [42], spectral entropy (SE) [43], singular value decomposition entropy (SVDE) [44], and approximate entropy (ApEn) and sample entropy (SampEn) [45], were all developed to represent the non-linear patterns of the EEG signals. Time domain features have less complexity; however, the non-stationary nature of the EEG makes the time domain features less representative and leads to degradation in the accuracy. Diverse frequency domain features were developed to lower the noise impact of the EEG data acquisition procedures. Frequency domain features include the Power Spectral Density (PSD) [46] and Short-Time Fourier Transform (STFT) [47], among others, and provide a better spectral estimation for informative EEG segments. Although frequency domain features are more noise robust in comparison to time domain features, frequency domain features have a higher computational complexity and are hence less cost-effective. Time-frequency domain features provide high noise robustness when compared to frequency domain features.
The development of neural network models for data classification has also helped researchers in improving the categorization capability of their models. The LSTM was used in [48] to extract features directly for the EEG patterns and then to fuse these. The LSTM has a low complexity and can be applied directly to time domain EEG data. However, the LSTM showed a low performance in emotion recognition. In [49], EEGNet was employed with better compactness and complexity compared to the LSTM. The EEG data were represented in a 3D shape and classified with a 3D-CNN in [50]. The Pearson Correlation Coefficient was employed to shuffle the EEG channels before feature extraction in [51]. In [52], the EEG features were extracted with a recurrent neural network (RNN) in a parallel mode to improve the accuracy of the model. The 2D spectrogram version of the EEG signal was described with various pre-trained DNN models, such as AlexNet [53], ResNet [54], LeNet [55], VGGNet [56], and InceptionV3 [57], and a high accuracy was obtained; however, the dimensionality was extremely high for all 64 EEG channels. The deep neural networks require expensive GPUs for feature extraction, which is not affordable and thus limits the deployment in many emotion identification applications. The details of various feature extraction methods are summarized in Table 1.

2.3. Feature Selection and Transformation

Feature reduction methods aim at reducing the dimension of the feature vector by either feature selection or transformation. The resulting feature variables should have high interclass variations and low intraclass variability. Feature reduction methods identify appropriate features with a strong relationship to class labels. The raw extracted feature vectors are usually subject to feature reduction approaches to reduce the memory requirements and complexity. Many researchers have adopted correlation-based feature reduction methods such as the correlation-based Analysis of Variance (ANOVA) and Kendall’s rank coefficient. The learning methods recently developed include sparse regression, and LASSO uses L1 regularization for reducing the dimension of the feature vector. The stochastic methods depend on the global minima for feature variable selection. Robust feature reduction methods, such as Particle Swarm Optimization (PSO) [58], the Genetic Algorithm (GA) [59], Simulated Annealing (SA) algorithm, Correlation Feature Selection (CFS), and Minimum Redundancy Maximum Relevance (mRMR), outperformed conventional methods, such as the PCA, LDA, and ICA. A summary of common feature reduction/transformation algorithms is provided in Table 1.

2.4. Classification Algorithms

The classification stage is the last one in the EEG processing pipeline. The performance of a classifier is measured by comparing the true class label with the estimated one by the classifier. Researchers have employed both shallow learning classifiers and deep learning classifiers for emotion recognition. The shallow learning classifiers include the SVM, kNN, DT, MLP, and others, while the deep learning classifier includes the CNN, DNN, DBN, and other modified versions. In [60], the SEED and DEAP datasets were classified with the SVM and KNN, achieving 93.8 and 77.4% classification accuracy, respectively. The differential evolution-based channel selection algorithm (DEFS_Ch) [61] emotion classification model employed the linear discriminant analysis (LDA) and achieved 86.85% accuracy. The 2D transformation of an EEG was obtained with Azimuthal Equidistant Projection (AEP) [62] and employed the DNN model, achieving 96.09% accuracy. In [63], hybrid handcrafted features were used with an SVM achieving 75.64% accuracy. The CapsNet [64] classification on DEAP achieved 68.2% accuracy. In [65], the flexible analytical wavelet transform (FAWT) features from the SEED dataset gave 83.3% accuracy. The reduced set of features obtained through recursive emotion feature elimination (RFE) gave 60.5 and 90.33% using the DEAP and SEED datasets, respectively. A brief summary of the most common classification methods is given in Table 1.

2.5. EEG Datasets for Emotion Recognition

Human emotion identification can be visual based, oral based, and biosignal based, depending upon the application of interest. The visual-based emotion is identified through the facial expression and body posture of the subject. The oral-based emotion is identified with the pitch and volume of the subject’s voice. Both the visual- and oral-based emotion identification methods are less complex; however, the methods are less effective as subjects can conceal their real emotions in facial expressions and voice tones [66]. Therefore, biosignals have been considered as real effective sources for human emotion identification. The multi-channel electrodes are attached to the human scalp at defined positions to record the neural activity in the brain. EEG databases consisting of various emotion classes of different age subjects have been developed and are publicly available. The emotional state of the subject is excited with the help of different inputs, including audio, video, and images. The databases mostly employed in the emotional categorization experiments include GAMEEMO [67], K-EmoCon [68], SEED [69], ASCERTAIN [70], DEAP [26], MAHNOB-HCI [71], and IADS [72]. The summary of the above-mentioned databases is available in Table 2.

3. Proposed Workflow

In this section, we discuss the detailed architecture of our proposed workflow. The proposed model consists of a pre-processing stage and a spectrogram transformation stage, followed by the feature extraction and the classification stages, as shown in Figure 2. The filtering process is performed with a bandpass Butterworth filter designed for the specified range of frequencies. The spectrogram transformation stage converts the 1D EEG signals into 2D spectrograms. The raw feature vector is obtained by fusing deep features from the GoogLeNet DNN and the handcrafted OMTLBP_SMC [73] (developed by the authors). The feature clustering is performed to find the vocabulary vectors from the training set. The final hybrid feature vector is the frequency of raw features with similarities to the vocabulary vector set. To avoid classifier bias, a simple KNN classifier is then used to estimate the class label for the emotional state of the subject. The different stages of the proposed workflow are now discussed in more detail.

3.1. BoDHF Validation Datasets

The proposed BoDHF was validated and tested on SEED [69] and DEAP [26] datasets. The SEED dataset consists of 15 subjects in total from a set of 62 electrodes. A total of 3 experiments were performed in different sessions with 24 trials in each. The duration of the SEED EEG signal varies in duration from 185 to 265 s. The DEAP dataset was collected from 32 subjects over 32 EEG electrodes, with 1 experiment in each of the 40 trials. The EEG channel data of 63 s was recorded over a sampling rate of 128 Hz.

3.2. BoDHF Data Pre-Processing

Noise in raw EEG signals is present due to both internal and external artifacts. The internal artifacts are due to eyeball movement, eye blinking, heart pumping, respiration, and muscle movements. The external noise rise is mainly due to electrode displacement. To remove both lower and higher frequency noises, a simple bandpass Butterworth filter is used for the range of 4–45 Hz. Next, and because brain activity produces non-stationary EEG patterns with varying frequency content with time, the 1D filtered EEG signal is converted into a 2D spectrogram using the Short-Time Fourier Transform. Small segments of the EEG signal are obtained using overlapping windows to analyze the temporal variations in the spectrum. The generalized expression for the STFT is given in Equation (1).
S ( m , k ) = i = 1 L 1 E ( i + f m τ ) w ( i ) e ( j 2 π L i k )
where S ( m , k ) denotes the mth component of the time-frequency frame of a spectrogram, k denotes the Fourier coefficient, L is the window length in samples i, w ( i ) is the ith sample of the overlapping window, and the f m is the shifting step of the overlap window. The output 2D spectrogram depends on the shape of the overlap window employed, such as Unimodal, Gaussian, and Symmetric. The time resolution of the spectrogram depends on the size of the overlapping window. A small window provides more time localization, while a large window decreases time resolution but increases frequency discrimination. The values at each index of time and frequency are represented by the amplitude of the spectrogram at that position. The expression for the amplitude is defined in Equation (2).
A ( m , k ) = 1 L | S ( m , k ) | 2
where A ( m , k ) represents the amplitude component of the spectrogram for a specified time-frequency frame. The spectrograms in Figure 3 show examples of the neutral, positive, and negative emotional states of a given subject.

3.3. BoDHF BoDHF Raw Feature Extraction

The proposed bag-of-hybrid-deep-features (BoHDF) extraction model consists of the pre-processing method followed by the raw features extraction stage. The raw feature vector is a fusion of the GoogLeNet fully connected layer feature values and the texture feature values obtained from the OMTLBP_SMC [73]. The hybrid combination of deep and handcrafted features extraction method combines the high-level convolutional features with a low-level pattern representation model to capture more distinctive interclass variations. The OMTLBP_SMC describes the pattern of micro-structures that exist inside the time-frequency image of each EEG signal. The OMTLBP_SMC is invariant to orientation and translational changes in the input data and provides robustness against noise artifacts. The OMTLBP_SMC adds distinctiveness to the neural features and enhances the performance of the overall model. The OMTLBP_SMC is extracted from a 3 × 3 patch of the spectrogram. The variable P shown in Figure 2 (feature extraction stage) denotes the overlapped multi-scale fused segment of the spectrogram.
P k 3 = i = 0 6 ( A 0 + 15 ( i ) + 45 ( k 2 ) 3 ) 7 k = 1 8
P k 2 = i = 0 4 ( A 0 + 22.5 ( i ) + 45 ( k 2 ) 2 ) 5 k = 1 8
P k 1 = A 0 + 45 ( k 1 ) 1 k = 1 8
where the symbol P denotes the segment obtained through overlapped multi-oriented pixel fusion process of the spectrogram’s amplitude value at radius 2 and 3 around the position of the pixels as shown in Equations (3) and (4). The variable i defines the amplitude values lying at a specific radius, which gives the kth angle value in the multi-scale fused patch. Sample points P k 3 , P k 2 , and P k 1 express the topological structure of 8 sample values having radius 3, 2, and 1, respectively. The multi-scale fused segment in Figure 2 is obtained by averaging the similar-oriented adjacent multi-oriented fused segment. The similar orientation values at radius 1 and 2 are fused together while the similar oriented at radius 2 and 3 are combined to obtain three topological structures shown in the proposed framework of the model. The multi-scale fusion is defined in Equation (6).
U ˜ v = i = ( v 1 ) v P d i 2 d = 1 8
where U ˜ v defines the multi-scale fused topological structure with the subscript v expressing the radius of the topological 8 structure shown in Figure 2. The P d i in Equation (6) represents the d t h orientation values of the segments’ multi-oriented fused patch.
The sign and magnitude components are obtained from U ˜ v using Local Difference Sign-Magnitude Transform (LDSMT) shown in Equations (7) and (8). In the third stage, the feature vectors are extracted from each topological structure as shown in Figure 2.
Q v = U ˜ v d C v d = 1 P
Q v = s v d × m v d d = 1 P and { s v d = s i g n ( Q v d ) m v d = | Q v d |
where Q v is the local difference value, U ˜ v d is the dth sample point at radius d, and C v represents the center amplitude value corresponding to segment U ˜ v . s v k denotes the dth sign operator with a value equal to 1 when d v k is larger than or equal to 0, and −1 otherwise. The m v represents the magnitude component, containing the absolute values of the d v . The scale invariant operator OMTLBP_S P , R r i u 2 is calculated by using the s v operator; the encoding of this is similar to LBP P , R r i u 2 . The P and R define the sample points and the radius values. The rotation invariant operator OMTLBP_M P , R r i u 2 shown in Equation (9) is obtained from d v .
O M T L B P _ M P , R r i u 2 = { η ( m v d , ρ m ) , if ζ P , R 2 P + 1 , otherwise
ζ P , R = | η ( m v P 1 , ρ m ) η ( m v 0 , ρ m ) | + d = 1 P 1 | η ( m v d , ρ m ) η ( m v d 1 , ρ m ) |
where η ( i , j ) is 1 when i larger than or equal to j and η ( x ) = 0 when i < j . The m v d denotes the d t h magnitude component, while the ρ m defines the average value obtained from the magnitude component of the segment at position(v) of the spectrogram. The OMTLBP_C is formulated as
O M T L B P _ C P , R = η ( C v , ρ I )
where ρ I is the average of the compete spectrogram. The joint combination of the three operators is expressed with OMTLBP_SMC P , R r i u 2 describing the complete texture feature of the spectrogram. The OMTLBP_SMC [73] and GoogLeNet features have been concatenated at the feature stage of the proposed workflow as shown in Figure 2.
The DNN feature vector is obtained from the GoogLeNet model (other network structures were also considered). GoogLeNet is a multi-layer deep neural network model initially described in [74]. The 22-layer GoogLeNet has reduced complexity and was found not to overfit the EEG data signals, hence outperforming other existing DNN models in the classification of 2D transformed EEG signals. The GoogLeNet is almost 12 times fewer parameters than Alexnet. The network was developed for easy deployment over low-power electronics, such as mobile phones. Its basic idea is to convolve in parallel different sizes from most accurate (1 × 1) to bigger filters (5 × 5). Overall, the GoogLeNet was used for our application as it trains faster than VGG, and the size of a pre-trained network is comparatively smaller than VGG. Moreover, the nine inception layers of the GoogLeNet boost the deep learning outputs and avoid overfitting. For our application, the 2D spectrogram images are resized to 224 × 224 to match the input layer of the GoogLeNet model. The spectrogram is fed into the GoogLeNet and processed through the convolutional and max-pooling layers of the network. The pooling layer of the model reduces the 3 × 3 and 5 × 5 blocks into a single value. The rectification is performed with the traditional RELU expression given in Equation (12).
R e L U = m a x 0 , x
where the R e L U is the rectified linear value, which reflects the maximum value between the input and 0. We have collected our desired feature values from the fully connected layer of the model, while the rest of the DNN stages were ignored. The resulting 1000 DNN features were concatenated to 600 OMTLBP feature vector to obtain the final raw feature vector of size 1600, as shown in Figure 2.
The feature extraction process is followed by a clustering stage. Here, the training data are used to create a vocabulary set. All extracted raw feature vectors are processed to create “k” groups of feature vectors. The k-clusters are created using an iterative process, described in Equations (13) and (14). The total of k-cluster centroids are initially selected, randomly represented with ψ j , with j ranging from 1 to k.
ψ j = i = 1 m ( x j i ) m
x j = arg m i n j | | x i ψ j | | 2
where ψ j denotes the centroid of cluster j, and the x j i defines the ith feature vector belonging to cluster j. The iterative process continues until the difference between the centroid in the ith and i 1 iteration reaches a certain threshold or limit. The cluster centroids for each class describe the vocabulary set of the dataset. The final histogram is the frequencies of occurrence of each vocabulary centroid in the raw feature data representing the final feature vector. The bag-of-hybrid-deep-features (BoHDF) vector for any emotional status is extracted by calculating the frequency of vocabulary in the raw features of spectrograms. The test features are classified using a simple K-nearest neighbor (KNN) and support vector machine (SVM) classifiers. Such classifiers are chosen given their simplicity, satisfactory performance, and to check the overall performance of the proposed workflow independently of the power of the final classification stage.

4. Experimental Results and Discussion

For our experimental setup, 32 channels out of 64 EEG electrodes were selected for the emotion recognition problem. The hybrid features consisting of the handcrafted and deep features were used with various classifiers, including the ensemble, decision tree, SVM, and KNN with a Gaussian kernel, cubic, quadratic, and weighted Gaussian, to identify the best combination of classifiers and kernels. An Intel®Core i7-6500U CPU with 2.50 GHz and 8 GB RAM with MATLAB software was used to implement and test the performance of the proposed workflow. The proposed framework was evaluated on the publicly available PhysioNet’s DEAP and SEED datasets. The datasets were split into 70% training and 30% testing segments with a 5-fold cross-validation procedure to train and test the classifier(s).
The classification performance of the proposed model was evaluated using two publicly available databases, namely the SEED and DEAP datasets. The BoHDF model was used for the feature extraction stage while the SVM, ensemble, tree, and KNN classifiers were used to classify the test EEG patterns into their respective emotion classes. The DNN models tested in this research work include the GoogLeNet, AlexNet, ResNet-50, ResNet-101, Inceptionv3, and ResNet models. The images of dimension 224 × 224 (from the spectrograms) were used as inputs to the different network’s first layer. The row feature vectors of dimension 1024 were then extracted at the fully connected layer. Finally, the OMTLBP_SMC feature vector of dimension 600 was added for the complete description of the EEG patterns. Different cluster sizes were considered, k = 8 , 10 , 12 , to find the optimal value of k. The cluster size 8 provided the highest classification accuracy on both datasets presented in Table 3. The performance of the proposed model is compared with the Multiband Feature Matrix (MFM) [64], Multivariate Empirical Mode Decomposition (MEMD) [75], Mel-Frequency Cepstral Coefficients (MFCC) [76], Recursive Feature Elimination (RFE) [77], differential entropy (DE) [58], bag of deep features (BoDF) [60], Empirical Mode Decomposition (EMD) [78], Shannon’s entropy [79], EMD combined with sample entropy (EMDSE) [47], spatial-temporal recurrent neural network (STRNN) [65], and Higher-Order Statistics (HOS) [80], and the results show that the proposed workflow consistently outperforms the existing state-of-the-art approaches on both the DEAP and SEED databases. The proposed model, consisting of GoogLeNet fully connected layer features, when tested on both the DEAP and SEED datasets, outperformed the state-of-the-art, as mentioned in Table 3. The proposed model in integration with GoogLeNet and the KNN classifier achieved the best performance of 93.83% and 95.95% on the DEAP and SEED databases, respectively.
In Table 4, different clusters were used to extract the BoDHF features integrated with different DNN features and classifiers, and the classification performance of the GoogLeNet and KNN on cluster size 8 has achieved the highest accuracy. The classification performance of the ensemble classifier provides an accuracy of 94.6% with GoogLeNet with cluster size 8, which is higher than AlexNet, ResNet-50, ResNet-101, and InceptionV3 which gave an accuracy of 94.2, 93.3, 93.2, and 93.5%, respectively. The tree classifier provided the highest accuracy of up to 92.6% with GoogLeNet and the tree classifier and a cluster size of 8; however, this is lower than that obtained with cluster size 12. The cluster size 12 on GoogLeNet and the tree classifier provided a 94.2% accuracy which is higher than the AlexNet, ResNet-50, ResNet-101, and InceptionV3 models with accuracies of 92.4, 91.1, 92.3, and 92.8%, respectively, for the same cluster size.
The classification performance of 95.1% was achieved with an SVM classifier integrated to the GoogLeNet features of the BoDHF on cluster size 8, which is highest in comparison to the other DNN models for the same classifier and cluster size. The AlexNet, ResNet-50, ResNet-101, and InceptionV3 models achieved a 92.0%, 92.8%, 91.5%, and 92.8% classification accuracy, respectively, for cluster size 8. The classification performance of the proposed model BoHDF when tested on various combinations of the integrated DNN feature extraction model and classifier has provided the highest performance over cluster size 8. Therefore, we have selected eight clusters in the proposed BoDHF model.

5. Conclusions

An EEG-based emotion recognition framework was proposed and evaluated on the DEAP and SEED datasets. Out of the 64 channels, only 32 of them were considered to reduce the cost and complexity. A hybrid combination of the handcrafted OMTLBP_SMC and the GoogLeNet DNN model was used to extract the features from the 2D spectrogram representing the EEG signals. Furthermore, the bag-of-features model with a cluster of size 8 was used to reduce the dimension of the developed hybrid model. The proposed model achieved 93.83% and 96.95% recognition accuracy on the DEAP and SEED databases, respectively, which was higher than recently reported works.
The limitation of the proposed BoDHF is its added complexity due to the feature extraction procedure comprising raw hybrid deep and handcrafted features and the vocabulary estimation for dimension reduction and adding robustness to the model. The future work in this project will include the development of merging the handcrafted layers in the DNN model to reduce the complexity.

Author Contributions

F.M.A. investigation and project administration and funding acquisition; F. software, validation, writing, and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by the Institutional Fund Projects under grant no. (IFPIP: 109-830-1443). Therefore, the authors gratefully acknowledge the technical and financial support from the Ministry of Education and King Abdul Aziz University, DSR, Jeddah, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset utilized in this work is already availbie on the physio net website with titles DEAP and SEEDs, where the details have already been provided.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cai, S.; Li, P.; Su, E.; Liu, Q.; Xie, L. A Neural-Inspired Architecture for EEG-Based Auditory Attention Detection. IEEE Trans. Hum.-Mach. Syst. 2022, 52, 668–676. [Google Scholar] [CrossRef]
  2. Medeiros, L.; Bosse, T.; Gerritsen, C. Can a Chatbot Comfort Humans? Studying the Impact of a Supportive Chatbot on Users’ Self-Perceived Stress. IEEE Trans. Hum.-Mach. Syst. 2021, 52, 343–353. [Google Scholar] [CrossRef]
  3. Nourmohammadi, A.; Jafari, M.; Zander, T.O. A survey on unmanned aerial vehicle remote control using brain-computer interface. IEEE Trans. Hum.-Mach. Syst. 2018, 48, 337–348. [Google Scholar] [CrossRef]
  4. Li, D.; Wang, Z.; Gao, Q.; Song, Y.; Yu, X.; Wang, C. Facial expression recognition based on Electroencephalogram and facial landmark localization. Technol. Health Care 2019, 27, 373–387. [Google Scholar] [CrossRef]
  5. Ho, N.; Yang, H.; Kim, S.; Lee, G. Multimodal approach of speech emotion recognition using multi-level multi attention based recurrent neural network. IEEE Access 2020, 8, 61672–61686. [Google Scholar] [CrossRef]
  6. Yang, Y.; Gao, Q.; Song, X.; Song, Y.; Mao, Z.; Liu, J. Facial expression and EEG fusion for investigating continuous emotions of deaf subjects. IEEE Sens. J. 2021, 21, 16894–16903. [Google Scholar] [CrossRef]
  7. Li, D.; Wang, Z.; Wang, C.; Liu, S.; Chi, W.; Dong, E.; Song, X.; Gao, Q.; Song, Y. The fusion of electroencephalography and facial expression for continuous emotion recognition. IEEE Access 2019, 7, 155724–155736. [Google Scholar] [CrossRef]
  8. Said, Y.; Barr, M. Human emotion recognition based on facial expressions via deep learning on high-resolution images. Multimed. Tools Appl. 2021, 80, 25241–25253. [Google Scholar] [CrossRef]
  9. Chowanda, A.; Sutoyo, R.; Tanachutiwat, S. Exploring text-based emotions recognition machine learning techniques on social media conversation. Procedia Comput. Sci. 2021, 179, 821–828. [Google Scholar] [CrossRef]
  10. Shu, L.; Xie, J.; Yang, M.; Li, Z.; Li, Z.; Liao, D.; Xu, X.; Yang, X. A review of emotion recognition using physiological signals. Sensors 2018, 18, 2074. [Google Scholar] [CrossRef]
  11. Chen, S.; Jiang, K.; Hu, H.; Kuang, H.; Yang, J.; Luo, J.; Chen, X.; Li, Y. Emotion Recognition Based on Skin Potential Signals with a Portable Wireless Device. Sensors 2021, 21, 1018. [Google Scholar] [CrossRef] [PubMed]
  12. Sharma, A.; Canavan, S. Multimodal physiological-based emotion recognition. In International Conference on Pattern Recognition; Springer: Cham, Switzerland, 2021; pp. 101–113. [Google Scholar]
  13. Zhang, T.; El Ali, A.; Wang, C.; Hanjalic, A.; Cesar, P. Corrnet: Fine-grained emotion recognition for video watching using wearable physiological sensors. Sensors 2021, 21, 52. [Google Scholar] [CrossRef] [PubMed]
  14. Gao, Q.; Yang, Y.; Kang, Q.; Tian, Z.; Song, Y. EEG-based Emotion Recognition with Feature Fusion Networks. Int. J. Mach. Learn. Cybern. 2022, 13, 421–429. [Google Scholar] [CrossRef]
  15. Xu, X.; Wei, F.; Zhu, Z.; Liu, J.; Wu, X. Eeg Feature Selection Using Orthogonal Regression: Application to Emotion Recognition. In Proceedings of the 45th International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1239–1243. [Google Scholar]
  16. Li, G.; Lee, C.H.; Jung, J.J.; Youn, Y.C.; Camacho, D. Deep learning for EEG data analytics: A survey. Concurr. Comput. Pract. Exp. 2020, 32, e5199. [Google Scholar] [CrossRef]
  17. Patel, P.; Annavarapu, R.N. EEG-based human emotion recognition using entropy as a feature extraction measure. Brain Inform. 2021, 8, 1–13. [Google Scholar] [CrossRef] [PubMed]
  18. Ahern, G.L.; Schwartz, G.E. Differential lateralization for positive and negative emotion in the human brain: EEG spectral analysis. Neuropsychologia 1985, 23, 745–755. [Google Scholar] [CrossRef] [PubMed]
  19. Alm, C.; Roth, D.; Sproat, R. Emotions from text: Machine learning for text-based emotion prediction. In Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing, Vancouver, BC, Canada, 6–8 October 2005; pp. 579–586. [Google Scholar]
  20. Scherer, K. Which emotions can be induced by music? what are the underlying mechanisms? and how can we measure them? J. New Music Res. 2004, 33, 239–251. [Google Scholar] [CrossRef] [Green Version]
  21. Hubert, W.; de Jong-Meyer, R. Autonomic, neuroendocrine, and subjective responses to emotion inducing film stimuli. Int. J. Psychophysiol. 1991, 11, 131–140. [Google Scholar] [CrossRef]
  22. Zheng, X.; Liu, X.; Zhang, Y.; Cui, L.; Yu, X. A portable HCI system oriented EEG feature extraction and channel selection for emotion recognition. Int. J. Intell. Syst. 2021, 36, 152–176. [Google Scholar] [CrossRef]
  23. Ansari-Asl, K.; Chanel, G.; Pun, T. A channel selection method for EEG classification in emotion assessment based on synchronization likelihood. In Proceedings of the 2007 15th European Signal Processing Conference, Poznan, Poland, 3–7 September 2007; pp. 1241–1245. [Google Scholar]
  24. Asghar, M.A.; Khan, M.J.; Rizwan, M.; Shorfuzzaman, M.; Mehmood, R.M. AI inspired EEG-based spatial feature selection method using multivariate empirical mode decomposition for emotion classification. Multimed. Syst. 2022, 28, 1275–1288. [Google Scholar] [CrossRef]
  25. Zheng, W.-L.; Lu, B.-L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Autom. Mental Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  26. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  27. Peng, H.; Wang, Y.; Chao, J.; Huo, X.; Majoe, D. Stability study of the optimal channel selection for emotion classification from EEG. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017; pp. 2031–2036. [Google Scholar]
  28. Wang, Z.M.; Hu, S.Y.; Song, H. Channel selection method for eeg emotion recognition using normalized mutual information. IEEE Access 2019, 7, 143303–143311. [Google Scholar] [CrossRef]
  29. Lan, T.; Erdogmus, D.; Adami, A.; Pavel, M.; Mathan, S. Salient EEG channel selection in brain computer interfaces by mutual information maximization. In Proceedings of the 27th Annual International Conference of IEEE Engineering in Medicine and Biology Society, Shanghai, China, 17–18 January 2006; pp. 7064–7067. [Google Scholar]
  30. Alotaiby, T.; El-Samie, F.E.A.; Alshebeili, S.A.; Ahmad, I. A review of channel selection algorithms for EEG signal processing. EURASIP J. Adv. Signal Process. 2015, 66, 1–21. [Google Scholar] [CrossRef] [Green Version]
  31. Ghaemi, A.; Rashedi, E.; Pourahimi, A.M.; Kamandar, M.; Rahdari, F. Automatic channel selection in EEG signals for classification of left or right hand movement in Brain Computer Interfaces using improved binary gravitation search algorithm. Biomed. Signal Process. Control 2017, 3, 109–118. [Google Scholar] [CrossRef]
  32. Lal, T.N.; Schröder, M.; Hinterberger, T.; Weston, J.; Bogdan, M.; Birbaumer, N.; Schölkopf, B. Support vector channel selection in BCI. IEEE Trans. Biomed. Eng. 2004, 51, 1003–1010. [Google Scholar] [CrossRef] [Green Version]
  33. Demir, F.; Sobahi, N.; Siuly, S.; Sengur, A. Exploring Deep Learning Features For Automatic Classification Of Human Emotion Using EEG Rhythms. IEEE Sens. J. 2021, 21, 14923–14930. [Google Scholar] [CrossRef]
  34. Abdulbaqia, A.S.; Younisb, M.T.; Younus, Y.T.; Obaidd, A.J. A hybrid technique for EEG signals evaluation and classification as a step towards to neurological and cerebral disorders diagnosis. Int. J. Nonlinear Anal. Appl. 2022, 13, 773–781. [Google Scholar]
  35. Ludwig, S.; Bakas, S.; Adamos, D.A.; Laskaris, N.; Panagakis, Y.; Zafeiriou, S. EEGminer: Discovering Interpretable Features of Brain Activity with Learnable Filters. arXiv 2021, arXiv:2110.10009. [Google Scholar]
  36. Mandhouj, B.; Cherni, M.A.; Sayadi, M. An automated classification of EEG signals based on spectrogram and CNN for epilepsy diagnosis. Analog. Integr. Circuits Signal Process. 2021, 108, 101–110. [Google Scholar] [CrossRef]
  37. Baygin, M.; Dogan, S.; Tuncer, T.; Barua, P.D.; Faust, O.; Arunkumar, N.; Abdulhay, E.W.; Palmer, E.E.; Acharya, U.R. Automated ASD detection using hybrid deep lightweight features extracted from EEG signals. Comput. Biol. Med. 2021, 134, 104548. [Google Scholar] [CrossRef]
  38. Wang, W.; Gill, E.W. Comparison of a modified periodogram and standard periodogram for current estimation by an hf surface radar. In Proceedings of the Ocean 2014-TAIPEI, Taipei, Taiwan, 7–10 April 2014; pp. 1–7. [Google Scholar]
  39. Jenke, R.; Peer, A.; Buss, M. Feature extraction and selection for emotion recognition from EEG. IEEE Trans. Affect. Comput. 2014, 5, 327–339. [Google Scholar] [CrossRef]
  40. Li, M.; Lu, B.-L. Emotion classification based on gamma-band EEG. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 1223–1226. [Google Scholar]
  41. Liu, Y.; Sourina, O.; Nguyen, M.K. Real-time EEG-based emotion recognition and its applications. In Transaction on Computational Science XII; Springer: Berlin/Heidelberg, Germany, 2011; pp. 256–277. [Google Scholar]
  42. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef]
  43. Inouye, T.; Shinosaki, K.; Sakamoto, H.; Toi, S.; Ukai, S.; Iyama, A.; Katsuda, Y.; Hirano, M. Quantification of EEG irregularity by use of the entropy of the power spectrum. Electroencephalogr. Clin. Neurophysiol. 1991, 79, 204–210. [Google Scholar] [CrossRef] [PubMed]
  44. Sleigh, J.W.; Olofsen, E.; Dahan, A.; De Goede, J.; Steyn-Ross, D.A. Entropies of the EEG: The effects of general anaesthesia. Paper presented at the 5th International Conference on Memory, Awareness and Consciousness, New York, NY, USA, 1–3 June 2001. [Google Scholar]
  45. Hosseini, S.A.; Naghibi-Sistani, M.B. Emotion recognition method using entropy analysis of EEG signals. Int. J. Image Graph. Signal Process. 2011, 3, 30. [Google Scholar] [CrossRef] [Green Version]
  46. Al-Fahoum, A.S.; Al-Fraihat, A.A. Methods of EEG signal features extraction using linear analysis in frequency and time-frequency domains. ISRN Neurosci. 2014, 2014, 730218. [Google Scholar] [CrossRef] [Green Version]
  47. Zhang, Y.; Ji, X.; Zhang, S. An approach to EEG-based emotion recognition using combined feature extraction method. Neurosci. Lett. 2016, 633, 152–157. [Google Scholar] [CrossRef]
  48. Ahmed, M.Z.I.; Sinha, N. EEG-based emotion classification using LSTM under new paradigm. Biomed. Phys. Eng. Express 2021, 7, 065018. [Google Scholar] [CrossRef]
  49. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef] [Green Version]
  50. Salama, E.S.; El-Khoribi, R.A.; Shoman, M.E.; Shalaby, M.A.W. EEG-based emotion recognition using 3D convolutional neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 329–337. [Google Scholar] [CrossRef] [Green Version]
  51. Wen, Z.; Xu, R.; Du, J. A novel convolutional neural networks for emotion recognition based on EEG signal. In Proceedings of the 2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Shenzhen, China, 15–17 December 2017; pp. 672–677. [Google Scholar]
  52. Yang, Y.; Wu, Q.; Qiu, M.; Wang, Y.; Chen, X. Emotion recognition from multi-channel EEG through parallel convolutional recurrent neural network. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  53. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classifcation with deep convolutional neural networks. In Advances in Neural Information Processing Systems; ACM: New York, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  55. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  56. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  57. Szegedy, C.; Vanhoucke, V.; Iofe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  58. Nakisa, B.; Rastgoo, M.N.; Tjondronegoro, D.; Chandran, V. Evolutionary computation algorithms for feature selection of EEG-based emotion recognition using mobile sensors. Expert Syst. Appl. 2018, 93, 143–155. [Google Scholar] [CrossRef]
  59. Rodionov, A.S.; L’vov, A.A. Comparison of linear, nonlinear and feature selection methods for EEG signal classification. In Proceedings of the International Conference on Actual Problems of Electron Devices Engineering, 2004 (APEDE 2004), Saratov, Russia, 15–16 September 2004; pp. 436–439. [Google Scholar] [CrossRef]
  60. Asghar, M.A.; Khan, M.J.; Fawad; Amin, Y.; Rizwan, M.; Rahman, M.; Badnava, S.; Mirjavadi, S.S. EEG-Based Multi-Modal Emotion Recognition using Bag of Deep Features: An Optimal Feature Selection Approach. Sensors 2019, 19, 5218. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Al-Qazzaz, N.K.; Sabir, M.K.; Ali, S.; Ahmad, S.A.; Grammer, K. Effective EEG Channels for Emotion Identification over the Brain Regions using Differential Evolution Algorithm. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019. [Google Scholar]
  62. Ozdemir, M.A.; Degirmenci, M.; Guren, O.; Akan, A. EEG based Emotional State Estimation using 2-D Deep Learning Technique. In Proceedings of the 2019 Medical Technologies Congress (TIPTEKNO), Izmir, Turkey, 3–5 October 2019. [Google Scholar]
  63. Guo, K.; Yu, H.; Chai, R.; Nguyen, H.; Su, S.W. A Hybrid Physiological Approach of Emotional Reaction Detection Using Combined FCM and SVM Classifier. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019. [Google Scholar]
  64. Chao, H.; Dong, L.; Liu, Y.; Lu, B. Emotion Recognition from Multiband EEG Signals Using CapsNet. Sensors 2019, 19, 2212. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Zhang, T.; Zheng, W.; Cui, Z.; Zong, Y.; Li, Y. Spatial-Temporal Recurrent Neural Network for Emotion Recognition. IEEE Trans. Cybern. 2019, 49, 839–847. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Liu, Y.; Fu, G. Emotion recognition by deeply learned multi-channel textual and EEG features. Future Gener. Comput. Syst. 2021, 119, 1–6. [Google Scholar] [CrossRef]
  67. Alakus, T.B.; Gonen, M.; Turkoglu, I. Database for an emotion recognition system based on EEG signals and various computer games GAMEEMO. Biomed. Signal Process. Control 2020, 60, 101951. [Google Scholar] [CrossRef]
  68. Park, C.Y.; Cha, N.; Kang, S.; Kim, A.; Khandoker, A.H.; Hadjileontiadis, L.; Oh, A.; Jeong, Y.; Lee, U. K-EmoCon, a multimodal sensor dataset for continuous emotion recognition in naturalistic conversations. Sci. Data 2020, 7, 1–16. [Google Scholar] [CrossRef]
  69. Duan, R.-N.; Zhu, J.-Y.; Lu, B.-L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar]
  70. Subramanian, R.; Wache, J.; Abadi, M.K.; Vieriu, R.L.; Winkler, S.; Sebe, N. ASCERTAIN: Emotion and personality recognition using commercial sensors. IEEE Trans. Affect. Comput. 2018, 9, 147–160. [Google Scholar] [CrossRef]
  71. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef] [Green Version]
  72. Bradley, M.M.; Lang, P.J. The International Affective Digitized Sounds Affective Ratings of Sounds and Instruction Manual; Tech. Rep. B-3; Univ. Florida: Gainesville, FL, USA, 2007; pp. 29–46. [Google Scholar]
  73. Khan, M.J.; Riaz, M.A.; Shahid, H.; Khan, M.S.; Amin, Y.; Loo, J.; Tenhunen, H. Texture representation through overlapped multi-oriented tri-scale local binary pattern. IEEE Access 2019, 7, 66668–66679. [Google Scholar]
  74. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  75. Mert, A.; Akan, A. Emotion recognition from EEG signals by using multivariate empirical mode decomposition. Pattern Anal. Appl. 2016, 21, 81–89. [Google Scholar] [CrossRef]
  76. Gupta, V.; Chopda, M.D.; Pachori, R.B. Cross-Subject Emotion Recognition Using Flexible Analytic Wavelet Transform From EEG Signals. IEEE Sens. J. 2019, 19, 2266–2274. [Google Scholar] [CrossRef]
  77. Li, X.; Song, D.; Zhang, P.; Zhang, Y.; Hou, Y.; Hu, B. Exploring EEG features in cross-subject emotion recognition. Front. Neurosci. 2018, 12, 162. [Google Scholar] [CrossRef] [Green Version]
  78. Eftekhar, A.; Toumazou, C.; Drakakis, E.M. Empirical mode decomposition: Real-time implementation and applications. J. Signal Process. Syst. 2013, 73, 43–58. [Google Scholar] [CrossRef]
  79. Vijayan, A.E.; Sen, D.; Sudheer, A.P. EEG-based emotion recognition using statistical measures and auto-regressive modeling. In Proceedings of the 2015 IEEE International Conference on Computational Intelligence and Communication Technology, Ghaziabad, India, 13–14 February 2015; pp. 587–591. [Google Scholar]
  80. Chakladar, D.D.; Chakraborty, S. EEG based emotion classification using “correlation based subset selection”. Biol. Inspired Cognitive Archit. 2018, 24, 98–106. [Google Scholar] [CrossRef]
Figure 1. EEG electrodes placement on the human scalp.
Figure 1. EEG electrodes placement on the human scalp.
Sensors 23 00498 g001
Figure 2. Proposed workflow.
Figure 2. Proposed workflow.
Sensors 23 00498 g002
Figure 3. Two-dimensional spectrogram representing different emotions from the SEED dataset (Channel 1).
Figure 3. Two-dimensional spectrogram representing different emotions from the SEED dataset (Channel 1).
Sensors 23 00498 g003
Table 1. Summary of frequently used techniques for emotion classification using EEG patterns.
Table 1. Summary of frequently used techniques for emotion classification using EEG patterns.
Pre-ProcessingFeature ExtractionFeature ReductionClassification
Short-Time Fourier TransformPower Spectral Density of Alpha, Beta, Gamma, Theta, and DeltaMaximum Redundancy Maximum Relevance (mRMR)Convolutional Neural Network (CNN)
Down-samplingEntropyLinear Discriminant Analysis (LDA)Deep Neural Network (DNN) (different variants)
Band-pass FilteringMean, Variance, and Standard DeviationPrincipal Component Analysis (PCA)Recurrent Neural Network (RNN)
Adaptive FilteringSlope Sign ChangesCorrelation-based Feature Selector (CFS)Generative Adversarial Network (GAN)
Channel SelectionHilbert–Hung TransformNon-Negative Matrix Factorization (NNMF) and OthersSupport Vector Machine (SVM)
Spectrogram ConversionPower of DWT of Sub-bands-K- Nearest Neighbor (KNN)
Wavelet TransformRecursive Energy Efficiency of DW-Naive Bayes (NB)
Discrete Cosine Transform and OthersHigh-Order Crossing and Others Decision Tree (DT) and Others
Table 2. Publicly available databases for emotion identification.
Table 2. Publicly available databases for emotion identification.
DatasetYearSubjects (M/F)Age RangeStimuliAcquisition Device
GAMEEMO [67]202028(20–27)Emotional game videoEPOC+14 channel Emotiv
K-EmoCon [68]202032 (20/12)(19–36)Video footage of the debateNeuroSky Mind Wave Headset
SEED [69]201715 (7/8)-Film clips64-channels device
ASCERTAIN [70]201658 (37/21)Avg. 3036 movie clipsDry Electrode EEG device
DEAP [26]201232 (16/16)(19–37)VideoBiosemi Active Two
MAHNOB-HCI [71]201127 (11/16)(19–40)Video, imagesBiosemi Active Two
IADS [72]2007--Sounds-
Table 3. Comparison of our proposed model with recent emotion identification methods.
Table 3. Comparison of our proposed model with recent emotion identification methods.
Ref.MethodDatasetChannelsClassifierAccuracy (%)
[64]MFMDEAP18CapsNet68.2
[75]MEMDDEAP12ANN75
[75]MEMDDEAP12KNN67
[76]MFCCSEED12SVM83.5
[77]RFEDEAP12SVM60.5
[58]DEDEAP32PNN79.3
[60]BODFDEAP32SVM77.4
[60]BODFDEAP32KNN73.6
[78]EMDDEAP62SVM63.4
[78]EMDDEAP62KNN57.3
[79]Shannon’s entropyDEAP-Multi-class SVM94.09
[47]EMD + sample entropyDEAP-SVM94.98
[47]EMD + sample entropyDEAP-Multi-class SVM93.2
Proposed alg.AlexNetDEAP28SVM (Fine Gaussian)92.1
Proposed alg.AlexNetDEAP28KNN (Weighted)93.2
Proposed alg.ResNet-50DEAP40SVM (Fine Gaussian)91.5
Proposed alg.ResNet-50DEAP40KNN (Weighted)92.7
Proposed alg.ResNet-101DEAP29SVM (Fine Gaussian)91.8
Proposed alg.ResNet-101DEAP29KNN (Weighted)91.5
Proposed alg.InceptionV3DEAP32SVM (Cubic)91.9
Proposed alg.InceptionV3DEAP32KNN (Weighted)92.4
Proposed alg.GoogLeNetDEAP32KNN (Weighted)93.83
[76]MFCCSEED12Random Forest72.07
[65]STRNNSEED62CNN89.5
[77]RFESEED18SVM90.4
[60]BODFSEED62SVM93.8
[60]BODFSEED62KNN91.4
[80]HOSSEED34SVM82.1
[78]EMDSEED62SVM56
[78]EMDSEED62KNN52.6
Proposed alg.AlexNetSEED28SVM (Fine Gaussian)93.7
Proposed alg.AlexNetSEED28KNN (Weighted)94.6
Proposed alg.ResNet-50SEED40SVM (Fine Gaussian)92.8
Proposed alg.ResNet-50SEED40KNN (Weighted)93.9
Proposed alg.ResNet-101SEED29SVM (Fine Gaussian)92.4
Proposed alg.ResNet-101SEED29KNN (Weighted)92.8
Proposed alg.InceptionV3SEED32SVM (Cubic)92.7
Proposed alg.InceptionV3SEED32KNN (Weighted)93.3
Proposed alg.GoogLeNetSEED32KNN (Weighted)96.95
Table 4. Performance comparison of the proposed method integrated with various DNN models and different cluster sizes using the SEED dataset.
Table 4. Performance comparison of the proposed method integrated with various DNN models and different cluster sizes using the SEED dataset.
ClustersDNN ModelClassifierAccuracy (%)ClustersDNN ModelClassifierAccuracy (%)ClustersDNN ModelClassifierAccuracy (%)
12AlexNetEnsemble93.710AlexNetEnsemble92.48AlexNetEnsemble94.2
12AlexNetTree92.410AlexNetTree92.08AlexNetTree92.1
12AlexNetSVM93.310AlexNetSVM93.38AlexNetSVM92.0
12AlexNetKNN94.610AlexNetKNN93.78AlexNetKNN92.4
12ResNet-50Ensemble93.710ResNet-50Ensemble92.48ResNet-50Ensemble93.3
12ResNet-50Tree91.110ResNet-50Tree90.28ResNet-50Tree91.1
12ResNet-50SVM92.810ResNet-50SVM92.08ResNet-50SVM92.8
12ResNet-50KNN92.810ResNet-50KNN92.48ResNet-50KNN92.4
12ResNet-101Ensemble92.410ResNet-101Ensemble92.88ResNet-101Ensemble93.2
12ResNet-101Tree92.310ResNet-101Tree92.08ResNet-101Tree92.1
12ResNet-101SVM92.010ResNet-101SVM92.48ResNet-101SVM91.5
12ResNet-101KNN92.010ResNet-101KNN92.48ResNet-101KNN92.8
12InceptionV3Ensemble94.210InceptionV3Ensemble92.08InceptionV3Ensemble93.5
12InceptionV3Tree92.810InceptionV3Tree94.38InceptionV3Tree91.5
12InceptionV3SVM92.410InceptionV3SVM92.48InceptionV3SVM92.8
12InceptionV3KNN92.410InceptionV3KNN92.08InceptionV3KNN91.5
12GoogLeNetEnsemble92.010GoogLeNetEnsemble93.38GoogLeNetEnsemble94.6
12GoogLeNetTree94.210GoogLeNetTree91.18GoogLeNetTree92.6
12GoogLeNetSVM92.110GoogLeNetSVM92.08GoogLeNetSVM95.1
12GoogLeNetKNN92.710GoogLeNetKNN93.68GoogLeNetKNN96.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alotaibi, F.M.; Fawad. An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status. Sensors 2023, 23, 498. https://doi.org/10.3390/s23010498

AMA Style

Alotaibi FM, Fawad. An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status. Sensors. 2023; 23(1):498. https://doi.org/10.3390/s23010498

Chicago/Turabian Style

Alotaibi, Fahad Mazaed, and Fawad. 2023. "An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status" Sensors 23, no. 1: 498. https://doi.org/10.3390/s23010498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop