Next Article in Journal
Threat Analysis and Distributed Denial of Service (DDoS) Attack Recognition in the Internet of Things (IoT)
Next Article in Special Issue
Machine Learning-Based Feature Selection and Classification for the Experimental Diagnosis of Trypanosoma cruzi
Previous Article in Journal
Effect of Phase Noise on the Optical Millimeter-Wave Signal in the DWDM-RoF System
Previous Article in Special Issue
Bidimensional and Tridimensional Poincaré Maps in Cardiology: A Multiclass Machine Learning Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning

by
Stanisław Saganowski
Department of Artificial Intelligence, Wrocław University of Science and Technology, 50-370 Wrocław, Poland
Electronics 2022, 11(3), 496; https://doi.org/10.3390/electronics11030496
Submission received: 31 December 2021 / Revised: 22 January 2022 / Accepted: 2 February 2022 / Published: 8 February 2022
(This article belongs to the Special Issue Machine Learning in Electronic and Biomedical Engineering)

Abstract

:
Bringing emotion recognition (ER) out of the controlled laboratory setup into everyday life can enable applications targeted at a broader population, e.g., helping people with psychological disorders, assisting kids with autism, monitoring the elderly, and general improvement of well-being. This work reviews progress in sensors and machine learning methods and techniques that have made it possible to move ER from the lab to the field in recent years. In particular, the commercially available sensors collecting physiological data, signal processing techniques, and deep learning architectures used to predict emotions are discussed. A survey on existing systems for recognizing emotions in real-life scenarios—their possibilities, limitations, and identified problems—is also provided. The review is concluded with a debate on what challenges need to be overcome in the domain in the near future.

1. Introduction

Emotions are a basic component of our life, just like breathing or eating. They are also responsible for a majority of our decisions [1]. For that reason, emotion recognition is an important and profitable research problem. The ability to recognize emotions can help in emotion-based disorders [2], autism [3,4], monitoring our well-being [5,6] and mental health [7], controlling stress [8], human–computer interaction [9], recommendation systems [10,11], and computer games [12,13]. At the same time, emotion recognition is a very ambitious task, as it connects several disciplines, i.e., psychology, electronics/sensors, signal processing, and machine learning; see Figure 1a.
For years, psychologists have analyzed people and the processes taking place between the affective situation and human reaction and behavior. This allowed them identify different perspectives on emotions [14] and define various models such as basic emotion models [15,16,17], appraisal models [18,19], psychological construction models [20,21], and social construction models [22,23]. As a consequence, there is no single commonly accepted definition of emotions.
Electronics created small and effective sensors that capture images/videos, record audio, and collect physiological signals, as seen in Figure 1b. Signal processing allows us to efficiently process the collected data, improving their quality and obtaining informative features. Machine learning, in turn, enables us to analyze vast amounts of data, automatically extract patterns, create affective states’ representations, and thus recognize or even predict emotions or other affective states.
This work provides the state of the art in the fields of wearable sensors, signal processing, and machine learning models and techniques, adequate for the emotion recognition from speech, facial images, and physiological signals. The crucial aspects of each domain are explained with examples. Rather than providing the details of each method/system, their general idea is presented, positioned in the context of the problem, and the reference to the source article is provided to keep the article concise.
The primary focus of this review is on emotion recognition from the physiological signals because it can be performed continuously in everyday life using wearables, as opposed to facial and speech emotion recognition. However, the latter modalities are also covered in this work. The abbreviations used in the article are listed in the Acronyms section at the end of the paper.

2. Sensors and Devices

Different modalities can be utilized to perform emotion recognition, but first, the appropriate sensors have to be utilized to capture these modalities.

2.1. Emotion Recognition from Physiology

Among the physiological signals, the EEG provides the highest accuracy of emotion recognition [24,25,26]. However, due to the EEG electrodes sensitivity, at this point, it is impossible to utilize EEG in the field. Therefore, researchers focused on wearable devices, which provide various biosignals and other environmental data [27]. Primarily, ECG, BVP, and EDA signals are used [5,9,12]. Those signals can be further enriched with ACC, GYRO, SKT, and RSP [2,7,8,28,29], as well as with UV, GPS, and MIC data [30].
Numerous devices for obtaining physiological signals are available on the market. Precise medical-grade devices, such as the BioPac MP160 and the ProComp Infiniti, can be used for laboratory studies. They are large, wired, and complicated, but they are also handled by a trained technician. Devices for studies in the field, however, must be small, wearable, and easy to operate by the participant. There are many types of wearables to choose from, e.g., smart rings, wristbands, smart bands, fit bands, smartwatches, armbands, chest straps, chest patches, headbands, smart glasses, smart clothes, and more.
The choice of a particular type and model depends on the study we plan to perform [31]. For instance, devices monitoring cardiac and electrodermal activity are the best choice for emotion recognition in the field. Comfortable EEG headbands can be used for meditation studies. Dedicated devices are necessary for analyzing sleep, while commercial smartwatches are sufficient for monitoring physical activity. When selecting a device for studies in the field, the following factors should be taken into consideration:
  • What physiological signals do we need to collect? Does the device provide them in a raw format?
  • What is the signal recording frequency? Is it appropriate for our problem?
  • How portable is the device? What types of physical activities does the device have to handle?
  • How are the data obtained from the device? By cable or directly transferred to the cloud? Do we need to integrate the existing study system with the device?
  • How convenient and easy to use is the device? How long should the battery last?
Table 1 provides an overview of various types of wearables, their sensors, and physiological signals and other data they provide. The author manually investigated the devices marked with *. For other devices, the official producers’ or other external information were used. Most devices included in the table were introduced over the past five years. Additionally, a few older devices, which were, or still are, frequently utilized in affective studies, are covered, e.g., Empatica E4 released in 2015 and BodyMedia SenseWear introduced in 2003.
Smartwatches seem to be the most convenient device for the participants and offer some extra features. Moreover, people already have a long history of wearing watches and, thus, are more willing to accept such devices. As for the downside of smartwatches, their HR and HRV readings are not very precise when participants are in motion [32]. More accurate measurements can be obtained with chest straps [33], which, in turn, are not very comfortable. Perhaps, a compromise would be a chest patch such as Vital Patch, although it can only run on a battery for a limited time and still requires more validation studies.
Other devices analyzed within this work were Emotiv Epoc+*, NeuroSky, emWave2, Honor Band 4*, Xiaomi Mi Band 5*, Xiaomi Mi Band 3*, Polar A370*, Fitbit Blaze*, Samsung Galaxy Fit E, Samsung Gear Live, Philips DTI-2, Sony SmartBand 2, and 20 others. They do not offer access to the data or have other disadvantages; hence, they were not included in Table 1.

2.2. Facial Emotion Recognition

Emotion recognition from facial expressions requires a camera to obtain the face image. Ideally, the face would be turned toward the camera, without any obstacles (hair, eyeglasses, and face mask), and the image would be captured in high resolution and good lighting conditions. Such criteria are easy to fulfill in a laboratory environment but very troublesome in daily life. It is possible to use a smartphone or laptop camera, but this limits the emotion recognition to situations when the subject is in front of the camera. McDuff et al. performed a field study utilizing laptop cameras to analyze facially expressed affection [34]. They recorded about 120 people for approximately eight days between 7 a.m. and 7 p.m. and found out that people were in front of the computer on average for 19% of this time. However, there are still use-cases in which ER from the face can be helpful, e.g., in online learning to ensure students are focused and engaged instead of being bored or frustrated [35].

2.3. Speech Emotion Recognition

In the case of ER from speech, the microphone is responsible for recording audio signals. Again, it is easier to obtain clean, undisturbed audio without background noise in a laboratory setup. It is also possible at home when we interact with speaker-based home assistants, such as Alexa or Google Home. Outdoor, however, the environmental sounds and other people’s utterances are most likely captured alongside the main speaker’s voice. Nevertheless, speech emotion recognition in the field might still be possible with appropriate filtering. For example, Lu et al. proposed the StressSense, a smartphone application that can detect stress in diverse real-life conversational situations, both indoor and outdoor [36].

3. Signal Processing and Transformation

The goal of signal processing is to eliminate interference and distortions from physiological signals which are not related to emotional characteristics and might bias the predictive ability of the models. The possible sources of signal artifacts are peripheral electromagnetic fields impacting electronic circuits within wearables, mixing of signals generated by different organs (e.g., the heart, brain, or muscle electrical activities), temporary deactivation or saturation of sensors, electronic noise generated inside the circuits, body motions, and movements and adjustments in sensor contact with the skin. The common approaches to reducing signal distortions are based on the known properties of the physiological signals. For example, the amplitude of the ECG signal should not exceed values 0.5–5 mV, whereas the ECG frequency spectrum is usually within 0.05–150 Hz values. On the other hand, the BVP amplitude is usually within 20–300 mmHg and the BVP frequency spectrum within 0–15 Hz. Alternatively, the data-driven approach can be applied. We can distinguish four main groups of methods for removing artifacts: (1) filtration and smoothing; (2) decomposition and removing undesirable components; (3) normalization; and (4) winsorization.
According to the Nyquist–Shannon sampling theorem, before converting the processed voltage into a digital signal, initial antialiasing filtering should be performed. The application of additional filtering (both analog and digital) and smoothing emphasize desired frequency components while reducing unwanted ones. High-pass filters, for example, eliminate slowly wandering components; low-pass filters minimize high-frequency disturbances; pass-band filters focus on specific frequency ranges; and signal-smoothing techniques prioritize low-frequency components.
Decomposition facilitates the separation of desired and undesired components. The wavelet transform (WT), which is available in continuous (CWT), discrete (DWT), and fast (FWT) variations, turns a nonstationary input onto the coefficient components of various scales (also connected to frequency ranges) by using a hand-picked family of wavelets to detect local correlations. The first component, in most cases, is a measurement noise or an electromyogram anomaly that can be eliminated afterward. Independent component analysis (ICA) is yet another method for distinguishing independent input signals (mixed while recorded by multiple sensors).
Normalization allows us to retain the energy of signals at a consistent level, especially when data from multiple devices are merged. Interpolation replaces damaged parts of the signal while accounting for the statistical features of some or all neighboring data. Lastly, winsorization reduces extreme values to reduce the effect of possibly spurious outliers.
Lee et al. showed that the application of the signal processing and transformation methods, i.e., ICA, fast Fourier transform (FFT), truncated singular value decomposition, can reduce motion artifacts from the BVP signal recorded with a wearable device, resulting in more precise HR readings, even during intense exercises [37,38]. Masinelli et al. proposed the SPARE (spectral peak recovery) algorithm for BVP signals pulse-wave reconstruction [39]. The SPARE performs signal decomposition with singular spectrum analysis, spectral estimation with sparse signal reconstruction and FFT, harmonic relationship estimation, and reconstruction by applying a narrow bandpass filter to the resulting components and summing them up. The validation performed by the authors showed a 65% improvement in the detection of different biomarkers from the BVP signal.
Signal processing and transformation can also be helpful in feature extraction. Feng et al. utilized wavelet-based features to classify emotions from the EDA signal [3]. They compared four mother wavelet candidates, i.e., Daubechies, Coiflets, Symlets, and C-Morlet and found the C-Morlet to be the best. Zhao et al. applied FFT to obtain subbands of BVP and HRV signals collected with Empatica E4 device [40]. They also used dynamic threshold difference (DTD) to obtain HRV from BVP.

4. Machine Learning Models and Techniques

Emotion recognition usually involves a massive amount of data, e.g., speech or physiological signals sampled with high-frequency, facial images (or videos) captured in a high resolution. Processing such extensive sequential data requires complex structures, adequate methods, and significant computing power. Currently, the best approach is to utilize deep neural network architectures capable of reflecting temporal dynamic behavior of affective data, thus providing rich representational possibilities. The network architectures explicitly designed to capture the temporal information of data are the recurrent neural networks (RNNs). Besides emotion recognition, RNNs are used in problems such as time-series prediction, speech recognition, machine translation, and robot control. However, in their classic design, the RNNs have gradient vanishing and exploding problems. Architectures introducing gates (gated recurrent unit—GRU, long short-term memory—LSTM) and shortcuts (residual networks—ResNet) were proposed to overcome these problems.
Another deep neural network architecture helpful in emotion recognition is convolutional neural networks (CNNs). CNNs are excellent in image vision problems; hence, they are convenient in face detection and analysis tasks and can also be applied to signals’ spectrograms. A common approach is to mix and combine several architectures to benefit from their advantages.
In order to develop machine learning models capable of emotion recognition, we need data obtained from sensors to be labeled with affective states. This is usually accomplished by asking the participants to provide a self-assessment. The labeled dataset is commonly divided into three parts: one for training the model, another one for validating/optimizing the model, and the last one for testing the performance of the model. The deep learning models may require hundreds, thousands, or tens of thousands of samples to perform well. By contrast, the affective datasets are most often sparse and contain up to 50 participants (up to a few thousand samples).
The following sections describe the most popular deep learning architectures for emotion recognition problems and several techniques for better model creation and adjustment. Additionally, some of the recent emotion recognition studies utilizing state-of-the-art deep learning architectures are summarized in Table 2 and briefly described in the text. In most cases, the performance of emotion recognition is also reported. Please note that it is provided only to give some general idea of what level of performance is achieved in the literature. The results should not be compared without carefully analyzing the details of each work. Studies on emotion recognition can be different in many ways, e.g., they may vary in: (1) the dataset used; (2) the emotional model applied; (3) the machine learning approach adopted; (4) the number of classification classes and their distribution; (5) the validation strategy (e.g., user-independent vs. user-dependent); (6) whether the results are provided for train, validation, or test set; and (7) the performance quality measure.

4.1. Residual Networks

In 2015, He et al. proposed the ResNet [58] architecture, which facilitates creating very deep neural networks without concerns about performance and the vanishing gradient issue. All thanks to shortcuts that omit one or more layers. Later, He et al. refined ResNet with a pre-activation residual block, enabling even deeper networks [59]. The ResNet takes an image as the input and includes convolution layers. It is especially helpful in image processing problems, such as face recognition and landmark detection. It can also be applied to physiological or speech data once the signal is represented as an image or its features, e.g., power spectral density or other characteristics from the time or frequency domains.
For instance, Li and Lima used ResNet-50 (containing 50 layers) to perform facial expression recognition [46]. They achieved average accuracy of 95 ± 1.4% in recognizing seven classes: happy, sad, fearful, angry, surprised, disgusted, and neutral. Efremova et al. managed to embed a 20-layer ResNet model on a mobile device to perform real-time face and emotion recognition [48]. They distinguish positive, weak positive, neutral, weak negative, and negative classes and obtained 87% of accuracy on the test set. Bargal et al. concatenated 91-layer ResNet architecture with two VGG networks [60] to perform emotion recognition from videos [50]. They argue that this provides some form of regularization. They identify six basic emotions (anger, disgust, fear, happiness, sad, and surprise) and neutral state with 57% of accuracy on the test set.
Li et al. utilized ResNet to investigate speech emotion recognition [56]. They replaced the last layer with classifiers and observed the best results with the SVM classifier (on average 90% accuracy). They used EMODB [61], a publicly available dataset that offers about 500 utterances with anger, boredom, disgust, fear, happiness, sadness, and neutral attitude. Fan et al. proposed PyResNet to classify emotions from speech [51]. The name comes from modifying the second layer of the ResNet with pyramid convolution, which should reduce the issue of uncertain time position of accurate speech information. They collected almost 150,000 utterances labeled with numerous emotions (anger, happiness, sadness, disappointment, boredom, disgust, excitement, fear, surprise, normal, and others). PyResNet based on the ResNet152 achieved 62% of weighted accuracy on the test set in speech classification. Their dataset is publicly available.
Various modifications of the ResNet architecture were proposed. Gjoreski et al. introduced StresNet [57] (originally named STRNet), which aims to capture a temporal and spectral representation of the signal. In their work, StresNet and other deep architectures (CNN and LSTM) were used to assess the possibility of monitoring driver distractions from physiological and visual signals. They obtained 75% of F-measure with StresNet on the test set and the participant independent scenario in recognizing whether distraction occurred (binary classification).

4.2. Long Short-Term Memory

To mitigate the gradient vanishing and exploding problems of the RNNs, the long short-term memory [62] (LSTM) architecture has been proposed. The LSTM cells contain forget gates that control how much information is passed on. This also allows capturing long-term temporal dependencies of the data, making LSTMs very attractive in sequential processing tasks, e.g., reasoning affective state from the physiological signals. The LSTM input layer requires three-dimensional data that refer to sequence (samples), time steps, and features.
Awais et al. applied LSTM architecture to classify four emotions (amusement, boredom, relaxation, and fear) based on physiological signals [41]. Their multimodal approach achieved results above 93% of F-measure value. They used CASE [63], a publicly available dataset with a rich spectrum of signals, i.e., ECG, BVP, GSR, RSP, SKT, and EMG. Tizzano et al. used LSTM to recognize happy, sad, and neutral affective states in subjects that were listening to music or watching a short movie [44]. They claimed to obtain 99% accuracy. The physiological data (heart rate, 3-axis acceleration, and angular velocity) were obtained with wearable devices—Samsung Gear 2 and Polar H7. Nakisa et al. focused on the LSTM hyperparameter optimization in high/low arousal/valence classification task [25]. As an input, they used EEG and BVP physiological signals recorded with wearable devices—the Emotiv Insight headset and Empatica E4 wristband. They found differential evolution to be the best optimization algorithm. On average, they accomplished 67 ± 9.3% of accuracy in 4-class classification problem. Dar et al. utilized CNN and LSTM to classify high/low arousal/valence from EEG, ECG, and EDA signals [42]. They claimed 91% and 99% accuracy on DREAMER and AMIGOS datasets, respectively. Wang et al. introduced a dual-sequence LSTM architecture that is capable of processing two sequences of data simultaneously [52]. In their case, the mel-spectrogram focused on the time axis, and the mel-spectrogram favoring frequency axis is considered to classify utterances into happy, angry, sad, and neutral states. They achieved 72.7 ± 0.7% in mean weighted accuracy.
The LSTM architecture can also be used in the bidirectional version—Bi-LSTM, which has two recurrent layers. This allows one to propagate information in two directions (forward and backward) and is helpful for sequencing sequence tasks. Moreover, the LSTM utilized as encoder–decoder is often extended with the attention mechanizm. It allows passing the information from all the encoder nodes to all the decoder nodes, in contrast to the classic scheme where only the last encoder node forwards the condensed information to the first decoder node.
Sepas-Moghaddam et al. used VGG16 and bidirectional version of the LSTM, further enhanced with the attention mechanism to recognize happy, angry, surprised, and neutral states from facial images [47]. They obtained 87.6 ± 5.4% accuracy in subject-specific and 80.3 ± 9% in subject-independent validation. Yu and Kim proposed an attention-LSTM-attention model for classifying happy, angry, sad, and neutral states from 5-second speech samples [53]. They first generated new features with the attention mechanism and then applied bidirectional LSTM, and finally perform weighted pooling by an attention mechanism. Their approach achieved 67.7 ± 3.4% in weighted accuracy. Zhao et al. developed a model that combines attention-based bidirectional LSTM with attention-based fully convolutional networks to obtain spatial-temporal features from the speech spectrogram [55]. They validated their architecture on two datasets to classify (1) angry, happy, sad, and neutral states, and (2) anger, emphatic, neutral, positive, and rest states. They achieved 67% and 49% of unweighted accuracy, respectively, and concluded that deep representations are competitive to feature representations in speech emotion recognition tasks.

4.3. Convolutional Neural Networks and Fully Convolutional Networks

The idea of using convolutional neural network (CNN) to handwritten character recognition was proposed in the 1990s [64], but only with graphics processing units (GPUs) implementation could CNN be applied more broadly and effectively. The CNN performs excellent at image classification problems due to its pattern recognition abilities. The CNN comprises (1) convolutional layers that apply filters to extract features, (2) pooling layers that reduce the dimensionality, and (3) fully connected layers that aggregate information from final feature maps.
The Fully Convolutional Network [65] (FCN) architecture was proposed for semantic segmentation in images. In contrast to CNN, FCN allows input images to have various resolutions. Additionally, the FCN does not flatten the convoluted image and instead keeps it as 1 × 1 × depth layer to proceed with deconvolution (upsampling) to the original size. The idea is similar to the encoder (convolution part) and decoder (deconvolution), where we lose some information in favor of general representation (e.g., semantic segmentation without details about segments).
In emotion recognition problems, the CNN and FCN architectures are mostly used for facial analysis in photos and videos. Once we convert the signals into spectrograms or input as a vector, they can also be used for emotion recognition from speech or physiological signals.
Cheng et al. proposed an architecture based on FCN and CNN to assess valence value from facial video recordings [49]. They designed the super-resolution FCN architecture to ensure the robustness of the model in case the video is of low quality or low bit rate, e.g., a video being streamed in real-time. In the best setup, they reported 0.121 of RMSE. Zhang et al. created attention-based FCN to classify speech into four classes, i.e., happy, sad, angry, and neutral [54]. Applying FCN allowed them to use speech fragments of various lengths without segmentation, which would have been necessary in the case of CNN. No need for segmentation also means there is no risk of losing potential information. The authors claimed that using the attention mechanism makes it easier for the model to identify the region of the speech spectrogram indicating the affective state. They accomplished 70% in weighted accuracy. Santamaria-Granados et al. utilized CNN to perform 2-level arousal and 2-level valence classification using ECG and EDA physiological signals [45]. They achieved 76% accuracy for arousal and 75% for valence. Song et al. utilized CNN to perform 3-level arousal and 3-level valence classification using EEG, ECG, EDA, Resp, and SKT physiological signals [43]. They attained 62% and 58% accuracy for arousal and valence, respectively. Kanjo et al. enhanced the CNN architecture with the temporal dimension of the data by appending the LSTM cell after the fully connected layer [30]. The classification task was a 5-level valence. As an input, they used the tensor of HR, EDA, Temp, ACC, and environmental data, collected with wristband Microsoft Band 2, in a real-life scenario. They obtained 95% of accuracy (95% F-measure, 0.291 RMSE) on the test set.

4.4. End-to-End Deep Learning Approach

In the classical feature-based approach to model creation, we manually extract features from the available data. This, however, requires expert domain knowledge to (1) preprocess signal or image (e.g., signal: wavelet/Fourier transform, spectral analysis; image: adjusting resolution, removing Gaussian noise), (2) obtain relevant features (e.g., signal: morphology, statistical, and nonlinear measures; image: facial landmarks, action units), and (3) select only valuable ones. Even then, it is possible that we do not capture all the complex dependencies and characteristics of the affective modalities. Thanks to more advanced deep neural network architectures and more powerful computers, it has become possible to create end-to-end networks that retrieve complex features from the raw signal or image themselves.
Dzieżyc et al. compared ten deep neural network architectures (ResNet, StresNet, FCN, MLP, Encoder, Time-CNN, CNN-LSTM, MLP-LSTM, MCDCNN, and Inception) in the end-to-end affect recognition from physiological signals task [66]. They found that the networks perform better when arousal and valence classes are well separated (strongly differentiating stimuli) and worse when the class boundary is blurred. The best results were achieved with FCN, ResNet, and StresNet. Schmidt et al. utilized CNN to build an end-to-end architecture [67]. They provided BVP, EDA, ACC, and SKT physiological signals as input. Their end-to-end model performed better (45.5 ± 1.8% F-measure) than the feature-based model (43.8 ± 2.0% F-measure) for high/low valence/arousal classification. Zhao et al. proposed an end-to-end visual–audio attention network (VAANet), which consists of 2D and 3D ResNets, and several CNNs with attention [68]. Their architecture can recognize multiple emotions from videos. The validation on the VideoEmotion-8 (eight classes: anger, disgust, fear, sadness, anticipation, joy, surprise, and trust) and Ekman-6 (six classes: anger, disgust, fear, sadness, joy, and surprise) benchmarks indicated state-of-the-art performance—54.5% and 55.3% accuracy, respectively. Sun et al. used residual CNN as end-to-end architecture in the speech emotion recognition task [69]. A part of the architecture is responsible for recognizing gender. Sun validated his model on three datasets with different languages (Mandarin, English, and German) and outperformed models based on the classical features, spectrograms, and other non-end-to-end deep learning solutions, achieving up to 90.3% accuracy. Harper and Southern fed the interbeat interval (IBI) signal into CNN and Bi-LSTM separately and concatenated networks’ output to classify valence into low or high [70]. They performed very little processing, i.e., extracted IBI from HR and applied z-scoring; hence, their approach can be considered end-to-end. Their architecture outperformed the feature-based model obtaining up to 90% accuracy and 88% in F-measure value.

4.5. Representation Learning

Most of the works presented in this article use the advantages of learning data representation. Usually, it is capturing/extracting features from the data. However, when learning deep networks, the representation of signals or images can have more benefits, such as improving signal quality, generating missing samples or the whole signal, and translating one signal into another.
A denoising autoencoder (DAE), facilitating encoder–decoder architecture with convolutional layers, can be used to obtain a clean uncorrupted signal from the corrupted input signal. Chiang et al. compared the noise reduction abilities of DAE, CNN, and FCN for ECG signal and found DAE to be the best [71].
The generative adversarial networks (GANs) are great for generative modeling. They consist of two neural networks that compete against each other. The generator network produces artificial samples from a random distribution, whereas the discriminator network tries to recognize which samples are real and which are artificial. Adib et al. investigated five different models from the GAN family for ECG signal generation task [72]. They compared their effectiveness in normal cardiac cycles and found that the classic GAN performed best. Sun et al. utilized LSTM and GAN networks to generate abnormal ECG signal [73].
Samyoun et al. applied representation learning to translate signals/features from the wrist device into the signals/features from the chest device [74]. They used GANs to translate homogeneous signals (EDA from the wrist into EDA from the chest), Bi-LSTM to translate heterogeneous signals (BVP from the wrist into ECG and Resp from the chest), and MLP to translate features of uncorrelated signals (all wrist features into EMG from chest features). They showed that stress detection with translated features has a competitive performance to stress detection with the original signals. This opens the possibility of using more user-friendly wrist devices instead of inconvenient chest devices in real-life affective studies.

4.6. Model Personalization

An important aspect to consider in the emotion recognition task is model generalization and personalization. On the one hand, we would like to have a predictive model that can be applied irrespectively to the participants’ demographic and physiological characteristics. This would allow recognizing emotions in people that did not provide any prior data or are interacting with the system for the first time. On the other hand, people are very different in terms of psychological and physiological elements when it comes to expressing and perceiving emotions. Hence, a personal model might perform better than the general one.
The most straightforward approach to personalization is creating a separate model for each user. However, it requires a large number of personal samples. Udovičić et al. achieved better results with the per-person model than with the general model in the arousal and valence classification from EDA and BVP signals [75]. Tian et al. clustered participants into five groups based on their personality traits [76]. Such per-group personalization was enough to improve the model’s accuracy in relation to the general model for arousal and valence classification using the EEG signal by 10.2% and 9.1%, respectively. Taylor et al. proposed a multitask learning neural network architecture that enables model personalization in the assessment of the mood, stress, and health of individuals in the next day [77]. Rather than creating a separate model for each user, the multitask ability enables personalization. They obtained up to 82.2% accuracy (0.818 AUC). Can et al. demonstrate that creating a separate model for each participant outperformed per-group personalization and general model in stress classification task [78]. They achieved a gain of up to 17 percentage points.

5. Existing Systems

The iMotions is a platform with the broadest spectrum of services targeted at affective research and analysis for scientific and industry clients [79]. For both types of clients, the platform is paid. According to the pricing plan, it is very expensive – for academics, it is ca. $3.5k per module, and a module is usually related to a single modality. The software can collect and synchronize data from many sensors, among others: EDA, EEG, ECG, EMG, and eye-tracking and facial recognition devices. The iMotions offer integration with the most popular devices on the market produced by Emotiv, Empatica, Shimmer, and Tobii. It also includes a signal processing module that can perform basic processing such as obtaining HR and HRV from ECG, detecting peaks in EDA, obtaining PSD from EEG, and others. Moreover, a mobile module is available that facilitates conducting field studies. However, there are yet no use-cases to validate this module. At the current stage, the most valuable benefits of the platform are collecting self-assessments and data from various devices and performing analyses. With more features (especially signal processing algorithms), the platform would gain attractiveness.
Gloor et al. developed the Happimeter, an Android application for monitoring emotions in daily life [80]. The system requires a smartwatch that measures HR and collects self-assessments and location data. Once an appropriate amount of data is collected, the system starts predicting the user’s mood. The Happimeter offers interesting insights, such as who or what location makes us (un)happy, active, or stressed. It also includes the social aspect—we can build a network of friends and monitor their mood. The application was developed for Wear OS smartwatches and is publicly available. The advantage of the Happimeter is that it creates a personalized model to perform predictions. On the other hand, the HR is collected infrequently (the exact frequency is not specified) and as a simple averaged value instead of HRV, BVP, or other continuous, rich signal. As a result, the Happimeter is more likely to recognize the overall long-term mood rather than the specific emotions experienced at a specific moment. Moreover, it can only measure three levels of mood, activity, and stress. The researchers give a little information about the performed machine learning—feature extraction and model creation and validation. The system was utilized by the creators in several studies, e.g., to assess employee satisfaction (happiness, activity, and stress) [81], a person’s creativity [82], and even personal moral values [83]. However, in most of these studies, the authors took part in the study as subjects, which can lead to bias and can question the quality of the results [84].
Tripathi et al. proposed an EmoWare framework for personalized video recommendations [85]. They utilized the reinforcement learning approach to extract users’ individual behavior and a deep-bidirectional RNN to predict relevant movies. The system has not been validated in real life. In general, most proposed systems do not leave the laboratory walls, i.e., they are not validated in daily life. There are also cases when the field validation is performed improperly. For instance, Fortune et al. [86,87] claimed that facial- and EEG-based emotion recognition of employees in real-time is possible. In fact, in three points of the day (before the work, in the middle of the day, and after the work), the authors showed short affective movies to the employees and measured their EEG and EDA at those moments. Later, reactions to the stimuli were classified into positive vs. negative states. Hence, the authors recognize emotions in response to the consumed stimuli. The employees’ emotions were not connected with the work-related conditions or environment, or type of job they perform. The authors’ claims of recognizing emotions at work were exaggerated, as the context of the work is barely relevant here.
Hernandez et al. looked over some of the recent emotion recognition commercial applications in market research, human resources, transportation, defense and security, education, and mental health areas [88]. They discussed several issues of existing systems, e.g., (1) systems describe their functionality insufficiently, often hiding the technical limitations; (2) a limited amount of data exists that can be used to train models; (3) users may over-rely on the systems, which can lead to the physical or emotional injury; (4) privacy, freedom, and other human rights may be abused by using/applying such systems; (5) and more generally, emotions do not have agreed definition and are difficult to label. The authors propose the ethical guidelines that should help mitigate the recognized issues.
Huge corporations offering services to millions of people noticed the importance of including affective aspects in their products. For example, Netflix and Spotify offer search engines that consider how the content affects the emotional state of the user [89,90]. The next step is probably to automate such solutions, e.g., by creating personalized playlists tailored to our affective state at a given moment [91].

6. Futures of Emotion Recognition

Emotion recognition studies are most likely to be continued in three main directions, i.e., ER from facial expressions, speech, and physiology, because each approach has its unique advantages and applications. The camera-based ER from the face is comfortable and accurate enough when sitting in front of a computer, although it can be deceived by controlling facial expressions. The ER from speech is an ideal solution for voice assistants and human–robot communication that is soon a reality. The personalized models could be stored locally on the devices, thus improving the privacy and security of such a solution. The ER from physiology has the most significant potential to become the most popular and broadly applied approach. Comfortable and user-friendly wearables monitoring our vital functions in almost any condition are available and already pervasive in our life. At the same time, physiology is the most difficult to manipulate; hence, it can be considered ground truth for other approaches. Another advantage of physiology is the possibility of noncontact measurement, e.g., a camera analyzing the change in the skin texture or color [92,93] or radio-frequency analysis for breath and heart rate detection [94].
Several challenges need to be tackled to enable precise and reliable affective reasoning in daily life. Currently, the most significant imperfection of sensors measuring physiology is the inaccuracy of measurement in motion. An unreliable measurement undermines any further effort to recognize emotions during movement. These imperfections can be partially reduced with advanced signal processing and transformation methods. However, such a fundamental issue should be fixed at the sensor level, i.e., new ways of measurements prone to motion have to be developed. Simultaneously, a universally accepted protocol of a new device validation against medical-level hardware should be proposed.
In addition, wearable devices should become even smaller, even more convenient, and even more useful for users while offering long battery life, convenient data transfer, and reliability. Particularly prominent directions seem to be smart clothes, smart prosthesis [95], and other everyday items that can be made smart. Until then, most probably, smartwatches will be the primary source of physiological data and the foundation of affective field studies. They are cheap, useful for users, have many sensors (BVP, ACC, and GYRO), allow for integration with systems through a custom-made application and, above all, are already omnipresent. Their only disadvantage is the lack of an EDA sensor, which can be easily fixed.
In the signal processing and transformation domain, more algorithms for artifacts and outliers removal and reconstruction of signals recorded in motion are necessary. Especially promising are the wavelet-based methods that extract affective patterns from the signals.
In the machine learning field, new deep learning architectures customized for affective studies can be developed (such as in the case of the StresNet). Such architectures would preferably reflect complex relations across emotions and their personal character. While very little has been investigated so far, a multilabel approach to emotion classification could potentially reflect the natural coexistence of emotions in real life. Further research on model generalization and personalization is also necessary. Particularly, new strategies for combining the general underlying human patterns with the significant individual differences are required.
Large and high-quality data sets must be collected to develop better predictive models. This, in turn, requires simple and accurate methods to annotate the signal with the affective label during everyday life. The current solutions include developing and validating short, yet rich in information, self-assessments that may be completed on the smartwatch or smartphone, such as ESM [96] or EMA [97]. Another option is to assist the participant with a pretrained model that recognizes states of arousal or intense emotions [27,98].
The ethical part of affective studies is a couple of years behind the technological advancements and still insufficiently discussed. We need more analysis, debate, and legal actions to ensure that affective technology is safe for the people and is not abused.

7. Conclusions

In recent years, we have experienced a number of breakthroughs in sensors and machine learning domains that make it possible to move the affective studies out of the lab into real life. Although there are still challenges to overcome, the first systems are already being used in everyday life. Further improvements, however, are necessary to enable the development of more precise and reliable systems.
This paper discusses the current state of the art in sensors, signal processing and transformation, and deep machine learning areas. Commercially available wearables and other devices used to capture traces of human emotions are presented. The role of signal processing and transformation in emotion recognition is explained with examples. The most prominent deep learning architectures and various techniques to improve model efficiency are extensively discussed. Finally, the research directions with the highest opportunity to improve emotion recognition are suggested.

Funding

This work was partially supported by the National Science Centre, Poland, project no. 2020/37/B/ST6/03806; by the statutory funds of the Department of Artificial Intelligence, Wroclaw University of Science and Technology; and by the Polish Ministry of Education and Science–the CLARIN-PL Project.

Acknowledgments

The author would like to thank members of the Emognition research group for their contribution to this article.

Conflicts of Interest

The author declares no conflict of interest.

Acronyms

The list of acronyms used in this article:
Signals, Sensors, and Data From Wearables
AcronymFull Name
ACCaccelerometer
ALambient light
ALTaltimeter
ATambient temperature
BARbarometer
BPblood pressure
BVPblood volume pulse
CALcalories burned
ECGelectrocardiogra(ph/m)
EEGelectroencephalogra(ph/m)
EMGelectromyogra(ph/m)
GSRgalvanic skin response
GYROgyroscope
HRheart rate
HRVheart rate variability
IBIinterbit interval
MAGmagnetometer
MICmicrophone
PPGphotoplethysmograph
PPIpeak-to-peak interval
RRIR-R interval
RSPrespiration rate
SKTskin temperature
SpO2blood oxygen saturation
STPno. of steps
TERMtermometer
UVultraviolet
Deep Learning Architectures and Signal Processing
AcronymFull Name
Bi-bidirectional-
CNNconvolutional neural networks
DAEdenoising autoencoder
DTDdynamic threshold difference
FCNfully convolutional networks
FFAfast Fourier transform
GANgenerative adversarial networks
GRUgated recurrent unit
ICAindependent component analysis
LSTMlong short-term memory
MCDCNNmultichannel deep convolutional neural networks
MLPmultilayer perceptron
ResNetresidual networks
RNNrecurrent neural networks
SPAREspectral peak recovery
StresNetspectrotemporal residual networks
VGGvisual geometry group
WTwavelet transform

References

  1. O’Brien, D.T. Thinking, Fast and Slow; Farrar, Straus and Giroux: New York, NY, USA, 2012. [Google Scholar]
  2. He, C.; Yao, Y.J.; Ye, X.S. An emotion recognition system based on physiological signals obtained by wearable sensors. In Wearable Sensors and Robots; Springer: Berlin, Germany, 2017; pp. 15–25. [Google Scholar]
  3. Feng, H.; Golshan, H.M.; Mahoor, M.H. A wavelet-based approach to emotion classification using EDA signals. Expert Syst. Appl. 2018, 112, 77–86. [Google Scholar] [CrossRef]
  4. Pollreisz, D.; TaheriNejad, N. A simple algorithm for emotion recognition, using physiological signals of a smart watch. In Proceedings of the 39th Annual International Conference of the Ieee Engineering in Medicine and Biology Society (EMBC), Jeju Island, Korea, 11–15 July 2017; pp. 2353–2356. [Google Scholar]
  5. Fernández-Aguilar, L.; Martínez-Rodrigo, A.; Moncho-Bogani, J.; Fernández-Caballero, A.; Latorre, J.M. Emotion detection in aging adults through continuous monitoring of electro-dermal activity and heart-rate variability. In Proceedings of the International Work-Conference on the Interplay Between Natural and Artificial Computation, Tenerife, Spain, 3–7 June 2019; pp. 252–261. [Google Scholar]
  6. Hu, L.; Yang, J.; Chen, M.; Qian, Y.; Rodrigues, J.J. SCAI-SVSC: Smart clothing for effective interaction with a sustainable vital sign collection. Futur. Gener. Comput. Syst. 2018, 86, 329–338. [Google Scholar] [CrossRef]
  7. Albraikan, A.; Hafidh, B.; El Saddik, A. iAware: A real-time emotional biofeedback system based on physiological signals. IEEE Access 2018, 6, 78780–78789. [Google Scholar] [CrossRef]
  8. Schmidt, P.; Reiss, A.; Duerichen, R.; Marberger, C.; Van Laerhoven, K. Introducing wesad, a multimodal dataset for wearable stress and affect detection. In Proceedings of the 20th ACM International Conference on Multimodal Interaction, Boulder, CO, USA, 16–20 October 2018; pp. 400–408. [Google Scholar]
  9. Setiawan, F.; Khowaja, S.A.; Prabono, A.G.; Yahya, B.N.; Lee, S.L. A framework for real time emotion recognition based on human ans using pervasive device. In Proceedings of the 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; pp. 805–806. [Google Scholar]
  10. Álvarez, P.; Zarazaga-Soria, F.J.; Baldassarri, S. Mobile music recommendations for runners based on location and emotions: The DJ-Running system. Pervasive Mob. Comput. 2020, 67, 101242. [Google Scholar] [CrossRef]
  11. Tkalčič, M. Emotions and personality in recommender systems: Tutorial. In Proceedings of the 12th ACM Conference on Recommender Systems, Vancouver, BC, Canada, 2 October 2018; pp. 535–536. [Google Scholar]
  12. Nalepa, G.J.; Kutt, K.; Giżycka, B.; Jemioło, P.; Bobek, S. Analysis and use of the emotional context with wearable devices for games and intelligent assistants. Sensors 2019, 19, 2509. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Xu, T.; Yin, R.; Shu, L.; Xu, X. Emotion recognition using frontal EEG in VR affective scenes. In Proceedings of the MTT-S International Microwave Biomedical Conference (IMBioC), Nanjing, China, 6–8 May 2019; pp. 1–4. [Google Scholar]
  14. Gross, J.J.; Feldman Barrett, L. Emotion generation and emotion regulation: One or two depends on your point of view. Emot. Rev. 2011, 3, 8–16. [Google Scholar] [CrossRef] [PubMed]
  15. Damasio, A.R. The Feeling of What Happens: Body and Emotion in the Making of Consciousness; Mariner Books: Boston, MA, USA, 1999. [Google Scholar]
  16. Ekman, P. Universals and cultural differences in facial expressions of emotions. W: J. Cole (red.). Nebr. Symp. Motiv. 1971, 19, 207–283. [Google Scholar]
  17. Plutchik, R. A general psychoevolutionary theory of emotion. In Theories of Emotion; Elsevier: Amsterdam, The Netherlands, 1980; pp. 3–33. [Google Scholar]
  18. Frijda, N.H. The Emotions; Cambridge University Press: Cambridge, UK, 1986. [Google Scholar]
  19. Gross, J.J. Emotion regulation: Current status and future prospects. Psychol. Sci. 2015, 26, 1–26. [Google Scholar] [CrossRef]
  20. Barrett, L.F. The future of psychology: Connecting mind to brain. Perspect. Psychol. Sci. 2009, 4, 326–339. [Google Scholar] [CrossRef] [Green Version]
  21. James, W. The emotions. Nature 1922, 110, 730–731. [Google Scholar]
  22. Averill, J.R. A constructivist view of emotion. In Theories of Emotion; Elsevier: Amsterdam, The Netherlands, 1980; pp. 305–339. [Google Scholar]
  23. Mesquita, B.; Boiger, M. Emotions in context: A sociodynamic model of emotions. Emo. Rev. 2014, 6, 298–302. [Google Scholar] [CrossRef] [Green Version]
  24. Chen, T.; Yin, H.; Yuan, X.; Gu, Y.; Ren, F.; Sun, X. Emotion recognition based on fusion of long short-term memory networks and SVMs. Digit. Signal Process. 2021, 117, 103153. [Google Scholar] [CrossRef]
  25. Nakisa, B.; Rastgoo, M.N.; Rakotonirainy, A.; Maire, F.; Chandran, V. Long short term memory hyperparameter optimization for a neural network based emotion recognition framework. IEEE Access 2018, 6, 49325–49338. [Google Scholar] [CrossRef]
  26. Soroush, M.; Maghooli, K.; Setarehdan, S.; Nasrabadi, A. A Review on EEG signals based emotion recognition. Int. Clin. Neurosc. J. 2017, 4, 118. [Google Scholar] [CrossRef]
  27. Saganowski, S.; Dutkowiak, A.; Dziadek, A.; Dzieżyc, M.; Komoszyńska, J.; Michalska, W.; Polak, A.; Ujma, M.; Kazienko, P. Emotion recognition using wearables: A systematic literature review-work-in-progress. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; pp. 1–6. [Google Scholar]
  28. Schmidt, P.; Reiss, A.; Dürichen, R.; Van Laerhoven, K. Labelling Affective States in the Wild: Practical Guidelines and Lessons Learned. In Proceedings of the 2018 ACM International Joint Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing and Wearable Computers, Singapore, 8–12 October 2018; pp. 654–659. [Google Scholar]
  29. Maier, M.; Marouane, C.; Elsner, D. DeepFlow: Detecting Optimal User Experience From Physiological Data Using Deep Neural Networks. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 2108–2110. [Google Scholar]
  30. Kanjo, E.; Younis, E.M.; Ang, C.S. Deep learning analysis of mobile physiological, environmental and location sensor data for emotion detection. Inf. Fusion. 2019, 49, 46–56. [Google Scholar] [CrossRef]
  31. Saganowski, S.; Kazienko, P.; Dzieżyc, M.; Jakimów, P.; Komoszyńska, J.; Michalska, W.; Dutkowiak, A.; Polak, A.; Dziadek, A.; Ujma, M. Consumer Wearables and Affective Computing for Wellbeing Support. In Proceedings of the 17th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, Darmstadt, Germany, 7–9 December 2020. [Google Scholar]
  32. Bent, B.; Goldstein, B.A.; Kibbe, W.A.; Dunn, J.P. Investigating sources of inaccuracy in wearable optical heart rate sensors. NPJ Digit. Med. 2020, 3, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Speer, K.E.; Semple, S.; Naumovski, N.; McKune, A.J. Measuring heart rate variability using commercially available devices in healthy children: A validity and reliability study. EJIHPE 2020, 10, 390–404. [Google Scholar] [CrossRef] [Green Version]
  34. McDuff, D.; Jun, E.; Rowan, K.; Czerwinski, M. Longitudinal Observational Evidence of the Impact of Emotion Regulation Strategies on Affective Expression. IEEE Trans. Affect. Comput. 2019, 12, 636–647. [Google Scholar] [CrossRef]
  35. Mukhopadhyay, M.; Pal, S.; Nayyar, A.; Pramanik, P.K.D.; Dasgupta, N.; Choudhury, P. Facial emotion detection to assess Learner’s State of mind in an online learning system. In Proceedings of the 5th International Conference on Intelligent Information Technology, Hanoi, Vietnam, 13–22 February 2020; pp. 107–115. [Google Scholar]
  36. Lu, H.; Frauendorfer, D.; Rabbi, M.; Mast, M.S.; Chittaranjan, G.T.; Campbell, A.T.; Gatica-Perez, D.; Choudhury, T. Stresssense: Detecting stress in unconstrained acoustic environments using smartphones. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing, Pittsburgh, PA, USA, 5–8 September 2012; pp. 351–360. [Google Scholar]
  37. Lee, J.; Kim, M.; Park, H.K.; Kim, I.Y. Motion artifact reduction in wearable photoplethysmography based on multi-channel sensors with multiple wavelengths. Sensors 2020, 20, 1493. [Google Scholar] [CrossRef] [Green Version]
  38. Lee, H.; Chung, H.; Ko, H.; Lee, J. Wearable multichannel photoplethysmography framework for heart rate monitoring during intensive exercise. IEEE Sens. J. 2018, 18, 2983–2993. [Google Scholar] [CrossRef]
  39. Masinelli, G.; Dell’Agnola, F.; Valdés, A.A.; Atienza, D. SPARE: A spectral peak recovery algorithm for PPG signals pulsewave reconstruction in multimodal wearable devices. Sensors 2021, 21, 2725. [Google Scholar] [CrossRef]
  40. Zhao, B.; Wang, Z.; Yu, Z.; Guo, B. EmotionSense: Emotion recognition based on wearable wristband. In Proceedings of the Symposia and Workshops on Ubiquitous, Autonomic and Trusted Computing, UIC-ATC, Guangzhou, China, 8–12 October 2018; pp. 346–355. [Google Scholar]
  41. Awais, M.; Raza, M.; Singh, N.; Bashir, K.; Manzoor, U.; ul Islam, S.; Rodrigues, J.J. LSTM based Emotion Detection using Physiological Signals: IoT framework for Healthcare and Distance Learning in COVID-19. IEEE Internet Things J. 2020, 8, 16863–16871. [Google Scholar] [CrossRef]
  42. Dar, M.N.; Akram, M.U.; Khawaja, S.G.; Pujari, A.N. Cnn and lstm-based emotion charting using physiological signals. Sensors 2020, 20, 4551. [Google Scholar] [CrossRef] [PubMed]
  43. Song, T.; Lu, G.; Yan, J. Emotion recognition based on physiological signals using convolution neural networks. In Proceedings of the 12th International Conference on Machine Learning and Computing, Shenzhen, China, 19–21 June 2020; pp. 161–165. [Google Scholar]
  44. Tizzano, G.R.; Spezialetti, M.; Rossi, S. A Deep Learning Approach for Mood Recognition from Wearable Data. In Proceedings of the IEEE International Symposium on Medical Measurements and Applications (MeMeA), Bari, Italy, 1 June–1 July 2020; pp. 1–5. [Google Scholar]
  45. Santamaria-Granados, L.; Munoz-Organero, M.; Ramirez-Gonzalez, G.; Abdulhay, E.; Arunkumar, N. Using deep convolutional neural network for emotion detection on a physiological signals dataset (AMIGOS). IEEE Access 2018, 7, 57–67. [Google Scholar] [CrossRef]
  46. Li, B.; Lima, D. Facial expression recognition via ResNet-50. Int. J. Cogn. Comput. Eng. 2021, 2, 57–64. [Google Scholar] [CrossRef]
  47. Sepas-Moghaddam, A.; Etemad, A.; Pereira, F.; Correia, P.L. Facial emotion recognition using light field images with deep attention-based bidirectional LSTM. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3367–3371. [Google Scholar]
  48. Efremova, N.; Patkin, M.; Sokolov, D. Face and emotion recognition with neural networks on mobile devices: Practical implementation on different platforms. In Proceedings of the 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; pp. 1–5. [Google Scholar]
  49. Cheng, B.; Wang, Z.; Zhang, Z.; Li, Z.; Liu, D.; Yang, J.; Huang, S.; Huang, T.S. Robust emotion recognition from low quality and low bit rate video: A deep learning approach. In Proceedings of the Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017; pp. 65–70. [Google Scholar]
  50. Bargal, S.A.; Barsoum, E.; Ferrer, C.C.; Zhang, C. Emotion recognition in the wild from videos using images. In Proceedings of the 18th ACM International Conference on Multimodal Interaction, Tokyo, Japan, 12–16 November 2016; pp. 433–436. [Google Scholar]
  51. Fan, W.; Xu, X.; Xing, X.; Chen, W.; Huang, D. LSSED: A large-scale dataset and benchmark for speech emotion recognition. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 641–645. [Google Scholar]
  52. Wang, J.; Xue, M.; Culhane, R.; Diao, E.; Ding, J.; Tarokh, V. Speech emotion recognition with dual-sequence LSTM architecture. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 6474–6478. [Google Scholar]
  53. Yu, Y.; Kim, Y.J. Attention-LSTM-attention model for speech emotion recognition and analysis of IEMOCAP database. Electronics 2020, 9, 713. [Google Scholar] [CrossRef]
  54. Zhang, Y.; Du, J.; Wang, Z.; Zhang, J.; Tu, Y. Attention based fully convolutional network for speech emotion recognition. In Proceedings of the Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Honolulu, HI, USA, 12–15 November 2018; pp. 1771–1775. [Google Scholar]
  55. Zhao, Z.; Bao, Z.; Zhao, Y.; Zhang, Z.; Cummins, N.; Ren, Z.; Schuller, B. Exploring deep spectrum representations via attention-based recurrent and convolutional neural networks for speech emotion recognition. IEEE Access 2019, 7, 97515–97525. [Google Scholar] [CrossRef]
  56. Li, Z.; Li, J.; Ma, S.; Ren, H. Speech emotion recognition based on residual neural network with different classifiers. In Proceedings of the IEEE/ACIS 18th International Conference on Computer and Information Science (ICIS), Beijing, China, 17–19 June 2019; pp. 186–190. [Google Scholar]
  57. Gjoreski, M.; Gams, M.Ž.; Luštrek, M.; Genc, P.; Garbas, J.U.; Hassan, T. Machine learning and end-to-end deep learning for monitoring driver distractions from physiological and visual signals. IEEE Access 2020, 8, 70590–70603. [Google Scholar] [CrossRef]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 18–20 June 2016; pp. 770–778. [Google Scholar]
  59. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  60. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  61. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.F.; Weiss, B. A Database of German Emotional Speech. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.130.8506&rep=rep1&type=pdf (accessed on 1 January 2022).
  62. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  63. Sharma, K.; Castellini, C.; van den Broek, E.L.; Albu-Schaeffer, A.; Schwenker, F. A dataset of continuous affect annotations and physiological signals for emotion analysis. Sci. Data 2019, 6, 1–13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  65. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  66. Dzieżyc, M.; Gjoreski, M.; Kazienko, P.; Saganowski, S.; Gams, M. Can we ditch feature engineering? end-to-end deep learning for affect recognition from physiological sensor data. Sensors 2020, 20, 6535. [Google Scholar] [CrossRef] [PubMed]
  67. Schmidt, P.; Dürichen, R.; Reiss, A.; Van Laerhoven, K.; Plötz, T. Multi-target affect detection in the wild: An exploratory study. Proceedings of The 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing, London, UK, 9–13 September 2019; pp. 211–219. [Google Scholar]
  68. Zhao, S.; Ma, Y.; Gu, Y.; Yang, J.; Xing, T.; Xu, P.; Hu, R.; Chai, H.; Keutzer, K. An End-to-End visual-audio attention network for emotion recognition in user-generated videos. arXiv 2020, arXiv:2003.00832. [Google Scholar] [CrossRef]
  69. Sun, T.W. End-to-end speech emotion recognition with gender information. IEEE Access 2020, 8, 152423–152438. [Google Scholar] [CrossRef]
  70. Harper, R.; Southern, J. A bayesian deep learning framework for end-to-end prediction of emotion from heartbeat. arXiv 2020, arXiv:1902.03043. [Google Scholar] [CrossRef] [Green Version]
  71. Chiang, H.T.; Hsieh, Y.Y.; Fu, S.W.; Hung, K.H.; Tsao, Y.; Chien, S.Y. Noise reduction in ECG signals using fully convolutional denoising autoencoders. IEEE Access 2019, 7, 60806–60813. [Google Scholar] [CrossRef]
  72. Adib, E.; Afghah, F.; Prevost, J.J. Synthetic ECG Signal Generation Using Generative Neural Networks. arXiv 2021, arXiv:2112.03268. [Google Scholar]
  73. Sun, H.; Zhang, F.; Zhang, Y. An LSTM and GAN Based ECG Abnormal Signal Generator. In Advances in Artificial Intelligence and Applied Cognitive Computing; Springer: Berlin, Germany, 2021; pp. 743–755. [Google Scholar]
  74. Samyoun, S.; Mondol, A.S.; Stankovic, J.A. Stress detection via sensor translation. In Proceedings of the 16th International Conference on Distributed Computing in Sensor Systems (DCOSS), Marina del Rey, CA, USA, 25–27 May 2020; pp. 19–26. [Google Scholar]
  75. Udovičić, G.; Ðerek, J.; Russo, M.; Sikora, M. Wearable emotion recognition system based on GSR and PPG signals. In Proceedings of the 2nd International Workshop on Multimedia for Personal Health and Health Care, New York, NY, USA, 23 October 2017; pp. 53–59. [Google Scholar]
  76. Tian, Z.; Huang, D.; Zhou, S.; Zhao, Z.; Jiang, D. Personality first in emotion: A deep neural network based on electroencephalogram channel attention for cross-subject emotion recognition. Royal Soc. Open Sci. 2021, 8, 201976. [Google Scholar] [CrossRef]
  77. Taylor, S.; Jaques, N.; Nosakhare, E.; Sano, A.; Picard, R. Personalized multitask learning for predicting tomorrow’s mood, stress, and health. IEEE Trans. Affect. Comput. 2017, 11, 200–213. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  78. Can, Y.S.; Chalabianloo, N.; Ekiz, D.; Fernandez-Alvarez, J.; Riva, G.; Ersoy, C. Personal stress-level clustering and decision-level smoothing to enhance the performance of ambulatory stress detection with smartwatches. IEEE Access 2020, 8, 38146–38163. [Google Scholar] [CrossRef]
  79. iMotions Platform. Available online: https://imotions.com (accessed on 1 January 2022).
  80. Gloor, P.A.; Colladon, A.F.; Grippa, F.; Budner, P.; Eirich, J. Aristotle said “happiness is a state of activity”—Predicting mood through body sensing with Smartwatches. J. Syst. Sci. Syst. Eng. 2018, 27, 586–612. [Google Scholar] [CrossRef]
  81. Roessler, J.; Gloor, P.A. Measuring happiness increases happiness. JCSS 2021, 4, 123–146. [Google Scholar] [CrossRef] [Green Version]
  82. Sun, L.; Gloor, P.A.; Stein, M.; Eirich, J.; Wen, Q. No Pain No Gain: Predicting Creativity Through Body Signals. In Digital Transformation of Collaboration; Springer: Berlin, Germany, 2019; pp. 3–15. [Google Scholar]
  83. Sun, L.; Gloor, P.A. Measuring Moral Values with Smartwatch-Based Body Sensors. In Digital Transformation of Collaboration; Springer: Berlin, Germany, 2019; pp. 51–66. [Google Scholar]
  84. Budner, P.; Eirich, J.; Gloor, P.A. Making you happy makes me happy-Measuring Individual Mood with Smartwatches. Available online: http://healthdocbox.com/Psychology_and_Psychiatry/68305189-Making-you-happy-makes-me-happy-measuring-individual-mood-with-smartwatches.html (accessed on 1 January 2022).
  85. Tripathi, A.; Ashwin, T.; Guddeti, R.M.R. EmoWare: A context-aware framework for personalized video recommendation using affective video sequences. IEEE Access 2019, 7, 51185–51200. [Google Scholar] [CrossRef]
  86. Fortune, E.; Yusuf, Y.; Zornes, S.; Loyo Lopez, J.; Blocker, R. Assessing Induced Emotions in Employees in a Workplace Setting Using Wearable Devices. In Proceedings of the 2020 Design of Medical Devices Conference, Minneapolis, MA, USA, 6–9 April 2020; p. V001T09A004. [Google Scholar]
  87. Fortune, E.; Yusuf, Y.; Blocker, R. Measuring Arousal and Emotion in Healthcare Employees Using Novel Devices. In Proceedings of the IEEE 20th International Conference on Bioinformatics and Bioengineering (BIBE), Cincinnati, OH, USA, 26–28 October 2020; pp. 835–838. [Google Scholar]
  88. Hernandez, J.; Lovejoy, J.; McDuff, D.; Suh, J.; O’Brien, T.; Sethumadhavan, A.; Greene, G.; Picard, R.; Czerwinski, M. Guidelines for Assessing and Minimizing Risks of Emotion Recognition Applications. In Proceedings of the 9th International Conference on Affective Computing and Intelligent Interaction (ACII), Nara, Japan, 28 September–1 October 2021; pp. 1–8. [Google Scholar]
  89. Emotional Movies-Netflix. Available online: https://ir.netflix.net/ir-overview/profile/default.aspx (accessed on 22 January 2022).
  90. Panda, R.; Redinho, H.; Gonçalves, C.; Malheiro, R.; Paiva, R.P. How Does the Spotify API Compare to the Music Emotion Recognition State-of-the-Art? In Proceedings of the 18th Sound and Music Computing Conference (SMC 2021), Axea sas/SMC Network. 29 June–1 July 2021; pp. 238–245. [Google Scholar] [CrossRef]
  91. Álvarez, P.; de Quirós, J.G.; Baldassarri, S. A Web System Based on Spotify for the automatic generation of affective playlists. In Cloud Computing, Big Data & Emerging Topics; Springer: Berlin, Germany, 2020; pp. 124–137. [Google Scholar]
  92. McDuff, D.; Nishidate, I.; Nakano, K.; Haneishi, H.; Aoki, Y.; Tanabe, C.; Niizeki, K.; Aizu, Y. Non-contact imaging of peripheral hemodynamics during cognitive and psychological stressors. Sci. Rep. 2020, 10, 10884. [Google Scholar] [CrossRef]
  93. Uchida, M.; Akaho, R.; Ogawa-Ochiai, K.; Tsumura, N. Image-based measurement of changes to skin texture using piloerection for emotion estimation. Artif. Life Robot. 2019, 24, 12–18. [Google Scholar] [CrossRef]
  94. Zhao, M.; Adib, F.; Katabi, D. Emotion recognition using wireless signals. In Proceedings of the 22nd Annual International Conference on Mobile Computing and Networking, New York, NY, USA, 3–7 October 2016; pp. 95–108. [Google Scholar]
  95. Richter-Lunn, K. Incognito: Sensorial Interpretations of Covert Physiological Signals for Therapeutic Mediation. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 2021. [Google Scholar]
  96. Myin-Germeys, I.; Kuppens, P. The Open Handbook of Experience Sampling Methodology. In The Open Handbook of Experience Sampling Methodology; Independently Publisher: Chicago, IL, USA, 2021; pp. 1–311. [Google Scholar]
  97. Smyth, J.M.; Stone, A.A. Ecological momentary assessment research in behavioral medicine. J. Happiness Stud. 2003, 4, 35–52. [Google Scholar] [CrossRef]
  98. Saganowski, S.; Behnke, M.; Komoszyńska, J.; Kunc, D.; Perz, B.; Kazienko, P. A system for collecting emotionally annotated physiological signals in daily life using wearables. In Proceedings of the 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW), Nara, Japan, 28 September–1 October 2021; pp. 1–3. [Google Scholar]
Figure 1. (a) The interdisciplinary background of emotion recognition; (b)modalities used to recognize emotions.
Figure 1. (a) The interdisciplinary background of emotion recognition; (b)modalities used to recognize emotions.
Electronics 11 00496 g001
Table 1. The recent commercially available wearable devices, their sensors measuring physiological signals, and the data they provide; (*) marks wearables tested by the author.
Table 1. The recent commercially available wearable devices, their sensors measuring physiological signals, and the data they provide; (*) marks wearables tested by the author.
DeviceTypeRelease DateSensorsPhysiological Raw SignalsOther Data
Apple Watch 7SmartwatchOctober 2021ACC, AL, ALT, BAR, ECG, GPS, GYRO, MIC, PPGBVP, ECG, SpO2ACC, AL, ALT, BAR, BP, GPS, GYRO, HR, MIC, RSP, SKT, STP
Fitbit charge 5SmartbandSeptember 2021ACC, BAR, GPS, GYRO, PPG-ACC, BAR, GYRO, HR
Samsung Galaxy Watch 4 *SmartwatchAugust 2021ACC, AL, BAR, ECG, GPS, GYRO, MIC, PPGBVPACC, AL, BAR, GPS, GYRO, HR, MIC, PPI, STP
Huawei Watch 3SmartwatchJune 2021ACC, AL, BAR, GPS, GYRO, MAG, MIC, PPG, TERM-ACC, AL, BAR, GPS, GYRO, HR, MAG, MIC, SKT, STP
Fitbit SenseSmartwatchSeptember 2020ACC, AL, ALT, GPS, GYRO, MIC, PPG, TERM-ACC, BAR, GYRO, HR
Samsung Galaxy Watch 3 *SmartwatchAugust 2020ACC, AL, BAR, ECG, GPS, GYRO, MIC, PPGBVPACC, AL, BAR, GPS, GYRO, HR, MIC, PPI, STP
Apple Watch 5 *SmartwatchSeptember 2019ACC, BAR, ECG, GPS, GYRO, MIC, PPG-ACC, BAR, CAL, GPS, GYRO, HR, MIC, STP
Fossil Gen 5 *SmartwatchAugust 2019ACC, AL, ALT, GPS, GYRO, MIC, PPGBVPACC, AL, ALT, GPS, GYRO, HR, MIC, STP
Garmin Fenix 6X ProSmartwatchAugust 2019ACC, AL, ALT, GPS, GYRO, PPG, SpO2BVP, SpO2ACC, AL, ALT, GPS, GYRO, HR, STP
Samsung Galaxy Watch *SmartwatchAugust 2019ACC, AL, BAR, GPS, GYRO, MIC, PPGBVPACC, AL, BAR, GPS, GYRO, HR, MIC, STP
Polar OH1ArmbandMarch 2019ACC, PPGBVPACC, PPI
Garmin HRM-DUALChest strapJanuary 2019ECGECGRRI
Muse 2 *HeadbandJanuary 2019ACC, EEG, GYRO, PPG, SpO2BVP, EEG, SpO2ACC, GYRO, HR
Fitbit Charge 3 *FitbandOctober 2018ACC, ALT, GYRO, PPG-ACC, ALT, HR
Garmin VivoActive 3 Music *SmartwatchJune 2018ACC, BAR, GPS, GYRO, PPG-ACC, CAL, HR, PPI, RSP, STP
Oura ring *Smart ringApril 2018ACC, GYRO, PPG, TERM-HR, PPI, SKT
Moodmetric *Smart ringDecember 2017ACC, EDAEDASTP
DREEMHeadbandJune 2017ACC, EEG, PPG, SpO2BVP, EEG, SpO2ACC, HR
AidlabChest strapMarch 2017ACC, ECG, RSP, TERMECG, RSP, SKTActivities, HR, HRV, STP
Polar H10Chest strapMarch 2017ACC, ECGECGACC, RRI
VitalPatchChest patchMarch 2016ACC, ECG, TERMECG, SKTHR, RRI, RSP, STP
Emotiv InsightHeadbandOctober 2015ACC, EEG, GYRO, MAGEEGACC, GYRO, MAG
Empatica E4*Wristband2015ACC, EDA, PPG, TERMBVP, EDA, SKTACC, HR, PPI
Microsoft Band 2SmartbandOctober 2014ACC, AL, ALT, BAR, EDA, GYRO, PPG, TERM, UVBVP, EDA, SKTACC, AL, ALT, BAR, CAL, GYRO, HR, PPI, STP, UV
BodyMedia SenseWearArmband2003ACC, EDA, TERMEDA, SKTACC
Table 2. The recent emotion recognition studies utilizing state-of-the-art deep learning architectures.
Table 2. The recent emotion recognition studies utilizing state-of-the-art deep learning architectures.
ModalityReferenceArchitectureClassification/Regression Problem Considered
Phys. signalsAwais et al., 2021 [41]LSTM4-class: amused, bored, feared, relaxed
Dar et al., 2020 [42]CNN + LSTM4-class: high/low arousal/valence
Song et al., 2020 [43]CNN2 × 3-class: calm/medium/excited arousal; unpleasant/neutral/pleasant valence
Tizzano et al., 2020 [44]LSTM3-class: happy, neutral, sad
Kanjo et al., 2018 [30]CNN + LSTM5-class: level of valence (from 0 to 4)
Nakisa et al., 2018 [25]LSTM4-class: high/low arousal/valence
Santamaria-Granados et al., 2018 [45]CNN2 × binary: high/low arousal/valence
Facial expressionLi and Lima, 2021 [46]ResNet-507-class: angry, disgusted, fearful, neutral, happy, sad, surprised
Sepas-Moghaddam et al., 2020 [47]VGG16 + Bi-LSTM + attention4-class: angry, happy, neutral, surprised
Efremova et al., 2019 [48]ResNet-205-class: positive, weak positive, neutral, weak negative, negative
Cheng et al., 2017 [49]FCN + CNNregression: valence value
Bargal et al., 2016 [50]ResNet-91 with 2 × VGG7-class: angry, disgusted, fearful, neutral, happy, sad, surprised
SpeechFan et al., 2021 [51]PyResNet: ResNet with pyramid convolution4-class: angry, neutral, happy, sad
Wang et al., 2020 [52]dual-sequence LSTM4-class: angry, neutral, happy, sad
Yu and Kim, 2020 [53]attention-LSTM-attention4-class: angry, neutral, happy, sad
Zhang et al., 2019 [54]FCN-attention4-class: angry, neutral, happy, sad
Zhao et al., 2019 [55]attention-Bi-LSTM + attention-FCN4-class: angry, neutral, happy, sad; 5-class: angry, emphatic, neutral, positive, resting
Li et al., 2019 [56]ResNet7-class: angry, bored, disgusted, fearful, neutral, happy, sad
Visual + phys. signalsGjoreski et al., 2020 [57]StresNet, CNN, LSTMbinary (driver distraction)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Saganowski, S. Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning. Electronics 2022, 11, 496. https://doi.org/10.3390/electronics11030496

AMA Style

Saganowski S. Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning. Electronics. 2022; 11(3):496. https://doi.org/10.3390/electronics11030496

Chicago/Turabian Style

Saganowski, Stanisław. 2022. "Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning" Electronics 11, no. 3: 496. https://doi.org/10.3390/electronics11030496

APA Style

Saganowski, S. (2022). Bringing Emotion Recognition Out of the Lab into Real Life: Recent Advances in Sensors and Machine Learning. Electronics, 11(3), 496. https://doi.org/10.3390/electronics11030496

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop