Next Article in Journal
A Robust Laser Stripe Extraction Method for Structured-Light Vision Sensing
Next Article in Special Issue
Experimental Verification of Objective Visual Fatigue Measurement Based on Accurate Pupil Detection of Infrared Eye Image and Multi-Feature Analysis
Previous Article in Journal
Process Parameters for FFF 3D-Printed Conductors for Applications in Sensors
Previous Article in Special Issue
Sparse Spatiotemporal Descriptor for Micro-Expression Recognition Using Enhanced Local Cube Binary Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Emotion Variation from Controlling Contrast of Visual Contents through EEG-Based Deep Emotion Recognition

1
Division of Software Convergence, Sangmyung University, Seoul 03016, Korea
2
Department of Computer Science, Sangmyung University, Seoul 03016, Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(16), 4543; https://doi.org/10.3390/s20164543
Submission received: 29 June 2020 / Revised: 4 August 2020 / Accepted: 6 August 2020 / Published: 13 August 2020
(This article belongs to the Special Issue Sensor Applications on Emotion Recognition)

Abstract

:
Visual contents such as movies and animation evoke various human emotions. We examine an argument that the emotion from the visual contents may vary according to the contrast control of the scenes contained in the contents. We sample three emotions including positive, neutral and negative to prove our argument. We also sample several scenes of these emotions from visual contents and control the contrast of the scenes. We manipulate the contrast of the scenes and measure the change of valence and arousal from human participants who watch the contents using a deep emotion recognition module based on electroencephalography (EEG) signals. As a result, we conclude that the enhancement of contrast induces the increase of valence, while the reduction of contrast induces the decrease. Meanwhile, the contrast control affects arousal on a very minute scale.

Graphical Abstract

1. Introduction

We watch and appreciate various visual contents such as movies, dramas, animation etc. The scenes from the contents evoke diverse emotional responses from us. The emotions from the scene can be categorized through Russell’s emotion model such as excitement, happy, pleased, peaceful, calm, gloomy, sad, fear and suspense [1]. The story line of visual contents is regarded as one of the most important factors that evoke a specific emotional response. For example, the battle scene between the fisher and sharks from an animation “The old man and the sea” evokes suspense. The scene of the settled forks from an animation “The man who plant trees” evokes peaceful or calm. We regard the story line of visual contents as latent factor of an emotional response, since it is the most important and major factor that affects the emotional responses. In addition to the story line, many researchers have questioned whether the representation of visual contents such as colors or contrasts can affect the emotional responses. Since color of a visual contents such as color of a cloth is very hard to change, we concentrate on the contrast of a scene, which affects the whole scene of the contents.
The purpose of our research is to examine how the emotional responses from visual contents are affected from the change of the contrast of a scene. Enhancing or reducing the contrast is expected to increase or decrease the emotional responses, respectively. In many applications, positive emotional responses such as happy or peaceful are magnified to delight the audiences further. Negative responses such as sad or fear are required to get reduced to relieve the audiences who feel uncomfortable for the scenes.
The assumption of this study that the emotional responses can be affected from the change of the contrast is backgrounded from biological observation. The lower or higher contrast can affect the number of photons arriving at the retina of an eye. Since the difference of the intensities of the photons is reduced, the brain reflecting on the neurons of retina can produce different biosignals, since the different biosignals can be interpreted as the change of valence and arousal. The relationship between the contrast and emotional responses should be examined in a very careful approach.
To reveal the relationship between the contrast and emotional response, we collect some scenes from various visual contents of three representative emotions: neutral, positive and negative. Then, we control the contrast of the scenes in both directions, either enhancing or reducing. For each scene manipulated, we employ human participants to appreciate the contents, while their EEG biosignals are captured and processed using a deep learning-based emotion recognition model. We prove that the contrast change of a visual contents affect the arousal or valence estimated from the emotion recognition model. The overview of our framework is illustrated in Figure 1.
The estimation of the relations between contrast and valance/arousal by participants’ annotation is a subjective approach. For a reliable estimation of the relation, a series of objective approaches that employ the biosignals estimated from the participants. Many researchers employ diverse biosignals including EEG, electrocardiography (ECG), electromyography (EMG), photoplethysmogram (PPG), respiration rate (RSP) and galvanic skin response (GSR), etc. Since people cannot control their own biosignals, the biosignal-based approaches for estimating valence and arousal gather higher confidence than user survey-based approaches.
Furthermore, we carefully experiment the change of emotion recognition for the visual contents and compare the magnitude of emotion change due to the contrast control. From this experiment, we can suggest how to control visual contents to strengthen or weaken the emotion of the contents.
The change of emotion is expressed by the value of valence and arousal, which are measured through EEG biosignal. Actually, many researchers have presented a series of models that measure valence and arousal from EEG signal. Among these works, some of them employ classical machine learning schemes including Bayesian [2], dual tree-complex wavelet packet transform (DT-CWPT ) [3], decision fusion [4], support vector machine (SVM) [5,6], hidden Markov Model (HMM) [7] and brain computer interface (BCI) [8]. Some recent models employ deep learning schemes such as long short term memory (LSTM) recurrent neural network (RNN) [9], convolutional recurrent neural network (CRNN) [10], 3D CNN [11], LSTM [12] and multi-columned CNN [13]. The models using conventional machine learning schemes show about 70% accuracy, while the recent deep learning-based models show about 80% accuracy.
This paper is organized as follows. In Section 2, we briefly review several works for deep learning-based emotion recognition model and relationship between chroma and human emotion. In Section 3, we present outlines of our experiment including how sample movie clips are selected from visual contents and contrast of these clips are controlled. We explain our emotion recognition model and experiment strategy in Section 4, and present analysis and limitation of our work in Section 5. Finally, we draw conclusion of our work and suggest the direction of future work in Section 6.

2. Related Work

In this section, we describe previous studies around deep learning, emotion recognition approaches and emotional responses from visual contents.

2.1. Deep Learning-Based Emotion Recognition Models

2.1.1. Early Models

Researchers employed handcrafted features to apply them to a variety of machine learning algorithms such as support vector machine (SVM) and principal component analysis (PCA). They showed acceptable accuracy, but the performance of their framework heavily relied on feature selection.
Many researchers have employed deep learning-based approach which does not require delicate feature selection process. Jirayucharoensak et al. [14] proposed a model of recognizing emotion from unstationary EEG signals. They used stacked autoencoder (SAE) to learn features from EEG signals. To minimize unstationary effect, they exploited PCA that extracts the most important component and covariate shift adaptation, which is also effective to avoid overfitting.
Khosrowabadi et al. [15] presented a multi-layer feedforward network model to distinguish emotion from unstationary EEG signals. The network model consists of spectral filtering layer for analyzing input signals and shift register memory layers. They identified emotion and employed arousal and valence level to represent the emotion according to Russell’s model.

2.1.2. Deep Learning-Based Models

The early frameworks for emotion recognition have a common limitation in selecting features by human experts’ hands. As the progress of convolutional neural network (CNN), many researchers have overcome such a limitation. Salama et al. [11] presented 3D CNN model to recognize emotion from multi-channel structured EEG signals. Since EEG signals are spatio-temporal, their 3D CNN is a quite proper structure to be fed the signals. They also showed that data augmentation phase is helpful to improve the performance of emotion recognition and avoid overfitting. They tested the proposed model on the DEAP dataset and showed 88.49% and 87.44% accuracy with regard to arousal and valence, respectively.
Moon et al. [16] proposed a CNN-based model for recognizing emotion from EEG signals. Their main distinguishing point is connectivity features which is useful to describe synchronous activation in many different regions of brain. Therefore, they effectively acquired a variety of asymmetric brain activity patterns which is a key role to recognize emotion. They were not solely dependent on CNN structure, but showed that the partial feature selection contributes to improving the CNN-based approaches.
Croce et al. [17] classified brain and artifactual independent components by employing CNN-based model. Although, their research goal was not emotion recognition, but independent component analysis is essential to conventional feature selection approach. They conducted heuristic selection of machinery hyperparameters and CNN-based self-selection of the interesting features. They showed acceptable performance with cross validation 92.4% for EEG.
Yang et al. [13] proposed multi-channel structured CNN model to recognize emotion from unstationary EEG signals. Their model consisted of several independent recognizing modules, where they designed each module based on DenseNet [18]. For the final decision, they merged the results from the modules. They compared their performance with the existing studies and showed better accuracy for valence and arousal. They also extended their emotion recognizer to apply the task of distinguishing emotional responses to photographs and artwork [19].
Zhang et al. [20] presented an effective spatial attention map (SAM) for weighing multi-hierarchical convolutional features. SAM is useful for suppressing filter values corresponding to background features. They also proposed multi-model adaptive response fusion (MAF) approach which is for adaptive weighted fusion of multiple response maps generated by attentional feature.
Recurrent neural network (RNN) is widely used for processing time series data such as writing. Since EEG signals are time-sequential, many researchers have employed RNN to recognize emotion from the signals. Li et al. [21] proposed emotion recognizer based on RNN structure. They focused on three types of properties of EEG signals: frequency, spatial and temporal. To interpret the EEG signals in frequency and spatial domain, they made extraction of rational asymmetry (RASM). They extracted temporal correlation through LSTM RNN structure. They tested the proposed model by employing DEAP dataset and showed 76.67% mean accuracy.
Xing et al. [12] developed a model for recognizing emotion by employing multi-channel EEG. Their framework consists of an emotion time model and a linear EEG mixing model. The framework separate source of EEG signals from collected EEG signals, which improve the performance of classifying. The technical background of EEG mixing model is based on emotion timing model and stacked autoencoder (SAE). They tested their model using DEAP dataset and show the 81.10% and 74.38% accuracy for valence and arousal, respectively.
Several researchers employed conventional feature selection model to improve deep learning-based model. Some researchers often applied deep learning-based model to the other machine learning framework, other researchers used several different deep learning-based models.
Yoo et al. [22] presented a neural network-based recognizer for six types of emotional states including happy, joy, fear, anger, sadness and despair. They tested their proposed model for a variety of multi-modal bio-signals including EEG, PPG, GSR and ECG.
Yang et al. [23] proposed hybrid neural network-based model by combining RNN and CNN for recognizing emotion. They employed baseline signals in pre-processing phase, which improved the performance of recognition. Such pre-processing method resembles conventional feature-based techniques. They also had the proposed model to learn the way of representing unstationary EEG signals for recognizing emotion. In their hybrid model, the RNN module makes an extraction of contextual information from signals and the CNN module measures inter-channel correlations among EEG signals.

2.1.3. Relationship Between Chroma and Human Emotion

In early days, Adams and Osgood [24] presented a classic research on the affective meaning of color, which pioneers to reveal the relationship between color and affections.
Recently, Suk et al. [25] researched characteristics of emotional responses to hue, lightness and chroma. In this study, subjects are asked for assessing their emotional response against various color stimulus. The result shows chroma level and lightness significantly affect valence and arousal. Although the result support that color stimulus could incur emotional response, the study relies on subjective self assessment whereas our research utilizes objective EEG signal.
Jun et al. [26] performed a study to find out how can we modify an image’s color characteristics in order to enhance its emotional impact. Measurements are performed to how emotional responses are changed due to different color and contrast, also regarding contexts of the image. They make use of skin conductance and heartbeat rate variation to quantify emotion. In the result, color and contrast clearly affect emotional response, but aspect of the emotional impact is different amongst context of the image. The study confirms our priori that emotional response is affected by colors and gives future study improvements.
Rajae-Joordens [27] insisted the effect of hue, lightness and saturation of colored light upon human arousal and valence is significant. In the research, both objective and subjective measurements for emotional response under various colored lights are performed. Although the result proves features of color seem to affect human emotion, discrepancy between results from objective and subjective measurements also explains why there are different aspect of change in emotional responses between various studies.
With recent advances in deep learning, various image transfer techniques are introduced. Peng et al. [28] studied how emotional responses are being changed by transferring image to another. In general, image transfer techniques lead in changed color tone and texture. A quantitative evaluation is performed to find out differences in emotions from against original image to image transferred one. The result shows image transference, in terms of color and texture, evokes some change in emotional response, albeit causal analysis against color or contrast are not discussed.

3. Contrast Change of Visual Contents

3.1. Sampling Clips of Three Representative Emotions

We selected three well-known animations for our study: Loving Vincent [29], The old man and the sea [30], and A man who plants trees [31]. These animations are recognized as distinctive rendering styles. We selected the artistically styled contents, since we concentrated on the color variation of the contents. Since the color variation of artistic styled image was narrower than that of real photograph, the effect of contrast control could be more drastic for artistic styled images. Therefore, the effect of contrast control on artistically styled images for emotion variation could be more dominant than that on real photographs. We will extend the target visual contents to real photographs in future work.
From the selected visual contents, we sampled three different scenes, each of which contain positive, neutral and negative emotion, respectively. For this purpose, we sampled 10 scenes from each of the contents and hired 10 human participants to mark the emotion of the scene for the three categories: positive, neural, and negative. We also asked the participants to suggest the background of their selection. From the votes of the participants, we selected the scenes of highest votes for our experiment. The emotions from the sampled clips for three emotions are illustrated in Figure 2, which were estimated from the model we describe in Section 4. The emotions from the clips were grouped into three different emotions: positive, neutral and negative. From Figure 2, we concluded that the emotions marked by 10 human participants on the clips were correct.

3.1.1. Scenes of Positive Emotion

From Loving Vincent, the scene where Armand talks to Marguerite, the daughter of Dr. Gachet, about good memories of Vincent van Gogh was selected as the scene of most positive emotion. The major reason of the selection is that this scene that reminds of a positive memory about Vincent evokes positive emotion.
From The old man and the sea, the scene of the fantasy that a boy is playing with fishes in the sea was selected. The major reason of the selection is that playing with fishes in the sea is a fantasy from childhoods.
From A man who plants trees, the scene of happy people living in a village filled with a lot of trees was selected. The major reason is that this last scene of the animation shows how the effort of a man can make people happy and a village flourish.
The scenes of positive emotion are illustrated in Figure 3.

3.1.2. Scenes of Neutral Emotion

Most of the participants selected the scenes that happened early in the contents as scenes of neutral emotion. Since the early scenes introduced the actors of the contents and explain their situations, the emotional response from these scenes tended to be neutral.
From Loving Vincent, the scene where Armand is visiting Auvers was selected. They selected this scene since visiting Auvers is an action that does not evoke any emotion responses.
From The old man and the sea, the opening scene where a boy visits the old man and talks about his dream was selected. Similar to the Loving Vincent scene, talking about a dream does not evoke either positive or negative emotions.
From A man who plants trees, the scene where the narrator encounters Elzeard Bouffier and visits his house was selected. Similar to the above contents, visiting a house does not evoke emotional responses.
The scenes of neutral emotion are illustrated in Figure 4.

3.1.3. Scenes of Negative Emotion

Most of the participants selected scenes of conflict as the scenes of negative emotion. The existence of unpleasant objects such as blood accelerated the negative emotion.
From Loving Vincent, the scene where Louise Chevalier, the maid of Dr. Gachet speaks ill of Vincent van Gogh was selected as a scene of negative emotion. Since most of the participants had good feelings about Vincent, they felt unpleasant and negative about the conversation between Louise and Armand.
From The old man and the sea, the scene of bloody struggle between sharks that came to hunt the tuna caught by the old man was selected as negative. The blood spread on the sea added the negative emotion.
From A man who plants trees, the scene that describes the process of falling down in a village due to the harsh weather was selected as a scene of negative emotion.
The scenes of negative emotion are illustrated in Figure 5.

3.2. Controlling the Contrast of the Scenes

We applied linear contrast control scheme, which has been widely used in many image processing applications. The reason is that both contrast enhancement and contrast reduction were required. Since we did not want to affect color factor in the contrast control, we converted the color in RGB space into HSV space and applied contrast control for the S (saturation) and V (value). After the control, we re-converted the color in HSV space into RGB space.
A contrast controlled value of saturation, denoted as S , was estimated as follows:
S = S + ( S S a v g ) α ,
where S a v g is the average value of the saturation, and α is a control parameter. We set α as a negative value for contrast reduction. We control the contrast of the scenes from the selected visual contents. The result of the control is shown in Appendix A. We set α = 0.1 for enhancement and α = 0.1 for reduction.

4. Implementation and Results

4.1. Implementation Detail

We implemented our model in a personal computer with Intel Core i7 9600 CPU (Intel, San Jose, CA, USA), 16 GB main memory, and nVidia GeForce TitanX GPU (nVidia, Santa Clara, CA, USA). The model is implemented using Python with Pytorch library. Our multi-column structured emotion recognition model was derived from our previous research [13]. The structure of our emotion recognition model is illustrated in Figure 6. In this work, we build our recognition model with five recognizing modules, since Yang et al. revealed that five modules showed the best performance in recognizing emotion.
Hence the model was trained with a DEAP dataset [32], we provided a similar setup with the DEAP. We used LiveAmp 32 and LiveCap [33], which allowed us to set up 32 channels following a standard 10/20 system [34].

4.2. Data Collection

We followed the data collection strategy in [13]. For a training, we downsampled the EEG signals of DEAP dataset to 128 Hz and applied a 4.0–4.5 Hz band pass filter. As a result, we sampled 128 × 60 samples from each trial for 40 channels. Among them, we excluded eight channels for the normalization. 32 channels of EEG signal became one row of our 2D input. Therefore, we sampled 32 consecutive samples from a single input for our model. Figure 7 illustrates the construction of a single input data. For a test, we similarly built 32 × 32 input data using the EEG signals captured from participants.

4.3. Deep Emotion Recognition Model

As we described before, three scenes were selected from each animation. We assumed that contrast would more or less affects emotional response by the mood of the scene. 30 participants were asked to watch each of the positive, neutral, and negative scenes from an animation, then move to next animation after 30 s of rest. After nine playbacks, another nine playbacks were followed with increased contrast, then last nine with decreased contrast. Each playthrough, consisting of 27 playbacks, was preceded by 3 s of baseline recording. All recordings were started/stopped by human control, allowing less than 1 s of error.

4.4. Model Training

To streamline our data to DEAP dataset, similar preprocessing is taken. We used EEGLab with MATLAB to preprocess those recordings alongside the channel location file. The data were downsampled to 128 Hz with a bandpass frequency filter from 4.0–45.0 Hz was applied. The order of the channels was same as that of DEAP. The data were segmented into 27 trials. Our emotion recognition model got preprocessed data and produced estimation upon valence and arousal level for each trial.
Our model was trained using the DEAP dataset according to the strategy presented by [13]. For training and validation, DEAP dataset, which was composed of the EEG signals captured from 32 participants, was segmented into three subsets: 22 participants for training, five for validation, and five for testing. Since each participant executed 40 experiments, the experiment data in the datasets for training, validation and test were 880, 200, and 200, respectively. Furthermore, we sampled 32 consecutive values from equally-spaced different positions of an experiment data whose distance was 7680 / ( 32 k ) . Note that k is the number of modules of our model. We set k as 5 in this study. In total, the number of training, validation and test samples for our model was 42,240, 9600 and 9600.
In many related literatures, various models have been presented to estimate valence and arousal from EEG biosignal. They employ either classic machine learning techniques [2,3,4,5,6,7,8] or deep learning techniques [9,10,11,12,13]. According to [13], the machine learning technique-based schemes present 71.66% accuracy for valence and 69.37% for arousal in average, and the deep learning-based schemes present 81.4% and 80.5%, respectively. These research trends present a strong background for us to employ a deep learning-based model for estimating valence and arousal using the EEG signals captured from users. We present the precision, recall and F1 score of our model in Table 1, which was estimated in [13].

4.5. Experiment

We hired 30 human participants for our experiments and group them into three groups. The distribution of gender and age of the participants for each group are suggested in Table 2. Each group was required to watch three movie clips. We organized the movie clips into nine categories as illustrated in Figure 8. The groups represented by three colors such as green, blue and yellow were assigned to movie clips of different emotions.
A participant of a group was guided to watch three movie clips that belong to the same category. The order of three movie clips was shuffled randomly. Each clip was played for 60∼90 s. After watching three movie clips that belonged to one category, the participants had a day break to neutralize their emotional responses. The schedule of our experiment is illustrated in Figure 9.
Their emotional responses expressed by their EEG signals were processed through our multi-stage emotion recognition module to extract valence and arousal. We present the valence and arousal extracted from our experiment in Figure 10, Figure 11 and Figure 12, respectively.

5. Analysis

5.1. Analysis 1: The Comparison of Groundtruth Emotions

In the first analysis, we compared the valence and arousal of the original three emotions suggested in Figure 10, Figure 11 and Figure 12 through t-test. As illustrated in Table 3, p values for the valence of the pair of neutral and negative and neutral and positive from the three contents were less than 0.01, which denotes that the difference of valence was significant. Meanwhile, p values for the arousals were all greater than 0.01, which denotes that the difference of valence was not significant. From the t-test, we concluded that the movie clips we selected from the three contents showed distinctive valence and similar arousal.

5.2. Analysis 2: The Difference of Emotions from the Control of Contrast Affects Valence

In the second analysis, we compared the change of valence and arousal due to the control of contrast. As illustrated in Figure 13, we recognized that the enhancement of contrast (up) increased the valence of the contents and the reduction of contrast (down) decreased the valence for all the three status of the emotion (neutral, positive and negative). Even though the amount of increase or decrease of the valence varied according to the content, the increase or decrease happened for all the contents.
On the contrary, the arousal did not show a specific change from the control of contrast. As illustrated in Figure 14, the control of contrast produced very minute change of arousal. Therefore, we conclude that the control of contrast affected the valence only. Furthermore, the enhancement of contrast increased valence and the reduction decreased valence. We executed t-test for the valence and arousal suggested in Figure 13 and Figure 14 and demonstrated that the valence was different significantly and that the arousal was similar. The result of the t-test is suggested in Table 4.

5.3. Analysis 3: Comparison of Enhanced Emotions to the Original Emotions

In this analysis, we compared the enhanced valence from the control of contrast to the valence of original contents. In Figure 15, the enhanced valence and reduced valence from the contents of originally neutral state were compared with the valence from the original positive and the original negative, respectively. Figure 15b shows the comparison for three contents and Figure 15c shows the comparison for the averaged valence from the three contents. In the graphs, we concluded that the enhanced or reduced amount of the valence did not reach the valence of original contents. For example, the valence from the originally neutral emotion was enhanced due to the contrast up, while the amount of the increased valence did not reach the valence of originally positive emotion.
The same result was observed for the originally positive and negative emotion. In Figure 16, we compared the reduced emotion due to contrast down from originally positive emotion to the neutral emotion. Even though the valence was reduced from the positive emotion, the valence was still greater than the valence of neutral emotion.
Similarly, in Figure 17, the emotion due to contrast up from originally negative emotion was enhanced, which was recognized in the increase of valence. In this case, the valence enhanced from the negative emotion was still less than the valence of neutral emotion.

5.4. Analysis 4: Mostly Changed and Leastly Changed Contents

We compared the contents whose valence changed in the greatest scale in Figure 18 and in the least scale in Figure 19. In Figure 18, we present three scenes whose emotions changed in the greatest scale. As visually recognized, these contents contained landscape scenes with natural lights. Since these scenes contained many components of various color and tone, the increase of the contrast enriched the colorful expression of the scene, which enhanced the valence of the original contents. The decrease of the contrast, on the contrary, blunted the colorful expression of the scene, which reduced the valence of the original contents.
The scenes in Figure 19 were affected in the least scale from the control of contrast. Since these scenes were either fantastically illuminated scenes or indoor scenes, the colors in the components of the scenes were not diverse. These monotonous colors were less affected by the control of contrast, which resulted in the least enhanced or reduced valences from the original emotions.

5.5. Analysis 5: Existing Study

Since most other works considered mono-color as stimuli, it was hard to directly compare our result with other models. Jun et al. [26] pointed out that contents and contextual information may affect emotional response alongside with color itself, therefore we carefully chose certain scenes with emotional backgrounds. These scenes contained various colors and context, then raised leads in mixed emotional response. Our model was to make “scene-wise” change in contrast, rather than changing specific colors.

5.6. Limitation

An unexpected result of our contrast control scheme is that the intensity of the image also varies. Enhancing the contrast may increase the average intensity of the image and reducing may reduce the average intensity. In Figure 20, we sampled an image from three contents and estimate the change of contrast and average intensity. The change of contrast was measured by sorting the pixels according to their intensities and subtracting the average intensity of highest 10% pixels from that of darkest 10% pixels. The results are illustrated in the rightmost column of Figure 20.
Even though the average intensity of an image varied, we still argued that the increase or decrease of valence comes from the change of contrast rather than the average intensity in the following two points. The first point is that the magnitude of the change of contrast is greater than the magnitude of the change of average intensity. Since the change of contrast shows greater magnitudes than the change of average intensity, we argued that the change of valence is affected by the change of contrast rather than the change of average intensity.
Our second point is from existing literatures. Even though brightness is known to one of the cause of emotional response, the study of Wilms and Oberfeld [35] indicates that brightness is relatively weak driver of the change of emotional response. It causes limited increase for arousal and valence only for zero to lowly saturated colors. Therefore, we could assume that brightness is not the major driver of observed effect.
Our third point is to employ user-annotated valence-arousal values. Considering user-annotated valence-arousal values with the valence-arousal values estimated from EEG signals through emotion recognition model can improve the confidence on the performance of our approach. However, our experiment on participants collected only their EEG signals. Our next experiment on estimating human emotion using EEG signals will collect user annotated valence-arousal values as well as EEG biosignals and compare them to improve the confidence of our experiment.

6. Conclusions and Future Work

In this paper, we have proved that the contrast control of visual contents affect the valence of the people who watch the contents. The enhancement of contrast increases valence, while the reduction decreases. Arousal is not affected by the change of contrast. We have proved our argument by extracting EEG biosignals from human participants and by recognizing valence and arousal using a deep emotion recognition model.
We have a plan to extend this study to review the relations between other properties of contents such as color with the emotion of visual contents. We also have a plan to apply the result of this study in manipulating visual contents to satisfy users.

Author Contributions

Conceptualization, H.Y., J.H., and K.M; methodology, H.Y.; software, J.H.; validation, J.H.;formal analysis, J.H.;investigation, K.M., resources, J.H.; data curation, J.H.; writing–original draft preparation, H.Y.; writing–review and editing, K.M.; visualization, K.M.; supervision, K.M.; project administration, K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The contrast-controlled scenes from the selected visual contents.
Figure A1. The contrast-controlled positive scenes.
Figure A1. The contrast-controlled positive scenes.
Sensors 20 04543 g0a1
Figure A2. The contrast-controlled neutral scenes.
Figure A2. The contrast-controlled neutral scenes.
Sensors 20 04543 g0a2
Figure A3. The contrast-controlled negative scenes.
Figure A3. The contrast-controlled negative scenes.
Sensors 20 04543 g0a3

References

  1. Russell, J. Evidence for a three-factor theory of emotions. J. Res. Personal. 1977, 11, 273–294. [Google Scholar] [CrossRef]
  2. Yoon, H.J.; Chung, S.Y. EEG-based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm. Comput. Biol. Med. 2013, 43, 2230–2237. [Google Scholar] [CrossRef]
  3. Naser, D.S.; Saha, G. Recognition of emotions induced by music videos using DT-CWPT. In Proceedings of the Indian Conference on Medical Informatics and Telemedicine (ICMIT), Kharagpur, India, 28–30 March 2013; pp. 53–57. [Google Scholar]
  4. Rozgic, V.; Vitaladevuni, S.N.; Prasad, R. Robust the EEG emotion classification using segment level decision fusion. In Proceedings of the IEEE Conference of Acoustics, Speech, and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 1286–1290. [Google Scholar]
  5. Zhuang, X.; Rozgic, V.; Crystal, M. Compact unsupervised EEG response representation for emotion recognition. In Proceedings of the IEEE-EMBS International Conference on Biomedical and Health Informatics, Valencia, Spain, 1–4 June 2014; pp. 736–739. [Google Scholar]
  6. Arnau-Gonzalez, P.; Arevalillo-Herrez, M.; Ramzan, N. Fusing highly dimensional energy and connectivity features to identify affective states from EEG signals. Neurocomputing 2017, 244, 81–89. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, J.; Hu, B.; Xu, L.; Moore, P.; Su, Y. Feature-level fusion of multimodal physiological signals for emotion recognition. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, Washington, DC, USA, 9–12 November 2015; pp. 395–399. [Google Scholar]
  8. Atkinson, J.; Campos, D. Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers. Expert Syst. Appl. 2016, 47, 35–41. [Google Scholar] [CrossRef]
  9. Alhagry, S.; Fahmy, A.A.; El-Khoribi, R.A. Emotion recognition based on EEG using LSTM recurrent neural network. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 355–358. [Google Scholar] [CrossRef] [Green Version]
  10. Li, X.; Song, D.; Zhang, P.; Yu, G.; Hou, Y.; Hu, B. Emotion recognition from multi-channel EEG data through convolutional recurrent neural network. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, Shenzhen, China, 15–18 December 2016; pp. 352–359. [Google Scholar]
  11. Salama, E.S.; El-Khoribi, R.A.; Shoman, M.E.; Shalaby, M.A.E. EEG-based emotion recognition using 3D convolutional neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 329–337. [Google Scholar] [CrossRef]
  12. Xing, X.; Li, Z.; Xu, T.; Shu, L.; Hu, B.; Xu, X. SAE+LSTM: A New framework for emotion recognition from multi-channel EEG. Front. Nuerorobotics 2019, 13, 37. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, H.; Han, J.; Min, K. A Multi-Column CNN Model for Emotion Recognition from EEG Signals. Sensors 2019, 19, 4736. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Jirayucharoensak, S.; Pan-Ngum, S.; Israsena, P. EEG-based emotion recognition using deep learning network with principal component based covariate shift adaptation. Sci. World J. 2014, 2014, 627892. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Khosrowabadi, R.; Chai, Q.; Kai, K.A.; Wahab, A. ERNN: A biologically inspired feedforward neural network to discriminate emotion from EEG signal. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 609–620. [Google Scholar] [CrossRef] [PubMed]
  16. Moon, S.-E.; Jang, S.; Lee, J.-S. Convolutional neural network approach for EEG-based emotion recognition using brain connectivity and its spatial information. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2556–2560. [Google Scholar]
  17. Croce, P.; Zappasodi, F.; Marzetti, L.; Merla, A.; Pizzella, V.; Chiarelli, A.M. Deep Convolutional Neural Networks for Feature-Less Automatic Classification of Independent Components in Multi-Channel Electrophysiological Brain Recordings. IEEE Trans. Biom. Eng. 2019, 66, 2372–2380. [Google Scholar] [CrossRef] [PubMed]
  18. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  19. Yang, H.; Han, J.; Min, K. Distinguishing emotional responses to photographs and artwork using a deep learning-based approach. Sensors 2019, 19, 5533. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Zhang, T.; Zheng, W.; Cui, Z.; Zong, Y.; Li, Y. Spatial-temporal recurrent neural network for emotion recognition. IEEE Trans. Cybern. 2019, 49, 839–847. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Li, Z.; Tian, X.; Shu, L.; Xu, X.; Hu, B. Emotion Recognition from EEG Using RASM and LSTM. Commun. Comput. Inf. Sci. 2018, 819, 310–318. [Google Scholar]
  22. Yoo, G.; Seo, S.; Hong, S.; Kim, H. Emotion extraction based on multi-bio-signal using back-propagation neural network. Multimed. Tools Appl. 2018, 77, 4925–4937. [Google Scholar] [CrossRef]
  23. Yang, Y.; Wu, Q.; Qiu, M.; Wang, Y.; Chen, X. Emotion recognition from multi-channel EEG through parallel convolutional recurrent neural network. In Proceedings of the International Joint Conference on Neural Networks, Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  24. Adams, F.; Osgood, C. A cross-cultural study of the affective meaning of color. J. Cross-Cult. Psychol. 1973, 4, 135–156. [Google Scholar] [CrossRef]
  25. Suk, H.-J.; Irtel, H. Emotional response to color across media. Color Res. Appl. 2010, 35, 64–77. [Google Scholar] [CrossRef]
  26. Jun, J.; Ou, L.-C.; Oicherman, B.; Wei, S.; Luo, M.; Nachilieli, H.; Staelin, C. Psychophysical and psychophysiological measurement of image emotion. In Proceedings of the 18th Color and Imaging Conference, San Antonio, TX, USA, 8–12 November 2010; pp. 121–127. [Google Scholar]
  27. Rajae-Joordens, R. The Effects of Colored Light on Valence and Arousal. In Sensing Emotions; Westerink, J., Krans, M., Ouwerkerk, A., Eds.; Springer: Dordrecht, The Netherlands, 2010; Volume 12, pp. 65–84. [Google Scholar]
  28. Peng, K.-C.; Chen, T.; Sadovnik, A.; Gallagher, A. A Mixed Bag of Emotions: Model, Predict, and Transfer Emotion Distributions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 860–868. [Google Scholar]
  29. Kobiela, D.; Welchman, H. Loving Vincent. 2017. Available online: https://en.wikipedia.org/wiki/Loving_Vincent (accessed on 14 June 2020).
  30. Petrov, A. The Old Man and the Sea. 1999. Available online: https://en.wikipedia.org/wiki/The_Old_Man_and_the_Sea_(1999_film) (accessed on 26 May 2020).
  31. Back, F. The Man Who Plants Trees. 1987. Available online: https://en.wikipedia.org/wiki/The_Man_Who_Planted_Trees_(film) (accessed on 16 June 2020).
  32. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  33. BCI+: LiveAmp. Compact Wireless Amplifier for Mobile EEG Applications. BCI+ Solutions by Brain Products. Available online: bci.plus/liveamp/ (accessed on 12 December 2019).
  34. Klem, G.H.; Lüders, H.O.; Jasper, H.H.; Elger, C. The ten-twenty electrode system of the International Federation. The International Federation of Clinical Neurophysiology. Electroencephalogr. Clin. Neurophysiol. Suppl. 1999, 52, 3–6. [Google Scholar]
  35. Wilms, L.; Oberfeld, D. Color and emotion: Effects of hue, saturation, and brightness. Psychol. Res. 2018, 82, 896–914. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overview of our framework: three groups of participants watch three different categories of contents: contrast enhanced, original, and contrast reduced. We process their EEG signals through our multi-columned emotion recognition model and estimate their valence and arousal.
Figure 1. The overview of our framework: three groups of participants watch three different categories of contents: contrast enhanced, original, and contrast reduced. We process their EEG signals through our multi-columned emotion recognition model and estimate their valence and arousal.
Sensors 20 04543 g001
Figure 2. Ground truth emotions estimated from the clips sampled from the three contents.
Figure 2. Ground truth emotions estimated from the clips sampled from the three contents.
Sensors 20 04543 g002
Figure 3. The scenes of positive emotion.
Figure 3. The scenes of positive emotion.
Sensors 20 04543 g003
Figure 4. The scenes of neutral emotion.
Figure 4. The scenes of neutral emotion.
Sensors 20 04543 g004
Figure 5. The scenes of negative emotion.
Figure 5. The scenes of negative emotion.
Sensors 20 04543 g005
Figure 6. Our multi-columned emotion recognition model with five recognizing modules: Each module is composed of four convolutional layers, one max pooling layer and four fully connected layers. The activation function we employ is a tanh function, which produces values in ( 1 , 1 ) . The result of each module is a paired value of valence and arousal.
Figure 6. Our multi-columned emotion recognition model with five recognizing modules: Each module is composed of four convolutional layers, one max pooling layer and four fully connected layers. The activation function we employ is a tanh function, which produces values in ( 1 , 1 ) . The result of each module is a paired value of valence and arousal.
Sensors 20 04543 g006
Figure 7. The construction of input data: 32 consecutive samples captured from 32 EEG channels constructs a single input vector for a recognizing module. Since we use five modules in our recognition model, five consecutive samples are captured.
Figure 7. The construction of input data: 32 consecutive samples captured from 32 EEG channels constructs a single input vector for a recognizing module. Since we use five modules in our recognition model, five consecutive samples are captured.
Sensors 20 04543 g007
Figure 8. Nine categories of movie clips combined by contents of three emotions and three contrast variation. The color of each category corresponds to the human participant group.
Figure 8. Nine categories of movie clips combined by contents of three emotions and three contrast variation. The color of each category corresponds to the human participant group.
Sensors 20 04543 g008
Figure 9. The schedule for experiment of three groups.
Figure 9. The schedule for experiment of three groups.
Sensors 20 04543 g009
Figure 10. The valence and arousal extracted from the visual contents of neutral emotion.
Figure 10. The valence and arousal extracted from the visual contents of neutral emotion.
Sensors 20 04543 g010
Figure 11. The valence and arousal extracted from the visual contents of positive emotion.
Figure 11. The valence and arousal extracted from the visual contents of positive emotion.
Sensors 20 04543 g011
Figure 12. The valence and arousal extracted from the visual contents of negative emotion.
Figure 12. The valence and arousal extracted from the visual contents of negative emotion.
Sensors 20 04543 g012
Figure 13. The change of valence from controlling contrast.
Figure 13. The change of valence from controlling contrast.
Sensors 20 04543 g013
Figure 14. The change of arousal from controlling contrast.
Figure 14. The change of arousal from controlling contrast.
Sensors 20 04543 g014
Figure 15. Comparison of the enhanced emotions from neutral and the original positive and negative emotions.
Figure 15. Comparison of the enhanced emotions from neutral and the original positive and negative emotions.
Sensors 20 04543 g015
Figure 16. Comparison of the enhanced emotions from positive and the original neutral emotion.
Figure 16. Comparison of the enhanced emotions from positive and the original neutral emotion.
Sensors 20 04543 g016
Figure 17. Comparison of the enhanced emotions from negative and the original neutral emotion.
Figure 17. Comparison of the enhanced emotions from negative and the original neutral emotion.
Sensors 20 04543 g017
Figure 18. The contents whose emotions are mostly changed.
Figure 18. The contents whose emotions are mostly changed.
Sensors 20 04543 g018
Figure 19. The contents whose emotions are leastly changed.
Figure 19. The contents whose emotions are leastly changed.
Sensors 20 04543 g019
Figure 20. The change of contrast and intensity according to the contrast control.
Figure 20. The change of contrast and intensity according to the contrast control.
Sensors 20 04543 g020
Table 1. The precision, recall and F1 score.
Table 1. The precision, recall and F1 score.
Precision (%)Recall (%)F1 Score (%)
Valence Arousal Valence Arousal Valence Arousal
85.5786.6680.1881.0282.7983.75
Table 2. Gender and age distribution of the participants.
Table 2. Gender and age distribution of the participants.
GenderAge
FemaleMale < 20 20 s 30 s > 40
group 1642710
group 2551710
group 3551810
Table 3. p values for the valences and arousals between neutral, positive and negative emotions from the three contents.
Table 3. p values for the valences and arousals between neutral, positive and negative emotions from the three contents.
Loving VincentThe Old Man and the SeaThe Man Who Plants Trees
val.arou.val.arou.val.arou.
Neutral and Positive 3.84936 × 10 14 0.82521 2.09928 × 10 11 0.39830 7.92079 × 10 12 1.0
Neutral and Negative 4.30645 × 10 14 0.10998 2.28149 × 10 11 0.01567 6.66321 × 10 12 0.27336
Table 4. p values for the valences and arousals between original contents and contrast controlled contents for three original emotion.
Table 4. p values for the valences and arousals between original contents and contrast controlled contents for three original emotion.
OriginalEmotionLoving VincentThe Old Man and the SeaThe Man Who Plants Trees
EmotionsControlval.arou.val.arou.val.arou.
neutraloriginal & up 7.766 × 10 14 1.0 5.879 × 10 7 1.0 5.342 × 10 4 0.5527
original & down 6.859 × 10 12 0.7343 7.803 × 10 9 0.7941 3.451 × 10 5 0.6606
positiveoriginal & up 1.754 × 10 10 0.4495 2.194 × 10 8 0.5248 1.754 × 10 10 0.8598
original & down 9.166 × 10 12 0.8502 1.025 × 10 6 0.8198 9.166 × 10 12 0.3604
negativeoriginal & up 3.296 × 10 8 0.2025 1.695 × 10 12 0.2526 3.773 × 10 7 0.8279
original & down 1.618 × 10 6 0.0716 4.981 × 10 11 0.0332 2.108 × 10 6 0.7278

Share and Cite

MDPI and ACS Style

Yang, H.; Han, J.; Min, K. Emotion Variation from Controlling Contrast of Visual Contents through EEG-Based Deep Emotion Recognition. Sensors 2020, 20, 4543. https://doi.org/10.3390/s20164543

AMA Style

Yang H, Han J, Min K. Emotion Variation from Controlling Contrast of Visual Contents through EEG-Based Deep Emotion Recognition. Sensors. 2020; 20(16):4543. https://doi.org/10.3390/s20164543

Chicago/Turabian Style

Yang, Heekyung, Jongdae Han, and Kyungha Min. 2020. "Emotion Variation from Controlling Contrast of Visual Contents through EEG-Based Deep Emotion Recognition" Sensors 20, no. 16: 4543. https://doi.org/10.3390/s20164543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop