Next Article in Journal
Traffic-Aware Secured Cooperative Framework for IoT-Based Smart Monitoring in Precision Agriculture
Previous Article in Journal
Application of Teager–Kaiser Energy Operator in the Early Fault Diagnosis of Rolling Bearings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

N-Step Pre-Training and Décalcomanie Data Augmentation for Micro-Expression Recognition

Department of Artificial Intelligence, Kyungpook National University, Daegu 41566, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6671; https://doi.org/10.3390/s22176671
Submission received: 11 July 2022 / Revised: 19 August 2022 / Accepted: 28 August 2022 / Published: 3 September 2022
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:
Facial expressions are divided into micro- and macro-expressions. Micro-expressions are low-intensity emotions presented for a short moment of about 0.25 s, whereas macro-expressions last up to 4 s. To derive micro-expressions, participants are asked to suppress their emotions as much as possible while watching emotion-inducing videos. However, it is a challenging process, and the number of samples collected tends to be less than those of macro-expressions. Because training models with insufficient data may lead to decreased performance, this study proposes two ways to solve the problem of insufficient data for micro-expression training. The first method involves N-step pre-training, which performs multiple transfer learning from action recognition datasets to those in the facial domain. Second, we propose Décalcomanie data augmentation, which is based on facial symmetry, to create a composite image by cutting and pasting both faces around their center lines. The results show that the proposed methods can successfully overcome the data shortage problem and achieve high performance.

1. Introduction

Humans reveal personal feelings, intentions, and emotional conditions through their facial expressions. Generally, a person reveals emotions through explicit macro-expressions that last between 0.25 and 4 s. During these periods, the emotions expressed on the face and the actual feelings felt coincide. Conversely, when a person unconsciously reveals a hidden emotion in fractional time (e.g., 0.25 s), it is considered to be a micro-expression. These are likely to be missed or misinterpreted, even in laboratory settings. Figure 1 shows a comparison between the micro- and macro-expressions.
In 1966, Haggard et al. [1] first proposed the concept of micro-expressions. About three years later, Ekman et al. [2] witnessed this phenomenon while researching lie detection using interview videos of psychologists and patients. Robust micro-expression recognition systems are used in various fields, such as criminal recognition, lie detection, and psychological diagnosis. Owing to the broad applicability of micro-expression recognition, many studies have been conducted in recent years. These studies are primarily divided into hand-crafted feature-based methods [3,4,5,6,7,8] and deep learning-based methods [9,10,11,12,13,14]. The hand-crafted feature extraction was widely used at the beginning of the development of micro-expression detection technology because it extracts direct-designed features according to the given learning method.
On the other hand, the deep learning-based method automatically extracts features to solve specific tasks. Because such models are more convenient and achieve higher performance than hand-crafted feature-based methods, more studies using deep learning-based methods have recently emerged. Although micro-expression samples are limited and have a low intensity which causes difficulties, deep learning-based networks, such as convolutional neural networks (CNNs) and long short-term memory models, have achieved excellent progress.
Even though micro-expression datasets [4,15,16] consist of sequential frames, most studies thus far have used static images or optical flows [13,17,18,19], and few have used the frame sequence as input to the network [9,20]. Additionally, some studies have simultaneously used macro-expression data for learning with sequential inputs. However, to the best of our knowledge, no studies have transferred a model pre-trained on other more extensive datasets, such as action recognition datasets, to micro-expression datasets several times.
Therefore, in this paper, we propose a novel framework to overcome the insufficient number of clips in the micro-expression dataset. Our main contributions can be summarized as follows:
  • N-step pre-training is proposed which use extensive datasets in addition to facial datasets;
  • We propose a new augmentation method specialized for facial data, called “Décalcomanie”;
  • The combination of N-step pre-training and Décalcomanie augmentation outperforms state-of-the-art approaches in micro-expression recognition.
The remainder of this paper is organized as follows: Section 2 describes the related studies on micro-expression recognition. Section 3 describes the data preprocessing methods prior to training and the details of the proposed approach. We present the experimental settings and results in Section 4, and we conclude this paper in Section 5.

2. Related Work

In facial expression recognition, two methods are mainly used: hand-crafted feature-based methods and deep learning-based methods. In the hand-crafted feature extraction, LBP variants-based methods [3,4] and optical flow-based methods [13,19] are widely used. However, because micro-expression images have little movement and implied change, it is difficult for the hand-crafted feature-based method to achieve high performance.
Deep learning-based methods have performed better than hand-crafted feature-based methods in the field of computer vision in recent years. Therefore, studies using the deep learning-based method have been actively proposed even in micro-expression recognition. Wu et al. [9] used a three-stream combining 2D and 3D convolutional neural network to classify expressions and proposed two variants: intermediate fusion and late fusion. Xia et al. [10] improved the performance by adversarial learning using both the micro-expression dataset and the macro-expression dataset. Recently, Hung et al. [20] proposed multilevel transfer learning for a macro-expression dataset. They used related domain datasets for pre-training and performed experiments on the macro-expression dataset.
In this paper, we also propose a deep learning-based micro-expression recognition method. The methods mentioned earlier have not adequately utilized the pre-trained model and are dependent on additional information such as landmarks and optical flow. Therefore, we propose a novel method that uses a pre-trained model that learns large amounts of data across the N-step and a transformation method specialized in facial datasets called Decalcomanie. We emphasize that we utilize massive datasets with unrelated domains to facial in multistep pre-training, although Hung et al. [20] only utilized relevant domain datasets.

3. Proposed Method

Insufficient data may lead to poor model performance. This section presents two methods to overcome the problem: N-step pre-training and Décalcomanie data augmentation. Before explaining the proposed methods in detail, we first show the preprocessing methods used to deal with the sequential data. The framework of the proposed method is shown in Figure 2.

3.1. Preprocessing

3.1.1. Oversampling with Synthetic Samples

The micro-expression datasets have an imbalanced distribution. For example, in the Spontaneous Micro-expression (SMIC) dataset, the negative class has 70 samples, the positive class has 51, and the surprise class has 43. Training a model with such an imbalanced data may lead to overfitting. To handle this problem, Wu et al. [9] used an extension of the synthetic minority oversampling technique (SMOTE) [21] algorithm. They modified the SMOTE algorithm by using two samples to generate a new one using the following formula:
x n e w , i = λ · x 1 , i + ( 1 λ ) · x 2 , i ,
where x 1 , i and x 2 , i are random samples belonging to the same category, and x 1 , i indicates the i-th frame of the sequence of sample x 1 . λ is sampled from a uniform distribution (0, 1). They attempted to solve the class imbalance by creating a new sample belonging to the insufficient class using the algorithm. We also adopted this method to alleviate the imbalance problem.

3.1.2. Frame Interpolation

The micro-expression dataset comprises sequence samples having different numbers of frames. Before training, all frames of the samples must be fixed at the same number to capture the appropriate temporal information. We adopted a linear interpolation method, provided by PyTorch, to fix the number of frames. The method samples down or up to the desired size through linear interpolation. We used this method because it is easy to implement, and there is no redundancy or loss of information.

3.2. N-Step Pre-Training

Transfer learning uses a pre-trained model to solve another task. Many studies have attempted to improve performance by transferring a pre-trained model that learned macro-expression data first to another micro-expression case. However, to the best of our knowledge, no studies have applied multistep transfer learning using massive sequential datasets that have even unrelated domains for micro-expression datasets (e.g., Kinetics-400 [22], UCF101 [23]), and a large-scale visual recognition dataset (e.g., ImageNet [24]). Although the micro-expression dataset has a different domain from the action recognition dataset, it is the same in that it contains temporal and spatial information. Therefore, we assumed that, if the model learns this helpful information first, there will be a benefit in the micro-expression data.
Multistep pre-training is composed of multiple-transfer learning. With one dataset, we can train the model from scratch and replace the fully-connected layer with a randomly initialized one. We can then fine-tune the model again on another dataset. A method of performing such transfer learning multiple times on different datasets is the multistep pre-training method that we propose. However, we emphasize that N-step pre-training does not simply attempt to implement multiple transfer learning; the sequence of datasets used in multiple steps is important. Figure 3 shows the procedure of two-step pre-training. Here, we train the model on the action recognition dataset first and then on the macro-expression dataset again. Lastly, we fine-tune the model on the micro-expression dataset.

3.3. Décalcomanie Augmentation

According to Mandal et al. [25], the left side of the human face is more expressive of middle-intensity happiness and minimum-intensity happiness and sadness. The right side is more expressive in the most intense expressions of happiness and sadness. Because both sides of the face reveal slightly different information about emotion, we devised Décalcomanie augmentation. Originally, Décalcomanie was a drawing technique that draws only half of a picture, folds it, and transcribes it to the other side, similar to the stamping principle. Décalcomanie augmentation also divides the face in half and transcribes back to the other side to create a new sample. Figure 4 illustrates the procedure of Décalcomanie augmentation.
As the video clips of micro-expression datasets used in this paper are not inverted, as when looking in a mirror, the left face and right face of the subject are originally right-side and left-side, respectively. However, to facilitate understanding and avoid confusion, we refer to the left and right faces as the left and right sides when looking at the video clip. If Décalcomanie is applied, the left- and right-side frames can be obtained alongside the original ones. Because three kinds of the frames can be obtained, a new training process was required to learn the additional samples. Therefore, we propose two versions of training process: (1) shared backbone and multiple losses, and (2) fusion with shared backbone.

3.3.1. Shared Backbone and Multiple Losses

This version shares the backbone network and adds up the losses obtained by feed-forwarding each frame to the model. We used four subset cases consisting of OLR, OL, OR, and LR, where O, L, and R denote the original, left-side, and right-side frames, respectively. For each case, we calculated the training loss using Equation (2):
L O L R = λ O · L o r i g i n a l + λ L · L l e f t + λ R · L r i g h t , L O L = λ O · L o r i g i n a l + λ L · L l e f t , L O R = λ O · L o r i g i n a l + λ R · L r i g h t , L L R = λ L · L l e f t + λ R · L r i g h t ,
where L is cross-entropy loss. λ O , λ L , and λ R are hyperparameters multiplied by each corresponding loss, and we set the sum of each λ in each equation to one. Figure 5 illustrates the procedure for using LR frames.
In this version, we tested the model’s performance with two cases: (1) using only the original frames and (2) using all frames as input. In the latter case, s o f t m a x ( · ) was applied after summing the output values obtained by feed-forwarding each input to the shared backbone.

3.3.2. Fusion with Shared Backbone

Each feature extracted from the shared backbone was concatenated and forwarded to a single linear layer or multilayer perceptron (MLP). We set the hidden dimensions of the MLP to 256. Like the shared backbone and multiple losses version, we used OLR, OL, OR, and LR, which are possible input cases, and the combination of the frames used during training was the same for testing. Figure 6 shows the procedure of fusion with the shared backbone using LR frames.

4. Experiments

First, we summarize the experimental setup and then verify the effectiveness of each proposed method. Lastly, we show the performance of the model when the two proposed methods are combined.

4.1. Setup

Here, we summarize the dataset details used in the experiment, the metrics used to evaluate the model performance, and the implementation details in order.

4.1.1. Datasets

Facial Dataset

Facial expressions can be divided into two main categories: macro-expression and micro-expression. Macro-expressions appear on our faces without any oppression when emotions are usually expressed. Macro-expressions are easy to recognize because of their long duration (0.25–4 s), and the emotion to be revealed is clearly expressed. Examples of each dataset we used are shown in Figure 7. The details of the macro-expression datasets are as follows:
  • Extended Cohn–Kanade (CK+) [26] initially contained 593 video sequences of 123 subjects aged 18 to 50 with diverse genders and heritage. The samples were collected at 30 fps with a resolution of 640 × 490 or 640 × 480 pixels. These samples were divided into seven categories: anger, disgust, fear, sadness, contempt, happiness, and surprise.
  • Oulu-CASIA [27] consists of six emotion classes (i.e., anger, disgust, fear, sadness, happiness, and surprise) from 80 people between 23 and 58 years old. The camera recorded the expressions at 25 fps with a resolution of 320 × 240 pixels. This dataset was collected under two conditions: near-infrared (NIR) and visible (VIS) light systems. VIS images consist of three different versions: dark, strong, and weak. The dark versions are created when samples are taken in a dark environment. The strong version is the case where the emotion of the subject’s expression stands out, and the weak version is the opposite.
Micro-expressions, alternatively, are unconsciously short facial expressions typically made under stress. They appear for only 0.5 s and are very subtle, making it difficult to judge emotions, even when intentionally observed. There are many micro-expression datasets, but high quality ones include SMIC [4], CASME2 [15], and SAMM [16]. These datasets were recorded in a laboratory environment, and subjects were asked to maintain a “poker face” without showing emotions as much as possible under different stimuli. The details of each dataset are as follows:
  • SMIC [4] consists of samples recorded by 100 fps high-speed cameras and samples recorded at standard speed with 25 fps of both VIS and NIR light ranges. Each subset is referred to as SMIC-HS, SMIC-VIS, and SMIC-NIR. The SMIC-HS consists of 164 samples taken from 16 subjects, each assigned a label of “negative”, “positive”, or “surprise”.
  • CASME2 [15] is an improved version of the existing micro-expression dataset, CASME [28]. CASME2 was filmed using a high-speed 250 fps camera and crops only the face part at a size of 280 × 340 pixels. The dataset has 247 samples generated from 26 subjects, but some expressions are unevenly distributed because they were difficult to derive under laboratory conditions. It provides five classes of micro-expressions.
  • SAMM [16] comprises 159 samples collected from 32 participants. Although other datasets lack ethnic diversity, the SAMM dataset consists of 13 different ethnicities; the average age was 33.24 years, and the gender distribution was almost identical. Each sample was collected at a high speed of 200 fps with a high resolution of 2040 × 1088. The labels for each sample were emotion labels designated by participants through surveys.

Action Recognition Dataset

Because the facial dataset is sequential, the model learns spatial and temporal information during training. The final model performance can be higher if a large-scale action recognition dataset is used to learn spatiotemporal information in advance before learning the facial dataset.
  • Kinetics-400 [22] consists of 400 human action classes and has at least 400 video clips per class. These videos were taken from YouTube and focus on human actions. The dataset covers a broad range of classes that are largely divided into person (e.g., drawing, laughing, and fist-pumping), person-person (e.g., hugging and shaking hands), and person–object actions (e.g., mowing the lawn and washing dishes). Unlike facial datasets, the entire human body is often displayed on a video.
  • UCF-101 [23] consists of 101 action categories and videos collected from YouTube. It contains 13,320 videos and is one of the most challenging datasets because the videos were filmed under significant variations of camera motion. The categories are largely classified into five types: human–object interaction, body motion only, human–human interaction, playing music instruments, and sports. Similar to Kinetics, UCF-101 is a human action recognition dataset, but the proportion of videos capturing face or upper body is slightly higher.

4.1.2. Evaluation Metrics

In this paper, we use the leave-one-subject-out (LOSO) cross-validation protocol to evaluate the model. We adopted this protocol to make up for the micro-expression dataset having a biased number of samples per subject. If K subjects are in the dataset, the LOSO protocol divides the experiment into K folds. It uses one subject as the test set and the remaining K 1 subjects as the training set, which leads to K experiments. Meanwhile, the micro-expression datasets also have a biased distribution of samples for emotion classes, providing an imbalanced distribution. To reduce the potential bias, we used weighted average recall (WAR), unweighted average recall (UAR), and unweighted F1 score (UF1) as evaluation standards.
UF 1 = 1 C c = 1 C F 1 c , UAR = 1 C c = 1 C T P c N c , WAR = C T P c c = 1 C N c
where
F 1 = 2 T P c 2 T P c + F P c + F N c
In Equation (3), C is the number of classes, and T P , F N , and F P are the true positive, false negative, and false positive, respectively. N c is the total number of samples in category C.

4.1.3. Implementation Details

We used one NVIDIA RTX A6000 48 GB GPU per experiment. For N-step pre-training experiments, we trained 3D-ResNet-50 [29] for 30 epochs using Adam [30] with β 1 = 0.9 , β 2 = 0.999 , a batch size of 30, and a learning rate of 0.0001, which decayed by 10 at the 13th, 18th, and 22nd epoch. We resized all frames to 112. In the case of Décalcomanie experiments, we trained 3D-ResNeXt-101 [29] for 100 epochs. This is different in that the learning rate is decayed by 10 at the 30th, 60th, and 80th epochs; however, all other hyperparameters are the same as in the pre-training experiment. In the case of OLR frames, we set λ O , λ L , and λ R to 0.4, 0.4, and 0.2, respectively. For OL, OR, and LR, we set λ to 0.5. We used synthetic samples generated using extended SMOTE with the N-step pre-training experiments, but not in the Décalcomanie experiments. Because synthetic samples have mixed faces, cutting them in half and combining them can cause noisy representations. Furthermore, because the number of frames in each video sample of the micro-expression datasets is different, it was necessary to fix the datasets’ frames to capture the temporal information. We used the linear interpolation to set the number of frames to their average number. As a result, we set the video lengths of SMIC, SAMM, and CASME2 to 34, 74, and 66, respectively.

4.2. Effect of N-Step Pre-Training

We conducted an N-step pre-training experiment for a micro-expression dataset using 3D-ResNet-50. The datasets used as source datasets were Kinetics-400, UCF101, ImageNet, and macro-expression datasets CK+ and Oulu-CASIA. As Kinetics-400 and ImageNet are massive, we use public pre-trained models on those without direct learning to reduce learning costs. We used SMIC and SAMM as target datasets in this experiment. We showed a visual comparison of the UF1 and UAR performance of each N-step pre-trained model in Figure 8. We also presented the numerical results in Table 1. For convenience, We refer to ImageNet, Kinetics-400, UCF101, and a combination of CK+ and Oulu-CASIA as IN, Kinetics, UCF, and Macro, respectively. We analyzed the result with a particular focus on UF1 and UAR. Through an experimental investigation, we found three interesting results.
First, in one-step pre-training, we found that the model pre-trained on IN performed worse than the scratch model, and in the case of the rest of the datasets, they performed better, and their performance was higher in the order of kinetics, UCF, and Macro. Although IN is large-scale, it hinders model performance because it is composed of still images without temporal information. In addition, Kinetic has a higher proportion of frames focusing on the whole body, and UCF has a higher proportion of frames with the upper body and face than Kinetics. Hence, UCF is slightly closer than Kinetics, although its domain is clearly different from the facial dataset. Therefore, we assumed for that reason that the performance of the model pre-trained on UCF was higher than that of the model pre-trained on Kinetics. For macro-expression datasets with a similar domain to micro-expression datasets, the performance was the highest among one-step pre-training methods.
Second, we analyzed whether the order of datasets used for pre-training affects performance in N-step pre-training. In Table 1, the last second lines show the experimental results obtained by switching UCF and Macro with the dataset learned in the last step. Training the model on Macro in the last step was better than UCF, suggesting that the dataset order of the pre-training process could achieve higher performance as it proceeded in the order of datasets with domains like the micro-expression domain.
Finally, when the macro-expression dataset was used for pre-training in the last step, we compared the Scratch, Macro, Kinetics → Macro, IN → UCF → Macro, and Kinetics → UCF → Macro to evaluate whether performance improves as the number of steps increases. As a result, we confirmed that, with more steps, performance improved. Furthermore, when comparing IN → UCF → Macro and Kinetics → UCF → Macro, which used the same three-step pre-training, the performance of Kinetics → UCF → Macro was better than that of IN → UCF → Macro, which is presumed to have been caused by the presence or absence of temporal information, as mentioned.

4.3. Effect of Décalcomanie Augmentation

We trained 3D-ResNeXt-101 from scratch without using the pre-trained model solely to check the performance of Décalcomanie data augmentation. We did not use the aforementioned oversampling method. The datasets used in experiments are SMIC, CASME II, and SAMM. We resized each frame to 112 and applied scale, rotation, and horizontal flip augmentation in addition to Décalcomanie.
The results of the shared backbone and multiple loss on SMIC, SAMM, and CASME II using 3D-ResNeXt-101 from scratch are shown in Table 2. The first line of the table means the result when we did not apply Décalcomanie augmentation. First, we tested if Décalcomanie augmentation can be used as test time augmentation. During training time, we only used the original frame as input and used various frames created via Décalcomanie such as OLR, OL, and OR during test time. The top four lines of Table 2 prove that Décalcomanie can be utilized as test time augmentation. The performances got higher for every dataset when our proposed augmentation method was used as test time augmentation. For example, when test time augmentation was not used for the SMIC dataset, UF1, UAR, and WAR were 0.5833, 0.5680, and 0.5671, respectively. However, UF1, UAR, and WAR rose to 0.5985, 0.5929, and 0.6098 when OLR frames created using the Décalcomanie augmentation were input during the test, meaning that the performance was increased by 0.0152, 0.0246, and 0.0427, respectively.
Then, we also conducted experiments to verify the effect of Décalcomanie as data augmentation during training. In most of the experiments, models with Décalcomanie achieved higher performance than those without it. In addition, overall, the model with a single linear layer obtained higher performance than the model with a multilayer perceptron. In the experiments using SMIC and CASME II, the performance improvement was the highest in the sub-experiment using OLR frames as input. We hypothesize that these results occurred because emotional expressions were well represented on both sides of the face. Unlike the results, in the case of SAMM, the performance improvement was the greatest when using the OR frames. When we compared SAMM to other datasets, the subjects’ expressions were significantly more emotionally inhibited. Although it is difficult to detect changes due to minimal facial changes, we assume that the reason for the highest performance when inserting the OR frames is that there was a little more minimum intensity on the right side of the subject’s face.
The results of the fusion with the shared backbone are shown in Table 3. Overall, the fusion with shared backbone version does not have a higher overall performance than the shared backbone and multiple losses version. However, there was a performance improvement compared to when Décalcomanie was not applied. Using a single linear as a classifier performed better than using multilayer perceptron in most cases. Due to the small number of data, the performance of the deeper model is not higher. In this experiment, UF1, UAR, and WAR were highest in SMIC and CASME II when OL or OLR frames were used, and in the case of SAMM, when OR frames were used achieved the highest performance. As can be seen from the results, it can be assumed that SMIC and CASME II contained facial expression information evenly on both faces. SAMM also showed more minor performance improvements than did the other two datasets because emotional expression was further suppressed.
To further research the effectiveness of the Décalcomanie method, we compared it with other augmentation methods. The results are shown in Table 4. First, we compared it with basic augmentation methods usually used. We rotated images with 30 degrees, resized images with scale ( 1.1 , 1.1 ) , or flipped images horizontally. The results of applying primary augmentations show low performances, which means that the simple transformation of the frames does not have much effect. When we applied the SMOTE algorithm that produces the synthetic samples used by Wu et al. [9] with simple augmentations, the performance of all the metrics was over 0.6. However, when we applied our proposed augmentation method, Decalcomanie, the models’ performance was much higher than when using SMOTE. This result demonstrates that Décalcomanie has a remarkable effect on facial data.

4.4. Overall Results

Here, we combined the N-step pre-training and Décalcomanie data augmentation to evaluate the overall results. We used the best combination in the N-step pre-training experiment (Kinetics → UCF → Macro). 3D-ResNeXt-101 was used as the backbone network. After N-step pre-training, we fine-tuned the pre-trained model on the SMIC dataset and evaluated its performance.
Table 5 shows the results. As stated, shared backbone and multiple losses experiments had better performance than did fusion with shared backbone experiments. Earlier, the performance was highest on the SMIC dataset when experimenting with OLR or OL frames, showing the same results in this experiment. In the case of shared backbone and multiple losses, when both N-step pre-training and the Décalcomanie method were applied compared to Scratch, UF1, UAR, and WAR increased by +0.2104, +0.2267, and +0.2256, respectively. For fusion with shared backbone experiments, the performance was lower than shared backbone and multiple losses; however, the performance was still significantly better than scratch.
We compared our proposed method with other state-of-the-art approaches which used the LOSO protocol and the same number of classes for a fair comparison. Since the results of the shared backbone and multiple losses were better than the fusion with shared backbone in the previous experiments, we used the figures of the shared backbone and multiple losses version for comparison. Refs. [3,4,13,19] are hand-crafted feature-based methods and [10,31,32,33] are deep learning-based methods which used additional datasets in addition to the micro-expression dataset. The comparison with these methods is shown in Table 6. We achieve the best performance when our proposed methods, such as N-step pre-training and Décalcomanie data augmentation, are combined.

5. Conclusions

In this paper, we proposed N-step pre-training and Décalcomanie augmentation to avoid the data shortage problem in micro-expression recognition. In N-step pre-training, we transferred the model multiple times on various datasets in a specific order. Furthermore, we devised a data augmentation method specialized for the face called Décalcomanie. We evaluate the effectiveness of each method on micro-expression datasets. When combining N-step pre-training and Décalcomanie augmentation, the experimental results show that our proposed methods outperform the state-of-the-art methods on the SMIC dataset.

Author Contributions

The contribution of the authors for this publication article are as follows: C.L.: methodology, software, conceptualization, writing—original draft, writing—review and editing. J.H.: investigation, writing—review and editing. H.J.: conceptualization, project administration, supervision, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (2019-0-00330, Development of AI Technology for Early Screening of Infant/Child Autism Spectrum Disorders based on Cognition of the Psychological Behavior and Response). Furthermore, this research was also supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.2021-0-02068, Artificial Intelligence Innovation Hub).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Haggard, E.A.; Isaacs, K.S. Micromomentary facial expressions as indicators of ego mechanisms in psychotherapy. In Methods of Research in Psychotherapy; Springer: Berlin/Heidelbegr, Germany, 1966; pp. 154–165. [Google Scholar]
  2. Ekman, P.; Friesen, W.V. Nonverbal leakage and clues to deception. Psychiatry 1969, 32, 88–106. [Google Scholar] [CrossRef] [PubMed]
  3. Huang, X.; Wang, S.J.; Zhao, G.; Piteikainen, M. Facial micro-expression recognition using spatiotemporal local binary pattern with integral projection. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 1–9. [Google Scholar]
  4. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–6. [Google Scholar]
  5. Takalkar, M.A.; Xu, M.; Chaczko, Z. Manifold feature integration for micro-expression recognition. Multimed. Syst. 2020, 26, 535–551. [Google Scholar] [CrossRef]
  6. Wang, Y.; See, J.; Phan, R.C.W.; Oh, Y.H. Lbp with six intersection points: Reducing redundant information in lbp-top for micro-expression recognition. In Computer Vision—ACCV 2014, Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; Springer: Cham, Switzerland, 2014; pp. 525–537. [Google Scholar]
  7. Huang, X.; Zhao, G.; Hong, X.; Zheng, W.; Pietikäinen, M. Spontaneous facial micro-expression analysis using spatiotemporal completed local quantized patterns. Neurocomputing 2016, 175, 564–578. [Google Scholar] [CrossRef]
  8. Huang, X.; Zhao, G. Spontaneous facial micro-expression analysis using spatiotemporal local radon-based binary pattern. In Proceedings of the 2017 International Conference on the Frontiers and Advances in Data Science (FADS), Xi’an, China, 23–25 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 159–164. [Google Scholar]
  9. Wu, C.; Guo, F. TSNN: Three-Stream Combining 2D and 3D Convolutional Neural Network for Micro-Expression Recognition. IEEJ Trans. Electr. Electron. Eng. 2021, 16, 98–107. [Google Scholar] [CrossRef]
  10. Xia, B.; Wang, W.; Wang, S.; Chen, E. Learning from Macro-expression: A Micro-expression Recognition Framework. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2936–2944. [Google Scholar]
  11. Xie, H.X.; Lo, L.; Shuai, H.H.; Cheng, W.H. Au-assisted graph attention convolutional network for micro-expression recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2871–2880. [Google Scholar]
  12. Li, Y.; Huang, X.; Zhao, G. Micro-expression action unit detection with spatial and channel attention. Neurocomputing 2021, 436, 221–231. [Google Scholar] [CrossRef]
  13. Li, J.; Wang, Y.; See, J.; Liu, W. Micro-expression recognition based on 3D flow convolutional neural network. Pattern Anal. Appl. 2019, 22, 1331–1339. [Google Scholar] [CrossRef]
  14. Kim, D.H.; Baddar, W.J.; Ro, Y.M. Micro-expression recognition with expression-state constrained spatio-temporal feature representations. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 382–386. [Google Scholar]
  15. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef] [PubMed]
  16. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. Samm: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2016, 9, 116–129. [Google Scholar] [CrossRef]
  17. Farnebäck, G. Two-frame motion estimation based on polynomial expansion. In Image Analysis, Proceedings of the Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 363–370. [Google Scholar]
  18. Zhou, L.; Mao, Q.; Xue, L. Dual-inception network for cross-database micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  19. Liong, S.T.; See, J.; Wong, K.; Phan, R.C.W. Less is more: Micro-expression recognition from video using apex frame. Signal Process. Image Commun. 2018, 62, 82–92. [Google Scholar] [CrossRef]
  20. Hung, J.C.; Chang, J.W. Multi-level transfer learning for improving the performance of deep neural networks: Theory and practice from the tasks of facial emotion recognition and named entity recognition. Appl. Soft Comput. 2021, 109, 107491. [Google Scholar] [CrossRef]
  21. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  22. Kay, W.; Carreira, J.; Simonyan, K.; Zhang, B.; Hillier, C.; Vijayanarasimhan, S.; Viola, F.; Green, T.; Back, T.; Natsev, P.; et al. The kinetics human action video dataset. arXiv 2017, arXiv:1705.06950. [Google Scholar]
  23. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  24. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef]
  25. Mandal, M.K.; Asthana, H.S.; Pandey, R. Asymmetry in emotional face: Its role in intensity of expression. J. Psychol. 1995, 129, 235–241. [Google Scholar] [CrossRef] [PubMed]
  26. Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 94–101. [Google Scholar]
  27. Zhao, G.; Huang, X.; Taini, M.; Li, S.Z.; PietikäInen, M. Facial expression recognition from near-infrared videos. Image Vis. Comput. 2011, 29, 607–619. [Google Scholar] [CrossRef]
  28. Yan, W.J.; Wu, Q.; Liu, Y.J.; Wang, S.J.; Fu, X. CASME database: A dataset of spontaneous micro-expressions collected from neutralized faces. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–7. [Google Scholar]
  29. Kataoka, H.; Wakamiya, T.; Hara, K.; Satoh, Y. Would mega-scale datasets further enhance spatiotemporal 3D CNNs? arXiv 2020, arXiv:2004.04968. [Google Scholar]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Song, B.; Li, K.; Zong, Y.; Zhu, J.; Zheng, W.; Shi, J.; Zhao, L. Recognizing spontaneous micro-expression using a three-stream convolutional neural network. IEEE Access 2019, 7, 184537–184551. [Google Scholar] [CrossRef]
  32. Liu, J.; Zheng, W.; Zong, Y. SMA-STN: Segmented Movement-Attending Spatiotemporal Network forMicro-Expression Recognition. arXiv 2020, arXiv:2010.09342. [Google Scholar]
  33. Xia, B.; Wang, S. Micro-Expression Recognition Enhanced by Macro-Expression from Spatial-Temporal Domain. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, Virtual, 19–26 August 2021; Zhou, Z.H., Ed.; pp. 1186–1193. [Google Scholar] [CrossRef]
Figure 1. Examples of macro-expression (left) and micro-expression (right). Macro-expression samples have high intensity, while micro-expression samples show little change in facial expression.
Figure 1. Examples of macro-expression (left) and micro-expression (right). Macro-expression samples have high intensity, while micro-expression samples show little change in facial expression.
Sensors 22 06671 g001
Figure 2. Framework of our proposed method (three-step pre-training with Décalcomanie augmentation). First, we train the model from scratch on Kinetics-400 and transfer it to UCF101. Next, we train the model on the macro-expression dataset. Finally, the model is fine-tuned on the micro-expression dataset by applying the Décalcomanie augmentation method. In each step, we replace the extant fully-connected (FC) layer of the model with a randomly initialized FC layer. CNN, convolutional neural network.
Figure 2. Framework of our proposed method (three-step pre-training with Décalcomanie augmentation). First, we train the model from scratch on Kinetics-400 and transfer it to UCF101. Next, we train the model on the macro-expression dataset. Finally, the model is fine-tuned on the micro-expression dataset by applying the Décalcomanie augmentation method. In each step, we replace the extant fully-connected (FC) layer of the model with a randomly initialized FC layer. CNN, convolutional neural network.
Sensors 22 06671 g002
Figure 3. Procedure of two-step pre-training. We trained the model with the massive action recognition dataset, initialized the FC layer, and transferred the model to the macro-expression dataset. The pre-trained model was fine-tuned to the micro-expression training set again and then evaluated against the test set to obtain the final performance. CNN, convolutional neural network.
Figure 3. Procedure of two-step pre-training. We trained the model with the massive action recognition dataset, initialized the FC layer, and transferred the model to the macro-expression dataset. The pre-trained model was fine-tuned to the micro-expression training set again and then evaluated against the test set to obtain the final performance. CNN, convolutional neural network.
Sensors 22 06671 g003
Figure 4. Procedure of Décalcomanie augmentation. First, we cut all the frames in half. We flip the cut left and right face frames, respectively, and concatenate the half-frame before flipping it to create new frames. We denote the new frame made of only the left faces and the new frame made of only the right faces as the L and R frames, respectively.
Figure 4. Procedure of Décalcomanie augmentation. First, we cut all the frames in half. We flip the cut left and right face frames, respectively, and concatenate the half-frame before flipping it to create new frames. We denote the new frame made of only the left faces and the new frame made of only the right faces as the L and R frames, respectively.
Sensors 22 06671 g004
Figure 5. Procedure of “shared backbone and multiple losses” version when using LR frames. Each input frame shares the network and obtains each loss for each input frame. The training cost is calculated by combining each loss.
Figure 5. Procedure of “shared backbone and multiple losses” version when using LR frames. Each input frame shares the network and obtains each loss for each input frame. The training cost is calculated by combining each loss.
Sensors 22 06671 g005
Figure 6. Procedure of “fusion with the shared backbone” version when using LR frames. Each input was forwarded to the shared backbone network, and we obtained a k-dimensional vector for each. After concatenating all vectors, we fed-forward the representation vector to calculate the training loss. CNN, convolutional neural network; FC, fully-connected.
Figure 6. Procedure of “fusion with the shared backbone” version when using LR frames. Each input was forwarded to the shared backbone network, and we obtained a k-dimensional vector for each. After concatenating all vectors, we fed-forward the representation vector to calculate the training loss. CNN, convolutional neural network; FC, fully-connected.
Sensors 22 06671 g006
Figure 7. Samples of macro- and micro-expression datasets. CK+ and Oulu-CASIA are macro-expression datasets, and SMIC, CASME2, and SAMM are micro-expression datasets. (a) CK+; (b) Oulu-CASIA; (c) SMIC; (d) CASME2; (e) SAMM.
Figure 7. Samples of macro- and micro-expression datasets. CK+ and Oulu-CASIA are macro-expression datasets, and SMIC, CASME2, and SAMM are micro-expression datasets. (a) CK+; (b) Oulu-CASIA; (c) SMIC; (d) CASME2; (e) SAMM.
Sensors 22 06671 g007
Figure 8. Visualization of N-step pre-training performance (UF1 and UAR). In both cases, the performance of Kinetics → UCF → Macro achieves the best performance (Kin.: Kinetics-400).
Figure 8. Visualization of N-step pre-training performance (UF1 and UAR). In both cases, the performance of Kinetics → UCF → Macro achieves the best performance (Kin.: Kinetics-400).
Sensors 22 06671 g008
Table 1. Experimental results of N-step pre-training (IN: ImageNet, Kinetics: Kinetics-400, UCF: UCF101, Macro: a combination of CK+ and Oulu-CASIA). (a) When training the 3D-ResNet-50 on the SMIC. (b) When training the 3D-ResNet-50 on the SAMM.
Table 1. Experimental results of N-step pre-training (IN: ImageNet, Kinetics: Kinetics-400, UCF: UCF101, Macro: a combination of CK+ and Oulu-CASIA). (a) When training the 3D-ResNet-50 on the SMIC. (b) When training the 3D-ResNet-50 on the SAMM.
(a)(b)
Pre-Trained Step UF1 UAR WAR Pre-Trained Step UF1 UAR WAR
Scratch00.55400.55420.5549scratch00.41750.40910.5515
IN10.51040.55490.5122IN10.38500.40770.4412
Kinetics10.56380.58400.5610Kinetics10.42740.41740.5588
UCF10.57210.57730.5793UCF10.43790.45270.5441
Macro10.59970.59880.6098Macro10.45810.45390.5294
Kinetics → Macro20.60600.60660.5976Kinetics → Macro20.46060.47120.5441
IN → UCF → Macro30.63880.63460.6585IN → UCF → Macro30.46140.47410.5588
Kinetics → Macro → UCF30.58930.58930.5915Kinetics → Macro → UCF30.46230.46370.5515
Kinetics → UCF → Macro30.65610.67330.6463Kinetics → UCF → Macro30.46770.48330.5515
Table 2. Experimental results of Décalcomanie (Shared Backbone and Multiple Losses). (a) When training the 3D-ResNeXt-101 on the SMIC. (b) When training the 3D-ResNeXt-101 on the CASME II. (c) When training the 3D-ResNeXt-101 on the SAMM.
Table 2. Experimental results of Décalcomanie (Shared Backbone and Multiple Losses). (a) When training the 3D-ResNeXt-101 on the SMIC. (b) When training the 3D-ResNeXt-101 on the CASME II. (c) When training the 3D-ResNeXt-101 on the SAMM.
(a)(b)(c)
Train Test UF1 UAR WAR Train Test UF1 UAR WAR Train Test UF1 UAR WAR
OO0.58330.56800.5671OO0.45860.45350.5181OO0.49670.49940.6066
OLR0.59850.59290.6098OLR0.51440.51440.5382OLR0.53330.55180.6177
OL0.58560.57980.6159OL0.49730.49870.5261OL0.51600.53930.5882
OR0.58840.59180.5915OR0.48090.48750.5141OR0.54150.56090.6103
OLRO0.62420.64540.6220OLRO0.55810.55270.6104OLRO0.54130.54190.6103
OLR0.64410.63690.6524OLR0.46070.46720.5783OLR0.47290.46530.5809
OLO0.61140.63520.6098OLO0.51650.52290.5944OLO0.56470.55800.6471
OL0.62250.63260.6220OL0.54320.54220.6104OL0.56990.56680.6471
ORO0.62830.63030.6280ORO0.48090.48820.5622ORO0.52760.52650.6250
OR0.62930.61250.6280OR0.50180.49870.5663OR0.60110.61570.6691
LRO0.61500.61250.6159LRO0.54190.53990.6064LRO0.51690.51730.5956
LR0.62270.63640.6220LR0.49330.49200.5783LR0.50870.49890.6029
Table 3. Experimental For the MLP column, × denotes that a single fully-connected layer is used for the classfier. results of Décalcomanie (Fusion with Shared Backbone). (a) When training the 3D-ResNeXt-101 on the SMIC. (b) When training the 3D-ResNeXt-101 on the CASME II. (c) When training the 3D-ResNeXt-101 on the SAMM. “×” in the MLP column indicates that a single fully-connected layer is used for the classifier.
Table 3. Experimental For the MLP column, × denotes that a single fully-connected layer is used for the classfier. results of Décalcomanie (Fusion with Shared Backbone). (a) When training the 3D-ResNeXt-101 on the SMIC. (b) When training the 3D-ResNeXt-101 on the CASME II. (c) When training the 3D-ResNeXt-101 on the SAMM. “×” in the MLP column indicates that a single fully-connected layer is used for the classifier.
(a)(b)(c)
Train
/Test
MLP UF1 UAR WAR Train
/Test
MLP UF1 UAR WAR Train
/Test
MLP UF1 UAR WAR
O×0.58330.56800.5671O×0.45860.45350.5181O×0.49670.49940.6066
0.55090.55070.55490.42350.43850.51810.47120.47530.5662
OLR×0.63460.64380.6463OLR×0.48340.49740.5588OLR×0.50610.51590.6176
0.60680.60240.62200.48000.47030.55820.50930.52080.6029
OL×0.60600.61770.6098OL×0.46970.45490.5261OL×0.49430.50180.5368
0.61720.61970.61590.44110.45490.52610.46260.51520.5055
OR×0.62280.62610.6280OR×0.46560.47200.5141OR×0.59380.61260.6691
0.58660.59350.59150.45590.47240.49400.48950.49120.5809
LR×0.59170.59920.6037LR×0.45360.43850.5422LR×0.50050.50560.5735
0.59190.58560.60980.46950.46750.51810.50480.54500.5441
Table 4. Comparison between various data augmentation techniques on the SMIC dataset. SBML, Shared backbone and multiple losses; FSB, Fusion with a shared backbone.
Table 4. Comparison between various data augmentation techniques on the SMIC dataset. SBML, Shared backbone and multiple losses; FSB, Fusion with a shared backbone.
MethodUF1UARWAR
RandomRotation0.57680.57810.5976
RandomResize0.57340.57730.5731
RandomHorizontalFlip0.57190.57660.5792
SMOTE [9] (+Rotation, Resize, Flip)0.60630.60060.6220
Décal-SBML (+Rotation, Resize, Flip)0.64410.63690.6524
Décal-FSB (+Rotation, Resize, Flip)0.63460.64380.6463
Table 5. Experimental results when training 3D-ResNeXt-101 using both N-step pre-training and Décalcomanie on the SMIC dataset. (a) Shared Backbone and Multiple Losses. (b) Fusion with a shared backbone. “×” in the MLP column denotes that a single fully-connected layer is used for the classifier.
Table 5. Experimental results when training 3D-ResNeXt-101 using both N-step pre-training and Décalcomanie on the SMIC dataset. (a) Shared Backbone and Multiple Losses. (b) Fusion with a shared backbone. “×” in the MLP column denotes that a single fully-connected layer is used for the classifier.
(a)(b)
Train Test UF1 UAR WAR Train/Test MLP UF1 UAR WAR
Scratch0.58330.56800.5671O×0.58330.56800.5671
OLRO0.79370.79470.7927OLR×0.75660.75540.7561
OLR0.73820.74110.73780.72280.72090.7195
OLO0.73500.75420.7317OL×0.75510.76730.7500
OL0.71560.72260.71340.70430.72620.7012
ORO0.78580.78930.7805OR×0.72620.72270.7317
OR0.77190.78280.76830.75070.75060.7561
LRO0.74600.75480.7439LR×0.74540.74700.7317
LR0.72050.72570.71950.73410.74110.7317
Table 6. Comparison to other methods on the SMIC dataset. The performance of the proposed method achieved the highest value when both N-step pre-training and Décalcomanie were used (font in bold).
Table 6. Comparison to other methods on the SMIC dataset. The performance of the proposed method achieved the highest value when both N-step pre-training and Décalcomanie were used (font in bold).
MethodUF1UARWAR
LBP-TOP [4]--0.4878
STLBP-IP [3]--0.5951
Bi-WOOF [19]0.6200-0.6220
3D flow-based CNN [13]--0.5549
TSCNN [31]0.7236-0.7274
MTMNet [10]0.76800.7440-
SMA-STN [32]0.7683-0.7744
MiNet [33]0.7780-0.7860
Proposed0.79370.79470.7927
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, C.; Hong, J.; Jung, H. N-Step Pre-Training and Décalcomanie Data Augmentation for Micro-Expression Recognition. Sensors 2022, 22, 6671. https://doi.org/10.3390/s22176671

AMA Style

Lee C, Hong J, Jung H. N-Step Pre-Training and Décalcomanie Data Augmentation for Micro-Expression Recognition. Sensors. 2022; 22(17):6671. https://doi.org/10.3390/s22176671

Chicago/Turabian Style

Lee, Chaehyeon, Jiuk Hong, and Heechul Jung. 2022. "N-Step Pre-Training and Décalcomanie Data Augmentation for Micro-Expression Recognition" Sensors 22, no. 17: 6671. https://doi.org/10.3390/s22176671

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop