Next Article in Journal
Theoretical and Experimental Studies on the Signal Propagation in Soil for Wireless Underground Sensor Networks
Next Article in Special Issue
Affective Latent Representation of Acoustic and Lexical Features for Emotion Recognition
Previous Article in Journal
Noise-Resistant CECG Using Novel Capacitive Electrodes
Previous Article in Special Issue
An Adaptive Face Tracker with Application in Yawning Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive 3D Model-Based Facial Expression Synthesis and Pose Frontalization

1
Center for Imaging Media Research, Korea Institute of Science and Technology, Seoul 02792, Korea
2
Department of Smart IT, Hanyang Women’s University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(9), 2578; https://doi.org/10.3390/s20092578
Submission received: 1 April 2020 / Revised: 26 April 2020 / Accepted: 28 April 2020 / Published: 1 May 2020
(This article belongs to the Special Issue Sensor Applications on Emotion Recognition)

Abstract

:
Facial expressions are one of the important non-verbal ways used to understand human emotions during communication. Thus, acquiring and reproducing facial expressions is helpful in analyzing human emotional states. However, owing to complex and subtle facial muscle movements, facial expression modeling from images with face poses is difficult to achieve. To handle this issue, we present a method for acquiring facial expressions from a non-frontal single photograph using a 3D-aided approach. In addition, we propose a contour-fitting method that improves the modeling accuracy by automatically rearranging 3D contour landmarks corresponding to fixed 2D image landmarks. The acquired facial expression input can be parametrically manipulated to create various facial expressions through a blendshape or expression transfer based on the FACS (Facial Action Coding System). To achieve a realistic facial expression synthesis, we propose an exemplar-texture wrinkle synthesis method that extracts and synthesizes appropriate expression wrinkles according to the target expression. To do so, we constructed a wrinkle table of various facial expressions from 400 people. As one of the applications, we proved that the expression-pose synthesis method is suitable for expression-invariant face recognition through a quantitative evaluation, and showed the effectiveness based on a qualitative evaluation. We expect our system to be a benefit to various fields such as face recognition, HCI, and data augmentation for deep learning.

1. Introduction

The human face conveys a range of biological and social characteristics. Specifically, facial expressions are one of the crucial factors in understanding an individual’s emotional state and intentions in human communication, and facial expression recognition (FER) has attracted considerable attention from academia and commercial industries. Although facial expression recognition has clear advantages of being a natural and non-intrusive way to analyze human emotions, humans exhibit non-basic, subtle and rather complex affective states that are not limited to prototypical facial expressions (fear, happiness, surprise, anger, disgust, and sadness) [1,2]. A human face has an extremely complex geometric form [3], and it is not easy to obtain and analyze human expressions quantitatively. In addition, one of the main obstacles to obtaining an accurate facial expression is an extreme facial pose that changes the facial appearance. In fact, such change in facial appearance is greater than the inter-variations among different people. Pose changes are fundamentally caused by rigid three-dimensional (3D) head movements, and handling pose variations in a two-dimensional (2D) image space is therefore an ill-posed problem [4]. To overcome these problems, a prominent method is to use 3D facial information.
Three-dimensional facial reconstruction technologies have contributed greatly to the synthesis of various changes in expression and facial pose captured in un unconstrained environment. Modeling a 3D face with a single RGB camera incurs the lowest cost and has minimal requirements. As mentioned above, a single-view reconstruction is ill-posed [4], and numerous efforts have relied on prior models. By representing 3D faces with a linear combination in 3DMM [5], many 3D reconstruction methods have been developed. Offering a higher shape and texture accuracy using advanced scanning devices, a Basel Face Model [6] is a publicly available database based on a nonrigid interactive closest point [7] method to align 3D scans directly to a UV space. In addition, because a single-mode Gaussian in the underlying 3DMM is unable to represent a real-world distribution, Gaussian Mixture 3DMM can describe various ethnic groups as a mixture of Gaussian subpopulations through a separate mean and common covariance [8]. A multilinear face model [9] supports a decoupled space of combined variations, such as the identities, expressions, and visemes on the facial shape and provides separate control for these attributes. It is, therefore, widely used in face tracking and facial animations [10,11,12,13]. Deep-learning-based approaches [14,15] have recently been studied to overcome the limitations of a linear 3DMM representation power and the need for 3D facial scans for the learning process. Using these promising 3D-based methods, researchers have used a 3D face to correct the input poses toward the frontal view. This enables pose-invariant facial recognition based on a synthesis of novel views and engenders significant improvements in recognition [16,17,18,19].
In addition, it is difficult to naturally change a facial expression in a photograph. In this case, traditional 2D warping techniques cause unnatural results when applied to create significant changes in facial expressions such as a wide opening of the mouth. As with the pose problem, because 2D image warping methods do not correctly account for changes in facial geometry and viewpoint, a number of approaches have been developed to model and animate natural facial expressions in a 3D space [3,20,21]. Although physical-based methods for modeling the muscles, skin, and skull have yielded convincing results [22,23], they have limitations for practical use owing to intensive calculations and a complex structure. In fact, recent state-of-the-art technologies used in facial expression synthesis are closely linked to 3D face capturing and real-time animation techniques. In general, using a linear model, facial expressions can be effectively modeled and adjusted with an approximate range of valid expressions [24,25]. Based on linear models, numerous studies have focused on building expression blendshape rigs following the FACS [26]. By estimating the expression coefficients with the constructed expression blendshapes for an input expression, researchers can create animations of actors and for face puppeteering [10,27,28]. In addition, generative and adversarial networks have become more prevalent for synthesizing various expressions in the wild [29,30].
However, even with the breakthrough of facial modeling technology, there is still room for improvement in facial synthesis and modeling against various changes (in pose and expression) tha severely alter the appearance of the face. Most face modeling approaches take a nearly frontal image as an input or generate a discrete number of expressions determined based on the content of the dataset. To effectively analyze and acquire facial expressions, we propose a method for facial synthesis that simulates the pose and expression from unconstrained 2D face images with 3D face models. First, we reconstruct a 3D face using a multilinear model [11] from an arbitrary image view point and then reposition it to a frontal pose. With this fitting process, we also handle the lack of correspondence between 2D and 3D landmarks. After 3D modeling for an input expression, any new facial expressions of the same identity are generated through a weighted linear combination of the blendshape sets. Next, we cope with the texture loss issue when rendering faces in a frontal view. For a realistic synthesis of a facial appearance, not only geometrical changes in facial pose and expression but also details of the facial texture, such as wrinkles, should be presented on the resulting face accompanied with changes in shape. In this study, we built a facial wrinkle database to derive more natural changes in facial expression by synthesizing the proper wrinkles to match the target expression.
The major contributions of this study can be summarized as follows:
  • We propose a contour-fitting method to fit a 3D model to an input image to manage self-occluded facial regions, such as facial contours owing to a variety of expressions and poses through a 3D landmark reassignment through an iterative fitting process.
  • We propose a method for generating images of natural facial expressions by resolving the problem of inconsistent wrinkles that occur when transferring various facial expressions. We built a wrinkle and patch database, from which wrinkle features comparable to the input expressive faces can be found, and their corresponding wrinkle patches are then synthesized with the transferred expressive facial images.
The rest of this paper is organized as follows: Section 2 describes the proposed methods for acquiring a 3D facial geometry and generating various facial expressions. In Section 3, we detail the face recognition conducted for a quantitative evaluation of the manipulated pose and expression faces, as well as a user study conducted for a qualitative analysis. Finally, some concluding remarks are presented in Section 4.

2. Proposed Method

2.1. 3D Fitting Using a Multilinear Model

To capture a face with a joint space of facial attributes, namely identity and expression, we use a multilinear face model based on a large 3D model [11] following FACS [26]. This facial DB is composed of 150 individuals, each with 47 expressions in 11,510 vertices. We represent the models as a 3D tensor, and each dimension of the tensor denotes the identities, expressions, and vertices, respectively. We apply a higher-order singular value decomposition [31] to factorize the tensor using the following equation:
V = C × 2 w i d t × 3 w e x p t ,
where C is the core tensor, w i d and w e x p are the column vectors of identity and expression parameters, respectively [11]. Therefore, any facial expression of any person (V) can be generated. Figure 1 shows a created mesh.
As described in [11], we can truncate the insignificant components and obtain a reduced model of the dataset as follows:
V = C r e d u c e d × 2 U ^ i d t × 3 U ^ e x p t ,
U ^ i d and U ^ e x p are truncated left singular vectors of the identity-flattening and expression-flattening of the 3D facial matrix, respectively. However, the fitted mesh with the core tensor with small identities does not resemble the input face well. Figure 2 shows the fitted meshes from an input image using the core tensor from different numbers of identities and 47 expressions. As shown, the core tensor in increasing numbers of identities increases the fitting quality. Therefore, we fully use the core tensor with 150 identities of 47 expressions to obtain a better reconstruction.
Fitting a 3D face to a 2D face is a process that involves minimizing the matching errors between the facial landmarks on the 3D model and on a 2D face while estimating camera (rotation, translation, and scale) and model parameters (identity and expression). The 2D facial landmarks are extracted using a Dlib face detector [32] and 3D facial landmarks are manually selected. The Dlib is a modern C++ toolkit including machine learning algorithms and provides functions to detect the face and identifies the facial landmarks automatically. We assume the camera projection is a weak perspective, P. Then, a vertex, v i , of the reference 3D model (V) can be simply projected onto the image space, p i , as follows:
p i = s × P × R × v i + t ,
where s is a scaling factor, R is a global rotation, and t describes a translation in the 2D image space, respectively. Hence, the differences in position between the 2D and 3D landmarks are minimized as follows:
E i = 1 2 s × P × R × ( C × 2 w i d t × 3 w e x p t ) ( i ) + t K ( i ) 2 ,
where K ( i ) is the i-th landmark in the input image. The L-BFGS (Limited-memory Broyden-Fletcher-Goldfarb-Shanno) algorithm [33] is applied in an iterative manner to calculate all unknown parameters (∼200), i.e., the pose (scale, translation, and rotation) and facial attribute ( w i d and w e x p ) within a few seconds. The L-BFGS is an optimization algorithm for solving large problems; it belongs to quasi-newton methods and has the advantage of saving memory since it estimates an approximate Hessian inverse matrix with only a few vectors. The 3D reconstruction of a single image requires approximately 1.2 s on average with 100 test face images (with various expressions). Finally, the texture from the input face is projected onto the fitted mesh. The results are shown in Figure 3.

2.1.1. Pose Fitting Process

In the 3D reconstruction process, it is desirable to find precise positions of 3D landmarks corresponding to 2D facial landmarks because it can directly reduce the matching errors. Unfortunately, in uncontrolled environments, most facial images have pose variations, and they are self-occluded. Therefore, the landmarks on a 2D facial contour may not be entirely consistent with their 3D locations, which results in a failure of accurate 3D model fitting. Figure 4 shows landmark inconsistencies of self-occluded regions during the 3D fitting process from variations in pose.
To handle this problem and achieve a fully automatic and adaptive landmark selection, we present a pose fitting method that uses the distance information between landmarks on the input face and their corresponding 3D landmarks on a 3D shape. First, all 3D facial vertices are projected onto a 2D image plane, and the convex hull of all 3D facial vertices is constructed. Each normal direction of a contour landmark of an input face is then used to find the closest point on the convex hull. When a point contacts the convex hull, we consider it a contact point. In the case of a non-frontal pose, the actual face contour landmarks are occluded by a 2D projection. Thus, the observed landmarks are located along the visible facial shape. For this reason, the corresponding contour features on a 3D model should be visible in the current pose. To confirm the visibility, the normal direction of the mesh of V in the current view is checked. If the normal direction is opposite, the vertex is invisible at that position and we exclude the point from the contour features. Consequently, in the visible vertices, the closest vertex from the contact point is updated as a new contour point.
The proposed contour fitting algorithm is described in Algorithm 1.
Algorithm 1. Contour fitting algorithm
Input: 2D face landmarks K ( i ) ( i = 1 , , 68 ) and 3D mesh V.
Output: Final 3D facial landmarks ( v i ).
1: Set V based on an initial guess.
2: Repeat:
3: Project all vertices of V onto an image plane and construct a convex hull of the projected V.
4: Calculate normal directions of the contour landmarks in the image.
5: Consider a point contacting the convex hull as a contact point.
6: Find the closest vertex of V from the contact point only if the vertex is visible at the current view.
7: Update it as the new vertex of V.
8: Update V with Equation (4).
9: Continue Until: the matching errors are minimized in Equation (4).
The yellow lines in Figure 5b,c represent the facial landmarks and the blue shape is the convex hull of the facial geometry. The red lines in Figure 5b show the contour features projected onto the face contours, and Figure 5c describes the fitted shape. Figure 6 illustrates 3D faces with and without using the proposed fitting method.

2.1.2. Frontalization

The expression transformation should work well when the target face is not nearly in a frontal view with a neutral expression. To frontalize the pose, we align the 3D head model to a front view by taking the inverse transformation matrix, which is estimated during the 3D model fitting. However, as described in Section 2.1, the head rotation engenders invisible facial regions through a 3D to 2D projection. To solve this issue, we borrow the pixels on the opposite side of the invisible facial region and fill in the occluded areas using a seamless blending method [34]. Examples of frontalized faces are shown in Figure 7.

2.2. Facial Expression Generation

As described in Section 2.1, we use FaceWarehouse [11], which follows the FACS [26], which anatomically models facial muscle movements that correspond to an emotion and has been used to guide the construction of the example expression basis. In this study, the 3D face prior is decomposed to obtain the expression-specific row vector ( U e x p in (2)) that acts as a blendshape basis of the database [35,36]. In general, a blendshape model generates a facial movement as a linear combination of numerous facial expressions basis. In contrast to PCA models, the blendshapes have an advantage in that an expression given by different people corresponds to a similar set of basis weights [10,25]. Thus, many facial animation approaches apply this advantage to effectively transfer expressions from a source face to a target face [10,27]. In addition, a multilinear model supports a decomposition of the blendshape basis that gives independent control over the combined facial attributes such as the identity, expression, or viseme [9]. Based on such considerations, we generate various expressions from an arbitrary image input using parametrical weight control with person-specific blendshapes or precalculated expression weights of the existing faces while maintaining the identity in (1). We use the neutral expression basis from the dataset to create a resting facial expression. Figure 8 shows examples of synthesized faces.

2.2.1. Building of Expressive Wrinkle Table

When a generic 3D mesh is fitted to an input image, the facial geometry and texture information that are mapped into each vertex are obtained simultaneously. However, a rendering issue remains because the texture image acquired during the 3D face fitting is used as is. For example, muscle movements on a smiling or frowning face cause wrinkles around the lips and eyes, and even when the input geometry is changed to other expressions such as a neutral expression, these wrinkles remain in the texture. Figure 9 shows the remaining wrinkles in a synthesized face.
To generate realistic facial expressions, we propose a method that derives changes in natural expression by synthesizing facial wrinkles adaptively. To do so, we build the wrinkle and patch tables from the defined wrinkle regions. Facial wrinkles typically appear on both sides of the eyes and mouth, glabella, and forehead, and the forehead wrinkles are excluded owing to occlusions resulting from the hair in this study. We define five wrinkle regions on each facial image, as described in Figure 10, and the image patches are then extracted from the regions.
To extract wrinkle patches and features in identical conditions, our normalization process is conducted on the corrected pose-expression face in account to shape, illumination, rotation, scale, and RGB color data [37]. A neutral expressive image with the same identity that corresponds to an input with other expressions also undergoes a normalization process for selecting the patches in the same wrinkle area. The wrinkle patches and features (obtained from the corrected images), and the patches (from the neutral faces) for each wrinkle, are grouped to configure the expressive wrinkle table. Figure 11 illustrates the construction of the expressive wrinkle table.
We built the facial wrinkle tables from an image with a pixel resolution of 2592 × 1728 using eight facial expressions (neutral, angry, surprised, smiling, disgusted, frowning, sad and screaming) from 400 people ranging from 10 to 80 years in age. Figure 12 shows example expressions in our database.
As shown in Figure 9, the remaining wrinkles in the transformed face induce an unnatural result. To remove these expressive wrinkles from the face, we use neutral image patches in the expressive wrinkle table. We call this procedure expression neutralization, and Figure 13 describes the expression neutralization process.
Specifically, the input wrinkle features are extracted in the regions shown in Figure 10 using the region-specific Gabor filters [37]. For each wrinkle, we find the index and corresponding neutral patches of the most similar wrinkle pattern based on the Euclidean distance between the wrinkle feature vector of an input face and the wrinkle feature vectors in the expressive wrinkle table. These patches are then synthesized directly in the transformed neutral face.

2.2.2. Expression Synthesis with Expressive Wrinkles

If we want to change the smiling input to an expression of disgust, the wrinkles according to this expression, such as the glabella, both sides of the eyes, and nasolabial folds, will need to be synthesized. To do so, we can obtain each wrinkle patch from the disgust wrinkle table that corresponds to the wrinkle feature index found in the smile wrinkle table. Then, wrinkles extracted by the Gabor filters from the patches are blended into the disgusted face image using Poisson editing [34]. Figure 14 illustrates the wrinkle synthesis process.
More specifically, when a face is input, its expression is neutralized after 3D fitting. We then extract the wrinkle features after the normalization process. We found that if people have similar expressive wrinkle shapes within a certain facial area, the probability will be high that similar expressive wrinkles will appear when they are making other expressions. Based on this observation, we search each corresponding wrinkle patch by measuring the feature similarity between wrinkles in the input and DB, and the wrinkle patches are then synthesized. In other words, for each wrinkle, we find the index of the most similar wrinkle pattern based on the Euclidean distance between the wrinkle feature vector of the input face ( f I ) and the wrinkle feature vectors in the table ( f T a b l e ) as follows:
arg min n f ( I m ) f ( T a b l e m , n ) 2
where m is the m-th wrinkle, as shown in Figure 10a and n is the wrinkle index in the table. To remove the wrinkles in the neutralized face, we blend the neutral patches that correspond to the wrinkle features into the neutralized face (Figure 14).
If the facial wrinkles are not properly shown or are weakly formed, an unwanted expression transformation may occur. To solve this problem, all five wrinkle patches can be handled together to find the most similar wrinkle set, the equation of which is as follows:
arg min n m = 1 5 λ m f ( I m ) f ( T a b l e m , n ) 2 ( m = 1 5 λ m = 1 )
where λ m indicates the wrinkle weights.
Figure 15 shows the results of an expressive wrinkle synthesis. Figure 15b describes the target expressive image after the expression neutralization process. Figure 15c shows the synthesized glabella wrinkles and nasolabial folds in the target face. More examples are shown in Figure 16.

3. Experiment

3.1. Comparison Pose Fitting

In this paper, we propose a new adaptive pose fitting method. The contour matching process directly affects the quality of the 3D fitting. To prove that our method performs well for a variety of postures, we compare our method to two of the latest methods: EOS [38] and HPEN [16]. HPEN and EOS are successful PCA-based 3D reconstruction methods from a single image. To reduce the fitting errors due to 2D-3D contour landmarks inconsistency, HPEN assumes that a human face is roughly a cylinder and uses the parallel auxiliary points for horizontal landmark marching on the obscured half of the face. Also, EOS finds the candidate contour vertices that are lied in the image points stored in a kd-tree by a nearest neighbour search.
We exploit 337 human images from Multi-PIE [39] with four poses (0 ° , 15 ° , 30 ° , 45 ° ) and six expressions (neutral, smiling, disgusted, surprised, squinting, and screaming). We project the resulting 3D face onto the input image and calculate the average Euclidean errors of the projected contour and ground-truth points, which are respectively annotated in the input image. More specifically, p i m g ( i ) comprises the ground-truth features and p p r o j ( i ) comprises the projected 3D facial vertices.
p i m g ( i ) = ( x i m g ( 1 ) , y i m g ( 1 ) ) ( x i m g ( n ) , y i m g ( n ) ) , p p r o j ( i ) = ( x p r o j ( 1 ) , y p r o j ( 1 ) ) ( x p r o j ( n ) , y p r o j ( n ) ) .
Next, the errors of the contour features on the input image and of the projected contour features are calculated. First, the inter-ocular distance (D) is measured as the Euclidean distance between the outer corners of the eyes as follows:
D = p i m g ( l e f t e y e c o r n e r ) p i m g ( r i g h t e y e c o r n e r ) .
Then, the average point-to-point Euclidean error is normalized by the inter-ocular distance as follows [40]:
E r r o r = 1 N p i m g ( i ) p p r o j ( i ) D , N = 17 .
The results are shown in Figure 17. The position errors of our method across the poses and expressions are lower than EOS [38]. EOS only calculates the occluded contour vertices on the half of the whole facial contour when the input face is not front. Our fitting method considers the occluded contours as the entire face, not the contour of the half of the face. Since the contour of the faces in moderate poses, changes except the extreme pose where a right or left facial contour is completely visible, the 3D contour landmark must be changed accordingly. The images with smiling, squinting, disgusted, and neutral expressions are more accurately fitted by the proposed method than with HPEN [16]; however, the results of the scream and surprise expressions at 45° are slightly lower than those of HPEN.

3.2. Face Recognition

In this paper, we present pose frontalization and expression correction methods for capturing and synthesizing various facial expressions of an arbitrary input. To demonstrate the effectiveness of our method, we apply a face recognition process between the synthesized images and gallery images. Experiments were conducted on the Multi-PIE face database [39], which is a publicly available dataset from Carnegie Mellon University. The database consists of 750,000 images from 337 subjects captured under various poses, illumination conditions, and expressions. Each subject has six expressions (neutral, smiling, surprised, squinting, disgusted, and screaming) and 15 poses (a yaw angle ranging from 180 ° to 180 ° ) under 19 illuminations. To compute the similarity between the query and gallery images, we extract high-dimensional feature vectors [41] and multi-scale local binary pattern descriptors [42] from the dense facial landmarks. In addition, we compress these features using a PCA to reduce the cost of high computational and storage expenses. A joint Bayesian approach [43] is then applied to compute the similarity score. A face is modeled with the feature representation f as the sum of the inter- and intra-classes from the training data as f = μ + ϵ , where μ is a face id entity and ϵ denotes intra-class variations. The similarity score is computed using a log-likelihood ratio as follows:
E r r o r = l o g P ( f 1 , f 2 | H i n t e r ) P ( f 1 , f 2 | H i n t r a ) .

3.3. Recognition Performance

3.3.1. Performance: Pose Frontalization

In this experiment, we evaluated only the proposed frontalization method. We used 337 subjects and three poses ( 15 ° , 30 ° and 45 ° ) in a frontal illumination with a neutral expression in all sessions. The first 171 subjects were used as the training set. The training images included all frontal and neutral expression images from each of the 171 subjects, and the remaining 166 subjects were used for testing. In the testing, all frontalized images from the proposed frontalization method were used for the probe images. In addition, a single frontal and neutral expression image of each subject from the first session(The Multi-PIE database items were collected in four sessions over the span of five months. Each session has a frontal image with a neutral expression and uniform light. We selected these images in the first session as the gallery images.) was used to compose the gallery sets.
Table 1 shows the performance comparison of the proposed method and the state-of-the-art approach [16]. The performance was measured using the Rank-1 recognition rate. For the frontalization, the recognition of the proposed method is higher than that of the state-of-the-art approach with small variations in pose ( 15 ° and 30 ° ). Nevertheless, our frontalization rates are lower than those of the state-of-the-art method for a large pose ( 45 ° ) owing to incorrect landmarks on the images. This is because the facial boundary features often overlap with the corners of the eyes or the nostril features under large variations in pose, which leads to poor frontalization results with erroneous 3D facial modeling.

3.3.2. Performance: Pose Frontalization and Expression Neutralization

In this experiment, we evaluated the frontalization and expression correction methods based on our proposed pose fitting method. We used 337 subjects with 6 expressions and 4 poses (0 ° , 15 ° , 30 ° and 45 ° ) under frontal illumination conditions for all sessions. During testing, all neutral images frontalized and synthesized by the proposed method were used for the probe images. The other configurations were the same as those described in Section 3.3.1. Table 2 shows the performance of the proposed method. Our method significantly increases the recognition performance at yaw angles of 30 ° and 45 ° compared with the results of the original images. When comparing the performance with the state-of-the-art approach [16], our method achieves higher recognition rates under most of the expressions (smiling, squinting, and disgusted) and poses (0 ° , 15 ° and 30 ° ). However, our performance is slightly lower for one expression (screaming) and a large pose variation ( 45 ° ).

3.3.3. User Study

For a quantitative analysis, we evaluated the performance of the proposed algorithms subjectively with the help of 20 volunteers. The 20 volunteers where graduate students in their 20s and 30s who had no background in the field of face synthesis. The evaluation scale was calculated by MOS (mean of opinion score). We randomly selected images from 14 subjects that are not frontal and synthesized them into various poses and expressions with the proposed method. We separated the results with the subject ID. The input photo and the composite photos are printed it on papers and then each of which was distributed to 20 volunteers. The volunteers scored MOS for the synthetic photos. Specifically, the degree to which the synthetic face appears as the input face ID is evaluated by a score of 1 to 5. If the composite photo looks completely different from the subject ID on the input photo, it is rated as the lowest point of 1, and a score of 5 if it looks very similar. In this experiment, we scored MOS 4.7 for the 20 evaluators.

4. Conclusions

Facial expressions are important for understanding the emotional states and intentions of other people. This paper presented a guide to facilitate an analysis of facial expressions by encoding and manipulating facial expressions based on the FACS parameters. Our system can model the geometric information in a 3D space from a single image using the proposed fitting process to cope with arbitrary expressions and poses. In addition, the acquired input facial expression is parametrically changed into various facial expressions that maintain the identity based on person-specific blendshapes. At a texture level, expression wrinkles suitable for the target expression are selected and synthesized using the proposed wrinkle-adaptive method for a realistic expression synthesis and to avoid a loss of personal characteristics. We demonstrated the plausible transformation from one face to another with different expressions and poses and proved the effectiveness through quantitative and qualitative evaluations. We expect that the proposed approach will be applicable to numerous applications, such as HCI, face recognition, and data augmentation for deep learning. In the future, we plan to conduct challenging research to improve reconstruction accuracy in more extreme poses and expressions while improving resolution on low quality face images such as CCTV images from a distance.

Author Contributions

Conceptualization, Y.-J.H. and I.-J.K.; Methodology, Y.-J.H. and S.E.C.; Software, Y.-J.H.; Validation, G.P.N.; Formal analysis, H.C.; Investigation, J.C.; Resources, I.-J.K.; Data curation, G.P.N. and Y.J.H.; Writing—Original draft preparation, Y.-J.H.; Writing—review and editing, I.-J.K. and H.C.; Visualization, Y.-J.H.; Supervision, I.-J.K.; Project administration, I.-J.K.; Funding acquisition, I.-J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by R&D program for Advanced Integrated-intelligence for IDentification (AIID) through the National Research Foundation of Korea (NRF) funded by Ministry of Science and ICT, grant number (2018M3E3A1057288).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gunes, H.; Schuller, B. Categorical and dimensional affect analysis in continuous input: Current trends and future directions. Image Vis. Comp. 2013, 31, 120–136. [Google Scholar] [CrossRef]
  2. Ekman, P.; Friesen, W.V. Constants across cultures in the face and emotion. J. Pers. Soc. Psychol. 1971, 17, 124–129. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Pighin, F.; Hecker, J.; Lischinski, D.; Szeliski, R.; Salesin, D.H. Synthesizing realistic facial expressions from photographs. In Proceedings of the SIGGRAPH ’98: 25th Annual Conference on Computer Graphics and Interactive Techniques, Orlando, FL, USA, 19–24 July 1998; pp. 75–84. [Google Scholar]
  4. Kemelmacher-Shlizerman, I.; Seitz, S.M. Face reconstruction in the wild. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1746–1753. [Google Scholar]
  5. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In Proceedings of the SIGGRAPH ’99: 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 187–194. [Google Scholar]
  6. Paysan, P.; Knothe, R.; Amberg, B.; Romdhani, S.; Vetter, T. A 3D face model for pose and illumination invariant face recognition. In Proceedings of the IEEE Conference on Advanced Video and Signal Based Surveillance, Genova, Italy, 2–4 September 2009; pp. 296–301. [Google Scholar]
  7. Amberg, B.; Romdhani, S.; Vetter, T. Optimal step nonrigid ICP algorithms for surface registration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  8. Koppen, P.; Feng, Z.H.; Kittler, J.; Awais, M.; Christmas, W.; Wu, X.J.; Yin, H.F. Gaussian mixture 3D morphable face model. Pattern Recognit. 2018, 74, 617–628. [Google Scholar] [CrossRef]
  9. Vlasic, D.; Brand, M.; Pfister, H.; Popović, J. Face transfer with multilinear models. ACM Trans. Graph. 2005, 24, 426–433. [Google Scholar] [CrossRef]
  10. Cao, C.; Weng, Y.; Lin, S.; Zhou, K. 3D shape regression for real-time facial animation. ACM Trans. Graph. 2013, 32, 41:1–41:10. [Google Scholar] [CrossRef]
  11. Cao, C.; Weng, Y.; Zhou, S.; Tong, Y.; Zhou, K. FaceWarehouse: A 3D facial expression database for visual computing. IEEE Trans. Vis. Comput. Graph. 2014, 20, 413–425. [Google Scholar]
  12. Cao, C.; Hou, Q.; Zhou, K. Displaced dynamic expression regression for real-time facial tracking and animation. ACM Trans. Graphics 2014, 33, 1–10. [Google Scholar] [CrossRef]
  13. Shi, F.; Wu, H.T.; Tong, X.; Chai, J. Automatic acquisition of high-fidelity facial performances using monocular videos. ACM Trans. Graph. 2014, 33, 1–13. [Google Scholar] [CrossRef]
  14. Tran, L.; Liu, X. On learning 3D face morphable model from in-the-wild images. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 1. [Google Scholar] [CrossRef] [Green Version]
  15. Gecer, B.; Ploumpis, S.; Kotsia, I.; Zafeiriou, S. GANFIT: Generative adversarial network fitting for high fidelity 3D face reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1155–1164. [Google Scholar]
  16. Zhu, X.; Lei, Z.; Yan, J.; Yi, D.; Li, S.Z. High-fidelity pose and expression normalization for face recognition in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 787–796. [Google Scholar]
  17. Zhao, J.; Cheng, Y.; Xu, Y.; Xiong, L.; Li, J.; Zhao, F.; Jayashree, K.; Pranata, S.; Shen, S.; Xing, J.; et al. Towards pose invariant face recognition in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2207–2216. [Google Scholar]
  18. Cao, J.; Hu, Y.; Zhang, H.; He, R.; Sun, Z. Learning a high fidelity pose invariant model for high-resolution face frontalization. In Proceedings of the Advances in Neural Information Processing Systems 31, Montréal, QC, Canada, 3–8 December 2018; pp. 2867–2877. [Google Scholar]
  19. Shi, L.; Song, X.; Zhang, T.; Zhu, Y. Histogram-based CRC for 3D-aided pose-invariant face recognition. Sensors 2019, 19, 759. [Google Scholar] [CrossRef] [Green Version]
  20. Blanz, V.; Basso, C.; Poggio, T.; Vetter, T. Reanimating faces in images and video. Comput. Graph. Forum 2003, 22, 641–650. [Google Scholar] [CrossRef]
  21. Yang, F.; Bourdev, L.; Shechtman, E.; Wang, J.; Metaxas, D. Facial expression editing in video using a temporally-smooth factorization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 861–868. [Google Scholar]
  22. Lee, Y.; Terzopoulos, D.; Waters, K. Realistic modeling for facial animation. In Proceedings of the SIGGRAPH’95: 22nd Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 6–11 August 1995; pp. 55–62. [Google Scholar]
  23. Terzopoulos, D.; Waters, K. Physically-based facial modelling, analysis, and animation. J. Vis. Comput. Animat. 1990, 1, 73–80. [Google Scholar] [CrossRef]
  24. Chuang, E.; Bregler, C. Performance driven facial animation using blendshape interpolation. Tech. Rep. 2002, 2, 3. [Google Scholar]
  25. Lewis, J.P.; Anjyo, K.; Rhee, T.; Zhang, M.; Pighin, F.H.; Deng, Z. Practice and theory of blendshape facial models. In Proceedings of the Eurographics 2014—State of the Art Reports, Strasbourg, France, 7–11 April 2014; pp. 199–218. [Google Scholar]
  26. Ekman, P.; Friesen, W. Facial Action Coding System: A Technique for the Measurement of Facial Movement; Consulting Psychologist Press: Palo Alto, CA, USA, 1978. [Google Scholar]
  27. Thies, J.; Zollhöfer, M.; Stamminger, M.; Theobalt, C.; Nießner, M. Face2Face: Real-time face capture and reenactment of RGB videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2387–2395. [Google Scholar]
  28. Kim, H.; Garrido, P.; Tewari, A.; Xu, W.; Thies, J.; Niessner, M.; Pérez, P.; Richardt, C.; Zollhöfer, M.; Theobalt, C. Deep video portraits. ACM Trans. Graphics 2018, 37. [Google Scholar] [CrossRef]
  29. Choi, Y.; Choi, M.J.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
  30. Zhou, Y.; Shi, B.E. Photorealistic facial expression synthesis by the conditional difference adversarial autoencoder. In Proceedings of the International Conference on Affective Computing and Intelligent Interaction, San Antonio, TX, USA, 23–26 October 2017; pp. 370–376. [Google Scholar]
  31. De Lathauwer, L.; De Moor, B.; Vandewalle, J. A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl. 2000, 21, 1253–1278. [Google Scholar] [CrossRef] [Green Version]
  32. King, D.E. Dlib-Ml: A machine learning toolkit. J. Mach. Learn. Res. 2009, 10, 1755–1758. [Google Scholar]
  33. Liu, D.C.; Nocedal, J. On the limited memory BFGS method for large scale optimization. Math. Program. Ser. A B 1989, 45, 503–528. [Google Scholar] [CrossRef] [Green Version]
  34. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. ACM Trans. Graph. 2003, 22, 313–318. [Google Scholar] [CrossRef]
  35. Vasilescu, M.A.O.; Terzopoulos, D. Multilinear analysis of image ensembles: TensorFaces. In Proceedings of the European Conference on Computer Vision, Copenhagen, Denmark, 28–31 May 2002; pp. 447–460. [Google Scholar]
  36. Wang, H.; Ahuja, N. Facial expression decomposition. In Proceedings of the IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 958–965. [Google Scholar]
  37. Choi, S.E.; Jo, J.; Lee, S.; Choi, H.; Kim, I.J.; Kim, J. Age face simulation using aging functions on global and local features with residual images. Expert Syst. 2017, 80, 107–125. [Google Scholar] [CrossRef]
  38. Huber, P.; Hu, G.; Tena, R.; Mortazavian, P.; Koppen, W.P.; Christmas, W.J.; Rätsch, M.; Kittler, J. A multiresolution 3D morphable face model and fitting framework. In Proceedings of the Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Rome, Italy, 27–29 February 2016; pp. 79–86. [Google Scholar]
  39. Gross, R.; Matthews, I.; Cohn, J.; Kanade, T.; Baker, S. Multi-PIE. In Proceedings of the 8th IEEE International Conference on Automatic Face Gesture Recognition, Amsterdam, The Netherlands, 17–19 September 2008; pp. 1–8. [Google Scholar]
  40. Cristinacce, D.; Cootes, T.F. Feature detection and tracking with constrained local models. Pattern Recognit. 2006, 41, 929–938. [Google Scholar]
  41. Chen, D.; Cao, X.; Wen, F.; Sun, J. Blessing of dimensionality: High-dimensional feature and its efficient compression for face verification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3025–3032. [Google Scholar]
  42. Ojala, T.; Pietikäinen, M.; Matti, T.; Mäenpää, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  43. Chen, D.; Cao, X.; Wang, L.; Wen, F.; Sun, J. Bayesian face revisited: A joint formulation. In Proceedings of the European Conference on Computer Vision, Firenze, Italy, 7–13 October 2012; pp. 566–579. [Google Scholar]
Figure 1. 3D face representation through the core tensor, identity and expression weights.
Figure 1. 3D face representation through the core tensor, identity and expression weights.
Sensors 20 02578 g001
Figure 2. Fitted shapes with different numbers of identities: (a) Input image. (b) Fitted meshes with 10 identities of 47 expressions. (c) 50 identities of 47 expressions. (d) 100 identities of 47 expressions. (e) 150 identities of 47 expressions.
Figure 2. Fitted shapes with different numbers of identities: (a) Input image. (b) Fitted meshes with 10 identities of 47 expressions. (c) 50 identities of 47 expressions. (d) 100 identities of 47 expressions. (e) 150 identities of 47 expressions.
Sensors 20 02578 g002
Figure 3. Resulting 3D faces: (a) Input images. (b) 3D faces.
Figure 3. Resulting 3D faces: (a) Input images. (b) 3D faces.
Sensors 20 02578 g003
Figure 4. Differences in feature position between 2D and 3D faces with respect to pose changes: (a) Input images. (b) 3D faces. Green dots are 2D landmarks and magenta points are the corresponding 3D feature positions.
Figure 4. Differences in feature position between 2D and 3D faces with respect to pose changes: (a) Input images. (b) 3D faces. Green dots are 2D landmarks and magenta points are the corresponding 3D feature positions.
Sensors 20 02578 g004
Figure 5. Proposed fitting process.
Figure 5. Proposed fitting process.
Sensors 20 02578 g005
Figure 6. 3D reconstruction results with and without the proposed fitting method: (a) Input image. (b) 3D face without proposed method. (c) 3D face with proposed method.
Figure 6. 3D reconstruction results with and without the proposed fitting method: (a) Input image. (b) 3D face without proposed method. (c) 3D face with proposed method.
Sensors 20 02578 g006
Figure 7. Frontalized faces: (a,c) Input images. (b,d) Rendered frontalized faces.
Figure 7. Frontalized faces: (a,c) Input images. (b,d) Rendered frontalized faces.
Sensors 20 02578 g007
Figure 8. Synthesized facial expression results: (a) Input image. (b) Expression of synthesized faces.
Figure 8. Synthesized facial expression results: (a) Input image. (b) Expression of synthesized faces.
Sensors 20 02578 g008
Figure 9. Remaining wrinkles after expression synthesis process: (a) Input image. (b) Expression synthesized image.
Figure 9. Remaining wrinkles after expression synthesis process: (a) Input image. (b) Expression synthesized image.
Sensors 20 02578 g009
Figure 10. Wrinkle set and Gabor filter set: (a) Wrinkle set. (b) Gabor filters for wrinkle set [37].
Figure 10. Wrinkle set and Gabor filter set: (a) Wrinkle set. (b) Gabor filters for wrinkle set [37].
Sensors 20 02578 g010
Figure 11. Process of building expression wrinkle table.
Figure 11. Process of building expression wrinkle table.
Sensors 20 02578 g011
Figure 12. Example facial expressions from out database.
Figure 12. Example facial expressions from out database.
Sensors 20 02578 g012
Figure 13. Expression neutralization process.
Figure 13. Expression neutralization process.
Sensors 20 02578 g013
Figure 14. Wrinkle synthesis process on target facial expression image.
Figure 14. Wrinkle synthesis process on target facial expression image.
Sensors 20 02578 g014
Figure 15. Wrinkle synthesis results: (a) Input face. (b) Target expression after expression neutralization. (c) Synthesized wrinkles in target face. (d) Synthesized wrinkles (red boxes).
Figure 15. Wrinkle synthesis results: (a) Input face. (b) Target expression after expression neutralization. (c) Synthesized wrinkles in target face. (d) Synthesized wrinkles (red boxes).
Sensors 20 02578 g015
Figure 16. Synthesized expression wrinkles in faces: (a,c) Input images. (b,d) Synthesized expression wrinkles in target faces.
Figure 16. Synthesized expression wrinkles in faces: (a,c) Input images. (b,d) Synthesized expression wrinkles in target faces.
Sensors 20 02578 g016
Figure 17. Comparative results on the Multi-PIE dataset [39] for our method and the state-of the-art methods, EOS [38] and HPEN [16].
Figure 17. Comparative results on the Multi-PIE dataset [39] for our method and the state-of the-art methods, EOS [38] and HPEN [16].
Sensors 20 02578 g017
Table 1. Rank-1 recognition rates for Multi-PIE under variations in pose achieved by the proposed frontalization method.
Table 1. Rank-1 recognition rates for Multi-PIE under variations in pose achieved by the proposed frontalization method.
Pose/MethodOriginal ImageHPEN [16]Proposed Method
0 ° 99.2647--
15 ° 97.398597.887598.8435
30 ° 94.832596.77596.817
45 ° 75.544693.157892.5905
Table 2. Rank-1 recognition rates for Multi-PIE in pose frontalization and expression neutralization with combined pose and expression variations. “Smile#” images are from both session 1 and session 3.
Table 2. Rank-1 recognition rates for Multi-PIE in pose frontalization and expression neutralization with combined pose and expression variations. “Smile#” images are from both session 1 and session 3.
Original ImageHPEN [16]Proposed
Smile#SurpriseSquintDisgustScreamSmile#SurpriseSquintDisgustScreamSmile#SurpriseSquintDisgustScream
0 ° 95.915293.210398.692893.918993.403996.410194.155898.039295.23059496.823394.211397.480995.873993.4333
15 ° 91.370189.493896.815292.542492.451693.370190.987296.362491.830493.526394.531091.247597.804793.758192.1842
30 ° 90.068981.300187.666685.013070.265391.222286.755691.461590.207585.751891.910985.564090.367289.280983.6842
45 ° 78.813169.245375.704476.601365.757989.025785.627488.431389.469380.333389.526384.354988.752789.760878.6436

Share and Cite

MDPI and ACS Style

Hong, Y.-J.; Choi, S.E.; Nam, G.P.; Choi, H.; Cho, J.; Kim, I.-J. Adaptive 3D Model-Based Facial Expression Synthesis and Pose Frontalization. Sensors 2020, 20, 2578. https://doi.org/10.3390/s20092578

AMA Style

Hong Y-J, Choi SE, Nam GP, Choi H, Cho J, Kim I-J. Adaptive 3D Model-Based Facial Expression Synthesis and Pose Frontalization. Sensors. 2020; 20(9):2578. https://doi.org/10.3390/s20092578

Chicago/Turabian Style

Hong, Yu-Jin, Sung Eun Choi, Gi Pyo Nam, Heeseung Choi, Junghyun Cho, and Ig-Jae Kim. 2020. "Adaptive 3D Model-Based Facial Expression Synthesis and Pose Frontalization" Sensors 20, no. 9: 2578. https://doi.org/10.3390/s20092578

APA Style

Hong, Y. -J., Choi, S. E., Nam, G. P., Choi, H., Cho, J., & Kim, I. -J. (2020). Adaptive 3D Model-Based Facial Expression Synthesis and Pose Frontalization. Sensors, 20(9), 2578. https://doi.org/10.3390/s20092578

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop