Next Article in Journal
Co3O4/Al-ZnO Nano-composites: Gas Sensing Properties
Next Article in Special Issue
Three-D Wide Faces (3DWF): Facial Landmark Detection and 3D Reconstruction over a New RGB–D Multi-Camera Dataset
Previous Article in Journal
A Novel Method for Early Gear Pitting Fault Diagnosis Using Stacked SAE and GBRBM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Histogram-Based CRC for 3D-Aided Pose-Invariant Face Recognition

1
The School of Computer and Communication Engineering, Jiangsu University, Zhenjiang 212000, China
2
Jiangsu Province Engineering Laboratory of Pattern Recognition and Computational Intelligience, Jiangnan University, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(4), 759; https://doi.org/10.3390/s19040759
Submission received: 16 December 2018 / Revised: 8 February 2019 / Accepted: 9 February 2019 / Published: 13 February 2019
(This article belongs to the Special Issue Sensor Applications on Face Analysis)

Abstract

:
Traditional Collaborative Representation-based Classification algorithms for face recognition (CRC) usually suffer from data uncertainty, especially if it includes various poses and illuminations. To address this issue, in this paper, we design a new CRC method using histogram statistical measurement (H-CRC) combined with a 3D morphable model (3DMM) for pose-invariant face classification. First, we fit a 3DMM to raw images in the dictionary to reconstruct the 3D shapes and textures. The fitting results are used to render numerous virtual samples of 2D images that are frontalized from arbitrary poses. In contrast to other distance-based evaluation algorithms for collaborative (or sparse) representation-based methods, the histogram information of all the generated 2D face images is subsequently exploited. Second, we use a histogram-based metric learning to evaluate the most similar neighbours of the test sample, which aims to obtain ideal result for pose-invariant face recognition using the designed histogram-based 3DMM model and online pruning strategy, forming a unified 3D-aided CRC framework. The proposed method achieves desirable classification results that are conducted on a set of well-known face databases, including ORL, Georgia Tech, FERET, FRGC, PIE and LFW.

1. Introduction

In recent years, with the huge development in computer sciences and informatics, representation-based methods (RBMs) have been given significant attention and garnered considerable achievements [1,2] in pattern recognition. The collaborative representation-based classification (CRC) method [3,4] is considered as the one of the most representative RBMs and has also been successfully applied in various research fields, including feature representation, image classification [5], machine learning and biometrics. Moreover, it has been proven that CRC has potential advantages in the applications of sparse representation [6,7] and low-rank representation [8].
In particular, the basic idea of CRC is to reconstruct a test sample using the dictionary set that includes all training samples across each subject [9]. The category is subsequently obtained by evaluating the collaborative capacity on the query face image across each subject and the test image is assigned to the class that has the minimum reconstruction error. Although CRC can effectively obtain the collaborative capacity from the dataset, it cannot handle the case in which the training samples of the same subject contain variations in illumination, expression and occlusion and lead to the huge difference between each other. In view of this kind of condition, Liu et al. [10] proposed a kernel CRC (KCRC) approach to improve the performance of SRC. A hierarchical collaborative representation-base classification(HCRC) [11] was proposed to extend representation-based method. It was proved that the accuracy of CRC can be improved substantially by minimizing not only the Euclidean distance between a query face and its approximator but also the Euclidean distances from the approximator to training faces in each class. Zhu et al. [12] performed CRC on patches and combined the recognition outputs of all patches to a designed multi-scale patch based collaborative representation for face recognition. Rodriguez [13] developed a discriminative dictionary by means of a framework of sparse representation-based method, which is applied to code the test sample for recognition task. A multiple-kernel sparse representation method for supervised and unsupervised learning was proposed by Thiagarajan et al. [14]. Meanwhile, aim to reduce computational cost in face occlusion problem, Yang [15] studied a Gabor occlusion dictionary. Cheng et al. [16] developed an 1 -graph for image classification, while Qiao et al. [17] constructed a subspace to reserve the 1 -graph for face recognition. In addition, Yanget et al. [18] synthesized a latent dictionary learning with SRC framework for face recognition. In [19], a method of assembling the similarity and distinctiveness of features into CRC was proposed by Yang et al. They also presented a more general model for classification. Liu et al. [20] improved the accuracy of CRC by evaluating the reconstruction error of the test sample. Recently, Xu et al. [21,22,23,24] proposed a series of simple-to-achieve sets of synthesized training samples for the goal of augmenting training set. Due to the fact that the traditional dictionaries built by raw images cannot fully represent the appearance of test samples, especially in the case of pose variations, more recently, deep learning has also been widely used for robust face recognition across pose variations and shown very promising results [25,26,27].
From the above-mentioned studies, we can discover that the ability of RBMs to uncover information of discrimination depends on the process of collaborative (sparse)-constrained signal reconstruction. An effective dictionary must be that it is preserving most important properties of signal with various types. Nevertheless, these algorithms can fail to exploit the valuable properties of signals with various types; this mainly results from redundancy and uncertainty issues, which lead to incorrect classification results. The rationale of the representation-based method can be summarized as follows: it is the fact the test sample with high correlation to training samples from the subjects of a dictionary will be quite reasonable to classify it into the class with high collaborative representation ability [28].
Despite extensive studies on the representation-based classification model, less attention has been focused on the evaluation of the histogram statistical metric information from the synthesized images generated by 3DMM for the purposes of competitive training data selection and pose-invariant face classification. In this paper, we focus on a new CRC method that comprehensively combines the merits from both the 3D morphable face model and the pruning strategy under the histogram statistical model. The proposed method provides three main contributions:
  • We use fitting process of 3DMM to reconstruct the 3D shape and texture for 2D face images. By this means, we can synthesize a lot of virtual 2D face samples whose poses are frontalized from the arbitrary poses.
  • We reduce the redundancy and uncertainty of the synthesized dictionary by designing a pruning strategy. The histogram statistical metric information of all the generated 2D face images is subsequently exploited and evaluated. We then discard the training images with small contributions to represent a test sample.
  • We propose a complete H-CRC framework to perform pose-invariant face recognition by discovering the most representative training samples from the synthesized dictionary using 3DMM generated faces and the histogram-based metric learning.
Compared with the conventional CRC algorithm, the proposed method can be considered as an evaluation and pruning strategy under the H-CRC framework. Among the different variation types of face images, pose-invariant face classification is one of the most challenging tasks. To this end, a great number of algorithms have been explored during the past decades, including deep learning-based methods which can achieve amazing performance with a huge training face dataset and a well-designed CNN model. However, the training of a deep neural network usually require a huge amount of training data and expensive GPU devices. For deep learning-based approaches, we have already made some studies and compared them with the 3DMM-based approach. As shown in our previous studies [29,30], the performance of DNN-based approaches decreases rapidly as the degree of pose variation increases, especially when the head rotation is larger than 30 degree. In addition, for smaller pose variations such as 0 and 15 , 3DMM-based approach also obtained competitive results as compared with DNN-based approaches. In this paper, we explored the use of a 3DMM in generating virtual training samples, as well as design of a histogram-based pruning strategy to achieve the effective statistical information of all the generated 2D face images, for the robust pose-invariant CRC-based face classification. It should be noted that the proposed method is based on Non-deep-learning-based strategy, which might not continuously outperform the deep learning-based method which is driven by a huge training data. But the designed histogram-based CRC using 3DMM in this paper can swiftly address the pose-invariant issue especially in the case of under-sampled face recognition.
The remainder of this paper is organized as follows: Section 2 briefly outlines the CRC method, which serves as a background to the proposed method detailed in Section 3. Section 4 offers analytical elaborations of the proposed method. Section 5 reports comprehensive experimental results obtained on several well-known face databases. Finally, concluding comments are presented in Section 6.

2. Outline of CRC

Input a K × M training samples set { x 1 , 1 , , x K , M } to build a dictionary. Where K denotes the number of classes and M stands for the number of samples in each subject, a test sample y R P can be approximated by the linear combination of all the training samples:
y m = 1 M k = 1 K α k , m x k , m ,
where α k , m is the entry of the coefficient vector corresponding to the mth training sample in the kth class x k , m R P and P is the dimensionality of a training or test sample. The entry α k , m indicates the response of the corresponding gallery set to represent the query sample y . Equation (1) can be compactly re-described as follows:
min α 2 s . t . y = X α .
where X = [ x 1 , 1 , , x K , M ] R P × K M denotes the dictionary matrix which contains all the training samples and α = [ α 1 , 1 , , α K M ] T stands for the coefficient vector that is estimated by solving the 2 -norm minimization problem. The optimization of Equation (2) is a typical least square problem, and α can be achieved by
α = ( X T X + μ I ) 1 X T y ,
where μ and and I respectively denote a small positive constant and the identity matrix.
We can evaluate the propensity of the kth class to represent the test sample after obtaining the coefficient vector, as below:
c k = m = 1 M α k , m x k , m ,
where c k stands for the reconstructed signal of the test sample merely using the linear combination of training samples of the kth class. The reconstruction error for the test sample using the kth class can be obtained by:
E ( y ) k = y c k 2 2 ,
and the label of the test sample y is estimated by
L a b e l ( y ) = argmin k { E ( y ) k } .

3. The Proposed Method

As demonstrated in Section 2, although some improved representation-based classification methods can exploit important characteristics among complicated data sets, capturing certain possible variations in the statistical distribution information in each class space is still nontrivial [31]. Moreover, traditional methods cannot avoid the influence of data uncertainty effectively in a dataset in which the samples from subjects are informative and redundant. The classification performance may become compromised because of this condition. This motivates us to propose a novel histogram-based CRC (H-CRC) for 3D-aided pose-invariant face classification to evaluate the statistical distribution information across each subject. The schematic diagram of the designed algorithm is shown in Figure 1.

3.1. Pose Normalization by Use of 3DMM

First, we perform the pose normalization by means of a 3DMM that consists of a texture and a shape models. The texture and shape models are trained by projecting 3D shape (x, y, z) coordinates and associated texture (r, g, b) values onto two separate PCA spaces. Given an input image, 3DMM can recover the information of the input image, the shape, texture, camera (pose) and illumination parameters, via a fitting process. Once these parameters are estimated, the input face can be rendered in any given virtual view by adjusting the camera parameters. Note that the occluded pixels in the input image can be reconstructed using the estimated texture. For face recognition tasks, pose normalization is performed by converting a face under arbitrary poses to a frontal pose. Once the input image is fitted, pose normalization and frontal face rendering can be achieved by adjusting the camera parameters to a frontal view.
More specifically, a vertex v = [ x 3 d , y 3 d , z 3 d ] T R 3 of a 3D shape is projected onto the 2D coordinates s = [ x 2 d , y 2 d ] T via a camera projection. We can divided the projection process into two parts: a rigid 3D transformation T r : R 3 R 3 and a perspective projection T p : R 3 R 2 :
T r : v = R v + τ ,
T p : s = o x + f v x v z o y f v y v z ,
in above formulation, R R 3 × 3 is the rotation matrix, τ R 3 represents a spatial translation, [ o x , o y ] T denotes the optical axis of the image-plane position, and f is the focal length.
As a consequence, by setting different parameters of camera { R , τ , f } , the face image of any different pose can be rendered from the obtained 3D texture and shape. In this work, we perform frontalization; therefore, we only render an image to the frontal view. As shown in Figure 2, some 2D face images are rendered from an input sample using 3DMM.

3.2. Histogram Measurement for Training Data Optimization

From a probability prospective, the grey level of an input image can be viewed as the frequency of different grey scale pixel values. Thereby, the histogram corresponds to the probability density function p ( r ) , and the probability distribution function P ( r ) represents an integral of the probability density function.
Assume that we have a discrete grey scale image x, and n i denotes the number of occurrences of grey level i. Thus, the probability of a grey level i in the picture can be obtained as follows:
p r ( x ) = p ( r = x ) = n i n ( 0 i < M )
where M denotes the total number of gray levels in the image (typical value is defined as 256), n stands for the total number of pixels of the image, and p r ( i ) is the histogram probability of each image for pixel value i, which is normalized to the range 0 , 1 . The probability density function p r and the probability distribution function P ( r ) can be obtained as:
P ( r ) = 0 r p ( r ) d r
The histogram can be viewed as a one-dimensional column H , in which the rth element h r indicates the number of pixels corresponding to the luminance value.
H = h 0 , h 1 , , h 255
The second step of our method is to calculate the residual of the test sample’s histogram for each training sample by
D k ( y ) = H H j 2 2 ( j = 1 , 2 , , m ) ,
where m is the number of training samples; H j ( j = 1 , 2 , , m ) denotes the corresponding histogram of each training sample x j ( j = 1 , 2 , , m ) ; and H represents the histogram of the test sample. Thus, we can obtain m measurements between the test sample and each training database, that is, the evaluation of the histogram reconstruction is adopted for the whole training set. We then sort the achieved histogram measurements in ascending order as follows:
Ω = ( d 1 , d 2 , , d m )
By Equation (13), we choose the L best histogram measurements with smaller Euclidean distances to optimize the training set. Thus, the competitive training images x 1 , x 2 , , x L that are closely associated with the test image are finally achieved. We then use these training samples that have the best representation capability to reconstruct the query sample and perform effective classification.

3.3. Perform Classification Using Obtained Nearest Neighbors

The third step of the designed H-CRC algorithm is to represent the test sample which can be described as a linear combination of the obtained L nearest neighbours samples. In this phase, the following equation can be approximately satisfied by:
y = β 1 x 1 + + β L x L ,
where x i ( i = 1 , 2 , , L ) stands for the identified L nearest neighbours and β i ( i = 1 , 2 , , L ) denotes the coefficients. Equation (14) can be re-described as
y = X β ,
where β = [ β 1 , , β L ] T and X = [ x 1 , , x L ] . We can solve β using β = X 1 y while X is a nonsingular square matrix; otherwise, the coefficient β can be obtained by β = ( X T X + μ I ) 1 X T y , here, μ denotes a small positive constant and I stands for the identity matrix.
Considering that the nearest neighbours sample of test sample may be obtained from different subjects, we evaluate the sum of the contributions of these neighbours across each class for reconstruction of the test sample, we then assign the test sample to the class label that has the smallest reconstruction error. Concretely, if all the neighbours from the kth class ( k = 1 , , C ) are ξ s , , ξ t , then the sum of the contribution that is measured for representing the test image of the kth subject is given as below:
y k = ω s ξ s + + ω t ξ t
The reconstruction error of y k from y is described as follows:
D k ( y ) = y y k 2 2
Evidently, the above formula allows the reconstruction error between the query sample and each y k to be evaluated in a fair way by measuring y y k 2 2 . We can conclude that a smaller deviation D k implies a greater contribution to representing the query image; and y k could be classified into the class that generates the smallest deviation.

3.4. The Detailed H-CRC Algorithm

The proposed H-CRC algorithm is described as Algorithm 1:
Algorithm 1: The Proposed H-CRC Algorithm.
1. input: A dictionary that contains training set and a test sample y .
2. preprocessing: Perform the pose normalization using a 3D morphable model.
3. fori = 1 to N (the number of test samples) do
4. Calculate the histogram residual of the test sample for each training sample by Equation (12).
5. Choose the best L histogram measurements with the smallest Euclidean distances for the test sample by Equation (13).
6. Represent the test sample using a linear combination of the obtained L nearest neighbors by Equation (14).
7. Evaluate the sum of the contributions of achieved neighbours across each class for representing the test image by Equation (16).
8. Perform the classification that produces the smallest reconstruction error by Equation (17).
End for
Return: The final recognition rate.

4. Analysis of the Proposed Method

In this section, we will discuss about the characteristics, rationale and advantages of proposed method. The basic idea of our method is to develop a histogram-based CRC to generate a better representation of the test sample for 3D-aided pose-invariant face classification.
(1) Histogram measures with pose normalization
One of the key ideas of the proposed method is to perform the pose normalization using a 3DMM for converting a face under arbitrary poses to a frontal pose. According to the pixel distribution regularities for training images, the samples from the same subject should have similar pixel distributions. This motivates us to use a statistical histogram to intuitively describe the pixel-based features of the rendered image.
To demonstrate how the designed algorithm works, Figure 3, Figure 4, Figure 5 and Figure 6 show the histogram difference among several face images evaluated on the ORL face dataset. Specifically, we randomly choose three training samples, A, B and C, to compare with their corresponding histograms. Note that images A and B are chosen from the same subject, sample C is selected from another subject. To demonstrate the merit of the proposed method, sample A represents a frontal face image, and sample B reflects a significant change in pose variations. As shown in Figure 3, Figure 4 and Figure 5, we note that the similarities of the pixel distributions of samples A and B are much lower than those of samples B and C; this easily leads to misclassification errors. Figure 6 shows the histogram of a synthesized image using a 3DMM to convert sample B with a large pose angle to a frontal view. Evidently, samples A and D have high similarity in their pixel distributions. Therefore, the pixel distribution information reflected by the histogram can be used to evaluate faulty training samples that adversely impact the representation-based classification.
(2) The Basis of Histogram Measurement for Optimum Sample Selection
In this method, we use a histogram to evaluate the pixel distribution of each sample for the purpose of training data pruning. In general, the range of the grey scale pixel values of the histogram is from 0 to 255, representing a total of 256 levels on the gray scale. Thus, a description of an image is transformed for grey scale analysis for the corresponding histogram. To judge the similarity of two images, measuring the Euclidean distance between them represents the most favourable method. Generally, the smaller the distance between two samples, the higher the similarity between the two samples.
As discussed in the previous section, the feasibility of employing histogram information is used to analyse the similarity between two samples. In this section, we will further discuss how we use the histogram measurement to evaluate the competitive samples. Specifically, the histogram of each image can be thought of as a one-dimensional column, where the rth element h r indicates the number of pixels corresponding to the luminance value. Thus, the evaluation of the similarity between two images can be easily transformed into the histogram measurement in terms of their variations in the pixel distributions of images. Meanwhile, in this method, an optimum selection scheme is presented for obtaining the competitive training samples.
As shown in Figure 7, the proposed method gradually achieves better classification results on both the FRGC and PIE datasets, especially in the condition that the large portions of the training samples are eliminated. It is noted that the best classification result can be achieved under our method when using a 90% elimination proportion. Compared with other schemes for distance measurements, the histogram-based estimation can achieve a better tradeoff between time consumption and space complexity. The reasons for this are two fold. On the one hand, for a 256-dimensional vector, the complexity of the distance evaluation is tolerable. On the other hand, the statistics of all the pixel scale values can be used to reduce the loss of information when using histograms to perform training data pruning. To this end, we will prove our hypothesis via a specific instance that is detailed in the next section.
(3) A Specific Instance
In this section, we explain the proposed method by an empirical way. We evaluated on an ORL subset that includes the first 5 classes. We use the first 3 face images per subject for training and the remaining ones are used for test. Thus we created a training set of 15 images and a test set with 35 images. The example faces of this specific instance are shown in Figure 8; it is noted that the test sample is selected from the 3rd class of the subset. Using a different method, Figure 9, Figure 10 and Figure 11 indicate the comparison results of reconstruction error which is calculated by use of training samples of the respective class in the subset for representing a test sample.
Specifically, Figure 9 shows an example of a classification ordered by the residual degree between the test sample and each original training subject. We can see that the test sample reserving the minimum reconstruction error is assigned to the label of the 5th class, highlighted in red, which leads to incorrect classification results.
To reveal the nature of the designed method in this paper, the contribution of each subject to represent a query face image is measured in terms of residual. The subjects with the larger reconstruction error are eliminated from the original subset. Figure 10 shows the re-calculation of the reconstruction error of the test image by each remaining class using an elimination scheme based on the Euclidean distance measurement. Compared with the conventional method as shown in Figure 10, the residual of the test image by the 3rd subject is reduced when fewer classes in the original subset is used; however, the error remains slightly higher than that of the 1st class and leads to misclassification results.
The robust classification result is shown in Figure 11. The result is attributed to the histogram-based CRC mode, through which we can select competitive training samples to generate an optimum subset. Figure 11 shows the residual of the test image by each remaining class when using the histogram-based measurement. We note that the test sample with the minimum residual is assigned to the label of the 3rd class, which leads to correct classification results.
Furthermore, to verify how the designed algorithm works, Figure 12 shows the recognition rates of different learning stages of the method, evaluated on a FRGC face subset that includes 100 persons with 30 different face images of each class. In this experiment, the first 10 training images per subject are selected as training samples, and the remaining 20 ones are used for test. As shown in Figure 12, the respective recognition rate is measured in terms of different dictionary learning stages, including conventional CRC, histogram-based CRC, histogram-based-elimination CRC and histogram-based-elimination CRC using 3DMM. From this experimental result, we can see that the joint combination of the histogram measurements and the pose normalization using 3DMM in our method can significantly improve the classification performance. In addition, the proposed method creates a dictionary that is learned from a dynamical optimization process, increasing the capacity of the representation to reconstruct input signals faithfully.
As discussed above, two aspects can give rise to the superiority of the proposed method. First, we generate the histogram-based measurement both from the test samples and the training set. Because the statistical histogram can be used to intuitively describe the pixel-based features of the images, a set of faulty training subjects with a less competitive representation capacity can be eliminated by this histogram-based metric learning. Second, we reconstruct the 3D shape and texture of each image by fitting 3DMM to raw iamges of 2D in the training set; thus, pose normalization and frontal face rendering can be achieved by adjusting the camera parameters to a frontal view.

5. Experimental Results

In this section, the comprehensive experimental results on a set of well-known face datasets are conducted, which includes ORL [32], FERET [33], GT [34], CMU-PIE [35], LFW [36], FRGC [37] and AR [38]. The facial images from these selected datasets were obtained with variations in illumination, expression and pose. Note that the proposed method is performed using the remaining competitive training samples that represent 10% of all samples that were frontalized by 3DMM.
The ORL face database [32] was collected by the Olivetti Research Laboratory in Cambridge. The dataset contains 40 distinct persons, and each subject includes 10 images. Those images were captured at different time period, with varying environment conditions, including facial illuminations, expressions and details. All individual images are in the upright and frontal position. Each image is resized to 56 by 46. Figure 13 shows a few image samples from ORL.
The FERET dataset [33] was created by the US Department of Defense through the DARPA program, which has become a benchmark database for the evaluation of face classification techniques. Our proposed algorithm was evaluated on a subset of FERET, which includes 1400 images of 200 individuals with seven different images of each person. We resized each face image in FERET to 40 by 40. Figure 14 shows a few image samples from FERET.
The Georgia Tech (GT) face dataset [34] was produced by the Georgia Institute of Technology. This database includes 50 individuals which were taken over two or three sessions. All face images across each class in GT were captured by 15 colours with a cluttered background. The images show frontal and/or tilted faces with different illumination, expression and pose. We resized all of these images with different sizes to 40 by 30. Figure 15 a few image samples from GT.
The CMU-PIE dataset [35] contains 41,368 images of 68 persons. All pictures in the dataset include mixed intraclass variations which are introduced based on 3 types of interference (each individual has 43 illuminations, 4 expressions and 13 poses). CMU-PIE has also become a benchmark database to evaluate face classification algorithms. In this paper, we perform our proposed method on a subset of CMU-PIE, including 6800 images of 68 persons with 100 different images (ten poses and ten illuminations) of each class. The images in the subset were resized to 100 by 100. Figure 16 shows a few image samples from CMU-PIE.
LFW [36] is considered as one of the most challenging datasets, which contains 13233 face images of 5749 persons of different gender, ages, etc. This face image dataset was characterized by abundant variations, including pose, illumination and expression variations. Our method is evaluated on a subset of LFW, including 1580 images of 158 persons with 10 different images of each class. Each face image in LFW subset was resized to 64 by 64. Figure 17 shows a few image samples from LFW.
The FRGC version 2 dataset [37] is viewed as a face recognition grand challenge database. It includes controlled and uncontrolled colour face images. The controlled images have good image quality, whereas the uncontrolled ones is with poor quality. All pictures of FRGC were taken under complex backgrounds. In this paper, we use 100 persons for experiments and each person contain 30 different face images. The pictures were resized to 80 by 80. Figure 18 shows a few images of FRGC.
The AR face database [38] contains about 4000 face images of 126 individuals, which consists of the frontal faces with different facial expressions, illuminations and disguises. There are two sessions and each session has 13 face images per subject. Figure 19 shows a few images of AR.

5.1. Results on ORL

We repeated our experiment 20 times and measured the recognition rate of different face classification algorithms on the ORL database. In each round of the experiment, θ ( θ = 2 , 3 , 4 ) training samples per subject were randomly selected for training, and the remaining ones were used for test.
The recognition rate of the proposed method is compared to sparse representation-based classification (SRC) [39], collaborative representation-based classification (CRC) [40], linear regression-based classification (LRC) [41], extended SRC (ESRC) [42], regularized robust coding (RRC) [43], relaxed collaborative representation (RCR) [19], TPTSR [44], SLC-ADL [45], Two-Step LSRC [46], DSRL2 [47] and other classical subspace-based classification methods, including Complete LDA (CLDA) [48], Locality Preserving Projection (LPP) [49], Marginal Fisher Analysis (MFA) [50], Orthogonal Laplacianface (OLPP) [51] and Supervised Orthogonal Discriminant Subspace Projection (SODSP) [52].
Table 1 shows the average recognition rate and standard deviation for each algorithm, regardless of the different numbers of training images involved. From Table 1, we can see that our method achieves much better results than other algorithms in terms of accuracy.

5.2. Results on FERET

For the FERET database, we also repeated our experiments 20 times. In each round of experiment, θ ( θ = 3 , 4 , 5 ) samples per subject were randomly chosen for training, and the remaining ones were used for test. Hence, a training set of 200 × θ samples and a test set with 200 × ( 7 θ ) samples were generated.
We then used the histogram metric strategy to obtain the top 10 % competitive training samples from the synthesized dictionary using 3DMM for classification. The classification result of our method is compared to various sparse or collaborative representation-based algorithms, including LRC [41], CRC [40], ESRC [42], RCR [19], RRC [43], TPTSR [44], Two-Step LSRC [46], SLC-ADL [45], L1LS [53], DALM [54], Homotopy [55], FISTA [56] and DSRL2 [47].
As shown in Table 2, the proposed method consistently achieves much better classification results than the other methods, regardless of the different numbers of training samples involved.

5.3. Results on GT

For the GT database, we also repeated 20 times experiments and made comparison on different algorithms in terms of average recognition accuracy. We randomly select θ ( θ = 3 , 4 , 5 ) samples per subject for training, and used the remaining samples for test. Table 3 show the comparison of various methods, including LRC [41], CRC [40], ESRC [42], RCR [19], RRC [43], TPTSR [44], Two-Step LSRC [46], SLC-ADL [45], L1LS [53], DALM [54], Homotopy [55], FISTA [56] and DSRL2 [47].
From the experimental results of Table 3, we can draw a conclusion that our method achieves the best recognition rates of 68.88%, 72.69% and 79.88% to the corresponding θ ( θ = 3 , 4 , 5 ) training samples, which beats other classical sparse or collaborative representation-based classification methods.

5.4. Results on PIE

For the PIE dataset, the experiments were also repeated 20 times. In each round, θ ( θ = 5 , 10 , 15 ) images per subject were randomly chosen for training, and the remaining ones were used for test. Table 4 shows the recognition rates of different sparse representation-based methods, including LRC [41], CRC [40], ESRC [42], TPTSR [44], Two-Step LSRC [46], SLC-ADL [45] and the proposed method. According to these figures, we can see that the proposed method achieves much better classification performance than those of achieved by other methods across all different sizes of training sets.

5.5. Results on FRGC

For FRGC, we used a similar split as that on the PIE database to create the training and test sets. We demonstrated the experimental results of different methods including LRC [41], CRC [40], ESRC [42], TPTSR [44], Two-Step LSRC [46], SLC-ADL [45] and the proposed method in Table 5. It is noted that we also used same executed strategy as that on the above mentioned datasets and reported the average recognition rate as final result. As shown in Table 5, the proposed method achieves 79.96%, 93.45% and 95.47% recognition rates, regardless of different sizes of training samples involved, which are all better than those achieved by the other traditional methods. This is mainly because the collaborative representation is performed based on the histogram statistical metric information among the arbitrary pose variations.

5.6. Results on LFW

For the LFW database, we randomly selected θ ( θ = 1 , 2 , 3 , 4 ) images per subject as training samples, and used the remaining images for test. We also used same executed strategy as that on the above mentioned datasets. The experimental result of the proposed method is compared to LRC [41], CRC [40], ESRC [42], TPTSR [44], Two-Step LSRC [46] and SLC-ADL [45], which is shown in Table 6.
From Table 6, we can see that the proposed method beats the other algorithms. Noted that the performances achieved by the our method on LFW and FRGC are much better than those on the other databases. This is attributed to the fact that these two datasets contain more huge variations in appearance than the other databases. In this case, the advantage of our method is more evident compared to other algorithms.
The total of experimental results mentioned above demonstrate that the proposed method can achieve more effective and stable recognition accuracy than the other traditional methods, regardless of the numbers of training samples involved.

5.7. Comparison of Computation Time

To indicate the computational efficiency of the proposed method, we measured the performance of different face classification algorithms in terms of computation time, evaluated on ORL using the first 4 images of each subject as the training samples. The running time (in seconds) of our method is compared to the different sparse or collaborative representation-based methods, including CRC, LRC, ESRC, TPTSR, Two-Step LSRC, SLC-ADL, and other typical subspace-based classification methods such as PCA and CLDA. As shown in Figure 20, the proposed method consumes 5.21 s, which represents a better computational performance than other classical sparse or collaborative representation-based methods. Although the running time of our method is slightly higher than that of the classical PCA and CLDA algorithms, the acceptability of the computational complexity of our method in practice is entirely unaffected. Moreover, it should be noted that the computation process of our method is split into two stages. One step generates a number of virtual 2D face images, which are frontalized from the arbitrary pose variations; this step can be performed offline. The other step performs pose-invariant face classification using histogram statistical metric information and is completed online.

5.8. Experiment Comparisons with Some Deep Learning Based Methods

In this section, we compared the proposed method with some state-of-the-art deep models such as NN-CNN [57], VGG-SVM [58] and VGG-3DPD-CRC [4]. The experiments were conducted on some well-known challenging face datasets, including AR, LFW and FRGC. To be more specific, we first applied the VGG-FACE model to all the training and test images for robust facial features extraction, using computation of the forward propagation to obtain the convolution features. To make a comparison, the classical NN, SVM approaches and the recent 3DPD-CRC method have been used. The experiments were conducted on multiple face datasets. In practical applications, deep neural networks are able to extract robust features of images with appearance variations. With such robust features, even using a very simple classifier, such as the SVM classifier, can work well. It is worth noting that the proposed method does not aim at beating deep neural networks for robust feature extraction. In fact, the proposed algorithm should be treated as a more powerful sparse representation based classifier that can be jointly used with deep neural networks. For each subject of different dataset, a single image is randomly selected for training and the remaining ones are used for test. We repeated our experiment 10 times and measured the performance of different face classification algorithms in terms of recognition rate, which are reported in Table 7.
Moreover, in our method, using the histogram descriptor to replay raw images is a common and effective method to reduce data redundancy. If the raw image is not high-dimensional, the histogram with 256 bins is sufficient enough to effectively represent the original data. However, it might lead to a poor classification performance as the data size increases. To verify this notion, we used another partitioned histogram strategy [11] to enhance the face classification accuracy. To be more specific, we divided each image into 25 blocks and counted the histogram of each blocks. The final vector is jointed by all blocks to be the evaluation descriptor of dictionary pruning scheme. The results of SRC, CRC, NN-CNN, VGG-SVM, VGG-3DPD-CRC, the proposed VGG-H1-CRC using classical histogram descriptor and the enhanced version of VGG-H1-CRC, i.e., VGG-H2-CRC using partitioned histogram strategy are shown in Table 7. As shown in Table 7, the proposed VGG-H2-CRC beats all other methods in terms of face recognition accuracy across all different head rotations, including VGG-H1-CRC. This is attributed to the fact that the more robust collaborative representation is performed based on the 3DMM generated faces allied with the partitioned histogram-based pruning scheme.

5.9. Specific Experiment with Various Head Poses

To verify the effectiveness of the proposed method in pose variations, we designed an verified experiment to make sensitivity analysis across various head poses. To be more specific, the proposed H-CRC using VGG-FACE model, two traditional representation-based classification methods, i.e., SRC and CRC, and three deep learning-based methods, i.e., NN-CNN [57], VGG-SVM [58] and VGG-3DPD-CRC [4], were evaluated on a subset of PIE, for the reason that the PIE database contains much more variations in appearance than the others. The subset was chosen by 2992 images of 68 individuals with 11 pose variations, including 0 (pose 06), ±22.5 (pose 05, 07), ±45 (pose 04, 08), ±67.5 (pose 02, 03, 09, 10) and ±90 (pose 01, 11) in yaw rotations, and 4 illumination variations per subject. We used 4 frontal images of each subject to generate the training dictionary and the remaining images were used for test. In the proposed method, we rendered 18 synthesized face images for each example to perform dictionary augmentation. The virtual face images were synthesized from ±10 to ±90 in yaw with the interval of 10 .
The comparison results of these two classical representation-based methods including SRC and CRC, as well as three deep learning-based methods, such as NN-CNN, VGG-SVM, VGG-3DPD-CRC and our VGG-H-CRC algorithm are shown in Table 8. According to this table, we can see that the proposed VGG-H-CRC method performs much better than the others in terms of face classification accuracy across all different head rotations.

6. Conclusions

In this paper, a novel histogram-based CRC for 3D-aided pose-invariant face recognition is developed. The designed method is aim to solve the under-sampled face classification problem including pose-invariant. The proposed method is designed to explore the use of a 3DMM in generating virtual training samples, as well as design of a histogram-based pruning strategy to achieve the effective statistical information of all the generated 2D face images, for the robust pose-invariant CRC-based face classification, which makes CRC robust in terms of face recognition rate while also being highly efficient in terms of computation time. We believe that our promising results can encourage future study on generating more meaningful pixel-based information for collaborative capabilities and better representation-based classification solutions.

Author Contributions

Conceptualization, X.S. and L.S.; methodology, X.S.; software, T.Z.; validation, X.S., T.Z. and Y.Z.; formal analysis, L.S.; investigation, L.S. and X.S.; resources, T.Z.; data curation, T.Z..; writing—original draft preparation, L.S.; writing—review and editing, X.S.; visualization, X.S.; supervision, Y.Z.; project administration, X.S.; funding acquisition, X.S.

Funding

This work was partially supported by the National Key Research and Development Program of China (Grant No. 2017YFC1601800, 2016YFD0401204), the National Natural Science Foundation of China (Grant No. 61876072), the Natural Science Foundation of Jiangsu Province (Grant No. BK20161135), China Postdoctoral Science Foundation (Grant no. 2018T110441, 2016M590407), Six Talent Peaks Project in Jiangsu Province (Grant No. XYDXX-012) and the 111 Project (Grant No. B12018).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, M.; Yang, W.; Jiang, J.; Wu, Y.; Zhang, Y.; Chen, W.; Feng, Q. Brain extraction based on locally linear representation-based classification. Neuroimage 2014, 92, 322–339. [Google Scholar] [CrossRef] [PubMed]
  2. Gao, Y.; Ma, J.; Yuille, A.L. Semi-Supervised Sparse Representation Based Classification for Face Recognition With Insufficient Labeled Samples. IEEE Trans. Image Process. 2017, 26, 2545–2560. [Google Scholar] [CrossRef] [PubMed]
  3. Cheng, Y.; Jin, Z.; Gao, T.; Chen, H.; Kasabov, N. An improved collaborative representation based classification with regularized least square (CRC–RLS) method for robust face recognition. Neurocomputing 2016, 215, 250–259. [Google Scholar] [CrossRef]
  4. Song, X.; Feng, Z.H.; Hu, G.; Kittler, J.; Wu, X.J. Dictionary Integration Using 3D Morphable Face Models for Pose-Invariant Collaborative-Representation-Based Classification. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2734–2745. [Google Scholar] [CrossRef] [Green Version]
  5. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  6. Zhang, Z.; Xu, Y.; Yang, J.; Li, X.; Zhang, D. A Survey of Sparse Representation: Algorithms and Applications. IEEE Access 2015, 3, 490–530. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, J.; Lu, C.; Wang, M.; Li, P.; Yan, S.; Hu, X. Robust Face Recognition via Adaptive Sparse Representation. IEEE Trans. Cybern. 2017, 44, 2368–2378. [Google Scholar] [CrossRef] [PubMed]
  8. Zheng, C.H.; Hou, Y.F.; Zhang, J. Improved sparse representation with low-rank representation for robust face recognition. Neurocomputing 2016, 198, 114–124. [Google Scholar] [CrossRef]
  9. Song, X.; Yang, X.; Shao, C.; Yang, J. Parity symmetrical collaborative representation-based classification for face recognition. Int. J. Mach. Learn. Cybern. 2017, 8, 1–8. [Google Scholar] [CrossRef]
  10. Liu, W.; Lu, L.; Li, H.; Wang, W.; Zou, Y. A novel kernel collaborative representation approach for image classification. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 4241–4245. [Google Scholar]
  11. Vo, D.M.; Lee, S.W. Robust face recognition via hierarchical collaborative representation. Inf. Sci. 2017, 432, 332–346. [Google Scholar] [CrossRef]
  12. Zhu, P.; Zhang, L.; Hu, Q.; Shiu, S.C.K. Multi-scale patch based collaborative representation for face recognition with margin distribution optimization. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012. [Google Scholar]
  13. Rodriguez, F.; Sapiro, G. Sparse Representations for Image Classification: Learning Discriminative and Reconstructive Non-Parametric Dictionaries; Technical Report; University of Minnesota: Minneapolis, MN, USA, 2008. [Google Scholar]
  14. Thiagarajan, J.J.; Ramamurthy, K.N.; Spanias, A. Multiple kernel sparse representations for supervised and unsupervised learning. IEEE Trans. Image Process. 2014, 23, 2905–2915. [Google Scholar] [CrossRef]
  15. Yang, M.; Zhang, L. Gabor Feature Based Sparse Representation for Face Recognition with Gabor Occlusion Dictionary. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 448–461. [Google Scholar]
  16. Cheng, B.; Yang, J.; Yan, S.; Fu, Y. Learning with 1-graph for image analysis. IEEE Trans. Image Process. 2010, 19, 858–866. [Google Scholar] [CrossRef] [PubMed]
  17. Qiao, L.; Chen, S.; Tan, X. Sparsity preserving projections with applications to face recognition. Pattern Recognit. 2010, 43, 331–341. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, M.; Dai, D.; Shen, L.; Gool, L.V. Latent Dictionary Learning for Sparse Representation Based Classification. In Proceedings of the Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4138–4145. [Google Scholar]
  19. Yang, M.; Zhang, L.; Zhang, D.; Wang, S. Relaxed collaborative representation for pattern classification. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 2224–2231. [Google Scholar]
  20. Liu, Z.; Pu, J.; Huang, T.; Qiu, Y. A novel classification method for palmprint recognition based on reconstruction error and normalized distance. Appl. Intell. 2013, 39, 307–314. [Google Scholar] [CrossRef]
  21. Xu, Y.; Zhang, B.; Zhong, Z. Multiple representations and sparse representation for image classification. Pattern Recognit. Lett. 2015, 68, 9–14. [Google Scholar] [CrossRef]
  22. Xu, Y.; Li, Z.; Yang, J.; Zhang, D. A Survey of Dictionary Learning Algorithms for Face Recognition. IEEE Access 2017, 5, 8502–8514. [Google Scholar] [CrossRef]
  23. Xu, Y.; Li, Z.; Zhang, B.; Yang, J.; You, J. Sample diversity, representation effectiveness and robust dictionary learning for face recognition. Inf. Sci. 2017, 375, 171–182. [Google Scholar] [CrossRef]
  24. Xu, Y.; Zhang, Z.; Lu, G.; Yang, J. Approximately symmetrical face images for image preprocessing in face recognition and sparse representation based classification. Pattern Recognit. 2016, 54, 68–82. [Google Scholar] [CrossRef]
  25. Zhu, Z.; Ping, L.; Wang, X.; Tang, X. Deep Learning Identity-Preserving Face Space. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013. [Google Scholar]
  26. Zhu, Z.; Ping, L.; Wang, X.; Tang, X. Multi-view perceptron: A deep model for learning face identity and view representations. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  27. Yim, J.; Jung, H.; Yoo, B.I.; Choi, C.; Park, D.; Kim, J. Rotating your face using multi-task deep neural network. In Proceedings of the Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  28. Song, X.; Feng, Z.H.; Hu, G.; Yang, X.; Yang, J.; Qi, Y. Progressive sparse representation-based classification using local discrete cosine transform evaluation for image recognition. J. Electron. Imaging 2015, 24, 053010. [Google Scholar] [CrossRef] [Green Version]
  29. Hu, G.; Fei, Y.; Chan, C.H.; Deng, W.; Christmas, W.; Kittler, J.; Robertson, N.M. Face Recognition Using a Unified 3D Morphable Model. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016; Volume 9912. [Google Scholar]
  30. Koppen, P.; Feng, Z.-H.; Kittler, J.; Awais, M.; Christmas, W.; Wu, X.-J.; Yin, H.-F. Gaussian Mixture 3D Morphable face model. Pattern Recognit. 2018, 74, 617–628. [Google Scholar] [CrossRef]
  31. Song, X.; Feng, Z.H.; Hu, G.; Wu, X.J. Half-Face Dictionary Integration for Representation-Based Classification. IEEE Trans. Cybernet. 2015, 47, 142–152. [Google Scholar] [CrossRef]
  32. Samaria, F.S.; Harter, A.C. Parameterisation of a stochastic model for human face identification. In Proceedings of the Second IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 5–7 December 1994; pp. 138–142. [Google Scholar]
  33. Phillips, P.J.; Moon, H.; Rizvi, S.A.; Rauss, P.J. The FERET evaluation methodology for face-recognition algorithms. PAMI 2000, 22, 1090–1104. [Google Scholar] [CrossRef] [Green Version]
  34. Nefian, A. Georgia Tech Face Database. Available online: http://www.cl.cam.ac.uk/Research/DTG/attarchive:pub/data/att_faces.zip (accessed on 1 March 2013).
  35. Sim, T.; Baker, S.; Bsat, M. The CMU pose, illumination, and expression database. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1615–1618. [Google Scholar] [Green Version]
  36. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled Faces in the Wild: A Database for Studying Face Recognition in Unconstrained Environments. In Proceedings of the Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition, Marseille, France, 17 October 2008. [Google Scholar]
  37. Phillips, P.J.; Flynn, P.J.; Scruggs, T.; Bowyer, K.W.; Chang, J.; Hoffman, K.; Marques, J.; Min, J.; Worek, W. Overview of the Face Recognition Grand Challenge. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 947–954. [Google Scholar]
  38. AR Database. Available online: http://cobweb.ecn.purdue.edu/aleix/aleix-face-DB.html (accessed on 10 September 2009).
  39. Wright, J.; Yang, A.Y.; Ganesh, A.; Sastry, S.S.; Ma, Y. Robust face recognition via sparse representation. PAMI 2009, 31, 210–227. [Google Scholar] [CrossRef] [PubMed]
  40. Zhang, L.; Yang, M.; Feng, X.; Ma, Y.; Zhang, D. Collaborative representation based classification for face recognition. arXiv, 2012; arXiv:1204.2358. [Google Scholar]
  41. Naseem, I.; Togneri, R.; Bennamoun, M. Linear regression for face recognition. PAMI 2010, 32, 2106–2112. [Google Scholar] [CrossRef] [PubMed]
  42. Deng, W.; Hu, J.; Guo, J. Extended SRC: Undersampled face recognition via intraclass variant dictionary. PAMI 2012, 34, 1864–1870. [Google Scholar] [CrossRef] [PubMed]
  43. Yang, M.; Zhang, L.; Yang, J.; Zhang, D. Regularized robust coding for face recognition. IEEE Trans. Image Process. 2013, 22, 1753–1766. [Google Scholar] [CrossRef] [PubMed]
  44. Xu, Y.; Zhang, D.; Yang, J.; Yang, J.Y. A two-phase test sample sparse representation method for use with face recognition. IEEE Trans. Circuits Syst. Video Technol. 2011, 21, 1255–1262. [Google Scholar]
  45. Wang, J.; Guo, Y.; Guo, J.; Li, M.; Kong, X. Synthesis linear classifier based analysis dictionary learning for pattern classification. Neurocomputing 2017, 238, 103–113. [Google Scholar] [CrossRef]
  46. Shao, C.; Song, X.; Feng, Z.H.; Wu, X.J.; Zheng, Y. Dynamic dictionary optimization for sparse-representation-based face classification using local difference images. Inf. Sci. 2017, 393, 1–14. [Google Scholar] [CrossRef] [Green Version]
  47. Xu, Y.; Zhong, Z.; Yang, J.; You, J.; Zhang, D. A New Discriminative Sparse Representation Method for Robust Face Recognition via l2 Regularization. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2233–2242. [Google Scholar] [CrossRef]
  48. Lu, G.F.; Zou, J.; Wang, Y. Incremental complete LDA for face recognition. Pattern Recognit. 2012, 45, 2510–2521. [Google Scholar] [CrossRef]
  49. He, X.; Yan, S.; Hu, Y.; Niyogi, P.; Zhang, H.J. Face recognition using Laplacianfaces. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 328–340. [Google Scholar] [PubMed] [Green Version]
  50. Yan, S.; Xu, D.; Zhang, B.; Zhang, H.J.; Yang, Q.; Lin, S. Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 40–51. [Google Scholar] [CrossRef] [PubMed]
  51. Cai, D.; He, X.; Han, J.; Zhang, H.J. Orthogonal laplacianfaces for face recognition. IEEE Trans. Image Process. 2006, 15, 3608–3614. [Google Scholar] [CrossRef]
  52. Chen, Y.; Xu, X.H. Supervised orthogonal discriminant subspace projects learning for face recognition. Neural Netw. 2014, 50, 33–46. [Google Scholar] [CrossRef] [PubMed]
  53. Koh, K.; Kim, S.; Boyd, S.; Lin, Y. L1 Ls: A Simple MATLAB Solver for ℓ1-Regularized Least Squares Problems. 2007. Available online: http://www. stanford.edu/boyd/l1_ls (accessed on 4 November 2008).
  54. Yang, A.Y.; Zhou, Z.; Balasubramanian, A.G.; Sastry, S.S.; Ma, Y. Fast 1-Minimization Algorithms for Robust Face Recognition. IEEE Trans. Image Process. 2013, 22, 3234–3246. [Google Scholar] [CrossRef]
  55. Yang, A.Y.; Sastry, S.S.; Ganesh, A.; Ma, Y. Fast 1-minimization algorithms and an application in robust face recognition: A review. In Proceedings of the 17th IEEE International Conference on Image Processing (ICIP), Hong Kong, China, 12–15 September 2010. [Google Scholar]
  56. Beck, A.; Teboulle, M. A Fast Iterative Shrinkage—Thresholding Algorithm for Linear Inverse Problems. Siam J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  57. Wu, S.H.; Zhong, Y.W.; Yao, J. 2, 3-Di (2-pyridyl)-5-phenylpyrazine: A NN-CNN-Type Bridging Ligand for Dinuclear Transition-Metal Complexes. Chem. Asian J. 2013, 8, 1504–1513. [Google Scholar] [CrossRef]
  58. Vijayalakshmi, A.; Rajesh Kanna, B. Deep learning approach to detect malaria from microscopic images. In Multimedia Tools and Applications; Springer: New York, NY, USA, 2019; pp. 1–21. [Google Scholar]
Figure 1. The schematic diagram of the proposed method.
Figure 1. The schematic diagram of the proposed method.
Sensors 19 00759 g001
Figure 2. Some 2D face images rendered from input face images using a 3D morphable model (3DMM).
Figure 2. Some 2D face images rendered from input face images using a 3D morphable model (3DMM).
Sensors 19 00759 g002
Figure 3. Sample A and its histogram.
Figure 3. Sample A and its histogram.
Sensors 19 00759 g003
Figure 4. Sample B and its histogram.
Figure 4. Sample B and its histogram.
Sensors 19 00759 g004
Figure 5. Sample C and its histogram.
Figure 5. Sample C and its histogram.
Sensors 19 00759 g005
Figure 6. Sample D and its histogram.
Figure 6. Sample D and its histogram.
Sensors 19 00759 g006
Figure 7. Classification results with different elimination scales of the training samples.
Figure 7. Classification results with different elimination scales of the training samples.
Sensors 19 00759 g007
Figure 8. Some example faces of the ORL subset.
Figure 8. Some example faces of the ORL subset.
Sensors 19 00759 g008
Figure 9. Reconstruction error of a test image using the original ORL subset, which leads to incorrect classification results.
Figure 9. Reconstruction error of a test image using the original ORL subset, which leads to incorrect classification results.
Sensors 19 00759 g009
Figure 10. Reconstruction error of the test sample by each remaining class using an elimination scheme based on the Euclidean distance measurement, which still leads to incorrect classification results.
Figure 10. Reconstruction error of the test sample by each remaining class using an elimination scheme based on the Euclidean distance measurement, which still leads to incorrect classification results.
Sensors 19 00759 g010
Figure 11. Histogram residual between the test sample and each remaining class obtained by histogram-based measurements. This shows that the test sample with the minimum reconstruction error is assigned to the label of the 3rd class, which leads to correct classification results.
Figure 11. Histogram residual between the test sample and each remaining class obtained by histogram-based measurements. This shows that the test sample with the minimum reconstruction error is assigned to the label of the 3rd class, which leads to correct classification results.
Sensors 19 00759 g011
Figure 12. Different learning stages, denoted as Stage1: conventional collaborative representation-based classification (CRC), Stage2: histogram-based CRC, Stage3: histogram-based-elimination CRC and Stage4: histogram-based-elimination CRC using 3DMM.
Figure 12. Different learning stages, denoted as Stage1: conventional collaborative representation-based classification (CRC), Stage2: histogram-based CRC, Stage3: histogram-based-elimination CRC and Stage4: histogram-based-elimination CRC using 3DMM.
Sensors 19 00759 g012
Figure 13. Sample images from the ORL face database.
Figure 13. Sample images from the ORL face database.
Sensors 19 00759 g013
Figure 14. Sample images from the FERET face image database.
Figure 14. Sample images from the FERET face image database.
Sensors 19 00759 g014
Figure 15. Sample images from the Georgia Tech face database.
Figure 15. Sample images from the Georgia Tech face database.
Sensors 19 00759 g015
Figure 16. Sample images from the CMU-PIE face image database.
Figure 16. Sample images from the CMU-PIE face image database.
Sensors 19 00759 g016
Figure 17. Sample images from the LFW face image database.
Figure 17. Sample images from the LFW face image database.
Sensors 19 00759 g017
Figure 18. Sample images from the FRGC face image database.
Figure 18. Sample images from the FRGC face image database.
Sensors 19 00759 g018
Figure 19. Sample images from the AR face image database.
Figure 19. Sample images from the AR face image database.
Sensors 19 00759 g019
Figure 20. Running time of different methods.
Figure 20. Running time of different methods.
Sensors 19 00759 g020
Table 1. Face recognition rates (%) of different methods on ORL database.
Table 1. Face recognition rates (%) of different methods on ORL database.
MethodNumber of Training Samples
234
CLDA [48]79.47 ± 3.1183.39 ± 1.5786.75 ± 2.42
LPP [49]80.93 ± 2.7187.78 ± 2.1691.33 ± 1.42
MFA [50]83.72 ± 3.2791.42 ± 1.8994.88 ± 0.98
OLPP [51]84.31 ± 3.1791.82 ± 2.0995.50 ± 1.44
SODSP [52]84.75 ± 3.0692.61 ± 1.4895.91 ± 0.93
CRC [40]84.16 ± 2.4790.65 ± 1.6092.73 ± 0.93
LRC [41]84.61 ± 2.0890.29 ± 1.7593.47 ± 1.22
SRC [39]82.62 ± 2.0788.74 ± 1.7291.02 ± 1.39
ESRC [42]83.08 ± 2.1988.14 ± 1.9493.08 ± 1.82
TPTSR [44]83.12 ± 2.2288.57 ± 1.6693.75 ± 1.42
SLC-ADL [45]71.87 ± 2.8477.14 ± 2.1883.33 ± 1.77
Two-Step LSRC [46]85.32 ± 2.4189.84 ± 2.0694.18 ± 2.40
DSRL2 [47]86.21 ± 2.1389.71 ± 1.8694.28 ± 1.47
Our method95.41 ± 2.0598.14 ± 1.1999.17 ± 0.85
Table 2. Face recognition rates (%) of different methods on FERET database.
Table 2. Face recognition rates (%) of different methods on FERET database.
MethodNumber of Training Samples
345
CRC [40]50.41 ± 2.7158.40 ± 2.5764.50 ± 2.43
LRC [41]69.78 ± 2.6077.60 ± 2.4681.12 ± 2.03
ESRC [42]54.13 ± 2.5671.33 ± 2.7176.50 ± 2.36
RRC [43]42.93 ± 2.8453.74 ± 2.8270.21 ± 2.63
RCR [19]45.12 ± 2.7051.02 ± 2.6559.82 ± 2.48
TPTSR [44]57.25 ± 2.5365.12 ± 2.4978.66 ± 2.15
SLC-ADL [45]49.75 ± 3.0668.33 ± 3.4873.75 ± 2.93
Two-Step LSRC [46]58.75 ± 2.6777.13 ± 2.6179.24 ± 2.90
L1LS [53]59.43 ± 2.8276.27 ± 2.7782.63 ± 2.29
Homotopy [55]54.14 ± 2.6672.67 ± 2.5677.45 ± 2.42
DALM [54]59.79 ± 2.9475.65 ± 2.8080.91 ± 2.69
FISTA [56]38.90 ± 2.9950.54 ± 2.9658.95 ± 2.92
DSRL2 [47]61.69 ± 2.7379.21 ± 2.4681.16 ± 2.47
Our method86.91 ± 2.2588.12 ± 2.4188.77 ± 2.67
Table 3. Face recognition rates (%) of different methods on GT database.
Table 3. Face recognition rates (%) of different methods on GT database.
MethodNumber of Training Samples
345
CRC [40]46.62 ± 2.8148.51 ± 2.7651.75 ± 2.41
LRC [41]53.53 ± 2.1457.20 ± 2.5760.22 ± 1.92
ESRC [42]47.67 ± 2.0251.64 ± 2.2953.85 ± 2.37
RRC [43]44.13 ± 2.8243.44 ± 2.8945.70 ± 2.74
RCR [19]36.25 ± 2.8737.89 ± 2.8141.20 ± 2.84
TPTSR [44]58.50 ± 2.6265.82 ± 2.5175.82 ± 2.15
SLC-ADL [45]41.53 ± 3.1349.09 ± 2.4852.83 ± 3.04
Two-Step LSRC [46]59.16 ± 2.1867.81 ± 1.9776.08 ± 1.93
L1LS [53]51.40 ± 2.9052.47 ± 2.7961.66 ± 2.95
Homotopy [55]45.74 ± 2.7249.64 ± 2.7652.46 ± 2.83
DALM [54]52.75 ± 2.9157.61 ± 2.8461.30 ± 2.68
FISTA [56]48.93 ± 3.0251.99 ± 2.9353.05 ± 2.99
DSRL2 [47]54.12 ± 2.7856.66 ± 2.7261.82 ± 2.61
Our method68.88 ± 2.7772.69 ± 2.8079.88 ± 2.11
Table 4. Face recognition rates ( % ) of different methods on PIE database.
Table 4. Face recognition rates ( % ) of different methods on PIE database.
MethodNumber of Training Samples
51015
CRC [40]31.8643.5352.94
LRC [41]31.0849.4152.65
ESRC [42]33.2950.9254.76
TPTSR [44]59.9064.8574.41
SLC-ADL [45]28.8243.0245.29
Two-Step LSRC [46]62.5567.7579.14
Our method82.3591.2993.82
Table 5. Face recognition rates (%) of different methods on FRGC database.
Table 5. Face recognition rates (%) of different methods on FRGC database.
MethodNumber of Training Samples
51015
CRC [40]62.3676.6080.80
LRC [41]40.5254.3562.73
ESRC [42]66.5479.0782.93
TPTSR [44]63.9277.0079.50
SLC-ADL [45]36.2948.5359.91
Two-Step LSRC [46]67.9581.1885.88
Our method79.9693.4595.47
Table 6. Face recognition rates (%) of different methods on LFW database.
Table 6. Face recognition rates (%) of different methods on LFW database.
MethodNumber of Training Samples
1234
CRC [40]8.3411.5814.0315.63
LRC [41]4.677.8710.8913.01
ESRC [42]9.0614.1617.2319.97
TPTSR [44]10.0917.7919.5923.81
SLC-ADL [45]3.867.6910.8413.82
Two-Step LSRC [46]11.1718.3521.4225.40
Our method20.1834.1344.4152.53
Table 7. Face recognition rates (%) of deep learning-based approaches on different datasets.
Table 7. Face recognition rates (%) of deep learning-based approaches on different datasets.
DatasetNN-CNNVGG-SVMVGG-3DPD-CRCVGG-H1-CRCVGG-H2-CRC
AR97.2997.6398.8699.1299.30
LFW54.1154.9260.2462.7364.22
FRGC91.4391.8493.9295.0896.13
Table 8. Face recognition rates (%) with various head poses.
Table 8. Face recognition rates (%) with various head poses.
Pose Number01020304050708091011
Method
SRC254718227475
CRC355718218584
NN-CNN11133327464139322928
VGG-SVM13153542304842353132
VGG-3DPD-CRC33345849695951413839
VGG-H-CRC35376052736354454244

Share and Cite

MDPI and ACS Style

Shi, L.; Song, X.; Zhang, T.; Zhu, Y. Histogram-Based CRC for 3D-Aided Pose-Invariant Face Recognition. Sensors 2019, 19, 759. https://doi.org/10.3390/s19040759

AMA Style

Shi L, Song X, Zhang T, Zhu Y. Histogram-Based CRC for 3D-Aided Pose-Invariant Face Recognition. Sensors. 2019; 19(4):759. https://doi.org/10.3390/s19040759

Chicago/Turabian Style

Shi, Liang, Xiaoning Song, Tao Zhang, and Yuquan Zhu. 2019. "Histogram-Based CRC for 3D-Aided Pose-Invariant Face Recognition" Sensors 19, no. 4: 759. https://doi.org/10.3390/s19040759

APA Style

Shi, L., Song, X., Zhang, T., & Zhu, Y. (2019). Histogram-Based CRC for 3D-Aided Pose-Invariant Face Recognition. Sensors, 19(4), 759. https://doi.org/10.3390/s19040759

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop