Next Article in Journal
Independent Random Recurrent Neural Networks for Infrared Spatial Point Targets Classification
Next Article in Special Issue
Feature Extraction with Discrete Non-Separable Shearlet Transform and Its Application to Surface Inspection of Continuous Casting Slabs
Previous Article in Journal
Retrieval of Chemical Oxygen Demand through Modified Capsule Network Based on Hyperspectral Data
Previous Article in Special Issue
Determination of the Optimal State of Dough Fermentation in Bread Production by Using Optical Sensors and Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Sensor Face Registration Based on Global and Local Structures

1
Hebei Engineering Research Center for Assembly and Inspection Robot, North China Institute of Aerospace Engineering, Langfang 065000, China
2
Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing 100192, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2019, 9(21), 4623; https://doi.org/10.3390/app9214623
Submission received: 10 September 2019 / Revised: 27 October 2019 / Accepted: 28 October 2019 / Published: 30 October 2019
(This article belongs to the Special Issue Intelligent Processing on Image and Optical Information)

Abstract

:
The work reported in this paper aims at utilizing the global geometrical relationship and local shape feature to register multi-spectral images for fusion-based face recognition. We first propose a multi-spectral face images registration method based on both global and local structures of feature point sets. In order to combine the global geometrical relationship and local shape feature in a new Student’s t Mixture probabilistic model framework. On the one hand, we use inner-distance shape context as the local shape descriptors of feature point sets. On the other hand, we formulate the feature point sets registration of the multi-spectral face images as the Student’s t Mixture probabilistic model estimation, and local shape descriptors are used to replace the mixing proportions of the prior Student’s t Mixture Model. Furthermore, in order to improve the anti-interference performance of face recognition techniques, a guided filtering and gradient preserving image fusion strategy is used to fuse the registered multi-spectral face image. It can make the multi-spectral fusion image hold more apparent details of the visible image and thermal radiation information of the infrared image. Subjective and objective registration experiments are conducted with manual selected landmarks and real multi-spectral face images. The qualitative and quantitative comparisons with the state-of-the-art methods demonstrate the accuracy and robustness of our proposed method in solving the multi-spectral face image registration problem.

1. Introduction

Image fusion can often analyze and extract the complementary information of multi-sensor data, and compose a robust or informative image, which can provide more complex and detailed target scene representation [1,2,3]. Since the fusion process of multi-sensor images is in accordance with human visual perception, it has an important research significance for object detection and target recognition in the areas of remote sensing image, medical imaging analysis, military target detection and video surveillance [4,5,6]. E.g., face recognition is one of the major applications of multi-sensor data, such as visible and thermal infrared images. Visible images can provide face texture details with high spatial resolution, while thermal infrared images are not likely to be interfered with by illumination variation or face disguise with high thermal contrast. Therefore, it is beneficial for face recognition to fuse the multi-sensor data, which can combine the advantages of texture detail and thermal radiation information in the multi-spectral face images.
However, image registration is a prerequisite for the success of multi-spectral image fusion, which is an essential and challenging step in the process of image fusion research [7]. Currently, multi-spectral images registration can generally be implemented in two ways: Hardware-based registration and software-based registration. In order to ensure that the multi-spectral images are strictly geometrically aligned, hardware-based registration can be realized based on the coaxial catadioptric optical system through the beam splitter [8]. But the sophisticated and cumbersome imaging sensor equipment for generating co-registered image pairs may not be practical in many face recognition scenarios, due to its high cost and low availability. In contrast, software-based registration can capture dual band images simultaneously and independently by utilizing different off-the-shelf low cost wavebands sensors. Since no additional hardware is needed, it may be more appropriate for practical detection and recognition scenarios compared to the hardware-based registration [9]. In general, software-based registration can be classified into two categories: Area-based and feature-based approaches [10]. Area-based methods attempt to deal directly with the image intensity values to detect, for instance, cross-correlation [11], Fourier transform and mutual information [12,13]. On the contrary, feature-based methods attempt to indirectly extract salient structures or features in the images, such as points of high curvature, corners, strong edges, intersections of lines, structural contours and silhouettes within the images [14,15,16]. For visible and thermal infrared image pairs usually involve quite different intensity values, and the area-based methods are good usually only on a selected small region of the images. Instead, salient structures with a strong edge could often represent a significant common feature of the heterogeneous images. In view of this, this paper focuses on the feature-based registration methods for the visible and thermal infrared face images.
Feature-based registration methods first extract different feature point sets of salient structures in the multi-spectral images, and then the registration problem can be simplified to determine the correct correspondence and the inherent spatial transformation between two point sets of extracted features. Unlike the face recognition problem [17,18], the multi-sensor face registration problem has not enough common features available to build a convolutional neural network (CNNs) framework to predict the classification model; a popular strategy is to regard the alignment of two point sets as probability density estimation, e.g., the Gaussian Mixture Model (GMM) [19,20,21] and Student’s t Mixture Model (SMM) [22,23,24]. Because GMM formulations are intensively used, and the heavily tailed SMM is more robust and precise against noise and outliers than the classical GMM, it is a potential research direction to take advantage of SMM for image registration methods [25,26,27]. In general, both GMM and SMM probabilistic methods are intended to exploit global relationships in the point sets. However, the neighborhood structures among the feature points also contribute to the alignment of two point sets. Thus, another interesting point matching strategy would be aimed first at using local neighborhood structures as the feature descriptor to recover the feature point correspondences, e.g., Shape Context (SC) [28] and Inner-Distance Shape Context (IDSC) [29]. But for multi-spectral face images registration, the local neighborhood structures of feature points are not discriminative enough, because it will inevitably include a number of mismatching points. To address this issue, in this paper we modify the SMM probabilistic framework and take full advantage of the global and local structures during the multi-spectral face images registration process.
Our contribution in this paper includes the following three aspects: Firstly, in order to keep the global and local structures of feature point sets in the registration process, we use IDSC as the local shape descriptors of feature point sets, and treat the confidence of feature matching as the mixing proportion of the finite mixture model, which aims to describe the local shape and global geometrical relationship of feature point sets in a new SMM probabilistic framework. Secondly, in order to make the multi-spectral fusion image hold more apparent details of a visible image and the thermal radiation information of an infrared image, we propose a guided filtering and gradient preserving image fusion strategy for multi-spectral face images fusion. It can improve the robustness of face recognition techniques in varying illumination conditions. Thirdly, we construct a simple multi-spectral imaging system in order to acquire higher resolution multi-spectral face images from visible and thermal infrared cameras simultaneously.

2. The Proposed Registration Method for Visible and Infrared Face Images

For the software-based registration problem, a basic process of infrared and visible images fusion mainly includes the following procedures: Feature extraction and description, feature point sets registration, image transformation and interpolation and finally image fusion [1,30]. As is shown in Figure 1, the first step is to extract robust common salient features that might be preserved in the multi-spectral face images. Considering the fact that visible and infrared images are characterizations of two different modalities, we use edge maps to represent the common salient features of multi-spectral face images. After obtaining the salient features of multi-spectral images, the edge maps can usually be discretized as two feature point sets. Thus, the registration procedure is converted into determining the correct correspondence and estimating the spatial transformation between the feature point sets for multi-spectral face images. Furthermore, the other visible image points can be interpolated using a smoothing Thin-Plate Spline (TPS) [31] based on the pixel coordinates of the thermal infrared image. Finally, the multi-spectral image fusion process takes place in the output image, and the fused image can preserve both the apparent details of the visible image and the thermal radiation information of the infrared image, which will greatly improve the accuracy and efficiency of detection and recognition. In this paper, our attention mainly focuses on the following two aspects: Feature point sets registration and multi-spectral face images fusion.

2.1. Inner-Distance Shape Context for 2D Feature Descriptors

In order to establish the initial correspondences between the two feature point sets of visible and infrared face images, we first use IDSC as the local feature descriptors, and solve the point sets matching problem through a bipartite graph method [32]. Compared to the SC feature matching method, the IDSC feature matching method uses inner-distance to replace the Euclidean distance, which is robust to articulation shape and part structure. For example, Figure 2 shows the schematic illustration of face silhouette feature descriptors based on IDSC, in the IDSC feature histograms from Figure 2c, the horizontal axis nθ = 12 denotes the numbers of orientation bins and the vertical axis nd = 5 denotes the numbers of logarithm distance bins.
As can be seen from Figure 2c, the IDSC feature histograms for all marked points are quite different, and the excellent discriminative performance demonstrates the inner-distance’s ability to capture local structures in the point sets. Consider a point p i in the point set X and a point q j in the point set Y. Let Ci,j denote the cost of matching p i to q j , h X , p i ( k ) and h Y , q j ( k ) denote the normalized histogram of the relative coordinates of the remaining points, respectively, and K be the number of histogram bins. As the IDSC is represented based on the histogram, it can be measured by the χ2 test statistic:
C i , j = C ( p i , q j ) = 1 2 k = 1 K [ h X , p i ( k ) - h Y , q j ( k ) ] 2 h X , p i ( k ) + h Y , q j ( k ) ,
In general, the smaller the matching cost Ci,j is, the more similar the local appearance at points p i and q j are. Once the cost matrix C for all correspondence points in the point set X and Y is obtained, the correspondences Ω between the two point sets are an instance of an assignment problem, which can be solved by the Hungarian method.

2.2. Student’s t Mixtures for Feature Point Set Registration

In order to fully exploit the global structures in the feature point sets, we treat the alignment of two point sets as a probability density estimation of an SMM distribution with the IDSC local feature descriptors. Compared to the GMM distribution, the main advantage of SMM is that Student’s t distribution has a heavier tail than the exponentially decaying tail of Gaussian distribution. As is shown in the Figure 3, each component of the SMM has an additional parameter υ called the degrees of freedom, which can change the shape of a Student’s t distribution curve, and hence SMM provides a more robust model than the classical GMM.
Given a point set XN×D = (x1, …, xN)T, which can be considered as an observation datum, the other point set YM×D = (y1, …, yM)T is treated as the SMM centroids, where N and M represent the number of points in X and Y, respectively, and D represents the dimension of each point. Then the goal of registration is to align the centroids point set Y to the observation data point set X. Typically, the point sets contain noise and outliers, which can be supposed to be a uniform distribution 1/N, and the weight of the uniform distribution is denoted as η ∈ [0, 1]. Let wij ∈ [0, 1] denote a mixing proportion corresponding to the ith component of the SMM for point xj ( i = 1 M w i j =   1 ), then the mixture probability density function can take the form:
f ( x | ψ ) = η 1 N + ( 1 η ) i = 1 M ω i j f ( x j | y i , Σ i , υ i ) ,
where ψ = (ψ1, ψ2, …, ψM) is mixture parameter set with ψi = (wij, yi, i, υi), and f (xj|yi, i, υi) is the Student’s t distribution probability density function for the ith component of SMM; we express that:
f ( x j | y i , Σ i , υ i ) = Γ ( υ i + D 2 ) | Σ i | 1 2 ( π υ i ) 1 2 D Γ ( υ i 2 ) [ 1 + d ( x j , y i ; Σ i ) υ i ] υ i + D 2 ,
where d(xj, yi; i) is the Mahalanobis squared distance between observation point xj and centroid point yi, while ∑i, Γ and υ represent the covariance matrix, Gamma function and degrees of freedom, respectively.

2.3. Multi-Spectral Face Registration Using Global and Local Structures

In order to estimate the mixture parameter ψ in the mixture probability density function (2), according to [22,23], we first rewrite the function (2) as a complete data logarithm likelihood function ln L C ( ψ ) = j = 1 N ln i = 1 M w i j f ( x j | y i , Σ i , υ i ) , and then use the Expectation Maximization (EM) to train the logarithm likelihood function. In general, the EM algorithm can be divided into the following two steps: Expectation step (E-step) and Maximization step (M-step), and the iterative process proceeds by alternating between the E- and M-steps until convergence.
E-step: We first use the current parameter ψ(k) to calculate the posterior probability τ i j ( k ) , and according to the Bayesian Theorem, it can be expressed in terms of the observation data xj belonging to the ith component of the SMM
τ i j ( k ) = w i j ( k ) f ( x j | y i ( k ) , Σ i ( k ) , υ i ( k ) ) f ( x j | ψ ( k ) ) ,
Next we need to calculate the kth iteration conditional expectation of the logarithm likelihood function ln L C ( ψ ) , which can be calculated as follows:
Q ( ψ | ψ ( k ) ) = Q 1 ( w | ψ ( k ) ) + Q 2 ( υ | ψ ( k ) ) + Q 3 ( y , Σ | ψ ( k ) ) ,
where Q1, Q2 and Q3 are respectively written as follows:
Q 1 ( w | ψ ( k ) )   =   j = 1 N i = 1 M τ i j ( k ) ln w i j Q 2 ( υ | ψ ( k ) )   =   j = 1 N i = 1 M τ i j ( k ) ( ln Γ ( υ i 2 ) + υ i 2 ln ( υ i 2 ) + υ i 2 { j = 1 N ( ln u i j ( k ) u i j ( k ) ) + γ ( υ i ( k ) + D 2 ) ln ( υ i ( k ) + D 2 ) } ) Q 3 ( y , Σ i | ψ ( k ) ) = j = 1 N i = 1 M τ i j ( k ) ( D 2 ln ( 2 π ) 1 2 ln | Σ i | + D 2 ln u i j ( k ) u i j ( k ) x j y i 2 2 Σ i ) ,
with the gamma in Q2 that denotes the Digamma function, and the uij(k) in Q3 that represents the conditional expectation about additional missing data in the complete-data; it can be written as:
u i j ( k ) = υ i ( k ) + D υ i ( k ) + D ( x j , y i ( k ) ; Σ i ( k ) ) ,
In addition, we use the displacement function ρ to denote the non-rigid transformation in the point set: X = T(Y, ρ) = Y + ρ(Y), and add the regularization term φ(ρ) to the Q3, which can enforce the smoothness of the displacement function in the alignment of two point sets. So this Q3 term can be rewritten as:
Q ˜ 3 = Q 3 ( ρ , Σ i ( k + 1 ) | ψ ( k ) ) + λ 2 φ ( ρ ) ,
where λ is the regularization parameter and controlling the trade-off of two terms in Equation (7). By reproducing Kernel Hilbert Space and Fourier transformation, the displacement function ρ can be denoted as:
ρ ( y j ) = i = 1 M h i G ( y i , y j ) , G ( y i , y j ) = e 1 2 y i y j β 2 ,
where G is a Gaussian kernel matrix, β determines the width of the smoothing Gaussian filter, and the Equation (8) can be conveniently denoted in matrix form as: ρ(Y) = GH, where HM×D = (h1, …, hM) is a matrix of coefficients. Thus, the Equation (7) can be expanded in the following matrix form:
Q ˜ 3 = j = 1 N i = 1 M τ i j ( k ) ( D 2 ln ( 2 π ) 1 2 ln | Σ i | + D 2 ln u i j ( k ) u i j ( k ) x j y i 2 2 Σ i ) + λ 2 t r ( ( H T ) ( k ) G H ( k ) ) ,
M-step: On the M-step at the (k + 1)th iteration of the EM algorithm, considering that Q1, Q2 and Q3 in Equation (5) can be computed independently of each other, thus the maximization of the objective function Q1, Q2 and Q3 with respect to the parameters w, υ, , ρ can be operated separately. The mixing proportion for SMM is updated by our consideration of the first term Q1. Given the feature descriptors in Section 2.1, it can obtain the coarse correspondences Ω between the point sets X and Y. For an observation data xj, we define ι (0 ≤ ι ≤1) as a confidence by IDSC feature matching, and then the mixing proportion wij by incorporating the local structures among neighboring points can be updated by the following rule:
w i j ( k + 1 ) = { j = 1 N τ i j ( k ) N , ( x j y i ) Ω ι ,      ( x j y i ) Ω ,
If observation data xj does not have a corresponding point yi in the label Ω, the mixing proportion wij(k+1) is given by the average of the posterior probabilities τ i j ( k ) of the SMM component membership. If the observation datum xj corresponds to the centroid point yi in the label Ω, the wij(k+1) is given by a constant confidence ι.
Note that in Equation (10) the mixing proportion wij does not only depend on prior assignment by local structures, but also depends upon the posterior probability of observation data xj belonging to the ith component of the SMM by global structures. In addition, we obtain degree of freedom υ by taking the corresponding derivative of Q2 to zero, and the updated value υ(k+1) is a solution of the following equation:
1 - γ ( υ i ( k + 1 ) 2 ) + ln ( υ i ( k + 1 ) 2 ) + j = 1 N τ i j ( k ) ( ln u i j ( k ) u i j ( k ) ) j = 1 N τ i j ( k ) + γ ( υ i ( k ) + D 2 ) ln ( υ i ( k ) + D 2 ) = 0 ,
Then, to update the estimates of covariance matrices and displacement function ρ in Equation (7), we also need to take the derivative of Q ˜ 3 , and the update values can be denoted as:
H ( k + 1 ) = ( d i a g ( P ( k ) 1 ) G + λ Σ i ( k ) I ) 1 ( P ( k ) X d i a g ( P ( k ) 1 ) Y ) , P i j ( k ) = τ i j ( k ) u i j ( k ) Σ i ( k + 1 ) = j = 1 N τ i j ( k ) u i j ( k ) x j T ( y i , ρ ) 2 j = 1 N τ i j ( k + 1 ) , ( i = 1 , , N )
where 1 is a column vector of all ones, and I is an identity matrix, diag(•) denotes a diagonal matrix and G(i,•) denotes the column vector in the kernel matrix G. Furthermore, for our centroid point yi in the point set Y, the non-rigid transformation T(Y, ρ) = Y + ρ(Y) can be expressed by Equation (8) as follows:
T ( y i , ρ ) = y i + G ( i , · ) H i ( k + 1 ) ,
At the end, the iterative process alternates between E-steps and M-steps until satisfying the following convergence condition:
L C ( ψ ( k + 1 ) ) L C ( ψ ( k ) ) L C ( ψ ( k + 1 ) ) ε ,
where ε is a convergence threshold. After a number of repeatability testing, we use the following setting for the parameters η = 0.1, ι = 0.8, λ = 3, β = 2, ε = 10−5 in our experiment. Furthermore, to register the visible and infrared images accordingly, the image transformation and interpolation is performed on the visible image. Since our multi-spectral face registration method is implemented by using global and local structures, then the registration method is to be named as Face Registration using the Global and Local Structure (FR-GLS) in the rest of this paper.

3. The Proposed Fusion Strategies for Visible and Infrared Face Images

After obtaining the accurate alignment of the visible and infrared images, the second problem is then to solve the multi-spectral images fusion. In order to make sure that the fused image can preserve both the apparent details of the visible image and thermal radiation information of the infrared image, we propose the fusion strategies based on Guided Filtering and Gradient preserving (GF-GP). Firstly, we use two-scale image decomposition and reconstruction with a guided filtering algorithm to get an initial fusion image, and then the output fusion image is optimized by combining the intensity distribution of the infrared image and the intensity variation of the visible image. According to [33], since each source image can be separated into a base layer containing the large-scale details, and the detail layer containing the small-scale details, the two-scale image fusion FG is given by the guided filtering algorithm as follows:
F G = B ¯ + D ¯ ,
where the fused base layer B ¯ and the fused detail layer D ¯ can be denoted as a weighted average cost of the base Bn and detail Bn layers of different source images:
B ¯ = n = 1 N W n B B n , D ¯ = n = 1 N W n D D n ,
where W n B and W n D represent the refined weight maps of the base Bn and detail layers Dn via guided image filtering respectively. In order to optimize the initial fusion image FG, here the visible, infrared and fused images are defined by V, I and F, respectively. Given that the thermal radiation information in the infrared image is typically characterized by the pixel intensity values, the pixel intensity values are quite different between the target and background. Thus, we constrain the fused image F to have the similar pixel intensity distribution with the infrared source image I, and the optimization problem can be formulated as:
δ 1 = F I ,
where ‖•‖ denotes the l1 norm. Besides, given that the pixel intensity distribution in the same physical location might be discrepant for infrared and visible face images, they are manifestations of two different phenomena. Note that the detailed information of object edge and texture is mainly characterized by the pixel gradient values in the visible image. Hence, we constrain the fused image F to have the similar pixel gradient values with the visible source image V, the optimization problem can also be formulated as:
δ 2 = F V ,
where denotes the gradient operator. By combining the Equations (17) and (18), the optimization problem of the fusion image F can be formulated as minimizing the following objective function:
δ F = δ 1 + α δ 2 = F I + α F V ,
where α is a weighting factor, the optimal F could be found by using gradient descent strategy. Furthermore, it is important to note that if the α is small, the fusion image preserves the more thermal radiation information of the infrared image, otherwise, the fusion image preserves more edge and texture information of the visible image. We set the weighting coefficient α = 5 in our experiment, because it can achieve good subjective visual quality in most cases.

4. Experimental Results

To demonstrate the robustness and efficiency of our proposed registration method in solving the multi-spectral face image registration problem, subjective and objective evaluation experiments are conducted with the UTK-IRIS standard multi-spectral face database [34] and self-constructed multispectral face data sets. The experiments are performed on a desktop with 3.3 GHz Intel Core™ i5-4590 CPU, 8 GB memory and MATLAB Code.

4.1. Registration on Real Face Images with the UTK-IRIS Database

In this section, the performance of face image registration with our FR-GLS method is evaluated by using the public UTK-IRIS Database. The images in the database have a spatial resolution of 320 × 240 pixels and the database contains individuals with various poses, facial expressions and illumination variations. We first give an intuitive impression of registration and fusion results, and then we provide a quantitative comparison with three typical methods such as Coherent Point Drift (CPD) [19], Regularized Gaussian Fields (RGF) [20] and SMM [25], based on the specified landmarks of visible and thermal infrared image pairs. The reasons for choosing these three comparison methods rest on the following two considerations. On the one hand, CPD and SMM use both global structures based on finite mixture models to parameterize the transformations. On the other hand, besides the global relationships in the feature point sets, RGF uses local features descriptor such as SC to recover the accurate transformation.
However, our FR-GLS has two major advantages compared to RGF: (i) Unlike the local features descriptor used in RGF, we use a more robust local features descriptor such as IDSC to initialize the correspondences between the two shape features; (ii) In order to preserve the global and local structures of feature point sets during the registration process, a heavy tail distribution in our FR-GLS is used to replace the traditional Gaussian distribution in the RGF method. Furthermore, we use the IDSC descriptor to assign the mixing proportion wij of the SMM.

4.1.1. Qualitative Evaluation

In order to obtain a qualitative evaluation of the registration results issued by the software-based registration process in Figure 1, several typical frame pairs of four individuals involving different poses and illumination changes are given in Figure 4a–d. Specifically, the original visible images and thermal infrared images are shown in the first and second row, respectively. Then the middle row shows the edge maps extracted by the canny edge detector. We use the red point set to denote the edge map of the thermal infrared images, and the blue point set corresponds to the edge map of the visible images. Subsequently, the visible image is aligned to the thermal infrared image by our FR-GLS method, and the registration performance is demonstrated by the checkerboard in the fourth row, where the aligned visible image is superimposed onto the corresponding thermal infrared image. We can see that the seams between two grids in the checkerboard are natural, and this demonstrates the feasibility and effectiveness of the proposed registration method qualitatively. Finally, the last row represents the fusion results of the visible and thermal IR images by our GF-GP method. We can see that the fused images in the last row contain both the apparent details of the visible image and the thermal radiation information of the infrared image, which can be clearly seen from noses, cheeks, ears, mouths and glasses. Therefore, it will be beneficial for the performance of follow-up face detection and recognition under the case of illumination changes or disguise.

4.1.2. Quantitative Evaluation

In order to give a quantitative comparison of our FR-GLS and other typical methods, such as CPD [19], RGF [20] and SMM [25], we further utilize several visible and infrared image pairs of the four individuals in Figure 4, and manually selected several landmarks and their correspondences (about 20 pairs) in each visible and infrared image pairs as ground truth. For example, the landmarks involve the salient features like the cheeks, eyes, nose, mouth, eyebrows, ears and glasses, which are highly distinguished. Referring to the metric used in [20,21,35], the recall rate is proposed to be a qualitative indicator for evaluating the registration results on all landmarks set pairs of an individual. Here the recall rate, or true positive rate, is defined as the proportion of the true positive correspondences between the landmark pairs to the ground truth correspondences, and a true positive correspondence is counted when the pair falls within a given accuracy threshold of Euclidean distance. Here the Euclidean distance can be defined as 2-norm between a landmark in the aligned feature point set of visible images and the corresponding landmark in the feature point set of infrared images. The recall curves plot the recall rates under different threshold values.
Figure 5 shows the recall rates against different threshold values using the registration methods of CPD, SMM, RGF and FR-GLS, considering four individuals. Each individual contains about 100 hundred-image pairs corresponding to every spectrum. By the definition of recall, we can see that the FR-GLS and RGF methods for the global and local structures are significantly superior to the CPD and SMM methods for the global structures only based on finite mixture models. Moreover, the curves of our FR-GLS method are consistently above the curves of the RGF method for different accuracy thresholds and individuals, which could be attributable to the combination of the robust SMM and IDSC local feature descriptor.
The average registration errors of each individual for different registration methods are summarized in Table 1. We again see that the FR-GLS and RGF methods with the global and local structures significantly outperform the CPD and SMM methods just with finite mixture models.
Moreover, our FR-GLS method can achieve consistently better performance compared to the RGF method. The average registration errors of this RGF method are about 2.40, 2.24, 2.24 and 1.92 pixels on the four individuals, respectively. By contrast, the average registration errors of our FR-GLS method are reduced to about 2.25, 2.06, 1.99 and 1.78 pixels, respectively. In addition, The SMM method with SMM has a slightly better performance than the CPD method with GMM, where the average registration errors are 2.88, 2.40, 2.56 and 2.13 pixels, respectively. Whereas, the average registration errors of the CPD method are up to 3.04, 2.58, 2.73 and 2.35 pixels, respectively. It can quantitatively verify the heavy-tailed property of SMM.

4.2. Registration on Real Face Images of Self-Built Multi-Spectral Database

In order to demonstrate the performance of our registration in the high-resolution multi-spectral imaging system, we built a four degree of freedom (DOF) pan and tilt head vision platform for multispectral database acquisition, which can be seen in Figure 6.
This 4-DOF pan and tilt head vision platform can realize up-down pitch movement and left-right yaw movement, respectively. Moreover, a Belgium Raven-640-Analog thermal infrared camera with 10 mm short lens is used to acquire a 74° × 59° field of view (FOV) infrared image (Figure 6a left), while a DAHENG MER-310-12UC color camera with an Optotune EL-10-30-Ci variable-focus liquid lens is used to acquire a 28° × 22° FOV visible image (Figure 6a right). Meanwhile, a visible image is always taken simultaneously with the infrared image to form a multi-spectral image pair, and the visible and infrared images are both compressed into 640 × 480 pixels. The multi-spectral vision equipment and image acquisition scenario is shown in Figure 6b. The self-built multi-spectral database is composed of 20 different individuals involving different illumination conditions and pose variations. In addition, for each individual, the sampled images are captured from 13 different angles (tilt head from −90° to +90°) and two illumination conditions (dark and light).

4.2.1. Qualitative Comparison

Referring to the qualitative evaluation in Section 4.1, Figure 7 shows an intuitive impression of registration and fusion results on six typical frame pairs in the self-built multi-spectral database.
As shown in the Figure 7 above. Firstly, several original visible images and thermal infrared images of six individuals are given in Figure 7a,b, respectively, and those face images involve different poses and illumination changes. In addition, the image pairs in the bright light are shown in the first three rows, while the image pairs in the dark condition are shown in the last three rows with different poses. Then, Figure 7c shows the discrete point set of edge maps, which are extracted by the canny edge detector. For distinguishing each spectral image, we also use the blue point set to denote the edge map of the visible image, and the red point set corresponds to the edge map of the thermal infrared image. We can see that contour point sets of the face images in the edge map pairs are significantly different for the thermal infrared and visible images. Those discrepancies will result in a lot of outlier and noise for the edge maps alignment, especially in the textured regions of hairs and clothes. Hence, it is hard to match the two edge map pairs accurately.
Furthermore, note that we just focus on the registration results of the face region, the redundant information, such as the edge points of clothes, which is not related to face detection, might be ignored during the matching. Subsequently, the visible image is aligned to the thermal infrared image by our FR-GLS method, and the registration performance is demonstrated by checkerboard images in Figure 7d, where the aligned visible image is superimposed to the corresponding thermal infrared image. In addition to the uninterested regions of hairs and clothes, we can see from Figure 7d that the seams between two grids in the checkerboard images are natural, and this can demonstrate the feasibility and effectiveness of the proposed registration method qualitatively. Finally, for further verification, Figure 7e shows the fusion results of the visible and thermal infrared images by our GF-GP method. Compared to the original thermal infrared images in the second column (Figure 7b), we can see that the fused images in the last column (Figure 7e) seem to be sharpened, which contain both the apparent details of the visible image and the thermal radiation information of the infrared image. Thus, it will be beneficial for follow-up face detection and recognition in case of illumination changes, e.g., the outdoor environment illumination changed from bright to dark.

4.2.2. Quantitative Comparison

To have a quantitative evaluation, we also manually select a set of landmarks in the multi-spectral images (about 40 pairs of landmarks for each spectral image), and then treat the recall on all real face images of an individual as the metric. Figure 8 shows the recall comparison results of the four methods on six individuals, where each individual contains about 52 image pairs.
We can see from the Figure 8 above, FR-GLS (mauve fork) and RGF (green star) methods with the global and local structures are significantly superior to the CPD and SMM methods with global structures, and the curve of the Student’s t distribution model (red square) is mainly above the curve of the Gaussian distribution model (blue triangle). Furthermore, the curve of our FR-GLS method, combined with SMM and IDSC, is consistently above those of the other three methods. In conclusion, for the six pairs of typical multi-spectral images in Figure 7, the average registration errors of CPD, SMM, RGF and our FR-GLS method are about 2.40, 2.24, 2.01 and 1.76 pixels, respectively.

5. Conclusions

In this paper, we have introduced a novel multi-sensor face images registration method. It uses the IDSC to describe the local feature of a face image, which is more stable than SC, and then the Student’s t Mixture probabilistic model is used to estimate the transformation between visible and infrared images by combining global and local structures of feature point sets. In order to verify the correctness of the registration method intuitively, we propose a multi-spectral image fusion strategy based on guided filtering and gradient preserving. Experimental results on a standard real face database and self-built multi-spectral face database demonstrate that the proposed method is able to achieve much more registration accuracy compared to other state-of-the-art registration methods. Therefore, it will be beneficial to improve the reliability of a fusion-based face recognition system. In the future, given that the ultimate goal of registration and fusion is to enhance the recognition rate, a series of more comprehensive and scientific evaluations for multi-sensor face images registration will be conducted through some combination of the latest deep neural network face recognition method.

Author Contributions

Conceptualization, W.L. and W.Z.; Methodology, W.L.; Software, W.Z.; Validation, N.L. and W.Z.; Formal Analysis, W.L.; Investigation, W.L.; Resources, M.D.; Data Curation, W.L.; Writing-Original Draft Preparation, W.L.; Writing-Review & Editing, W.L. and W.Z.; Visualization, W.L.; Supervision, N.L. and M.D.; Project Administration, X.L.; Funding Acquisition, M.D and X.L.

Funding

This research was funded by National Natural Science Foundation of China, grant number No. 51475046, No. 51475047 and the National High-tech R&D Program of China. grant number No. 2015AA042308.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  2. Ma, J.; Yu, W.; Liang, P. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  3. Huang, Y.; Bi, D.; Wu, D. Infrared and visible image fusion based on different constraints in the non-subsampled shearlet transform domain. Sensors 2018, 18, 1169. [Google Scholar] [CrossRef]
  4. Ma, J.; Chen, C.; Li, C. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  5. Paramanandham, N.; Rajendiran, K. Infrared and visible image fusion using discrete cosine transform and swarm intelligence for surveillance applications. Infrared Phys. Technol. 2018, 88, 13–22. [Google Scholar] [CrossRef]
  6. Zhu, Z.; Qi, G.; Chai, Y. A geometric dictionary learning based approach for fluorescence spectroscopy image fusion. Appl. Sci. 2017, 7, 161. [Google Scholar] [CrossRef]
  7. Singh, R.; Vatsa, M.; Noore, A. Integrated multilevel image fusion and match score fusion of visible and infrared face images for robust face recognition. Pattern Recognit. 2008, 41, 880–893. [Google Scholar] [CrossRef] [Green Version]
  8. Vizgaitis, J.N.; Hastings, A.R. Dual band infrared picture-in-picture systems. Opt. Eng. 2013, 52, 061306. [Google Scholar] [CrossRef]
  9. Kong, S.G.; Heo, J.; Boughorbel, F. Multiscale fusion of visible and thermal IR images for illumination-invariant face recognition. Int. J. Comput. Vis. 2007, 71, 215–233. [Google Scholar] [CrossRef]
  10. Oliveira, F.P.M.; Tavares, J.M.R.S. Medical image registration: A review. Comput. Method Biomech. 2014, 17, 73–93. [Google Scholar] [CrossRef]
  11. Avants, B.B.; Epstein, C.L.; Grossman, M. Symmetric diffeomorphic image registration with cross-correlation: Evaluating automated labeling of elderly and neurodegenerative brain. Med. Image Anal. 2008, 12, 26–41. [Google Scholar] [CrossRef] [Green Version]
  12. Pan, W.; Qin, K.; Chen, Y. An adaptable-multilayer fractional fourier transform approach for image registration. IEEE Trans. Pattern Anal. 2008, 31, 400–414. [Google Scholar] [CrossRef]
  13. Zhuang, Y.; Gao, K.; Miu, X. Infrared and visual image registration based on mutual information with a combined particle swarm optimization—Powell search algorithm. Optik Int. J. Light Electron Opt. 2016, 127, 188–191. [Google Scholar] [CrossRef]
  14. Sotiras, A.; Davatzikos, C.; Paragios, N. Deformable medical image registration: A survey. IEEE Trans. Med. Imaging 2013, 32, 1153–1190. [Google Scholar] [CrossRef]
  15. Viergever, M.A.; Maintz, J.B.A.; Klein, S. A survey of medical image registration–under review. Med. Image Anal. 2016, 33, 140–144. [Google Scholar] [CrossRef]
  16. Moghbel, M.; Mashohor, S.; Mahmud, R. Review of liver segmentation and computer assisted detection/diagnosis methods in computed tomography. Artif. Intell. Rev. 2018, 50, 497–537. [Google Scholar] [CrossRef]
  17. Zhang, Y.; Zhao, D.; Sun, J. Adaptive convolutional neural network and its application in face recognition. Neural Process. Lett. 2016, 43, 389–399. [Google Scholar] [CrossRef]
  18. Zhang, K.; Zhang, Z.; Li, Z. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal. Proc. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  19. Myronenko, A.; Song, X. Point set registration: Coherent point drift. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2262–2275. [Google Scholar] [CrossRef]
  20. Ma, J.; Zhao, J.; Ma, Y. Non-rigid visible and infrared face registration via regularized Gaussian fields criterion. Pattern Recognit. 2015, 48, 772–784. [Google Scholar] [CrossRef]
  21. Tian, T.; Mei, X.; Yu, Y. Automatic visible and infrared face registration based on silhouette matching and robust transformation estimation. Infrared Phys. Technol. 2015, 69, 145–154. [Google Scholar] [CrossRef]
  22. Peel, D.; McLachlan, G.J. Robust mixture modelling using the t distribution. Stat. Comput. 2000, 10, 339–348. [Google Scholar] [CrossRef]
  23. Gerogiannis, D.; Nikou, C.; Likas, A. Robust Image Registration using Mixtures of t-distributions. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  24. Gerogiannis, D.; Nikou, C.; Likas, A. The mixtures of Student’s t-distributions as a robust framework for rigid registration. Image Vis. Comput. 2009, 27, 1285–1294. [Google Scholar] [CrossRef]
  25. Zhou, Z.; Zheng, J.; Dai, Y. Robust non-rigid point set registration using student’s-t mixture mode. PloS ONE 2014, 9, e91381. [Google Scholar]
  26. Zhou, Z.; Tu, J.; Geng, C. Accurate and robust non-rigid point set registration using student’s t mixture model with prior probability modeling. Sci. Rep. UK 2018, 8, 8742. [Google Scholar] [CrossRef]
  27. Maiseli, B.; Gu, Y.; Gao, H. Recent developments and trends in point set registration methods. J. Vis. Commun. Image Represent. 2017, 46, 95–106. [Google Scholar] [CrossRef]
  28. Belongie, S.; Malik, J.; Puzicha, J. Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar] [CrossRef] [Green Version]
  29. Ling, H.; Jacobs, D.W. Shape classification using the inner-distance. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 286–299. [Google Scholar] [CrossRef]
  30. Li, H.; Wu, X.J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef]
  31. Chui, H.; Rangarajan, A. A new point matching algorithm for non-rigid registration. Comput. Vis. Image Underst. 2003, 89, 114–141. [Google Scholar] [CrossRef]
  32. Riesen, K.; Bunke, H. Approximate graph edit distance computation by means of bipartite graph matching. Image Vis. Comput. 2009, 27, 950–959. [Google Scholar] [CrossRef]
  33. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [PubMed]
  34. UTK-IRIS Database. Available online: http://www.cse.ohio-state.edu/otcbvs-bench/ (accessed on 10 September 2015).
  35. Dwith, C.; Ghassemi, P.; Pfefer, T. Free-form deformation approach for registration of visible and infrared facial images in fever screening. Sensors 2018, 18, 125. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Visible and infrared face images fusion basic process for software-based registration.
Figure 1. Visible and infrared face images fusion basic process for software-based registration.
Applsci 09 04623 g001
Figure 2. Schematic illustration of face silhouette feature descriptors based on Inner-Distance Shape Context (IDSC). (a) Bellman-Ford shortest path graph built using face silhouette landmark points; (b,c) Four marked points and their corresponding IDSC feature histograms.
Figure 2. Schematic illustration of face silhouette feature descriptors based on Inner-Distance Shape Context (IDSC). (a) Bellman-Ford shortest path graph built using face silhouette landmark points; (b,c) Four marked points and their corresponding IDSC feature histograms.
Applsci 09 04623 g002
Figure 3. The Student’s t distribution for various degrees of freedom.
Figure 3. The Student’s t distribution for various degrees of freedom.
Applsci 09 04623 g003
Figure 4. Registration and fusion results of our method on four typical unregistered visible/infrared image pairs in the database of UTK-IRIS. (a) Charles; (b) Heo; (c) Meng; (d) Sharon.
Figure 4. Registration and fusion results of our method on four typical unregistered visible/infrared image pairs in the database of UTK-IRIS. (a) Charles; (b) Heo; (c) Meng; (d) Sharon.
Applsci 09 04623 g004
Figure 5. Quantitative comparisons of multispectral face image pairs with four individuals in Figure 4. (a) Charles; (b) Heo; (c) Meng; (d) Sharon.
Figure 5. Quantitative comparisons of multispectral face image pairs with four individuals in Figure 4. (a) Charles; (b) Heo; (c) Meng; (d) Sharon.
Applsci 09 04623 g005
Figure 6. 4-DOF pan and tilt head vision platform for multi-spectral database acquisition: (a) Design mechanical model of vision platform; (b) Image acquisition scenario.
Figure 6. 4-DOF pan and tilt head vision platform for multi-spectral database acquisition: (a) Design mechanical model of vision platform; (b) Image acquisition scenario.
Applsci 09 04623 g006
Figure 7. Registration and fusion results of our method considering six typical unregistered visible and infrared image pairs in the self-built multi-spectral face datasets. (a) Visible image; (b) Thermal infrared image; (c) Canny edge maps; (d) Superimposed checkerboard pattern of visible and infrared images; (e) Fusion results.
Figure 7. Registration and fusion results of our method considering six typical unregistered visible and infrared image pairs in the self-built multi-spectral face datasets. (a) Visible image; (b) Thermal infrared image; (c) Canny edge maps; (d) Superimposed checkerboard pattern of visible and infrared images; (e) Fusion results.
Applsci 09 04623 g007
Figure 8. Quantitative comparisons of multi-spectral face image pairs with 6 individuals in Figure 7.
Figure 8. Quantitative comparisons of multi-spectral face image pairs with 6 individuals in Figure 7.
Applsci 09 04623 g008
Table 1. The average registration errors comparison of CPD, SMM, RGF and our FR-GLS method on four typical individuals in Figure 4.
Table 1. The average registration errors comparison of CPD, SMM, RGF and our FR-GLS method on four typical individuals in Figure 4.
MethodsThe Average Registration Errors
CharlesHeoMengSharon
CPD [19]3.042.582.732.35
SMM [25]2.882.402.562.13
RGF [20]2.402.242.241.92
FR-GLS2.252.061.991.78

Share and Cite

MDPI and ACS Style

Li, W.; Dong, M.; Lu, N.; Lou, X.; Zhou, W. Multi-Sensor Face Registration Based on Global and Local Structures. Appl. Sci. 2019, 9, 4623. https://doi.org/10.3390/app9214623

AMA Style

Li W, Dong M, Lu N, Lou X, Zhou W. Multi-Sensor Face Registration Based on Global and Local Structures. Applied Sciences. 2019; 9(21):4623. https://doi.org/10.3390/app9214623

Chicago/Turabian Style

Li, Wei, Mingli Dong, Naiguang Lu, Xiaoping Lou, and Wanyong Zhou. 2019. "Multi-Sensor Face Registration Based on Global and Local Structures" Applied Sciences 9, no. 21: 4623. https://doi.org/10.3390/app9214623

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop