**Training Set Enlargement Using Binary Weighted Interpolation Maps for the Single Sample per Person Problem in Face Recognition**

**Yonggeol Lee <sup>1</sup> and Sang-Il Choi 2,\***


Received: 22 August 2020; Accepted: 21 September 2020; Published: 23 September 2020

**Abstract:** We propose a method of enlarging the training dataset for a single-sample-per-person (SSPP) face recognition problem. The appearance of the human face varies greatly, owing to various intrinsic and extrinsic factors. In order to build a face recognition system that can operate robustly in an uncontrolled, real environment, it is necessary for the algorithm to learn various images of the same person. However, owing to limitations in the collection of facial image data, only one sample can typically be obtained, causing difficulties in the performance and usability of the method. This paper proposes a method that analyzes the changes in pixels in face images associated with variations by extracting the binary weighted interpolation map (B-WIM) from neutral and variational images in the auxiliary set. Then, a new variational image for the query image is created by combining the given query (neutral) image and the variational image of the auxiliary set based on the B-WIM. As a result of performing facial recognition comparison experiments on SSPP training data for various facial-image databases, the proposed method shows superior performance compared with other methods.

**Keywords:** image generation; weighted interpolation map; binarization; single sample per person

### **1. Introduction**

Face recognition technology is used to identify individuals from their captured facial images by leveraging a labeled database containing people's identities. Compared with other types of biometric recognition, face recognition is less invasive and does not require a subject to be in proximity to or in contact with a sensor, making the method widely applicable to user identification, e-commerce, access control, surveillance, and human–computer interaction. However, because variations caused by extrinsic factors (e.g., illumination and pose) and intrinsic factors (e.g., facial expression, age, and accessories) are very large, it is difficult to robustly recognize a face under uncontrolled conditions [1,2]. To deal with these variations, facial recognition methods have been studied under the assumption that several images can be made available for each person, and high-performance methods have been built using vast databases of this nature (e.g., VGGface2 [3], Tufts Face [4], UMDfaces [5]), MegaFace [6], and LFW [7,8] databases).

However, for many large-scale face recognition applications (e.g., passport authentication, drivers' license identification, and police investigations), the training data required to learn algorithms do not offer many samples per person. In many cases, there is only a single sample per person (SSPP) available [9,10]. For example, law enforcement agencies have constructed databases of facial images (i.e., mug-shots) for decades. The related datasets comprise frontal face images under steady illumination and blank expressions. However, owing to cost and privacy issues, these databases are rarely augmented with extra multi-conditional candid photos. Furthermore, it is known that

criminals usually attempt to disguise their identities when committing a crime [11,12]. Even if they do not, it remains very difficult for systems to match active faces with the collected neutral images. As such, the dearth of learnable data restricts the use of feature-extraction and various other supervised methods [13–16].

To solve the SSPP problem, several methods of enlarging training datasets have been proposed to generate new images from a given one. The theory of the evolution of technology suggests that such datasets can be expanded using existing means [17–20]. In E(PC)2 A2+ [21], extended from (PC)2 A [22], an image and its corresponding half-, first-, and second-order projected images were used as the training set. In the (2D)<sup>2</sup> PCA method [23], new images were generated by simultaneously applying two-directional principal component analysis (PCA) [24] in the row and column directions of 2D images. In the SPCA+ method [25], the training set was enlarged by combining the original image linearly with its derived image obtained by perturbing the image matrix's singular values. In [26], concatenated left- and right-side images obtained from the symmetry of the face were used as training samples. In [27], images were generated using a symmetry transform for the intraclass and a linear combination for the interclass. In the interclass relationship (ICR) [28], data were generated by a weighted combination of (at least) two images in the training set. The ICR rectified the underestimated intraclass and overestimated the interclass. In MVI [29], the training set was enlarged by generating multiple low-resolution virtual images using a single high-resolution image. In SRGES [30], images were generated by adding mean images of the difference between neutral and variational images for each variation in the auxiliary set to the query image. In [31], occluded images were generated by using a weighted interpolation map and an auxiliary set. The weighted interpolation map represented the degrees of changes in pixels at the same positions between an image and its occluded version. The degree of changes was measured using the standard deviation of the difference between neutral and variational images in the auxiliary set. When generating a new image, the pixels at positions of large differences were replaced with the pixel values of the average image of the occluded images in the auxiliary set.

In this paper, we propose binary weighted interpolation maps (B-WIM) to enlarge the training set for face recognition. Generally, the occurrence of variations leads the local pixels to change in the face image. Supposing it is possible to grasp the change in local pixels between the original and varied images, it then becomes possible to capture the characteristics of the image changes caused by the variation. By analyzing these characteristics, the proposed method can maintain most of the characteristics of the neutral image while replacing only the changed areas with the pixel values of the variational one. For this, we first construct an auxiliary set consisting of neutral images and their variational images. Then, the normalized weighted interpolation map is extracted by using the log-scaled standard deviation of the absolute difference between the neutral images and the corresponding variational images in the auxiliary set. Each element of the weighted interpolation map reflects the degree of change caused by the variation in individual pixels, and a B-WIM is obtained via binarization.

When generating a new image for a given query (neutral) image, the variational image corresponding to the neutral image having the highest correlation with the query image is selected from the auxiliary set. Then, a new image is generated by combining the query and selected variational images. The overall procedure of the proposed method is shown in Figure 1.

The idea for the proposed method was motivated by the ICR concept and the weighted interpolation map method, which are face-generating frameworks. However, unlike ICR, by simply increasing the number of images by the weight combinations of two images, the proposed method has the advantage of generating a natural image with a specific variation. Additionally, the proposed method creates an image of higher quality than the weighted interpolation map (WIM) method by selecting the neutral image and the variational image corresponding to the query image.

Face recognition experiments are evaluated using the following criteria. First, we measure the change in the face recognition rate according to the degree of variation of different databases. Second, the face recognition performance is analyzed using unsupervised and supervised learning methods. Finally, the overall face recognition rates of all methods are assessed. We compare the proposed method with other methods dealing with the SSPP problem: WIM, ICR, E(PC2)A+, SPCA+, (2D)2PCA, SLC, MVI, and SRGES. The results of the experiment show that the proposed method exhibits high face recognition performance for all criteria.

The remainder of this paper is organized as follows. Section 2 explains the proposed method for generating data and describes each procedure in detail. The experimental face recognition results are described in Section 3, and the discussion and conclusion follow.

**Figure 1.** Overall procedure of the proposed method.

### **2. Proposed Method**

Using ICR [28], a new image, **I***new*, is generated using a weighted combination of neutral images, **I***<sup>i</sup>* and **I***j*, as

$$\mathbf{I}\_{new} = (1 - \lambda) \cdot \mathbf{I}\_i + \lambda \cdot \mathbf{I}\_j, \quad \text{where } 0 \le \lambda \le 1 \tag{1}$$

In Equation (1), the weight, *λ*, decides the ratio of **I***<sup>i</sup>* and **I***<sup>j</sup>* to be reflected in **I***new*. If *λ* is 0.5, the two images are assumed to have been reflected equally. In the case of ICR, it is easy to generate images by applying a single parameter for all pixels. However, the changes in some areas within the image will not be reflected accurately, owing to variations. Extrinsic factors, such as occlusions, cause changes in the neutral image. For example, pixels around the eyes change significantly when wearing sunglasses. On the other hand, areas unrelated to these variations generally retain the pixel information of the neutral image. Therefore, it is necessary to obtain the weights for each pixel.

In this paper, we propose a method to enlarge the training set. A new image, **I***new*, with variations, is generated via a combination of the neutral image, **I**, and the variational image, ˆ **I**, derived from **I** by referring to the ICR. The method of generating a new image is redefined as follows:

$$I\_{new}(\mu, \upsilon) = (1 - \mathcal{B}(\mu, \upsilon))I(\mu, \upsilon) + \mathcal{B}(\mu, \upsilon)I(\mu, \upsilon). \tag{2}$$

In Equation (2), <sup>B</sup>(*u*, *<sup>v</sup>*) is the pixel weight at a certain position in both **<sup>I</sup>** and <sup>ˆ</sup> **I**. A new image, **I***new*, is generated, containing the variations only in some areas while maintaining the characteristics of the neutral image as much as possible.

### *2.1. Binary Weighted Interpolation Maps (B-WIM)*

The occurrence of variations changes the pixel values of the neutral image. Figure 2 shows the difference between **I** and ˆ **I**, caused by facial expression variations. In the aligned face images, the area around the mouth where the smile occurs significantly changes the pixel value of **I**.

**Figure 2.** (**a**) Neutral image; (**b**) variational image; (**c**) absolute difference between neutral and variational image.

Absolute difference is the difference between (**<sup>I</sup>** <sup>∪</sup> <sup>ˆ</sup> **<sup>I</sup>**) and (**<sup>I</sup>** <sup>∩</sup> <sup>ˆ</sup> **I**). The standard deviation represents the degree of statistical variation from the difference between the pixels of these images

$$\begin{aligned} \mathcal{M}(u,v) &= \log \left( \frac{1}{m-1} \sqrt{\sum\_{i=1}^{m} (|\not{I}\_{i}(u,v) - I\_{i}(u,v)| - \mu(u,v))^{2}} \right)\_{\ast} \\ \mu(u,v) &= \frac{1}{m} \sum\_{i=1}^{m} |\not{I}\_{i}(u,v) - I\_{i}(u,v)| \end{aligned} \tag{3}$$

where subscripts *i*(= 1, 2, .., *m*) denote the *i*th individual. In Equation (3), when the value is very large according to the degree of change, some pixels are saturated in the generated image. Therefore, the normalized **M** is calculated as follows:

$$\mathbf{M} = \frac{\mathbf{M} - \min(\mathbf{M})}{\max(\mathbf{M}) - \min(\mathbf{M})}.\tag{4}$$

Figure 3a–f shows each **M** for facial expressions, such as angry, afraid, disgusted, sad, smiling, and surprised. Facial expressions are related to the activation of a distinct set of facial muscles [32,33]. When smiling, pixels around the mouth, which are related to the levator anguli oris muscle, change significantly when compared with the neutral image [34,35].

**Figure 3.** Weighted interpolation maps (**M**) and binary weighted interpolation maps (**B**) for facial expressions. (*AN*: angry, *AF*: afraid, *D I*: disgusted, *SA*: sad, *SM*: smiling, *SU*: surprised).

In a previous work [31], the WIM was extracted to measure the degree of change between neutral and variational images in pixels, and the query image and mean-variational image from the WIM were combined. However, the WIM has a problem in that the weight values of the locations associated with variations may be relatively lower in the normalization process when the maximum value obtained from Equation (3) is too large. If M(*u*, *v*) is 0.5, the location where the variation has occurred is not replaced by the pixel value of ˆ*I*(*u*, *v*). In this case, the pixel values for *I*(*u*, *v*) and ˆ*I*(*u*, *v*) will be mixed equally at *Inew*(*u*, *v*). This is a type of noise. To overcome this, we define B-WIMs **B** as follows:

$$\mathcal{B}(u,v) = \begin{cases} 1 & \mathcal{M}(u,v) > \theta \\ 0, & \mathcal{M}(u,v) \le \theta \end{cases} \tag{5}$$

In Equation (5), depending on the threshold, (*θ*), B(*u*, *v*) has a logical value of 0 or 1 (Figure 3g–l). The pixel value of *I*(*u*, *v*) is fully reflected when the value of B(*u*, *v*) is 0. Conversely, if B(*u*, *v*) becomes 1, the pixel value of *Inew*(*u*, *v*) will completely replace the pixel value of ˆ*I*(*u*, *v*). Accordingly, the WIM problem can be solved.

We use the structural similarity (SSIM) index [36] to find the optimal *θ*. The SSIM index evaluates a distorted image with respect to a reference image to quantify their structural similarity [37].

If *θ* is 0, all elements of **B** have a value of 1. Thus, **I***new* is obtained from Equation (2), which becomes the same as the variational image (ˆ **I**) in the auxiliary set. However, if *θ* is 1, the **I***new* is identical to the query image (**J**). To find the value of *θ* to generate a new image in which the variation is reflected in a balanced manner while maintaining the unique identity of the query image, we investigate SSIM(**I**,**I***new*), SSIM(ˆ **I**,**I***new*) of **I**, and ˆ **I** in the auxiliary set by increasing *θ* from 0 to 1, respectively. As shown in Figure 4, two SSIMs are balanced when *θ* is between 0.5 and 0.7. Thus, we set *θ* from 0.5 to 0.7, depending on the type of variation.

**Figure 4.** Structural similarity (SSIM) results for each variation (**a**) angry; (**b**) afraid; (**c**) disgusted; (**d**) sad; (**e**) smiling; (**f**) surprised.

Figure 5 shows the image samples generated by applying the B-WIM constructed from *θ* for the "smiling" variation for images in the auxiliary set. It is visually confirmed that the variations in the generated image are included while *θ* is less than 0.7.

**Figure 5.** Generated images from **I** and ˆ **I** for each threshold (*θ*).

### *2.2. Generation of New Images from a Query Image*

The new image (**I***new*) can be generated from a query image (**J**) as follows:

$$I\_{nvw}(\mu, \upsilon) = (1 - \mathcal{B}(\mu, \upsilon)) f(\mu, \upsilon) + \mathcal{B}(\mu, \upsilon) f(\mu, \upsilon). \tag{6}$$

Unlike during the phase of B-WIM extraction, a variational image (ˆ **J**) derived from **J** cannot be obtained in the phase of image generation. Therefore, ˆ **J** must be replaced with another image in a separate auxiliary set. In WIM, the mean image for each variation in the auxiliary set is used as ˆ **J**. Although it can be applied equally to all query images, morphological elements may be lost if the variation's own changes are large. For example, mufflers can be worn in various ways depending on a person's personality. Moreover, the designs of mufflers are also very diverse. Therefore, the mean image cannot preserve the form of all mufflers.

In this study, we select the neutral image (i.e., nearest neighbor) of the auxiliary set with a minimum Euclidean distance (L2-norm) from the query image based on the whole pixel [28].

$$d(\mathbf{I}, \mathbf{J}) = ||\mathbf{I} - \mathbf{J}||\_2^2 \tag{7}$$

Then, ˆ **J** is replaced by ˆ **I***id*, derived from **I***id*, where *id* is the index with a minimum distance from **J**(min(*d*(**I**,**J**))). Equation (6) is redefined as

$$I\_{new}(\mu, \upsilon) \approx (1 - \mathcal{B}(\mu, \upsilon)) f(\mu, \upsilon) + \mathcal{B}(\mu, \upsilon) f\_{id}(\mu, \upsilon). \tag{8}$$

Finally, **I***new* is generated from Equation (8). The overall procedure of the proposed method is summarized as follows:


### **3. Experiments**

### *3.1. Database*

In the experiment, all images were aligned to 80 × 80 pixels via affine transformation based on manually detected eye coordinates. Then, the image was compensated using histogram equalization [38].

We used the Bosphorus [32] and RaFD [39] databases in our face recognition experiments (Table 1). The Bosphorus database comprises images captured under seven different facial expression conditions from 58 subjects and includes various expressions, such as neutral, angry, disgusted, afraid, sad, smiling, and surprised. The neutral images ("indexed 5") were selected to generate new images, and the remaining images were used for the face recognition test. The RaFD database contains 536 images captured from 67 subjects. Each subject provided images of eight facial expressions (i.e., neutral, angry, contemptuous, disgusted, afraid, sad, smiling, and surprised). We used a neutral image ("indexed 6") to generate new images, and the remaining facial expression images were used for testing. Both databases applied practiced expressions using a Facial Action Coding System (FACS) [33] specialist. Furthermore, all subjects were tightly controlled through negative feedback to acquire the required activation of action units (AU).

**Table 1.** Characteristics of each facial expression database.


### *3.2. Face Recognition Results*

We compared the proposed method with other methods dealing with the SSPP problem (i.e., WIM, ICR, E(PC2)A+, SPCA+, (2D)<sup>2</sup> PCA, SLC, MVI, and SRGES methods). The proposed method, WIM, and SRGES generated as many images as the number of variations contained in the auxiliary set. With the ICR, the number of generated images depended on the *k* neighbors in the training set and the feature extraction method used by each database. In E(PC2)A+, the half-, first-, and second-order projected images of each neutral image were used as the training set. In SPCA+, seven images were enlarged from different *n*-order singular values for each neutral image in the training set. In (2D)<sup>2</sup> PCA, an image was generated using a two-directional PCA in the row and column directions of the 2D images. In SLC, 11 images from the neutral image were added to the training set, which included symmetric images and linear combinations of virtual images. In MVI, four low-resolution images (size 40 × 40, 26 × 26, 20 × 20, and 16 × 16) are generated from the neutral image and various scaling factors (i.e., 2, 3, 4, 5, respectively).

In this study, the face recognition performance of all methods was evaluated based on the given criteria. First, we measured the change in the face recognition rate according to the degree of variation in each database. Both databases contained similar facial expression variations, but they differed in the intensity of the facial expressions. In the Bosphorus database, the AUs were captured at their given peak intensity levels. In the RaFD database, there were large deviations in the intensities of expressions according to subject. Thus, the RaFD database was closer to the real

world than the Bosphorus one. Second, the face recognition performance was analyzed according to unsupervised and supervised learning methods. An unsupervised learning-based PCA [24] and supervised learning-based discriminant common vector (DCV) [40] were used to extract the features for face recognition. PCA extracted (N + N − 1) features, including the number of images of training data (N ) and the number of enlarged images (N ) from itself. DCV extracted (*c* − 1) features, where *c* was the total number of classes, regardless of the number of images. In the face recognition experiment, the recognition rates were measured using the maximum number of features extracted from each method. If a given set had been modeled properly, it could be expected to show high performance, regardless of the two methods. When evaluating the face recognition performance, the one nearest-neighbor rule was used with the *l*<sup>2</sup> norm as a classifier.

In this study, two protocols for face recognition were used [41]: "*Closed Set*" and "*Open Set*", according to the auxiliary set. These are described as follows:


First, we divided the given databases into face recognition and auxiliary sets. A total of 30 subjects from all subjects in each database were used for the face recognition set, and the others were used for the auxiliary set. Among the 210 and 240 images for 30 subjects in the Bosphorus and RaFD databases, respectively, neutral images were used as training data to construct PCA and DCV feature spaces for face recognition, and variational images were generated using the proposed methods from these images to enlarge the training set.

Table 2 shows the face recognition results for the "*Closed Set*" protocol. In the experimental results, the proposed method, WIM, and (2D)2PCA presented similar facial-recognition results within the same database, regardless of the feature-extraction manner. The other methods had a face recognition rate difference of up to 21.11% between each manner. Additionally, the proposed method and WIM showed high performance in the face recognition results according to the degree of variation by database. For the rest of the methods, the face recognition performance decreased by more than 15.87∼33.17% as the degree of variation increased. Figure 6 shows the recognition rates for a different number of DCV features. The proposed method gives a recognition rate of 96.11% and 93.33%, with 29 features for the Bosphorus and RaFD databases, respectively. As can be seen from Figure 6, the proposed method shows a comparable or better recognition performance to the other methods, regardless of the number of features. Finally, we confirmed that the proposed method was excellent in the absolute comparison of face recognition rates for both databases. Because the proposed method consistently showed high face recognition performance regardless of the various criteria, it could be inferred that a new image was generated by reflecting the various variations from the given neutral images.


**Table 2.** Face recognition results in *Closed Set* protocol.

Generally, the basic emotion group consists of angry, disgusted, afraid, sad, smiling, and surprised [42]. For the "*Open Set*" protocol, the auxiliary set containing six defined facial expressions consists of the AR [43], CK+ [44], Jaffe [45], PF07 [46], and Yale [47] databases. Additionally, it included various races and genders. To measure the degree of change from the variations, only subjects without glasses (occlusion) were used to construct the auxiliary set. The selected subjects had images of both neutral and facial expressions.

**Figure 6.** Face recognition results for a various number of DCV features in "*Closed Set*" protocol.

The AR database contained images from 85 subjects (37 males and 48 females) of different races [43]. We selected four facial expressions: neutral, angry, smiling, and screaming. The CK+ database contained 84 subjects from many different races. Image sequences contained changes in facial expressions over time. The neutral image at the start time and the facial expression image at the end time comprised the auxiliary set. We used seven facial expressions (i.e., neutral, angry, afraid, disgusted, sad, smiling, and surprised), except for "contemptuous." The Jaffe database included 10 subjects (only females) of Asian ethnic groups. Seven facial expressions of each subject were taken (e.g., neutral, angry, afraid, disgusted, sad, smiling, and surprised). The PF07 database contained the images of 200 subjects (100 males and 100 females) of Asian ethnic groups, all of whom provided four images with different facial expression conditions (i.e., neutral, angry, smiling, and surprised). The Yale database included 15 subjects (14 males and a female) of many different races. We used four facial expressions (i.e., neutral, sad, smiling, and surprised), excluding those with eyes closed or winking (Table 3).


**Table 3.** Facial expressions in each database in the auxiliary set.

Table 4 shows the face recognition results with a separate auxiliary set. Because the auxiliary set was constructed from separate databases, the face recognition experiment used images of all the subjects contained in each database.


**Table 4.** Face recognition results in *Open Set* protocol.

Depending on the degree of variation, the differences in face recognition rates were measured in the order of the proposed method: (5.77%), SRGES (8.76%), WIM (10.37%), SLC (10.69%), (2D)2PCA (15.38%), MVI (16.87%), ICR (17.38%), E(PC)2A2+ (24.77%), and SPCA+ (30.83%). For the criteria, the proposed method and SLC maintained high performance, whereas the remaining methods showed differences in face recognition performance. Generally, the proposed method showed the highest face recognition rates. Figure 7 shows the recognition rates for a different number of DCV features. The proposed method gives a recognition rate of 88.51% with 57 features and 87.21% with 66 features for the Bosphorus and RaFD databases, respectively. As can be seen from Figure 7, the proposed method shows the best recognition performance compared to the other methods for all number of features. This experiment also confirmed the superiority of the proposed method for each criterion.

On the other hand, from the results of Tables 2 and 4, it can be seen that the recognition rate in the "*Closed Set*" protocol was about 10% higher than that in the "*Open Set*" protocol. It is generally known that face recognition performance decreases as the number of subjects to be recognized increases [48]. In our experiment, however, we think the main reason for the difference between the results of Tables 2 and 4 is that, in the "*Closed Set*" protocol, the images included in the auxiliaries set had homogeneous characteristics, because they were taken under similar conditions of resolution, camera type, lighting conditions, etc. However, in the "*Open Set*" protocol, the auxiliary set comprised images from various kinds of databases, which differ from the query images.

**Figure 7.** Face recognition results for a various number of DCV features in "*Open Set*" protocol.

### **4. Discussion and Conclusions**

Building a face recognition system that works robustly in various environments involves difficulties in securing the data needed to learn recognition algorithms. Moreover, large-scale face recognition applications typically use databases that contain SSPPs. A single image is not sufficiently representative for face recognition. The SSPP problem makes using the feature extraction method in a supervised manner quite difficult, because the interclass variations are unknown. To overcome this, several methods have been proposed. However, there have been limitations in that these methods did not reflect facial characteristics that could have various variations.

We proposed an image generation method that uses a B-WIM that leverages the fact that the pixels of specific parts of the neutral face image vary significantly compared with other areas when there is an environmental variation in face recognition. The B-WIM statistically reflects the change in individual pixel values caused by the variation from the neutral and variational images included in the auxiliary set. For a given query image (neutral image), the proposed method creates a new variational image that reflects the characteristics of the variation while maintaining the unique characteristics of the face in the query image based on B-WIM. Through this, a training dataset containing only one sample per person can be made into a richer set that includes variational images for each person, further improving the performance of the face recognition system.

The proposed method has the following advantages. The proposed method does not require a large amount of computation or a large dataset for creating new images. When the number of pixels in an image is *n*, while SPCA+ has the complexity of *O*(*n*2), the complexity of the proposed method is *O*(*n*). Some methods, such as ICR, E(PC2)A+, (2D)<sup>2</sup> PCA, SLC, and MVI, require similar computations as the proposed method but do not address specific variations. In contrast, the proposed method generates high-quality variational images for query images in real-time, effectively improving the performance of existing face recognition systems at a low cost. Face recognition experiments using Bosphorus and RaFD databases showed that the proposed method outperformed the existing methods for solving the SSPP problem. In addition to general facial recognition algorithms, images generated using the proposed method can be utilized in the study of various facial images, including the fake image detection algorithms [49,50].

On the other hand, by comparing the recognition rates for two protocols of face recognition, "*Closed Set*" and "*Open Set*," we found that the quality of the image created using the proposed method was affected by the images included in the auxiliary set. Although the proposed method can effectively generate new images for a specific variation, it does not control the degree of variation or handle more than two variations simultaneously. It is expected that the small sample-size problems, including the SSPP problem, can be solved more effectively by subdividing the degree of variation

within the proposed method's algorithmic structure and applying the interpolation maps for two or more types of variations together. We leave these problems to future works.

**Author Contributions:** Y.L. and S.-I.C. designed the experiments and drafted the manuscript. All authors have read and agreed to the published version of the manuscript.

**Funding:** The present research was supported by a National Research Foundation of Korea(NRF) grant funded by the Korean government (MSIT) (No. 2018R1A2B6001400) and the MSIT(Ministry of Science and ICT), Korea, under the ICAN(ICT Challenge and Advanced Network of HRD) program(IITP-2020-2020-0-01824) supervised by the IITP(Institute of Information & Communications Technology Planning & Evaluation).

**Conflicts of Interest:** The authors declare no conflict of interest.

### **Abbreviations**


### **References**


© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
