*Article* **Development of a Face Prediction System for Missing Children in a Smart City Safety Network**

**Ding-Chau Wang 1, Zhi-Jing Tsai 2, Chao-Chun Chen <sup>2</sup> and Gwo-Jiun Horng 3,\***


**Abstract:** Cases of missing children not being found are rare, but they continue to occur. If the child is not found immediately, the parents may not be able to identify the child's appearance because they have not seen their child for a long time. Therefore, our purpose is to predict children's faces when they grow up and help parents search for missing children. DNA paternity testing is the most accurate way to detect whether two people have a blood relation. However, DNA paternity testing for every unidentified child would be costly. Therefore, we propose the development of the Face Prediction System for Missing Children in a Smart City Safety Network. It can predict the faces of missing children at their current age, and parents can quickly confirm the possibility of blood relations with any unidentified child. The advantage is that it can eliminate incorrect matches and narrow down the search at a low cost. Our system combines StyleGAN2 and FaceNet methods to achieve prediction. StyleGAN2 is used to style mix two face images. FaceNet is used to compare the similarity of two face images. Experiments show that the similarity between predicted and expected results is more than 75%. This means that the system can well predict children's faces when they grow up. Our system has more natural and higher similarity comparison results than Conditional Adversarial Autoencoder (CAAE), High Resolution Face Age Editing (HRFAE) and Identity-Preserved Conditional Generative Adversarial Networks (IPCGAN).

**Keywords:** face aging; generative adversarial network; StyleGAN2; FaceNet; missing child

#### **1. Introduction**

We will divide this paper into three subsections to describe the introduction: "Status of Missing Children's Cases", "Problems of Current Face Aging Methods", and "Contribution".

#### *1.1. Status of Missing Children's Cases*

According to the Federal Bureau of Investigation's National Crime Information Center (NCIC) Missing Person and Unidentified Person Statistics, there will be 365,348 children missing in the United States in 2020 [1]. According to the National Crime Agency's Missing Persons Statistics, it is estimated that over 65,800 children will go missing between 2019 and 2020 [2].

From the above statistical results, it is known that there has always been a situation of missing children. Therefore, we built the Face Prediction System for Missing Children to predict the face of children when they grow up and help parents or police search, detect, and identify missing children. Moreover, we applied our system to the Smart City Safety Network. According to S. P. Mohanty et al.'s description [3], a smart city includes smart infrastructure, smart transportation, smart energy, smart health care, and smart technology. The key to transforming traditional cities into smart cities is information and

**Citation:** Wang, D.-C.; Tsai, Z.-J.; Chen, C.-C.; Horng, G.-J. Development of a Face Prediction System for Missing Children in a Smart City Safety Network. *Electronics* **2022**, *11*, 1440. https:// doi.org/10.3390/electronics11091440

Academic Editor: Byung-Gyu Kim

Received: 16 February 2022 Accepted: 25 April 2022 Published: 29 April 2022

**Publisher's Note:** MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

**Copyright:** © 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https:// creativecommons.org/licenses/by/ 4.0/).

communication technology (ICT). Smart cities use ICT to solve a variety of urban problems. In addition, M. Lacinák et al. emphasized the importance of safe cities [4,5]. They describe that every smart city must also be a safe city, and a safe city should be regarded as a part of a smart city. A safe city system should include the following features: smart safety systems for surveillance, search, detection and identification, etc. The purpose of our system conforms to the concept of smart city safety, and then we use the concepts of IoE and AIoT to implement our system and form a network. More details can be found in Section 3.

To predict future faces currently, face aging image generation methods in the field of machine learning can be used. However, the existing face aging model only considers the facial features that only older people have. In fact, head size and genetics can also affect appearance. Therefore, current face aging methods cannot well predict children's faces when they grow up. Section 1.2 describes the problems of current face aging methods in detail.

#### *1.2. Problems of Current Face-Aging Methods*

We refer to many face-aging image generation methods, such as CAAE [6] extended by VAE, F-GAN [7], HRFAE [8] and IPCGAN [9] extended by GANs. Details can be found in Section 2.3. Overall, these face-aging methods add or smooth some irregular wrinkles on the face, making the generated results appear older or younger. However, these methods do not work for children under 12; they only work for adults. According to medical research [3–10], the period from 0 years old to adolescence is the fastest-growing period for human beings. The appearance (including facial appearance, body shape, etc.) will change greatly. Therefore, if we only consider the facial features that only older people have and do not consider other factors (such as head size and genetics, etc.), it is impossible to predict children's faces when they grow up.

Figure 1 illustrates the problems of current face-aging methods. In Figure 1, (a) is the original image, whose age is between 0 and 20 years old, while (b)~(d) is the face image converted by using the original image through the Group-GAN model; (b) is between 20 and 40 years old, (c) is between 40 and 60 years old, and (d) is over 60 years old.

**Figure 1.** Group-GAN face aging results.

We can see from Figure 1 that the child and the adult are transformed from (a) to (b)~(d), respectively. Two problems arise during this transition:

1. The difference between the child and adult transition from (a) to (b) is small, especially when the child (b) does not look like they are between 20 and 40 years old. The reason is that most aging models only consider the facial texture and do not consider that the head's size will also change with age;

2. The result of a child's transition from (a) to (d) is very unnatural, while the transition of an adult from (a) to (d) is relatively natural. The reason is not only because the size of the head is not taken into account, but also because people grow most rapidly before puberty (children under the age of 12), so the appearance changes very greatly. Therefore, it is not enough to only consider the facial texture for face prediction.

Overall, because existing aging models only consider the appearance characteristics that older adults have, they cannot predict children's future faces.

We propose a Face Prediction System for Missing Children. In addition to taking into account the facial features of children before their disappearance, it also considers the facial features of their blood relatives. According to genetics [10–13], life on earth mainly uses DNA in the blood as genetic material so that the offspring will have the parent's traits (traits including appearance and disease, etc.). In addition, according to Mendelian inheritance [14,15], human looks are mainly determined by genetics, which means that children are born with traits such as appearance and diseases that are part of their parents. Our system takes into account human genetics and then estimates and predicts children's future faces. It is a more reasonable estimate than other aging models. No one has proposed this prediction method to date.

#### *1.3. Contribution*

Suppose a parent wants to search, detect and identify their child from a group of unidentified children. If every unidentified child had a DNA paternity test, there would be a lot of cost and time waiting for the test. If parents used our system, they could quickly confirm the possibility of blood relation with any unidentified child. Additionally, there are the following benefits:


Overall, we have the following two contributions:


#### **2. Related Work**

#### *2.1. Generative Adversarial Networks (GANs)*

GANs [15–17] are unsupervised learning networks trained only through images without labels. A GAN is mainly composed of a generator and discriminator network. The final goal of a GAN is for the generator to randomly create real-looking images that cannot be distinguished from training images. Many scholars have developed different methods and applications based on the concept of GANs. For example, in the field of image generation, the Progressive GAN [18] proposed by NVIDIA Tero Karras et al. can randomly generate high-resolution images. There is also StyleGAN [19] proposed in 2019, which follows the training network of Progressive GAN and has the function of style conversion, which can control the changes of different styles of images.

However, due to the droplet and phase artifact problems in StyleGAN, StyleGAN2 [20] was proposed in 2020 to solve the above problems and make the output results more natural. StyleGAN has had a huge impact on image generation and editing, and many scholars have used StyleGAN for different studies. For example, Image2StyleGAN [21,22], SEAN [23], Editing in Style [24], StyleFlow [25], Pixel2style2pixel [26], StyleCLIP [27]

and StyleMapGAN [28] are all image generation and editing methods developed based on StyleGAN.

Figure 2 is the generator architecture of StyleGAN [18–20]. The generator of StyleGAN consists of the mapping network and the synthesis network. The mapping network is a non-linear network using an 8-layer MLP. Its input is the latent code (or latent variable) *z* in latent space *Z*, and the output is the intermediate latent code (or dlatents) *w* in intermediate latent space *W*. The latent space is simply a representation of compressed data in which similar kinds of data points will be closer in the latent space. Latent space is useful for learning data features and finding simpler data representations for analysis. We can interpolate data in the latent space and use our model's decoder to 'generate' data samples. The purpose of the mapping network is to convert the input *z* to *w*. Because the use of *z* to control image features is limited, a mapping network is needed to convert *z* to *w*, which is used to reduce the correlation between features and to control the generation of different images.

**Figure 2.** Generator architecture of StyleGAN [18–20].

The synthesis network is used to generate images of different styles and add affine transformation A and random noise B to each sub-network layer. A is used to control the style of the generated image, which can affect the pose of a face, identity features, etc. B is used for the details of the generated image and can affect details such as hair strands, wrinkles, skin color, etc.

Figure 3 shows the style mixing result of StyleGAN. Sources A and B are pre-trained models using StyleGAN to project the images into the corresponding latent space. Finally, the images are directly generated by the latent code. The coarse styles from source B mainly control the coarser low-resolution features (no more than 8 × 8), affecting posture, general hairstyle, facial shape, etc. The middle styles from source B mainly control the finer features of the middle resolution (16 × 16 to 32 × 32), including facial features, hairstyles, opening or closing of eyes, etc. The fine styles from source B mainly control the relatively high-quality, high-resolution features (64 × 64 to 1024 × 1024), affecting the color of eyes, hair, and skin.

**Figure 3.** Examples that were generated by a mixture of Source A and B latent codes using Style-GAN [19,20].

Table 1 shows the training results of StyleGAN and StyleGAN2. The ↑ indicates that higher is better and ↓ that lower is better. In 2020, StyleGAN2 improved the shortcomings of SyleGAN, including the droplet and phase artifact problem, and enhanced the quality of SyleGAN, including weight demodulation, path length regularization and no progressive growth. Because StyleGAN2 can generate higher quality images and train faster, this study mainly uses the pre-trained model of StyleGAN2 for style mixing and image projection processing.


**Table 1.** Training results of StyleGAN [19] and StyleGAN2 [20].

The " + " in the table represents the experimental results based on StyleGAN (A) plus (B) to (F) configurations.

#### *2.2. FaceNet*

FaceNet [29] is a unified framework for solving recognition and verification problems proposed by Google. According to Florian Schroff et al., FaceNet mainly uses convolutional neural networks to analyze face information and project the information into Euclidean space. The similarity between the two can be directly calculated by calculating the distance in the space. In 2015, FaceNet received the highest accuracy score of 99.63% in LFW [30] and received attention for this. At present, the development of face recognition and verification is quite mature, including OpenFace [31], Deep-Face [32], VGG-Face [33,34], DeepID [35–38], ArcFace [39] and Dlib [40], all of which have over 90% accuracy. The highest accuracy face-recognition model today is VarGFaceNet [41]. We used Euclidean distance (L2) to calculate distance and converted it into similarity.

#### *2.3. Image Generation Method of Face Aging*

In the field of image generation, variational autoencoder (VAE) [42] and generative adversarial network (GAN) [17] methods are the mainstream.

#### 2.3.1. VAE

Research methods extended by VAE include the adversarial autoencoder (AAE) [43,44], the conditional adversarial autoencoder (CAAE) [6], and the conditional adversarial consistent identity autoencoder (CACIAE) [45], etc. AAE is a training method that combines the encoder-decoder idea of VAE and the generator-discriminator of GAN. CAAE is a face-aging method proposed by Z. Zhang et al. It builds a discriminator based on AAE to make the generated images more realistic. CAAE can learn the face manifold and achieve smooth age progression and regression so that the results can appear more aged or younger. In addition, the CACIAE proposed by Bian et al. can reduce the loss of identity information, making the results more realistic and age-appropriate. In the experimental results, our system is compared with CAAE. Since CAAE only considers facial lines, it cannot predict the appearance of children when they grow up.

#### 2.3.2. GANs

The methods of synthesizing face images using GANs can be divided into two categories: translation-based and condition-based.

#### Translation-Based Method

The translation-based face image synthesis method converts any set of style images into another set of style images. This concept first came from Cycle-GAN [46], proposed by Zhu et al. Its advantage is that it does not require the pairing of two collection domains, making it available for face-style transfer, unlike pix2pix [47], which must have two or two. Only paired data can be used for training. The disadvantage is that it can only be converted between two domains, so later, Choi et al. proposed StarGAN [48], which can learn multiple domains and solve Cycle-GAN's problem.

In terms of face-aging models, Palsson et al. proposed F-GAN [7], based on the style transfer architecture developed by Cycle-GAN. F-GAN combines the advantages of Group-GAN and FA-GAN. When the age span is large (about 20 years old or more), because the effect is better, Group-GAN is used for face conversion, and FA-GAN is used on the contrary. The problem with F-GAN is that it cannot be converted naturally, and the image quality is low. After 2018, because StyleGAN provides an FFHQ dataset, it became easier to generate high-quality images, but it has artifact problems. Subsequently, Shen et al. proposed InterFaceGAN [49], which can semantically edit the learned latent semantic information (for example, changing age, gender and angle, etc.) and repair the artifacts in the image, making the resulting image more natural. Although it produces higher-quality images, it is not suitable for predicting the appearance of children because it only takes into account the texture of the face.

#### Condition-Based Method

The condition-based face image synthesis method can be regarded as a supervised GAN. It adds an additional condition to the generator's input and the discriminator. The condition can be a label or a picture, etc. The function guides the generator and the discriminator towards training on this condition. This concept first came from cGAN [50–52], proposed by Mirza et al. It has a better effect than the original GAN, so it has been widely used in the future.

In terms of face-aging models, Wang et al. proposed IPCGAN [9], an architecture that successfully generates new synthetic face images and preserves identities in specific age groups. It generates realistic, age-appropriate faces and guarantees that the synthesized faces have the same identity as the input image. In the experimental results, our system is compared with IPCGAN. Since IPCGAN only changes the facial lines, it cannot predict the appearance of children when they grow up.

In addition, HRFAE [8], proposed by Yao et al., combines age labels and latent vectors and can be used for face age editing on high-resolution images. The core idea is to create a latent space containing face identities and a feature modulation layer corresponding to the individual's age and then combine these two elements so that the generated output image is the specified target age. In the experimental results, our system is compared with HRFAE. Because HRFAE only considers facial lines, it cannot predict the appearance of children when they grow up.

#### **3. Method**

We propose a Face Prediction System for Missing Children, whose purpose is to predict children's future faces. It allows parents to quickly and inexpensively confirm the possibility of blood relation with any child. When parents search for missing children, our system helps to eliminate low similarity matches and narrow the search. Our system considers the respective features of the following two face images to predict the future face, including face images of the child before the disappearance and face images of the blood relatives. Our system combines StyleGAN2 and FaceNet methods to achieve prediction. StyleGAN2 is used to style mix two face images. FaceNet is used to compare the similarity of two face images. The input is an image of the missing child available before the disappearance and multiple images of family members related by blood. The output is a prediction result. More details can be found in Sections 3.1–3.4.

At the application level, we apply our Face Prediction System for Missing Children and the issues of searching for missing children to the concepts of IoE and AIoT, as shown in Figure 4, which will be described in detail below.

**Figure 4.** Our system is applied to the concepts of IoE and AIoT [53,54].

On the left side of Figure 4 is the IoE that combines machine-to-machine (M2M), people-to-people (P2P), and people-to-machine (P2M) connections. The difference between IoE and IoT is that IoT only focuses on the pillar of things, while IOE includes four pillars, namely things, people, process and data. IoE is the intelligent connection of the four

pillars. The definition of the process is to provide the right information to the right person or machine at the right time to make the connection between people, things, and data more valuable. M2M is defined as the transmission of data from one machine or thing to another machine, including sensors, robots, computers, mobile devices, etc. These M2M connections can be considered IoT. P2P is defined as the transmission of information from one person to another. At present, P2P is mainly realized by mobile devices (such as PCs, TVs, and smartphones) and social networks (such as Facebook, Twitter, and LinkedIn). P2M is defined as the transmission of information between people and machines. People conduct complex data analysis through machines to obtain useful key information and help people make informed decisions. The following will explain the application mode of our system with the concepts of M2M, P2P and P2M.


Our system mainly combines StyleGAN2 and FaceNet methods. StyleGAN2 is used to mix two images, and FaceNet is used to compare the similarity of the two images. The architecture of this system will be described in detail below.

#### *3.1. Overview of the System Architecture*

Figure 5 shows a flowchart of the image processing steps, divided into three main parts: data preprocessing, phase 1: filtering the best new face image, and phase 2: predicting the age and appearance of a missing child.


The details of these three parts will be described in order below.

**Figure 5.** Flowchart of image processing steps.

#### *3.2. Data Preprocessing*

Data preprocessing consists of the following two steps:


#### 3.2.1. Dlib Face Alignment Module

Figure 6 shows the flowchart of the Dlib Face Alignment Module, which mainly corrects and truncates each face in the original image and outputs it as a face image of size 1024 × 1024. The Dlib Face Alignment Module contains three functions: 'Face Detector', 'Facial Landmark Predictor' and 'Face Alignment'.

**Figure 6.** Dlib Face Alignment Module Flowchart.


#### 3.2.2. StyleGAN2 Project Image Module

Figure 7 shows the StyleGAN2 projection process, whose input is an image of a missing child (the child in Figure 6). The *Projection image* is a function provided by StyleGAN2, which can iterate continuously on an input image (missing child face), producing a very similar one (a projection). In Figure 6, the projected image when iteration is 1 is the projected image after StyleGAN2 training, which is the default image of StyleGAN2. When iterating the 1000th time, the result of the projected image is very similar to the input image (missing child face), so we stored the dlatents this time as a NumPy file to be used for StyleGAN2 style mixing or interpolation in the future.

**Figure 7.** StyleGAN2 projection image process.

#### *3.3. Phase 1: Filter the Best New Face Image*

The first phase of the system is mainly to filter the best new face images. The input data is a projected image of a missing child and several projected images of family members. The output is an image of the best new face, one of the 36 mixed faces. There are four modules in the first processing phase: the Similarity Sequence Module, StyleGAN2 Style Mixing Module, FaceNet Face Compare Module and Best New Face Filter Module.

#### 3.3.1. Similarity Sequence Module

The Similarity Sequence Module focuses on selecting the two family members most similar to the child from multiple family members. The missing child is first compared with each family member using FaceNet. All the similarities are ranked in descending order, and the images of the top two family members with the highest similarities are output.

#### 3.3.2. StyleGAN2 Style Mixing Module

The StyleGAN2 Style Mixing Module inputs the top two dlatents with the highest similarity in the first phase, and after the StyleGAN2 style mixing process, a total of 36 mixed new faces will be generated. For example, Figure 8 is the StyleGAN2 style mixing result.

**Figure 8.** StyleGAN2 Style Mixing result.

#### 3.3.3. Best New Face Filter Module

The Best New Face Filter Module mainly filters one of the 36 new faces mixed by StyleGAN2 as the best new face. In the first phase, this module mainly uses the similarity percentage metric to evaluate the advantages and disadvantages of 36 new faces. These 36 new faces *n*1, *n*2, *n*3, ··· , *n*36 will get a weight *W*1,*nk* , respectively. Then the system will rank *W*1,*nk* from small to large, and the minimum min*W*1,*nk* is the best new face.

$$\mathcal{W}\_{1,n\_k} = \mathcal{W}\_{1,P\_{n\_k}} + \mathcal{W}\_{1,S\_{n\_k}} \tag{1}$$

$$\left| \mathcal{W}\_{1, P\_{\pi\_k}} = \left| P\_{\mathbb{C}, Top1} - P\_{n\_k, Top1} \right| = \left| \frac{\mathcal{S}\_{\mathbb{C}, Top1}}{\mathcal{S}\_{\mathbb{C}, Top1} + \mathcal{S}\_{\mathbb{C}, Top2}} - \frac{\mathcal{S}\_{n\_k, Top1}}{\mathcal{S}\_{n\_k, Top1} + \mathcal{S}\_{n\_k, Top2}} \right| \tag{2}$$

$$\left| \mathcal{W}\_{1, \mathcal{S}\_{\mathcal{W}\_k}} = \left| \mathcal{S}\_{\mathcal{C}, \mathcal{Top1}} - \mathcal{S}\_{\mathcal{W}\_k, \mathcal{Top1}} \right| + \left| \mathcal{S}\_{\mathcal{C}, \mathcal{Top2}} - \mathcal{S}\_{\mathcal{n}\_k, \mathcal{Top2}} \right| \tag{3}$$

Here, Equation (1) is the formula for weight *W*1,*nk* (refer to Table 2 for symbolic meaning), which mainly calculates the similarity percentage of *<sup>W</sup>*1,*Pnk* and similarity of *W*1,*Snk* between children and 36 new faces. Then, *W*1,*Pnk* and *W*1,*Snk* is added and called the Similarity Percentage Metric. Equation (2) is the similarity percentage *W*1,*Pnk* , which mainly calculates *PC*,*Top*<sup>1</sup> and *PC*,*Top*1. The smaller the gap between the two, the better, indicating that the percentage value of the two is closer. Equation (3) is the similarity of *<sup>W</sup>*1,*Snk* ; the smaller the formula, the better, indicating that the new face image is more similar to the family. After calculating *W*1,*nk* , the system will sort each weight, and the smallest weight min*W*1,*nk* is the best new face.

**Table 2.** Symbol Definition.


#### *3.4. Phase 2: Predicting the Age and Appearance of a Missing Child*

The second phase of the system focuses on predicting the current age of the missing child. The input data are the best new face and the missing child image, and the output data are the prediction result. A total of four modules were used in the second phase of processing, in the order of Data Preprocessing, StyleGAN2 Style Mixing Module, FaceNet Face Compare Module and Best New Face Filter Module. The two modules, Data Preprocessing and StyleGAN2 Style Mixing Module, operate in the same way as the corresponding modules in Phase 1, while the other modules are different.

#### 3.4.1. FaceNet Face Compare Module

The FaceNet Face Compare Module mainly compares the best new face and the missing child image with each new face in the second phase. Finally, it records the similarity comparison information in the JSON file for subsequent analysis.

#### 3.4.2. Best New Face Filter Module

The Best New Face Filter Module mainly selects the best prediction result from 36 new faces in the second phase. In the second phase, the similarity percentage is mainly used to evaluate the advantages and disadvantages of 36 new faces. These 36 new faces *n*1, *n*2, *n*3, ··· , *n*36 will get a weight *W*2,*nk* , respectively. Then the system will rank *W*2,*nk* from small to large, and the minimum min*W*2,*nk* is the best new face.

$$\mathcal{W}\_{\mathcal{Z}, \mathfrak{m}\_k} = \left| \mathcal{S}\_{\mathcal{C}, B} - \mathcal{S}\_{\mathfrak{n}\_k, B} \right| \tag{4}$$

Equation (4) is the formula for weight *W*2,*nk* (refer to Table 2 for symbolic meaning), which is the similarity gap between 36 new faces and the best new face. The smaller *W*2,*nk* is, the more it means that the new face will be more similar to the best new face. After calculating *W*2,*nk* , the system will sort each weight and the smallest weight min*W*2,*nk* is the best new face.

#### **4. Experiment**

Figure 9 shows the experimental results of this system. The input data for this experiment were obtained from members of the same family. The first column in Figure 9 is the image of missing children. These three images are of different people; they are about 3 years old. The second column is images of family members or relatives of the missing children; the third column is our system, which contains the first and second phases. The input in the first phase is an image of the missing child available before their disappearance and multiple images of family members who are related by blood (dotted box in Figure 9), and the output is the best new face image, which is a mixed image of the facial features of two blood relatives. The input in the second phase is an image of the missing child available before their disappearance and an image of the best new face. The fourth column is the predicted results (output) of our system; the fifth column is the similarity comparison between the predicted result and the expected result, and the sixth column is the expected output, which is the ground truth, the faces of the missing children at the age of 20. The system mainly uses the face compare function provided by SKEye [55] for similarity comparison, and its similarity comparison refers to Algorithm 1. The predictions of the three sisters were compared to the expected output, and the results were 77%, 76% and 77%, respectively.

**Figure 9.** A comparison experiment of the similarity between the predicted image and the expected image.

From the physical appearance, it is difficult for humans to identify the gender of children under the age of three. We observe the child in the second row in Figure 9; she looks like a male, and the predicted image also looks like a male, but this does not affect the final similarity comparison of our prediction system. Because our system excludes human subjective judgments (including hairstyles) and only compares the similarity of facial features, the system will not be misled by physical appearance.


```
Input: F1: Features of the first face; F2: Features of the second face;
```
**Output:** *S*: Similarity between *F*<sup>1</sup> and *F*2;

```
1: Load F1 and F2;
2: Get F1 and F2 base64 code;
```

Figure 10 shows the comparison diagram of our system, CAAE [6], HRFAE [8] and IPCGAN [9]. The first line is the input child image; Line 2 is the expected output; Lines 3~6 are the prediction results of the system, CAAE, HRFAE and IPCGAN and the similarity comparison results with the expected output. The similarity can correspond to Table 3. It can be seen from Figure 10 that, compared with other aging models, this system can produce more natural and high-resolution images, and the prediction accuracy is the highest, about more than 75%, which means that this system can well predict the appearance of children when they grow up.

Our system works for non-special families, direct blood relatives, and images with intact and undamaged faces. The following will list the conditions that do not apply because these reasons may result in low similarity:


**Figure 10.** Comparison diagram of our system.

**Table 3.** Similarity comparison results of our system.


#### **5. Conclusions**

This study proposes a Face Prediction System for Missing Children, which can enable parents to quickly confirm whether they have the possibility of a parent-child relationship with any missing child, hoping to help parents find the missing child. The system combines FaceNet and StyleGAN2 methods to predict the appearance of missing children at their present age through similarity comparison and style mixing. Finally, we compare this system with other aging models, including CAAE, HRFAE and IPCGAN. Experiments show that this system has the highest prediction accuracy compared with other aging models, and the prediction results are of higher picture quality and natural.

**Author Contributions:** Conceptualization, G.-J.H.; methodology, Z.-J.T. and D.-C.W.; software, Z.- J.T.; validation, Z.-J.T.; investigation, Z.-J.T. and D.-C.W.; resources, D.-C.W.; writing—original draft preparation, Z.-J.T. and G.-J.H.; writing—review and editing, Z.-J.T. and G.-J.H.; supervision, C.- C.C.; project administration, G.-J.H. All authors have read and agreed to the published version of the manuscript.

**Funding:** This research received no external funding.

**Informed Consent Statement:** Informed consent was obtained from all participants involved in the study.

**Acknowledgments:** This work was supported in part by the Ministry of Science and Technology (MOST) of Taiwan under Grants MOST 110-2221-E-218-002 and in part by the "Allied Advanced Intelligent Biomedical Research Center, STUST" from Higher Education Sprout Project, Ministry of Education, Taiwan, and in part by the Ministry of Science and Technology (MOST) of Taiwan under Grant MOST 110-2221-E-218-007.

**Conflicts of Interest:** The authors declare no conflict of interest.

#### **References**

