Next Article in Journal
“In Vivo” Validation of 3D-Printed Innovative Surgical Template for Lumbar Spinal Arthrodesis
Next Article in Special Issue
Choice between Surgery and Conservative Treatment for Patients with Lumbar Spinal Stenosis: Predicting Results through Data Mining Technology
Previous Article in Journal
A Survey of Machine Learning Models in Renewable Energy Predictions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Chinese Font Generation System Reflecting Emotions Based on Generative Adversarial Network

1
School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
2
Major of Electrical Engineering and Electronics, Graduate School of Engineering, Kogakuin University, Tokyo 163-8677, Japan
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2020, 10(17), 5976; https://doi.org/10.3390/app10175976
Submission received: 2 August 2020 / Revised: 22 August 2020 / Accepted: 26 August 2020 / Published: 28 August 2020

Abstract

:
Manual font design is difficult and requires professional knowledge and skills to perform. Therefore, how to automatically generate the required fonts is a very challenging research task. On the other hand, there are few people who have studied the relationship between fonts and emotions, and common fonts generally cannot reflect emotional information. This paper proposes an Emotional Guidance GAN: an automatic Chinese font generation framework based on Generative Adversarial Network (GAN), which enables the generated fonts to reflect human emotional information. First, an elaborated questionnaire system was developed from Tencent company, which aims to quantitatively figure out the relationship between fonts and emotions. A visual expression recognition part is designed based on the trained model to provide a font generation module with conditional information. Moreover, the Emotional Guidance GAN (EG-GAN) with EM Distance and Gradient Penalty, as well as classification strategies, is proposed to generate new fonts with combined multiple styles that infer by an expression recognition module. The results of the evaluation experiments and the resolution of the synthesized font characters show the credibility of our model.

1. Introduction

Recently with the development of advertising and multiple media technology rushing into our daily lives, text and characters are generally taken, in the area of publishing-and-printing and screen display, to show some emotions. The reason why most of these application scenarios express emotions using fonts that are appropriate for the characters is because characters alone cannot express enough emotions to users, especially for hieroglyphs such as Chinese and Japanese. Therefore, during the process of font design, designers tend to incorporate emotional information in them. [1,2] show that characters with specific styles can effectively exert the expression function of emotions, and at the same time, it can be a designed font that can connect characters with people’s emotions and resonate with the audience. What is more, modern Chinese character font expressions need to adapt to the different visual pursuits of modern people according to different purposes, environments, and artistic styles. It is very practical to study the relationship between fonts and emotions. Although the fonts’ emotion analysis has been well studied for analyzing the expression of the fonts’ characteristic, which is deemed necessary to study the emotion that the font shows, quantitative research on the emotional intention is rarely utilized. Considering the discussion above, we propose a questionnaire system to combine both qualitative and quantitative aspects to exactly analyze the specific font attached to the corresponding facial expressions.
Different from western characters, common Chinese characters include more than 3500 characters, and each character has various fonts with different thickness, length, and straightness, as well as partial strokes. The design of Chinese fonts is very time-consuming and requires professional skill, artistic sense, and expertise in calligraphy. Compared to Latin alphabet font generation [3], as proposed by Hideaki Hayashi, the task of Chinese character generation is more complicated and difficult. Most previous works have focused on local strokes of characters rather than the whole. In contrast, Zi2Zi [4] is directly originated from the Pix2Pix [5] model, which uses paired images as training data to generate fonts automatically. However, the Zi2Zi model is hard to train due to mode collapse and instability. Since the performance of Generator and Discriminator is difficult to balance without knowing the metrics of the model training, some generated font images are somewhat blurred. Therefore, we propose the use of Wasserstein distance [6], which can reflect the training performance by measuring the difference between generated images and real images. We also employ Gradient Penalty [7] to enhance the stability of the model and improve the quality of the generated images.
Much research has been done to find what to recommend and how to choose appropriate fonts for users. Wada et al. [8] propose a system for automatically generating fonts reflecting Kansei. However, the proposed system operates based on the genetic algorithm that directly operates the parameters related to Kansei engineering many times to creates various kinds of fonts, which is complicated and time-consuming. Therefore, we propose a recommended automatic font generation system, which allows ordinary users who do not need professional font design capabilities to change the font to convey certain emotions. In order to obtain accurate style results, we incorporate a classification loss in the model. The novelty of this research is to generate a new font that more appropriately reflects the emotion contained in the input image.
Our main contributions of this paper can be summarized as follows:
(1)
We propose and design a questionnaire system to quantitatively and qualitatively study the relationship between fonts and facial expressions. Data analysis shows that the system has high credibility, and the result provides a dataset for further research.
(2)
In our model, we propose an Emotional Guidance GAN (EG-GAN) algorithm; employing an emotion-guided operation on the font generation module, the automatic Chinese font generation system is able to generate new style of Chinese fonts with corresponding emotions.
(3)
We incorporate EM Distance, Gradient Penalty, and a classification strategy to enable the font generation module to generate high-quality font images and make sure that each single font has a consistent style.
(4)
We conduct various experimental strategies on various Chinese font datasets. The experimental results are utilized as the basis for other questionnaires we propose to analyze, and it shows that generated fonts are credible with specific emotions.
The rest of this paper is shown as follows. Section 2 discusses about the review of related literatures, and a proposed questionnaire is elaborated in Section 3. A detailed model is presented in Section 4. Experimental results and analyses are described in Section 5, and the conclusion is drawn in Section 6.

2. Related Works

2.1. Font Emotion Research

The imagery and characteristics of Writing Characters have been keenly studied for decades. Most of this research focus on the evolution [9] and recognition [10], as well as the construction of Chinese characters [11]. Recently, some studies have demonstrated that fonts can reflect relative emotions [1,2,12,13]. Accordingly, these studies show that fonts can express some extra emotions besides characters’ meaning. [12] shows that a character’s features are based on the method that proposed which emotion is affected by the elements of character such as Genre, Serif, Tool kind, and Aspect ratio, which includes thickness, length, and straightness as well as the partial strokes. Wada et al. [8] propose a system to automatically generate font reflecting Kansei. In this paper, we focus on exploring which fonts can reflect facial emotions by using the questionnaire method to estimate the correlation between fonts and emotions. However, with regard to the deficiencies of these proposed methods of analyzing fonts’ emotion. Promotions are incorporated into our work to enhance the reliability of relationship between alternative fonts and emotions.

2.2. Generative Adversarial Network (GAN)

Previous studies [14,15] have introduced a generative model based on Generative Adversarial Network (GAN), which includes two parts as follows: Generator G to generate a distribution from a random noise, and the discriminator D is adversarially trained to distinguish the credible degree. However, traditional GAN tends to have a phenomenon of model collapse and gradient disappearance, which is elaborated in [6]. Various variants to improve the quality of GAN have been done in different tasks. For instance, the results of Least squares GANs (LS-GANs) [16], Wasserstein-GANs (WGANs) [6], and WGAN with gradient penalty (WGAN-GP) [7] show that using some useful tricks can be helpful to generate high-quality images.

2.3. Automatic Font Generation

Various research has been carried out in previous studies on automatic font generation. Different from Western letters generative tasks [3,17], Chinese characters generation tasks with the problems time-consuming and tough. Former studies are usually based on the hierarchical representation of strokes to represent Chinese characters [18,19]. In recent years, with the wide application of deeply generative model, the Generative Adversarial Network (GAN) [15] has been one of the very important models that have been applied for font generative tasks [4,20,21,22]. Zi2Zi [4], one of the automatic font generation methods, is based on paired images font generation, which considers each character as a whole and learns to transform between fonts. There are many studies like Combined cCVAE and cGAN [20] that carry out automatic font generation tasks.

2.4. Style Embedding Generation

The task of generating a specific style of fonts is studied by different means. Style migration based on the Convolutional Neural Network (CNN) [23] is employed to create a font with artistic style [24]. The task of handwritten font generation [20,21] mainly focuses on the characteristic of the font, which attracts the feature of resource input to generate a font with a specific style. Besides, some style learning tasks on font generation based on GAN use a style embedding strategy to generate new fonts [3,17,22,25]. However, since these generated new fonts are various and it is hard to guide the process of image synthesis, our proposed algorithm is useful and meaningful.

3. Questionnaire for the Relationship between Facial Expressions and Fonts

Common facial expressions are divided into eight categories [26] including ‘anger,’ ‘contempt,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise.’ As illustrated in Section 1, the Chinese fonts can accurately reflect emotions. In order to obtain a better comprehending of people’s emotional tendency towards fonts, we began to quantitatively and qualitatively research the relationship between facial expressions and fonts. The questionnaires are set based on the relationship between facial expressions and fonts.

3.1. Questionnaire Design

We selected 110 Chinese fonts in the font collection, and these different fonts composed of Simplified Chinese are designed by artists. Thereafter, we chose 30 fonts that could best express a certain degree of emotional meaning. The fonts described by artists included classic Chinese fonts, such as Songti style, Kaiti style, and Lishu style. After that, we created a survey based on five calligraphy experts with more than ten years of experience and 25 common respondents, asking them to select 10 fonts that could reflect emotions from all the given options. We collected the survey results of 30 respondents (consist of five calligraphy experts and 25 common respondents). For considering the importance of representative samples, we adopted the following weight assignment: for the choices of the calligraphy experts and the common respondents, the weights are set to 1.5 and 1, respectively, which are empirically determined. Finally, we obtained the 10 most selected fonts as shown in Figure 1 for the next questionnaire.
The eight facial expressions used in this questionnaire are posted on the Tencent platform according to the order of ‘anger,’ ‘contempt,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise.’ A total of 452 male and female respondents who usually have the opportunity to see different fonts in their daily lives were tasked to pick up one of the ten fonts. The content of the questions and the corresponding example are shown in Table 1 and Figure 2, respectively.

3.2. Questionnaire Results

Figure 3 shows the results of the questionnaire for emotions, and we can see that the results below show it is distinguishable for this questionnaire. It is found that font 3 is the most selected with 43.6% having chosen it, followed by font 7, which has been chosen by 19.9% of the participants for the emotion of anger. For the emotion of contempt, font 7 is the most selected with 45.4% having chosen it, followed by font 4, which has been chosen by 23.3% of the participants. For the emotion of happiness, font 6 is the most selected with 56.2% having chosen it, followed by font 10 with up to 13.5%. For the emotion of neutral, font 5 is the most selected with 76.1% having chosen it, followed by font 6, which has been chosen by 12.8% of the participants. For the emotion of surprise, font 8 is the most selected with 30.1% having chosen it, followed by font 1, which has been chosen by 13.5% of respondents. These results are relatively well distinguished. However, the choice of font for disgust, sadness, and fear is relatively unclear because the results’ difference in the questionnaire is small. Therefore, we choose font 9 because it is the highest option for all three emotions.
According to the results of the questionnaire, we get some correspondence as shown in Table 2. For example, we find “Anger” best corresponds to font 3, as well as “Contempt” to font 7, etc.

4. Architecture

In this task, we propose an Emotional Guidance GAN (EG-GAN) algorithm based on Zi2Zi merged with Earth Mover (EM) distance and the Gradient Penalty, as well as classification loss, which learns to map from random noise vector z, and we observe source font image x combined with two conditional singles that contain style information s and classification information f to y, G { x , s , f , z } y . The generator G is trained to produce domain images that cannot differentiate between the generated image and the real image by the corresponding trained discriminator, D, which is trained to recognize the generator’s output as well as possible. A bunch of paired images that are determined by the results of the questionnaire are fed to the font generation module. The training procedure of the font generation module is shown in Figure 4. Finally, with a guided information c ( c 1 , c 2 ) that contains c 1 and c 2 controlled by inputting the facial expression image, the generated Chinese fonts are taken to another questionnaire and other metrics to evaluate the creditable degree. The whole process is diagrammed in Figure 5.

4.1. Network Architectures

4.1.1. Facial Information Extraction Module

As is shown in Figure 6, a state-of-the-art model illustrated in [27,28] is adapted to recognize emotions, which takes facial image as an input and calculates the first two probabilities and makes a process of data adjustment to guide new font generation with a specific style. The probabilities of seven kinds of expressions including ‘anger,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise’ are calculated by the expression recognition module. Considering that the Fer2013 [29] expressions dataset is manually annotated and have certain errors, to make sure the generated fonts of Chinese character exactly reflects emotions, we select the first two probabilities ( c 1 , c 2 ) of the expression recognition module’s results to make sure that the two data c 1 , c 2 are adjusted into the standard data c 1 = c 1 / c 1 + c 2 and c 2 = c 2 / c 1 + c 2 usage of regularization as combined style labels for the process of font generation.

4.1.2. Font Generation Module

As illustrated in Section 1 of our paper, the Zi2Zi model faces some problems as a Chinese character generation model. On the one hand, the model is difficult to train, easy to collapse, and often unstable. On the other hand, generating images leads to relative low resolution. Considering those problems mentioned above, we adopt some strategies to solve these issues. We use EM Distance [6]:
W ( P r , P g ) = i n f ω ~ ( P r , P g ) E ( p 1 , p 2 ) ~ ω [ p 1 p 2 ] ,
which is continuous and differentiable, and which could solve the phenomenon of gradient disappearance. Where Π ( P r , P g ) is the set of all possible joint distributions combined by P r and P g , the P r is the distribution of real paired images (consisting of source font images and real target domain images), and the P g is the distribution of fake paired images (consisting of source font images and generated images). The marginal distributions of ω ( p 1 , p 2 ) are, respectively, P r and P g . Equation (1) can be interpreted as the minimum cost of transforming the distribution P r into the distribution P g . Meanwhile, as the EM Distance continues to shrink the closer the two distributions are, this can provide indicators to guide the process of model training and visible supervision over the process of training and judging the metrics model of effectiveness, and this can further the enhancement towards the revolution of generating images. Furthermore, Gradient Penalty [7] is employed to enhance the quality of generation images.
L g r a d i e n t _ p e n a l t y = E x ^ ~ P r , g [ ( x ^ D ( x ^ ) 2 1 ) 2 ] ,
where x ^ is uniformly sampled from the straight line between pairs of points sampled from P r and P g [7]. We use L1 distance rather than L2, as L1 encourages less blurring:
L L 1 ( G ) = E x , y , s , z [ y G ( x , s , z ) 1 ] .
We define the generator loss:
L g e n _ l o s s = E ( x , z ) ~ P g [ D ( G ( x , z ) ) ] ,
to encourage G to fool the D by the way of generating high-quality images as much as possible. The discriminator loss is
L d i s c _ l o s s = E ( x , z ) ~ P g [ D ( G ( x , z ) ) ] E ( x , z ) ~ P r [ D ( x , y ) ] .
We use constant loss [30] to assume that the generated image and real image should reside in the same space and close to each other.
L c o n s t _ l o s s = E ( x , s , z ) ~ P g [ G e n c o d e r ( x , z ) ] E ( x , y ) ~ P r [ G e n c o d e r ( x , y ) ] .
We use category loss [31] to solve the one-to-many function [32] by concatenating a non-trainable Gaussian noise as style information, and embedding s to the character embedding to generate a target character. We incorporate the style information s in the constant loss function and the category loss function.
L c a t e g o r y _ l o s s = L r e a l _ c a t e g o r y _ l o s s + L f a k e _ c a t e g o r y _ l o s s ,
L r e a l _ c a t e g o r y _ l o s s = E ( x , y ) ~ P r ( t log D ( x , s , y ) + ( 1 t ) log ( 1 D ( x , s , y ) ) ) ,
L f a k e _ c a t e g o r y _ l o s s = E ( x , s , z ) ~ P g ( t log D ( G ( x , s , z ) ) + ( 1 t ) log ( 1 D ( G ( x , s , z ) ) ) ) ,
where t denotes target domain labels. Although those loss functions can be used in our model to generate high-quality images, some fonts generation will still be blurred. Therefore, we incorporate a classification strategy [22] to make the discriminator easier to distinguish the style of the generated characters.
L d _ c l s = E ( x , y , t ) ~ P r ( log D ( t | y ) ) .
The D ( t | y ) is a probability distribution on target domain labels computed by the Discriminator.
L g _ c l s = E ( x , t , z , f ) ~ P g ( log D ( t | G ( x , t , z , f ) ) ) ,
where y , t , and f denote ground-truth images, ground-truth labels, and classification information, respectively.
Combining all the loss functions, the final objection function is
m i n G m a x D ( L G + L D + λ L g r a d i e n t _ p e n a l t y ) ,
and
L G = L g e n _ l o s s + ϕ L c o n s t _ l o s s + α L L 1 + β L f a k e _ c a t e g o r y _ l o s s + L g _ c l s
L D = L d i s c _ l o s s + γ L c a t e g o r y _ l o s s + L d _ c l s ,
where λ , α , β , γ , and   ϕ are hyper-parameters. The architecture of font generation module is shown in Figure 7.

5. Experimental Results and Analyses

5.1. Comparison Experiments and Results

In Section 4, we illustrate our approach and the detailed model of generating fonts to reflect emotions. We divide the task into three parts. The first one is the expression recognition module, which is up to the state-of-the-art commonly used in expressions recognition on FER2013 [29]. In this part, the probabilities of seven kinds of expressions including ‘anger,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise’ are calculated by the expression recognition module. Then the first two probabilities are adjusted into the standard data elaborated in Section 4.1.1. Intuitively, we can see the results of recognition and the generated font with corresponding emotions shown in Figure 8. The upper part of Figure 8 shows the results of expression recognition and the corresponding information after data processing. The bottom part of Figure 8 shows the characters of the source font and the generated characters mixed with two styles, respectively.
The font generation stage is performed into two parts: (1) the training process, and (2) the testing process. In the process of training, we adopt a strategy to split training into a two-step process. First, 27 fonts with 1000 characters are randomly sampled and are fed to the model described in Section 4.1.2; then, we freeze the parameters of the encoder. After that, we choose to fine-tune [4] to the six fonts with 3000 characters that are acquired from the questionnaire. By this way, the encoder learns the characters’ structure information with 27,000 characters in the first step, and the dedicated decoder can better focus on the characteristics of the target domain. Those loss functions elaborated in Section 4 are designed to make the proposed model generate high-quality images. We also confirm that the two-timescale update rule (TTUR) [33] is effective, and we advocate using it specifically to address slow learning in regularized discriminators. We set the α   =   100 ,   β   =   0.5 ,   ϕ   =   15 ,   and   γ   =   1 during the first step, and α   =   300 ,   β   =   0.5 ,   ϕ   =   150 ,   and   γ   =   0.9 during the second step. We employ the style information s and classification information f in these two steps of the training process. In the process of the test, the guided signal c is embedded into the generator G to synthesize the specific style of font. The guided signal c is adjusted from first two probabilities that are recognized in the process of expression recognition. The source font is Heiti style.
In our experiments, the EG-GAN model and EG-GAN 1 model are employed, respectively. “EG-GAN 1” represents the EG-GAN model without adopting the strategy of Gradient Penalty. Intuitively, compared with the Zi2Zi model, ours can successfully generate multiple style fonts with higher quality. The comparison results are shown in Figure 9.
Moreover, some evaluation metrics, structural similarity index (SSIM) [34,35], and peak signal-to-noise ratio (PSNR) [35], to show the quality of synthesized images, are employed in our experiment. SSIM is used to evaluate the aspects of luminance, contrast, and structure between two images. The higher the SSIM score, the clearer the description of image distortion. PSNR is a ratio between the real images and the reconstructed images to measure the quality of the images. We calculate the SSIM and the PSNR for grayscale images (8 bits). Input a reference image u and a test image v , both of size M × N , the PANR is defined by:
P S N R ( u , v ) = 10 l g ( 255 2 / M S E ( u , v ) ) ,
where
M S E ( u , v ) = 1 M N i = 1 M j = 1 N ( u i j v i j ) 2 .
The MSE represents the cumulative squared error. The SSIM is defined as:
S S I M ( u , v ) = l u m ( u , v ) c o n ( u , v ) s t r ( u , v ) ,
where
{ l u m ( u , v ) = 2 μ u μ v + 1 μ u 2 + μ v 2 + 1 c o n ( u , v ) = 2 σ u σ v + 2 σ u 2 + σ v 2 + 2 s t r ( u , v ) = σ u v + 3 σ u σ v + 3 .
The l u m ( u , v ) is the luminance function which calculates the proximity between μ u and μ v . Here the μ u and the μ v are the mean luminance of input two images. The c o n ( u , v ) is the contrast function which calculates the contrast of the input two images. Here the σ u and the σ v are the standard deviation of the input two images. The s t r ( u , v ) is the structure function which calculates the correlation of the input two images. Here the σ u v is the covariance between two images. The positive constants 1 , 2 and 3 are set to make sure that a non-null denominator.
A combination of the obtained six fonts image and every single font image is employed in our experiment. Table 3 shows the comparison results of the SSIM and PSNR metrics between Zi2Zi and our model in a combination of the obtained six fonts image. Table 4 shows the comparison results of the SSIM metric between Zi2Zi and our model in every single font image, and Table 5 shows the comparison results of the PSNR metric between Zi2Zi and our model in every single font image.
In order to confirm that employing the EM Distance can improve the convergence rate of the model, the sampled comparison results from training process of Zi2Zi and EG-GAN 1 are shown in Figure 10. We can see that our generated images have a higher quality and convergence rate in the same steps.
Meanwhile, we compare the details of EG-GAN 1,2 and EG-GAN 2 with EG-GAN 1 and EG-GAN, respectively. “EG-GAN 1,2” represents the EG-GAN model both without employing Gradient Penalty strategy and classification loss. “EG-GAN 2” represents the EG-GAN model without employing classification loss. We notice that the classification strategy is useful, and the comparison results are shown in Figure 11.
An evaluation experiment is conducted to verify the effectiveness of the proposed system. Specifically, the purpose is to verify whether the generated fonts can reflect the corresponding emotions. The experiment is in the form of a questionnaire posted on the Tencent platform that collects responses from 200 respondents for the generated six fonts shown in Figure 12, respectively. Respondents are asked to choose one font from options corresponding emotion. The results of this questionnaire present in Figure 13. According to the results of the evaluative questionnaire shown in Table 2, we get some correspondence, which are as follows: (1) Font 6 reflects on “Happiness”; (2) Font 3 reflects on “Anger”; (3) Font 8 reflects on “Surprise”; (4) Font 7 reflects on “Contempt”; (5) Font 5 reflects on “Neutral”; and (6) Font 9 reflects on “Disgust,” “Fear,” and “Sadness.” All these questions are answered more than 50%, as shown in Figure 13, which confirms that our algorithm is credible.

5.2. Discussion

The performance of the Generator and Discriminator is the key issue for ordinary GAN during the training process. However, it is difficult to balance the performance without knowing the metrics of the model training. In addition, in order to evaluate the difference between the real image and the generated image, the KL divergence and the JS divergence [6,7] for the Discriminator are used, but there is a problem of gradient vanishing during the initial training. The experiment results shown in Figure 9 and Figure 10 intuitively confirm the effectiveness of incorporating the strategies of EM Distance and Gradient Penalty. Meanwhile, SSIM and PSNR metrics are employed in our experiments to imply lower numerical differences compared with the baseline. Generally, the Zi2Zi model can achieve the function of one-to-many. However, since the performance of the classification function proposed previously only focuses on the individual information, some generated fonts are somewhat blurred. Therefore, we incorporate conditional information to make sure that each single font has a consistent style. In order to verify the effectiveness of the classification strategy, the comparison results are shown in Figure 11. Meanwhile, these generated fonts shown in Figure 12 are used as questions for the questionnaire, and results with high reliability are shown in Figure 13. These comparison experiments mentioned above verify the effectiveness of our proposed font generation system.

6. Conclusions

In this paper, we have proposed an automatic Chinese font generation algorithm reflecting emotions based on Generative Adversarial Network (GAN). A questionnaire is designed to study the relationship between fonts and emotions. According to the results of the questionnaire, the fonts and emotions are associated with each other. The Emotional Guidance GAN (EG-GAN) model combines the guided signal recognized by the expression recognition module with the font generation module to synthesize new fonts that reflect corresponding emotions. We incorporate EM Distance, Gradient Penalty, and TTUR, as well as classification loss tricks, to improve the resolution of the generated images and to reasonably demonstrate the generated fonts with quantitative styles by using the weight of the adjusted labels from the results of the expression recognition module. Some comparison experiments are employed to confirm the effect of our models. In addition to these, an evaluation experiment is carried out further proving the credibility of this task. Moreover, we can also get some inspirations to generate fonts with various styles contained in different scenes.

Author Contributions

Methodology, L.C. and F.L.; software, L.C.; data curation, H.C.; validation, W.Y. and J.C.; writing—original draft preparation, L.C.; writing—review and editing, F.L. and Q.C.; supervision, F.L. and Q.C.; funding acquisition, F.L. and Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The Programme for Professor of Special Appointment (Eastern Scholar) at Shanghai Institutions of Higher Learning, and also partially supported by JSPS KAKENHI Grant Number 15K00159.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, S.; Aizawa, K. Emotype: Expressing emotions by changing typeface in mobile messenger texting. Multimed. Tools Appl. 2019, 78, 14155–14172. [Google Scholar] [CrossRef] [Green Version]
  2. Amare, N.; Manning, A. Seeing typeface personality: Emotional responses to form as tone. In Proceedings of the 2012 IEEE International Professional Communication Conference, Orlando, FL, USA, 8–10 October 2012; pp. 1–9. [Google Scholar]
  3. Hayashi, H.; Abe, K.; Uchida, S. GlyphGAN: Style-consistent font generation based on generative adversarial networks. Knowl. Based Syst. 2019, 186, 104927. [Google Scholar] [CrossRef] [Green Version]
  4. Tian, Y.C. zi2zi: Master Chinese Calligraphy with Conditional Adversarial Networks. Available online: https://kaonashi-tyc.github.io/2017/04/06/zi2zi.html (accessed on 30 July 2020).
  5. Isola, P.; Zhu, J.Y.; Zhou, T.H.; Efros, A.A. Image-to-image translation with adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  6. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  7. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein GANs. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 5769–5779. [Google Scholar]
  8. Wada, A.; Hagiwara, M. Japanese font automatic creating system reflecting user’s kansei. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, Washington, DC, USA, 8 October 2003; pp. 3804–3809. [Google Scholar]
  9. Dai, R.W.; Liu, C.L.; Xiao, B.H. Chinese character recognition: History, status and prospects. Front. Comput. Sci. China 2007, 1, 126–136. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, C.L.; Jaeger, S.; Nakagawa, M. Online recognition of chinese characters: The state-of-the-art. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 198–213. [Google Scholar] [PubMed]
  11. Lan, Y.J.; Sung, Y.T.; Wu, C.Y.; Wang, R.L.; Chang, K.E. A cognitive-interactive approach to chinese characters learning: System design and development. In Learning by Playing. Game-Based Education System Design and Development. Edutainment 2009; Chang, M., Kuo, R.K., Chen, G.D., Hirose, M., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2009; pp. 559–564. [Google Scholar]
  12. Chuang, H.C.; Ma, M.Y.; Feng, Y.C. The features of chinese typeface and its emotion. In Proceedings of the International Conference on Kansei Engineering and Emotion Research, Paris, French, 2–4 March 2010. [Google Scholar]
  13. Shen, H.Y. The Image and the meaning of the chinese character for ‘enlightenment’. J. Anal. Psychol. 2019, 64, 32–42. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  15. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27(NIPS), Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  16. Mao, X.D.; Li, Q.; Xie, H.R.; Lau, R.Y.K.; Wang, Z.; Smolley, S.P. Least squares generative adversarial networks. arXiv 2016, arXiv:1611.04076. [Google Scholar]
  17. Yuan, D.J.; Feng, H.X.; Liu, T.L. Research on new font generation system based on generative adversarial network. In Proceedings of the 4th International Conference on Mechanical, Control and Computer Engineering (ICMCCE), Hohhot, China, 24–26 October 2019; pp. 18–21. [Google Scholar]
  18. Zong, A.; Zhu, Y. Strokebank: Automating personalized chinese handwriting generation. In Proceedings of the 28th AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 29–31 July 2014; pp. 3024–3029. [Google Scholar]
  19. Pan, W.Q.; Lian, Z.H.; Sun, R.J.; Tang, Y.M.; Xiao, J.G. Flexifont: A flexible system to generate personal font libraries. In Proceedings of the 2014 ACM Symposium on Document Engineering, Fort Collins, CO, USA, 16–19 September 2014; pp. 17–20. [Google Scholar]
  20. Kong, W.R.; Xu, B.C. Handwritten chinese character generation via conditional neural generative models. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–7 December 2017. [Google Scholar]
  21. Chang, B.; Zhang, Q.; Pan, S.Y.; Meng, L.L. Generating handwritten chinese characters using CycleGAN. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 199–207. [Google Scholar]
  22. Chen, J.F.; Ji, Y.L.; Chen, H.; Xu, X. Learning one-to-many stylised chinese character transformation and generation by generative adversarial networks. IET Image Process. 2019, 13, 2680–2686. [Google Scholar] [CrossRef]
  23. Balouchian, P.; Foroosh, H. Context-sensitive single-modality image emotion analysis: A unified architecture from dataset construction to CNN classification. In Proceedings of the 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 1932–1936. [Google Scholar]
  24. GitHub. Available online: https://github.com/yuweiming70/Style_Migration_For_Artistic_Font_With_CNN (accessed on 29 July 2020).
  25. Azadi, S.; Fisher, M.; Kim, V.; Wang, Z.W.; Shechtman, E.; Darrell, T. Multi-content gan for few-shot font style transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7564–7573. [Google Scholar]
  26. Langner, O.; Dotsch, R.; Bijlstra, G.; Wigboldus, D.H.J.; Hawk, S.T.; Knippenberg, A.V. Presentation and validation of the radboud faces database. Cogn. Emot. 2010, 24, 1377–1388. [Google Scholar] [CrossRef]
  27. GitHub. Available online: https://github.com/WuJie1010/Facial-Expression-Recognition.Pytorch (accessed on 29 July 2020).
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  29. Goodfellow, I.J.; Erhan, D.; Carrier, P.L.; Courville, A.; Mirza, M.; Hamner, B.; Cukierski, W.; Tang, Y.; Thaler, D.; Lee, D.-H.; et al. Challenges in representation learning: A report on three machine learning contests. Neural Netw. 2015, 64, 59–63. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Taigman, Y.; Polyak, A.; Wolf, L. Unsupervised cross-domain image generation. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  31. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier GANs. In Proceedings of the 34 International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  32. Johnson, M.; Schuster, M.; Le, Q.V.; Krikun, M.; Wu, Y.; Chen, Z.; Thorat, N.; Viégas, F.; Wattenberg, M.; Corrado, G.; et al. Google’s multilingual neural machine translation system: Enabling zero-shot translation. TACL 2017, 5, 339–351. [Google Scholar] [CrossRef] [Green Version]
  33. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6629–6640. [Google Scholar]
  34. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Horé, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
Figure 1. Options for questionnaire determined by five calligraphy experts and 25 common respondents.
Figure 1. Options for questionnaire determined by five calligraphy experts and 25 common respondents.
Applsci 10 05976 g001
Figure 2. A typical example of the questionnaire, which asking respondents to choose a font that corresponds to the emotion shown in the facial image (The emotion “anger” is shown here).
Figure 2. A typical example of the questionnaire, which asking respondents to choose a font that corresponds to the emotion shown in the facial image (The emotion “anger” is shown here).
Applsci 10 05976 g002
Figure 3. The main results of the questionnaire. For each emotion, the first two probabilities are marked, respectively.
Figure 3. The main results of the questionnaire. For each emotion, the first two probabilities are marked, respectively.
Applsci 10 05976 g003
Figure 4. The training procedure of the font generation module. (a) The generator, G, tries to generate high-quality images to fool the discriminator; (b) the discriminator, D, is trained to recognize the generator’s outputs as well as possibilities. The font generation module is trained by paired images (consists of source font images x and target domain images y that are ground-truth).
Figure 4. The training procedure of the font generation module. (a) The generator, G, tries to generate high-quality images to fool the discriminator; (b) the discriminator, D, is trained to recognize the generator’s outputs as well as possibilities. The font generation module is trained by paired images (consists of source font images x and target domain images y that are ground-truth).
Applsci 10 05976 g004
Figure 5. The diagram of the proposed font generation system. 1 * and 2 * imply the learning processes of the discriminator model with {source font image x, generated image} and {source font image x, real image} tuples, respectively. First, the system trains the font generation module until the discriminator cannot distinguish between real fonts and generated fonts. Then, the pre-trained model is used to recognize the facial expression of the input image, so as to guide the generator to generate a new font with specific emotions.
Figure 5. The diagram of the proposed font generation system. 1 * and 2 * imply the learning processes of the discriminator model with {source font image x, generated image} and {source font image x, real image} tuples, respectively. First, the system trains the font generation module until the discriminator cannot distinguish between real fonts and generated fonts. Then, the pre-trained model is used to recognize the facial expression of the input image, so as to guide the generator to generate a new font with specific emotions.
Applsci 10 05976 g005
Figure 6. Facial image is transformed into 48×48 gray images to feed the pre-trained recognition module. Then the probabilities of seven kinds of expressions including ‘anger,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise’ are calculated. The first two probabilities ( c 1 , c 2 ) are selected and finally adjusted into standard data c ( c 1 , c 2 ) as combined style labels to guide the process of font generation.
Figure 6. Facial image is transformed into 48×48 gray images to feed the pre-trained recognition module. Then the probabilities of seven kinds of expressions including ‘anger,’ ‘disgust,’ ‘fear,’ ‘happiness,’ ‘neutral,’ ‘sadness,’ and ‘surprise’ are calculated. The first two probabilities ( c 1 , c 2 ) are selected and finally adjusted into standard data c ( c 1 , c 2 ) as combined style labels to guide the process of font generation.
Applsci 10 05976 g006
Figure 7. The architecture of font generation. The features of the characters in the source font strive to have low-level information via the encoder. Then the abstracted information combines with style information s , classification information f , and guided information c ( c 1 , c 2 ) , the output of which can be regarded as the characteristics extracted in the combined target domains that depend on the guided information c ( c 1 , c 2 ) . The skip connections between encoder and decoder aim to share the low-level information. Finally, the generated images are decoded.
Figure 7. The architecture of font generation. The features of the characters in the source font strive to have low-level information via the encoder. Then the abstracted information combines with style information s , classification information f , and guided information c ( c 1 , c 2 ) , the output of which can be regarded as the characteristics extracted in the combined target domains that depend on the guided information c ( c 1 , c 2 ) . The skip connections between encoder and decoder aim to share the low-level information. Finally, the generated images are decoded.
Applsci 10 05976 g007
Figure 8. The results of expression recognition and the generated font with corresponding emotions. When inputting a facial image with specific emotions, a new font is generated by combining the first two probabilities of the emotions.
Figure 8. The results of expression recognition and the generated font with corresponding emotions. When inputting a facial image with specific emotions, a new font is generated by combining the first two probabilities of the emotions.
Applsci 10 05976 g008
Figure 9. The comparison results of fonts generated by various models. The images of the first row are the ground truth, and the second raw are generated by the Zi2Zi model. Images of the third and fourth rows are generated by EG-GAN 1 and the EG-GAN model, respectively.
Figure 9. The comparison results of fonts generated by various models. The images of the first row are the ground truth, and the second raw are generated by the Zi2Zi model. Images of the third and fourth rows are generated by EG-GAN 1 and the EG-GAN model, respectively.
Applsci 10 05976 g009
Figure 10. The comparison of EG-GAN 1 with Zi2Zi in the training process.
Figure 10. The comparison of EG-GAN 1 with Zi2Zi in the training process.
Applsci 10 05976 g010
Figure 11. The comparison results of fonts generated by our models. The images of the first and second rows are the characters of Font 9, and the third and fourth rows are of Font 6. The models used to generate the images of the first to fourth rows are EG-GAN 1,2, EG-GAN 1, EG-GAN 2 and EG-GAN, respectively.
Figure 11. The comparison results of fonts generated by our models. The images of the first and second rows are the characters of Font 9, and the third and fourth rows are of Font 6. The models used to generate the images of the first to fourth rows are EG-GAN 1,2, EG-GAN 1, EG-GAN 2 and EG-GAN, respectively.
Applsci 10 05976 g011
Figure 12. Fonts generated by our proposed model for reliability analysis.
Figure 12. Fonts generated by our proposed model for reliability analysis.
Applsci 10 05976 g012
Figure 13. Evaluation results of fonts generated by our proposed model.
Figure 13. Evaluation results of fonts generated by our proposed model.
Applsci 10 05976 g013
Table 1. Questions 1–4 aim to acquire basic information of respondents, which is useful to improve the credibility of questionnaire. Questions 5–12 are main content of questionnaire.
Table 1. Questions 1–4 aim to acquire basic information of respondents, which is useful to improve the credibility of questionnaire. Questions 5–12 are main content of questionnaire.
Question SequenceQuestion Content
Question 1What’s your name?
Question 2What’s your gender?
Question 3How about your experience on calligraphy?
Question 4Talking about fonts reflecting emotions
Question 5Choose a font best corresponds to “anger”.
Question 6Choose a font best corresponds to “contempt”.
Question 7Choose a font best corresponds to “disgust”.
Question 8Choose a font best corresponds to “fear”.
Question 9Choose a font best corresponds to “happiness”.
Question 10Choose a font best corresponds to “neutral”.
Question 11Choose a font best corresponds to “sadness”.
Question 12Choose a font best corresponds to “surprise”.
Table 2. Quantitative correspondence between fonts and emotions.
Table 2. Quantitative correspondence between fonts and emotions.
EmotionFont
AngerFont 3
ContemptFont 7
HappinessFont 6
NeutralFont 5
Disgust, Fear, SadnessFont 9
SurpriseFont 8
Table 3. Comparison of objective image quality metrics (SSIM, PSNR) of the proposed model.
Table 3. Comparison of objective image quality metrics (SSIM, PSNR) of the proposed model.
ModelSSIMPSNR
Zi2Zi0.885414.7226
EG-GAN 10.886514.7501
EG-GAN0.895315.3035
1 Denotes EG-GAN without Gradient Penalty.
Table 4. Comparison of SSIM metrics for each single font.
Table 4. Comparison of SSIM metrics for each single font.
ModelFont 3Font 5Font 6Font 7Font 8Font 9
Zi2Zi0.86700.90060.90490.88200.87940.8809
EG-GAN 10.86140.90360.91580.87320.88310.8835
EG-GAN0.86900.90820.90920.89970.89400.8915
1 Denotes EG-GAN without Gradient Penalty.
Table 5. Comparison of PSNR metrics for each single font.
Table 5. Comparison of PSNR metrics for each single font.
ModelFont 3Font 5Font 6Font 7Font 8Font 9
Zi2Zi14.236716.902216.335813.660813.715714.2021
EG-GAN 114.182817.146117.139113.178914.023114.4951
EG-GAN14.693217.362716.406414.804314.867115.0289
1 Denotes EG-GAN without Gradient Penalty.

Share and Cite

MDPI and ACS Style

Chen, L.; Lee, F.; Chen, H.; Yao, W.; Cai, J.; Chen, Q. Automatic Chinese Font Generation System Reflecting Emotions Based on Generative Adversarial Network. Appl. Sci. 2020, 10, 5976. https://doi.org/10.3390/app10175976

AMA Style

Chen L, Lee F, Chen H, Yao W, Cai J, Chen Q. Automatic Chinese Font Generation System Reflecting Emotions Based on Generative Adversarial Network. Applied Sciences. 2020; 10(17):5976. https://doi.org/10.3390/app10175976

Chicago/Turabian Style

Chen, Lu, Feifei Lee, Hanqing Chen, Wei Yao, Jiawei Cai, and Qiu Chen. 2020. "Automatic Chinese Font Generation System Reflecting Emotions Based on Generative Adversarial Network" Applied Sciences 10, no. 17: 5976. https://doi.org/10.3390/app10175976

APA Style

Chen, L., Lee, F., Chen, H., Yao, W., Cai, J., & Chen, Q. (2020). Automatic Chinese Font Generation System Reflecting Emotions Based on Generative Adversarial Network. Applied Sciences, 10(17), 5976. https://doi.org/10.3390/app10175976

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop