Next Article in Journal
Monitoring and Analysis of Deformation Evolution Law of Fault Activation Caused by Deep Mining in Shizishan Copper Mine, China
Previous Article in Journal
Physical Hazards in Aepyceros melampus Carcasses Killed for Meat Purposes by Aerial and Thoracic Shots
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An End-to-End Classifier Based on CNN for In-Air Handwritten-Chinese-Character Recognition

School of Computer Science and Technology, Anhui University of Technology, Maanshan 243032, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(14), 6862; https://doi.org/10.3390/app12146862
Submission received: 21 June 2022 / Revised: 3 July 2022 / Accepted: 5 July 2022 / Published: 7 July 2022
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
A convolutional neural network (CNN) has been successfully applied to in-air handwritten-Chinese-character recognition (IAHCCR). However, the existing models based on CNN for IAHCCR need to convert the coordinate sequence of a character into images. This conversion process increases training and classifying time, and leads to the loss of information. In order to solve this problem, we propose an end-to-end classifier based on CNN for IAHCCR in this paper, which, to knowledge, is novel for online handwritten-Chinese-character recognition (OLHCCR). Specifically, our model based on CNN directly takes the original coordinate sequence of an in-air handwritten-Chinese-character as input, and the output of the full connection layer is pooled by global average pooling to form a fixed-size feature vector, which is sent to softmax for classification. Our model can not only directly process coordinate sequences such as the models based on recurrent neural network (RNN), but can also obtain the global structure information of characters. We conducted experiments on two datasets, IAHCC-UCAS2016 and SCUT-COUCH2009. The experimental results show a comparison with existing CNN models based on image processing or RNN-based methods; our method does not require data augmentation techniques nor an ensemble of multiple trained models, and only uses a more compact structure to obtain higher recognition accuracy.

1. Introduction

In the field of image processing and pattern recognition, the recognition of online handwritten characters is important work and many impressive achievements have been achieved [1,2,3,4,5]. The data processed by online handwriting recognition is a series of coordinate sequences, which are usually used in human–computer interaction devices, such as handwriting tablets, smart phones and pads. The input range of the traditional human–computer interaction method is limited by the size of the touch device, and damage to the local area will make the entire device unusable. In this context, a new way of human–computer interaction, vision-based and in-handwriting, has attracted more and more researchers’ interest [6,7,8], e.g., we can use in-air handwriting to switch TV channels remotely or adjust the temperature of an air conditioner. Compared with the traditional online handwriting a using touch pad or wearable device, visual-based in-air handwriting experiences fewer space constraints and allows the writers to freely write in the air; the generated character has no pen-lift information and is finished in one stroke. The jitter of the stroke of a character is quite serious and the strokes overlap each other. These characteristics of in-air handwriting characters bring more difficulties for IAHCCR. Some examples of in-air handwriting Chinese characters and traditional handwritten Chinese characters can be seen in Figure 1.
As in-air handwriting is a new development of traditional online handwriting, the methods used for OLHCCR are available for IAHCCR. Traditional OLHCCR methods do not directly recognize the original coordinate sequence of an online handwritten Chinese character, but extract features from a preprocessed coordinate sequence according to specific classification rules and specific domain knowledge [9,10]. Before deep learning was introduced into OLHCCR, statistical features, classification algorithms based on statistical features and preprocessing algorithms had always been the research hotspots of OLHCCR, and show excellent performance in OLHCCR [11,12]. Related methods have been introduced into IAHCCR, e.g., linear or nonlinear normalization [9,11], eight-directional features [10], modified quadratic discriminant functions (MQDF) [13], etc.
In recent years, deep learning has achieved great success in the fields of computer vision and pattern recognition [14,15,16], and also been successfully applied to OLHCCR [17,18,19,20]. Compared with the above-mentioned traditional classification models, the models based on deep learning show an overwhelming advantage in OLHCCR. However, to our knowledge, all the existing models based on CNN for OLHCCR do not directly recognize the coordinate sequence of an online handwritten Chinese character, but convert the coordinate sequence into images or vectors [2,3,21,22,23]. As shown in Figure 2, this conversion process not only causes the loss of training and recognition time, but also only utilizes the spatial information of characters and loses the temporal information of the coordinate sequence, so it is difficult to obtain a higher recognition rate.
Since the CNN based model uses a large-scale convolution kernel, it often requires a large number of training patterns and leads to more memory consumption. Unlike the CNN models, the models based on RNN can directly process the coordinate sequences of online handwritten Chinese characters and outperform most CNN structures [4,24]. Although RNN-based models are suitable for processing sequence data, they are less suitable for processing long sequence data than CNN and ignore the global structures of online handwritten Chinese characters. In-air handwritten Chinese characters usually contain hundreds of points, so RNN-based models will consume a lot of computation time.
Based on the above analysis, we propose an end-to-end classifier based on CNN for IAHCCR in this paper which has the advantages of both CNN and RNN. First, we directly use the preprocessed coordinate sequences of online handwritten Chinese characters as the input of the network. Then, the coordinate sequences are converted into one-dimensional feature maps through the first layer of convolution, and then the range of the receptive field is expanded by stacking the number of convolution layers to obtain the contextual connection. Finally, global average pooling is applied to the output of the convolutional layer to obtain a fixed-size feature vector, which is sent to the fully connected layer to extract features and use softmax for classification. The end-to-end CNN can directly recognize the coordinate sequences of online handwritten Chinese characters. Compared with the existing CNN models, the end-to-end CNN does not need to convert the original data into images, nor does it need to design features combined with specific domain knowledge. Due to selecting the convolution kernel with a smaller scale, the end-to-end CNN needs fewer parameters and occupies less memory. Compared with the RNN-based models, the end-to-end CNN can learn the global information of online handwritten Chinese characters and adapt to different lengths of coordinate sequences.
The rest of the paper is organized as follows. Section 2 briefly introduces the related works. Section 3 introduces the proposed method at length. The experimental results are reported in Section 4. We conclude this paper in Section 5.

2. Related Works

Research on IAHCCR has been underway for several years, and a complete recognition method usually consists of three stages: preprocessing, feature extraction and classification. The work related to different stages will be described in detail below.
In the preprocessing stage, as mentioned in Section 1, IAHCCR can use the method of OLHCCR. Character normalization can reduce within-class variation and improve recognition accuracy. Traditional methods require the normalization of characters to a uniform size. Linear normalization, which causes character-shape changes, has been superseded by other methods [11]. For example, nonlinear normalization [9], pseudo-2D normalization and line-density projection interpolation [11]. End-to-end methods do not require normalization to a fixed-size vector but, rather, normalize the distribution of coordinate sequences [24].
Feature extraction and classification are two separate stages in the traditional IAHCCR method. In the stage of manual feature extraction, compared with the character image drawn using the original coordinates, the method of decomposing local strokes into different directions to form multiple feature maps achieves higher recognition accuracy, such as eight-directional feature maps [10]. Especially for IAHCCR, the whole Chinese character is more like a curve function defined on a two-dimensional plane that can be expanded by Taylor. Qu et al. [25] proposed higher-order directional features. Representing Chinese characters as directional features has been the standard approach for a long time. Classifiers are also very important for IAHCCR. To further improve the recognition rate and recognition speed, the learning vector quantization technique-based [26] multi-level classification technique is reported for IAHCCR in [27]. Qu et al. introduced locality-sensitive sparse representation-based classifiers (LSRC) [28] into IAHCCR, and achieved a higher recognition rate than MQDF [25]. In order to further improve the recognition accuracy of LSRC, a loss function is designed to minimize the reconstruction error of each training pattern and make the reconstruction of each training pattern as close as possible to the optimized prototype of its class is suggested, which significantly improves recognition accuracy [29].
As deep-learning techniques have made great achievements in other fields, the deep neural network was introduced into IAHCCR. In the deep-learning-based IAHCCR algorithm, feature extraction and classification are integrated. For IAHCCR, Qu et al. [23] proposed a nine-layer convolutional-neural-network model combined with data-augmentation technology which significantly improved the recognition rate. Like other image recognition tasks [30], the performance of this method requires a large amount of data to ensure the performance, so data-augmentation technology is needed to expand the data, although it also requires more memory consumption and a higher number of parameters. Ren et al. [20] proposed an end-to-end recognizer based on recurrent neural networks. This method directly recognizes character sequences without converting characters into feature vectors. In order to further improve the recognition accuracy, Ref. [4] proposed an RNN system with two new computing architectures added. Table 1 summarizes related works. In this paper, we combine the advantages of CNN and RNN to propose an end-to-end CNN model for directly recognizing sequences.

3. Proposed Method

Like other end-to-end recognition methods, the end-to-end CNN method consists of two parts, preprocessing and model architecture.

3.1. Preprocessing

As the writing styles of writers vary widely, the structure, position, shape, sampling-point density and stroke order of the finished in-air Chinese characters are different. These varied intra-class structures and the confusion between similar characters always result in a reduction in recognition accuracy [24]. In this paper, the primary purpose of the preprocessing is to eliminate redundant points and standardize the distribution of coordinate points, so as to improve the recognition accuracy for IAHCCR. The preprocessing steps in this paper are summarized as follows: (1) Remove redundant points in the coordinate sequence of in-air handwritten Chinese characters. (2) Normalize the coordinates to a unified coordinate system.

3.1.1. Remove Redundant Points

Any given in-air handwritten character P can be represented by its coordinate sequence as
P = [ ( x 1 , y 1 ) , , ( x t , y t ) , , ( x T , y T ) ]
where x t and y t are the XY coordinates of the tth point of P; and t = 1 , , T , T is the number of the coordinate points of P. For the coordinate sequence of the character, except for the first point, if the Euclidean distance between the tth point ( x t , y t ) and its adjacent point ( x t 1 , y t 1 ) is less than the given threshold L, i.e.,
( x t x t 1 ) 2 + ( y t y t 1 ) 2 < L
then the point ( x t , y t ) is deleted as a redundant point. In the following experiments, L varies corresponding to different handwritten character P, and can be computed by 0.015 × m a x { h , w } , where h and w are the space height and width of P, respectively.

3.1.2. Normalize Coordinates

Since in-air handwriting has fewer space constraints and an unstable writing position, the positions of the written Chinese characters vary: some are higher, some are lower, some are left and some are right. The end-to-end CNN directly takes a coordinate sequence as input, so the variation in coordinate-points distribution will greatly reduce the recognition rate. In order to decrease the variations in the spatial sizes and positions of characters, we employed coordinates normalization, following the method presented in [24]. Let I t = ( x t x t 1 ) 2 + ( y t y t 1 ) 2 be the distance between two consecutive points ( p t = ( x t , y t ) and p t 1 = ( x t 1 , y t 1 ) ). In order to normalize the coordinates to a standard interval, it is first necessary to estimate the mean of the coordinates projected on the coordinate axis. We can calculate the mean μ x and μ y of XY coordinates, respectively, by
μ x = 1 2 t = 2 T I t ( x t + x t 1 ) t = 2 T I t
μ y = 1 2 t = 2 T I t ( y t + y t 1 ) t = 2 T I t
The standard deviation δ x on the x axis can be estimated as
δ x = t = 2 T I t [ ( x t μ x ) 2 + ( x t μ x ) ( x t 1 μ x ) + ( x t 1 μ x ) 2 ] 3 × t = 2 T I t
The normalized trajectory needs to keep the original character shape and stroke writing direction, the characters are only scaled by the standard deviation δ x . For the tth point of P, we can obtain the normalized point ( x ¯ t , y ¯ t ) by
x ¯ t = x t μ x δ x , y ¯ t = y t μ y δ x
Some examples of processing through the above steps are shown in Figure 3. From Figure 3, we can see that the processed coordinates are evenly distributed on both sides of ( 0 , 0 ) and the number of points has been reduced from 723 to 355.

3.2. Designing End-to-End CNN Architecture

For the proposed architecture, a brief overview is first given. As shown in Figure 4, the preprocessed sequence of coordinates is used as input to the recognizer. A larger receptive field can be obtained by stacking convolutional layers (Conv1 and Conv2) with small convolution kernels and reducing the sequence length by downsampling, so that more discriminative features can be learned. The fully connected layer (FC) requires a fixed-size feature vector input, but the length of the coordinate sequence is variable. Therefore, it is necessary to average over the coordinate sequence to obtain a fixed-size feature vector. For the output of the convolutional layer, the sequence mean m i of each feature map Z i = [ z i 1 , , z i t , , z i T ] is estimated by
m i = 1 T t = 1 T z i t
Since the number of output feature maps is fixed, we can combine these means into a fixed-size feature vector f = [ m 1 , , m c , , m C ] , where C is the number of feature maps. Finally the softmax function is used to estimate the probability distribution of all classes. Next, we will introduce the specific configuration of end-to-end CNN in detail.
The configuration of the end-to-end classifier can be seen in Figure 5a. In Figure 5a, “Conv k: 2 × 3 , s:1, 64” denotes that the kernel size a convolutional layer is 2 × 3 , the stride is 1 and the number of feature maps is 64, respectively. PReLU stands for the parametric rectified linear unit [31]. “Max-pool k: 1 × 2 s:2” denotes that the size of max pooling is 1 × 2 , and the stride is 2. “Dropout 0.2” denotes that the dropout rate is 0.2 [32]. “FC 160” denotes a fully connected layer, and the number of channels is 160. “Block N” represents a residual block, which is illustrated in Figure 5b, N = 64 , 128 , 256 .
In more detail, the preprocessed in-air handwritten character P = [ p 1 , , p t , , p T ] R 2 × T are directly used as the input of the classifier, T varies according to each different character. In order to identify the coordinate sequence, the network needs to stack multiple convolutional layers to expand the receptive field, so that the model can extract as much spatial and temporal information of the sequence as possible. P is first processed using a convolutional layer in the time dimension to obtain a series of feature maps of 1 × T 1 , T 1 changes with T, the kernel size is 2 × 3 and the stride is 1. After that, dropout is employed to avoid overfitting, and the max pooling is used to reduce the sequence feature length, which can further increase the receptive field range and reduce the impact of zero padding on sequence recognition. Since the constructed network is very deep, the residual link [33] can efficiently train the network. As shown in Figure 5b, we construct a residual block which contains a total of three convolutional layers. In Figure 5b, “⊕” denotes the elementwise sum, three convolutional layers are denoted by “Conv1”, “Conv2”, and “Conv3”, respectively. Conv1 and Conv2 are used to directly extract sequence features with kernel size 1 × 3 and stride 1. Conv3 is designed to make the feature size of the input and output the same during the connection operation, the size of the convolution kernel is 1 × 1 , and the stride is 1. The residual block is repeated until the number of feature maps is increased from 64 to 256. Then, global average pooling (GAP) [34] is employed to obtain a fixed-size feature vector f. Finally, f is input into the fully connected layer (FC) and softmax is used to classify it.

4. Experiments

4.1. Datasets

We conducted experiments on the IAHCC-UCAS2016 [23] which is an in-air handwritten Chinese character dataset and the GB1 dataset in the SCUT-COUCH2009 [35] database, which is a traditional handwritten Chinese character dataset. The samples in IAHCC-UCAS2016 consist of projections on a 2D plane of a sequence of 3D coordinates recorded by a sensor worn on the fingertip. The samples in the GB1 dataset are the trajectory coordinates written directly on the tablet. Both datasets are publicly available. The GB1 dataset involves 3755 character classes of the first level set of GB2312-80 and each class has 188 patterns. The IAHCC-UCAS2016 covers 3811 Chinese character classes, and each class contains 115 samples. For each class, 80 % were randomly selected as training sets and the remaining 20 % as test sets.

4.2. Model Training Strategy

Our network structure was implemented based on PyTorch, initialized with the default parameters of the framework. The optimizer used Adam; the initial learning rate was set to 0.001. The learning rate was decreased when the accuracy on the training set no longer increased or increased slowly, with a decay rate of 0.1. All experiments were conducted on RTX-2080ti.

4.3. Comparison Experiments

In order to determine the appropriate mini-batch size under the condition that the learning rate is set to 0.001, the mini-batch size was set to 64, 128, 256 and 512, respectively. Similar to other studies in this field, and since the amount of data for each class in the dataset was equal, the accuracy criterion was used to evaluate the proposed method. The accuracy was calculated by
a c c u r a c y = T P + T N T P + T N + F N + F P
In Equation (8), T P is the number of samples that are correctly assigned to the goal class, T N is the number of samples that are correctly not assigned to the goal class, FN is the number of samples that are wrongly assigned to the goal class, and FP is the number of samples that are wrongly assigned to the other classes. We performed five-fold cross-validation on the dataset and reported the average recognition rate for each mini-batch size in Figure 6. The model converged at 20 epochs and performed best when the mini-batch size is set to 128.
For convolution operation on P, we want the size of the sequence before and after convolution operation to be same, so we need to zero-pad the sequence. We designed three padding methods, padding1, which is evenly filled at both ends of the sequence; padding2, which is filled at the head of the sequence; and padding3, padded at the end of the sequence. The change in the recognition rate during the training process is shown in Figure 7. The best recognition rates of the three methods are shown in Table 2. From Figure 7 and Table 2, we can see that padding1 is the best padding method. Similar to using CNN to recognize images, when CNN is used to recognize coordinate sequences, zero padding on the edge can effectively retain the information of the edge position. Compared with method 2 and method 3, method 1 has a higher recognition accuracy by 0.07 % and 0.28 % . Filling methods 2 and 3 both lose the feature of the edge position to some extent. For the fill method 3, in particular, the header position information of the character sequence is lost, which is quite important to recognize the character.
To verify the effectiveness of the end-to-end CNN, we compared the proposed method with traditional CNN architectures on both datasets. We used the nine-layer CNN presented in [23] as a benchmark (including Cov8d using eight-direction feature maps, CovCd1 using a combination of higher-order direction feature maps and curved feature maps, and CovCd2 using a combination of eight-direction feature maps, high-order eight-direction feature maps and curved feature maps). In addition, we also compared our method with end-to-end RNN (RNN1) [20] and RNN combined with new computing architectures (RNN2) [4] on the IAHCC-UCAS2016 dataset. In Table 3, Table 4, Table 5 and Table 6, the column “DA” indicates whether the data-augmentation technique was adopted during training, and the column “Ensemble” indicates whether the recognition decision was made by the ensemble of multiple trained models. As shown in Table 3 and Table 4, the traditional CNN that recognizes directional features [23] does not directly recognize coordinate sequences, but recognizes the extracted directional feature images. This indirect learning method will affect the recognition accuracy, and it is difficult to obtain a high recognition rate when the amount of data is insufficient, so it is necessary to use data-augmentation techniques to expand the data set during the training process. Our method does not use data augmentation techniques and achieves a recognition accuracy of 96.07 % on the IAHCC-UCAS2016 dataset and 98.02 % on the GB1 dataset. Although RNN can directly identify the coordinate sequence and extract the time-series features between the coordinates, it is difficult to consider the global spatial features. RNN [4,20] often use multiple models to jointly participate in recognition decisions to improve recognition accuracy. However, this strategy will exponentially increase the storage cost and the improvement in recognition accuracy is not ideal. As shown in Table 3, the recognition accuracy of RNN2-Ensemble is 0.8 % higher than that of RNN2, but the storage cost is about 3.67 times that of RNN2. Compared with RNN, end-to-end CNN can extract more discriminate spatiotemporal features more comprehensively, requires only a single model and the storage cost is only 6.48 MB.
We also compared our proposed method with traditional methods on both datasets. These methods include nearest prototype classifier (NPC) [36], nearest prototype classifier trained by MCE (NPC-MCE) [28], multistage classifiers (Multi1) [29], discriminative multistage classifiers (Multi2) [26], modified quadratic discriminant functions (MQDF) [25], locality-sensitive sparse representation-based classifiers (LSRC) [27] and locality-sensitive sparse representation toward optimized prototype classifier (LSROPC) [29]. Table 5 and Table 6 summarize the recognition performance of the various methods on both datasets. From Table 5 and Table 6, compared with traditional machine-learning methods, deep-learning technology has huge advantages, and the above methods all identify the features extracted by artificial means, which inevitably loses the timing information of the trajectory sequence to a certain extent. Therefore, using CNN to directly identify coordinate sequences can achieve an overwhelming performance improvement.

5. Conclusions

This paper proposes an end-to-end classifier based on CNN for IAHCCR. Our method achieves 96.08 % recognition accuracy on the IAHCC-UCAS2016 dataset with a storage cost of 6.48MB. Compared with the directional feature-extraction strategy that indirectly identifies temporal features, the direct identification of trajectory sequences can directly extract more discriminate temporal features to obtain better results, and no complex feature-extraction process is required. Unlike RNN, the macroscopic structure of trajectory sequences can also be considered. The experimental results show that the proposed method is very suitable for OLHCCR. In future work, we plan to explore more robust and efficient CNN architectures for IAHCCR.

Author Contributions

Conceptualization, X.Q.; methodology, X.Q. and M.H.; software, X.Q. and M.H.; validation, M.H.; resources, X.Q.; data curation, X.Q.; writing—original draft preparation, M.H.; writing—review and editing, X.Q., J.H. and X.W.; Funding acquisition, X.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China Youth Fund under Grant No. 61906003, and the University Synergy Innovation Program of Anhui Province under Grant No. GXXT-2021-004.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the datasets used in this research are publicly accessible. The IAHCC-UCAS2016 dataset is available at: http://cvmt.ucas.ac.cn/dataset/, accessed on 21 June 2022. The GB1 dataset in the SCUT-COUCH2009 database is available at: http://www.hcii-lab.net/data/scutcouch/, accessed on 21 June 2022.

Acknowledgments

We sincerely thank the editors and the reviewers for their valuable comments in improving this paper. We would also like to thank L.W. Jin et. al. for helping to provide the experimental data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, C.L.; Jaeger, S.; Nakagawa, M. Online recognition of chinese characters: The state-of-the-art. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 198–213. [Google Scholar]
  2. Yin, F.; Wang, Q.F.; Zhang, X.Y.; Liu, C.L. ICDAR 2013 Chinese handwriting recognition competition. In Proceedings of the 2013 International Conference on Document Analysis and Recognition, Washington, DC, USA, 25–28 August 2013; pp. 1464–1470. [Google Scholar]
  3. Gan, J.; Wang, W.Q.; Lu, K. A new perspective: Recognizing online handwritten Chinese characters via 1-dimensional CNN. Inform. Sci. 2019, 378, 375–390. [Google Scholar] [CrossRef]
  4. Ren, H.Q.; Wang, W.Q.; Liu, C.L. Recognizing online handwritten Chinese characters using RNNs with new computing architectures. Pattern Recognit. 2019, 93, 179–192. [Google Scholar] [CrossRef]
  5. Li, Y.; Qian, Y.; Chen, Q.C.; Hu, B.T.; Wang, X.L.; Ding, Y.X.; Ma, L. Fast and Robust Online Handwritten Chinese Character Recognition with Deep Spatial & Contextual Information Fusion Network. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  6. Xu, S.B.; XUE, Y.; CHEN, Y.Q. Quantitative Analyses on Effects from Constraints in Air-Writing. IEICE Trans. Inform. Syst. 2019, E120, 867–870. [Google Scholar] [CrossRef]
  7. Fu, Z.J.; Xu, J.S.; Zhu, Z.D.; Liu, A.X. Writing in the air with WiFi signals for virtual reality devices. IEEE Trans. Mobile Comput. 2019, 18, 473–484. [Google Scholar] [CrossRef]
  8. Gadekallu, T.R.; Srivastava, G.; Liyanage, M.; Iyapparaja, M.; Chowdhary, C.L.; Koppu, S.; Maddikunta, P.K.R. Hand gesture recognition based on a Harris Hawks optimized Convolution Neural Network. Comput. Electr. Eng. 2022, 100, 107836. [Google Scholar] [CrossRef]
  9. Bai, Z.L.; Huo, Q. A study of nonlinear shape normalization for online hand-written Chinese character recognition: Dot density vs. line density equalization. In Proceedings of the 2006 International Conference on Pattern Recognition, Hong Kong, China, 20–24 August 2006; pp. 921–924. [Google Scholar]
  10. Bai, Z.L.; Huo, Q. A study on the use of 8-directional features for online handwritten Chinese character recognition. In Proceedings of the 2005 International Conference on Document Analysis and Recognition, Seoul, Korea, 31 August–1 September 2005; pp. 262–266. [Google Scholar]
  11. Liu, C.L.; Marukawa, K. Pseudo two-dimensional shape normalization methods for handwritten Chinese character recognition Pattern Recognition. Inform. Sci. 2005, 38, 2242–2255. [Google Scholar]
  12. Liu, C.L.; Yin, F.; Wang, D.H.; Wang, Q.F. Online and offline handwritten Chinese character recognition: Benchmarking on new databases. Pattern Recognit. 2013, 46, 155–162. [Google Scholar] [CrossRef]
  13. Kimura, F.; Takashina, K.; Tsuruoka, S.; Miyake, Y. Modified quadratic discriminant functions and the application to Chinese character recognition. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 9, 149–153. [Google Scholar] [CrossRef]
  14. Lai, S.X.; Jin, L.W.; Yang, W.X. Toward high-performance online HCCR: A CNN approach with DropDistortion, path signature and spatial stochastic max-pooling. Pattern Recognit. Lett. 2017, 89, 60–66. [Google Scholar] [CrossRef] [Green Version]
  15. Hsieh, T.A.; Wang, H.M.; Lu, X.G.; Tsao, Y. WaveCRN: An Efficient Convolutional Recurrent Neural Network for End-to-end Speech Enhancement. IEEE Signal Proc. Let. 2020, 27, 2149–2153. [Google Scholar] [CrossRef]
  16. Oza, P.; Patel, V.-M. One-Class Convolutional Neural Network. IEEE Signal Proc. Let. 2019, 26, 277–281. [Google Scholar] [CrossRef] [Green Version]
  17. Yang, W.X.; Jin, L.W.; Tao, D.C.; Xie, Z.C.; Feng, Z.Y. Drop sample: A new training method to enhance deep convolutional neural networks for large-scale unconstrained handwritten chinese character recognition. Pattern Recognit. 2016, 58, 190–203. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, H.M.; Zhang, X.Y.; Yin, F.; Yang, Q.; Liu, C.L. Convolutional Prototype Network for Open Set Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2358–2370. [Google Scholar] [CrossRef]
  19. Zhang, X.Y.; Bengio, Y.S.; Liu, C.L. Online and Offline Handwritten Chinese Character Recognition: A Comprehensive Study and New Benchmark. Pattern Recognit. 2017, 61, 348–360. [Google Scholar] [CrossRef] [Green Version]
  20. Ren, H.Q.; Wang, W.Q.; Lu, K.; Zhou, J.S.; Yuan, Q.C. An end-to-end recognizer for in-air handwritten Chinese characters based on a new recurrent neural networks. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo, Hong Kong, China, 11–14 July 2017; pp. 841–846. [Google Scholar]
  21. Liu, X.; Hu, B.; Chen, Q.; Wu, X.; You, J. Stroke Sequence-Dependent Deep Convolutional Neural Network for Online Handwritten Chinese Character Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 4637–4648. [Google Scholar] [CrossRef] [Green Version]
  22. Parvizi, A.; Kazemifard, M.; Imani, Z. Fast Online Character Recognition Using a Novel Local-Global Feature Extraction Method. In Proceedings of the 2021 International Conference on Information and Knowledge Technology (IKT), Babol, Iran, 14–16 December 2021; pp. 183–187. [Google Scholar]
  23. Qu, X.W.; Wang, W.Q.; Lu, K.; Zhou, J.-S. Data augmentation and directional feature maps extraction for in-air handwritten Chinese character recognition based on convolutional neural network. Pattern Recognit. Lett. 2018, 111, 9–15. [Google Scholar] [CrossRef]
  24. Zhang, X.Y.; Yin, F.; Zhang, Y.M.; Liu, C.L.; Bengio, Y. Drawing and recognizing Chinese characters with recurrent neural network. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 849–862. [Google Scholar] [CrossRef] [Green Version]
  25. Qu, X.W.; Wang, W.Q.; Lu, K.; Zhou, J.S. High-order directional features and sparse representation based classification for in-air handwritten chinese character recognition. In Proceedings of the IEEE International Conference on Multimedia and Expo, Seattle, WA, USA, 11–15 July 2016; pp. 1–6. [Google Scholar]
  26. Liu, C.L.; Nakagawa, M. Evaluation of prototype learning algorithms for nearest-neighbor classifier in application to handwritten character recognition. Pattern Recognit. 2001, 34, 601–615. [Google Scholar] [CrossRef]
  27. Xu, N.; Wang, W.Q.; Qu, X.W. A discriminative classifier for in-air handwritten Chinese characters recognition. In Proceedings of the 2015 International Conference on Internet Multimedia Computing and Service, Quebec City, QC, Canada, 27–30 September 2015; pp. 16–19. [Google Scholar]
  28. Wei, C.P.; Chao, Y.W.; Yeh, Y.R.; Wang, Y.C.F. Locality-sensitive dictionary learning for sparse representation based classification. Pattern Recognit. 2013, 46, 1277–1287. [Google Scholar] [CrossRef]
  29. Qu, X.W.; Wang, W.Q.; Lu, K.; Zhou, J.S. In-air handwritten Chinese character recognition with locality-sensitive sparse representation toward optimized prototype classifier. Pattern Recognit. 2018, 78, 4783–4797. [Google Scholar] [CrossRef]
  30. XU, N.; WANG, W.Q.; QU, X.W. On-line sample generation for in-air written chinese character recognition based on leap motion controller. In Proceedings of the 2015 Pacific-Rim Conference on Multimedia, Gwangju, Korea, 16–18 September 2015; pp. 171–180. [Google Scholar]
  31. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  32. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  33. He, K.M.; Zhang, X.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26–30 June 2016; pp. 770–778. [Google Scholar]
  34. Lin, M.; Chen, Q.; Yan, S.C. Network in network. In Proceedings of the 2014 IEEE International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  35. Jin, L.-W.; Gao, Y.; Liu, G.; Li, Y.-Y. SCUT-COUCH2009—A comprehensive online unconstrained Chinese handwriting database and benchmark evaluation. Int. J. Doc. Anal. Recog. 2011, 14, 53–64. [Google Scholar] [CrossRef] [Green Version]
  36. Jin, X.B.; Liu, C.L.; Hou, X. Regularized margin-based conditional loglikelihood loss for prototype learning. Pattern Recognit. 2010, 43, 2428–2438. [Google Scholar] [CrossRef]
Figure 1. (a) In-air handwritten Chinese characters. (b) Traditional handwritten Chinese characters.
Figure 1. (a) In-air handwritten Chinese characters. (b) Traditional handwritten Chinese characters.
Applsci 12 06862 g001
Figure 2. Examples of in-air handwritten Chinese characters.
Figure 2. Examples of in-air handwritten Chinese characters.
Applsci 12 06862 g002
Figure 3. Example of normalization. The character (left) before the normalization and that (right) after the normalization are given, respectively.
Figure 3. Example of normalization. The character (left) before the normalization and that (right) after the normalization are given, respectively.
Applsci 12 06862 g003
Figure 4. Overview of end-to-end convolutional neural network architecture. The stride of the convolutional layer shown in the figure is 1, and the convolution kernel size is 3.
Figure 4. Overview of end-to-end convolutional neural network architecture. The stride of the convolutional layer shown in the figure is 1, and the convolution kernel size is 3.
Applsci 12 06862 g004
Figure 5. (a) The configuration of end-to-end convolutional neural network. (b) The architecture of residual convolution block.
Figure 5. (a) The configuration of end-to-end convolutional neural network. (b) The architecture of residual convolution block.
Applsci 12 06862 g005
Figure 6. The average recognition accuracy of different batchsizes on the IAHCC-UCAS2016 varied with the training epoch.
Figure 6. The average recognition accuracy of different batchsizes on the IAHCC-UCAS2016 varied with the training epoch.
Applsci 12 06862 g006
Figure 7. The recognition accuracy of different zero-pad methods on the IAHCC-UCAS2016 varies with the training epoch.
Figure 7. The recognition accuracy of different zero-pad methods on the IAHCC-UCAS2016 varies with the training epoch.
Applsci 12 06862 g007
Table 1. Summary of related works.
Table 1. Summary of related works.
MethodsTypesLimitations
Linear normalization [11]PreprocessingNeglecting the density of handwritten character sequence coordinates can lead to severe shape distortion
Nonlinear normalization [9]PreprocessingUnable to correct skew, local width or height imbalances in handwritten characters
Pseudo-2D normalization and line density projection interpolation [11]PreprocessingUsually dealing with image features
Coordinate normalization [24]PreprocessingUsually dealing with sequence features
Eight-directional feature maps [10]Feature extractionUsually applied to OLHCCR
Higher-order directional features. [25]Feature extractionUsually applied to IAHCCR
Learning-vector-quantization-technique-based multi-level classifier [27]ClassifierHigh computational cost and storage consumption
MQDF [25]ClassifierLow recognition accuracy and high storage cost
LSRC [28]ClassifierDifficulty obtaining discriminative features and low recognition accuracy
Locality-sensitive sparse representation toward optimized prototype classifier [29]ClassifierLow recognition accuracy and high storage cost
Nine-layer convolutional neural network model combined with data-augmentation technology [23]ClassifierRelying on manual feature extraction and large amounts of data
End-to-end recognizer based on recurrent neural networks [20]ClassifierLess efficient computational structure
RNN system with two new computing architectures added [4]ClassifierIt has difficultly effectively learning the global spatial features of sequences
Table 2. Accuracy of different padding methods on the IAHCC-UCAS2016 dataset.
Table 2. Accuracy of different padding methods on the IAHCC-UCAS2016 dataset.
MethodsAccuracy (%)
padding196.07
padding296.0
padding395.79
Table 3. Performance comparison of different deep-learning methods on the IAHCC-UCAS2016 dataset.
Table 3. Performance comparison of different deep-learning methods on the IAHCC-UCAS2016 dataset.
MethodsAccuracy (%)Storage (MB)DAEnsemble
Cov8d [23]91.6220.2NoNo
CovCd1 [23]92.3220.2YesNo
CovCd2 [23]92.9320.2YesNo
RNN1 [20]92.507.03NoNo
RNN1-Ensemble [20]93.4061.59NoYes
RNN2 [4]93.607.01NoNo
RNN2-Ensemble [4]94.4025.75NoYes
Proposed method96.076.48NoNo
Table 4. Performance comparison of different deep-learning methods on the GB1 dataset.
Table 4. Performance comparison of different deep-learning methods on the GB1 dataset.
MethodsAccuracy (%)Storage (MB)DAEnsemble
Cov8d [23]96.1019.4NoNo
CovCd1 [23]97.1519.4yesNo
CovCd2 [23]97.4319.4yesNo
Proposed method98.026.44NoNo
Table 5. Performance comparison of various methods on the IAHCC-UCAS2016 dataset.
Table 5. Performance comparison of various methods on the IAHCC-UCAS2016 dataset.
MethodsAccuracy (%)Storage (MB)DAEnsemble
NPC [36]86.908.52NoNo
NPC-MCE [28]88.938.52NoNo
Multi1 [29]88.28128.8NoNo
Multi2 [26]88.90128.8NoNo
MQDF [25]89.96191.11NoNo
LSRC [27]88.9331.78NoNo
LSROPC [29]91.0170.00NoNo
Proposed method96.076.48NoNo
Table 6. Performance comparison of various methods on the GB1 dataset.
Table 6. Performance comparison of various methods on the GB1 dataset.
MethodsAccuracy (%)Storage (MB)DAEnsemble
NPC [36]91.304.76NoNo
NPC-MCE [28]92.964.76NoNo
Multi1 [29]92.36124.5NoNo
Multi2 [26]93.42124.5NoNo
MQDF [25]95.21184.84NoNo
LSRC [27]94.4027.86NoNo
LSROPC [29]95.5065.53NoNo
Proposed method98.026.44NoNo
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hu, M.; Qu, X.; Huang, J.; Wu, X. An End-to-End Classifier Based on CNN for In-Air Handwritten-Chinese-Character Recognition. Appl. Sci. 2022, 12, 6862. https://doi.org/10.3390/app12146862

AMA Style

Hu M, Qu X, Huang J, Wu X. An End-to-End Classifier Based on CNN for In-Air Handwritten-Chinese-Character Recognition. Applied Sciences. 2022; 12(14):6862. https://doi.org/10.3390/app12146862

Chicago/Turabian Style

Hu, Mianjun, Xiwen Qu, Jun Huang, and Xuangou Wu. 2022. "An End-to-End Classifier Based on CNN for In-Air Handwritten-Chinese-Character Recognition" Applied Sciences 12, no. 14: 6862. https://doi.org/10.3390/app12146862

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop