Next Article in Journal
SNOWED: Automatically Constructed Dataset of Satellite Imagery for Water Edge Measurements
Previous Article in Journal
FragTracer: Real-Time Fragmentation Monitoring Tool for F2FS File System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Scene Text Recognition Based on Transformer

1
College of Information Science and Engineering, Xinjiang University, No. 777 Huarui Street, Urumqi 830017, China
2
Xinjiang Laboratory of Multi-Language Information Technology, Xinjiang University, No. 777 Huarui Street, Urumqi 830017, China
3
Xinjiang Multilingual Information Technology Research Center, Xinjiang University, No. 777 Huarui Street, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(9), 4490; https://doi.org/10.3390/s23094490
Submission received: 11 March 2023 / Revised: 12 April 2023 / Accepted: 3 May 2023 / Published: 5 May 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Scene text recognition (STR) has been a hot research field in computer vision, aiming to recognize text in natural scenes using computers. Currently, attention-based encoder–decoder frameworks struggle to precisely align feature regions with the target object when dealing with complex and low-quality images, a phenomenon known as attention drift. Additionally, with the rise of Transformer, the increasing size of parameters results in higher computational costs. In order to solve the above problems, based on the latest research results of Vision Transformer (ViT), we utilize an additional position-enhancement branch to alleviate attention drift and dynamically fused position information with visual information to achieve better recognition accuracy. The experimental results demonstrate that our model achieves a 3% higher average recognition accuracy on the test set compared to the baseline. Meanwhile, our model maintains the advantage of a small number of parameters and fast inference speed, achieving a good balance between accuracy, speed, and computational load.

1. Introduction

Optical character recognition (OCR) [1,2] refers to the printed characters from the paper document recognition proposed by German scientist Tausheck in 1929. With the development of OCR technology, people are satisfied with the recognition of the text in the document or the book and focus on text recognition in the actual scene, which is called scene text recognition (STR).
STR is the expansion and extension of traditional OCR technology in natural scene images and has broad application prospects. STR, such as automatic driving and license plate recognition, has been widely used daily [3,4,5,6]. Natural scene text extraction technology can be divided into two sub-tasks: scene text detection [7,8] and scene text recognition [9,10]. The task of scene text detection is to locate relevant characters in the natural scene. Scene text recognition is converting the text in the cropped picture into the corresponding characters. Although STR began to be active at the end of the 1990s, with the rise of deep neural networks in recent years, its accuracy has reached the commercial level.
Inspired by the machine translation of natural language processing [11], the coder–decoder framework of attention mechanism is widely used in text recognition in natural scenes and has significantly improved the performance [12]. This technology learns the mapping relationship between the input image and output sequence in a data-driven way. The model’s encoder usually uses CNN/Transformer to convert the image into a sequence of feature vectors. Each feature vector represents an area of the image called the attention area. The decoding part’s attention network (AN) [13] first calculates the alignment factor by referring to the history of the target character and the encoded feature vector to generate the composite vector to achieve the alignment between the attention area and the corresponding label. Then, the target character is generated according to the synthetic vector and the history of the target character. However, in dealing with some complex and low-quality images, such as image distortion and blurring, the relevant research [14] found that the existing attention-mechanism methods do not perform well in this situation. One of the main reasons is that the existing methods can not obtain an accurate alignment between the feature area and the target for such an image, which is called “attention drift”.
In order to improve the above problems, some studies [15,16] have used character position and semantic information to improve the model’s accuracy. However, with the development of deep learning, the number of layers of Resnet and Transformer is becoming deeper and deeper, more and more information is added, and the cost of the model has doubled. Most STR models ignore the speed and operation cost, which leads to a competitive performance of the model but a high consumption cost. Therefore, the speed and operation cost of the model are equally important [17]. An ideal model should not focus solely on the model’s performance, regardless of the cost required by the model.
In this work, we balanced recognition accuracy, model speed, and cost. Based on Vit-tiny, we built a lightweight scene text recognition model (LSTR) and designed position-enhancement and visual-enhancement modules, allowing the model to maintain a fast inference speed while achieving higher accuracy at a lower cost. The contributions are as follows:
1. The LSTR model improves the problem of attention drift and uses position information to align target labels with visual features, which improves the model performance.
2. We have achieved a balance in the size and accuracy of the model with good practical application ability and model parameters of only 7 m that can thoroughly learn the image features, obtain a good recognition effect, and perform at, on average, 13.6 ms per text image.
3. The LSTR is compared with the latest model, and the results show that the performance of our model is better than the existing methods.

2. Related Work

Traditional natural scene text recognition methods [18,19,20] focus on ordinary text, mostly adopting the bottom-up method, detecting a single character using a sliding window, then integrating these characters into the output text [21]. These methods rely on accurate character classifiers and cannot deal with complex situations well.
The modern method [22,23] is top-down and regards scene text recognition as a sequence-prediction problem. This method conducts end-to-end prediction for text sequences without single-character detection. Most of the recent work has focused on irregular text recognition because the text in natural scenes has a complex background, such as fuzziness, low contrast, and a variety of font styles. Therefore, the text-recognition algorithm must consider the image’s complex background and text diversity. We divide the text-recognition models of modern scenes into three categories:
Segmentation-based: The segmentation-based approach [24] treats STR as a semantic segmentation problem. Ca-fcn [25] proposes a two-dimensional approach to scene text recognition that uses a semantic segmentation network and a word-forming module to recognize text and predict the position of each character. TextScanner [26] has designed two branches, one for character classification and the other for character position and order prediction; by assigning the character positioning to different channels, the characters are naturally separated and sorted. It also proposes a new mutual monitoring mechanism, enabling the two branches to monitor and enhance each other using sequence-level annotations.
CTC-based: CTC-based methods [27,28] use connectionist temporal classification (CTC) to complete sequence recognition, and CRNN [29] uses CNN and RNN as model encoders to extract the text features in images. The CTC decoder is then used to predict the character class for each time step; however, the computation of RNN depends on the length of the input sequence, which takes a long time and can cause gradient disappearance or explosion. Gao et al. [30] replaced RNN with CNN and used a stacked convolution layer to effectively capture the input sequence’s context correlation, which made the computation complexity of the model lower and the parallel computation easier.
Attention-based: SAR [31] is a proposed model for better recognition of irregular text based on a 2D attention module to locate and identify characters individually. DAN [32] solves the problem that traditional attention mechanisms cannot align by decoupling the alignment operation from the history-decoding results. Due to the excellent performance of Transformer in natural language processing, more and more researchers are applying Transformer to STR. ViT [33] shows that Transformer performs significantly better in image classification than deep networks such as Resnet when pre-trained on large datasets and migrated to small and medium image-recognition datasets. It can significantly reduce the computational resources required for training. Deit [34] proves that ViT can achieve a better performance without requiring large datasets.
We not only aim to suppress attention drift but also focus on the speed and calculation cost of the model. Therefore, we designed a portable, fast, and effective model, LSTR, and proposed a position-enhancement module and a visual-enhancement module based on ViT so that the decoding process can automatically adjust the position between the feature area and the target. The experimental results show that our model has excellent competitiveness.

3. LSTR Model

To achieve higher accuracy with fewer parameters, we propose a Vision Transformer-based model called LSTR, as shown in Figure 1. In the encoder part, to control the model size, we use Vision Transformer as the backbone network to extract visual information. We also design a position branch to learn the character position information, which helps optimize the spatial alignment between the feature map and the target character area, improving the recognition accuracy and robustness. In the decoder part, we introduce the position-enhancement module (PEM) and visual-enhancement module (VAE), using the position-enhancement module to reinforce the position information and combining the enhanced position information with visual information through the visual-enhancement module. This enhances the model’s perception of text targets, alleviates attention drift, and thus improves the model’s accuracy.

3.1. Encoder

The encoder of the model mainly consists of two parts: the backbone network and the position branch. Next, we will provide a detailed introduction.

3.1.1. Backbone

We used a 12-layer ViT and positional branch as the encoder of LSTR, a Transformer structure similar to ViTSTR and DieT, except that we studied the input image size and finally selected the input image size of 32 ∗ 128.
As shown in Figure 2, the input picture x R H × W × C is divided into a series of non-overlapping blocks x p R N × P 2 C , where H and W represent the height and width of the image, respectively, C the number of channels, P P the patch dimension, and N the length of the patch sequence.
An encoder consists of L encoder blocks. In our model, L = 12 , and Figure 2 shows one of the encoder blocks. The input of the encoder is:
z 0 = x C l a s s ; x P 1 E ; x P 2 E ; X P N E + E p o s
where x C l a s s R 1 × D is the [ c l a s s ] embedding, E R P 2 C × D stands for the linear projection matrix, and E p o s R ( N + 1 ) × D is the positional embedding. The input of each encoder block passes through a layer normalization (LN) followed by a multi-head self-attention layer (MSA), then a multi-layer perceptron (MLP), and finally connects with the residuals. Among them, MSA determines the relationship between eigenvectors, and MLP consists of two linear layers with a Gelu activation function. The equation for the MSA blocks is:
z l = M S A L N z l 1 + Z l 1
The term l represents the number of encoder blocks, and the MLP equation is:
Z l = M L P L N z l + z l
The final output is embedded in Z l R ( N + 1 ) × D for subsequent recognition and as input to the visual-enhancement module.

3.1.2. Position Branch

In the task of scene text recognition, it is crucial not only to extract the visual features effectively, but also to incorporate the text position information. Therefore, we propose a position branch that encodes the text position information using positional encoding.
The model’s position branch encodes the characters’ position in the text using a single heat vector. “1” stands for appearing in the corresponding dimension; otherwise, “0”. Finally, the position embedding is obtained by using the following equations:
P E ( p o s , 2 i ) = sin p o s / 10 , 000 2 i / d m o d e l
P E ( p o s , 2 i + 1 ) = cos p o s / 10 , 000 2 i / d m o d e l
where p o s stands for the position information, i stands for the character position, and d is the dimension. In the P E matrix, the sin variable is added in the even position, and the cos variable is added in the odd position.
This position branch also includes a linear transformation layer and a normalization layer. By transforming and normalizing the positional encoding, the model can effectively utilize the positional information and improve the accuracy of text recognition. In addition, the position-branch design adopts the idea of residual connection, which enables the model to better utilize the previous information and further improve the performance of the position branch.

3.2. Decoder

In the decoding process, the position information will become weaker with an increase in the time step, leading to attention drift and lower accuracy. Therefore, as shown in Figure 3, we entirely use the position information and design a position-enhancement module, which uses the enhanced position information as a query vector to input the visual-enhancement module. Finally, the linear layer outputs the prediction characters.

3.2.1. Position-Enhancement Module

The input of the position-enhancement module is the position-embedding output of the position module in the encoder, and the enhanced position information is obtained through multi-head self-attention (MSA), multi-layer perceptron (MLP), residual connection, and layer normalization (LN). For example, in Equations (6)–(8), the query (Q), key (K), and value (V) of MSA are all positional embedding. As shown in Figure 4, the query vector of MSA adopts the upper triangular mask to prevent information disclosure.
MultiHead ( Q , K , V ) = C o n c a t head 1 , , head n W 0
head i = Attention Q w i Q , K w i K , V w i V
Q = K = V = P E
where Q , K , and V are the query, key, and value, respectively, and W i Q , W i K , and W i V are the weight matrices for the i-th attention head. Concat is a concatenation function, and W 0 is a trainable parameter matrix.
The output of the multi-head self-attention mechanism is added to the input query to obtain the residual connection result, which is then passed through layer normalization.
x = Q + Dropout ( MultiHead ( Q , K , V ) )
P = LayerNorm ( γ σ ( x μ ) + β )
where μ and σ represent the mean and standard deviation, respectively, of the vector in the sample dimension, and γ and β represent the learnable scaling and shifting factors, respectively. These parameters are learned during the training process.
Next, a feedforward network is used for feature extraction:
FFN ( P ) = W 2 ReLU ( W 1 ( P ) + b 1 ) + b 2
where W 1 , W 2 , b 1 , and b 2 are the weight matrices and biases for the feedforward network (FFN).
Finally, the output of the feedforward network is added to the previous result through residual connection and layer normalization, resulting in the enhanced position feature P.
P = LayerNorm ( P + Dropout ( FFN ( P ) ) )

3.2.2. Visual-Enhancement Module

The input of the visual-enhancement module in the position-enhanced visual transformer (PSE) consists of the visual information extracted by the Vision Transformer (ViT) in the encoder and the position information enhanced by the position-enhancement module. This input is processed with multi-head attention (MHA), multi-layer perceptron (MLP), residual connections, and layer normalization (LN) to obtain the visually enhanced features with improved positional information. In MHA, the query Q is the position feature P that has been enhanced by the position-enhancement module, and K and V are the visual features extracted using ViT. The interaction between the positional and visual information can effectively alleviate attention drift. Finally, the enhanced features V are obtained according to the enhancement Formulas (6)–(12) in PSE.
Subsequently, the enhanced visual information V is fed into a linear layer to obtain the final prediction probability y. Finally, the cross-entropy loss function is used to calculate the loss.
L C E ( y , t ) = i = 1 C t i log y i
where L is the loss, y represents the output of the model, t represents the ground truth label, and C represents the number of classes of the labels. t i represents the true label value of class i, while y i denotes the predicted probability value of class i.

4. Experimental Analysis

4.1. Datasets

Our work involves three datasets: synthetic datasets and real datasets for model training and standard datasets for model validation. We used not only MJ [35] and ST [36], two synthetic datasets (s), but also nine real datasets (R) from the integration from Bautista et al. [37] to extend the data during the training.
1
Synthetic datasets: Because of the high cost of labeling datasets, STR did not obtain sufficient datasets early on, so the synthetic datasets ST and MJ were widely used to train models.
  • MJSynth (MJ): It is a dataset of synthetic text images generated by rendering text onto natural images using a variety of fonts, sizes, and colors. The dataset contains over 9 million images, each of which includes one or more instances of synthetic text. The text in the images is diverse, including different languages, scripts, and text styles. The dataset is primarily designed for training and evaluating OCR systems;
  • SynthText (ST): It is another synthetic text dataset created by cropping the text from natural images and pasting it onto new backgrounds. The dataset was originally designed for scene text detection but has since been adapted for OCR. SynthText contains over 7 million synthetic images, each of which contains a single instance of text. The text in the images is mainly in English but includes examples of other languages. Unlike MJSynth, the text in SynthText is embedded in realistic scenes, such as street signs, shop windows, and posters.
2
Real datasets: With the development of STR, more and more real datasets have been accumulated, and Bautista and others have integrated these datasets, namely ArT [38], Coco [39], LSVT [40], MLT19 [41], RCTW17 [42], OpenVINO [43], TextOCR [44], ReCTS [45] and Uber [46].
  • ArT: A publicly available scene text recognition dataset that contains scene text images extracted from different works of art. The dataset consists of 9300 text images from various works of art and can be used for tasks such as character recognition, text detection, and text recognition;
  • Coco: A widely used computer vision dataset that contains images and annotations for various tasks such as object detection, image segmentation, scene understanding, etc. The scene text recognition task in the COCO dataset refers to identifying text in the image and converting it to a computer-readable text format.
  • LSVT: A large-scale scene text dataset from China that includes 100,000 images divided into training, validation, and test sets. The scene text in the dataset includes regular text, digits, letters, and Chinese characters.
  • MLT19: A large-scale scene text dataset from multiple Asian countries that includes 10,000 images divided into training, validation, and test sets. The scene text in the dataset includes regular text, digits, letters, and Chinese characters.
  • RCTW17: A scene text recognition dataset from China that includes 8940 images from street and outdoor scenes. The scene text in the dataset includes Chinese, English, and digits.
  • OpenVINO: A scene text recognition dataset released by Intel that includes 23,000 images divided into training, validation, and test sets. The scene text in the dataset includes digits, letters, and Chinese characters.
  • TextOCR: A scene text recognition dataset from India that includes 20,000 images from various scenes. The scene text in the dataset includes Indian languages, English, and digits.
  • ReCTS: A scene text recognition dataset from the Hong Kong University of Science and Technology that includes 6000 images from street and outdoor scenes. The scene text in the dataset includes Chinese and English.
  • Uber: A large-scale scene text dataset from Uber that includes 1 million images from different countries. The scene text in the dataset includes digits, letters, and Chinese characters.
As shown in Figure 5, the synthetic images are mostly operated on text, and the background is relatively simple. Compared with the synthetic image, the background of the real image is more complex, with a stronger contrast between light and dark, a larger difference in the background environment, and the partial occlusion and perspective effect of the image. These complex real images can make the model have better generalization and better deal with the text recognition of complex scenes. Therefore, we further verify the performance of the model in real scenes by combining with real datasets.
We evaluated our model using six widely used datasets, categorized into regular and irregular datasets according to their textual characteristics.
  • Regular datasets:
    • IIIT5K [47]: This dataset contains 5000 text images extracted from Google image search. These images are high-resolution and undistorted, with little noise or background interference;
    • SVT [48]: This dataset contains 804 outdoor street images collected from Google Street View. These images contain text in various fonts, sizes, and orientations, but with little interference and noise. This dataset is typically used to test OCR algorithm performance in recognizing text with diverse fonts, sizes, and orientations;
    • CUTE80 [49]: It contains 80 curved text images and is primarily used to test the performance of OCR algorithms in curved text recognition tasks. Each image contains 3–5 words with varying fonts, sizes, colors, and some noise and interference. This dataset is commonly used to evaluate the performance of curved text recognition algorithms, particularly for the challenging curved text recognition tasks.
  • Irregular datasets:
    • ICDAR2015 (IC15) [50]: This dataset contains many blurred, rotated, low-resolution images, including 2077 and 1811 versions. We chose to use the 1811 version, which discards some extremely distorted images;
    • SVTP [51]: This dataset contains 645 perspective text images extracted from Google Street View. These images contain text in various fonts, sizes, and orientations, but due to perspective distortion, the text lines may appear curved or distorted. This dataset is typically used to test OCR algorithm performance in processing perspective text images;
    • ICDAR2013 (IC13) [47]: This dataset contains 288 curved text images where the text lines are arranged along curved paths. These images contain text in various fonts, sizes, and orientations, as well as some noise and interference. This dataset is typically used to test OCR algorithm performance in processing text images arranged along curved paths.

4.2. Experimental Setup

By comparing the effect of text image size on the model, the height and width of the input image are 32 and 128, respectively. We used Deit-tiny as our encoder’s visual feature extraction module and kept the experimental configuration. Our experiment was performed on an Nvidia Tesla V100 GPU system with two graphics cards, a batch size of 240, and an iteration number of 700,000 steps. With the Adadelta Optimizer, the initial learning rate is 1. The learning rate decay strategy is cosine annealing LR.

4.3. Evaluation Metrics

In scene text recognition tasks, recognition accuracy is one of the important performance indicators to evaluate the algorithm. For English scene text recognition tasks, the word accuracy (WA) is typically adopted as a word-level accuracy metric to assess the performance of the model. WA refers to the ratio of the number of correctly recognized words to the total number of words. The accuracy demonstrated in subsequent experiments refers to the word accuracy.
A c c u r a c y = N u m b e r c o r r e c t N u m b e r t o t a l × 100 %
The number of model parameters, computation cost, and forward inference time are important indicators for evaluating model performance. The computation cost of a model is usually represented by the total number of floating-point operations per second (FLOPs), which reflects the computational complexity and speed of the model. In practical applications, the computation cost of the model needs to be controlled within a certain range to ensure the computing speed and power consumption of the application. Additionally, the forward inference time also reflects the speed and efficiency of the model in practical applications. For resource-constrained devices such as mobile devices, the model size and forward inference time need to be controlled within a certain range to ensure the response speed and resource usage of the application. Therefore, lightweight models have become a hot research topic in recent years.
When evaluating the computation cost of a model, the total number of floating-point operations in the model is usually used as an indicator. When evaluating the forward inference time of a model, the average time required for the model to infer a single image is often used as an indicator. This time can be calculated by averaging multiple forward inferences or measured using performance-testing tools. It should be noted that the model inference time may be affected by hardware devices and input image sizes, so these factors need to be considered when evaluating the model performance.

4.4. Comparison with Existing Technologies

Our model is built upon DeiT-tiny, which is an improved model that combines knowledge distillation and ViT. The baseline model only uses DeiT-tiny as the backbone network, without adding the position branch, position-enhancement module and visual-enhancement module. As shown in Table 1, we compared our model, LSTR, with the state-of-the-art approach to scene text recognition and summarized the results of each model across six datasets. It can be seen that LSTR with real datasets performs best in each dataset, improving by nearly 3% compared with SOTA. Our model is only 0.8% lower than the SOTA average without real dataset training. However, as Table 2 shows, their model parameters are nearly ten times as fast as ours, and their operating speeds are nowhere near as fast as ours. These data prove that our model is small, accurate, and has a speed advantage. At the same time, compared to the baseline model, our model achieves a 3% improvement in recognition accuracy while maintaining the same level of parameter size and inference speed.

4.5. Comparison with Lightweight Models

To determine whether a scene text recognition model is lightweight, factors such as the number of parameters, computational complexity, and inference speed should be considered. Generally, the computational complexity should be controlled to below several billion, while the inference speed should be fast enough to run on resource-limited devices.
In this paper, the model LSTR is compared with several lightweight models. CRNN is a classic model in scene text recognition, while ViTSTR is a simple single-level model architecture that only uses the encoder part of the Transformer and achieves an efficient recognition performance. MGP is an encoder–decoder structure, with the encoder also using Transformer and adopting a multi-granularity prediction strategy to improve the decoder of the model and enhance the recognition accuracy. Due to the size of the encoder, these two models are divided into tiny, small, and base versions, providing lightweight versions (tiny, small) while pursuing high accuracy.
As shown in Table 3, we compare LSTR with these models in terms of performance. Compared with CRNN, LSTR is slightly inferior in model speed and computational complexity, but has a 17.95% higher accuracy than CRNN. Compared with VitSTR, LSTR has higher accuracy than any of its versions, but is not as fast as ViTSTR and has a higher computational complexity. This is because we added the decoder and made corresponding improvements to the decoding part, sacrificing some model speed while increasing some computational complexity, resulting in a 12.25%, 8.15%, and 6.8% increase in recognition accuracy, respectively, significantly improving the performance of our model.
Compared with the lightweight versions of MGP (tiny, small), LSTR has an average accuracy that is 3.45% and 0.8% higher, respectively. Additionally, like MGP, LSTR improved the decoder part. However, our model is not only superior in lightweight modeling but also superior in model accuracy. These data demonstrate the excellence of our model in the realm of lightweight models.
Therefore, under the premise of being the same tiny version, our model achieved the best performance while maintaining the same level of parameters, time, and FLOPS. Compared with other models in the small and base versions, we still achieve a good performance, while also having significant advantages in terms of the parameters, time, and FLOPS.

4.6. Ablation Study

In VitSTR and other studies, the image size is 224 × 224, and the patch size is 16 × 16. However, as shown in Table 4, our experiment found that compared with the size of the 224 × 224 image, the size of the 32 × 128 image can retain more essential features of the text image and has more gain on the model. Therefore, our subsequent experiments are based on this.
The results of the ablation experiments in Table 5 show that we improved the accuracy by an average of 0.3% by adding a visual-enhancement module to the baseline, which indicates that the position information is effective for the augmentation of visual information, but that alone is not a significant increase. Therefore, we added a position-enhancement module to give the position branch the learning ability, further optimize the position feature, better represent the space layout between the characters, and ease the problem of attention drift. The experiments show that our recognition rate is improved by nearly 3% compared with the baseline, which shows our improvement is effective.

5. Conclusions

In order to improve the attention drift problem of scene text recognition models and control the training costs of the model, we constructed a model called LSTR based on Vision Transformer. First, a position branch is added and an attention mechanism is used to further align the position information and visual information, thereby alleviating the problem of attention drift. Second, it is a simple model structure that uses a tiny version and has only 7 M parameters but achieves a good accuracy, maintaining the lightweight characteristics of tiny models while greatly improving the recognition performance. Compared with the existing baseline models of scene text recognition, LSTR has achieved good results in terms of the accuracy, speed, and cost, making it more applicable and feasible in practical applications, especially in high-demand tasks such as scene text recognition that requires both accuracy and efficiency.

Author Contributions

X.L. designed the model and carried out the related experiments. X.L. and J.Z. wrote the manuscript. W.S. funded the research. M.X. and Y.L. reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China Joint Fund Project under Grants U1911401 and the National Natural Science Foundation of China Joint Fund Project under Grants U1603262.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Written informed consent for publication of this paper was obtained from all authors.

Data Availability Statement

The data that support the findings of this study are openly available in the public domain.

Conflicts of Interest

The authors declared no potential conflict of interest concerning this article’s research, authorship, and publication.

References

  1. Mandavia, K.; Badelia, P.; Ghosh, S.; Chaudhuri, A. Optical Character Recognition Systems for Different Languages with Soft Computing; Springer: Cham, Switzerland, 2017; Volume 352, pp. 9–41. [Google Scholar]
  2. Nagy, G. Twenty years of document image analysis in PAMI. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 38–62. [Google Scholar] [CrossRef]
  3. Patel, C.; Shah, D.; Patel, A. Automatic number plate recognition system (anpr): A survey. Int. J. Comput. Appl. 2013, 69, 21–33. [Google Scholar] [CrossRef]
  4. Laroca., R.; Cardoso., E.V.; Lucio., D.R.; Estevam., V.; Menotti., D. On the Cross-dataset Generalization in License Plate Recognition. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Vienna, Austria, 6–8 February 2022; Volume 5, pp. 166–178. [Google Scholar] [CrossRef]
  5. Hwang, W.; Kim, S.; Seo, M.; Yim, J.; Park, S.; Park, S.; Lee, J.; Lee, B.; Lee, H. Post-OCR parsing: Building simple and robust parser via BIO tagging. In Proceedings of the Workshop on Document Intelligence at NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  6. Limonova, E.; Bezmaternykh, P.; Nikolaev, D.; Arlazarov, V. Slant rectification in Russian passport OCR system using fast Hough transform. In Proceedings of the Ninth International Conference on Machine Vision (ICMV 2016), Nice, France, 18–20 November 2016; Volume 10341, pp. 127–131. [Google Scholar]
  7. Yao, C.; Bai, X.; Sang, N.; Zhou, X.; Zhou, S.; Cao, Z. Scene text detection via holistic, multi-channel prediction. arXiv 2016, arXiv:1606.09002. [Google Scholar]
  8. Liu, J.; Liu, X.; Sheng, J.; Liang, D.; Li, X.; Liu, Q. Pyramid mask text detector. arXiv 2019, arXiv:1903.11800. [Google Scholar]
  9. Jaderberg, M.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Reading text in the wild with convolutional neural networks. Int. J. Comput. Vis. 2016, 116, 1–20. [Google Scholar] [CrossRef]
  10. Borisyuk, F.; Gordo, A.; Sivakumar, V. Rosetta: Large scale system for text detection and recognition in images. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 71–79. [Google Scholar]
  11. Luong, T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar] [CrossRef]
  12. Lee, C.Y.; Osindero, S. Recursive recurrent nets with attention modeling for ocr in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2231–2239. [Google Scholar]
  13. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  14. Cheng, Z.; Bai, F.; Xu, Y.; Zheng, G.; Pu, S.; Zhou, S. Focusing attention: Towards accurate text recognition in natural images. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5076–5084. [Google Scholar]
  15. Zheng, T.; Chen, Z.; Fang, S.; Xie, H.; Jiang, Y.G. Cdistnet: Perceiving multi-domain character distance for robust text recognition. arXiv 2021, arXiv:2111.11011. [Google Scholar]
  16. Yue, X.; Kuang, Z.; Lin, C.; Sun, H.; Zhang, W. Robustscanner: Dynamically enhancing positional clues for robust text recognition. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 135–151. [Google Scholar]
  17. Atienza, R. Vision transformer for fast and efficient scene text recognition. In Proceedings of the Document Analysis and Recognition—ICDAR 2021: 16th International Conference, Lausanne, Switzerland, 5–10 September 2021; pp. 319–334. [Google Scholar]
  18. Neumann, L.; Matas, J. Real-time scene text localization and recognition. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3538–3545. [Google Scholar]
  19. Yao, C.; Bai, X.; Shi, B.; Liu, W. Strokelets: A learned multi-scale representation for scene text recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4042–4049. [Google Scholar]
  20. Yang, X.; He, D.; Zhou, Z.; Kifer, D.; Giles, C.L. Learning to read irregular text with attention mechanisms. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; Volume 1, p. 3. [Google Scholar]
  21. Lee, C.Y.; Bhardwaj, A.; Di, W.; Jagadeesh, V.; Piramuthu, R. Region-based discriminative feature pooling for scene text recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4050–4057. [Google Scholar]
  22. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar] [CrossRef]
  23. Bai, F.; Cheng, Z.; Niu, Y.; Pu, S.; Zhou, S. Edit probability for scene text recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1508–1516. [Google Scholar]
  24. Wan, Z.; Zhang, J.; Zhang, L.; Luo, J.; Yao, C. On vocabulary reliance in scene text recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11425–11434. [Google Scholar]
  25. Liao, M.; Zhang, J.; Wan, Z.; Xie, F.; Liang, J.; Lyu, P.; Yao, C.; Bai, X. Scene text recognition from two-dimensional perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8714–8721. [Google Scholar]
  26. Wan, Z.; He, M.; Chen, H.; Bai, X.; Yao, C. Textscanner: Reading characters in order for robust scene text recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12120–12127. [Google Scholar]
  27. Liu, W.; Chen, C.; Wong, K.Y.K.; Su, Z.; Han, J. Star-net: A spatial attention residue network for scene text recognition. In Proceedings of the BMVC, York, UK, 19–22 September 2016; Volume 2, p. 7. [Google Scholar]
  28. Wan, Z.; Xie, F.; Liu, Y.; Bai, X.; Yao, C. 2D-CTC for scene text recognition. arXiv 2019, arXiv:1907.09705. [Google Scholar]
  29. Shi, B.; Bai, X.; Yao, C. An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 2298–2304. [Google Scholar] [CrossRef] [PubMed]
  30. Gao, Y.; Chen, Y.; Wang, J.; Lu, H. Reading scene text with attention convolutional sequence modeling. arXiv 2017, arXiv:1709.04303. [Google Scholar]
  31. Li, H.; Wang, P.; Shen, C.; Zhang, G. Show, attend and read: A simple and strong baseline for irregular text recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8610–8617. [Google Scholar]
  32. Wang, T.; Zhu, Y.; Jin, L.; Luo, C.; Chen, X.; Wu, Y.; Wang, Q.; Cai, M. Decoupled attention network for text recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12216–12224. [Google Scholar]
  33. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
  34. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  35. Jaderberg, M.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Synthetic data and artificial neural networks for natural scene text recognition. arXiv 2014, arXiv:1406.2227. [Google Scholar]
  36. Gupta, A.; Vedaldi, A.; Zisserman, A. Synthetic data for text localisation in natural images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2315–2324. [Google Scholar]
  37. Bautista, D.; Atienza, R. Scene text recognition with permuted autoregressive sequence models. In Proceedings of the Computer Vision—ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; pp. 178–196. [Google Scholar]
  38. Yin, F.; Liu, C.L.; Lu, S. A new artwork dataset for experiments in artwork retrieval and recognition. In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; Volume 1, pp. 1220–1225. [Google Scholar]
  39. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  40. Liu, Y.; Jin, L.; Zhang, S.; Zhang, Z. LSVT: Large scale vocabulary training for scene text recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8774–8781. [Google Scholar]
  41. Liu, Y.; Chen, H.; Shen, C.; He, T.; Jin, L.; Wang, L.; Zhang, Z. Multi-lingual scene text dataset and benchmarks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7602–7611. [Google Scholar]
  42. Yin, F.; Zhang, Y.; Liu, C.L.; Lu, S. ICDAR2017 competition on reading Chinese text in the wild (RCTW-17). In Proceedings of the 2018 13th IAPR International Workshop on Document Analysis Systems (DAS), Vienna, Austria, 24–27 April 2018; pp. 178–183. [Google Scholar]
  43. Krylov, I.; Nosov, S.; Sovrasov, V. Open images v5 text annotation and yet another mask text spotter. In Proceedings of the Asian Conference on Machine Learning, Virtually, 17–19 November 2021; pp. 379–389. [Google Scholar]
  44. Mishra, A.; Shekhar, R.; Jawahar, C. Scene text recognition using higher order language priors. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 222–235. [Google Scholar]
  45. Shen, C.; Liu, X.; Yang, Y.; Jin, L.; Zhang, Z.; Chen, H.; Liu, F.; Zhang, Y.; Wu, W.; Liu, Y.; et al. ReCTS: A Large-Scale Reusable Corpus for Text Spotting. IEEE Trans. Multimed. 2019, 21, 1468–1481. [Google Scholar]
  46. Schneider, J.; Puigcerver, J.; Cissé, M.; Ginev, D.; Gruenstein, A.; Gutherie, D.; Jayaraman, D.; Kassis, T.; Kazemzadeh, F.; Llados, J.; et al. U-BER: A Large-Scale Dataset for Street Scene Text Reading. arXiv 2019, arXiv:1911.03585. [Google Scholar]
  47. Mishra, A.; Alahari, K.; Jawahar, C.V. Top-Down and Bottom-Up Cues for Scene Text Recognition. In Proceedings of the 2012 International Conference on Frontiers in Handwriting Recognition, Bari, Italy, 18–20 September 2012; pp. 359–364. [Google Scholar] [CrossRef]
  48. Wang, K.; Babenko, B.; Belongie, S. End-to-End Scene Text Recognition. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, Beijing, China, 18–21 September 2011; pp. 296–303. [Google Scholar] [CrossRef]
  49. Yin, X.Y.; Wang, X.; Zhang, P.; Wen, L.; Bai, X.; Lei, Z.; Li, S.Z. Robust Text Detection in Natural Images with Edge-Enhanced Maximally Stable Extremal Regions. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 151–156. [Google Scholar] [CrossRef]
  50. Karatzas, D.; Gomez-Bigorda, L.; Nicolaou, A.; Ghosh, S.; Bagdanov, A.; Iwamura, M.; Matas, J.; Neumann, L.; Chandrasekhar, V.; Lu, S.; et al. ICDAR 2015 Competition on Robust Reading. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015; pp. 1156–1160. [Google Scholar] [CrossRef]
  51. Quy Pham, T.; Kieu Nguyen, H.; Cheung, N.M. Recognizing text in perspective view of scene images based on unsupervised feature learning. Pattern Recognit. 2013, 46, 1834–1848. [Google Scholar]
Figure 1. Model architecture of the LSTR.
Figure 1. Model architecture of the LSTR.
Sensors 23 04490 g001
Figure 2. Transformer encoder block.
Figure 2. Transformer encoder block.
Sensors 23 04490 g002
Figure 3. Decoder frame.
Figure 3. Decoder frame.
Sensors 23 04490 g003
Figure 4. The query vector uses an upper triangular mask, where the ‘X’ values are assigned as ‘-inf’.
Figure 4. The query vector uses an upper triangular mask, where the ‘X’ values are assigned as ‘-inf’.
Sensors 23 04490 g004
Figure 5. Comparison between synthetic datasets and real datasets. The (left) is the synthetic datasets and the (right) is the real datasets.
Figure 5. Comparison between synthetic datasets and real datasets. The (left) is the synthetic datasets and the (right) is the real datasets.
Sensors 23 04490 g005
Table 1. Comparison of performance with existing technologies.
Table 1. Comparison of performance with existing technologies.
MethodYearTrain DataIC13SVTIIIT5kIC15SVTPCUTE80Avg
CRNN2017S89.480.181.865.365.961.574
RARE2016S92.385.48673.575.47180.6
TRBA2019S93.487.687.877.478.175.0283.22
SAR2019S9184.591.569.276.483.582.68
VisionLAN2021S95.791.795.883.78688.590.23
ViTSTR2021S93.287.788.478.581.881.385.15
ABI2021S97.493.596.28689.389.291.93
cdist2021S94.693.896.586.289.789.5891.73
SVTR2022S97.291.796.386.688.495.192.55
CornerTransformer2022S96.494.695.986.391.59292.78
baseline-S94.8691.893.7683.2683.4186.888.98
LSTR-S96.9694.1295.687.4188.8388.8391.95
LSTR-S, R98.696.4498.1390.9492.5597.2295.64
Table 2. Comparison of the parameters with existing technologies.
Table 2. Comparison of the parameters with existing technologies.
MethodYearAvgParameters ( 1 × 10 6 )Time (ms/Image)
CRNN2017748.36.3
RARE201680.610.818.8
TRBA201983.2249.622.8
SAR201982.6857.5120.0
VisionLAN202190.2332.828.0
ViTSTR202185.1585.59.8
ABI202191.9336.750.6
cdist202191.7365.4123.28
SVTR202292.5540.818.0
CornerTransformer202292.7885.7294.9
baseline-88.985.413.6
LSTR-91.957.113.6
LSTR *-95.647.113.6
* Training dataset is S and R.
Table 3. Performance comparison with lightweight models.
Table 3. Performance comparison with lightweight models.
MethodYearIC13SVTIIIT5kIC15SVTPCUTE80AvgParametersTimeFLOPS
CRNN201789.480.181.865.365.961.5748.36.31.4
ViTSTR-tiny202190.893.283..772.074.565.079.15.49.31.3
MGP-tiny202294.091.194.383.383.584.388.521.012.07.2
LSTR96.9694.1295.687.4188.8388.8391.957.113.65.7
ViTSTR-small202191.787.386.677.981.477.983.821.59.54.6
MGP-small202296.893.595.386.187.387.991.1552.612.225.4
ViTSTR-base202193.287.788.478.581.881.385.1585.59.817.6
MGP-base202297.394.796.487.291.090.392.81148.012.394.7
The unit of Parameters is ( 1 × 10 6 ), Time is (ms/image), and FLOPS is ( 1 × 10 9 ).
Table 4. Comparison of effects of different image sizes on performance.
Table 4. Comparison of effects of different image sizes on performance.
MethodImage Size (Patch)IC13_857SVTIIIT5kIC15SVTPCUTE80Avg
baseline32 × 128 (4 × 4)94.8691.8093.7683.2683.4186.8088.98
baseline224 × 224 (16 × 16)95.7990.1094.1082.7183.5683.6888.32
Table 5. Comparison of ablation experiments.
Table 5. Comparison of ablation experiments.
MethodIC13SVTIIIT5kIC15SVTPCUTE80Avg
baseline94.8691.8093.7683.2683.4186.8088.98
+VisAug95.2191.6594.3683.4385.2785.7689.28
+posAug+VisAug96.9694.1295.6087.4188.8388.8391.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luan, X.; Zhang, J.; Xu, M.; Silamu, W.; Li, Y. Lightweight Scene Text Recognition Based on Transformer. Sensors 2023, 23, 4490. https://doi.org/10.3390/s23094490

AMA Style

Luan X, Zhang J, Xu M, Silamu W, Li Y. Lightweight Scene Text Recognition Based on Transformer. Sensors. 2023; 23(9):4490. https://doi.org/10.3390/s23094490

Chicago/Turabian Style

Luan, Xin, Jinwei Zhang, Miaomiao Xu, Wushouer Silamu, and Yanbing Li. 2023. "Lightweight Scene Text Recognition Based on Transformer" Sensors 23, no. 9: 4490. https://doi.org/10.3390/s23094490

APA Style

Luan, X., Zhang, J., Xu, M., Silamu, W., & Li, Y. (2023). Lightweight Scene Text Recognition Based on Transformer. Sensors, 23(9), 4490. https://doi.org/10.3390/s23094490

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop