Next Article in Journal
Research on a 3D Point Cloud Map Learning Algorithm Based on Point Normal Constraints
Previous Article in Journal
Design, Calibration and Morphological Characterization of a Flexible Sensor with Adjustable Chemical Sensitivity and Possible Applications to Sports Medicine
Previous Article in Special Issue
Research on Point Cloud Acquisition and Calibration of Deep Hole Inner Surfaces Based on Collimated Ring Laser Beams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HubNet: An E2E Model for Wheel Hub Text Detection and Recognition Using Global and Local Features

Image Processing Center, School of Astronautics, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6183; https://doi.org/10.3390/s24196183
Submission received: 14 August 2024 / Revised: 13 September 2024 / Accepted: 19 September 2024 / Published: 24 September 2024

Abstract

:
Automatic detection and recognition of wheel hub text, which can boost the efficiency and accuracy of product information recording, are undermined by the obscurity and orientation variability of text on wheel hubs. To address these issues, this paper constructs a wheel hub text dataset and proposes a wheel hub text detection and recognition model called HubNet. The dataset captured images on real industrial production line scenes, including 446 images, 934 word instances, and 2947 character instances. HubNet is an end-to-end text detection and recognition model, not only comprising conventional detection and recognition heads but also incorporating a feature cross-fusion module, which improves the accuracy of recognizing wheel hub texts by utilizing both global and local features. Experimental results show that on the wheel hub text dataset, the HubNet achieves an accuracy of 86.5%, a recall of 79.4%, and an F1-score of 0.828, and the feature cross-fusion module increases the accuracy by 2% to 4%. The wheel hub dataset and the HubNet offer a significant reference for automatic detection and recognition of wheel hub text.

1. Introduction

Wheel hub, as a typical industrial die-casting component, is often marked with Latin characters and Arabic numerals using character embossing and laser marking techniques during the production process. These characters are arranged either transversely or in a circular pattern on the metal surface of the hub, recording critical information such as production serial number, date, part category, material, and process dimensions [1]. In actual production, due to the vast quantity of products, manually recording casting codes is inefficient, and errors are prone to occur over long periods of work.
The research on scene text detection and recognition tasks has a long history [2]. However, compared with normal scene texts, hub texts have low visual contrast, susceptibility to surface corrosion, and complex backgrounds with diverse orientations of the text. These characteristics significantly increase the difficulty of hub text detection and recognition. Therefore, existing text detection and recognition methods [3,4,5,6], which are designed to deal with natural scene text, have poor performances when directly applied to wheel hub text.
Current methods applied to text detection and recognition can primarily be categorized into traditional methods and deep learning methods. Traditional methods [7,8] often employ template matching algorithms for the identification and detection of cast characters, where image features are manually designed to locate and segment individual characters. Subsequently, character recognition is performed through feature classifiers based on machine learning, such as SVMs, random forests, etc. Traditional text recognition methods typically require substantial prior knowledge and can only be applied in environments with minimal interference, making them suitable for simple detection tasks only, which do not work well in the task of detecting and recognizing text on a wheel hub.
With the development of deep learning technology, deep neural networks are better able to extract image features of text, facilitating detection and recognition tasks in complex scenes. The text detection and recognition methods using deep neural networks are categorized into cascaded approaches and end-to-end approaches. The cascaded approach treats text detection [9,10,11,12,13,14,15,16] and text recognition [8,17,18,19] as two separate tasks, training a text detection network and a text recognition network independently, and then cascading the two networks to achieve the functionality of both detection and recognition. Due to the independent training of the text detection network and the text recognition network without any information exchange between them, the accuracy of the cascaded method tends to be relatively lower. In contrast, end-to-end methods [3,4,5,6,20,21] perform both detection and recognition simultaneously within a single network. These networks struggle to converge when trained on the wheel hub text images. because hub text exhibits low contrast and unclear engraving, which is difficult to detect and recognize the text at the same time.
To address the issues of low character contrast, unclear embossing, and diverse orientations in wheel hub character detection and recognition, this paper designed a wheel hub text detection and recognition model called HubNet and constructed a dataset to support the training of the model. Specifically, a wheel hub text dataset is constructed. The images in the dataset are collected from real industrial production line scenarios, featuring evenly distributed character categories and diverse orientations. Subsequently, based on existing natural scene text detection and recognition algorithms, we propose HubNet, a wheel hub character detection and recognition model utilizing global and local information. HubNet comprises a detection head, a recognition head, and a feature cross-fusion module. The detection head employs the state-of-the-art text detection network DBNet++ [22] as both the word detection branch and the character detection branch. Simultaneously, inspired by GLASS [23], a feature cross-fusion module is designed to integrate global and local features, thereby enhancing the accuracy of text recognition. The lightweight character recognition head is designed based on residual modules. The contributions of this paper are summarized as follows:
  • We establish a wheel hub text dataset, including 446 images, 934 word instances, and 2947 character instances. The images in the dataset originate from authentic factory production line scenes with uniformly distributed character categories and diverse orientations.
  • We propose HubNet, an end-to-end model for text detection and recognition utilizing both global and local features, combining directional information from global features and details from local features, which improves the model’s accuracy in text detection and recognition.
The remaining sections of this paper are structured as follows: Section 2 introduces the dataset constructed for this study, analyzes its characteristics, and presents the structure of HubNet; Section 3 details the experimental setup and validates the model through experimentation; Section 4 discusses the results and limitations of the proposed model.

2. Materials and Methods

2.1. Dataset

Currently, publicly available text detection and recognition datasets predominantly focus on text in natural scenarios [24,25,26,27], whereas text on a wheel hub exhibits notable differences in features compared with those in common environments. In support of text detection and recognition tasks for wheel hubs, we utilized line scan cameras to capture images of wheel hubs and, based on these images, constructed a wheel hub text dataset.
The wheel hub images in the dataset have a resolution of 5120 × 5120 pixels, as shown in the left image of Figure 1. Attempting to annotate these images directly would be quite challenging, and network training can be hindered by images with such high resolutions. To reduce the difficulty of dataset annotation, meet the requirements for network training, and ensure that any character can be fully displayed within a single image, we divide the original images into 11 × 11 images in an overlapping way, with a resolution of 512 × 512 pixels each. Images that turn out to be black after the division are discarded. Some of the images after division are shown in the right image of Figure 1. Part of the images in the dataset are shown in Figure 2.
The wheel hub text dataset produced for this paper exhibits the following characteristics and advantages:
Data collected from real production line scenarios: Unlike existing public datasets where images are sourced from natural scenarios, the hub dataset focuses on images captured specifically from industrial production line scenes. These images are gathered directly from authentic industrial production lines, offering essential data support for the task of text detection and recognition on wheel hubs under industrial conditions. The hub dataset comprises a total of 446 images, 934 word instances, and 2947 character instances, as detailed in Table 1.
The orientation of the text in the wheel hub text dataset is more diverse: Due to the fact that workpieces do not enter the production line at a fixed angle during the manufacturing process, the orientation of the workpieces varies in the images captured by the line scan camera on the production line, leading to diversity in the direction of the embossed characters on the workpieces. Several existing public datasets [24,25,26,27] focusing on natural scene text have already acknowledged the significance of text orientation in text detection and recognition tasks and have incorporated images with multi-directional text into their datasets. However, the orientation of the text in these existing datasets is still mostly concentrated within the range of −60° to +60°. The dataset constructed in this paper, targeting the true scenarios on the production line, features text orientations ranging from −180° to +180°, thus exhibiting a greater degree of directional diversity. The statistical chart is illustrated in Figure 3a.
The wheel hub text dataset features annotations at both the word and character levels: Most existing public datasets offer only word-level annotations, whereas recognition networks require character-level annotations. This discrepancy necessitates that state-of-the-art deep learning algorithms pre-train on synthetic datasets. The character-level annotations provided in the wheel hub text dataset enable deep learning models to be trained and fine-tuned at the character level. Additionally, within the word-level annotations, the distribution of aspect ratios is diverse, which is beneficial for detecting and recognizing characters with different aspect ratios on the wheel hub. This distribution is depicted in Figure 3b. Concurrently, in the character-level annotations, characters are evenly distributed across categories, avoiding the overrepresentation of any particular character type, as shown in Figure 3c.

2.2. Model

To detect and recognize wheel hub text in the context of an industrial production line scenario, this paper designs HubNet, a wheel hub text detection and recognition model using global and local features. The architecture of HubNet is illustrated in Figure 4.
HubNet consists of a backbone, a detection head, a feature cross-fusion module, and a recognition head. Its operational workflow is as follows: The network takes images containing wheel hub text as input. These images first pass through the backbone and obtain the feature map. The detection head yields word-level and character-level bounding polygons by using the feature map. and a global feature map. Subsequently, the character-level bounding polygons are used to crop the global feature map, obtaining the global feature F g . The original images are cropped according to these bounding polygons and fed into the recognition feature extractor, producing the local feature F l . Both F g and F l are then fed into the feature cross-fusion module, generating the fused feature F f . Ultimately, F f serves as the input to the recognition head, leading to the classification of each character. The output of the network comprises the locations of bounding polygons for words in the image and the category of each character.

2.2.1. Detection Head

Word-level detection should be coupled with recognition methods built on word-level optimization, which has a significantly larger search space than direct character classification [6]. This inevitably makes the subsequent recognition head more complicated and difficult to train by requiring a longer training time with a larger amount of training samples. Besides, though character-level detection can make the subsequent recognition task easier by dealing with it as a character classification task, it treats characters individually, raising the issue of how to determine which characters belong to the same word.
Therefore, the detection head designed in this paper is bifurcated into a word detection branch and a character detection branch, with both adopting the detector of the DBNet++ [22] model, which predicts the mask as well as the boundary of the text by using differentiable binarization. The detection head can predict the bounding polygons [28] of words and characters at the same time, while the recognition head can use the patches cropped by character polygons to classify the category of each character, and through the positioning of words and characters, it can be quite straightforward to determine which word each character belongs to.
In details, the word detection branch and the character detection branch have the same architecture but are trained on word level and character level, respectively. Both branches predict the probability map P and the threshold map T, and an approximate binary map B ^ is calculated by P and F using the following formula, where ( i , j ) indicates the coordinate point in the map:
B ^ i , j = 1 1 + e k ( P i , j T i , j ) ,
where k indicates the amplifying factor and it is set to 50 empirically. The differentiable binarization can not only help differentiate text regions from the background but also separate text instances that are closely joined [22].

2.2.2. Feature Cross-Fusion

There are currently two mainstream approaches for text recognition: The first is a modular cascaded method, as illustrated in Figure 5a, which cascades independent text detection and recognition models. The recognition model uses uniformly aligned and resized text crops as inputs in the vertical direction, abstracting away issues of scale and rotation. The second approach is a combined end-to-end (E2E) architecture, depicted in Figure 5b, which adds a recognition head that directly operates on the latent features of the detection model. Feature sampling replaces cropping, allowing for joint end-to-end training of detection and recognition.
As the E2E architecture has increasingly become the dominant structure, scale, and rotation-invariant image cropping has been supplanted by feature sampling, with sampled features becoming highly sensitive to both scale and rotation. While the joint optimization approach in E2E architectures has improved recognition performance for words that are not of extreme sizes or orientations, words at scale extremes and those with severe rotations tend to be poorly recognized within the E2E framework [23].
In order to leverage the strengths of both architectures, this paper proposes a feature cross-fusion module, as illustrated in Figure 6.
The feature cross-fusion module takes the global feature F g and the local feature F l as inputs; first, it concatenates F g and F l to obtain f c o n c a t with a shape of ( 2 C , W , H ) .
f c o n c a t = c o n c a t ( F g , F l ) ,
Subsequently generate the attention vector y R 2 C , as shown in the formula below:
y = v e c ( f c o n c a t ) T × v e c ( v T f c o n c a t ) .
Among them, v R 2 C is a learnable vector, and the v e c ( ) operation transforms a tensor of shape (C, W, H) into a matrix of shape (C, WH). The fused feature map is ultimately generated using the attention vector y, as shown in the formula below:
F f = f c o n c a t + y T f c o n c a t .
Ultimately, the fused feature map F f is fed into the recognition head to classify the characters in the image.
The feature cross-fusion module overcomes the drawbacks of both the cascaded text detection and recognition network (where the text detection network and the text recognition network are trained separately) and the end-to-end text detection and recognition network (where both networks are trained simultaneously), thereby enhancing the accuracy of character recognition.

2.2.3. Recognition Head

The character recognition algorithm designed in this paper adopts ResNet [29] as its base, taking the feature map F f R C × H × W as input, where the C is 256. This results in a classification outcome with 37 categories (1 background + 10 Arabic numerals + 26 alphabetic characters). The specific model of the network is illustrated in Figure 7.
The feature map F f R C × H × W , obtained after feature fusion, is first fed into four cascaded ResBlock modules. Following this, average pooling is applied to the feature map, resulting in a vector F p R C × 1 × 1 . This vector F p R C × 1 × 1 is then flattened and passed through a Multi-Layer Perceptron (MLP), yielding probabilities for 37 classes. Consequently, the final outcome of character classification is achieved.

2.2.4. Loss Function

The training of the wheel hub text detection and recognition network using global and local features designed in this chapter is divided into two sections: the detection head and the recognition head. The detection head and the recognition head are trained separately.
For the detection head, following [22], the loss function L d can be expressed as the weighted sum of the loss of the probability map L s , the loss of the binary map L b , and the loss of the threshold map L t .
L d = L s + α L b + β L t
Based on the values dictated by the loss function, the parameters α and β are respectively set to 1.0 and 1.0. Here, L s and L b denote the Binary Cross-Entropy (BCE) loss functions. L t represents the sum of L1 distances between predictions and labels within the dilated text polygon.
For the recognition head, the loss function L r employs the multi-class cross-entropy loss, defined as follows: where x represents the predicted probabilities, y is the ground truth label, w is the weight, C denotes the number of classes, and N is the number of samples in a batch:
L r = 1 N n = 1 N c = 1 C w c log e x p ( x n , c ) i = 1 C e x p ( x n , c ) y n , c

3. Experiment

3.1. Experimental Environment Configuration

Compared to traditional character detection and recognition methods, artificial intelligence and deep learning possess a larger parameter space and superior feature learning capabilities. Upon entering the era of big data, the deep learning process involves massive matrix multiplications and intricate non-linear computations, demanding hardware with substantial computational power and adequate storage capacity. We train and test the HubNet on an NVIDIA GeForce RTX 3090 GPU with 24 GB of memory. The experimental environment and hardware configuration used in this paper are detailed in Table 2.
In the process of developing the proposed model, this paper initially trains the detection head and the recognition head separately. Subsequently, both branches are integrated and undergo fine-tuning together.
For the detection head, we first utilize pre-trained weights obtained from 100,000 iterations of pre-training on the SynthText dataset. Following this, we fine-tune the detection head for 1200 epochs on the wheel hub text dataset constructed in this paper, with the batch size set to 16. We adopt a poly learning rate strategy, where the current iteration’s learning rate equals the initial learning rate multiplied by ( 1 i t e r m a x _ i t e r ) p o w e r , with the initial learning rate set to 0.007 and p o w e r equal to 0.9. We apply a weight decay of 0.0001 and momentum of 0.9. m a x _ i t e r represents the maximum number of iterations. Data augmentation for the training data includes random rotation with an angular range of (−10°, 10°), random cropping, and random flipping. All processed images are resized to 640 × 640 to enhance training efficiency.
For the recognition head, network parameters are initialized using Kaiming initialization [30], followed by 50 rounds of training on the hub character dataset constructed for this paper. The batch size for training is set to 64, with the input being the feature maps obtained from the backbone network and the final prediction being probability values for 37 classes. During the training of the recognition head, the initial learning rate is set to 0.001, and after every 20 rounds of training, the learning rate is reduced to one-tenth of its previous value.Data augmentation for the training data includes random offset of the bounding boxes with a range of (−5, 5) pixels, as well as random rotation with an angular range of (−180°, 180°). Throughout the training process of the detection head, no stretching operations are performed on the cropped feature maps to ensure that the details within the feature maps remain unchanged.
Finally, the pre-trained detection head and recognition head are combined for joint training. The number of training rounds is set to 50, with a learning rate of 0.0001. After every 20 rounds, the learning rate is decreased to 0.9 times the current learning rate. At this stage, data augmentation is applied to the input images, including random rotation within an angular range of (−10°, 10°), random cropping, random flipping, random lighting variations, Gaussian noise, and salt-and-pepper noise. The loss and accuracy curves of the training process are shown in Figure 8.

3.2. Analysis on Backbone and Feature Extractor

Deformable convolutions [31] provide flexible receptive fields for the backbone network with very little extra computational cost, especially advantageous for targets with large aspect ratios, such as word-level detections in text detection, and DBNet++ [22] improves model performance by adopting deformable convolutions. In the experiment, we also consider EfficientNet [32] as one of the candidate backbone networks. For the feature extractor, we conducted experiments with standard ResNet-18 and ResNet-18 that utilize deformable convolutions. Through experimentation on the dataset constructed in Section 2.1, it is observed, as shown in Table 3, that incorporating deformable convolutions boosts word-level detection accuracy by 4.3%, while the effect on character-level detection accuracy is not significantly influenced by whether the backbone network uses deformable convolutions. Subsequent ablation studies are carried out on a backbone network composed of a ResNet-50 with deformable convolutions, alongside a feature extractor based on a ResNet-18 utilizing standard convolutions.

3.3. Ablation Study on Feature Cross-Fusion Module

To validate the role of the feature cross-fusion module, we replace it with alternatives that use only global features, only local features, concatenate global and local features, or add global and local features together. As shown in Table 4, the feature cross-fusion module outperforms these alternatives by increasing recognition precision by 15.6%, 3.1%, 2.6%, and 1.7%, respectively. Further analysis reveals that using only global features for character recognition yields the lowest accuracy, since global features derived from the backbone network are more conducive to the binary classification task of text detection rather than the multiclass classification task of text recognition involving 37 categories. Using only local features for recognition neglects directional or contextual information present in global features, leading to a decrease in recognition success. Similarly, simply concatenating or summing global and local features does not match the performance of the feature cross-fusion module due to the mismatch in scale between global and local features, which requires harmonization through the feature cross-fusion process. Visualizations of the experiment results are displayed in Figure 9.

3.4. Quantitative Comparison

We compared withpare the proposed HubNet against four representative baselines, including SPTS [20], SwinTextSpotter [4], Mask TextSpotter v3 [3], and ABCNet [33]. We conduct the comparisons with other approaches on both detection and recognition at the word level. For a fair comparison, all methods are under the same train/test split. Experiment results are shown in Table 5, where we can see that HubNet outperforms the current methods on the Wheel Hub Text Dataset.
Networks such as ABCNet and MaskTextSpotter v3 perform poorly on the dataset. According to our analysis, the reason is that these networks attempt to detect and recognize hub text simultaneously, and unclearly embossed text makes it difficult for the networks to converge, thus leading to poorer performance. Due to the failure to integrate global and local information, SwinTextSpotter and SPTS v2 also exhibit poor recognition performance on texts with extreme orientations and unclear embossing.

4. Discussion

To deal with the challenge of automatically detecting and recognizing the text on the wheel hub, we set up a dataset consisting of wheel hub images, which do good for the search of text detection and recognition, and it can also be a completion of a natural scene text dataset.
HubNet outperforms existing models on the wheel dataset. The feature cross-fusion module combines global features with local features, improving the accuracy of text detection and recognition and allowing the model to successfully detect and recognize text with extreme orientations and unclear imprints, proving the significance of global and local features in text detection and recognition tasks.
However, despite surpassing current models, the recognition accuracy of HubNet remains contingent upon the clarity of the imprinted characters on the wheel hub. Moreover, the training methods of current models are complex, and cascaded training may lead to a decline in model accuracy. Future work will involve refining this model by exploring more advanced networks for the recognition head and adopting an end-to-end training approach, which may further elevate model accuracy, enabling efficient and precise end-to-end detection and recognition.

5. Conclusions

Compared to natural scene texts, characters on wheel hub present challenges due to erosion, low contrast, and varied orientations, making automated detection and recognition particularly difficult. To address these issues, a wheel hub text dataset was constructed, featuring a diverse distribution of character orientations and word aspect ratios with a relatively balanced distribution of character label categories, providing the data foundation for the research. Moreover, we proposed HubNet, a wheel hub text detection and recognition model consisting of a detection head and recognition head conventionally, and a feature cross-fusion module proposed in this paper, which utilizes global and local features simultaneously to improve the recognition accuracy. Experiments conducted on HubNet have shown the effectiveness of the model and the feature cross-fusion module, which achieves a favorable outcome on the wheel hub text dataset.

Author Contributions

Conceptualization, Y.Z. and C.M.; Methodology, Y.Z.; Resources, C.M.; Writing—original draft, Y.Z.; Writing—review & editing, C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are available at https://github.com/Zeng1376/WheelHubTextDetectionRecognition, accessed on 1 January 2024.

Acknowledgments

Deep gratitude is extended to Hongsheng He from the University of Alabama for the invaluable assistance rendered in the refinement and enhancement of the manuscript, and AI+Computed Optics Joint Laboratory of Vision Optics and Beihang University for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shabestari, B.N.; Miller, J.W.V.; Wedding, V. Low-cost real-time automatic wheel classification system. Mach. Vis. Appl. Archit. Syst. Integr. 1992, 1823, 218–222. [Google Scholar]
  2. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  3. Liao, M.; Pang, G.; Huang, J.; Hassner, T.; Bai, X. Mask textspotter v3: Segmentation proposal network for robust scene text spotting. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 706–722. [Google Scholar]
  4. Huang, M.; Liu, Y.; Peng, Z.; Liu, C.; Lin, D.; Zhu, S.; Yuan, N.; Ding, K.; Jin, L. Swintextspotter: Scene text spotting via better synergy between text detection and text recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 8–24 June 2022; pp. 4593–4603. [Google Scholar]
  5. Ye, M.; Zhang, J.; Zhao, S.; Liu, J.; Liu, T.; Du, B.; Tao, D. Deepsolo: Let transformer decoder with explicit points solo for text spotting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19348–19357. [Google Scholar]
  6. Xing, L.; Tian, Z.; Huang, W.; Scott, M.R. Convolutional character networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9126–9136. [Google Scholar]
  7. Wang, T.; Wu, D.J.; Coates, A.; Ng, A.Y. End-to-end text recognition with convolutional neural networks. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; IEEE: Washington, DC, USA, 2012; pp. 3304–3308. [Google Scholar]
  8. Matas, J.; Chum, O.; Urban, M.; Pajdla, T. Robust wide-baseline stereo from maximally stable extremal regions. Image Vis. Comput. 2004, 22, 761–767. [Google Scholar] [CrossRef]
  9. Liu, Y.; Jin, L. Deep matching prior network: Toward tighter multi-oriented text detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1962–1969. [Google Scholar]
  10. Liao, M.; Zhu, Z.; Shi, B.; Xia, G.S.; Bai, X. Rotation-sensitive regression for oriented scene text detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5909–5918. [Google Scholar]
  11. Zhou, X.; Yao, C.; Wen, H.; Wang, Y.; Zhou, S.; He, W.; Liang, J. East: An efficient and accurate scene text detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5551–5560. [Google Scholar]
  12. Liu, Y.; Jin, L.; Fang, C. Arbitrarily shaped scene text detection with a mask tightness text detector. IEEE Trans. Image Process. 2019, 29, 2918–2930. [Google Scholar] [CrossRef] [PubMed]
  13. Zhu, Y.; Chen, J.; Liang, L.; Kuang, Z.; Jin, L.; Zhang, W. Fourier contour embedding for arbitrary-shaped text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3123–3131. [Google Scholar]
  14. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  15. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  16. Liao, M.; Shi, B.; Bai, X.; Wang, X.; Liu, W. Textboxes: A fast text detector with a single deep neural network. Proc. AAAI Conf. Artif. Intell. 2017, 31, 4161–4167. [Google Scholar] [CrossRef]
  17. Wang, K.; Babenko, B.; Belongie, S. End-to-end scene text recognition. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1457–1464. [Google Scholar]
  18. Fang, S.; Xie, H.; Wang, Y.; Mao, Z.; Zhang, Y. Read like humans: Autonomous, bidirectional and iterative language modeling for scene text recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7098–7107. [Google Scholar]
  19. Qiao, Z.; Zhou, Y.; Yang, D.; Zhou, Y.; Wang, W. Seed: Semantics enhanced encoder-decoder framework for scene text recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13528–13537. [Google Scholar]
  20. Liu, Y.; Zhang, J.; Peng, D.; Huang, M.; Wang, X.; Tang, J.; Huang, C.; Lin, D.; Shen, C.; Bai, X.; et al. Spts v2: Single-point scene text spotting. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15665–15679. [Google Scholar] [CrossRef] [PubMed]
  21. Zhang, X.; Su, Y.; Tripathi, S.; Tu, Z. Text spotting transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 9519–9528. [Google Scholar]
  22. Liao, M.; Zou, Z.; Wan, Z.; Yao, C.; Bai, X. Real-time scene text detection with differentiable binarization and adaptive scale fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 919–931. [Google Scholar] [CrossRef] [PubMed]
  23. Ronen, R.; Tsiper, S.; Anschel, O.; Lavi, I.; Markovitz, A.; Manmatha, R. Glass: Global to local attention for scene-text spotting. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 249–266. [Google Scholar]
  24. Gupta, A.; Vedaldi, A.; Zisserman, A. Synthetic data for text localisation in natural images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2315–2324. [Google Scholar]
  25. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  26. Karatzas, D.; Gomez-Bigorda, L.; Nicolaou, A.; Ghosh, S.; Bagdanov, A.; Iwamura, M.; Matas, J.; Neumann, L.; Chandrasekhar, V.R.; Lu, S. ICDAR 2015 competition on robust reading. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1156–1160. [Google Scholar]
  27. Ch’ng, C.K.; Chan, C.S. Total-text: A comprehensive dataset for scene text detection and recognition. In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; IEEE: Piscataway, NJ, USA, 2017; Volume 1, pp. 935–942. [Google Scholar]
  28. Vatti, B.R. A generic solution to polygon clipping. Commun. ACM 1992, 35, 56–63. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  31. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  32. Tan, M. Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
  33. Liu, Y.; Chen, H.; Shen, C.; He, T.; Jin, L.; Wang, L. ABCNet: Real-time scene text spotting with adaptive bezier-curve network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9809–9818. [Google Scholar]
Figure 1. Collected wheel hub images (left) and image cropped at the red box (right).
Figure 1. Collected wheel hub images (left) and image cropped at the red box (right).
Sensors 24 06183 g001
Figure 2. Part of the images in the dataset.
Figure 2. Part of the images in the dataset.
Sensors 24 06183 g002
Figure 3. Analysis of Wheel Hub Text Dataset. (a) Distribution of character orientation in the dataset. (b) Distribution of word aspect ratio in the dataset. (c) Distribution of character labels in the dataset.
Figure 3. Analysis of Wheel Hub Text Dataset. (a) Distribution of character orientation in the dataset. (b) Distribution of word aspect ratio in the dataset. (c) Distribution of character labels in the dataset.
Sensors 24 06183 g003
Figure 4. Structure of HubNet. Images are fed through a backbone network to obtain feature maps, upon which word and character detection is performed. Following character detection, the feature maps are cropped to derive global features, and the cropped images are passed through a feature extractor for local feature acquisition. These global and local features undergo the feature cross-fusion module, after which the features are input into a recognition head to yield character recognition results, thus completing the task of wheel hub character detection and recognition.
Figure 4. Structure of HubNet. Images are fed through a backbone network to obtain feature maps, upon which word and character detection is performed. Following character detection, the feature maps are cropped to derive global features, and the cropped images are passed through a feature extractor for local feature acquisition. These global and local features undergo the feature cross-fusion module, after which the features are input into a recognition head to yield character recognition results, thus completing the task of wheel hub character detection and recognition.
Sensors 24 06183 g004
Figure 5. Two existing text detection and recognition frameworks. (a) Using local features only for text recognition. (b) Using global features only for text recognition.
Figure 5. Two existing text detection and recognition frameworks. (a) Using local features only for text recognition. (b) Using global features only for text recognition.
Sensors 24 06183 g005
Figure 6. Structure of the feature cross-fusion module.
Figure 6. Structure of the feature cross-fusion module.
Sensors 24 06183 g006
Figure 7. Structure of the recognition head.
Figure 7. Structure of the recognition head.
Sensors 24 06183 g007
Figure 8. Loss and accuracy curves during training.
Figure 8. Loss and accuracy curves during training.
Sensors 24 06183 g008
Figure 9. Visualization results of ablation studies on wheel hub text images. (a) Original Wheel Hub text images with location of text. (b) Proposed Model. (c) Model without FCF, using local features only. (d) Model without FCF, using global features only.
Figure 9. Visualization results of ablation studies on wheel hub text images. (a) Original Wheel Hub text images with location of text. (b) Proposed Model. (c) Model without FCF, using local features only. (d) Model without FCF, using global features only.
Sensors 24 06183 g009
Table 1. Dataset Statistics.
Table 1. Dataset Statistics.
ProjectQuantity
Image resolution512 × 512 pixels
Number of word instances in the dataset934
Number of character instances in the dataset2947
Table 2. Configuration of our experiment.
Table 2. Configuration of our experiment.
Software/HardwareVersion
CPUIntel® Core™ i7-14700KF
GPUNVIDIA GeForce RTX 3090
Operation SystemUbuntu 20.04
Python EnvironmentConda 24.12
Python versionPython3.8.19
Pytorch versionPytorch 1.10.1 + CUDA11.3
MMEngine version0.10.3
MMDetection version3.1.0
MMOCR version1.0.1
Table 3. Analysis on backbone and feature extractor (FE).
Table 3. Analysis on backbone and feature extractor (FE).
BackboneFEDetectionRecognition
Presicion Recall F1 Precision Recall F1
EfficientNetResNet18D90.789.590.180.873.476.9
EfficientNetResNet1890.789.590.182.073.677.6
ResNet50ResNet18D88.180.984.381.372.276.5
ResNet50ResNet1888.180.984.380.070.875.1
ResNet50DResNet18D92.490.891.685.578.782.0
ResNet50DResNet1892.490.891.686.579.482.8
The bold numbers represent the best experimental results.
Table 4. Ablation study on the proposed feature cross-fusion module.
Table 4. Ablation study on the proposed feature cross-fusion module.
FeaturesRecognition
Precision Recall F1
Global only70.964.467.5
Local only83.475.279.0
Add83.975.179.3
Concat84.877.681.0
Feature cross-fusion86.579.482.8
The bold numbers represent the best experimental results.
Table 5. Quantitative comparison of current methods.
Table 5. Quantitative comparison of current methods.
ModelVenueDetectionRecognition
Precision Recall F1 Precision Recall F1
Mask TextSpotter v3ECCV’202082.480.881.633.130.531.7
ABCNetCVPR’202056.557.356.912.410.911.6
SwinTextSpotterCVPR’202291.490.590.974.169.371.6
SPTS v2TPAMI’202391.291.891.579.272.775.8
Ours92.490.891.686.579.482.8
The bold numbers represent the best experimental results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, Y.; Meng, C. HubNet: An E2E Model for Wheel Hub Text Detection and Recognition Using Global and Local Features. Sensors 2024, 24, 6183. https://doi.org/10.3390/s24196183

AMA Style

Zeng Y, Meng C. HubNet: An E2E Model for Wheel Hub Text Detection and Recognition Using Global and Local Features. Sensors. 2024; 24(19):6183. https://doi.org/10.3390/s24196183

Chicago/Turabian Style

Zeng, Yue, and Cai Meng. 2024. "HubNet: An E2E Model for Wheel Hub Text Detection and Recognition Using Global and Local Features" Sensors 24, no. 19: 6183. https://doi.org/10.3390/s24196183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop