A Lightweight Convolutional Neural Network (CNN) Architecture for Traffic Sign Recognition in Urban Road Networks
Abstract
:1. Introduction
- An efficient CNN-based framework has been proposed for accurate traffic sign recognition, which addresses the computational constraints of existing systems. By optimizing the number of convolutional layers and filter size, the proposed framework reduces the number of trainable parameters and computational overhead. A systematic approach such as a grid search was employed to optimize the number of convolutional layers and filter sizes, resulting in the best performance with the fewest number of trainable parameters. The framework was designed to balance the complexity and computational efficiency of the model while maintaining high accuracy.
- The proposed framework addresses to the computational limitations of current TSR systems, making it more accessible and affordable for deployment in automotive systems with limited computational resources. Its minimal computational requirements and inference time enable it to operate efficiently and improve road safety, thereby reducing the incidence of accidents.
- To tackle the challenges posed by variations in traffic sign appearance, such as changes in lighting conditions, scale, perspective, viewing angle, blur, and shadow, as well as intra-category appearance variations and low inter-category variations, a variety of convolutional approaches have been explored. Furthermore, extensive research has been conducted on optimization strategies and design principles for Convolutional Neural Networks (CNNs) to design efficient TSRs.
- The proposed framework has been evaluated using two benchmark real-world datasets, GTSRB and BelgiumTS, which has provided confidence in its ability to operate effectively in practical settings. These evaluations provide a comprehensive understanding of the proposed system capabilities, enabling confidence in its potential for real-world deployment.
- A comprehensive comparative analysis with several state-of-the-art models is presented, such as GoogleNet, AlexNet, VGG16, VGG19, MobileNetv2, and ResNetv2. The evaluation parameters for comparison include accuracy, precision, recall, and f1-score. The analysis was conducted in detail and provides valuable insights into the performance of the proposed framework.
- Overall, the contribution of this research paper is significant as it helps to improve road safety by developing a practical and efficient TSR system that can operate in low-resource settings.
2. Related Work
3. Methodology
3.1. Dataset
3.1.1. GTSRB [9]
3.1.2. BelgiumTS [10]
3.2. Data Pre-Preocessing
3.2.1. Image Re-Scaling
3.2.2. Data Normalization
3.2.3. Data Augmentation
3.3. Proposed Methodology
3.4. Loss Function and Optimization Algorithm
4. Experimentation and Result
4.1. Experimental Setup
4.2. Key Performance Indicator
4.3. Experimental Results
4.3.1. Performance Results on GTSRB
4.3.2. Performance Results on BelgiumTS
5. Discussion
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Organization, W.H. Global Status Report on Road Safety 2018; World Health Organization: Geneva, Switzerland, 2018. [Google Scholar]
- Tao, X.; Gong, Y.; Shi, W.; Cheng, D. Object detection with class aware region proposal network and focused attention objective. Pattern Recognit. Lett. 2020, 130, 353–361. [Google Scholar] [CrossRef]
- Dewi, C.; Chen, R.C. Random forest and support vector machine on features selection for regression analysis. Int. J. Innov. Comput. Inf. Control 2019, 15, 2027–2037. [Google Scholar]
- Li, J.; Wang, Z. Real-time traffic sign recognition based on efficient CNNs in the wild. IEEE Trans. Intell. Transp. Syst. 2018, 20, 975–984. [Google Scholar] [CrossRef]
- Shahud, M.; Bajracharya, J.; Praneetpolgrang, P.; Petcharee, S. Thai traffic sign detection and recognition using convolutional neural networks. In Proceedings of the 2018 22nd International Computer Science and Engineering Conference (ICSEC), Chiang Mai, Thailand, 21–24 November 2018; pp. 1–5. [Google Scholar]
- Sanyal, B.; Mohapatra, R.K.; Dash, R. Traffic sign recognition: A survey. In Proceedings of the 2020 International Conference on Artificial Intelligence and Signal Processing (AISP), Amaravati, India, 10–12 January 2020; pp. 1–6. [Google Scholar]
- Tai, S.K.; Dewi, C.; Chen, R.C.; Liu, Y.T.; Jiang, X.; Yu, H. Deep learning for traffic sign recognition based on spatial pyramid pooling with scale analysis. Appl. Sci. 2020, 10, 6997. [Google Scholar] [CrossRef]
- Wang, X.; Guo, J.; Yi, J.; Song, Y.; Xu, J.; Yan, W.; Fu, X. Real-time and efficient multi-scale traffic sign detection method for driverless cars. Sensors 2022, 22, 6930. [Google Scholar] [CrossRef] [PubMed]
- Stallkamp, J.; Schlipsing, M.; Salmen, J.; Igel, C. The German traffic sign recognition benchmark: A multi-class classification competition. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 1453–1460. [Google Scholar]
- Mathias, M.; Timofte, R.; Benenson, R.; Van Gool, L. Traffic sign recognition—How far are we from the solution? In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; pp. 1–8. [Google Scholar]
- Youssef, A.; Albani, D.; Nardi, D.; Bloisi, D.D. Fast traffic sign recognition using color segmentation and deep convolutional networks. In Proceedings of the International Conference on Advanced Concepts for Intelligent Vision Systems, Lecce, Italy, 24–27 October 2016; Springer: Cham, Switzerland, 2016; pp. 205–216. [Google Scholar]
- Lai, Y.; Wang, N.; Yang, Y.; Lin, L. Traffic Signs Recognition and Classification based on Deep Feature Learning. In Proceedings of the ICPRAM, Funchal, Portugal, 16–18 January 2018; pp. 622–629. [Google Scholar]
- Wang, G.; Ren, G.; Wu, Z.; Zhao, Y.; Jiang, L. A robust, coarse-to-fine traffic sign detection method. In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; pp. 1–5. [Google Scholar]
- Chen, Y.; Xu, W.; Zuo, J.; Yang, K. The fire recognition algorithm using dynamic feature fusion and IV-SVM classifier. Clust. Comput. 2019, 22, 7665–7675. [Google Scholar] [CrossRef]
- Chen, Y.; Xiong, J.; Xu, W.; Zuo, J. A novel online incremental and decremental learning algorithm based on variable support vector machine. Clust. Comput. 2019, 22, 7435–7445. [Google Scholar] [CrossRef]
- Zeng, Y.; Xu, X.; Shen, D.; Fang, Y.; Xiao, Z. Traffic sign recognition using kernel extreme learning machines with deep perceptual features. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1647–1653. [Google Scholar] [CrossRef]
- Hechri, A.; Hmida, R.; Mtibaa, A. Robust road lanes and traffic signs recognition for driver assistance system. Int. J. Comput. Sci. Eng. 2015, 10, 202–209. [Google Scholar] [CrossRef]
- Hmida, R.; Ben Abdelali, A.; Mtibaa, A. Hardware implementation and validation of a traffic road sign detection and identification system. J. Real-Time Image Process. 2018, 15, 13–30. [Google Scholar] [CrossRef]
- Haque, W.A.; Arefin, S.; Shihavuddin, A.; Hasan, M.A. DeepThin: A novel lightweight CNN architecture for traffic sign recognition without GPU requirements. Expert Syst. Appl. 2021, 168, 114481. [Google Scholar] [CrossRef]
- Wong, A.; Shafiee, M.J.; Jules, M.S. MicronNet: A highly compact deep convolutional neural network architecture for real-time embedded traffic sign classification. IEEE Access 2018, 6, 59803–59810. [Google Scholar] [CrossRef]
- Zeng, Y.; Xu, X.; Fang, Y.; Zhao, K. Traffic sign recognition using deep convolutional networks and extreme learning machine. In Proceedings of the International Conference on Intelligent Science and Big Data Engineering, Suzhou, China, 14–16 June 2015; Springer: Cham, Switzerland, 2015; pp. 272–280. [Google Scholar]
- Cireşan, D.; Meier, U.; Masci, J.; Schmidhuber, J. A committee of neural networks for traffic sign classification. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 1918–1921. [Google Scholar]
- CireAan, D.; Meier, U.; Masci, J.; Schmidhuber, J. Multi-column deep neural network for traffic sign classification. Neural Netw. 2012, 32, 333–338. [Google Scholar] [CrossRef] [Green Version]
- Jin, J.; Fu, K.; Zhang, C. Traffic sign recognition with hinge loss trained convolutional neural networks. IEEE Trans. Intell. Transp. Syst. 2014, 15, 1991–2000. [Google Scholar] [CrossRef]
- Sun, W.; Du, H.; Nie, S.; He, X. Traffic sign recognition method integrating multi-layer features and kernel extreme learning machine classifier. Comput. Mater. Contin. 2019, 60, 147–161. [Google Scholar] [CrossRef]
- Wang, C. Research and application of traffic sign detection and recognition based on deep learning. In Proceedings of the 2018 International Conference on Robots & Intelligent System (ICRIS), Changsha, China, 26–27 May 2018; pp. 150–152. [Google Scholar]
- Ellahyani, A.; El Ansari, M.; Lahmyed, R.; Trémeau, A. Traffic sign recognition method for intelligent vehicles. JOSA A 2018, 35, 1907–1914. [Google Scholar] [CrossRef] [PubMed]
- Lim, K.; Hong, Y.; Choi, Y.; Byun, H. Real-time traffic sign recognition based on a general purpose GPU and deep-learning. PLoS ONE 2017, 12, e0173317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Qin, Z.; Yan, W.Q. Traffic-sign recognition using deep learning. In Proceedings of the International Symposium on Geometry and Vision; Springer: Cham, Switzerland, 2021; pp. 13–25. [Google Scholar]
- Temel, D.; Chen, M.H.; AlRegib, G. Traffic sign detection under challenging conditions: A deeper look into performance variations and spectral characteristics. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3663–3673. [Google Scholar] [CrossRef] [Green Version]
- Kamal, U.; Tonmoy, T.I.; Das, S.; Hasan, M.K. Automatic traffic sign detection and recognition using SegU-Net and a modified Tversky loss function with L1-constraint. IEEE Trans. Intell. Transp. Syst. 2019, 21, 1467–1479. [Google Scholar] [CrossRef]
- Xie, K.; Ge, S.; Ye, Q.; Luo, Z. Traffic sign recognition based on attribute-refinement cascaded convolutional neural networks. In Proceedings of the Advances in Multimedia Information Processing-PCM 2016: 17th Pacific-Rim Conference on Multimedia, Xi’an, China, 15–16 September 2016; Proceedings, Part I. Springer: Cham, Switzerland, 2016; pp. 201–210. [Google Scholar]
- Guo, J.; You, R.; Huang, L. Mixed vertical-and-horizontal-text traffic sign detection and recognition for street-level scene. IEEE Access 2020, 8, 69413–69425. [Google Scholar] [CrossRef]
- Luo, H.; Yang, Y.; Tong, B.; Wu, F.; Fan, B. Traffic sign recognition using a multi-task convolutional neural network. IEEE Trans. Intell. Transp. Syst. 2017, 19, 1100–1111. [Google Scholar] [CrossRef]
- Xing, J.; Yan, W.Q. Traffic sign recognition using guided image filtering. In Proceedings of the International Symposium on Geometry and Vision, Auckland, New Zealand, 28–29 January 2021; Springer: Cham, Switzerland, 2021; pp. 85–99. [Google Scholar]
- Mao, X.; Hijazi, S.; Casas, R.; Kaul, P.; Kumar, R.; Rowen, C. Hierarchical CNN for traffic sign recognition. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 19–22 June 2016; pp. 130–135. [Google Scholar]
- Liu, Z.; Qi, M.; Shen, C.; Fang, Y.; Zhao, X. Cascade saccade machine learning network with hierarchical classes for traffic sign detection. Sustain. Cities Soc. 2021, 67, 102700. [Google Scholar] [CrossRef]
- Sichkar, V.N. Real time detection and classification of traffic signs based on YOLO version 3 algorithm. J. Sci. Tech. Inf. Technol. Mech. Opt. 2020, 127, 418–424. [Google Scholar] [CrossRef]
- Megalingam, R.K.; Thanigundala, K.; Musani, S.R.; Nidamanuru, H.; Gadde, L. Indian traffic sign detection and recognition using deep learning. Int. J. Transp. Sci. Technol. 2022. [Google Scholar] [CrossRef]
- Bangquan, X.; Xiong, W.X. Real-time embedded traffic sign recognition using efficient convolutional neural network. IEEE Access 2019, 7, 53330–53346. [Google Scholar] [CrossRef]
- Valeja, Y.; Pathare, S.; Patel, D.; Pawar, M. Traffic Sign Detection using Clara and Yolo in Python. In Proceedings of the 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 19–20 March 2021; Volume 1, pp. 367–371. [Google Scholar]
- Zaklouta, F.; Stanciulescu, B. Real-time traffic-sign recognition using tree classifiers. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1507–1514. [Google Scholar] [CrossRef]
- Lopez-Montiel, M.; Orozco-Rosas, U.; Sanchez-Adame, M.; Picos, K.; Ross, O.H.M. Evaluation method of deep learning-based embedded systems for traffic sign detection. IEEE Access 2021, 9, 101217–101238. [Google Scholar] [CrossRef]
- Avramović, A.; Sluga, D.; Tabernik, D.; Skočaj, D.; Stojnić, V.; Ilc, N. Neural-network-based traffic sign detection and recognition in high-definition images using region focusing and parallelization. IEEE Access 2020, 8, 189855–189868. [Google Scholar] [CrossRef]
- Zhou, S.; Qiu, J. Enhanced SSD with interactive multi-scale attention features for object detection. Multimed. Tools Appl. 2021, 80, 11539–11556. [Google Scholar] [CrossRef]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Srivastava, R.K.; Greff, K.; Schmidhuber, J. Highway networks. arXiv 2015, arXiv:1505.00387. [Google Scholar]
- He, K.; Sun, J. Convolutional neural networks at constrained time cost. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5353–5360. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
- Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
- Zhang, J.; Xie, Z.; Sun, J.; Zou, X.; Wang, J. A cascaded R-CNN with multiscale attention and imbalanced samples for traffic sign detection. IEEE Access 2020, 8, 29742–29754. [Google Scholar] [CrossRef]
- Kuros, S.; Kryjak, T. Traffic Sign Classification Using Deep and Quantum Neural Networks. In Proceedings of the Computer Vision and Graphics: Proceedings of the International Conference on Computer Vision and Graphics ICCVG 2022; Springer: Cham, Switzerland, 2023; pp. 43–55. [Google Scholar]
- Xu, X.; Jin, J.; Zhang, S.; Zhang, L.; Pu, S.; Chen, Z. Smart data driven traffic sign detection method based on adaptive color threshold and shape symmetry. Future Gener. Comput. Syst. 2019, 94, 381–391. [Google Scholar] [CrossRef]
- Li, C.; Chen, Z.; Wu, Q.J.; Liu, C. Deep saliency detection via channel-wise hierarchical feature responses. Neurocomputing 2018, 322, 80–92. [Google Scholar] [CrossRef]
Layer Name | Output Size | Kernel Size | Strides | Activation | Number of Layers |
---|---|---|---|---|---|
Input | 100 × 100 × 3 | – | – | – | 1 |
Conv2d | 98 × 98 × 16 | 3 × 3 | 1 | ReLU | 1 |
Conv2d | 98 × 98 × 32 | 3 × 3 | 1 | ReLU | 2 |
MaxPooling2D | 49 × 49 × 32 | 2 × 2 | 2 | – | 1 |
Conv2d | 49 × 49 × 32 | 3 × 3 | 1 | ReLU | 1 |
Conv2d | 49 × 49 × 64 | 5 × 5 | 1 | ReLU | 2 |
Conv2d | 49 × 49 × 64 | 3 × 3 | 1 | ReLU | 1 |
MaxPooling2D | 24 × 24 × 64 | 2 × 2 | 2 | – | 1 |
Batch Normalization | 24 × 24 × 64 | – | – | – | 1 |
Conv2d | 24 × 24 × 64 | 3 × 3 | 1 | ReLU | 1 |
Conv2d | 24 × 24 × 128 | 5 × 5 | 1 | ReLU | 4 |
Conv2d | 24 × 24 × 256 | 3 × 3 | 1 | ReLU | 2 |
Global Average Pooling 2D | 24 × 24 × 128 | – | – | – | 1 |
Flatten | 256 | – | – | – | 1 |
Dense1 | 1024 | – | – | ReLU | 1 |
Dense2 | Number of Classes | – | – | Softmax | 1 |
Image Resolution Size | Time (in Seconds) | Resources Usage (in MByte) | Accuracy (%) | F1-Score (%) |
---|---|---|---|---|
50 × 50 | 578.0 | 9.2 | 96.97 | 96.97 |
80 × 80 | 1001.1 | 17.1 | 97.56 | 97.51 |
100 × 100 | 1264.2 | 20.7 | 98.41 | 98.42 |
Method | Accuracy (%) | Precison (%) | Recall (%) | F1-Score (%) | Parameters (in Millions) | Training Time (in Seconds) | Response Time (in millieconds) | Resources Usage (in MByte) |
---|---|---|---|---|---|---|---|---|
GoogleNet | 98.31 | 98.37 | 98.31 | 98.27 | 3.70 | 2983.7 | 94.80 | 22.3 |
AlexNet | 97.08 | 97.17 | 97.08 | 97.07 | 15.78 | 1638.5 | 73.05 | 28.0 |
VGG19 | 94.44 | 94.82 | 94.44 | 94.43 | 22.40 | 2454.8 | 90.68 | 23.7 |
VGG16 | 97.60 | 97.67 | 97.60 | 97.59 | 22.40 | 2442.9 | 96.21 | 24.1 |
MobileNetV2 | 97.30 | 97.45 | 97.30 | 97.30 | 12.77 | 1319.7 | 100.17 | 23.0 |
ResNet50V2 | 96.97 | 97.09 | 96.97 | 96.92 | 40.37 | 1974.9 | 105.93 | 26.5 |
Ours | 98.41 | 98.51 | 98.41 | 98.42 | 2.61 | 1264.2 | 74.34 | 20.7 |
Method | Precison (%) | Recall (%) | Missing Rate (%) | F1-Score (%) | Parameters (in Millions) |
---|---|---|---|---|---|
Li J [4] | 84.50 | 97.81 | 2.19 | 90.67 | 2.92 |
Kamal U [31] | 95.29 | 89.01 | 10.99 | 92.04 | 5 |
Faster R-CNN [50] | 96.10 | 86.30 | 13.70 | 90.94 | 16 |
Mask R-CNN [51] | 97.10 | 86.90 | 13.10 | 91.72 | 18 |
Cascaded R-CNN [52] | 96.80 | 88.60 | 11.40 | 92.52 | 27 |
Multiscale Cascaded R-CNN [53] | 98.30 | 85.00 | 15.00 | 91.71 | 55 |
Deep QNN [54] | 94.40 | 94.94 | 5.06 | 94.46 | AlexNet: 61.1 VGG16: 138.3 ResNet-50: 25.6 QP: 6 qubits QCNN: 10 qubits |
Xu X [55] | 93.96 | 95.27 | 4.73 | 94.61 | N/A |
Li C [56] | 90.70 | 81.7 | 18.3 | 86 | 4.5 |
Ours | 98.51 | 98.41 | 1.59 | 98.41 | 2.61 |
Method | Accuracy (%) | Precison (%) | Recall (%) | F1-Score (%) | Parameters (in Millions) |
---|---|---|---|---|---|
GoogleNet | 84.20 | 89.82 | 84.20 | 85.46 | 3.71 |
AlexNet | 70.71 | 80.77 | 70.71 | 72.02 | 15.79 |
VGG19 | 84.08 | 88.47 | 84.08 | 84.74 | 22.41 |
VGG16 | 69.12 | 79.09 | 69.12 | 69.82 | 22.41 |
MobileNetV2 | 36.98 | 71.89 | 36.98 | 35.75 | 12.77 |
ResNet50V2 | 12.46 | 40.15 | 12.46 | 12.86 | 40.37 |
Ours | 92.06 | 94.60 | 92.06 | 92.54 | 2.63 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Khan, M.A.; Park, H.; Chae, J. A Lightweight Convolutional Neural Network (CNN) Architecture for Traffic Sign Recognition in Urban Road Networks. Electronics 2023, 12, 1802. https://doi.org/10.3390/electronics12081802
Khan MA, Park H, Chae J. A Lightweight Convolutional Neural Network (CNN) Architecture for Traffic Sign Recognition in Urban Road Networks. Electronics. 2023; 12(8):1802. https://doi.org/10.3390/electronics12081802
Chicago/Turabian StyleKhan, Muneeb A., Heemin Park, and Jinseok Chae. 2023. "A Lightweight Convolutional Neural Network (CNN) Architecture for Traffic Sign Recognition in Urban Road Networks" Electronics 12, no. 8: 1802. https://doi.org/10.3390/electronics12081802
APA StyleKhan, M. A., Park, H., & Chae, J. (2023). A Lightweight Convolutional Neural Network (CNN) Architecture for Traffic Sign Recognition in Urban Road Networks. Electronics, 12(8), 1802. https://doi.org/10.3390/electronics12081802