Ship Infrared Automatic Target Recognition Based on Bipartite Graph Recommendation: A Model-Matching Method
Abstract
:1. Introduction
- We propose a novel SIATR model recommendation method, distinguished by its ability to adaptively match the optimal model based on sample attributes, which enhances adaptability to complex scenarios and improves overall performance;
- We have created a measure of SIATR model feature learning credibility. This measure, in combination with traditional accuracy metrics, provides a more comprehensive assessment of model usability;
- During the experimental validation phase, we analyze both recognition accuracy and feature learning reliability, as well as the relationship between the model resource consumption of the recommendation system.
2. Materials and Methods
2.1. Dataset
2.2. Proposed Approach
2.2.1. Preliminaries
2.2.2. Knowledge Construction
- For the i-th input image, the mask image is obtained using the method in Figure 2, and the attribute partition interval of the image is obtained using the following formula:
- Here, is used to determine the specific attribute subclass corresponding to the sample.
- Let be the predicted output formula of candidate ; calculate the predicted scores for image before and after masking using (2) and (3):
- Here, represents the predicted categories. Record the predicted scores in vector form using (4) and (5):
- Based on this, adopt a reward–penalty approach to obtain the weights and for the model with respect to the sample (to be separately introduced later).
- Construct the model–attribute weight matrix for image . Let the weight of the true attribute of the image be equal to and , and the weight of non-corresponding attributes be equal to 0, i.e., assigned according to the following equation:
- Combine the corresponding to images and compute the average using Equation (7) to obtain the final model–attribute weight matrix .
- Obtain each candidate model’s class prediction score matrix for the original image . The specific form of is shown in Equation (8):
- To simplify the data and focus on the interested categories, extract the sub-matrix under the category from , where represents the true category of image .
- Normalize the elements of using the softmax function, and denote the updated element as .
- Using the precision-optimized reward–penalty function to calculate the weights , such that the weights for correctly predicted models are positive and the weights for incorrectly predicted models are negative. In the following equation, represents the accuracy-optimized penalty factor:
- Obtain each candidate model’s class prediction score matrix for the masked image . The specific form of is shown in Equation (11):
- Calculate the Euclidean distance between and to measure the similarity of the model before and after image masking, and denote the result as .
- Using Equation (9), normalize the elemental values of , and denote the result as .
- When the original image is predicted correctly, an excellent candidate model’s decision should be less sensitive to the background region of the target, resulting in a smaller and, thus, should be assigned a larger weight. Conversely, when the prediction is incorrect, negative weights should be assigned, and smaller values of should incur greater penalties. To achieve this, the absolute value is introduced into Equation (13) to design a reward–penalty function for calculating the weight , where represents the credibility penalty factor.
2.2.3. Model-Adaptive Recommendation
3. Experiments and Results
3.1. Evaluation Metrics
3.2. Experimental Settings
3.3. Results and Discusion
3.3.1. The Recommendation System Aims to Improve the Accuracy
3.3.2. The Recommendation System Aims to Improve the Credibility
3.3.3. The Recommendation System Aims to Improve the Accuracy and Credibility
3.3.4. The Recommendation System with Reduced Resource Consumption
4. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A
Algorithm A1. The proposed SIATR-BGR model-matching method. |
Part 1: Knowledge construction Input: ship IR images. Output: Model–attribute weight matrix . 01: for to do 02: Obtain attribute interval according to Equation (1). 03: Generate using the method shown in Figure 2. 04: for to do 05: Calculate model output scores: and using Equations (2) and (3). 06: Construct vectors and according to Equations (4) and (5). 07: end for Reward–penalty calculation of weights: 08: Construct matrix and according to Equation (8) and Equation (11), respectively. 09: Extract submatrix from . 10: Calculate using Equations (9) and (10) and . 11: Construct matrix using Equation (12). 12: Calculate using Equation (9), Equation (13), and . 13: Construct matrix using , , and Equation (6). 14: end for 15: Construct matrix using Equation (7). Part 2: Model-adaptive recommendation Input: Unknown IR ship image . Output: Recommended model: . 16: Obtain attribute interval according to Equation (1). 17: Extract submatrix from . 18: Calculate using Equation (14). 19: Obtain using Equation (15). |
References
- He, Y.; Deng, B.; Wang, H.; Cheng, L.; Zhou, K.; Cai, S.; Ciampa, F. Infrared machine vision and infrared thermography with deep learning: A review. Infrared Phys. Technol. 2021, 116, 103754. [Google Scholar] [CrossRef]
- Pawar, S.; Gandhe, S. SAR (Synthetic Aperture Radar) Image Study and Analysis for Object Recognition in Surveillance. IJISAE 2023, 11, 552–573. [Google Scholar]
- Wang, N.; Wang, Y.; Er, M.J. Review on deep learning techniques for marine object recognition: Architectures and algorithms. Control. Eng. Pract. 2022, 118, 104458. [Google Scholar] [CrossRef]
- Özertem, K.A. Key features for ATA/ATR database design in missile systems. In Proceedings of the Automatic Target Recognition XXVII, Anaheim, CA, USA, 10–11 April 2017; pp. 106–111. [Google Scholar]
- Alves, J.A.; Rowe, N.C. Recognition of Ship Types from an Infrared Image Using Moment Invariants and Neural Networks; NAVAL Postgraduate School: Monterey, CA, USA, 2001. [Google Scholar]
- Luo, Q.; Khoshgoftaar, T.M.; Folleco, A. Classification of ships in surveillance video. In Proceedings of the 2006 IEEE International Conference on Information Reuse & Integration, Waikoloa, HI, USA, 16–18 September 2006; pp. 432–437. [Google Scholar]
- Li, H.; Wang, X. Automatic recognition of ship types from infrared images using support vector machines. In Proceedings of the 2008 International Conference on Computer Science and Software Engineering, Wuhan, China, 12–14 December 2008; pp. 483–486. [Google Scholar]
- Dargan, S.; Kumar, M.; Ayyagari, M.R.; Kumar, G. A survey of deep learning and its applications: A new paradigm to machine learning. Arch. Comput. Methods Eng. 2020, 27, 1071–1092. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Zhang, M.M.; Choi, J.; Daniilidis, K.; Wolf, M.T.; Kanan, C. VAIS: A dataset for recognizing maritime imagery in the visible and infrared spectrums. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 10–16. [Google Scholar]
- Karabayır, O.; Yücedağ, O.M.; Kartal, M.Z.; Serim, H.A. Convolutional neural networks-based ship target recognition using high resolution range profiles. In Proceedings of the 2017 18th International Radar Symposium (IRS), Prague, Czech Republic, 28–30 June 2017; pp. 1–9. [Google Scholar]
- Khellal, A.; Ma, H.; Fei, Q.J.S. Convolutional neural network based on extreme learning machine for maritime ships recognition in infrared images. Sensors 2018, 18, 1490. [Google Scholar] [CrossRef] [PubMed]
- Ren, Y.; Yang, J.; Guo, Z.; Zhang, Q.; Cao, H.J.E. Ship classification based on attention mechanism and multi-scale convolutional neural network for visible and infrared images. Electronics 2020, 9, 2022. [Google Scholar] [CrossRef]
- Huang, L.; Wang, F.; Zhang, Y.; Xu, Q. Fine-grained ship classification by combining CNN and swin transformer. Remote Sens. 2022, 14, 3087. [Google Scholar] [CrossRef]
- Zhang, Y.; Lei, Z.; Yu, H.; Zhuang, L.J.I.G.; Letters, R.S. Imbalanced high-resolution SAR ship recognition method based on a lightweight CNN. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
- Liu, B.; Xiao, Q.; Zhang, Y.; Ni, W.; Yang, Z.; Li, L. Intelligent recognition method of low-altitude squint optical ship target fused with simulation samples. Remote Sens. 2021, 13, 2697. [Google Scholar] [CrossRef]
- Sharifzadeh, F.; Akbarizadeh, G.; Seifi Kavian, Y. Ship classification in SAR images using a new hybrid CNN–MLP classifier. J. Indian Soc. Remote Sens. 2019, 47, 551–562. [Google Scholar] [CrossRef]
- Teixeira, E.; Araujo, B.; Costa, V.; Mafra, S.; Figueiredo, F. Literature Review on Ship Localization, Classification, and Detection Methods Based on Optical Sensors and Neural Networks. Sensors 2022, 22, 6879. [Google Scholar] [CrossRef] [PubMed]
- Li, J.; Yu, Z.; Yu, L.; Cheng, P.; Chen, J.; Chi, C. A Comprehensive Survey on SAR ATR in Deep-Learning Era. Remote Sens. 2023, 15, 1454. [Google Scholar] [CrossRef]
- Baorong, X.; Yu, Z.; Shuyi, F.; Xian, L.; Songfeng, D. Research of the infrared ship target recognition technology based on the complex background. In Proceedings of the 2018 IEEE/ACIS 17th International Conference on Computer and Information Science (ICIS), Singapore, 6–8 June 2018; pp. 850–854. [Google Scholar]
- Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
- Mohammed, A.; Kora, R. A comprehensive review on ensemble deep learning: Opportunities and challenges. J. King Saud Univ. -Comput. Inf. Sci. 2023, 35, 757–774. [Google Scholar] [CrossRef]
- Bai, X.; Wang, X.; Liu, X.; Liu, Q.; Song, J.; Sebe, N.; Kim, B. Explainable deep learning for efficient and robust pattern recognition: A survey of recent developments. Pattern Recognit. 2021, 120, 108102. [Google Scholar] [CrossRef]
- Tanimoto, S.L.; Itai, A.; Rodeh, M. Some matching problems for bipartite graphs. J. ACM 1978, 25, 517–525. [Google Scholar] [CrossRef]
- Westlake, S.T.; Volonakis, T.N.; Jackman, J.; James, D.B.; Sherriff, A. Deep learning for automatic target recognition with real and synthetic infrared maritime imagery. In Proceedings of the Artificial Intelligence and Machine Learning in Defense Applications II, Online, 21–25 September 2020; pp. 41–53. [Google Scholar]
- Ward, C.M.; Harguess, J.; Hilton, C. Ship classification from overhead imagery using synthetic data and domain adaptation. In Proceedings of the OCEANS 2018 MTS/IEEE Charleston, Charleston, SC, USA, 22–25 October 2018; pp. 1–5. [Google Scholar]
- Rizaev, I.G.; Achim, A. SynthWakeSAR: A synthetic sar dataset for deep learning classification of ships at sea. Remote Sens. 2022, 14, 3999. [Google Scholar] [CrossRef]
- Mathew, A.; Amudha, P.; Sivakumari, S. Deep learning techniques: An overview. Adv. Mach. Learn. Technol. Appl. Proc. AMLTA 2021, 1141, 599–608. [Google Scholar]
- Mumtaz, A.; Jabbar, A.; Mahmood, Z.; Nawaz, R.; Ahsan, Q. Saliency based algorithm for ship detection in infrared images. In Proceedings of the 2016 13th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 12–16 January 2016; pp. 167–172. [Google Scholar]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part I 13. pp. 818–833. [Google Scholar]
- Kortylewski, A.; He, J.; Liu, Q.; Yuille, A.L. Compositional convolutional neural networks: A deep architecture with innate robustness to partial occlusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 8940–8949. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
- Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. Mnasnet: Platform-aware neural architecture search for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2820–2828. [Google Scholar]
- Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
- Leon, F.; Floria, S.-A.; Bădică, C. Evaluating the effect of voting methods on ensemble-based classification. In Proceedings of the 2017 IEEE International Conference on INnovations in Intelligent Systems and Applications (INISTA), Gdynia, Poland, 3–5 July 2017; pp. 1–6. [Google Scholar]
Reference | Data Source | Method Description | Result |
---|---|---|---|
Khellal et al. [12] | IR | CNN + extreme learning machine | Compared to CNN based on backpropagation, the method has better accuracy and faster training speed. |
Ren et al. [13] | IR/ visible light | Multi-scale CNN + attention mechanism | The performance of the proposed method is superior to traditional machine learning methods and some CNN-based methods. |
Huang et al. [14] | Visible light | CNN + swin transformer | The parallel network structure can extract features more fully. It performs best among multiple image recognition models. |
Liu et al. [15] | Visible light | CNN + feature fusion mechanism + supplement sample size using simulation data | Compared to the CNN model as the backbone, the feature fusion mechanism and sample size complementation effectively and iteratively optimize the overall recognition performance. |
Sharifzadeh et al. [16] | SAR | CNN + multilayer perceptron | Compared with using CNN or multilayer perceptron alone for ship recognition, the hybrid approach can better extract features and achieve higher accuracy. |
Zhang et al. [17] | SAR | CNN + deep metric learning + gradually balanced sampling | The residual neural network embedded with the new mechanism performs more accurately on multiple public datasets. |
Elements | Attributes |
---|---|
Classes | Container freighter (CF), cruise ship (CS), warship (WS). |
Radiation factors | Time, air temperature, water temperature, maritime area, motion state, wind speed, and weather (sunny, rainy, foggy, etc.). |
Camera shooting factors | Distance, azimuth, zenith angle. |
Attributes | Attribute Division | |
---|---|---|
Classes | Interval | |
Time | Daytime | 8:00–17:00 |
Morning and vening | 5:00–8:00, 17:00–20:00 | |
Nighttime | 20:00–5:00 | |
Motion state | Moving | Speed greater than 0 |
Static | Speed equals 0 | |
Air temperature | Temperature_high | 0–10 °C |
Temperature_low | 10–20 °C | |
Zenith angle | Zenith_high | Angle greater than 45° |
Zenith_low | Angle less than 45° | |
Distance | Near | 3–6 km |
Far | 6–9 km |
CNN Model | Acc (%) | Prec (%) | Rec (%) | F1 (%) | EDMS |
---|---|---|---|---|---|
ResNet | 93.83 | 93.90 | 93.83 | 93.85 | 0.7455 |
SqueezeNet | 94.46 | 94.48 | 94.46 | 94.45 | 0.7386 |
DenseNet | 93.35 | 94.05 | 93.35 | 93.41 | 0.7382 |
MobileNet | 94.80 | 94.90 | 94.80 | 94.81 | 0.7507 |
MnasNet | 94.36 | 94.41 | 94.36 | 94.37 | 0.7405 |
ShuffleNet | 93.64 | 93.65 | 93.64 | 93.65 | 0.7049 |
Model/Recommendation System | Acc (%) | Prec (%) | Rec (%) | F1 (%) |
---|---|---|---|---|
SIATR-BGR () | 95.86 | 95.86 | 95.86 | 95.86 |
Hard Voting | 95.28 | 95.30 | 95.28 | 95.28 |
Soft Voting | 95.23 | 95.31 | 95.23 | 95.24 |
MobileNet | 94.80 | 94.90 | 94.80 | 94.81 |
Acc (%) | Prec (%) | Rec (%) | F1 (%) | EDMS | |
---|---|---|---|---|---|
SIATR-BGR (α = 0.4, β = 0.6. WU = 0.7) | 95.23 | 95.31 | 95.23 | 95.24 | 0.7578 |
Evaluation Metrics | DMA | DMA + DenseNet | DMA + ResNet | DMA + ShuffleNet | DMA + DenseNet + ResNet | DMA + DenseNet + ShuffleNet | DMA + ResNet+ ShuffleNet | All Models |
---|---|---|---|---|---|---|---|---|
Acc (%) | 95.62 | 95.71 | 95.81 | 95.76 | 95.81 | 95.76 | 95.86 | 95.86 |
Prec (%) | 95.62 | 95.72 | 95.81 | 95.76 | 95.81 | 95.76 | 95.86 | 95.86 |
Rec (%) | 95.62 | 95.71 | 95.81 | 95.76 | 95.81 | 95.76 | 95.86 | 95.86 |
F1 (%) | 95.62 | 95.71 | 95.81 | 95.75 | 95.81 | 95.75 | 95.86 | 95.86 |
Parameters (million) | 13.00 | 20.98 | 24.69 | 14.37 | 32.67 | 22.35 | 26.06 | 34.04 |
Memory size (MB) | 38.39 | 65.49 | 81.09 | 43.37 | 108.19 | 70.47 | 86.07 | 113.17 |
Evaluation Metrics | DMC | DMC + SqueezeNet | DMC + DenseNet | DMC + ShuffleNet | DMC + SqueezeNet + DenseNet | DMC + SqueezeNet + ShuffleNet | DMC + DenseNet + ShuffleNet | All Models |
---|---|---|---|---|---|---|---|---|
EDMS | 0.7657 | 0.7754 | 0.7737 | 0.7682 | 0.7781 | 0.7778 | 0.7778 | 0.7781 |
Parameters (million) | 23.45 | 24.69 | 31.43 | 24.82 | 32.67 | 26.06 | 32.8 | 34.04 |
Memory size (MB) | 78.30 | 81.09 | 105.4 | 83.28 | 108.19 | 86.07 | 110.38 | 113.17 |
Acc (%) | Prec (%) | Rec (%) | F1 (%) | EDMS | |
---|---|---|---|---|---|
SIATR-BGR (, , ) | 95.13 | 95.14 | 95.13 | 95.13 | 0.7552 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, H.; Liu, C.; Ma, J.; Sun, H. Ship Infrared Automatic Target Recognition Based on Bipartite Graph Recommendation: A Model-Matching Method. Mathematics 2024, 12, 168. https://doi.org/10.3390/math12010168
Zhang H, Liu C, Ma J, Sun H. Ship Infrared Automatic Target Recognition Based on Bipartite Graph Recommendation: A Model-Matching Method. Mathematics. 2024; 12(1):168. https://doi.org/10.3390/math12010168
Chicago/Turabian StyleZhang, Haoxiang, Chao Liu, Jianguang Ma, and Hui Sun. 2024. "Ship Infrared Automatic Target Recognition Based on Bipartite Graph Recommendation: A Model-Matching Method" Mathematics 12, no. 1: 168. https://doi.org/10.3390/math12010168
APA StyleZhang, H., Liu, C., Ma, J., & Sun, H. (2024). Ship Infrared Automatic Target Recognition Based on Bipartite Graph Recommendation: A Model-Matching Method. Mathematics, 12(1), 168. https://doi.org/10.3390/math12010168