An Optimized Class Incremental Learning Network with Dynamic Backbone Based on Sonar Images
Abstract
:1. Introduction
- Several representative CILNs designed for optical images are evaluated on the constructed sonar image dataset (SonarImage20). The result shows that all networks suffer from the catastrophic forgetting problem, confirming previous suspicions. The best-performing DER is selected as the baseline for subsequent improvements.
- An optimized backbone (OptResNet) based on ResNet18 is proposed to replace the backbone of the baseline. This modification improves the network’s feature extraction capabilities and recognition accuracy while significantly reducing the number of parameters.
- A series of new strategies are introduced to optimize the class incremental training strategies of the baseline, which enhances the stability and robustness of the network training. A novel CILN(DER-Sonar) that integrates the optimized backbone with the optimized class incremental training strategies is proposed, effectively addressing the catastrophic forgetting problem on the sonar image dataset.
2. Dataset
3. Methods
3.1. CIL Concept and Performance Evaluation for SonarImage20
3.2. Baseline
3.2.1. Backbone
3.2.2. Class Incremental Training Strategies
3.3. Optimized Class Incremental Learning Network (DER-Sonar)
3.3.1. Optimized Backbone
3.3.2. Optimized Class Incremental Training Strategies
Algorithm 1 Nearest Class Mean Classifier |
Input: // input image |
Require: // exemplar set |
Require: // weights of the network |
for do |
// mean of exemplars |
end for |
// nearest prototype |
Output: // prediction label |
4. Experiments
4.1. Implementation Details
4.2. Comparative Experiment
4.3. Ablation Experiment
4.3.1. Backbone
4.3.2. Class Incremental Training Strategies
5. Discussion
6. Conclusions
- Optimization for training time: although DER-Sonar introduces an additional training time overhead, we anticipate a significant reduction in this aspect with further refinement of the methods used, especially in evolving development environments.
- Dataset expansion: SonarImage20 provides a robust foundation, but expanding it with more diverse underwater scenarios can further test and refine the capabilities of DER-Sonar.
- Investigate transfer learning: investigating how DER sonar can benefit from transfer learning, especially from models trained on large optical image datasets, could be a promising direction.
- Address real-time challenges: since underwater target recognition often requires real-time processing, future iterations of DER-Sonar could focus on optimizing for real-time deployment in underwater vehicles.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Yu, X.; Xing, X.; Zheng, H.; Fu, X.; Huang, Y.; Ding, X. Man-Made Object Recognition from Underwater Optical Images Using Deep Learning and Transfer Learning. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1852–1856. [Google Scholar] [CrossRef]
- Neupane, D.; Seok, J. A Review on Deep Learning-Based Approaches for Automatic Sonar Target Recognition. Electronics 2020, 9, 1972. [Google Scholar] [CrossRef]
- Chen, R.; Chen, Y. Improved Convolutional Neural Network YOLOv5 for Underwater Target Detection Based on Autonomous Underwater Helicopter. J. Mar. Sci. Eng. 2023, 11, 989. [Google Scholar] [CrossRef]
- Masana, M.; Liu, X.; Twardowski, B.; Menta, M.; Bagdanov, A.D.; van de Weijer, J. Class-Incremental Learning: Survey and Performance Evaluation on Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 5513–5533. [Google Scholar] [CrossRef] [PubMed]
- Belouadah, E.; Popescu, A.; Kanellos, I. A comprehensive study of class incremental learning algorithms for visual tasks. Neural Netw. 2021, 135, 38–54. [Google Scholar] [CrossRef] [PubMed]
- Xu, H.; Yang, L.; Long, X. Underwater Sonar Image Classification with Small Samples Based on Parameter-based Transfer Learning and Deep Learning. In Proceedings of the 2022 Global Conference on Robotics, Artificial Intelligence and Information Technology (GCRAIT), Chicago, IL, USA, 30–31 July 2022; pp. 304–307. [Google Scholar] [CrossRef]
- Irfan, M.; Jiangbin, Z.; Iqbal, M.; Masood, Z.; Arif, M.H.; ul Hassan, S.R. Brain inspired lifelong learning model based on neural based learning classifier system for underwater data classification. Expert Syst. Appl. 2021, 186, 115798. [Google Scholar] [CrossRef]
- Li, Z.; Hoiem, D. Learning without Forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2935–2947. [Google Scholar] [CrossRef] [PubMed]
- Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. iCaRL: Incremental Classifier and Representation Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Douillard, A.; Cord, M.; Ollion, C.; Robert, T.; Valle, E. Podnet: Pooled outputs distillation for small-tasks incremental learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 86–102. [Google Scholar]
- Yan, S.; Xie, J.; He, X. DER: Dynamically Expandable Representation for Class Incremental Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 3014–3023. [Google Scholar]
- Wang, F.Y.; Zhou, D.W.; Ye, H.J.; Zhan, D.C. Foster: Feature boosting and compression for class-incremental learning. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 398–414. [Google Scholar]
- Zhou, D.W.; Wang, Q.W.; Ye, H.J.; Zhan, D.C. A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental Learning. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
- Kim, J.; Song, S.; Yu, S.C. Denoising auto-encoder based image enhancement for high resolution sonar image. In Proceedings of the 2017 IEEE Underwater Technology (UT), Busan, Republic of Korea, 21–24 February 2017; pp. 1–5. [Google Scholar] [CrossRef]
- Singh, D.; Valdenegro-Toro, M. The Marine Debris Dataset for Forward-Looking Sonar Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, QC, Canada, 10–17 October 2021; pp. 3741–3749. [Google Scholar]
- Preciado-Grijalva, A.; Wehbe, B.; Firvida, M.B.; Valdenegro-Toro, M. Self-Supervised Learning for Sonar Image Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 18–24 June 2022; pp. 1499–1508. [Google Scholar]
- Du, X.; Sun, Y.; Song, Y.; Sun, H.; Yang, L. A Comparative Study of Different CNN Models and Transfer Learning Effect for Underwater Object Classification in Side-Scan Sonar Images. Remote Sens. 2023, 15, 593. [Google Scholar] [CrossRef]
- Zhou, D.W.; Wang, Q.W.; Qi, Z.H.; Ye, H.J.; Zhan, D.C.; Liu, Z. Deep class-incremental learning: A survey. arXiv 2023, arXiv:2302.03648. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Wu, S.; Wang, G.; Tang, P.; Chen, F.; Shi, L. Convolution with even-sized kernels and symmetric padding. In Proceedings of the Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; Volume 32. [Google Scholar]
- Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
- Wu, Y.; He, K. Group Normalization. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
- Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
- Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
- Misra, D. Mish: A self regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
- Lim, S.; Kim, I.; Kim, T.; Kim, C.; Kim, S. Fast AutoAugment. In Proceedings of the Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; Volume 32. [Google Scholar]
- Cubuk, E.D.; Zoph, B.; Shlens, J.; Le, Q.V. Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 702–703. [Google Scholar]
- Müller, S.G.; Hutter, F. TrivialAugment: Tuning-Free Yet State-of-the-Art Data Augmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 774–782. [Google Scholar]
- He, J.; Zhu, F. Exemplar-Free Online Continual Learning. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 541–545. [Google Scholar] [CrossRef]
- Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019. [Google Scholar]
- Zhou, D.W.; Wang, F.Y.; Ye, H.J.; Zhan, D.C. PyCIL: A Python toolbox for class-incremental learning. Sci. China Inf. Sci. 2023, 66, 197101. [Google Scholar] [CrossRef]
CILN | Average Recognition Accuracy/% | Training Time/s |
---|---|---|
LwF | 67.13 | 1391 |
iCaRL | 79.34 | 1796 |
PodNet | 85.36 | 2817 |
DER(baseline) | 88.87 | 2412 |
FOSTER | 84.13 | 2665 |
MEMO | 88.01 | 1624 |
Ours_ LC | 96.18 | 4085 |
Ours_NCM | 96.30 | 4093 |
Upperbound | 99.60 | 5182 |
Backbone | Average Recognition Accuracy/% | Parameter Count/Million | Training Time/s |
---|---|---|---|
ResNet18 (baseline) | 88.87 | 11.821 | 2412 |
ResNet18 + SPConv | 91.06 | 10.515 | 3982 |
ResNet18 + SPConv + GN & Mish | 91.35 | 10.505 | 3716 |
ResNet18 + SPConv + GN & Mish + SENet | 91.68 | 10.505 | 3842 |
OptResNet (within all improvements) | 93.92 | 8.275 | 3792 |
Upperbound | 99.60 | 11.821 | 5182 |
Class Incremental Training Strategy | Average Recognition Accuracy/% | Training Time/s |
---|---|---|
Baseline | 88.87 | 2412 |
OptResNet | 93.92 | 3792 |
OptResNet + dynamic loss | 94.94 | 4070 |
OptResNet + dynamic loss + data augment | 96.18 | 4085 |
OptResNet + dynamic loss + data augment + NCM | 96.30 | 4093 |
Upperbound | 99.60 | 5182 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chen, X.; Liang, H. An Optimized Class Incremental Learning Network with Dynamic Backbone Based on Sonar Images. J. Mar. Sci. Eng. 2023, 11, 1781. https://doi.org/10.3390/jmse11091781
Chen X, Liang H. An Optimized Class Incremental Learning Network with Dynamic Backbone Based on Sonar Images. Journal of Marine Science and Engineering. 2023; 11(9):1781. https://doi.org/10.3390/jmse11091781
Chicago/Turabian StyleChen, Xinzhe, and Hong Liang. 2023. "An Optimized Class Incremental Learning Network with Dynamic Backbone Based on Sonar Images" Journal of Marine Science and Engineering 11, no. 9: 1781. https://doi.org/10.3390/jmse11091781
APA StyleChen, X., & Liang, H. (2023). An Optimized Class Incremental Learning Network with Dynamic Backbone Based on Sonar Images. Journal of Marine Science and Engineering, 11(9), 1781. https://doi.org/10.3390/jmse11091781