Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer
Abstract
:1. Introduction
- (1)
- In the first stage, we construct a source-free knowledge transfer module. It trains a generator to produce “source-like samples” using only the source domain model and the target domain model trained without labeled target domain data as supervision. Importantly, this process does not involve accessing source domain data. The “source-like samples” exhibit a style matching the source domain and content matching the target domain.
- (2)
- In the second stage, we employ a progressive joint training strategy to gradually train an adaptive model by inputting different proportions of “source-like samples” and target domain data. This process can be viewed as a means of data augmentation. Compared to directly applying target domain data to the source domain model, the “source-like samples” infused with source domain knowledge exhibit greater affinity to the model. Through iterative training, they effectively reduce domain discrepancies, thereby enhancing the model’s generalization performance.
- (1)
- We propose an unsupervised domain-adaptive vehicle re-identification method based on source-free knowledge transfer. Without the need to access source domain data, we utilize domain discrepancy information inherent in the source domain model and the target domain model to constrain a generator in generating “source-like samples.” These samples serve as a means of data augmentation to assist in model training for vehicle re-identification tasks.
- (2)
- We introduce “source-like samples” and a progressive joint training strategy for the target domain. These “source-like samples” are adapted to the same style as the source domain model and matched in content to the target domain data. They serve as an intermediate bridge between the source domain model and the target domain data, alleviating domain discrepancies and thus enhancing model performance.
2. Method
2.1. Pre-Trained Source Model and Target Model
2.2. Source-Free Image Generation Module
2.2.1. Knowledge Distillation Loss
2.2.2. Channel-Level Relational Consistency Loss
2.3. Progressive Joint Training Strategy
3. Experiment
3.1. Experimental Environment Settings
3.2. Datasets Setting and Evaluation Index Level
3.3. Experimental Results and Analysis
3.3.1. Experimental Results and Analysis on VehicleID→VeRi776
3.3.2. Experimental Results and Analysis on VeRi776→VehicleID
3.4. Ablation Experiment
3.4.1. Validation of the Loss Function of the Source-Free Image Generation Module
3.4.2. Validation of the Effectiveness of “Source-like Samples”
3.4.3. Validation of Progressive Joint Training Strategy
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Huang, L.; Yang, Q.; Wu, J.; Huang, Y.; Wu, Q.; Xu, J. Generated data with sparse regularized multi-pseudo label for person re-identification. IEEE Signal Process. Lett. 2020, 27, 391–395. [Google Scholar] [CrossRef]
- Bai, Y.; Lou, Y.; Gao, F.; Wang, S.; Wu, Y.; Duan, L.Y. Group-sensitive triplet embedding for vehicle reidentification. IEEE Trans. Multimed. 2018, 20, 2385–2399. [Google Scholar] [CrossRef]
- Zhao, Y.; Shen, C.; Wang, H.; Chen, S. Structural analysis of attributes for vehicle re-identification and retrieval. IEEE Trans. Intell. Transp. Syst. 2019, 21, 723–734. [Google Scholar] [CrossRef]
- Guo, H.; Zhu, K.; Tang, M.; Wang, J. Two-level attention network with multi-grain ranking loss for vehicle re-identification. IEEE Trans. Image Process. 2019, 28, 4328–4338. [Google Scholar] [CrossRef] [PubMed]
- Ma, H.; Li, X.; Yuan, X.; Zhao, C. Two-phase self-supervised pretraining for object re-identification. Knowl.-Based Syst. 2023, 261, 110220. [Google Scholar] [CrossRef]
- Song, L.; Wang, C.; Zhang, L.; Du, B.; Zhang, Q.; Huang, C.; Wang, X. Unsupervised domain adaptive re-identification: Theory and practice. Pattern Recognit. 2020, 102, 107173. [Google Scholar] [CrossRef]
- Yi, Z.; Zhang, H.; Tan, P.; Gong, M. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar]
- Lu, Z.; Lin, R.; He, Q.; Hu, H. Mask-aware pseudo label denoising for unsupervised vehicle re-identification. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4333–4347. [Google Scholar] [CrossRef]
- Wei, R.; Gu, J.; He, S.; Jiang, W. Transformer-Based Domain-Specific Representation for Unsupervised Domain Adaptive Vehicle Re-Identification. IEEE Trans. Intell. Transp. Syst. 2022, 24, 2935–2946. [Google Scholar] [CrossRef]
- Zhu, W.; Peng, B. Manifold-based aggregation clustering for unsupervised vehicle re-identification. Knowl.-Based Syst. 2022, 235, 107624. [Google Scholar] [CrossRef]
- Wang, Y.; Wei, Y.; Ma, R.; Wang, L.; Wang, C. Unsupervised vehicle re-identification based on mixed sample contrastive learning. Signal Image Video Process. 2022, 16, 2083–2091. [Google Scholar] [CrossRef]
- Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Venice, Italy, 22–29 October 2017; pp. 1125–1134. [Google Scholar]
- Peng, J.; Wang, Y.; Wang, H.; Zhang, Z.; Fu, X.; Wang, M. Unsupervised vehicle re-identification with progressive adaptation. arXiv 2020, arXiv:2006.11486. [Google Scholar]
- Zheng, A.; Sun, X.; Li, C.; Tang, J. Aware progressive clustering for unsupervised vehicle re-identification. IEEE Trans. Intell. Transp. Syst. 2021, 23, 11422–11435. [Google Scholar] [CrossRef]
- Wang, Y.; Peng, J.; Wang, H.; Wang, M. Progressive learning with multi-scale attention network for cross-domain vehicle re-identification. Sci. China Inf. Sci. 2022, 65, 160103. [Google Scholar] [CrossRef]
- Ge, Y.; Chen, D.; Li, H. Mutual mean-teaching: Pseudo label refinery for unsupervised domain adaptation on person re-identification. arXiv 2020, arXiv:2001.01526. [Google Scholar]
- Hermans, A.; Beyer, L.; Leibe, B. In defense of the triplet loss for person re-identification. arXiv 2017, arXiv:1703.07737. [Google Scholar]
- Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
- Huang, X.; Belongie, S. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1501–1510. [Google Scholar]
- Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
- Hoffman, J.; Tzeng, E.; Park, T.; Zhu, J.Y.; Isola, P.; Saenko, K.; Efros, A.; Darrell, T. Cycada: Cycle-consistent adversarial domain adaptation. In Proceedings of the International Conference on Machine Learning, Stockholm Sweden, 10–15 July 2018; pp. 1989–1998. [Google Scholar]
- Tung, F.; Mori, G. Similarity-preserving knowledge distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1365–1374. [Google Scholar]
- Li, Z.; Jiang, R.; Aarabi, P. Semantic relation preserving knowledge distillation for image-to-image translation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 648–663. [Google Scholar]
- Hou, Y.; Zheng, L. Visualizing adapted knowledge in domain transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13824–13833. [Google Scholar]
- Khan, K.; Rehman, S.U.; Aziz, K.; Fong, S.; Sarasvady, S. DBSCAN: Past, present and future. In Proceedings of the Fifth International Conference on the Applications of Digital Information and Web Technologies (ICADIWT 2014), IEEE, Chennai, India, 17–19 February 2014; pp. 232–238. [Google Scholar]
- Han, Y.; Huang, G.; Song, S.; Yang, L.; Wang, H.; Wang, Y. Dynamic neural networks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7436–7456. [Google Scholar] [CrossRef] [PubMed]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Liu, X.; Liu, W.; Mei, T.; Ma, H. Provid: Progressive and multimodal vehicle reidentification for large-scale urban surveillance. IEEE Trans. Multimed. 2017, 20, 645–658. [Google Scholar] [CrossRef]
- Liu, H.; Tian, Y.; Yang, Y.; Pang, L.; Huang, T. Deep relative distance learning: Tell the difference between similar vehicles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2167–2175. [Google Scholar]
- Fan, H.; Zheng, L.; Yan, C.; Yang, Y. Unsupervised person re-identification: Clustering and fine-tuning. ACM Trans. Multimed. Comput. Commun. Appl. TOMM 2018, 14, 1–18. [Google Scholar] [CrossRef]
- Deng, W.; Zheng, L.; Ye, Q.; Kang, G.; Yang, Y.; Jiao, J. Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 994–1003. [Google Scholar]
- Zhong, Z.; Zheng, L.; Li, S.; Yang, Y. Generalizing a person retrieval model hetero-and homogeneously. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 172–188. [Google Scholar]
- Zhong, Z.; Zheng, L.; Luo, Z.; Li, S.; Yang, Y. Invariance matters: Exemplar memory for domain adaptive person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 598–607. [Google Scholar]
Methods | VeRi776 | ||
---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | |
PUL [30] | 55.24 | 66.27 | 17.06 |
SPGAN [31] | 57.4 | 70.0 | 16.4 |
HHL [32] | 56.20 | 67.61 | 17.52 |
ECN [33] | 60.8 | 70.9 | 27.7 |
UDAP [6] | 73.9 | 81.5 | 35.8 |
PAL [13] | 68.17 | 79.91 | 42.04 |
Direct Transfer | 62.1 | 73.9 | 27.6 |
ours | 74.4 | 82.1 | 37.9 |
Methods | Test Size = 800 | Test Size = 1600 | ||||
---|---|---|---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | Rank-1 (%) | Rank-5 (%) | mAP (%) | |
PUL [30] | 40.03 | 46.03 | 43.9 | 33.83 | 49.72 | 37.68 |
CycleGAN [20] | 37.29 | 58.56 | 42.32 | 30.00 | 49.96 | 34.92 |
PAL [13] | 50.25 | 64.91 | 53.50 | 44.25 | 60.95 | 48.05 |
Direct | 39.56 | 56.03 | 43.01 | 35.01 | 50.84 | 39.17 |
Transfer ours | 52.76 | 67.29 | 58.33 | 47.65 | 63.83 | 53.72 |
Methods | Test Size = 2400 | Test Size = 3200 | ||||
---|---|---|---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | Rank-1 (%) | Rank-5 (%) | mAP (%) | |
PUL [30] | 30.90 | 47.18 | 34.71 | 28.86 | 43.41 | 32.44 |
CycleGAN [20] | 27.15 | 46.52 | 31.86 | 24.83 | 42.17 | 29.17 |
PAL [13] | 41.08 | 59.12 | 45.14 | 38.19 | 55.32 | 42.13 |
Direct | 31.05 | 48.52 | 34.72 | 28.12 | 42.98 | 31.99 |
Transfer ours | 43.87 | 62.43 | 50.42 | 41.77 | 60.42 | 47.29 |
Loss Function | VeRi776 | ||
---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | |
67.6 | 78.6 | 32.1 | |
70.2 | 80.3 | 34.4 | |
+ | 74.4 | 82.1 | 37.9 |
Type | VeRi776 | ||
---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | |
Supervised Learning | 91.4 | 96.2 | 70.5 |
Direct Transfer | 62.1 | 73.9 | 27.6 |
Source-Like Samples | 72.5 | 81.5 | 35.1 |
Joint Training | 74.4 | 82.1 | 37.9 |
Feed Ratio | VeRi776 | ||
---|---|---|---|
Rank-1 (%) | Rank-5 (%) | mAP (%) | |
1:5 | 67.4 | 76.5 | 31.3 |
1:4 | 69.5 | 78.9 | 33.2 |
1:3 | 72.6 | 80.9 | 36.1 |
1:2 | 74.4 | 82.1 | 37.9 |
1:1 | 73.6 | 81.7 | 37.4 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Song, Z.; Li, D.; Chen, Z.; Yang, W. Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer. Appl. Sci. 2023, 13, 11013. https://doi.org/10.3390/app131911013
Song Z, Li D, Chen Z, Yang W. Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer. Applied Sciences. 2023; 13(19):11013. https://doi.org/10.3390/app131911013
Chicago/Turabian StyleSong, Zhigang, Daisong Li, Zhongyou Chen, and Wenqin Yang. 2023. "Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer" Applied Sciences 13, no. 19: 11013. https://doi.org/10.3390/app131911013
APA StyleSong, Z., Li, D., Chen, Z., & Yang, W. (2023). Unsupervised Vehicle Re-Identification Method Based on Source-Free Knowledge Transfer. Applied Sciences, 13(19), 11013. https://doi.org/10.3390/app131911013