Next Article in Journal
New Quantum Private Comparison Using Bell States
Previous Article in Journal
Multipartite Correlations in Parikh–Wilczek Non-Thermal Spectrum
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Modality Person Re-Identification Method with Joint-Modality Generation and Feature Enhancement

1
School of Information and Cyber Security, People’s Public Security University of China, Beijing 100038, China
2
Key Laboratory of Security Prevention Technology and Risk Assessment of Ministry of Public Security, Beijing 100038, China
3
Beijing Public Security Bureau, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(8), 681; https://doi.org/10.3390/e26080681
Submission received: 24 June 2024 / Revised: 10 August 2024 / Accepted: 11 August 2024 / Published: 13 August 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
In order to minimize the disparity between visible and infrared modalities and enhance pedestrian feature representation, a cross-modality person re-identification method is proposed, which integrates modality generation and feature enhancement. Specifically, a lightweight network is used for dimension reduction and augmentation of visible images, and intermediate modalities are generated to bridge the gap between visible images and infrared images. The Convolutional Block Attention Module is embedded into the ResNet50 backbone network to selectively emphasize key features sequentially from both channel and spatial dimensions. Additionally, the Gradient Centralization algorithm is introduced into the Stochastic Gradient Descent optimizer to accelerate convergence speed and improve generalization capability of the network model. Experimental results on SYSU-MM01 and RegDB datasets demonstrate that our improved network model achieves significant performance gains, with an increase in Rank-1 accuracy of 7.12% and 6.34%, as well as an improvement in mAP of 4.00% and 6.05%, respectively.

1. Introduction

Person re-identification is an important issue in the field of computer vision [1], aimed at retrieving specific pedestrians from multiple non-overlapping cameras [2]. Traditional research on person re-identification mainly focuses on the application of technology under single modality, and most of the research is limited to the scene with sufficient light. With the continuous improvement in video surveillance security requirements, cameras with infrared modality switching function are gradually becoming popular and are applied in order to effectively cope with the limited working effect of visible cameras in night-time or harsh weather conditions. Aiming at the problem of different day and night lighting conditions, visible–infrared cross-modality person re-identification has been proposed [3].
Differently from traditional person re-identification under single modality, visible-infrared cross-modality person re-identification mainly studies the retrieval and matching of visible and infrared images of specific pedestrians in different horizons [4]. Due to the fact that visible images have three channels containing visible-light color information: red (R), green (G), and blue (B), while infrared images only have one channel containing near-infrared light intensity information, there is a fundamental difference in the images captured under the two modalities [5].
To alleviate this difference, the studies have been carried out based on three types of methods, including representation learning [6,7,8], metric learning [9,10,11], and modality conversion [12,13,14]. However, these methods tend to conduct cross-modality learning from the direct mapping of the two original modalities of visible images and infrared images, which makes the network model sensitive to parameters, difficult to converge, and requires a large amount of computational resources. Furthermore, existing methods fail to fully extract key channel features and spatial features of pedestrians in visible images and infrared images. These issues all result in low recognition accuracy of cross-modality person re-identification methods.
To address these issues, we propose a novel cross-modality person re-identification method based on the Attention Generalized mean pooling with Weighted triplet loss [15] (AGW) network model, which integrates modality generation and feature enhancement. The proposed method fully extracts shared features between visible images and infrared images, extracts more representative and generalized pedestrian features, accelerates the convergence speed of the network model, and improves the accuracy of cross-modality person re-identification. Our main contributions can be summarized as follows:
  • A lightweight network is used to generate intermediate modalities between visible images and infrared images as auxiliary modalities, enabling the network model to fully extract shared features between visible images and infrared images;
  • The feature enhancement method is utilized for optimizing feature extraction, and key feature information in pedestrian images is weighted and enhanced sequentially from both channel and spatial dimensions to enhance the efficiency and representation ability of the network model in utilizing pedestrian features;
  • The optimization strategy of centralizing gradient vectors is introduced to improve the generalization ability and training efficiency of the network model;
  • The experimental results on SYSU-MM01 and RegDB datasets show the superiority of the proposed method.

2. Related Work

The main challenge faced by cross-modality person re-identification is the modality differences between visible images and infrared images [16]. From the perspective of reducing the modality differences between visible images and infrared images, the visible-infrared cross-modality person re-identification methods can be summarized into three types: methods based on representation learning, methods based on metric learning, and methods based on modality conversion [3].
Methods based on representation learning mainly study how to design a reasonable network model architecture to extract discriminative and robust pedestrian features shared by visible images and infrared images, in order to reduce the modality differences between visible images and infrared images. In 2017, Wu et al. [6] proposed a deep-zero-padding data preprocessing method to train a single-stream network model. They first proposed cross-modality person re-identification and constructed the publicly available cross-modality person re-identification dataset SYSU-MM01; Yu et al. [7] proposed a dual-stream multi-branch network model based on fully correlated attention, which improves the color style robustness of the network model through a color-randomization data-augmentation algorithm, and enhances the discriminability of the extracted features by extracting multi-scale global features and local features; Fan et al. [8] proposed a clustering learning network model based on feature enhancement, which mines and enhances the subtle features of visible images and infrared images through global features and local features, and combines multi-level joint-clustering learning strategies to minimize modality differences and intra-class changes.
Methods based on metric learning mainly study how to reduce the modality differences between visible images and infrared images by designing a reasonable loss function to measure the similarity between two pedestrian images. Wang et al. [9] proposed a dual-path attention network model that enhances feature extraction through spatial dependencies between local features in pedestrian feature maps, and achieves center and boundary constraints on each class distribution through the proposed cross-modality dual-constraint loss; Zou et al. [10] proposed a heterogeneous center triplet loss based on angular distance, which not only solves the problem of selecting abnormal samples in traditional triplet loss, but also reduces the computational complexity of the network model; Zhang et al. [11] designed modality-consistency constraint loss and feature-center constraint loss, so that the modality-consistency constraint loss guides the network model to learn the invariant features between modalities, and the feature-center constraint loss supervises the network model to reduce intra-class feature differences and increase inter-class feature differences.
Methods based on modality conversion mainly study how to realize the mutual conversion of visible images and infrared images, utilizing a generative adversarial network [17] (GAN), and transform the cross-modality person re-identification into person re-identification under single modality to reduce the modality differences between visible images and infrared images. Wang et al. [12] proposed a double-layer difference-reduction learning method that decomposes mixed modalities and appearance differences. The difference-reduction subnetworks of image level and feature level are cascaded and jointly optimized in an end-to-end manner, while constraining modality and appearance differences; Wang et al. [13] proposed an end-to-end aligned generative adversarial network model, which reduces modality differences through a pixel alignment module and feature alignment module, and ensures identity consistency through a joint discrimination module; Zhang et al. [14] proposed a teacher–student generative adversarial network model based on different domains. By generating visible images into infrared images, the pre-trained teacher model generates feature maps to guide the student model in extracting discriminative features in the backbone network.
The aforementioned methods [6,7,8,9,10,11,12,13,14] address the issue of matching visible images and infrared images, to a certain extent. However, some challenges still remain unsolved, like low pedestrian-feature utilization efficiency and poor representation ability, insufficient network-model generalization ability, and slow convergence speed in cross-modality person re-identification. It is essential for the task to extract shared features of two modalities to effectively reduce the modality differences, and extract more representative pedestrian key channel features and spatial features and find a novel way to improve the training efficiency. Therefore, we propose a novel cross-modality person re-identification method to address the challenges above.

3. Methods

In this section, we first outline the overall framework of the proposed method in Section 3.1. Secondly, in Section 3.2, we explain the working principle of the lightweight modality generator [18] used in the proposed method. Then, in Section 3.3, we introduce the Convolutional Block Attention Module [19] (CBAM) incorporated into the feature extraction part of this work. Next, in Section 3.4, we describe the Gradient Centralization [20] (GC) optimization strategy adopted in the proposed method. Finally, in Section 3.5, we present all the loss functions that make up the network model.

3.1. Framework of the Proposed Method

On the basis of the AGW [15] network model, the proposed method consists of two parts: feature extraction and metric learning. As shown in Figure 1, the proposed method adopts a three-branch structure with ResNet50 [21], which embeds non-local neural networks [22] in Stage 2 and Stage 3, as the backbone network.
In the part of feature extraction, we input visible images into the lightweight modality generator for dimensionality reduction and augmentation, generating intermediate modalities between visible images and infrared images. Then, we jointly input them into the backbone network that shares weight parameters with other parts, except for the independent parameterized Stage 0, for feature extraction. In particular, we embedded the Convolutional Block Attention Module into Stage 1, weighting and enhancing key feature information sequentially from both channel and spatial dimensions. In the part of metric learning, all feature maps are constrained by the Generalized Mean pooling [23] (GeM) layer to output feature vectors and calculate the Weighted Regularization Triplet loss [15] (WRT). Afterwards, we used batch normalization [24] (BN) to normalize the feature vectors and mapped them using fully connected (FC) layers to calculate cross-entropy loss.
The network structure and overall process of the proposed method are shown in Figure 1, where GMP represents Generalized Mean pooling, BN represents batch normalization, and FC represents the fully connected layer. L triplet _ WRT and L cross - entropy represent the Weighted Regularization Triplet loss and cross-entropy loss, respectively.

3.2. Lightweight Modality Generator

The intermediate modality between visible images and infrared images can coordinate the connection between visible images and infrared images and establish a link, enabling the network model to fully extract the shared features between visible images and infrared images and thereby effectively reduce the modality differences between visible images and infrared images. The lightweight modality generator we utilized comprises two convolutional layers with kernel size 1 × 1 and a ReLU [25] activation layer. Compared to methods that use other auxiliary structures, such as generative adversarial network, the lightweight modality generator is more efficient and easy to optimize.
Using visible images as input, the intermediate modalities can be obtained from Equation (1). Among them, V represents the visible images, M represents the intermediate modalities, and g represents the lightweight modality generator, respectively.
M = g V
Specifically, the steps for generating intermediate modalities using the lightweight modality generator are shown in Figure 2. Firstly, the original three-channel visible image is mapped to a single-channel image after passing through the first 1 × 1 convolutional layer. This procedure compresses the features within the visible image into one channel. Subsequently, the single-channel image is passed into the ReLU activation layer, which effectively enhances the nonlinear expression capability. Finally, the single-channel image that passes through the ReLU activation layer is mapped to the three-channel intermediate modality image through another 1 × 1 convolutional layer. This procedure restores the number of channels in the image and preserves the crucial features extracted through the first two steps.
By generating intermediate modalities between visible images and infrared images, we effectively extracted shared features between these two images, thereby reducing the modality differences between visible images and infrared images. Using the intermediate modalities as auxiliary modalities, we connected visible images and infrared images, promoting the sharing of their features and making cross-modality learning easier.

3.3. Convolutional Block Attention Module

The introduction of attention mechanism can enable network models to focus on key information and features in data, thereby improving the efficiency and representation ability of network models in utilizing information. As a lightweight attention module, the Convolutional Block Attention Module can improve the feature representation ability and performance of network models in situations where training data and computing resources are limited. As shown in Figure 3, the Convolutional Block Attention Module consists of two modules: the channel attention module and the spatial attention module. Attention weights are inferred along the channel and spatial dimensions, and then multiplied with the input feature map to perform feature refinement operations.
Specifically, for the input feature map F, we pass it through the channel attention module to obtain the channel attention map M C ( F ) , which represents the evaluation of the importance of each channel in the input feature map F by the channel attention module. The operational steps of the channel attention module are shown in Equation (2), where σ represents the sigmoid function, AvgPool represents the global average-pooling operation, and MaxPool represents the global maximum-pooling operation.
M C F = σ M L P A v g P o o l F + M L P M a x P o o l F
We multiply the channel attention map M C ( F ) with the input feature map F of the channel attention module to obtain the input feature map F′ of the spatial attention module, as shown in Equation (3).
F = M C F F
After passing it through the spatial attention module, we obtain the spatial attention map M S ( F ) , which represents the evaluation of the importance of each pixel in the input feature map F′ by the spatial attention module. The operational steps of the spatial attention module are shown in Equation (4). Among them, σ represents the sigmoid function, f ( 7 × 7 ) represents the convolution operation with a filter size of 7 × 7 , AvgPool represents the global average-pooling operation, and MaxPool represents the global maximum-pooling operation.
M S F = σ f 7 × 7 A v g P o o l F ; M a x P o o l F
We multiply the spatial attention map M S ( F ) with the input feature map F′ of the spatial attention module to obtain the output feature map F″ of the input feature map F after being processed by the Convolutional Block Attention Module, as shown in Equation (5).
F = M S F F
We embed the Convolutional Block Attention Module after the third residual block in Stage 1 of the backbone network ResNet50, so that the improved network model can weight and enhance key feature information sequentially from both channel and spatial dimensions in sequence, thereby extracting more representative and generalized pedestrian features. The residual structure after adding the Convolutional Block Attention Module (CBAM) is shown in Figure 4.

3.4. Gradient Centralization

As an effective gradient optimization strategy, Gradient Centralization directly acts on the weight gradient, constraining the loss function in the network model by calculating the mean of the gradient vector, accelerating the convergence speed of the network model, improving the generalization ability of the network model, and enhancing the disturbance resistance of the loss function, as shown in Figure 5. Therefore, introducing Gradient Centralization into the Stochastic Gradient Descent [26] (SGD) optimizer regularizes the weight space and output-feature space, making the training process of the network model more stable and efficient.
Taking the fully connected layer as an example, in the case of obtaining the backpropagation gradient of the network model, the average value of each column vector in the gradient matrix is calculated, and then the average value is removed from each column vector to achieve centralization of each column vector. Finally, the centralized gradient matrix is returned, as shown in Equation (6):
Φ G C w p L = w p L 1 n q = 1 n w p , q L
Among them, w p L represent gradients, and w p represent weight vectors; p represents the p-th column vector in the gradient matrix, and q represents the q-th element in the p-th column vector.

3.5. Loss Function

We use the Weighted Regularization Triplet loss and cross-entropy loss as the metric loss and ID loss of the network model, respectively. The total loss function of the network model is shown in Equation (7), where L triplet _ WRT represents the Weighted Regularization Triplet loss and L cross - entropy represents the cross-entropy loss.
L total = L t r i p l e t _ W R T + L cross entropy
The Weighted Regularization Triplet loss can maintain the optimization of the relative distance between positive and negative sample pairs in traditional difficult samples for sampling triplet loss [27] without introducing any additional parameters, and has strong flexibility and adaptability. The calculation process of the Weighted Regularization Triplet loss is shown in Equation (8). Among them, ( i , j , k ) represents the difficult samples of sampling triplets within each training batch, P represents the positive sample set, N represents the negative sample set, d ij p represents the distance between the least similar positive sample pairs, d ik n represents the distance between the most similar negative sample pairs, d p represents the distance between each positive sample pair, and d n represents the distance between each negative sample pair.
L t r i p l e t _ W R T i , j , k = log 1 + exp w i p d i j p w i n d i k n w i p = exp d i j p p P exp d p w i n = exp d i k n n N exp d n
Cross-entropy loss anomaly is used in image classification tasks to evaluate the predictive performance of a network model by measuring the degree of similarity between its predictions and real data. The calculation process of cross-entropy loss is shown in Equation (9). Among them, N represents the number of samples, K represents the number of categories, yi,k represents the true category of the i-th sample, as k, and pi,k represents the probability of the predicted category of the i-th sample, as k.
L cross e n t r o p y = 1 N i = 1 N k = 1 K y i , k ln p i , k
We achieve constraints on the network model by using a joint supervision approach of the Weighted Regularization Triplet loss and cross-entropy loss.

4. Experiment and Analysis

4.1. Datasets and Evaluation Metric

We conducted experiments on two cross-modality person re-identification datasets, SYSU-MM01 [6] and RegDB [28], as shown in Table 1. Among them, the SYSU-MM01 dataset was captured by Sun Yat-sen University using 6 cameras, including 30,071 visible images and 15,792 infrared images of 491 pedestrians. The training set includes 11,909 infrared images and 22,258 visible images from 395 different pedestrians. In the test set, 3803 infrared images of 96 pedestrians were used as the query, and 301 randomly selected visible images were used as the gallery. The RegDB dataset was captured by a dual camera system consisting of a visible camera and a thermal imaging camera, containing 412 different pedestrians, each with 10 visible images and 10 infrared images. The training set includes 206 pedestrians, 2060 visible images, and 2060 infrared images. The test set includes 206 pedestrians, 2060 visible images, and 2060 infrared images.
We use the commonly used Cumulative Match Characteristic (CMC) and mean Average Precision [29] (mAP) in retrieval tasks as evaluation metrics for the network model. Among them, CMC is used to measure the network model’s ability to correctly identify target pedestrians during the retrieval process. It can intuitively reflect the performance of the network model by calculating the correct matching rate under different rankings. mAP is used to measure the average accuracy of network models in retrieving correct pedestrian identities. It can comprehensively evaluate the performance of the network model in all query scenarios by calculating the average accuracy of each retrieval in all queries.

4.2. Experimental Settings

The experiment was implemented on NVIDIA TITAN V GPU (manufacturer NVIDIA, sourced from Beijing, China) using PyTorch deep learning framework version 1.4.0. We used ResNet50, which removes the average pooling layer and fully connected layer, as the backbone network, and set the convolution step size of the last layer to 1. During the training process, we enhanced the image data through random cropping and horizontal flipping operations, and adjusted the image size to 288 × 144 . The number of epochs was set to 80, and the batch size was set to 72, including 24 visible images, 24 infrared images, and 24 intermediate modality images of 6 pedestrians. As shown in Equation (10), we used the dynamic learning rate and the Stochastic Gradient Descent optimizer with Gradient Centralization to update the gradient of the network model. The weight decay of the optimizer is set to 5 × 10 - 4 , and the momentum factor is set to 0.9.
l r epoch = 0.1 × e p o c h + 1 10 , e p o c h < 10 0.1 , 10 e p o c h < 20 0.01 , 20 e p o c h < 50 0.001 , e p o c h 50

4.3. Comparison with Existing Methods

We compared the performance of our proposed method with existing cross-modality person re-identification methods on the SYSU-MM01 and RegDB datasets, and the results are shown in Table 2. It can be seen that, compared with the baseline method, the proposed method has improved recognition accuracy on both the SYSU-MM01 and RegDB datasets, proving the effectiveness of the improved method. Compared with the existing cross-modality person re-identification methods in the table, the proposed method has significantly improved recognition accuracy on the SYSU-MM01 and RegDB datasets, proving its effectiveness.

4.4. Ablation Study

4.4.1. Using Lightweight Modality Generator

Visible images were input into the lightweight modality generator for dimensionality reduction and augmentation, generating intermediate modalities between visible images and infrared images, and were then input into the network model, together with visible images and infrared images. The performance evaluation results of the network model are shown in Table 3. It can be seen that the network model using the lightweight modality generator shows significant performance improvements on both the SYSU-MM01 and RegDB datasets. Among them, on the SYSU-MM01 dataset, Rank-1 improved by 5.47% and mAP improved by 1.16%; on the RegDB dataset, Rank-1 improved by 4.09% and mAP improved by 2.39%.

4.4.2. Incorporating Convolutional Block Attention Module

The Convolutional Block Attention Module was embedded into the backbone network ResNet50, and the performance evaluation results of the network model are shown in Table 4. It can be seen that the network model embedded with the Convolutional Block Attention Module shows modest performance improvements on both the SYSU-MM01 and RegDB datasets. Among them, on the SYSU-MM01 dataset, Rank-1 improved by 3.89% and mAP improved by 0.82%; on the RegDB dataset, Rank-1 improved by 3.01% and mAP improved by 2.10%.

4.4.3. Introducing Gradient Centralization

The Gradient Centralization algorithm was introduced into the Stochastic Gradient Descent optimizer, and the performance evaluation results of the network model are shown in Table 5. It can be seen that the network model introducing Gradient Centralization shows a slight improvement in performance on both the SYSU-MM01 and RegDB datasets. Among them, on the SYSU-MM01 dataset, Rank-1 improved by 0.09% and mAP improved by 0.23%; on the RegDB dataset, Rank-1 improved by 0.31% and mAP improved by 0.37%.

4.4.4. Ablation Experiments to Verify the Effectiveness of 3 Modules

To verify the effectiveness of various parts of the network model, we conducted ablation experiments on the SYSU-MM01 and RegDB datasets. The Rank-1 and mAP obtained from each group of experiments are shown in Table 6. It can be seen that using the lightweight modality generator, incorporating Convolutional Block Attention Module, and introducing Gradient Centralization simultaneously, the performance of the network model on the SYSU-MM01 and RegDB datasets is the best, with Rank-1 improving by 7.12% and 6.34%, respectively; MAP increased by 4.00% and 6.05%, respectively. The performance disparity of the network model on the SYSU-MM01 and RegDB datasets might be attributed to the variance in the sample size of the two datasets.

4.5. Visualization Analysis

The CMC of the proposed method on the RegDB dataset is shown in Figure 6. It can be seen that after using the lightweight modality generator, incorporating the Convolutional Block Attention Module, and introducing Gradient Centralization, the network model can perform better on the dataset. This indicates that the proposed method can effectively reduce the modality differences between visible images and infrared images, extract more representative and generalized pedestrian features, and improve the accuracy of person re-identification.
The loss curve of the proposed method on the RegDB dataset is shown in Figure 7. It can be seen that after introducing the Gradient Centralization algorithm into the Stochastic Gradient Descent optimizer, the loss of the first epoch is significantly reduced, and the convergence speed of the network model is accelerated. This indicates that the proposed method can improve the convergence speed and generalization ability of the network model.
The visualization results of the proposed method on the SYSU-MM01 and RegDB datasets are shown in Figure 8 and Figure 9, respectively. Among them, query is the image we want to search, and images 1–10 are the query results returned by the network model from the image database. The images with green borders are considered positive samples, meaning that they belong to the same pedestrian entity as the query image. In contrast, the images with red borders are considered negative samples, meaning that they do not belong to the same pedestrian entity as the query image.
It can be seen that for the simple cases of querying normal front or back infrared pedestrian images, the proposed method can basically retrieve all the correct samples. For more challenging cases, such as querying poor-quality infrared pedestrian images or infrared pedestrian images with occlusions such as backpacks, the proposed method may confuse some samples that are difficult to distinguish. This might be attributed to the considerable noise present in the queried infrared images and the occlusion of key pedestrian features such as backpacks, which leads to the inability of the network model to acquire comprehensive pedestrian-feature information.

5. Conclusions

In this paper, we proposed a novel cross-modality person re-identification method that integrates modality generation and feature enhancement. Firstly, the lightweight modality generator is used to generate intermediate modalities between visible images and infrared images, enabling the network model to fully extract shared features between visible images and infrared images; then, the Convolutional Block Attention Module (CBAM) is incorporated into the backbone network ResNet50, and key feature information is weighted and enhanced sequentially from both channel and spatial dimensions to enhance the efficiency and representation ability of the network model in utilizing pedestrian features; finally, the Gradient Centralization (GC) algorithm is introduced into the Stochastic Gradient Descent (SGD) optimizer to improve the convergence speed and generalization ability of the network model. We tested and evaluated the network model on the SYSU-MM01 and RegDB datasets, and the Rank-1 reached 54.62% and 77.84%, respectively. The mAP reached 51.65% and 86.18%, respectively. The above results demonstrate the superiority of the proposed cross-modality person re-identification method.
From the results of this article, it can be observed that it is worth investigating how to improve the computational efficiency of the network model through optimizing the structure and parameters. Additionally, how to employ large language models to facilitate the development of cross-modality person re-identification is also a challenging task. In future research, on the one hand, we will explore efficient cross-modality person re-identification under limited computing resources through techniques such as model compression and pruning. On the other hand, we will explore approaches to enhance the overall performance and robustness of the network model by utilizing the rich contextual information and prior knowledge provided by large language models.

Author Contributions

Conceptualization, Y.B.; methodology, Y.B.; formal analysis, Y.B. and R.W.; investigation, Y.B. and R.W.; writing—original draft preparation, Y.B.; writing—review and editing, Y.B. and Q.Z.; supervision, R.W., Z.Z., R.L. and M.W.; funding acquisition, R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Double First-Class Innovation Research Project for the People’s Public Security University of China, grant number 2023SYL08.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available online. These datasets were derived from the following public resources: [SYSU-MM01, RegDB].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, L. A Review of Person Re-Identification Based on Deep Learning. China Water Transp. (Second. Half Mon.) 2023, 23, 57–59. [Google Scholar] [CrossRef]
  2. Wang, S.; Xiao, S. Review of Person Re-identification. J. Beijing Inst. Technol. 2022, 48, 1100–1112. [Google Scholar]
  3. Liu, T.; Liu, Z. Overview of Cross Modality Person Re-Identification Research. Mod. Comput. Sci. 2021, 135–139. [Google Scholar] [CrossRef]
  4. Sun, Y.; Wang, R.; Zhang, Q.; Lin, R. A cross-modality person re-identification method for visible-infrared images. J. Beijing Univ. Aeronaut. Astronaut. 2022, 50, 2018–2025. [Google Scholar] [CrossRef]
  5. Han, C.; Pan, P.; Zheng, A.; Tang, J. Cross-Modality Person Re-Identification Based on Heterogeneous Center Loss and Non-Local Features. Entropy 2021, 23, 919. [Google Scholar] [CrossRef] [PubMed]
  6. Wu, A.; Zheng, W.S.; Yu, H.X.; Gong, S.; Lai, J. RGB-infrared cross-modality person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  7. Yu, W.; Zhao, Q.; Ji, T. Cross-modal pedestrian re-recognition based on color randomization and full related attention. Foreign Electron. Meas. Technol. 2023, 42, 10–16. [Google Scholar] [CrossRef]
  8. Fan, X.; Zhang, K.; Zhang, G.; Li, J. Cross-modal person re-identification algorithm based on multi-level joint clustering with feature enhancement. J. Electron. Meas. Instrum. 2024, 38, 94–103. [Google Scholar] [CrossRef]
  9. Wang, C.; Zhang, C.; Feng, Y.; Ji, Y.; Ding, J. Learning Visible Thermal Person Re-Identification via Spatial Dependence and Dual Constraint Loss. Entropy 2022, 24, 443. [Google Scholar] [CrossRef] [PubMed]
  10. Zou, Y.; Jiang, M. Multi-granularity cross-modality person re-identification with hetero-center angular constraints. Comput. Eng. Des. 2024, 45, 1210–1217. [Google Scholar]
  11. Zhang, J.; Chen, G. Visible-infrared Person Re-Identification Via Feature Constrained Learning. Prog. Laser Optoelectron. 2024, 61, 221–228. [Google Scholar] [CrossRef]
  12. Wang, Z.; Wang, Z.; Zheng, Y.; Chuang, Y.Y.; Satoh, S.I. Learning to reduce dual-level discrepancy for infrared-visible person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 618–626. [Google Scholar]
  13. Wang, G.; Zhang, T.; Cheng, J.; Liu, S.; Yang, Y.; Hou, Z. RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 3623–3632. [Google Scholar]
  14. Zhang, Z.; Jiang, S.; Huang, C.; Li, Y.; Da Xu, R.Y. RGB-IR cross-modality person ReID based on teacher-student GAN model. Pattern Recognit. Lett. 2021, 150, 155–161. [Google Scholar] [CrossRef]
  15. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep learning for person re-identification: A survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 2872–2893. [Google Scholar] [CrossRef] [PubMed]
  16. Liu, Y. A Review of Cross-Modal Person Re-Identification. Telev. Technol. 2022, 46, 9–11. [Google Scholar] [CrossRef]
  17. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  18. Li, D.; Wei, X.; Hong, X.; Gong, Y. Infrared-visible cross-modal person re-identification with an x modality. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4610–4617. [Google Scholar]
  19. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  20. Yong, H.; Huang, J.; Hua, X.; Zhang, L. Gradient centralization: A new optimization technique for deep neural networks. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part I 16. Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 635–652. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  23. Radenović, F.; Tolias, G.; Chum, O. Fine-tuning CNN image retrieval with no human annotation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1655–1668. [Google Scholar] [CrossRef] [PubMed]
  24. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning; PMLR: New York, NY, USA, 2015; pp. 448–456. [Google Scholar]
  25. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics; JMLR Workshop and Conference Proceedings; PMLR: New York, NY, USA, 2011; pp. 315–323. [Google Scholar]
  26. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  27. Hermans, A.; Beyer, L.; Leibe, B. In defense of the triplet loss for person re-identification. arXiv 2017, arXiv:1703.07737. [Google Scholar]
  28. Nguyen, D.T.; Hong, H.G.; Kim, K.W.; Park, K.R. Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 2017, 17, 605. [Google Scholar] [CrossRef]
  29. Chen, Y.C.; Zheng, W.S.; Lai, J.H.; Yuen, P.C. An asymmetric distance model for cross-view feature mapping in person reidentification. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 1661–1675. [Google Scholar] [CrossRef]
  30. Ye, M.; Lan, X.; Li, J.; Yuen, P. Hierarchical discriminative learning for visible thermal person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  31. Ye, M.; Lan, X.; Wang, Z.; Yuen, P.C. Bi-directional center-constrained top-ranking for visible thermal person re-identification. IEEE Trans. Inf. Forensics Secur. 2019, 15, 407–419. [Google Scholar] [CrossRef]
  32. Hao, Y.; Wang, N.; Li, J.; Gao, X. HSME: Hypersphere manifold embedding for visible thermal person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8385–8392. [Google Scholar]
  33. Ye, M.; Lan, X.; Leng, Q. Modality-aware collaborative learning for visible thermal person re-identification. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 347–355. [Google Scholar]
  34. Feng, Z.; Lai, J.; Xie, X. Learning modality-specific representations for visible-infrared person re-identification. IEEE Trans. Image Process. 2019, 29, 579–590. [Google Scholar] [CrossRef] [PubMed]
  35. Choi, S.; Lee, S.; Kim, Y.; Kim, T.; Kim, C. Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
Figure 1. Overall architecture of the proposed method. Among them, the initial Stage 0 includes the initial convolutional layer, batch normalization (BN) layer, ReLU layer, and Max pooling layer.
Figure 1. Overall architecture of the proposed method. Among them, the initial Stage 0 includes the initial convolutional layer, batch normalization (BN) layer, ReLU layer, and Max pooling layer.
Entropy 26 00681 g001
Figure 2. Structure of the lightweight modality generator.
Figure 2. Structure of the lightweight modality generator.
Entropy 26 00681 g002
Figure 3. Overview of Convolutional Block Attention Module.
Figure 3. Overview of Convolutional Block Attention Module.
Entropy 26 00681 g003
Figure 4. Residual structure of CBAM-ResNet.
Figure 4. Residual structure of CBAM-ResNet.
Entropy 26 00681 g004
Figure 5. Diagram of Gradient Centralization.
Figure 5. Diagram of Gradient Centralization.
Entropy 26 00681 g005
Figure 6. CMC curves of five methods on RegDB dataset.
Figure 6. CMC curves of five methods on RegDB dataset.
Entropy 26 00681 g006
Figure 7. Loss curves before and after introducing Gradient Centralization on RegDB dataset.
Figure 7. Loss curves before and after introducing Gradient Centralization on RegDB dataset.
Entropy 26 00681 g007
Figure 8. Visualization results output on the SYSU-MM01 dataset.
Figure 8. Visualization results output on the SYSU-MM01 dataset.
Entropy 26 00681 g008
Figure 9. Visualization results output on the RegDB dataset.
Figure 9. Visualization results output on the RegDB dataset.
Entropy 26 00681 g009
Table 1. Dataset condition.
Table 1. Dataset condition.
DatasetsNumber of
Pedestrians
Number of
Visible Cameras
Number of
Infrared Cameras
Number of Visible ImagesNumber of
Infrared Images
SYSU-MM014914230,07115,792
RegDB4121141204120
Table 2. Performance comparison of different methods on SYSU-MM01 and RegDB datasets.
Table 2. Performance comparison of different methods on SYSU-MM01 and RegDB datasets.
MethodSYSU-MM01RegDB
Rank-1mAPRank-1mAP
Zero-Padding [6]14.8015.9517.7518.90
HCML [30]14.3216.1624.4420.08
BDTR [31]27.8228.4234.6233.46
HSME [32]20.6823.1250.8547.00
D2RL [12]28.9029.2043.4044.10
MAC [33]33.2636.2236.4337.03
MSR [34]37.3538.1148.4348.67
AlignGAN [13]42.4040.7057.9053.60
Hi-CMD [35]34.9035.9070.9366.04
XIV-ReID [18]49.9250.7362.2160.18
baseline (AGW)47.5047.6570.5080.13
baseline + M_GEN + CBAM + GC (ours) 154.6251.6576.8486.18
1 In this table, M_GEN is an abbreviation for the modality generator, CBAM is an abbreviation for Convolutional Block Attention Module, and GC is an abbreviation for Gradient Centralization, as is the case in subsequent tables.
Table 3. Performance comparison of models using lightweight modality generator on SYSU-MM01 and RegDB datasets.
Table 3. Performance comparison of models using lightweight modality generator on SYSU-MM01 and RegDB datasets.
MethodSYSU-MM01RegDB
Rank-1mAPRank-1mAP
baseline47.5047.6570.5080.13
baseline + M_GEN52.9748.8174.5982.52
Table 4. Performance comparison of models incorporating Convolutional Block Attention Module on SYSU-MM01 and RegDB datasets.
Table 4. Performance comparison of models incorporating Convolutional Block Attention Module on SYSU-MM01 and RegDB datasets.
MethodSYSU-MM01RegDB
Rank-1mAPRank-1mAP
baseline47.5047.6570.5080.13
baseline + CBAM51.3948.4773.5182.23
Table 5. Performance comparison of models introducing Gradient Centralization on SYSU-MM01 and RegDB datasets.
Table 5. Performance comparison of models introducing Gradient Centralization on SYSU-MM01 and RegDB datasets.
MethodSYSU-MM01RegDB
Rank-1mAPRank-1mAP
baseline47.5047.6570.5080.13
baseline + GC47.5947.8870.8180.50
Table 6. Results of ablation experiments on SYSU-MM01 and RegDB datasets.
Table 6. Results of ablation experiments on SYSU-MM01 and RegDB datasets.
MethodSYSU-MM01RegDB
Rank-1mAPRank-1mAP
baseline47.5047.6570.5080.13
baseline + GC47.5947.8870.8180.50
baseline + CBAM51.3948.4773.5182.23
baseline + M_GEN52.9748.8174.5982.52
baseline + GC + CBAM51.5849.2773.2481.92
baseline + GC + M_GEN52.9150.0475.4184.54
baseline + CBAM + M_GEN53.8650.6375.7685.89
baseline + GC + CBAM + M_GEN54.6251.6576.8486.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bi, Y.; Wang, R.; Zhou, Q.; Zeng, Z.; Lin, R.; Wang, M. Cross-Modality Person Re-Identification Method with Joint-Modality Generation and Feature Enhancement. Entropy 2024, 26, 681. https://doi.org/10.3390/e26080681

AMA Style

Bi Y, Wang R, Zhou Q, Zeng Z, Lin R, Wang M. Cross-Modality Person Re-Identification Method with Joint-Modality Generation and Feature Enhancement. Entropy. 2024; 26(8):681. https://doi.org/10.3390/e26080681

Chicago/Turabian Style

Bi, Yihan, Rong Wang, Qianli Zhou, Zhaolong Zeng, Ronghui Lin, and Mingjie Wang. 2024. "Cross-Modality Person Re-Identification Method with Joint-Modality Generation and Feature Enhancement" Entropy 26, no. 8: 681. https://doi.org/10.3390/e26080681

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop