Next Article in Journal
A Brief Survey of Recent Advances and Methodologies for the Security Control of Complex Cyber–Physical Networks
Next Article in Special Issue
A Small Object Detection Method for Oil Leakage Defects in Substations Based on Improved Faster-RCNN
Previous Article in Journal
Scan Matching-Based Particle Filter for LIDAR-Only Localization
Previous Article in Special Issue
3D Object Detection for Self-Driving Cars Using Video and LiDAR: An Ablation Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Modality Person Re-Identification via Local Paired Graph Attention Network

1
Tianjin Key Laboratory of Wireless Mobile Communications and Power Transmission, Tianjin Normal University, Tianjin 300387, China
2
Department of Electronic and Electrical Engineering, University of Strathclyde, Glasgow G1 1QE, UK
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(8), 4011; https://doi.org/10.3390/s23084011
Submission received: 27 February 2023 / Revised: 7 April 2023 / Accepted: 13 April 2023 / Published: 15 April 2023

Abstract

:
Cross-modality person re-identification (ReID) aims at searching a pedestrian image of RGB modality from infrared (IR) pedestrian images and vice versa. Recently, some approaches have constructed a graph to learn the relevance of pedestrian images of distinct modalities to narrow the gap between IR modality and RGB modality, but they omit the correlation between IR image and RGB image pairs. In this paper, we propose a novel graph model called Local Paired Graph Attention Network (LPGAT). It uses the paired local features of pedestrian images from different modalities to build the nodes of the graph. For accurate propagation of information among the nodes of the graph, we propose a contextual attention coefficient that leverages distance information to regulate the process of updating the nodes of the graph. Furthermore, we put forward Cross-Center Contrastive Learning ( C 3 L ) to constrain how far local features are from their heterogeneous centers, which is beneficial for learning the completed distance metric. We conduct experiments on the RegDB and SYSU-MM01 datasets to validate the feasibility of the proposed approach.

1. Introduction

The purpose of person re-identification (ReID) [1,2,3,4] is to match pedestrians across multiple non-overlapping cameras, which could be considered to be a specific person-retrieval task. It is extensively applied in smart cities, autonomous driving, security surveillance, and so on. However, most person ReID methods focus on matching pedestrians captured by RGB cameras, and therefore they do not allow 24-h intelligent surveillance. To overcome this limitation, some researchers are dedicated to cross-modality person ReID.
Cross-modality person ReID [5,6,7,8,9] retrieves RGB pedestrian images from infrared (IR) pedestrian images and vice versa. It not only inherits the challenges of unimodality person ReID, such as the variations in postures, illumination, and camera view, but also possesses a large discrepancy between IR modality and RGB modality. The modality discrepancy results in an unreliable match due to different color and appearance information of IR images and RGB images.
Recently, some cross-modality person ReID methods have been proposed to learn feature representations and metric functions for both IR and RGB images. Regarding feature representations, many methods [10,11,12] extract modality-specific features and modality-shared features by designing dual-stream deep networks. Meanwhile, the local features [13,14,15] are also demonstrated to be effective for cross-modality person ReID. Furthermore, some approaches [16,17,18,19] apply graph convolution layers, which aggregate features from other pedestrian images to enhance the discriminative power of features. They treat each pedestrian image as a node of a graph and update the nodes based on their correlations, as shown in Figure 1a. However, they ignore the relationship between the pairs of IR images and RGB images when constructing the graph, therefore hindering the learning of discriminative features for pedestrian images of different modalities.
In metric learning, some works [20,21,22] propose various losses to minimize the distance between pedestrian images from different modalities for cross-modal person ReID. These losses target the learning of an embedding space for different modalities in which images of same-identity pedestrians are closer to each other and images of different-identity pedestrians are further away. For this purpose, these methods constrain the distance among the pedestrian images from different modalities [23,24] or the distance among the centers from different modalities [25,26]. However, they overlook the distance between the pedestrian image and its center of different modalities, which results in incomplete distance learning between different modalities.
In this paper, we propose a novel graph network entitled Local Paired Graph Attention Network (LPGAT) for cross-modality person ReID. This approach considers the correlation of paired pedestrian images from different modalities and local information in a uniform framework. Specifically, we design the proposed method as a two-stream network where each branch corresponds to each modality. To learn local features, we uniformly divide the feature maps for each stream in the horizontal direction. Later, we construct a graph using local features where each node is composed of the corresponding local features of paired pedestrian images with varying modalities, which is illustrated in Figure 1b. The inclusion criterion of the selected paired local features is that each paired local features come from two different modalities. For better propagation of information in the graph, we further propose the contextual attention coefficient which not only considers the node features but also the relationship between the nodes. Hence, the proposed LPGAT could directly learn the relationship between the paired local features from different modalities and accurately propagate the information between them.
Recently, the distance constraint among the distinct modality centers has verified the effectiveness of cross-modality person ReID. However, they do not constrain the features which are far from the centers. To overcome this, we propose Cross-Center Contrastive Learning ( C 3 L ) to reduce the distance between local features and their heterogeneous centers to narrow the gap between heterogeneous modalities. Combined with the constraint between heterogeneous centers, the proposed C 3 L helps learn the completed distance metric for IR images and RGB images, and therefore the discriminative features are obtained. The primary contributions are outlined below:
  • We propose LPGAT for cross-modality ReID. In contrast to previous approaches that only use pedestrian images from different modalities as the nodes of a graph, LPGAT uses the paired local features from different modalities as the nodes of a graph, thus alleviating the gap between the two modalities.
  • We propose C 3 L to constrain local features and their heterogeneous centers. In contrast to previous methods that only constrain the distance between the centers of different modalities, C 3 L constrains the features that are far from the center, thus narrowing the gap between heterogeneous modalities.
  • We compare the proposed method against state-of-the-art methods using two publicly accessible datasets, RegDB and SYSU-MM01, and our results demonstrate that the proposed method outperforms them.

2. Related Work

2.1. Cross-Modality Person ReID

To address the challenge of the modality gap for cross-modality person ReID, many methods are proposed to derive global or local features from heterogeneous pedestrian images. For the global features, Wu et al. [27] put forward the deep zero-padding model to learn complementary information from IR images and RGB images. Then, Ye et al. [20] introduced a dual-stream network to capture the globally shared information of IR images and RGB images. Chen et al. [6] raised the Neural Feature Search (NFS) to implement the feature-selection automation, which allows the network to filter the background noise and focus on the important portions of pedestrian images.
Since pedestrians’ partial information is crucial for cross-modality person ReID, several researchers have proposed learning local features from IR images and RGB images. Zhu et al. [21] and Sun et al. [22] developed a deep two-stream framework to capture local features to mitigate modality differences. Zhang et al. [28] proposed Dual-Alignment Part-aware Representation (DAPR) to simultaneously reduce the modality gap and learn discriminant features from the local and global aspects. In this paper, we adopt a two-steam deep network to learn local features, and aggregate paired local features of pedestrian images from different modalities.

2.2. Graph Attention Networks

Graph Convolutional Network (GCN) [29,30] has been proposed to handle the non-Euclidean data. It learns node features by propagating the information among nodes as well as their neighborhoods. Several vision-related tasks, including semantic segmentation [31] and face analysis [32], have widely applied GCN. Later, Graph Attention Network (GAT) [33,34] is further proposed to aggregate node features using attention weights.
Recently, more and more researchers have combined GCN or GAT with Convolutional Neural Network (CNN) for person ReID [16,19,35]. Ye et al. [16] put forward Dynamic Dual Attention Aggregation (DDAG), in which each pedestrian image is regarded as a node of the graph, and the relationship between the node and its neighborhoods is mined. Zhang et al. [19] treated each body part as a node of a graph and construct a graph using one pedestrian image to alleviate the intra-modality variations. In this paper, a graph is constructed using paired local features derived from distinct modalities, and the contextual attention coefficients are introduced for better propagation.

2.3. Contrastive Learning

The purpose of contrastive learning [36,37] is to learn discriminative features using image pairs, so that similar images are close to each other, while dissimilar ones are far away. It can be applied in both unsupervised and supervised learning, such as image classification [37] and object detection [38]. Recently, contrastive learning has been used in person ReID to improve the discrimination of features [39,40]. For example, Chen et al. [39] proposed Inter-instance Contrastive Encoding (ICE) to fully explore the relationship between different pedestrian images. Isobe et al. [41] presented the Cluster-wise Contrastive Learning (CCL) algorithm to learn noise-robust features for cross-domain person ReID. We propose C 3 L to reduce the distance between local features and their heterogeneous centers to reduce the modality gap for cross-modality person ReID, which is inspired by the applications of contrastive learning in cross-domain ReID.

3. Approach

In this section, an outline of the proposed method is first represented, which is depicted in Figure 2. The proposed method contains three key components, namely Local Feature Extractor, LPGAT Module, and C 3 L . Then, we introduce each of them in detail. Finally, we optimize the proposed approach.

3.1. Overview

Local Feature Extractor. The Local Feature Extractor is designed as a two-stream network where two individual ResNet-50 [42] are adopted as the backbone. Then, we divide the feature maps output by ResNet-50 horizontally and apply the global average pooling (GAP) to obtain the local features. Afterward, we apply the fully connected (FC) layers to reduce the dimension of local features, where the weights of FC layers are shared.
LPGAT Module. We regard the difference between paired local features of pedestrian images from different modalities as the node of a graph to learn the paired correlation of local features. Then, we update the nodes using the contextual attention coefficient which injects the distance information between the nodes into the process of information propagation.
Cross-Center Contrastive Learning. We introduce C 3 L to optimize the similarity between the local features and their heterogeneous centers. We then combine C 3 L with other metric functions to obtain the completed distance metric.

3.2. Local Feature Extractor

The Local Feature Extractor possesses two streams, and two individual pre-trained ResNet-50 are used as the backbone, where the stride of the convolution operation in the last layer is modified from 2 to 1. We feed the pedestrian images of IR modality and RGB modality into the two streams, respectively. Later, we obtain the feature maps of pedestrian images with the size of W × H × C , where W and H denote the width and the height of feature maps, and C is the number of channels. Afterward, we uniformly split the feature maps into P part-level stripes. We conduct GAP on each part-level stripe to obtain the local feature. The p-th local feature of the i-th RGB image is denoted as f i , p R R C × 1 , where p = 1 , , P . Similarly, the local feature of the IR image is denoted as f j , p I R C × 1 . Finally, we apply the FC layers with shared weights to reduce the dimension of the local features from C to D.

3.3. LPGAT Module

The local features have proved the robustness to the variances in viewpoints, poses, and so on [1,13]. Meanwhile, the paired features derived from distinct modalities facilitate the reduction of the modality gap. Hence, we propose to use the paired local features from the IR modality and RGB modality to construct the fully connected graph. The node of the fully connected graph is defined as:
g i , j p = f i , p R f j , p I
Hence, we obtain a fully connected graph G i p = { g i , 1 p , g i , 2 p , , g i , U p } for the p-th local feature of the i-th RGB image, where U is the number of nodes in the graph. Please note that the node of a graph can also be performed by subtracting the local feature of RGB modality from the local feature of IR modality.
After obtaining the graph, we need to calculate the attention coefficient to describe the correlation between different nodes. Many cross-modality person ReID approaches [16,17] calculate the attention coefficient between the nodes as:
α n , m p = e x p ( ( q g i , n p , g i , m p ) ) u = 1 U e x p ( ( q g i , n p , g i , u p ) )
where is a nonlinear operation performed by LeakyReLU, , is the concatenation operation, and q R 2 D × 1 is a learnable vector.
From Equation (2) we can see that it directly concatenates the nodes, but ignores the relationship between the nodes, which leads to inaccurate information propagation. Thus, we propose the contextual attention coefficient:
α n , m p = e x p ( ( q g i , n p , g i , m p ) ) u = 1 U e x p ( ( q g i , n p , g i , u p ) ) × k n , m p
k n , m p = e x p ( g i , n p g i , m p 2 ) + β
where · 2 indicates the Euclidean distance, and β is the hyperparameter. From Equation (4), we can see that the smaller distance between the nodes possesses a larger k n , m p , and therefore it produces a strong correlation between the nodes. Hence, the contextual attention coefficient is helpful for accurate information propagation. Please note that when we set k n , m p to 1, the contextual attention coefficient degenerates to the traditional attention coefficient. With the contextual attention coefficient, the node is represented as:
g ^ i , n p = m = 1 U α n , m p g i , m p
Finally, to further improve the representation ability, the node is updated as:
e i , n p = ϕ ( w g ^ i , n p , g i , n p )
where ϕ is the ELU activation function to learn a stable graph structure, and w R 2 D × 2 is a learnable matrix. We treat the optimization of LPGAT as a binary classification problem, and use the verification loss:
L g p = j = 1 U z i , j p log z ¯ i , j p ( 1 z i , j p ) log ( 1 z ¯ i , j p )
where z ¯ i , j p is the predicted probability of the j-th node of G i p , and z i , j p is the ground-truth of the j-th node of G i p . z i , j p = 1 indicates the paired local features derived from distinct modalities in the node are with the same identity, otherwise z i , j p = 0 .
In a word, we design the node of the graph in LPGAT as the paired local features derived from distinct modalities to effectively mitigate the discrepancy between IR modality and RGB modality. Furthermore, we inject the distance information using the contextual attention coefficient to propagate the information between the nodes accurately.

3.4. Cross-Center Contrastive Learning

As for cross-modality person ReID, learning the distance metric is an effective way to narrow the modality gap. Recently, the constraint on the centers of RGB modality and IR modality have achieved promising performance [21,25,26]. However, they overlook the distance between features and their heterogeneous centers resulting in some outliers in the learning process as shown in Figure 3a.
In this paper, we put forward C 3 L to force the local features to be close to the corresponding heterogeneous centers in the embedding space, and therefore the pedestrian images which have the same identity from distinct modalities are gathered as shown in Figure 3b.
The constraint between the p-th local feature of the i-th RGB image and its heterogeneous center is defined as:
L R p = log e x p { ( f i , p R ) c b , p I / τ } s = 1 S e x p { ( f i , p R ) c s , p I / τ } s . t . I D ( i , V ) = b
where τ > 0 is a scalar temperature parameter, S is the number of identities, and I D ( i , V ) indicates the identity of the i-th RGB image. Here, c b , p I is the center of the p-th local feature of the b-th identity for IR images, and it is defined as:
c b , p I = 1 O b i = 1 O b f i , p I s . t . I D ( i , I ) = b
where O b is the number of pedestrian images with the b-th identity, and I D ( i , I ) is the identity of the i-th IR image.
Similarly, the constraint between the p-th local feature of the i-th IR image and its heterogeneous center is defined as:
L I p = log e x p { ( f i , p I ) c b , p V / τ } s = 1 S e x p { ( f i , p I ) c s , p V / τ } s . t . I D ( i , I ) = b
where c b , p V is the heterogeneous center of f i , p I . In a word, the proposed C 3 L is formulated as:
L C p = L R p + L I p
The proposed C 3 L decreases the distance between the local features of IR images and their centers of RGB images, and so do the local features of IR images and their centers of RGB images. Hence, it clusters the local features of pedestrian images from different modalities with the same identity.

3.5. Optimization

For learning the completed distance metric, we exploit the proposed C 3 L as well as the heterogeneous center (HC) loss [21]. The HC loss aims to diminish the distance between the centers of IR modality and RGB modality with the same identity. The HC loss for the p-th local feature is denoted as L H C p . Additionally, we employ the cross-entropy loss to optimize the local features, and the cross-entropy loss for the p-th local feature is denoted as L C E p . Moreover, we use the validation loss L g p to optimize LPGAT and treat it as a binary classification task. Therefore, the overall loss of the proposed method is expressed as:
L = 1 P p = 1 P ( L C E p + λ 1 L H C p + λ 2 L g p + λ 3 L C p )
where λ 1 , λ 2 and λ 3 are the trade-off parameters to balance the importance between different losses.

4. Experiments

In this section, the evaluation protocol and datasets are first presented, followed by showing the implementation details of our experiments. After that, the experimental results are compared with the state-of-the-art methods, and ablation experiments are performed to evaluate the effectiveness of the key components of the presented approaches. Finally, the influence of several important parameters in the proposed method is analyzed.

4.1. Datasets

We perform trials on the SYSU-MM0 [27] and RegDB [43], which are two public cross-modal datasets.
SYSU-MM01, a massive dataset, is captured by four visible-light cameras as well as two NIR cameras in both outdoor and indoor settings. There are 491 pedestrian identities recorded in this dataset, and each pedestrian is photographed by two different cameras. Furthermore, 11,909 IR images and 22,258 RGB images of 395 identities are contained in the training set. During the testing phase, we perform our experiments on two settings, i.e., indoor search and all search. Each mode has 3803 query IR pedestrian images of 96 identities. Additionally, the gallery set for all-search settings contains 301 randomly selected pedestrian images taken by RGB cameras which are placed in outdoor and indoor environments, while the gallery set for the indoor search settings contains 112 randomly selected pedestrian images taken by RGB cameras which are placed in indoor environments.
RegDB consists of 8240 images from 412 identities, with each identity containing 10 IR images and 10 RGB images. The whole dataset is divided into two halves and used for training and testing, respectively, which the training set includes 2060 IR images and 2060 RGB images of 206 identities. As for the test set, there are 2060 query images of 206 identities and 2060 gallery images of 206 identities. Moreover, two evaluation settings are available, including Thermal-to-Visible (T-V) and Visible-to-Thermal (V-T).

4.2. Evaluation Metrics

The Cumulative Matching Characteristic (CMC) curve is a commonly used performance evaluation metric in the person re-identification task. It plots the probability of correctly matching the query image at different ranks. Specifically, the x-axis represents the rank of the retrieved image (i.e., 1st, 2nd, 3rd, etc.), and the y-axis represents the probability of correctly identifying the query image among the top k retrieved images. A larger area under the CMC curve indicates better performance. The mean Average Precision (mAP) is another commonly used performance evaluation metric in the person re-identification task. It measures the average precision of a set of queries. Specifically, it considers both the precision and recall of the retrieval results. A higher mAP value indicates the better performance.
In this paper, standard CMC and mAP are adopted as evaluation metrics to test the performance of the proposed method.

4.3. Implementation Details

We first resize the pedestrian image into 288 × 144 , then apply random cropping and shuffle a horizontal flip to augment the data. Meanwhile, we set the batch size to 64, where each batch is composed of 4 identities, and each identity consists of 8 RGB pedestrian images and 8 IR pedestrian images. The scalar temperature parameter τ in Equations (8) and (10) is set to 0.2. The hyperparameter β in Equation (4) is set to 2. To balance the importance between different losses, we set the trade-off parameters λ 1 , λ 2 , and λ 3 to 0.5, 0.4, and 0.5, respectively. We use stochastic gradient descent (SGD) optimizer to optimize the proposed method and fix the number of epochs to 60. The preliminary learning rate is set to 0.01 and decayed to 0.001 after 30 epochs. After that, We adopt the FC layers to reduce the local feature dimension to D = 512 , and the number of part-level stripes P is set to 6. In the testing phase, all local features are concatenated as the representation of a pedestrian image.

4.4. Comparisons with State-of-the-Art Methods

Table 1 and Table 2 show the results of the proposed methods on RegDB and SYSU-MM01 compared with the state-of-the-art methods, respectively. The compared methods mainly include the two-stream models (MSR [44], TONE [45], AGW [46]), GAN-based approaches (cmGAN [47], AlignGAN [48], JSIA-ReID [49]), modality aligning approaches (CMAlign [7]), and the metric learning (HCML [45], HSME [50]). In addition, some approaches learn the local features and meanwhile use the constraints between the heterogeneous centers (TSLFN+HC [21], WIT [22]).
Comparisons on SYSU-MM01. From Table 1, we can see that the proposed method achieves 61.89 % of Rank-1 accuracy and 60.12 % of mAP accuracy among the all-search setting, which exceeds NFS [6] and CMAlign [7] in terms of mAP accuracy by 4.67 % and 5.98 % , respectively. It is worth noting that the performance of our method exceeds that of DDAG [16] with respect to Rank-1 and mAP accuracy by 13.0 % and 13.4 % , respectively. This is because DDAG only uses pedestrian images of different modalities as the nodes of a graph, while our method uses the paired local features as the nodes of a graph. Compared with KSD [51], our method outperforms its Rank-1 and mAP accuracy by 1.3 % and 2.3 % , respectively. In addition, under the indoor search settings, the performance of our method surpasses that of WIT [22] by 5.8 % and 5.9 % regarding Rank-1 and mAP accuracy, respectively. This is because WIT uses the center constraint to pull images with the same identity to their cross-modality centers, but ignores constraining the features that are far from the centers. The proposed C 3 L overcomes this shortcoming by constraining the distance between local features and their heterogeneous centers.
The proposed method models the node of a graph with paired local features from different modalities, which outperforms other GAT models, such as DDAG. Furthermore, our approach yields superior performance to the other center-constrained approaches, i.e., TSLFN+HC [21] and WIT [22] on the indoor and all-search settings.
Comparisons on RegDB. We compare our LPGAT model with 13 different methods on the RegDB dataset. From the experimental results in Table 2, LPGAT shows the best performance compared with the other methods. Specifically, The proposed method obtains 89.37 % in Rank-1 accuracy and 78.74 % in mAP accuracy under the V-T mode, which surpasses the second-best method, i.e., WIT [22] with 4.37 % and 2.84 % in Rank-1 accuracy and mAP accuracy, respectively. In addition, under the T-V model, our method outperforms DDAG [16] and NFS [6] by 24 % and 8.4 % in Rank-1 accuracy, respectively, and surpasses them by 19.3 % and 5.7 % in mAP accuracy. Hence, it proves that our model has a strong generalization ability with different scenarios.
In conclusion, the proposed method yields superior performance on the two large-scale datasets, which demonstrates the good generalization capability of our approach.
Table 1. Comparisons with the state-of-the-art methods on SYSU-MM01 with two different settings. R-i indicates Rank-i. The bold indicates the best result.
Table 1. Comparisons with the state-of-the-art methods on SYSU-MM01 with two different settings. R-i indicates Rank-i. The bold indicates the best result.
SettingAll SearchIndoor Search
MethodR-1R-10R-20mAPR-1R-10R-20mAP
Zero-pad [27]14.8054.1271.3315.9520.5868.3885.7926.92
TONE [45]12.5250.7268.6014.4220.8268.8684.4626.38
HCML [45]14.3253.1669.1716.1624.5273.2586.7330.08
cmGAN [47]26.9767.5180.5627.8031.6377.2389.1842.19
HSME [50]20.6862.7477.9523.12----
BDTR [52]27.3266.9681.0727.3231.9277.1889.2841.86
eBDTR [52]27.8267.3481.3428.4232.4677.4289.6242.46
D 2 RL [53]28.9070.6082.4029.20----
MSR [44]37.3583.4093.3438.1139.6489.2997.6650.88
AlignGAN [48]42.4085.0093.7040.7045.9087.6094.4054.30
JSIA-ReID [49]38.1080.7089.9036.9043.8086.2094.2052.90
Xmodal [54]49.9289.7995.9650.73
MACE [10]51.6487.2594.4450.1157.3593.0297.4764.79
DDAG [16]54.7590.3995.8153.0261.0294.0698.4167.98
HAT [24]55.2992.1497.3653.8962.1095.7599.2069.37
TSLFN + HC [21]56.9691.5096.8254.9559.7492.0796.2264.91
DAPR [28]46.0087.9096.0043.9046.2089.2.0096.7055.80
WIT [22]59.2091.7096.5057.3060.7094.1098.4067.10
AGW [46]47.5084.3992.1447.6554.1791.1495.9862.97
FBP-AL [55]54.1486.0493.0350.20----
CMAlign [7]55.41--54.1458.46--66.33
NFS [6]56.9191.3496.5255.4562.7996.5399.0769.79
CPN [56]42.4887.1295.6244.90----
KSD [51]61.0793.1597.8658.7664.0995.7898.8970.57
Ours61.8993.5697.8660.1264.2496.5899.0871.04
Table 2. Comparisons with the state-of-the-art methods on RegDB for visible-infrared and infrared-visible settings. R-i indicates Rank-i. The bold indicates the best result.
Table 2. Comparisons with the state-of-the-art methods on RegDB for visible-infrared and infrared-visible settings. R-i indicates Rank-i. The bold indicates the best result.
SettingV-TT-V
MethodsR-1R-10R-20mAPR-1R-10R-20mAP
Zero-pad [27]17.7434.2144.3518.9016.6334.6844.2517.82
HCML [45]24.4447.5356.7820.0821.7045.0255.5822.24
BDTR [52]33.5658.6167.4332.7632.9258.4668.4331.96
eBDTR [52]34.6258.9668.7233.4634.2158.7468.6432.49
D 2 RL [53]43.4066.1076.3044.10----
MSR [44]48.4370.3279.9548.67----
HSME [50]50.8573.3681.6647.0050.1572.4081.0746.16
AlignGAN [48]57.90--53.60----
JSIA-ReID [49]48.50--49.3048.10--48.90
Xmodal [54]62.2183.1391.7260.18----
DDAG [16]69.3486.1491.4963.4668.0685.1590.3161.80
HAT [24]71.8387.1692.1667.5670.0286.4591.6166.30
MACE [10]72.3788.4093.5969.0972.1288.0793.0768.57
DAPR [28]61.5081.6088.7059.40----
AGW [46]70.10--66.40----
FBP-AL [55]73.9889.7193.6958.2470.0589.2293.8866.61
WIT [22]85.0096.9098.8075.90----
CMAlign [7]74.17--67.6472.43--65.46
NFS [6]80.5491.9695.0772.1077.9590.4593.6269.79
CPN [56]51.2971.1579.7949.37----
KSD [51]76.6690.1993.8469.6373.6489.2293.1067.41
Ours89.3797.6299.0878.7484.5195.8398.0173.75

4.5. Ablation Studies

We conduct ablation experiments on SYSU-MM01 with the all-search set-up to assess the effectiveness of each key component of our method. The detailed results are presented in Table 3. B represents the baseline which is implemented by the Local Feature Extractor and optimized by the HC loss and the cross-entropy loss. G A T indicates that the node of the graph is built by the single local feature of the pedestrian image, and L P G A T -k is the LPGAT module without using the contextual attention coefficient.
From Table 3, several conclusions can be drawn as follows. First, we observe that B + G A T exceeds B with 0.7 % in Rank-1 accuracy and 1.08 % in mAP accuracy, respectively. It indicates that aggregating the local features of different modalities can improve the discrimination of features. Second, B + L P G A T -k improves the performance compared with B + G A T , which demonstrates the effectiveness of using the paired local features from different modalities to build a graph. Third, the comparison between B + L P G A T and B + L P G A T -k proves the effectiveness of the contextual attention coefficient which is beneficial to obtain accurate information propagation. Fourth, B + C 3 L outperforms B by 4.46 % in Rank-1 accuracy and 3.67 % in mAP accuracy, respectively. The proposed C 3 L could facilitate the deep model to learn the completed distance metric by constraining the distance between local features and their heterogeneous centers to narrow the gap of different modalities. Finally, the performance is further improved when combining LPGAT and C 3 L , which demonstrates they could mutually reinforce.

4.6. Parameters Analysis

There are several key parameters in the proposed method. We evaluated the effect of different parameter values on all-search mode in SYSU-MM01, and the experimental results can be generalized to other cross-modality person ReID settings.
The impact of the hyperparameter β . We perform the experiments with different values of β in Equation (4) to evaluate the performance of the proposed method which is shown in Figure 4. From the figure, it can be seen that the performance peaks at β = 2, and it drops as β increases. Therefore, we set β to 2.
The impact of the scalar temperature parameter τ . The scalar temperature parameter τ is an important parameter that controls the range of similarity between the local features and their heterogeneous centers in Equations (8) and (10). The experimental results with different values of τ are shown in Figure 5 where the performance becomes better as τ increases and the performance drops when τ > 0.2 . Hence, the optimal value of τ is 0.2.
The impact of the trade-off parameters λ 1 , λ 2 and λ 3 . The trade-off parameters λ 1 , λ 2 and λ 3 in Equation (12) control the importance of different losses. To search the optimal values of λ 1 , λ 2 and λ 3 , we experimentally test different value combinations of them, and to conveniently display, we fix two parameters with the optimal values and show the influence of the other parameter. The results are shown in Figure 6 where we can see that when λ 1 = 0.5, λ 2 = 0.4 and λ 3 = 0.5 the performance is best.

5. Visualization

To intuitively verify the effectiveness of our method, we visualize the cosine similarity distribution of cross-modality positive and negative pairs (R-I positive and R-I negative) of B, B + L P G A T and B + C 3 L as shown in Figure 7. From the figure, we can see that the distribution of B + L P G A T and B + C 3 L are more separate than that of B. It demonstrates that LPGAT and C 3 L could improve the discrimination of features for cross-modality person ReID.

6. Conclusions

In this paper, we presented LPGAT for cross-modality person ReID to model the correlation between paired local features derived from distinct modalities. Meanwhile, we propose the contextual attention coefficient to ensure accurate information propagation on the graph. In addition, we propose C 3 L to decrease the modality gap for cross-modality person ReID by constraining the distance between local features and their heterogeneous centers. The results of experiments on two commonly used datasets demonstrate that the proposed approach surpasses the state-of-the-art approaches. In future work, we will extend our approach to video sequences for the cross-modal ReID domain. Considering that in practical ReID application scenarios, multiple tasks such as pedestrian attribute recognition and pose estimation often need to be performed simultaneously, the joint learning of multiple tasks will be considered in the future to make full use of multimodal information and improve the performance of pedestrian re-identification.

Author Contributions

Methodology, J.Z., Z.Z. and S.L.; Software, J.Z. and Q.D.; Formal analysis, S.L.; Resources, T.S.D.; Writing—original draft, J.Z.; Writing—review & editing, Q.D., Z.Z., S.L. and T.S.D.; Supervision, T.S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 62171321, Natural Science Foundation of Tianjin under Grant No. 20JCZDJC00180 and 22JCQNJC00010, the Scientific Research Project of Tianjin Educational Committee under Grant No. 2022KJ011 and Graduate Scientific Research Innovation Project of Tianjin.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used for training and evaluating the performance of our proposed approach are publicly available and can be accessed from [27,43].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sun, Y.; Zheng, L.; Yang, Y.; Tian, Q.; Wang, S. Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline). In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 480–496. [Google Scholar]
  2. He, T.; Shen, X.; Huang, J.; Chen, Z.; Hua, X.-S. Partial person re-identification with part-part correspondence learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 9105–9115. [Google Scholar]
  3. Fan, D.; Wang, L.; Cheng, S.; Li, Y. Dual branch attention network for person re-identification. Sensors 2021, 17, 5839. [Google Scholar] [CrossRef]
  4. Zhou, Y.; Liu, P.; Cui, Y.; Liu, C.; Duan, W. Integration of Multi-Head Self-Attention and Convolution for Person Re-Identification. Sensors 2022, 16, 6293. [Google Scholar] [CrossRef]
  5. Tian, X.; Zhang, Z.; Lin, S.; Qu, Y.; Xie, Y.; Ma, L. Farewell to mutual information: Variational distillation for cross-modal person re-identification. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 27–28 February 2021; pp. 1522–1531. [Google Scholar]
  6. Chen, Y.; Wan, L.; Li, Z.; Jing, Q.; Sun, Z. Neural feature search for rgb-infrared person re-identification. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 587–597. [Google Scholar]
  7. Park, H.; Lee, S.; Lee, J.; Ham, B. Learning by aligning: Visible-infrared person re-identification using cross-modal correspondences. In Proceedings of the 2021 International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 12046–12055. [Google Scholar]
  8. Zhao, Q.; Wu, H.; Zhu, J. Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification. Sensors 2023, 23, 1426. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, Z.; Wang, Y.; Liu, S.; Xiao, B.; Durrani, T.S. Cross-domain person re-identification using heterogeneous convolutional network. IEEE TCSVT 2022, 32, 1160–1171. [Google Scholar] [CrossRef]
  10. Ye, M.; Lan, X.; Leng, Q.; Shen, J. Cross-modality person re-identification via modality-aware collaborative ensemble learning. IEEE TIP 2020, 29, 9387–9399. [Google Scholar] [CrossRef]
  11. Cheng, D.; Li, X.; Qi, M.; Liu, X.; Chen, C.; Niu, D. Exploring cross-modality commonalities via dual-stream multi-branch network for infrared-visible person re-identification. IEEE Access 2020, 8, 12824–12834. [Google Scholar] [CrossRef]
  12. Cheng, Y.; Li, X.; Xiao, G.; Ma, W.; Gou, X. Dual-path deep supervision network with self-attention for visible-infrared person re-identification. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems (ISCAS), Daegu, Republic of Korea, 22–28 May 2021; pp. 1–5. [Google Scholar]
  13. Sun, Z.; Zhu, Y.; Song, S.; Hou, J.; Du, S.; Song, Y. The multi-layer constrained loss for cross-modality person re-identification. In Proceedings of the 2020 International Conference on Artificial Intelligence and Signal Processing (AISP), Amaravati, India, 10–12 January 2020; pp. 1–6. [Google Scholar]
  14. Zhang, L.; Du, G.; Liu, F.; Tu, H.; Shu, X. Global-local multiple granularity learning for cross-modality visible-infrared person reidentification. IEEE TNNLS 2021, 36, 1–11. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, T.; Wang, J.; Jiang, K.; Gu, X.; Wan, J. Visible Infrared Person Re-Identification via Global-Level and Local-Level Constraints. IEEE Access 2021, 9, 166339–166350. [Google Scholar] [CrossRef]
  16. Ye, M.; Shen, J.; Crandall, J.D.; Shao, L.; Luo, J. Dynamic dual-attentive aggregation learning for visible-infrared person re-identification. In Proceedings of the 2020 European Conference on Computer Vision (ECCV), Online, 17 July 2019; pp. 229–247. [Google Scholar]
  17. Cheng, Y.; Xiao, G.; Tang, X.; Ma, W.; Gou, X. Two-phase feature fusion network for visible-infrared person re-identification. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1149–1153. [Google Scholar]
  18. Wan, L.; Sun, Z.; Jing, Q.; Chen, Y.; Lu, L.; Li, Z. G2DA: Geometry-guided dual-alignment learning for RGB-infrared person re-identification. arXiv 2021, arXiv:2106.07853. [Google Scholar] [CrossRef]
  19. Zhang, J.; Li, X.; Chen, C.; Qi, M.; Wu, J.; Jiang, J. Global-Local Graph Convolutional Network for cross-modality person re-identification. Neurocomputing 2021, 452, 137–146. [Google Scholar] [CrossRef]
  20. Ye, M.; Wang, Z.; Lan, X.; Yuen, P.C. Visible thermal person re-identification via dual-constrained top-ranking. In Proceedings of the 2018 International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 13–19 July 2018; pp. 1092–1099. [Google Scholar]
  21. Zhu, Y.; Yang, Z.; Wang, L.; Zhao, S.; Hu, X.; Tao, D. Hetero-center loss for cross-modality person re-identification. IEEE TIP 2020, 29, 9387–9399. [Google Scholar] [CrossRef] [Green Version]
  22. Sun, J.; Li, Y.; Chen, H.; Peng, Y.; Zhu, X.; Zhu, J. Visible-infrared cross-modality person re-identification based on whole-individual training. Neurocomputing 2021, 440, 1–11. [Google Scholar] [CrossRef]
  23. Li, W.; Qi, K.; Chen, W.; Zhou, Y. Unified batch all triplet loss for visible-infrared person re-identification. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  24. Ye, M.; Shen, J.; Shao, L. Visible-infrared person re-identification via homogeneous augmented tri-modal learning. IEEE TIFS 2020, 16, 728–739. [Google Scholar] [CrossRef]
  25. Ling, Y.; Luo, Z.; Lin, Y.; Li, S. A Multi-Constraint Similarity Learning with Adaptive Weighting for Visible-Thermal Person Re-Identification. In Proceedings of the 2021 International Joint Conference on Artificial Intelligence (IJCA), Online, 11–17 July 2020; pp. 845–851. [Google Scholar]
  26. Kong, J.; He, Q.; Jiang, M.; Liu, T. Dynamic center aggregation loss with mixed modality for visible-infrared person re-identification. IEEE SPL 2021, 28, 2003–2007. [Google Scholar] [CrossRef]
  27. Wu, A.; Zheng, W.-S.; Yu, H.-X.; Gong, S.; Lai, J. RGB-infrared cross-modality person re-identification. In Proceedings of the 2017 International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5380–5389. [Google Scholar]
  28. Zhang, P.; Wu, Q.; Yao, X.; Xu, J. Beyond modality alignment: Learning part-level representation for visible-infrared person re-identification. IVC 2021, 108, 104–118. [Google Scholar] [CrossRef]
  29. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  30. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the 2017 Neural Information Processing Systems (NIPS), Los Angeles, CA, USA, 13–19 March 2017; pp. 1–11. [Google Scholar]
  31. Li, X.; Zhou, T.; Li, J.; Zhou, Y.; Zhang, Z. Group-wise semantic mining for weakly supervised semantic segmentation. In Proceedings of the 2021 AAAI Conference on Artificial Intelligence, Online, 17–18 May 2021; pp. 1984–1992. [Google Scholar]
  32. Wang, Z.; Zheng, L.; Li, Y.; Wang, S. Linkage based face clustering via graph convolution network. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1117–1125. [Google Scholar]
  33. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  34. Thekumparampil, K.K.; Wang, C.; Oh, S.; Li, L.-J. Attention-based graph neural network for semi-supervised learning. arXiv 2018, arXiv:1803.03735. [Google Scholar]
  35. Zhang, Z.; Zhang, H.; Liu, S. Person re-identification using heterogeneous local graph attention networks. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 27–28 February 2021; pp. 12136–12145. [Google Scholar]
  36. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised contrastive learning. In Proceedings of the 2020 Neural Information Processing Systems (NIPS), Hilton Midtown, NY, USA, 7–20 February 2020; pp. 18661–18673. [Google Scholar]
  37. Wang, P.; Han, K.; Wei, X.-S.; Zhang, L.; Wang, L. Contrastive learning based hybrid networks for long-tailed image classification. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 27–28 February 2021; pp. 943–952. [Google Scholar]
  38. Xie, E.; Ding, J.; Wang, W.; Zhan, X.; Xu, H.; Sun, P.; Li, Z.; Luo, P. Detco: Unsupervised contrastive learning for object detection. In Proceedings of the 2021 International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 8392–8401. [Google Scholar]
  39. Chen, H.; Lagadec, B.; Bremond, F. Ice: Inter-instance contrastive encoding for unsupervised person re-identification. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Montreal, QC, Canada, 10–17 October 2021; pp. 14940–14949. [Google Scholar]
  40. Si, T.; He, F.; Zhang, Z.; Duan, Y. Hybrid contrastive learning for unsupervised person re-identification. IEEE TMM 2022. [Google Scholar] [CrossRef]
  41. Isobe, T.; Li, D.; Tian, L.; Chen, W.; Shan, Y.; Wang, S. Towards discriminative representation learning for unsupervised person re-identification. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 27–28 February 2021; pp. 8526–8536. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Los Angeles, NV, USA, 26–27 June 2016; pp. 770–778. [Google Scholar]
  43. Nguyen, D.T.; Hong, H.G.; Kim, K.W.; Park, K.R. Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 2017, 17, 605. [Google Scholar] [CrossRef]
  44. Feng, Z.; Lai, J.; Xie, X. Learning modality-specific representations for visible-infrared person re-identification. IEEE TIP 2019, 29, 579–590. [Google Scholar] [CrossRef] [PubMed]
  45. Ye, M.; Lan, X.; Li, J.; Yuen, P. Hierarchical discriminative learning for visible thermal person re-identification. In Proceedings of the 2018 AAAI Conference on Artificial Intelligence, Louisiana, LA, USA, 2–7 February 2018; pp. 7501–7508. [Google Scholar]
  46. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep learning for person re-identification: A survey and outlook. IEEE TPAMI 2021, 44, 2872–2893. [Google Scholar] [CrossRef] [PubMed]
  47. Dai, P.; Ji, R.; Wang, H.; Wu, Q.; Huang, Y. Cross-modality person re-identification with generative adversarial training. In Proceedings of the 2018 International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 13–19 July 2018; pp. 677–683. [Google Scholar]
  48. Wang, G.A.; Zhang, T.; Cheng, J.; Liu, S.; Yang, Y.; Hou, Z. RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3623–3632. [Google Scholar]
  49. Wang, G.A.; Zhang, T.; Yang, Y.; Cheng, J.; Chang, J.; Liang, X.; Hou, Z.-G. Cross-modality paired-images generation for RGB-infrared person re-identification. In Proceedings of the 2020 AAAI Conference on Artificial Intelligence, Hilton New York Midtown, NY, USA, 7–12 February 2020; pp. 12144–12151. [Google Scholar]
  50. Hao, Y.; Wang, N.; Li, J.; Gao, X. HSME: Hypersphere manifold embedding for visible thermal person re-identification. In Proceedings of the 2019 AAAI Conference on Artificial Intelligence, Hilton New York Midtown, NY, USA, 15–17 July 2019; pp. 8385–8392. [Google Scholar]
  51. Zhou, Y.; Li, R.; Sun, Y.; Dong, K.; Li, S. Knowledge self-distillation for visible-infrared cross-modality person re-identification. In Proceedings of the 2022 International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 9–11 May 2022; pp. 1–15. [Google Scholar]
  52. Ye, M.; Lan, X.; Wang, Z.; Yuen, P.C. Bi-directional center-constrained top-ranking for visible thermal person re-identification. IEEE TIFS 2020, 15, 407–419. [Google Scholar] [CrossRef]
  53. Wang, Z.; Wang, Z.; Zheng, Y.; Chuang, Y.-Y.; Satoh, S.i. Learning to reduce dual-level discrepancy for infrared-visible person re-identification. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 618–626. [Google Scholar]
  54. Li, D.; Wei, X.; Hong, X.; Gong, Y. Infrared-visible cross-modal person re-identification with an x modality. In Proceedings of the 2020 AAAI Conference on Artificial Intelligence, Hilton New York Midtown, NY, USA, 7–12 February 2020; pp. 4610–4617. [Google Scholar]
  55. Wei, Z.; Yang, X.; Wang, N.; Gao, X. Flexible body partition-based adversarial learning for visible infrared person re-identification. IEEE Trans 2021, 33, 4676–4687. [Google Scholar] [CrossRef]
  56. Liu, J.; Song, W.; Chen, C.; Liu, F. Cross-modality person re-identification via channel-based partition network. In Proceedings of the 2022 International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 9–11 May 2022; pp. 1–13. [Google Scholar]
Figure 1. (a) Most methods for cross-modality person ReID treat each pedestrian image as a node of a graph. (b) The proposed method treats paired local features from different modalities as a node of a graph. We distinguish the pedestrian identities using different colors, with the same color indicating the same pedestrian. Circles and triangles are used to represent the IR and RGB modalities, respectively.
Figure 1. (a) Most methods for cross-modality person ReID treat each pedestrian image as a node of a graph. (b) The proposed method treats paired local features from different modalities as a node of a graph. We distinguish the pedestrian identities using different colors, with the same color indicating the same pedestrian. Circles and triangles are used to represent the IR and RGB modalities, respectively.
Sensors 23 04011 g001
Figure 2. The framework of our approach. We first apply the Local Feature Extractor to obtain the local features from different modalities. Then, we propose the LPGAT module to learn the correlation between the paired local features from different modalities. The same color indicates the same pedestrian, and the circle and the triangle represent IR and RGB modalities, respectively. We also use the proposed C 3 L to optimize the network, which constrains the distance between the local features and their heterogeneous centers.
Figure 2. The framework of our approach. We first apply the Local Feature Extractor to obtain the local features from different modalities. Then, we propose the LPGAT module to learn the correlation between the paired local features from different modalities. The same color indicates the same pedestrian, and the circle and the triangle represent IR and RGB modalities, respectively. We also use the proposed C 3 L to optimize the network, which constrains the distance between the local features and their heterogeneous centers.
Sensors 23 04011 g002
Figure 3. (a) The constraint on the centers of RGB modality and IR modality. However, there are some outliers in the learning process. (b) The proposed C 3 L constrains the local features and their heterogeneous centers, which alleviates the influence of outliers. The red arrows denote the constraints between features, whereas the dotted circles signify outliers. The stars indicate the centers, and the circle and the triangle represent the RGB modality and IR modality, respectively.
Figure 3. (a) The constraint on the centers of RGB modality and IR modality. However, there are some outliers in the learning process. (b) The proposed C 3 L constrains the local features and their heterogeneous centers, which alleviates the influence of outliers. The red arrows denote the constraints between features, whereas the dotted circles signify outliers. The stars indicate the centers, and the circle and the triangle represent the RGB modality and IR modality, respectively.
Sensors 23 04011 g003
Figure 4. The influence of the hyperparameter β in mAP accuracy and Rank-1 accuracy.
Figure 4. The influence of the hyperparameter β in mAP accuracy and Rank-1 accuracy.
Sensors 23 04011 g004
Figure 5. The influence of the scalar temperature parameter τ in Rank-1 accuracy and mAP accuracy.
Figure 5. The influence of the scalar temperature parameter τ in Rank-1 accuracy and mAP accuracy.
Sensors 23 04011 g005
Figure 6. The influence of the trade-off parameters λ 1 , λ 2 and λ 3 in mAP accuracy and Rank-1 accuracy.
Figure 6. The influence of the trade-off parameters λ 1 , λ 2 and λ 3 in mAP accuracy and Rank-1 accuracy.
Sensors 23 04011 g006
Figure 7. The distributions of the cosine similarity scores of positive and negative pairs of (a) B, (b) B + L P G A T and (c) B + C 3 L . The x axis shows the cosine similarity scores between RGB images and IR images, and the y axis shows the frequency statistics of the cosine similarity score.
Figure 7. The distributions of the cosine similarity scores of positive and negative pairs of (a) B, (b) B + L P G A T and (c) B + C 3 L . The x axis shows the cosine similarity scores between RGB images and IR images, and the y axis shows the frequency statistics of the cosine similarity score.
Sensors 23 04011 g007
Table 3. Ablation studies on SYSU-MM01 on the all-search setting. The bold indicates the best result.
Table 3. Ablation studies on SYSU-MM01 on the all-search setting. The bold indicates the best result.
MethodsR-1R-10R-20mAP
B56.2691.4196.5455.16
B + G A T 56.9691.8996.7556.24
B + L P G A T -k58.0192.8397.6756.63
B + L P G A T 59.0793.2597.6857.42
B + C 3 L 60.7293.5397.8358.83
Ours (B + L P G A T + C 3 L )61.8993.5697.8660.12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, J.; Dong, Q.; Zhang, Z.; Liu, S.; Durrani, T.S. Cross-Modality Person Re-Identification via Local Paired Graph Attention Network. Sensors 2023, 23, 4011. https://doi.org/10.3390/s23084011

AMA Style

Zhou J, Dong Q, Zhang Z, Liu S, Durrani TS. Cross-Modality Person Re-Identification via Local Paired Graph Attention Network. Sensors. 2023; 23(8):4011. https://doi.org/10.3390/s23084011

Chicago/Turabian Style

Zhou, Jianglin, Qing Dong, Zhong Zhang, Shuang Liu, and Tariq S. Durrani. 2023. "Cross-Modality Person Re-Identification via Local Paired Graph Attention Network" Sensors 23, no. 8: 4011. https://doi.org/10.3390/s23084011

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop