Next Article in Journal
Quantitative Dynamic Allodynograph—A Standardized Measure for Testing Dynamic Mechanical Allodynia in Chronic Limb Pain
Next Article in Special Issue
Deep-Learning-Aided Evaluation of Spondylolysis Imaged with Ultrashort Echo Time Magnetic Resonance Imaging
Previous Article in Journal
Explainable Risk Prediction of Post-Stroke Adverse Mental Outcomes Using Machine Learning Techniques in a Population of 1780 Patients
Previous Article in Special Issue
Computer Vision Technology for Monitoring of Indoor and Outdoor Environments and HVAC Equipment: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph Sampling-Based Multi-Stream Enhancement Network for Visible-Infrared Person Re-Identification

1
College of Computer and Information Science, Chongqing Normal University, Chongqing 401331, China
2
School of Computer Engineering, Weifang University, Weifang 261061, China
3
School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan 430081, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(18), 7948; https://doi.org/10.3390/s23187948
Submission received: 23 August 2023 / Revised: 11 September 2023 / Accepted: 13 September 2023 / Published: 18 September 2023
(This article belongs to the Special Issue Multi-Modal Data Sensing and Processing)

Abstract

:
With the increasing demand for person re-identification (Re-ID) tasks, the need for all-day retrieval has become an inevitable trend. Nevertheless, single-modal Re-ID is no longer sufficient to meet this requirement, making Multi-Modal Data crucial in Re-ID. Consequently, a Visible-Infrared Person Re-Identification (VI Re-ID) task is proposed, which aims to match pairs of person images from the visible and infrared modalities. The significant modality discrepancy between the modalities poses a major challenge. Existing VI Re-ID methods focus on cross-modal feature learning and modal transformation to alleviate the discrepancy but overlook the impact of person contour information. Contours exhibit modality invariance, which is vital for learning effective identity representations and cross-modal matching. In addition, due to the low intra-modal diversity in the visible modality, it is difficult to distinguish the boundaries between some hard samples. To address these issues, we propose the Graph Sampling-based Multi-stream Enhancement Network (GSMEN). Firstly, the Contour Expansion Module (CEM) incorporates the contour information of a person into the original samples, further reducing the modality discrepancy and leading to improved matching stability between image pairs of different modalities. Additionally, to better distinguish cross-modal hard sample pairs during the training process, an innovative Cross-modality Graph Sampler (CGS) is designed for sample selection before training. The CGS calculates the feature distance between samples from different modalities and groups similar samples into the same batch during the training process, effectively exploring the boundary relationships between hard classes in the cross-modal setting. Some experiments conducted on the SYSU-MM01 and RegDB datasets demonstrate the superiority of our proposed method. Specifically, in the VIS→IR task, the experimental results on the RegDB dataset achieve 93.69% for Rank-1 and 92.56% for mAP.

1. Introduction

Person re-identification (Re-ID) [1,2,3,4,5,6,7] is a complex computer vision task that focuses on matching individuals across non-overlapping camera views. The main objective is to associate images or videos of the same person while maintaining a low recall rate, thereby reducing the likelihood of incorrect matches. Effective Re-ID techniques have significant applications in various domains, such as surveillance, security, and public safety. With the increasing demand for Re-ID, there is a need to match infrared person images captured under challenging lighting conditions with visible person images. Consequently, VI Re-ID [8,9,10,11,12,13] garners significant attention from both the industry and academia.
Besides the intra-modal variations already present in single-modal Re-ID, a key challenge in the VI Re-ID task is how to reduce the modality discrepancy between visible and infrared images of the same identity. Existing research approaches primarily rely on modal transformation methods. These methods generate cross-modal or intermediate-modal images corresponding to person images to convert heterogeneous modalities into a unified modality, thereby reducing the modal discrepancy. Specifically, Generative Adversarial Networks (GANs) [14] and encoder–decoder structures [15,16] are commonly introduced in these methods. However, the transformation from infrared images to visible images is ill-posed, which may introduce additional noise and fail to generate accurate visible images. Moreover, GAN-based models [17] often overlook the relationships between global features or local features of person images in the VI Re-ID task, leading to limited modal adaptability in their methods.
To enhance the method’s adaptability to cross-modal challenges, some recent researchers apply modal-shared feature learning to the VI Re-ID task, which projects visible and infrared images into a specific shared embedding space, achieving cross-modal feature alignment. These approaches can be further divided into global feature learning and local feature learning. Specifically, global feature learning represents a person image as a single feature vector, which is suitable for capturing overall person identity information. On the other hand, local feature learning uses a set of feature vectors based on parts or regions to represent the image, allowing for a more detailed capture of the local features of person images. In addition, the two-stream convolutional neural network architecture is commonly applied in such methods and combined with loss functions (such as identity loss, triplet loss, etc.) for constraint.
Although these methods have achieved good results in alleviating the modality discrepancy, they still have certain limitations: (1) Existing modal-shared feature learning methods typically focus on exploring either global or local feature representations and rarely combine the advantages of both features. Moreover, due to the nature of infrared images, which only contain a single channel reflecting objects’ thermal radiation, key features like color cannot be utilized for cross-modal matching. Directly extracting features from infrared images may also suffer from interference from identity-irrelevant information. (2) These methods all adopt some basic sampling techniques, such as random sampling and uniform sampling, which do not consider the relationships and similarities between samples with different modalities. Additionally, in VI Re-ID tasks, there are limited features that can be extracted when dealing with infrared modality retrieval, resulting in numerous similar features among samples from different classes. Consequently, conventional sampling methods struggle to capture the subtle differences between these similar features.
To address the aforementioned two issues, we propose a novel method named the Graph Sampling-based Multi-stream Enhancement Network (GSMEN) in this paper. It is noteworthy that when humans engage in the visual judgment of infrared surveillance, they heavily rely on contour information. Despite the absence of color and texture features in infrared images, contour and shape information remains clear and visible, as depicted in Figure 1a with the contour image. It is evident that contours exhibit certain cross-modal invariance between visible and infrared images [18]. Additionally, as contours provide a holistic representation of a person rather than localized features, their global features can better capture the characteristic information. This observation motivates us to extend the global features of contours to the local features obtained from modal-shared feature learning, with the aim of enhancing the feature representation capability and reducing the cross-modal discrepancy between visible and infrared modalities. Consequently, the Contour Expansion Module (CEM) is proposed to fuse the contour-enhanced features with local features, resulting in improved matching performance for cross-modal image pairs.
Then, to tackle the challenge of exploring the boundaries between hard classes in VI Re-ID task, an efficient batch sampling technique is introduced, known as the Cross-modality Graph Sampler (CGS). Specifically, CGS involves constructing nearest neighbor relationship graphs for all classes in the visible and infrared modalities at the beginning of each epoch and then combining them. Subsequently, CGS conducts batch sampling by randomly selecting a sample as the anchor and choosing its top-k nearest neighboring samples of different classes, each class containing the same number of S instances, as illustrated in Figure 1b. Therefore, CGS ensures that the samples within a batch are mostly similar to each other, providing informative and challenging examples for discriminative learning. This sampler aims to explore the boundary relationships between hard classes and enhance the discriminative power of the learned model.
In summary, our contributions in the paper are as follows:
  • To enhance feature representation and reduce the cross-modal discrepancy between visible and infrared modalities, we propose the Contour Expansion Module (CEM), which combines the global features of contours with the local features obtained from modal-shared feature learning. To the best of our knowledge, this method represents the first attempt at tackling the VI Re-ID task.
  • To explore the boundaries between hard classes, we introduce the Cross-modality Graph Sampler (CGS). The sampler constructs nearest neighbor relationship graphs separately for the visible and infrared modalities, and then combines them for batch sampling. This sampling strategy ensures that samples within a batch are mostly similar to each other, providing informative and challenging examples for discriminative learning.
  • We conduct experiments on large-scale VI Re-ID datasets, SYSU-MM01, and RegDB. The results demonstrate that our method achieves significant improvements in matching performance and modal adaptability.

2. Related Work

VI Re-ID aims to address not only the challenges of handling intra-modal differences but also the cross-modal disparities arising from heterogeneous images. Therefore, alleviating cross-modal disparities is crucial, as they can exacerbate existing intra-modal differences.
To tackle these challenges, researchers attempt the modality-shared feature learning approach [19,20,21,22,23], which focuses on extracting discriminative and robust features from heterogeneous modalities for the model’s learning process. For instance, Wu et al. [8] introduce the large and challenging benchmark dataset (SYSU-MM01) and propose a deep one-stream zero-padding network for RGB-IR image matching. Additionally, Fu et al. [19] present a cross-modality neural architecture search method to enhance the effectiveness of neural network structures for VI Re-ID tasks. Furthermore, Zheng et al. [24] adopt eight attributes as annotation information in the PAENet to learn detailed semantic attribute information.
In recent years, image generation-based methods [14,25] are also applied to the VI Re-ID task, aiming to narrow the gap between visible and infrared modalities by adopting modality auxiliary information. For this purpose, Dai et al. [13] introduce a framework based on Generative Adversarial Networks (GANs) [14] for cross-modal image generation and propose cmGAN for feature learning. Similarly, Wei et al. [26] propose a comprehensive modality generation module that combines features from different modalities to create a new modality, effectively integrating multi-modal information. Additionally, Lu et al. [25] introduce the Progressive Modality-shared Transformer (PMT), which employs grayscale images as auxiliary modalities to enhance the reliability and commonality of visual features across different modalities, addressing the negative effects of modality disparities.
Furthermore, it is worth noting that among various types of information in cross-modal images (such as color, texture, and contour), contour information is crucial in cross-modal retrieval, as it exhibits strong modality invariance. However, previous researchers did not consider using it as auxiliary information to alleviate modality differences. Therefore, we attempt to introduce it as auxiliary enhancement information to improve the image-matching capability in VI Re-ID tasks.

3. Method

Our VI Re-ID framework is based on a two-stream convolutional neural network [4] but incorporates contour information to enhance the model’s cross-modal adaptability. The method we propose is outlined as shown in Figure 2, and the specific steps are as follows: First, the Cross-modality Graph Sampler (Section 3.1) samples the dataset with the aim of including categories that are close in distance into the same batch. The obtained samples consist of two modalities: visible and infrared modalities, both of which are obtained through contour extraction to produce their respective contour images. Next, these four categories (two types of original images and their corresponding contour images) of images are separately input into their corresponding backbone networks. Finally, the resulting sample features and contour features are fused in the Contour Expansion Module (Section 3.2) to mitigate the differences between the visible and infrared modalities.

3.1. Contour Expansion Module

The goal of contour detection is to identify pixels in the image that correspond to regions with significant changes in grayscale values. Recently, there have been studies [23,25] applying contour detection to object detection and semantic segmentation, achieving success. The challenge in the VI Re-ID task lies in the significant differences between the visible and infrared modalities. Additionally, contour information exhibits strong modality invariance. Hence, this inspired us to apply contour detection to the VI Re-ID task to alleviate the modal discrepancies. First, the pre-trained SCHP (Self-Correction Human Parsing) [29] is adopted as the contour detector to segment person contour maps from the images. Taking the visible image for example, the contour detection of visible image X v i s is below:
X v i s c = σ ( X v i s ) ,
where σ ( ) denotes the contour detector, X v i s c represents the person contour map generated from the visible image X v i s . Then, the obtained contour information needs to be integrated into the original image information. However, the fusion methods [30,31,32] vary in diversity. Therefore, this paper investigates the impact of different fusion methods on the experiments, including Element-wise addition and concatenation, as shown in Figure 3. Specifically, Element-wise addition emphasizes the employment of contour feature to supplement person image-related semantic information, while Element-wise concatenation expands the feature dimension without losing the respective information of person image and contour. Here are the specific fusion methods employed in this study.
Element-wise addition. As shown in Figure 3b,c, we, respectively perform feature addition before the CNN and after Conv Block n  ( 1 n 5 ) . The specific formula is as follows:
F i v i s m = F i v i s + F i v i s c ,
where F i v i s m represent visible feature after merging contour feature F i v i s c to basic visible feature F i v i s . i { R G B ,   conv-1 ,   ,   conv-5 } represents the fusion methods shown in Figure 3b,c. Specifically, F R G B v i s m represents the three-channel features obtained after transforming the RGB image into RGB data.
Element-wise concatenation. As depicted in Figure 3a, we augment the local features by incorporating the global feature of the visible image contours. Firstly, the visible image and the contour image are separately processed through CNN to obtain features F conv-5 v i s c × h × w and F conv-5 v i s c c × h × w , respectively. Next, the output features are subjected to Generalized Mean Pooling for local and global feature pooling, resulting in features F l o c a l v i s c × h / p and F g o b a l v i s c c :
F l o c a l v i s = GEMPooling ( F conv-5 v i s , ( h / p , w ) ) ,
F g o b a l v i s = GEMPooling ( F conv-5 v i s c , ( h , w ) ) ,
where GEMPooling ( z , ( x , y ) ) applies Generalized Mean Pooling [33] to z using a two-dimensional scale with the height of x and width of y. p is the number of local body parts in visible image. Then, 1 × 1 convolutional layers are utilized to adjust the number of feature channels in F l o c a l v i s and F g o b a l v i s c to C. Finally, by concatenating the local feature F v i s of the visible image with its contour image’s global feature F v i s c , the new visible feature F v i s m is obtained:
F v i s m = concat ( F v i s , F v i s c ) ,
where concat ( e , f ) represents the concatenation of feature e and feature f. Considering the comparative experiments in Section 4.5.2, Element-wise concatenation better accomplishes contour enhancement than other ways and is chosen as the fusion method for contour information in CEM.

3.2. Cross-Modality Graph Sampler

Both DG Re-ID [34,35,36,37,38] and VI Re-ID [5,9,10,11,12] face the challenge of modality differences. Furthermore, conventional sampling methods exhibit significant randomness, making it insufficient to distinguish the boundaries between hard classes. In contrast, the CGS sampler effectively addresses this limitation by focusing on grouping similar samples into the same batch. Combining with Figure 4, the details of CGS should be introduced below.
Before each epoch, we calculate the distances or similarities between classes using the latest trained model and then construct a graph encompassing all classes. This approach allows us to leverage the relationships between classes for informative sampling. To illustrate, one image per class is randomly chosen to form a smaller sub-dataset. Next, the feature F v i s m C × d should be extracted thought the latest trained model, where C represents the total number of training classes and d is the feature dimension. Subsequently, the pairwise Euclidean distances between all the selected samples are computed by the feature F v i s m C × d . As a result, we obtain a distance matrix d i s t v C × C that encompasses all classes:
d i s t v = ϕ ( ( F v i s m , F v i s m ) , dim = 0 ) ,
where ϕ ( ( x , y ) , dim = z ) represents the pairwise Euclidean distances calculated for the feature vectors x and y after aligning them along the zth dimension. Similarly, the process is applied to the infrared modality sample set:
d i s t i = ϕ ( ( F i n f m , F i n f m ) , dim = 0 ) ,
where F i n f m C × d is the feature extracted from Infrared samples. Afterwards, to obtain the neighboring classes across different modalities, the overall class distance matrix d i s t can be obtained by adding matrices d i s t v and d i s t i together:
d i s t = d i s t v + d i s t i .
Later, the top k 1 nearest neighboring classes need to be denoted by N ( c ) = { x i | i = 1 ,   ,   k 1 } from each class c , where k is the number of classes to sample in each mini-batch. Subsequently, a graph G = V , E with V = c | c = 1 ,   ,   C represents the vertices, where each class corresponds to one node, and E   = { c 1 ,   c 2   |   c 2 N c 1 } representing the edges. Finally, according to the graph G, we perform random sampling of S instances per class to create a mini-batch containing B = K × S samples for training. This approach allows us to establish connections between classes based on their proximity, enabling informative sampling for our training process.

3.3. Loss Function

Triplet loss L t r i [39] and Cross-entropy loss L i d are the fundamental losses for image classification tasks. Moreover, the Barlow Twins loss L s s d [40] is a self-supervised learning loss function, which is also introduced into our method to improve its performance. The overall loss is computed as follow:
L = L i d + L t r i + L s s d

4. Experiments

4.1. Datasets and Evaluation Protocol

4.1.1. Datasets

SYSU-MM01 [8] dataset consists of 491 different identities captured by four visible cameras and two infrared cameras. It encompasses two search modes: All-Search mode and Indoor-Search mode. Specifically, in the All-Search mode, the gallery set comprises all images captured by the visible cameras, allowing researchers to explore scenarios where all available visible cameras are employed for Re-ID tasks. In contrast, the Indoor-Search mode utilizes images from indoor visible cameras as the gallery set, which is employed for studying Re-ID tasks in indoor environments. The training set comprises 19,659 visible (VIS) images and 1792 infrared (NIR) images, providing a diverse collection of data covering 395 distinct person identities. The test set consists of 3803 infrared images from 96 different person identities, serving as the query set.
RegDB [41] dataset comprises pairs of images captured by visible and infrared cameras. It contains images of 412 different identities, with each identity having 10 visible images and 10 infrared images. These images are captured by a pair of cameras that overlap with each other, providing a comprehensive set of data for evaluation. Additionally, in order to effectively validate various methods, the dataset offers two testing protocols: Infrared-to-Visible (IR-to-VIS) and Visible-to-Infrared (VIS-to-IR).

4.1.2. Evaluation Protocol

To assess the performance of both datasets, we employ standard evaluation protocols, which incorporate Cumulative Matching Characteristics (CMC) [42], and mean Average Precision (mAP) [43] as evaluation metrics. Specifically, we conducted ten tests and computed the average results across these tests.

4.2. Implementation

The proposed methodology utilizes the PyTorch deep learning framework and is implemented on an NVIDIA RTX 3090 GPU. Building upon existing VI Re-ID methods, a pretrained ResNet-50 [28] is employed as a backbone network. During training, all images should be resized to the dimensions of 288 × 144 and data augmentation techniques (random cropping and random horizontal flipping) [44] are introduced.
The training process involves using the stochastic gradient descent (SGD) optimizer with a momentum value of 0.9. The initial learning rate is set to 0.01, and a warm-up strategy is employed to adjust the learning rate. Specifically, the learning rate is initialized to 0.01 and undergoes 10 decays, each occurring every twenty epochs. The training is stopped after 60 epochs. The number p of local body parts in Formula (3) is set to 6.

4.3. Comparison with State-of-the-Art Methods

In this section, a comparison between the proposed method and state-of-the-art VI Re-ID approaches is conducted, including Cross-modal feature learning: DDAG [21], AGW [4], cm-SSFT [45], GLMC [22] (best method), MPANet [12], LBA [46], CM-NAS [19], MMN [47], MID [48]; Modal transformation: cmGAN [13], AlignGAN [49], Xmodal [17], SFANet [50], AGMNet [51] (best method), and PMT [25]. According to Table 1, the results on the two datasets demonstrate that the proposed GSMEN outperforms state-of-the-art methods, achieving the outstanding performance. The specific comparisons are below:
Evaluations on SYSU-MM01. In comparison with the state-of-the-art Cross-modal feature learning method GLMC [22], our approach demonstrates significant superiority in both the All-Search and Indoor-Search modes. Specifically, in the All-Search mode, our method achieves a higher mAP and Rank-1 accuracy by 7.11% and 8.6%, respectively, compared to GLMC. Similarly, in the Indoor-Search mode, our method outperforms GLMC with a mAP and Rank-1 accuracy higher by 6.11% and 11.37%, respectively.
Furthermore, when compared to the current best Cross-modal feature learning method AGMNet [51], our approach also achieves remarkable performance gains. In the All-Search mode, our method surpasses AGMNet with a higher mAP and Rank-1 accuracy by 4.43% and 3.34%, respectively. Similarly, in the Indoor-Search mode, our method outperforms AGMNet with a mAP and Rank-1 accuracy higher by 1.83% and 3.54%, respectively.
Evaluations on RegDB. In comparison with the best-performing method in the VIS-to-IR mode, our approach demonstrates remarkable superiority. Specifically, our method achieves a higher mAP by 4.86% in this mode, showcasing its effectiveness in handling the cross-modal matching between visible and infrared images.
Similarly, in the IR-to-VIS mode, our method outperforms the best-performing method with a significantly higher mAP by 10.67%. This result highlights the exceptional capability of our approach to effectively address the challenges of cross-modal matching between infrared and visible images.
These impressive performance demonstrates the versatility and effectiveness of our proposed method in handling VI Re-ID task.

4.4. Ablation Study

In this section, a comprehensive ablation study is conducted to thoroughly evaluate the contributions of the Contour-oriented Enhancement Module (CEM) and the Cross-modality Graph Sampler (CGS) in our proposed approach. By systematically adding or removing these modules, we investigate their individual impacts on the performance of our model. The results are presented in Table 2, where shows the mAP and Rank-1 accuracy for each experimental setting.
First, we establish a baseline model [4] that comprises basic feature extraction and re-ranking with K-reciprocal [52]. In the subsequent analysis, the All-Search evaluation protocol on the SYSU-MM01 dataset [8] is applied as the benchmark for comparison. The performance of the baseline model achieves a mAP of 70.54% and a Rank-1 accuracy of 72.97%.
In the next step, to assess the influence of the CEM module, we integrate it into the baseline model. The inclusion of the CEM module results in significant performance gains, with mAP and Rank-1 accuracy increasing by 7.74% and 8.51%, respectively. This demonstrates that the CEM module effectively enhances feature representation and reduces modality discrepancy, contributing to the overall improvement in VI Re-ID performance.
Then, the impact of the CGS module is evaluated by incorporating it into the baseline model. The addition of the CGS module also leads to notable performance improvements, with mAP and Rank-1 accuracy increasing by 2.45% and 1.61%, respectively. The CGS module facilitates informative and challenging sample selection, effectively optimizing the training data and further enhancing the model’s discriminative capability.
Finally, we examine the combined effect of both the CEM and CGS modules by integrating them into the baseline model simultaneously. Remarkably, this joint integration yields remarkable performance enhancements, with mAP and Rank-1 accuracy increasing by 10.24% and 9.73%, respectively. The synergistic interplay between CEM and CGS reinforces the feature representation and sample selection aspects, leading to substantial overall improvements in the VI Re-ID task.
The ablation study demonstrates the effectiveness and significance of both the CEM and CGS modules in our proposed approach. The CEM module successfully leverages contour information to enhance feature representation, while the CGS module optimizes the sampling strategy for informative and challenging examples. By understanding the individual contributions of these modules, our study offers valuable insights into the design of a robust and efficient VI Re-ID task.

4.5. Comparison Experiment

4.5.1. Comparison Experiment of Sampling Methods

In our comparative experiments on different sampling methods, namely Random Sampler, Uniform Sampler, and our proposed Cross-modality Graph Sampler (CGS), we observe significant differences in their performances from Figure 5. Specifically, our CGS sampling method outperformed both Random Sampling and Uniform Sampling by a notable margin. The mAP achieved with CGS is 1.81% higher than Random Sampling and 2.41% higher than Uniform Sampling on the SYSU-MM01 dataset with All-Search mode. This result clearly demonstrates the superiority of our CGS sampling approach in improving the overall performance of the VI Re-ID task. Compared to other sampling methods that exhibit randomness, CGS leverages the relationships among classes and ensures that instances within a batch are mostly similar, providing informative and challenging examples for discriminative learning. By incorporating such informative sampling, our method is better able to handle cross-modal challenges and effectively capture subtle differences between similar features, leading to the improved performance observed in the experiments (Table 3).

4.5.2. Comparison Experiment of Fusion Methods

Besides, considering Section 3.1, for the method of fusing contour information, comparative experiments need to be conducted on different feature fusion methods, such as Element-wise addition and Element-wise concatenation (EC). Therefore, some comparative experiments are shown in Figure 5a, which can help determine which fusion method performs best in the task. Specifically, when our model adopts the Element-wise Concatenation fusion method, the mAP and Rank-1 accuracy of 72.97% and 70.54% is achieved, which outperformed other fusion methods and showed the best overall fusion performance. Taking into account the above analysis, the Element-wise Concatenation fusion method is employed for integrating contour information in this paper.

4.5.3. Comparison Experiment of the Contour Detectors

For the selection of contour extraction methods, two options are considered: Canny edge detection and Self-Correction Human Parsing (SCHP). Therefore, a comparative analysis is conducted, and the specific results are shown in Table 4. Specifically, under the Indoor-Search mode, the SCHP method outperformed the Canny edge detection method with a Rank-1 accuracy improvement of 7.93% and an mAP improvement of 5.65%. These results indicate that the SCHP method is better suited for contour extraction in our approach.

4.5.4. Comparison Experiment of the Output Channel

Furthermore, before employing the Element-wise Concatenation fusion method to merge the contour global features with the local features obtained from modal-shared feature learning, the output channel values of the contour global features are also worth our attention. In Figure 5b, we compare the output channel values of contour global features from 1 × 1 Conv. Our model achieves better performance when the output channel is set to 512. This result indicates that setting the output channel to 512 enhances the representation capability of our model for fusing contour information. This finding confirms the significance of selecting an appropriate output channel value for the effective utilization of contour global features and the enhancement of model performance.

4.6. Qualitative Analysis

In this section, we compare our proposed method with the AGW [4] approach using the SYSU-MM01 dataset. For the comparison, two sample images are selected, one depicting the frontal view and the other showing the rear view of individuals, as query samples. The Rank-10 visualization results are presented in Figure 6.
Upon analyzing the results, we observe that our method, with the inclusion of contour information and the utilization of the CGS sampler, effectively improves the retrieval performance. Specifically, in the case of the AGW method, there are instances of two erroneous matches in the rearview image retrieval, where these two images belong to the same class. In contrast, our method achieves correct Rank-10 results for all samples, indicating the superiority of the CGS sampler in distinguishing between similar samples from different classes. Moreover, when using the frontal view as a query sample, the AGW method shows three incorrect matches. These errors can be attributed to the similarity in backgrounds between these samples and the query sample. However, there are noticeable differences in the contour information between them. After enhancing the contour information in our method, only one matching error was observed, showcasing the significant role of contour assistance in enhancing matching capability.
Overall, these qualitative analyses demonstrate that the integration of contour information and the utilization of the CGS sampler effectively address the challenges posed by modal discrepancies and improve the precision and accuracy of the VI Re-ID task.

5. Conclusions

In this paper, we propose the Graph Sampling-based Multi-stream Enhancement Network (GSMEN) for the VI Re-ID task. The GSMEN integrates contour information with the globally shared contour features obtained from modal-shared feature learning. This integration aims to enhance feature representation and reduce cross-modal discrepancy. Our approach introduces the Contour Expansion Module (CEM) for fusing contour-enhanced features with local features and the Cross-modality Graph Sampler (CGS) for effective batch sampling. Experimental results on large-scale datasets demonstrate significant improvements in matching performance and modal adaptability. Our contributions include the novel CEM approach and the efficient CGS sampler, which show promising potential for VI Re-ID in various applications.

Author Contributions

Conceptualization, J.J. and J.X.; Data curation, J.J., J.X. and W.Z.; Formal analysis, J.J., J.X., W.Z., R.R., R.W., T.L. and S.X.; Funding acquisition, W.Z.; Investigation, J.J., J.X., W.Z., R.R., R.W., T.L. and S.X.; Methodology, J.J., W.Z. and R.R.; Project administration, W.Z.; Resources, J.J.; Software, J.J.; Supervision, W.Z. and R.R.; Validation, R.W. and T.L.; Writing—original draft, J.J.; Writing—review and editing, W.Z., R.R. and J.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the following grants and funding sources: Natural Science Foundation of Chongqing (No. 2023NSCQ-MSX1645); Science and Technology Research Program of Chongqing Municipal Education Commission (No. KJQN202200551); Key Project for Science and Technology Research Program of Chongqing Municipal Education Commission (No. KJZD-K202100505); Chongqing Technology Innovation and Application Development Project (No. cstc2020jscx-msxmX0190); Chongqing Normal University Foundation (No. 21XLB026).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used for training and evaluating the performance of our proposed approach are publicly available and can be accessed from [8,39].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, D.; Wang, L.; Cheng, S.; Li, Y. Dual branch attention network for person re-identification. Sensors 2021, 17, 5839. [Google Scholar] [CrossRef] [PubMed]
  2. Zhou, Y.; Liu, P.; Cui, Y.; Liu, C.; Duan, W. Integration of Multi-Head Self-Attention and Convolution for Person Re-Identification. Sensors 2022, 16, 6293. [Google Scholar] [CrossRef] [PubMed]
  3. Zhao, Q.; Wu, H.; Zhu, J. Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification. Sensors 2023, 23, 1426. [Google Scholar] [CrossRef] [PubMed]
  4. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep learning for person re-identification: A survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 2872–2893. [Google Scholar] [CrossRef]
  5. Shu, X.; Li, G.; Wang, X.; Ruan, W.; Tian, Q. Semantic-guided pixel sampling for cloth-changing person re-identification. IEEE Signal Process Lett. 2021, 28, 1365–1369. [Google Scholar] [CrossRef]
  6. Kong, J.; He, Q.; Jiang, M.; Liu, T. Dynamic center aggregation loss with mixed modality for visible-infrared person re-identification. IEEE Signal Process Lett. 2021, 28, 2003–2007. [Google Scholar] [CrossRef]
  7. Zhang, S.; Zhang, L.; Wang, W.; Wu, X. AsNet: Asymmetrical network for learning rich features in person re-identification. IEEE Signal Process Lett. 2020, 27, 850–854. [Google Scholar] [CrossRef]
  8. Wu, A.; Zheng, W.S.; Yu, H.X.; Gong, S.; Lai, J. RGB-infrared cross-modality person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5380–5389. [Google Scholar]
  9. Ye, M.; Ruan, W.; Du, B.; Shou, M.Z. Channel augmented joint learning for visible-infrared recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 13567–13576. [Google Scholar]
  10. Zhao, Z.; Liu, B.; Chu, Q.; Lu, Y.; Yu, N. Joint color-irrelevant consistency learning and identity-aware modality adaptation for visible-infrared cross modality person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; pp. 3520–3528. [Google Scholar]
  11. Zhang, Y.; Wang, H. Diverse Embedding Expansion Network and Low-Light Cross-Modality Benchmark for Visible-Infrared Person Re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–23 June 2023; pp. 2153–2162. [Google Scholar]
  12. Wu, Q.; Dai, P.; Chen, J.; Lin, C.W.; Wu, Y.; Huang, F.; Ji, R. Discover cross-modality nuances for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 27–28 February 2021; pp. 4330–4339. [Google Scholar]
  13. Dai, P.; Ji, R.; Wang, H.; Wu, Q.; Huang, Y. Cross-modality person re-identification with generative adversarial training. In Proceedings of the 2018 International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 677–683. [Google Scholar]
  14. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process Syst. 2014, 27, 53–65. [Google Scholar]
  15. Liang, T.; Jin, Y.; Gao, Y.; Liu, W.; Feng, S.; Wang, T.; Li, Y. Cmtr: Cross-modality transformer for visible-infrared person re-identification. arXiv 2021, arXiv:2110.08994. [Google Scholar]
  16. Jiang, K.; Zhang, T.; Liu, X.; Qian, B.; Zhang, Y.; Wu, F. Cross-modality transformer for visible-infrared person re-identification. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 480–496. [Google Scholar]
  17. Li, D.; Wei, X.; Hong, X.; Gong, Y. Infrared-Visible Cross-Modal Person Re-Identification with an X Modality. In Proceedings of the 2020 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 4610–4617. [Google Scholar]
  18. Ye, M.; Shen, J.; Shao, L. Visible-infrared person re-identification via homogeneous augmented tri-modal learning. IEEE Trans. Inf. Foren. Sec. 2021, 16, 728–739. [Google Scholar] [CrossRef]
  19. Fu, C.; Hu, Y.; Wu, X.; Shi, H.; Mei, T.; He, R. CM-NAS: Cross-modality neural architecture search for visible-infrared person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 11823–11832. [Google Scholar]
  20. Wu, A.; Zheng, W.S.; Gong, S.; Lai, J. RGB-IR person re-identification by cross-modality similarity preservation. Int. J. Comput. Vis. 2020, 128, 1765–1785. [Google Scholar] [CrossRef]
  21. Ye, M.; Shen, J.; Crandall, J.D.; Shao, L.; Luo, J. Dynamic dual-attentive aggregation learning for visible-infrared person re-identification. In Proceedings of the 2020 European Conference on Computer Vision, Online, 17 July 2019; pp. 229–247. [Google Scholar]
  22. Zhang, L.; Du, G.; Liu, F.; Tu, H.; Shu, X. Global-local multiple granularity learning for cross-modality visible-infrared person reidentification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–11. [Google Scholar] [CrossRef]
  23. Wang, G.A.; Zhang, T.; Yang, Y.; Cheng, J.; Chang, J.; Liang, X.; Hou, Z.G. Cross-modality paired-images generation for RGB-infrared person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12144–12151. [Google Scholar]
  24. Zheng, A.; Pan, P.; Li, H.; Li, C.; Luo, B.; Tan, C.; Jia, R. Progressive attribute embedding for accurate cross-modality person re-id. In Proceedings of the 30th ACM International Conference on Multimedia, New York, NY, USA, 10–14 October 2022; pp. 4309–4317. [Google Scholar]
  25. Lu, H.; Zou, X.; Zhang, P. Learning Progressive Modality-Shared Transformers for Effective Visible-Infrared Person Re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; pp. 1835–1843. [Google Scholar]
  26. Wei, Z.; Yang, X.; Wang, N.; Gao, X. Syncretic modality collaborative learning for visible infrared person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 225–234. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Los Angeles, NV, USA, 26–27 June 2016; pp. 770–778. [Google Scholar]
  28. Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
  29. Li, P.; Xu, Y.; Wei, Y.; Yang, Y. Self-correction for human parsing. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 3260–3271. [Google Scholar] [CrossRef]
  30. Kim, S.; Kim, S.; Min, D.; Sohn, K. Laf-net: Locally adaptive fusion networks for stereo confidence estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 205–214. [Google Scholar]
  31. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  32. Karim, S.; Tong, G.; Li, J.; Qadir, A.; Farooq, U.; Yu, Y. Current advances and future perspectives of image fusion: A comprehensive review. Inf. Fusion 2023, 90, 185–217. [Google Scholar] [CrossRef]
  33. Lin, B.; Zhang, S.; Yu, X. Gait recognition via effective global-local feature representation and local temporal aggregation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14648–14656. [Google Scholar]
  34. Liao, S.; Shao, L. Graph sampling based deep metric learning for generalizable person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, CA, USA, 16–20 June 2020; pp. 7359–7368. [Google Scholar]
  35. Jin, X.; Lan, C.; Zeng, W.; Chen, Z.; Zhang, L. Style normalization and restitution for generalizable person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, CA, USA, 16–20 June 2020; pp. 3143–3152. [Google Scholar]
  36. Zheng, K.; Liu, J.; Wu, W.; Li, L.; Zha, Z.J. Calibrated feature decomposition for generalizable person re-identification. arXiv 2021, arXiv:2111.13945. [Google Scholar]
  37. Zhao, Y.; Zhong, Z.; Yang, F.; Luo, Z.; Lin, Y.; Li, S.; Sebe, N. Learning to generalize unseen domains via memory-based multi-source meta-learning for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 27–28 February 2021; pp. 6277–6286. [Google Scholar]
  38. Choi, S.; Kim, T.; Jeong, M.; Park, H.; Kim, C. Meta batch-instance normalization for generalizable person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 27–28 February 2021; pp. 3425–3435. [Google Scholar]
  39. Liu, H.; Tan, X.; Zhou, X. Parameter sharing exploration and hetero-center triplet loss for visible-thermal person re-identification. IEEE Trans. Multimed. 2020, 23, 4414–4425. [Google Scholar] [CrossRef]
  40. Zbontar, J.; Jing, L.; Misra, I.; LeCun, Y.; Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. In Proceedings of the International Conference on Machine Learning, Chongqing, China, 9–11 July 2021; pp. 12310–12320. [Google Scholar]
  41. Nguyen, D.T.; Hong, H.G.; Kim, K.W.; Park, K.R. Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 2017, 17, 605. [Google Scholar] [CrossRef] [PubMed]
  42. Paisitkriangkrai, S.; Shen, C.; Van Den Hengel, A. Learning to rank in person re-identification with metric ensembles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1846–1855. [Google Scholar]
  43. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable person re-identification: A benchmark. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  44. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 13001–13008. [Google Scholar]
  45. Lu, Y.; Wu, Y.; Liu, B.; Zhang, T.; Li, B.; Chu, Q.; Yu, N. Cross-modality person re-identification with shared-specific feature transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, CA, USA, 16–20 June 2020; pp. 13379–13389. [Google Scholar]
  46. Park, H.; Lee, S.; Lee, J.; Ham, B. Learning by aligning: Visible-infrared person re-identification using cross-modal correspondences. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12046–12055. [Google Scholar]
  47. Zhang, Y.; Yan, Y.; Lu, Y.; Wang, H. Towards a unified middle modality learning for visible-infrared person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 788–796. [Google Scholar]
  48. Huang, Z.; Liu, J.; Li, L.; Zheng, K.; Zha, Z. Modality-Adaptive Mixup and Invariant Decomposition for RGB-Infrared Person Re-Identification. arXiv 2022, arXiv:2203.01735. [Google Scholar] [CrossRef]
  49. Wang, G.A.; Zhang, T.; Cheng, J.; Liu, S.; Yang, Y.; Hou, Z. RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3623–3632. [Google Scholar]
  50. Liu, H.; Ma, S.; Xia, D.; Li, S. SFANet: A Spectrum-Aware Feature Augmentation Network for Visible-Infrared Person ReIdentification. IEEE Trans. Neural Netw. Learn. Sys. 2023, 34, 1958–1971. [Google Scholar] [CrossRef] [PubMed]
  51. Liu, H.; Xia, D.; Jiang, W. Towards homogeneous modality learning and multi-granularity information exploration for visible-infrared person re-identification. IEEE J. Sel. Top. Signal Process 2023, 17, 545–559. [Google Scholar] [CrossRef]
  52. Zhong, Z.; Zheng, L.; Cao, D.; Li, S. Re-ranking person re-identification with k-reciprocal encoding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Venice, Italy, 22–29 October 2017; pp. 1318–1327. [Google Scholar]
  53. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Visible images and infrared images utilize the extended contour information obtained through contour detection to alleviate the modality discrepancy. Consequently, it becomes easier to match the same person between the visible and infrared modalities. (b) Different shapes represent different classes in the dataset. The CGS sampler first selects one class as an anchor. Next, it identifies the top-k nearest neighboring classes based on their distances to the anchor class. These selected neighboring classes are then included in the same batch for training.
Figure 1. (a) Visible images and infrared images utilize the extended contour information obtained through contour detection to alleviate the modality discrepancy. Consequently, it becomes easier to match the same person between the visible and infrared modalities. (b) Different shapes represent different classes in the dataset. The CGS sampler first selects one class as an anchor. Next, it identifies the top-k nearest neighboring classes based on their distances to the anchor class. These selected neighboring classes are then included in the same batch for training.
Sensors 23 07948 g001
Figure 2. Framework of GSMEN. (a) CGS samples the samples into N batches separately. (b) The inputs are divided into three categories: visible images, infrared images, and contour images obtained through contour detection from the two modalities. (c) ResNet-50 [27] is introduced as the base backbone network, supplemented with Non-local Attention [28] to enhance feature extraction. (d) The CEM integrates local features from both Visible and Infrared modalities with global feature from the contour modality.
Figure 2. Framework of GSMEN. (a) CGS samples the samples into N batches separately. (b) The inputs are divided into three categories: visible images, infrared images, and contour images obtained through contour detection from the two modalities. (c) ResNet-50 [27] is introduced as the base backbone network, supplemented with Non-local Attention [28] to enhance feature extraction. (d) The CEM integrates local features from both Visible and Infrared modalities with global feature from the contour modality.
Sensors 23 07948 g002
Figure 3. Fusion methods for contour information.
Figure 3. Fusion methods for contour information.
Sensors 23 07948 g003
Figure 4. Framework of the CGS. The feature distances ( d i s t v and d i s t i ) for each class of samples in the visible and infrared modalities are obtained based on the latest trained model. Then, they are merged to obtain the cross-modal distances, denoted as d i s t . Next, the nearest neighbor classes are grouped into the same batch based on the distance d i s t to complete the sampling process.
Figure 4. Framework of the CGS. The feature distances ( d i s t v and d i s t i ) for each class of samples in the visible and infrared modalities are obtained based on the latest trained model. Then, they are merged to obtain the cross-modal distances, denoted as d i s t . Next, the nearest neighbor classes are grouped into the same batch based on the distance d i s t to complete the sampling process.
Sensors 23 07948 g004
Figure 5. Training and testing on SYSU-MM01 dataset with All-Search mode.
Figure 5. Training and testing on SYSU-MM01 dataset with All-Search mode.
Sensors 23 07948 g005
Figure 6. Rank-10 results of AGW and our methods. Training on SYSU-MM01 dataset.
Figure 6. Rank-10 results of AGW and our methods. Training on SYSU-MM01 dataset.
Sensors 23 07948 g006
Table 1. Comparing data (%) between our method and other VI Re-ID methods. Red and bold signify the best result, while blue indicates the second-best result.
Table 1. Comparing data (%) between our method and other VI Re-ID methods. Red and bold signify the best result, while blue indicates the second-best result.
MethodVenueSYSU-MM01RegDB
All-SearchIndoor-SearchVIS to IRIR to VIS
R-1R-10R-20mAPR-1R-10R-20mAPR-1R-10R-20mAPR-1R-10R-20mAP
cmGAN [13]IJCAI 1826.9767.5180.5627.8031.6377.2389.1842.19--------
AlignGAN [49]ICCV 1942.4085.0093.7040.7045.9087.6094.4054.3057.90--53.656.30--53.40
Xmodal [17]AAAI 2049.9289.7995.9650.73--------62.21--60.18
DDAG [21]ECCV 2054.7590.3995.8153.0261.0294.1698.4167.9869.3486.1991.4963.4668.0685.1590.3161.80
AGW [4]TPAMI 2147.5084.3992.1447.6554.1791.1495.9862.97--------
cm-SSFT [45]CVPR 2061.6089.2093.9063.2070.5094.9097.7072.6072.30--72.9071.00--71.70
GLMC [22]TNNLS 2164.3793.9097.5363.4367.3598.1099.7774.0291.8497.8698.9881.4291.1297.8698.6981.06
MPANet [12]CVPR 2170.5896.2198.8068.2476.6498.2199.5780.9582.80--87.7083.70--80.90
LBA [46]ICCV 2155.41--54.1458.46--66.3374.17--67.6472.43--65.46
SFANet [50]CVPR 2165.7492.9897.0560.8371.6096.6099.4580.0576.3191.0294.2768.0070.1585.2489.2763.77
CM-NAS [15]CVPR 2161.9992.8797.2560.0267.0197.0299.3272.9584.5495.1897.8580.3282.5794.5197.3778.31
MMN [47]ACM 2170.6096.2099.0066.9076.2097.2099.3079.6091.6097.7098.9084.1087.5096.0098.1080.50
MID [48]AAAI 2260.2792.90-59.4064.8696.12-70.1287.4595.73-84.8584.2993.44-81.41
AGMNet [51]TNNLS 2369.6396.2798.8266.1174.6897.5199.1478.3088.4095.1096.9481.4585.3494.5697.4881.19
PMT [25]AAAI 2367.5395.3698.6464.9871.6696.7399.2576.5284.83--76.5584.16--75.13
GSMENOurs 72.9798.9399.4370.5478.2297.3899.2480.1393.6998.0499.1392.5691.4198.0998.7792.08
Table 2. Ablation experiment results of our method. Training on SYSU-MM01 dataset. The bold indicates the best result.
Table 2. Ablation experiment results of our method. Training on SYSU-MM01 dataset. The bold indicates the best result.
SettingsAll-SearchIndoor-Search
R-1mAPR-1mAP
B62.7360.8166.3267.01
B + CEM71.2468.5777.0878.26
B + CGS65.1862.4268.5469.94
B + CGS + CEM72.9770.5478.2280.13
Table 3. Comparison experiments of different sampling methods. Training on SYSU-MM01 dataset. The bold indicates the best result.
Table 3. Comparison experiments of different sampling methods. Training on SYSU-MM01 dataset. The bold indicates the best result.
SettingsAll-SearchIndoor-Search
R-1mAPR-1mAP
Random Sampler72.0268.7377.1878.54
Uniform Sampler71.2868.1376.6278.77
Cross-modality Graph Sampler72.9770.5478.2280.13
Table 4. Comparison experiments of different contour detectors. Training on SYSU-MM01 dataset. The bold indicates the best result.
Table 4. Comparison experiments of different contour detectors. Training on SYSU-MM01 dataset. The bold indicates the best result.
SettingsAll-SearchIndoor-Search
R-1mAPR-1mAP
Canny edge detection [53]71.2367.5970.2974.48
SCHP [29]72.9770.5478.2280.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, J.; Xiao, J.; Wang, R.; Li, T.; Zhang, W.; Ran, R.; Xiang, S. Graph Sampling-Based Multi-Stream Enhancement Network for Visible-Infrared Person Re-Identification. Sensors 2023, 23, 7948. https://doi.org/10.3390/s23187948

AMA Style

Jiang J, Xiao J, Wang R, Li T, Zhang W, Ran R, Xiang S. Graph Sampling-Based Multi-Stream Enhancement Network for Visible-Infrared Person Re-Identification. Sensors. 2023; 23(18):7948. https://doi.org/10.3390/s23187948

Chicago/Turabian Style

Jiang, Jinhua, Junjie Xiao, Renlin Wang, Tiansong Li, Wenfeng Zhang, Ruisheng Ran, and Sen Xiang. 2023. "Graph Sampling-Based Multi-Stream Enhancement Network for Visible-Infrared Person Re-Identification" Sensors 23, no. 18: 7948. https://doi.org/10.3390/s23187948

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop