Next Article in Journal
Survey of Software-Implemented Soft Error Protection
Next Article in Special Issue
Bagged Tree and ResNet-Based Joint End-to-End Fast CTU Partition Decision Algorithm for Video Intra Coding
Previous Article in Journal
A Contemporary Review on Utilizing Semantic Web Technologies in Healthcare, Virtual Communities, and Ontology-Based Information Processing Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Visible-Infrared Person Re-Identification: A Comprehensive Survey and a New Setting

1
School of Computer Science and Artificial Intelligence, Wuhan University of Technology, Wuhan 430070, China
2
School of Electronics Engineering and Computer Science, Peking University, Beijing 100091, China
3
School of Computer Science and Information Engineering, Hubei University, Wuhan 430062, China
4
School of Computer Science, Wuhan University, Wuhan 430072, China
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(3), 454; https://doi.org/10.3390/electronics11030454
Submission received: 29 December 2021 / Revised: 27 January 2022 / Accepted: 29 January 2022 / Published: 3 February 2022

Abstract

:
Person re-identification (ReID) plays a crucial role in video surveillance with the aim to search a specific person across disjoint cameras, and it has progressed notably in recent years. However, visible cameras may not be able to record enough information about the pedestrian’s appearance under the condition of low illumination. On the contrary, thermal infrared images can significantly mitigate this issue. To this end, combining visible images with infrared images is a natural trend, and are considerably heterogeneous modalities. Some attempts have recently been contributed to visible-infrared person re-identification (VI-ReID). This paper provides a complete overview of current VI-ReID approaches that employ deep learning algorithms. To align with the practical application scenarios, we first propose a new testing setting and systematically evaluate state-of-the-art methods based on our new setting. Then, we compare ReID with VI-ReID in three aspects, including data composition, challenges, and performance. According to the summary of previous work, we classify the existing methods into two categories. Additionally, we elaborate on frequently used datasets and metrics for performance evaluation. We give insights on the historical development and conclude the limitations of off-the-shelf methods. We finally discuss the future directions of VI-ReID that the community should further address.

1. Introduction

Person re-identification (ReID) is a fundamental building block in various tasks of computer vision, such as intelligent surveillance, video analysis [1], and criminal investigation [2]. With the advancement of intelligent monitoring and the enormous expansion of video data in recent years, conventional human power has been challenging and insufficient to deal with intricate surveillance scenarios. ReID aims at searching for a given individual across disjoint cameras. Numerous algorithms designed for ReID have been proposed with impressive results on some publicly available datasets, e.g., 98.1% and 94.5% Rank-1 accuracy on Market-1501 [3] and DukeMTMC-ReID [4] datasets, respectively [5]. However, the images captured by visible cameras may be unavailable in a dark environment. In such a case, infrared imaging equipment, which does not rely on visible light, should be applied. In 2017, Wu et al. [6] first introduced visible-infrared person re-identification (VI-ReID) and proposed a dataset named SYSU-MM01.
As shown in Figure 1a, for a certain pedestrian, the images of the corresponding identity (ID) should be matched from the other modality set. In addition to the common challenges, e.g., low-resolution, viewpoint change, pose variation, and occlusion, VI-ReID is an effortful problem that encounters additional modality discrepancy due to the significant differences between the two modalities. The two modalities can be considered heterogeneous data, as visible images contain three rich color information channels. In contrast, infrared images only include one channel with near-infrared light intensity information. Additionally, from the aspect of the imaging principle, the two modalities have differences in terms of the wavelength range. Moreover, the datasets are relatively single and small in scale. Some works expanded the VI-ReID datasets, but these datasets cannot be disclosed because of privacy issues.
To improve the practical application ability of VI-ReID, researchers previously achieved remarkable progress on VI-ReID. We divide existing methods into two categories—non-generative-based and generative-based—which were proposed in [7]. As shown in Figure 2a, the non-generative-based model mainly utilizes conventional methods, including feature representation learning and distance metric learning, to maximize the similarity between two images with the same ID and minimize the similarity between two images with different IDs [8,9,10]. In contrast, Figure 2b shows a generative-based model that unifies the modality on the data level, bridging the gap between two heterogeneous modalities [11,12].
To the best of our knowledge, almost all VI-ReID systems evaluated their performance based on the setting as shown in Figure 1a. However, this may not be in line with the actual scene. Taking the visible image V as an example, it may be more similar with some negative visible samples than positive infrared samples. The existing testing setting removes all visible images in the gallery to avoid this challenge. In this paper, we propose a novel testing setting that is closer to the practical scene. As shown in Figure 1b, instead of containing images from only one modality in query and gallery, we simultaneously put visible and infrared images into the query and gallery. This setting makes VI-ReID more challenging. Existing works created various two-stream architectures to learn modality-specific information in order to alleviate the cross-modality discrepancy. However, this kind of two-stream network may not extract effective features of visible and infrared images simultaneously in our new setting. Considering the realistic value of this setting, we believe that researchers should pay more attention to it.
In recent years, many excellent review papers have appeared in ReID. For example, Wang et al. [13] considered four different cross-modality application scenarios: low-resolution, infrared, sketch, text and then analyzed typical approaches. Ye et al. [8] categorized related works into closed-world ReID and open-world ReID, and proposed a strong baseline named AGW. Leng et al. [14] sorted out the papers in open-world ReID based on specific application scenarios. Inspired by them, we conduct a thorough overview for VI-ReID.
Our contributions are threefold:
  • We propose a new testing setting which is closer to practical application scenarios and conduct preliminary experiments to verify the significant challenges of the new setting.
  • We compare VI-ReID with ReID in detail and provide a thorough review of VI-ReID techniques, including datasets and performance metrics.
  • We conclude the necessary components of networks and discuss possible future directions of VI-ReID.

2. Visible-Infrared Person Re-Identification

2.1. ReID vs. VI-ReID

Generally, there is just visible modality in ReID, while VI-ReID contains two modalities: visible and infrared. As all know, visible images have three channels containing rich color information, while infrared images contain intensity information with the red channel only. As shown in Figure 3, there is a noticeable modality gap between visible and infrared images [11]. As the inter-modality discrepancy is substantially greater than the intra-modality discrepancy, bridging the modality gap between the two heterogeneous modalities is a major aspect of VI-ReID research.
For ReID, it only faces challenges of intra-modality, e.g., people’s appearance change, viewpoint change, and occlusion. In contrast, VI-ReID confronts not only the difficulties that appear in ReID, but also the cross-modality discrepancy. The networks designed for ReID are not suitable for VI-ReID since the solutions to intra- and inter-modality discrepancies are completely different.
To our knowledge, the performance gap between VI-ReID and ReID is also large. Ref. [15], for example, achieved 95.7% rank-1 on Market-1501, while the rank-1 of [10] just reached 70.58% on the SYSU-MM01 dataset. The performance of VI-ReID is far lower than that of ReID. However, the VI-ReID is more valuable in practical application scenarios, and we should pay more attention to it.

2.2. A New Testing Setting

In a VI-ReID dataset, V = { X v i } i = 1 N v and T = { X t i } i = 1 N t represent the visible and infrared images, respectively, where N v and N t denote the number of samples in a single modality, respectively. Every image has a corresponding ID label y { Y i } i = 1 N p , where N p denotes the number of IDs. Given a certain image as the query, the purpose of VI-ReID is to match images with the same label from the other modality according to the similarity.
However, this setting is not in line with practical scenarios. Just imagine that, given an image of a criminal who has been escaped for several days, we have to search for him via cross visible and infrared cameras. The off-the-shelf methods may not be useful in such a case. Instead of containing images from only one modality, as shown in Figure 1b, we set the probe P = V p T p and gallery G = V g T g to simultaneously contain visible and infrared images, where ∪ denotes union. V p , V g , and T p , T g are the mutually exclusive subsets of V and T, respectively. When researchers evaluate methods with our new setting, the images with the same ID and modality as the probe cannot appear in the gallery to avoid the impact of ReID in the same modality. When training with this setting, the mainstream dual-branch network structure may not extract effective features because of the effect of mixed modalities. The P and G first generate features F p = { F i p } i = 1 N and F g = { F i g } i = 1 N through the feature extraction module, and then they are matched by the feature matching module.

3. VI-ReID Methods

According to our investigation, there are no other types of articles published on mainstream conferences or journals except those that are deep-learning-based. Hence, only deep-learning-based approaches are included in this review. For the non-generative-based model, we subdivide the model into feature learning, metric learning, and training strategy. For the generative-based model, we subdivide the model into modality translation and extra modality. Besides, we introduce some methods using other technologies. We also summarize some algorithms intended for general ReID or other domains that perform well in VI-ReID. Some methods may be appropriate for multiple categories; however, we will take them to the most suitable position.

3.1. Milestones of Existing VI-ReID Studies

VI-ReID has achieved significant progress in a variety of areas thanks to the unwavering efforts of artificial intelligence researchers. We introduce these crucial milestones for VI-ReID following a timeline and present them in Figure 4. Note that the main basis of a paper selected as a milestone is its citations. We select the paper with highest citations among all papers in a category after dividing the papers into different categories.

3.2. Non-Generative-Based Model

3.2.1. Feature Representation Learning

It aims to extract robust and discriminative features to help the VI-ReID system correctly classify images into different fine-grained classes. We review three kinds of feature representation learning strategies.
Global Feature Representation Learning. As far as we know, most existing methods focus on extracting global features. An illustration is shown in Figure 5a. To obtain modality specific information, Feng et al. [16] established two individual branches for visible and infrared images, respectively. In [17], the authors thought only learning shared features means a massive loss of information, which reduces the difference of features. Therefore, they proposed a cross-modality shared-specific feature transfer algorithm. Ye et al. [18] pointed out that the consistency at the feature and classifier levels is essential when dealing with modality differences. To learn discriminative representations in each modality, Wei et al. [19] developed an attention-lifting mechanism. Wang et al. [20] excavated spatial and channel information of images to reduce the discrepancy between two heterogeneous modalities.
In addition, some works extract ID-invariant features by disentanglement to boost the performance. To achieve more robust retrieval for VI-ReID, Pu et al. [21] disentangled an ID-discriminable and an ID-ambiguous cross-modality feature subspace, respectively. In [22], the authors thought existing methods do not explicitly ignore spectrum information that is not related to VI-ReID. As a result, they disentangled the spectrum information in order to maximize invariant ID information while minimizing the influence of spectrum information. Zhao et al. [23] learned color-irrelevant features through color-irrelevant consistency learning and aligned the ID-level feature distributions by the ID-aware modality adaptation. Hao et al. [24] confused two modalities to learn modality irrelevant representation. In [10], the authors extracted modality irrelevant features by channel attention-guided instance normalization (IN).
Local Feature Representation Learning. As shown in Figure 5b, compared to the global feature, the local feature is more focused on the differences in details. Lin et al. proposed an attribute–person recognition network to make full use of the information contained in attributes [25]. Hao et al. [26] replaced global features with part-level features so that fine-grained camera-invariant information can be extracted. In [27], the authors proposed an adaptive body partition model for automatically detecting and distinguishing effective component representations. Liu et al. [28] presented a network that jointly learns global and local features to cope with viewpoint change and pose variation. Ye et al. [29] excavated contextual cues at the intra-modality components and cross-modality graph levels. Wang et al. [30] utilized global features and partial features to realize the complement of global information and detailed information. To select useful features, Wei et al. [31] designed a flexible body partition module to distinguish part representations automatically. Zhang et al. concatenated the global feature and local feature to create a more powerful feature descriptor [32]. In [33], aiming to eliminate the interference of background information, the authors exploited the knowledge of human body parts to extract robust features. Wu et al. [10] utilized pattern alignment to discover nuances in different patterns. Zhang et al. [34] also made an attempt to discover semantic differences between contrastive features by cross correlation.
Auxiliary Feature Representation Learning. Ye et al. [35] exploited auxiliary information, including the distribution of cross-modality features and contextual information, to bridge the gap between heterogeneous modalities. In [36], the authors designed camera-based batch normalization (BN) to guarantee an invariant input distribution independent of all cameras.

3.2.2. Metric Learning

The purpose of metric learning is to guide feature representation learning. We will go through some prevalent loss functions and training strategies.
Loss Function Design. Generally, researchers design different loss functions to solve targeted problems based on the observed phenomenon. A large cross-modal discrepancy and intra-modal variations generated by varied camera angles, human postures, etc., impact the VI-ReID. As shown in Figure 6a, the function of identity loss is to classify a sample into a correct class in the training phase, which is widely used. For contrastive loss, as shown in Figure 6b, it mainly constrains the training of Siamese networks. For instance, Ye et al. [37] proposed a hierarchical cross-modal matching model, which jointly optimized the modality shared and -specific matrix, aiming at the problem of perspective changing when different cameras record a person. To minimize the difference between same modality and cross similarities, Wu et al. [38] guided the learning of cross-modality similarity by same-modality similarity.
Figure 6c shows the triplet loss, which is contributed to pull the distance between positive sample pairs and push the distance between negative sample pairs. The samples with the same ID form clusters in feature space. The approach constrains the features by a set of triplets to obtain high performance [39]. Wang et al. [40] proposed an improved triplet loss to realize matching a video by an image. Ye et al. [9] proposed a bi-directional dual-constrained top-ranking loss to guide the feature learning objectives. Then, they improved this work by replacing the similarity between two samples with similarity between sample and center [41]. To alleviate the strict constraint of classical triplet loss, Liu et al. [2] proposed an improved triplet loss with the mode of center to center instead of instance to instance. Zhang et al. mitigated the modality discrepancy by mapping the heterogeneous representations into a common space [42]. To learn an angularly separable common feature space, Ye et al. [1] constrained the angles between feature vectors. Cai et al. [43] proposed a dual-modality hard mining triplet-center loss (DTCL) which can reduce computational cost and mine hard triplet samples. In order to eliminate the effect of inconsistent feature distribution in different modalities, Zhang et al. [44] mapped the feature space to angular space and proposed several loss functions to conduct specific angular metric learning.
Figure 6d shows quadruplet loss, which is an improved version of triplet loss. It adds relative distance between the samples with different IDs. In [45], current approaches, according to the authors, primarily combine classification and metric learning to train models in order to generate discriminative and robust representations. However, these methods ignore the relationship between the classification and feature embedding subspaces. The authors presented a hyperspherical manifold-embedded network with classification and recognition constraints based on this information. Jia et al. [46] utilized the similarity transitivity to tackle the problem of mismatching hard positive samples.
Training Strategy. To incorporate different loss functions into an organic whole, researchers have proposed different training strategies. Dai et al. [47] proposed a generative adversarial training strategy to deal with the lack of discriminative information. Ye et al. [48] observed that existing VI-ReID learning strategies ignore the discriminative information of different modalities. Therefore, they presented a modality aware collaborative learning strategy to deal with the gap between two modalities in both the feature level and classifier level. Zhang et al. [49] proposed a mutual learning module that provides a bi-directional transfer between two modalities, aiming at excavating useful information from them. Ling et al. [50] thought most existing methods constrain the similarity of the instance or class level, which is inadequate to make full use of the hidden relationships in cross-modality data. Hence, they proposed multi-constraint similarity learning from instance to instance, instance to class center, and class center to class center. Gao et al. [51] proposed a learning strategy for joint optimization of a single modality and unified modality spaces.

3.3. Generative-Based Model

The generative-based model mainly utilizes generative adversarial network (GAN) or encoder–decoder module to realize the mutual translation between the two modalities. Then, the methods of ReID are used to constrain the appearance of discrepancy.

3.3.1. Modality Translation

In recent years, GAN-based modality translation has gradually become popular. As shown in Figure 7a, modality translation includes infrared to visible and visible to infrared. In contrast, some works disentangled ID-discriminative and ID-excluding factors, and then generated image pairs to extract highly discriminative features.
For infrared to visible, this kind of method can be regarded as image colorization that has been extensively used in various fields [52]. To our knowledge, there is little work in the literature using colorization. Zhong et al. [53] bridged the gap between the two modalities by fusing the features of original infrared images and generated fake visible images. After that, Zhong et al. [11] improved the performance by pixel-wise transformation, which can retain original structure information.
For visible to infrared, Kniaz et al. [54] matched the fake infrared images generated by GAN with the gallery images to mitigate the modality discrepancy. Wang et al. added a pixel alignment module based on feature alignment module [47] to further reduce the gap between the two modalities [12]. However, Liu et al. [55] thought that those methods employing GAN to generate fake images destroy the structure information of generated images and introduce plenty of noise. Hence, they replaced fake images generated by GAN with grayscale images with three channels.
For dual translation, as shown in Figure 7b, it encodes the visible and infrared modalities into a consistent space to eliminate the effect of modality style. It then generates fake cross-modality image pairs with the same ID. In 2019, Wang et al. [56] first generated visible-infrared image pairs by disentanglement and mapped them into a unified space. Analogous to [56], the idea of disentanglement is also indicated in [57,58,59,60]. Among them, Choi et al. [57] encoded the prototype and the attribute separately to generate fake images containing invariant features. Meanwhile, [58,59] acquired visible-infrared image pairs by feature disentanglement and [60] added unseen IDs to generate discriminative features based on [58]. In [61], the network extracted appearance invariant features by generating corresponding fake images.

3.3.2. Extra Modal

Aside from modality translation, some works alleviated the modality discrepancy by introducing an additional third modality. In 2020, Li et al. [62] first introduced an “X” modality as the middle modality to eliminate the cross-modality discrepancy. Subsequently, Huang et al. [63] learned the shared features of images from both modalities to guide the generation of extra images. Ye et al. [64] bridged the gap between the two modalities by generating 3-channel grayscale images. Miao et al. [65] introduced two novel relevant modalities to investigate modality invariant representations. In [66], the authors reduced the cross-modality discrepancy by fusing the two modalities. Wei et al. [67] combined information from visible and infrared images to generate syncretic modality, which can help the network extract modality invariant representations. Zhang et al. [68] projected the images from both modalities into a consolidated subspace to mitigate the modality discrepancy.

3.4. Other Methods

Besides the aforementioned methods, some works also alleviated the impact of a large modality discrepancy by introducing some other technologies. Almost all existing works bridge the gap between the two modalities by manually designing feature extraction modules. Such a manually designed routine usually requires plenty of domain knowledge and practical experience. Therefore, Fu et al. [69] and Chen et al. [70] proposed a cross-modality neural architecture search method and a neural feature search method, respectively, to automatically realize the process of feature extraction. Inspired by the information bottleneck (IB), Tian et al. [71] designed a new strategy that can preserve sufficient label information while simultaneously getting rid of task-irrelevant details. Liang et al. [72] thought that the high cost of labeling person IDs in datasets greatly limits the development of supervised models. Hence, they proposed an unsupervised homogeneous–heterogeneous approach for the unsupervised visible-infrared problem. In [73], the authors used distance metrics instead of a fully connected layer to learn discriminative features. Ye et al. [74] decomposed three channels of visible images and excavated the relationship between each individual channel and infrared image.
In addition, as the tasks of ReID and VI-ReID are identical on the whole, some networks designed for ReID or other similar tasks are also valid on VI-ReID. For instance, Ye et al. [8] proposed a strong baseline for ReID, as it also shows excellent performance on VI-ReID. Jin et al. [75] combined the information removed by IN to achieve high performance. Methods aiming at solving problems of further related domains can also be applied to VI-ReID. For example, Yang et al. [76] proposed an unsupervised graph alignment method that aligns both data representations and distribution structures across the source and target domains, aiming at general cross-domain visual feature representations. The method [77] mitigates the negative effects of noise similarities in cross-modality retrieval by intra-modality distributions. These methods perform excellently on the corresponding tasks; therefore, we can learn from their ideas.

3.5. Summary

From the perspective of method categories, we make the following summaries:
  • Different methods have different strengths. Non-generative-based model are dedicated to mitigating the modality gap on the feature-level (e.g., [9]), while generative-based models pay more attention to the pixel level (e.g., [11]). Compared to non-generative-based model, there is either an information loss or introducing noise in unifying the modality. However, a generative-based model can avoid the impact of color information. A more detailed summary about mainstream works is shown in Table 1.
  • Combining with other techniques is a growing trend. To acquire more discriminative features, some researchers combined this task with some universal techniques (e.g., [70]), and there are also methods (e.g., [71]) that treat this issue from a fresh perspective.
  • The existing setting is not in line with practical application scenarios. In some cases, the modality discrepancy is larger than the differences among IDs. However, the existing testing settings avoid this challenge by putting only single modal images into the gallery.

4. Experimental Results

4.1. Datasets

We first review two prevalent VI-ReID datatsets (RegDB [78] and SYSU-MM01 [6]). Some pedestrian image samples from two datasets are shown in Figure 8.
RegDB [78] contains 412 different IDs, which are classified into 254 females and 158 males, and each ID corresponds to 10 visible images and 10 infrared images. From the samples shown in the first four columns of Figure 8, we can see clear differences between the images captured by two different cameras in terms of color and exposure. Generally, the dataset is randomly split into two halves for training and testing, respectively, according to the evaluation protocol in [37]. In the testing phase, the images from one modality are utilized as a query, while the gallery contains the images from the other modality. The final result is the average of 10 repeated operations.
SYSU-MM01 [6] is a public dataset for VI-ReID proposed in 2017. It contains four cameras for capturing visible images and two for capturing infrared images. Camera 1 and camera 2 are put in two bright rooms, and camera 4 and camera 5 are placed in bright outdoor scenes to capture visible images. Infrared cameras 3 and 6 are placed in a room and outdoor scene, respectively, to capture infrared images without light. There are, in total, 287,628 visible images and 15,792 infrared images of 491 different IDs in SYSU-MM01. As shown in Figure 8, the images in SYSU-MM01 are unpaired in terms of pose, viewpoint, etc.

4.2. Evaluation Metrics

Evaluation metrics play an important role when we want to test the pros and cons of a system. There are two widely used metrics for VI-ReID, named cumulative matching characteristics (CMC) [79] and mean average precision (mAP) [3].
CMC. Rank-r represents the probability that a correct match appears in the top-r search results ranked by confidence. For single shot, this is accurate. However, for multi-shot, CMC [79] cannot accurately represent a model’s discriminability, as it only examines the first match of ranked result.
mAP. The other widely used metric, mAP [3], is a more comprehensive metric for measuring the performance of the VI-ReID algorithm. It reflects how forward all images with the same ID and the probe in the gallery are in the ranked sequence. Therefore, when we face the problem that two algorithms have equal performance in searching the first match, it can address it effectively. However, when a hard sample appears, mAP may still have difficulties evaluating a better one between two algorithms.

4.3. Analysis of the State of the Art with Existing Setting

The performance results of state-of-the-art methods on RegDB and SYSU-MM01 are shown in Table 2 and Table 3, respectively. From the Table 2, we observe that [2] achieves superior performance rank1/mAP 91.05%/83.28% for visible to thermal query setting on RegDB. The main improvement comes from two aspects: replacing global-level features with part-level features and utilizing center-based triplet loss instead of instance-based triplet loss. As the images with the same ID but different modalities from RegDB are entirely aligned, the part-level features are more effective. In contrast, the images are not aligned well on SYSU-MM01. Hence, it is not as big a boost on SYSU-MM01. Moreover, the improvement in the loss function also plays a key role in performance enhancement. There are also some other works committed to this improvement, e.g., Ye et al. [9] adjusted the instance-to-instance-based triplet loss to the instance-to-class-center-based loss, and the performance was significantly improved on SYSU-MM01.
As shown in Table 3, MPANet [10] performs best on SYSU-MM01 [10]. As the infrared modality contains limited information, the difference among the infrared IDs is extremely inconspicuous. Most existing methods deal with the cross-modality discrepancy by proposing novel loss functions or introducing other modalities. In addition to addressing the modality discrepancy, MPANet exploits the nuances among different infrared images to extract more discriminative features.

4.4. Results of the State-of-the-Arts with New Setting

To evaluate the new proposed testing setting, we propose new testing datasets based on RegDB and SYSU-MM01, named RegDB_Mix and SYSU-MM01_Mix, respectively. The reconstructed datasets have the same number of identities and images with original datasets. Rather than putting images of the two modalities into the query and gallery respectively, we mix visible and infrared images, and remove the images in the gallery which have the same modality and identity as the images in the query. We train and test these approaches on a single NVIDIA Tesla P100 GPU. The other settings are consistent with those in the original paper. The multi-shot results on the two datasets are presented in Table 4. Compared to Table 2 and Table 3, we observe that the Rank-1 and mAP both have a great degree of decline. Specially, on RegDB_Mix, the CM-NAS [69] achieves the Rank-1 accuracy of 42.50% and mAP of 41.73%, approximately only half the value of Rank-1 and mAP on RegDB with the visible to infrared mode. On SYSU-MM01_Mix, the CM-NAS achieves the Rank-1 accuracy of 36.48% and mAP of 29.51%, significantly dropping the Rank-1 accuracy by 25.51% and mAP by 30.51% on SYSU-MM01 with all-search and single-shot modes.
To better present the challenges posed by the new setting, we randomly select 10 IDs from the testing set to visualize the distributions of learned features by t-SNE [84]. Here, we choose AGW [8] to extract features of the selected images. As shown in Figure 9, most the features extracted by AGW can be clustered well. However, compared to Figure 9a, many infrared images with different IDs (e.g., blue, red, and yellow) are gathered in Figure 9b. This means that pedestrian images with a certain ID are more likely to be influenced by the images with other different IDs. In fact, this is also more in line with reality, as infrared images contain less information. Hence, they are more difficult to discern by VI-ReID systems.

5. Conclusions and Future Directions

With the increase in functional application requirements, VI-ReID has attracted some researchers’ attention. This paper presents a comprehensive survey of VI-ReID. We first compare it with ReID in detail to show the different challenges of VI-ReID. With powerful deep learning techniques, VI-ReID has achieved remarkable progress, and we divide the existing methods into two categories: non-generative-based and generative-based methods. For the non-generative-based model, we analyze the method in terms of feature learning, metric learning, and training strategy. In contrast, the generative-based model applies modality translation to bridge the modality gap. Finally, we describe standard datasets in detail, evaluation metrics, and performance of the state-of-the-art methods on two datasets.
From Table 2 and Table 3, we observe that the performance of VI-ReID on two public datasets has improved a lot in recent years. Meanwhile, the complexity of networks architecture has also increased. Among the existing network architectures, feature learning and metric learning are the essential modules. The primary function of feature learning is to extract modality specific and -shared features. Recently, some works aiming to extract effective features have become more popular, including global-local features fusion. Here, the distance between two features with the same ID would be pulled, while the distance between two features with different IDs would be pushed by distance metric learning.
From the experimental results, we observe the following directions in VI-ReID:
  • One-stream network architecture. In terms of testing baseline, we believe that the new testing baseline with a more practical setting is more valuable to research than the existing setting. Considering that existing two-stream network architectures cannot validly solve the challenges of the new setting, a one-stream network that can extract more robust and effective features of two heterogeneous modalities may be a trend.
  • Weakly supervised or self-supervised. Considering the difficulties of obtaining a sufficient amount of high-confidence data, we should concentrate on those data with no labels or low label confidence. The approaches, such as those of [85,86] of leveraging this kind of data to address related issues is highly advanced in ReID. We believe that numerous works about weakly supervised or self-supervised data will appear in VI-ReID in the future.
  • Transfer learning. As the number of neural networks grows, the structures become more and more complicated, we expect that the neural network can draw on some current resources when facing comparable tasks. Further research on transfer learning, which has been widely used in ReID [87,88,89], may be a great direction in VI-ReID.
Limitations. First of all, this review draws on the authors’ summary of the literature analysis. Although we aim to be objective in the analysis process, we still cannot avoid a robust subjective tone. Thus, all descriptions are built on personal opinions. Moreover, this survey classifies the networks according to the criteria in [8]. In contrast, we focus on VI-ReID, which accounts for a small percentage of [8]. Finally, this review only covers the research results published in mainstream conferences or journals in this field. The main reason is perhaps that these articles sufficiently represent the research methods and research trends in the area.

Author Contributions

Conceptualization, X.Z. and Z.W.; methodology, H.Z. and X.Z.; software, H.Z.; resources, X.Z.; writing—original draft preparation, H.Z.; writing—review and editing, X.Z., W.H., K.J., W.L. and Z.W.; funding acquisition, X.Z. and W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Department of Science and Technology, Hubei Provincial People’s Government under grants 2021CFB513 and 2021CFB281.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the anonymous reviewers and the editor for their careful reviews and constructive suggestions to help us improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ye, H.; Liu, H.; Meng, F.; Li, X. Bi-Directional Exponential Angular Triplet Loss for visible-Infrared Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 1583–1595. [Google Scholar] [CrossRef] [PubMed]
  2. Liu, H.; Tan, X.; Zhou, X. Parameters Sharing Exploration and Hetero-Center based Triplet Loss for Visible-Thermal Person Re-Identification. IEEE Trans. Multimed. 2020, 23, 4414–4425. [Google Scholar] [CrossRef]
  3. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable Person Re-identification: A Benchmark. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  4. Ristani, E.; Solera, F.; Zou, R.S.; Cucchiara, R.; Tomasi, C. Performance Measures and a Data Set for Multi-target, Multi-camera Tracking. In Proceedings of the 2016 Springer European Conference on Computer Vision Workshops, Amsterdam, The Netherlands, 8–16 October 2016; pp. 17–35. [Google Scholar]
  5. Wang, G.; Lai, J.; Huang, P.; Xie, X. Spatial-Temporal Person Re-Identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 8933–8940. [Google Scholar]
  6. Wu, A.; Zheng, W.; Yu, H.; Gong, S.; Lai, J. RGB-Infrared Cross-Modality Person Re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5390–5399. [Google Scholar]
  7. Baltrusaitis, T.; Ahuja, C.; Morency, L. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 423–443. [Google Scholar] [CrossRef] [Green Version]
  8. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep Learning for Person Re-identification: A Survey and Outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2021. [Google Scholar] [CrossRef] [PubMed]
  9. Ye, M.; Wang, Z.; Lan, X.; Yuen, P.C. Visible Thermal Person Re-Identification via Dual-Constrained Top-Ranking. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 1092–1099. [Google Scholar]
  10. Wu, Q.; Dai, P.; Chen, J.; Lin, C.; Wu, Y.; Huang, F.; Ji, R. Discover Cross-Modality Nuances for Visible-Infrared Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 4330–4339. [Google Scholar]
  11. Zhong, X.; Lu, T.; Huang, W.; Ye, M.; Jia, X.; Lin, C. Grayscale Enhancement Colorization Network for Visible-infrared Person Re-identification. IEEE Trans. Circuits Syst. Video Technol. 2021. [Google Scholar] [CrossRef]
  12. Wang, G.; Zhang, T.; Cheng, J.; Liu, S.; Yang, Y.; Hou, Z. RGB-Infrared Cross-Modality Person Re-Identification via Joint Pixel and Feature Alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3622–3631. [Google Scholar]
  13. Wang, Z.; Wang, Z.; Zheng, Y.; Wu, Y.; Zeng, W.; Satoh, S. Beyond Intra-modality: A Survey of Heterogeneous Person Re-identification. In Proceedings of the 2020 29th IJCAI International Joint Conference Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 4973–4980. [Google Scholar]
  14. Leng, Q.; Ye, M.; Tian, Q. A Survey of Open-World Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1092–1108. [Google Scholar] [CrossRef]
  15. Chen, X.; Fu, C.; Zhao, Y.; Zheng, F.; Song, J.; Ji, R.; Yang, Y. Salience-Guided Cascaded Suppression Network for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3297–3307. [Google Scholar]
  16. Feng, Z.; Lai, J.; Xie, X. Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 579–590. [Google Scholar] [CrossRef]
  17. Lu, Y.; Wu, Y.; Liu, B.; Zhang, T.; Li, B.; Chu, Q.; Yu, N. Cross-Modality Person Re-Identification With Shared-Specific Feature Transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13376–13386. [Google Scholar]
  18. Ye, M.; Lan, X.; Leng, Q.; Shen, J. Cross-Modality Person Re-Identification via Modality-Aware Collaborative Ensemble Learning. IEEE Trans. Image Process. 2020, 29, 9387–9399. [Google Scholar] [CrossRef]
  19. Wei, X.; Li, D.; Hong, X.; Ke, W.; Gong, Y. Co-Attentive Lifting for Infrared-Visible Person Re-Identification. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1028–1037. [Google Scholar]
  20. Wang, H.; Zhao, J.; Zhou, Y.; Yao, R.; Chen, Y.; Chen, S. AMC-Net: Attentive modality-consistent network for visible-infrared person re-identification. Neurocomputing 2021, 463, 226–236. [Google Scholar] [CrossRef]
  21. Pu, N.; Chen, W.; Liu, Y.; Bakker, E.M.; Lew, M.S. Dual Gaussian-based Variational Subspace Disentanglement for Visible-Infrared Person Re-Identification. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2149–2158. [Google Scholar]
  22. Kansal, K.; Subramanyam, A.V.; Wang, Z.; Satoh, S. SDL: Spectrum-Disentangled Representation Learning for Visible-Infrared Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 3422–3432. [Google Scholar] [CrossRef]
  23. Zhao, Z.; Liu, B.; Chu, Q.; Lu, Y.; Yu, N. Joint Color-irrelevant Consistency Learning and Identity-aware Modality Adaptation for Visible-infrared Cross Modality Person Re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; pp. 3520–3528. [Google Scholar]
  24. Hao, X.; Zhao, S.; Ye, M.; Shen, J. Cross-Modality Person Re-Identification via Modality Confusion and Center Aggregation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 16403–16412. [Google Scholar]
  25. Lin, Y.; Zheng, L.; Zheng, Z.; Wu, Y.; Hu, Z.; Yan, C.; Yang, Y. Improving person re-identification by attribute and identity learning. Pattern Recognit. 2019, 95, 151–161. [Google Scholar] [CrossRef] [Green Version]
  26. Hao, Y.; Wang, N.; Gao, X.; Li, J.; Wang, X. Dual-alignment Feature Embedding for Cross-modality Person Re-identification. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 57–65. [Google Scholar]
  27. Wei, Z.; Yang, X.; Wang, N.; Song, B.; Gao, X. ABP: Adaptive Body Partition Model For Visible Infrared Person Re-Identification. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME 2020), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
  28. Liu, S.; Zhang, J. Local Alignment Deep Network for Infrared-Visible Cross-Modal Person Re-identification in 6G-Enabled Internet of Things. IEEE Internet Things J. 2020, 8, 15170–15179. [Google Scholar] [CrossRef]
  29. Ye, M.; Shen, J.; Crandall, D.J.; Shao, L.; Luo, J. Dynamic Dual-Attentive Aggregation Learning for Visible-Infrared Person Re-identification. In Proceedings of the 2020 16th European Conference Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 229–247. [Google Scholar]
  30. Wang, P.; Zhao, Z.; Su, F.; Zhao, Y.; Wang, H.; Yang, L.; Li, Y. Deep Multi-Patch Matching Network for Visible Thermal Person Re-Identification. IEEE Trans. Multimed. 2021, 23, 1474–1488. [Google Scholar] [CrossRef]
  31. Wei, Z.; Yang, X.; Wang, N.; Gao, X. Flexible Body Partition-Based Adversarial Learning for Visible Infrared Person Re-Identification. IEEE Trans. Neural Netw. Learn. Syst. 2021. [Google Scholar] [CrossRef]
  32. Zhang, L.; Du, G.; Liu, F.; Tu, H.; Shu, X. Global-Local Multiple Granularity Learning for Cross-Modality Visible-Infrared Person Reidentification. IEEE Trans. Neural Netw. Learn. Syst. 2021. [Google Scholar] [CrossRef]
  33. Dai, H.; Xie, Q.; Li, J.; Ma, Y.; Li, L.; Liu, Y. Visible-infrared Person Re-identification with Human Body Parts Assistance. In Proceedings of the 2021 International Conference on Multimedia Retrieval, Taipei, Taiwan, 21–24 August 2021; pp. 631–637. [Google Scholar]
  34. Zhang, S.; Yang, Y.; Wang, P.; Liang, G.; Zhang, X.; Zhang, Y. Attend to the Difference: Cross-Modality Person Re-Identification via Contrastive Correlation. IEEE Trans. Image Process. 2021, 30, 8861–8872. [Google Scholar] [CrossRef] [PubMed]
  35. Ye, M.; Cheng, Y.; Lan, X.; Zhu, H. Improving Night-Time Pedestrian Retrieval With Distribution Alignment and Contextual Distance. IEEE Trans. Ind. Inform. 2020, 16, 615–624. [Google Scholar] [CrossRef]
  36. Zhuang, Z.; Wei, L.; Xie, L.; Ai, H.; Tian, Q. Camera-based Batch Normalization: An Effective Distribution Alignment Method for Person Re-identification. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 374–387. [Google Scholar] [CrossRef]
  37. Ye, M.; Lan, X.; Li, J.; Yuen, P.C. Hierarchical Discriminative Learning for Visible Thermal Person Re-Identification. In Proceedings of the 2018 32th AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 7501–7508. [Google Scholar]
  38. Wu, A.; Zheng, W.; Gong, S.; Lai, J. RGB-IR Person Re-identification by Cross-Modality Similarity Preservation. Int. J. Comput. Vis. 2020, 128, 1765–1785. [Google Scholar] [CrossRef]
  39. Ding, S.; Lin, L.; Wang, G.; Chao, H. Deep feature learning with relative distance comparison for person re-identification. Pattern Recognit. 2015, 48, 2993–3003. [Google Scholar] [CrossRef] [Green Version]
  40. Wang, G.; Lai, J.; Xie, X. P2SNet: Can an Image Match a Video for Person Re-Identification in an End-to-End Way? IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 2777–2787. [Google Scholar] [CrossRef]
  41. Ye, M.; Lan, X.; Wang, Z.; Yuen, P.C. Bi-Directional Center-Constrained Top-Ranking for Visible Thermal Person Re-Identification. IEEE Trans. Inf. Forensics Secur. 2020, 15, 407–419. [Google Scholar] [CrossRef]
  42. Zhang, P.; Xu, J.; Wu, Q.; Huang, Y.; Zhang, J. Top-Push Constrained Modality-Adaptive Dictionary Learning for Cross-Modality Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4554–4566. [Google Scholar] [CrossRef]
  43. Cai, X.; Liu, L.; Zhu, L.; Zhang, H. Dual-modality hard mining triplet-center loss for visible infrared person re-identification. Knowl. Based Syst. 2021, 215, 106772. [Google Scholar] [CrossRef]
  44. Zhang, Q.; Lai, J.; Xie, X. Learning Modal-Invariant Angular Metric by Cyclic Projection Network for VIS-NIR Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 8019–8033. [Google Scholar] [CrossRef]
  45. Hao, Y.; Wang, N.; Li, J.; Gao, X. HSME: Hypersphere Manifold Embedding for Visible Thermal Person Re-Identification. In Proceedings of the 2019 33th AAAI Conference on Artificial Intelligence, New Orleans/Honolulu, HI, USA, 27 January–1 February 2019; pp. 8385–8392. [Google Scholar]
  46. Jia, M.; Zhai, Y.; Lu, S.; Ma, S.; Zhang, J. A Similarity Inference Metric for RGB-Infrared Cross-Modality Person Re-identification. In Proceedings of the 2020 29th IJCAI International Joint Conference Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 1026–1032. [Google Scholar]
  47. Dai, P.; Ji, R.; Wang, H.; Wu, Q.; Huang, Y. Cross-Modality Person Re-Identification with Generative Adversarial Training. In Proceedings of the 2018 27th IJCAI International Joint Conference Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 677–683. [Google Scholar]
  48. Ye, M.; Lan, X.; Leng, Q. Modality-aware Collaborative Learning for Visible Thermal Person Re-Identification. In Proceedings of the 2019 27th ACM International Conference Multimedia, Nice, France, 21–25 October 2019; pp. 347–355. [Google Scholar]
  49. Zhang, Z.; Wang, S. Visible Thermal Person Re-identification via Mutual Learning Convolutional Neural Network in 6G-Enabled Visual Internet of Things. IEEE Internet Things J. 2020, 8, 15259–15266. [Google Scholar] [CrossRef]
  50. Ling, Y.; Luo, Z.; Lin, Y.; Li, S. A Multi-Constraint Similarity Learning with Adaptive Weighting for Visible-Thermal Person Re-Identification. In Proceedings of the 2021 30th IJCAI International Joint Conference Artificial Intelligence, Montreal, QC, Canada, 19–27 August 2021; pp. 845–851. [Google Scholar]
  51. Gao, Y.; Liang, T.; Jin, Y.; Gu, X.; Liu, W.; Li, Y.; Lang, C. MSO: Multi-Feature Space Joint Optimization Network for RGB-Infrared Person Re-Identification. In Proceedings of the ACM International Conference Multimedia, ChengDu, China, 20–24 October 2021; pp. 5257–5265. [Google Scholar]
  52. Anwar, S.; Tahir, M.; Li, C.; Mian, A.; Khan, F.S.; Muzaffar, A.W. Image Colorization: A Survey and Dataset. arXiv 2020, arXiv:2008.10774. [Google Scholar]
  53. Zhong, X.; Lu, T.; Huang, W.; Yuan, J.; Liu, W.; Lin, C. Visible-infrared Person Re-identification via Colorization-based Siamese Generative Adversarial Network. In Proceedings of the 2020 ACM International Conference Multimedia Retrieval, Dublin, Ireland, 8–11 June 2020; pp. 421–427. [Google Scholar]
  54. Kniaz, V.V.; Knyaz, V.A.; Hladuvka, J.; Kropatsch, W.G.; Mizginov, V. ThermalGAN: Multimodal Color-to-Thermal Image Translation for Person Re-identification in Multispectral Dataset. In Proceedings of the 2018 European Conference Computer Vision Workshops, Munich, Germany, 8–14 September 2018; pp. 606–624. [Google Scholar]
  55. Liu, H.; Ma, S.; Xia, D.; Li, S. SFANet: A Spectrum-Aware Feature Augmentation Network for Visible-Infrared Person Re-Identification. IEEE Trans. Neural Netw. Learn. Sys. 2021. [Google Scholar] [CrossRef]
  56. Wang, Z.; Wang, Z.; Zheng, Y.; Chuang, Y.; Satoh, S. Learning to Reduce Dual-Level Discrepancy for Infrared-Visible Person Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference Computer Vision Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 618–626. [Google Scholar]
  57. Choi, S.; Lee, S.; Kim, Y.; Kim, T.; Kim, C. Hi-CMD: Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification. In Proceedings of the2020 IEEE/CVF Conference Computer Vision Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10254–10263. [Google Scholar]
  58. Wang, G.; Zhang, T.; Yang, Y.; Cheng, J.; Chang, J.; Liang, X.; Hou, Z. Cross-Modality Paired-Images Generation for RGB-Infrared Person Re-Identification. In Proceedings of the 2020 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12144–12151. [Google Scholar]
  59. Hu, B.; Liu, J.; Zha, Z.J. Adversarial Disentanglement and Correlation Network for RGB-Infrared Person Re-Identification. In Proceedings of the 2021 IEEE International Conference on Multimedia Expo, Shenzhen, China, 5–9 July 2021. [Google Scholar]
  60. Wang, G.; Yang, Y.; Zhang, T.; Cheng, J.; Hou, Z.; Tiwari, P.; Pandey, H.M. Cross-modality Paired-images Generation and Augmentation for Visible-infrared Person Re-identification. Neural Netw. 2020, 128, 294–304. [Google Scholar] [CrossRef]
  61. Chen, Y.; Zhang, S.; Qi, Z. MAENet: Boosting Feature Representation for Cross-Modal Person Re-Identification with Pairwise Supervision. In Proceedings of the2020 ACM International Conference Multimedia Retrieval, Dublin, Ireland, 8–11 June 2020; pp. 442–449. [Google Scholar]
  62. Li, D.; Wei, X.; Hong, X.; Gong, Y. Infrared-Visible Cross-Modal Person Re-Identification with an X Modality. In Proceedings of the 2020 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 4610–4617. [Google Scholar]
  63. Huang, Y.; Wu, Q.; Xu, J.; Zhong, Y.; Zhang, P.; Zhang, Z. Alleviating Modality Bias Training for Infrared-Visible Person Re-Identification. IEEE Trans. Multimed. 2021. [Google Scholar] [CrossRef]
  64. Ye, M.; Shen, J.; Shao, L. Visible-Infrared Person Re-Identification via Homogeneous Augmented Tri-Modal Learning. IEEE Trans. Inf. Forensics Secur. 2021, 16, 728–739. [Google Scholar] [CrossRef]
  65. Miao, Z.; Liu, H.; Shi, W.; Xu, W.; Ye, H. Modality-aware Style Adaptation for RGB-Infrared Person Re-Identification. In Proceedings of the 30th IJCAI International Joint Conference Artificial Intelligence, Montreal, QC, Canada, 19–27 August 2021; pp. 916–922. [Google Scholar]
  66. Ling, Y.; Zhong, Z.; Luo, Z.; Rota, P.; Li, S.; Sebe, N. Class-Aware Modality Mix and Center-Guided Metric Learning for Visible-Thermal Person Re-Identification. In Proceedings of the 28th ACM International Conference Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 889–897. [Google Scholar]
  67. Wei, Z.; Yang, X.; Wang, N.; Gao, X. Syncretic Modality Collaborative Learning for Visible Infrared Person Re-Identification. In Proceedings of the 2021 IEEE/CVF International Conference Computer Vision, Virtual, 11–17 October 2021. [Google Scholar]
  68. Zhang, Y.; Yan, Y.; Lu, Y.; Wang, H. Towards a Unified Middle Modality Learning for Visible-Infrared Person Re-Identification. In Proceedings of the 2021 ACM International Conference Multimedia, ChengDu, China, 20–24 October 2021; pp. 788–796. [Google Scholar]
  69. Fu, C.; Hu, Y.; Wu, X.; Shi, H.; Mei, T.; He, R. CM-NAS: Cross-Modality Neural Architecture Search for Visible-Infrared Person Re-Identification. In Proceedings of the 2021 IEEE/CVF International Conference Computer Vision, Virtual, 11–17 October 2021. [Google Scholar]
  70. Chen, Y.; Wan, L.; Li, Z.; Jing, Q.; Sun, Z. Neural Feature Search for visible-Infrared Person Re-Identification. In Proceedings of the 2021 IEEE/CVF Conference Computer Vision Pattern Recognition, Virtual, 19–25 June 2021. [Google Scholar]
  71. Tian, X.; Zhang, Z.; Lin, S.; Qu, Y.; Xie, Y.; Ma, L. Farewell to Mutual Information: Variational Distillation for Cross-Modal Person Re-Identification. In Proceedings of the 2021 IEEE/CVF Conference Computer Vision Pattern Recognition, Virtual, 19–25 June 2021. [Google Scholar]
  72. Liang, W.; Wang, G.; Lai, J.; Xie, X. Homogeneous-to-Heterogeneous: Unsupervised Learning for RGB-Infrared Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 6392–6407. [Google Scholar] [CrossRef] [PubMed]
  73. Tekeli, N.; Can, A.B. Distance Based Training for Cross-Modality Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference Computer Vision Workshops, Seoul, Korea, 27–28 October 2019; pp. 4540–4549. [Google Scholar]
  74. Ye, M.; Ruan, W.; Du, B.; Mike, Z. Channel Augmented Joint Learning for Visible-Infrared Recognition. In Proceedings of the 2021 IEEE/CVF International Conference Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  75. Jin, X.; Lan, C.; Zeng, W.; Chen, Z.; Zhang, L. Style Normalization and Restitution for Generalizable Person Re-Identification. In Proceedings of the 2020 IEEE/CVF Conference Computer Vision Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3140–3149. [Google Scholar]
  76. Yang, B.; Yuen, P.C. Cross-Domain Visual Representations via Unsupervised Graph Alignment. In Proceedings of the 2019 33th AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 5613–5620. [Google Scholar]
  77. Yang, F.; Wang, Z.; Xiao, J.; Satoh, S. Mining on Heterogeneous Manifolds for Zero-Shot Cross-Modal Image Retrieval. In Proceedings of the 2020 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12589–12596. [Google Scholar]
  78. Nguyen, D.T.; Hong, H.G.; Kim, K.W.; Park, K.R. Person Recognition System Based on a Combination of Body Images from Visible Light and Thermal Cameras. Sensors 2017, 17, 605. [Google Scholar] [CrossRef]
  79. Wang, X.; Doretto, G.; Sebastian, T.; Rittscher, J.; Tu, P.H. Shape and Appearance Context Modeling. In Proceedings of the 2007 IEEE/CVF International Conference Computer Vision, Rio de Janeiro, Brazil, 14–20 October 2007; pp. 1–8. [Google Scholar]
  80. Liu, H.; Cheng, J.; Wang, W.; Su, Y.; Bai, H. Enhancing the discriminative feature learning for visible-thermal cross-modality person re-identification. Neurocomputing 2020, 398, 11–19. [Google Scholar] [CrossRef] [Green Version]
  81. Fan, X.; Luo, H.; Zhang, C.; Jiang, W. Cross-Spectrum Dual-Subspace Pairing for RGB-infrared Cross-Modality Person Re-Identification. arXiv 2020, arXiv:2003.00213. [Google Scholar]
  82. Park, H.; Lee, S.; Lee, J.; Ham, B. Learning by Aligning: Visible-Infrared Person Re-identification using Cross-Modal Correspondences. arXiv 2021, arXiv:2108.07422. [Google Scholar]
  83. Zhu, Y.; Yang, Z.; Wang, L.; Zhao, S.; Hu, X.; Tao, D. Hetero-Center loss for cross-modality person Re-identification. Neurocomputing 2020, 386, 97–109. [Google Scholar] [CrossRef] [Green Version]
  84. Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  85. Wang, G.; Wang, G.; Zhang, X.; Lai, J.; Yu, Z.; Lin, L. Weakly Supervised Person Re-ID: Differentiable Graphical Learning and a New Benchmark. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2142–2156. [Google Scholar] [CrossRef]
  86. Wang, G.; Wang, K.; Wang, G.; Torr, P.; Lin, L. Solving Inefficiency of Self-supervised Representation Learning. In Proceedings of the 2021 IEEE/CVF International Conference Computer Vision, Virtual, 11–17 October 2021. [Google Scholar]
  87. Lin, Y.; Wu, Y.; Yan, C.; Xu, M.; Yang, Y. Unsupervised Person Re-identification via Cross-Camera Similarity Exploration. IEEE Trans. Image Process. 2020, 29, 5481–5490. [Google Scholar] [CrossRef]
  88. Wu, Y.; Lin, Y.; Dong, X.; Yan, Y.; Bian, W.; Yang, Y. Progressive Learning for Person Re-Identification with One Example. IEEE Trans. Image Process. 2019, 28, 2872–2881. [Google Scholar] [CrossRef] [PubMed]
  89. Wu, Y.; Lin, Y.; Dong, X.; Yan, Y.; OuYang, W.; Yang, Y. Exploit the Unknown Gradually: One-Shot Video-Based Person Re-Identification by Stepwise Learning. In Proceedings of the 2018 IEEE/CVF Conference Computer Vision Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5177–5186. [Google Scholar]
Figure 1. Comparison of two testing settings. Images with the same color of bounding boxes denote the same person identity. (a) The query and gallery only contain images from single modality. (b) Both the query and gallery contain images from two modalities.
Figure 1. Comparison of two testing settings. Images with the same color of bounding boxes denote the same person identity. (a) The query and gallery only contain images from single modality. (b) Both the query and gallery contain images from two modalities.
Electronics 11 00454 g001
Figure 2. Illustration of two different pipelines of VI-ReID.
Figure 2. Illustration of two different pipelines of VI-ReID.
Electronics 11 00454 g002
Figure 3. Comparison of visible three-channel brightness-gradient histograms of visible-infrared image pairs with the same ID on RegDB dataset.
Figure 3. Comparison of visible three-channel brightness-gradient histograms of visible-infrared image pairs with the same ID on RegDB dataset.
Electronics 11 00454 g003
Figure 4. Milestones of existing methods of VI-ReID. Since the problem was proposed, researchers have proposed various methods to bridge the modality discrepancy. Note that the top row and bottom row denote a generative-based model and non-generative-based model, respectively.
Figure 4. Milestones of existing methods of VI-ReID. Since the problem was proposed, researchers have proposed various methods to bridge the modality discrepancy. Note that the top row and bottom row denote a generative-based model and non-generative-based model, respectively.
Electronics 11 00454 g004
Figure 5. Two alternative ways for learning feature representations. (a) Global feature, learning modality specific and -shared feature representations. (b) Local feature, learning part-aggregated local feature.
Figure 5. Two alternative ways for learning feature representations. (a) Global feature, learning modality specific and -shared feature representations. (b) Local feature, learning part-aggregated local feature.
Electronics 11 00454 g005
Figure 6. Four widely used loss functions. Different shapes denote different IDs, while the different color represents different modality. (a) Identity loss, (b) contrastive loss, (c) triplet loss, (d) quadruplet loss. Many works employ their combinations.
Figure 6. Four widely used loss functions. Different shapes denote different IDs, while the different color represents different modality. (a) Identity loss, (b) contrastive loss, (c) triplet loss, (d) quadruplet loss. Many works employ their combinations.
Electronics 11 00454 g006
Figure 7. Three kinds of widely used methods in the literature. (a) Infrared to visible [11,53] and visible to infrared [12,54]; (b) more works generate visible-infrared image pairs, employing their combination [56,57,58,59,60].
Figure 7. Three kinds of widely used methods in the literature. (a) Infrared to visible [11,53] and visible to infrared [12,54]; (b) more works generate visible-infrared image pairs, employing their combination [56,57,58,59,60].
Electronics 11 00454 g007
Figure 8. Pedestrian image samples derived from RegDB [78] and SYSU-MM01 [6]. Each column denotes the same ID, and the top row and bottom row represent visible and infrared images, respectively.
Figure 8. Pedestrian image samples derived from RegDB [78] and SYSU-MM01 [6]. Each column denotes the same ID, and the top row and bottom row represent visible and infrared images, respectively.
Electronics 11 00454 g008
Figure 9. Visualization of the feature extracted with AGW [8] distributions. A total of 10 IDs are randomly selected from the testing set of SYSU-MM01. Here, samples with the same color indicate they are of the same person. The markers “circle” and “square” represent the images from infrared and visible modalities, respectively.
Figure 9. Visualization of the feature extracted with AGW [8] distributions. A total of 10 IDs are randomly selected from the testing set of SYSU-MM01. Here, samples with the same color indicate they are of the same person. The markers “circle” and “square” represent the images from infrared and visible modalities, respectively.
Electronics 11 00454 g009
Table 1. A summary of non-generative-based and generative-based models.
Table 1. A summary of non-generative-based and generative-based models.
TypeStrengthWeaknessReference
Non-Generative
Global Feature LearningModality specific and -shared featureMiss some important nuances[16,17,23]
Local Feature LearningWell-aligned part featuresSensitive to noise[26,29,30]
Contrastive LossIncrease inter-class variance of classifierGenerally used for Siamese network[37,38]
Triplet LossForm clustersLack of sufficient constraints to inter-class[1,9,41]
Generative
Single-TranslationAvoid color effectsIntroduce noisy or miss some information[11,12]
Dual-TranslationNeglect ID-irrelevant informationIntroduce noise[56,57]
Extra ModalPull the distance between two modalitiesMiss some information[62,64]
OthersIndependent of manual designLarge time cost[69,70]
Table 2. Rank-r accuracy (%) and mAP (%) performance of state-of-the-art methods on RegDB. Bold numbers are the best results.
Table 2. Rank-r accuracy (%) and mAP (%) performance of state-of-the-art methods on RegDB. Bold numbers are the best results.
ApproachVenueVisible to InfraredInfrared to Visible
r = 1r = 10r = 20mAPr = 1r = 10r = 20mAP
Zero-Padding [6]ICCV’1717.7534.2144.3518.9016.6334.6844.2517.82
HCML [37]AAAI’1824.4447.5356.7820.0821.7045.0255.5822.24
BDTR [9]IJCAI’1833.4758.4267.5231.8332.7257.9668.8631.10
MAC [48]ACM MM’1936.4362.3671.6337.0336.2061.6870.9936.63
D2RL [56]CVPR’1943.4066.1076.3044.10----
HSME [45]AAAI’1950.8573.3681.6647.0050.1572.4081.0746.16
AlignGAN [12]ICCV’1957.90--53.6056.30--53.40
DFE [26]ACM MM’1970.1386.3291.9669.1467.9985.5691.4166.70
eBDTR [41]TIFS’2034.6258.9668.7233.4634.2158.7468.6432.49
MSR [16]TIP’2048.4370.3279.9548.67----
JSIA-ReID [58]AAAI’2048.50--49.3048.10--48.90
EDFL [80]Neurocomputing’2052.5872.1081.4752.9851.8972.0981.0452.13
XIV-ReID [62]AAAI’2062.2183.1391.7260.18----
FMSP [38]IJCV’2065.0783.71-64.50----
DDAG [29]ECCV’2069.3486.1991.4963.4668.0685.1590.3161.80
Hi-CMD [57]CVPR’2070.9386.39-66.04----
cm-SSFT [17]CVPR’2072.30--72.9071.00--71.70
MACE [18]TIP’2072.3788.4093.5969.0972.1288.0793.0768.57
DG-VAE [21]ACM MM’2072.9786.89-71.78----
CoAL [19]ACM MM’2074.1290.2394.5369.87----
SIM [46]IJCAI’2074.47--75.2975.24--78.30
CDP [81]TIP’2165.0083.5089.6062.70----
expAT [1]TIP’2166.48--67.3167.45--66.51
CPN [44]TIP’2168.5984.8198.3369.20----
AGW [8]TPAMI’2170.05--66.37----
HAT [64]TIFS’2171.8387.1692.1667.5670.0286.4591.6166.30
FMI [71]CVPR’2173.20--71.6071.80--70.10
MSO [51]ACM MM’2173.6088.60-66.9074.6088.70-67.50
LbA [82]ICCV’2174.17--67.6472.43--65.46
SFANet [55]TNNLS’2176.3191.0294.2768.0070.1585.2489.2763.77
CICL [23]AAAI’2178.80--69.4077.90--69.40
MCLNet [24]ICCV’2180.3192.7096.0373.0775.9390.9394.5969.49
NFS [70]CVPR’2180.5491.9695.0772.1077.9590.4593.6269.79
GECNet [11]TCSVT’2182.3392.7295.4978.4578.9391.9995.4475.58
MPANet [10]CVPR’2183.70--80.9082.80--80.70
SMCL [67]ICCV’2183.93--79.8383.05--78.57
CM-NAS [69]CVPR’2184.5495.1897.8580.3282.5794.5197.3778.31
CAJ [74]ICCV’2185.0395.4997.5479.1484.7595.3397.5177.82
MPMN [30]TMM’2186.5696.6898.2882.9184.6295.5197.3379.49
HCTL [2]TMM’2191.0597.1698.5783.2889.3096.4198.1681.46
AMC-Net [20]Neurocomputing’2191.2198.1699.2281.6189.0397.6299.2779.85
MMN [68]ACM MM’2191.6097.7098.9084.1087.5096.0098.1080.50
Table 3. Rank-r accuracy (%) and mAP (%) performances of state-of-the-art methods on SYSU-MM01. Bold numbers are the best results.
Table 3. Rank-r accuracy (%) and mAP (%) performances of state-of-the-art methods on SYSU-MM01. Bold numbers are the best results.
ApproachAll SearchIndoor Search
Single-ShotMulti-ShotSingle-ShotMulti-Shot
r = 1r = 10r = 20mAPr = 1r = 10r = 20mAPr = 1r = 10r = 20mAPr = 1r = 10r = 20mAP
Zero-Padding [6]14.854.171.316.019.161.478.410.920.668.485.826.924.475.991.318.6
HCML [37]14.353.269.216.2----24.573.386.730.1----
BDTR [9]17.055.472.019.7------------
cmGAN [47]26.967.580.627.831.572.785.022.331.677.289.242.237.080.992.132.8
TCMDL [42]16.958.876.619.3----21.671.487.932.3----
HSME [45]20.762.778.023.1------------
D2RL [56]28.970.682.429.2------------
SDL [22]28.170.283.729.0----32.680.590.739.6----
MAC [48]33.279.090.136.2----33.482.593.745.0----
AlignGAN [12]42.485.093.740.751.589.495.733.945.987.694.454.357.192.797.445.3
DFE [26]48.788.995.348.654.691.696.842.152.389.995.959.759.694.598.150.6
eBDTR [41]27.867.381.328.4----32.577.489.642.5----
Hi-CMD [57]34.977.6-35.9------------
MSR [16]37.383.493.338.143.986.995.730.539.689.397.750.946.693.698.840.1
JSIA-ReID [58]38.180.789.936.945.185.793.829.543.886.294.252.952.791.196.442.7
XIV-ReID [62]49.989.896.050.7------------
MACE [18]51.687.394.450.1----57.493.097.564.8----
CML [66]51.892.797.751.256.394.198.143.455.094.499.463.760.496.999.553.5
DDAG [29]54.890.495.853.0----61.094.198.468.0----
SIM [46]56.9--60.9------------
HC [83]57.091.596.855.0----59.792.196.264.9----
DG-VAE [21]59.593.8-58.5------------
cm-SSFT [17]61.689.293.963.263.491.295.762.070.594.997.772.673.096.399.172.4
CDP [81]38.082.391.738.4------------
expAT [1]38.676.686.438.6------------
AGW [8]47.5--47.7----54.2--63.0----
GECNet [11]53.489.995.751.8----60.694.398.162.9----
HAT [64]55.392.197.453.9----62.195.899.269.4----
LbA [82]55.4--54.1----58.5--66.3----
NFS [70]56.991.396.555.563.594.497.848.662.896.599.169.870.097.799.561.5
CICL [23]57.294.398.459.360.795.298.652.666.698.899.774.773.899.499.968.3
CPN [44]57.392.697.156.963.193.997.450.759.394.598.466.766.397.499.858.5
MSO [51]58.792.1-56.465.994.4-49.663.196.6-70.372.197.8-61.7
FMI [71]60.094.298.158.8----66.196.699.473.0----
HCTL [2]61.793.197.257.5----63.491.795.368.2----
CM-NAS [69]62.092.997.360.068.794.998.453.567.097.099.373.076.598.799.965.1
MCLNet [24]65.493.397.162.0----72.696.799.276.6----
SFANet [55]65.793.097.060.8----71.696.699.580.0----
SMCL [67]67.492.996.861.872.290.794.354.968.896.698.875.679.695.398.066.6
CAJ [74]69.995.798.566.9----76.397.999.580.4----
MMN [68]70.696.299.066.9----76.297.299.379.6----
MPANet [10]70.696.298.868.275.697.999.462.976.798.299.681.084.299.799.975.1
Table 4. Rank-1 accuracy (%) and mAP (%) performance of state-of-the-art methods on RegDB_Mix and SYSU-MM01_Mix. Bold numbers are the best results. All approaches listed in this table are reproduced by us.
Table 4. Rank-1 accuracy (%) and mAP (%) performance of state-of-the-art methods on RegDB_Mix and SYSU-MM01_Mix. Bold numbers are the best results. All approaches listed in this table are reproduced by us.
ApproachVenueRegDB_MixSYSU-MM01_Mix
r = 1r = 10r = 20mAPr = 1r = 10r = 20mAP
DDAG [29]ECCV’2036.2852.0559.4133.8532.1655.0566.2725.44
LbA [82]ICCV’2138.5755.0762.3835.9834.8065.3877.2727.46
GECNet [11]TCSVT’2141.7657.9464.6637.4922.2960.7175.5814.19
AGW [8]TPAMI’2142.0257.0363.9538.1130.8764.8177.1223.70
CM-NAS [69]CVPR’2142.5051.3756.0441.7336.4854.0862.7029.51
HCTL [2]TMM’2158.4170.8576.2152.5933.6252.2461.4726.97
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, H.; Zhong, X.; Huang, W.; Jiang, K.; Liu, W.; Wang, Z. Visible-Infrared Person Re-Identification: A Comprehensive Survey and a New Setting. Electronics 2022, 11, 454. https://doi.org/10.3390/electronics11030454

AMA Style

Zheng H, Zhong X, Huang W, Jiang K, Liu W, Wang Z. Visible-Infrared Person Re-Identification: A Comprehensive Survey and a New Setting. Electronics. 2022; 11(3):454. https://doi.org/10.3390/electronics11030454

Chicago/Turabian Style

Zheng, Huantao, Xian Zhong, Wenxin Huang, Kui Jiang, Wenxuan Liu, and Zheng Wang. 2022. "Visible-Infrared Person Re-Identification: A Comprehensive Survey and a New Setting" Electronics 11, no. 3: 454. https://doi.org/10.3390/electronics11030454

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop