Next Article in Journal
Knowledge Graph Multi-Hop Question Answering Based on Dependent Syntactic Semantic Augmented Graph Networks
Previous Article in Journal
Prompt-Enhanced Generation for Multimodal Open Question Answering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MHDNet: A Multi-Scale Hybrid Deep Learning Model for Person Re-Identification

School of Mathematical Sciences, Jiangsu University, Zhenjiang 212013, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(8), 1435; https://doi.org/10.3390/electronics13081435
Submission received: 21 February 2024 / Revised: 31 March 2024 / Accepted: 7 April 2024 / Published: 10 April 2024
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)

Abstract

:
The primary objective of person re-identification is to identify individuals from surveillance videos across various scenarios. Conventional pedestrian recognition models typically employ convolutional neural network (CNN) and vision transformer (ViT) networks to extract features, and while CNNs are adept at extracting local features through convolution operations, capturing global information can be challenging, especially when dealing with high-resolution images. In contrast, ViT rely on cascaded self-attention modules to capture long-range feature dependencies, sacrificing local feature details. In light of these limitations, this paper presents the MHDNet, a hybrid network structure for pedestrian recognition that combines convolutional operations and self-attention mechanisms to enhance representation learning. The MHDNet is built around the Feature Fusion Module (FFM), which harmonizes global and local features at different resolutions. With a parallel structure, the MHDNet model maximizes the preservation of local features and global representations. Experiments on two person re-identification datasets demonstrate the superiority of the MHDNet over other state-of-the-art methods.

1. Introduction

Person re-identification (Re-ID) is a critical field within computer vision that utilizes techniques to detect individuals in images or videos [1,2]. It can be considered a subfield within the domain of image retrieval. In recent years, Re-ID has emerged as a crucial research area due to its diverse applications in criminal investigations and computer vision. The primary goal of Re-ID is to efficiently retrieve and recognize specific individuals, thereby facilitating automatic identification in surveillance systems with non-overlapping fields across multiple cameras. However, challenges such as occlusion [3,4], illumination, human posture [5], and other confounding factors [6] can significantly impact the effectiveness of Re-ID in practical surveillance systems.
Early research on person re-identification focuses on extracting feature representations to distinguish individual identities [7,8] and designing an effective distance measurement to establish image similarity [9,10,11]. With the development of deep learning, the focus of early research on Re-ID has shifted towards developing robust feature representations [12,13,14] to distinguish individual identities, alongside the exploration of effective distance measurement techniques tailored to establish image similarity. Person re-identification based on deep learning can divide three aspects from the extracted image features: (1) Global feature learning is widely used in person re-identification. However, using attention mechanisms [15,16] to extract a global feature vector may cause the network to neglect local details, which are also important for accurate Re-ID. (2) Local feature learning is achieved by dividing the pedestrian image into smaller blocks to extract features from various parts and enhance the network’s performance. However, using methods like grid segmentation and pose division can increase computational complexity, especially when using the pose division technique which may require an additional pose estimation network to improve performance. (3) Integrating global and local features through fusion learning [13,17] can help the network capture both global and local information to improve the identification of pedestrians [18,19,20]. Although this approach leads to better semantic feature extraction, it can increase the model’s complexity and network parameters. Additionally, it is important to note that current methods mainly concentrate on deep semantic features and may not consider other relevant factors.
Current research focuses primarily on the output features of high layers, overlooking the learning potential of shallow components for detecting subtle details of a person. Furthermore, the use of multi-loss joint training enables evaluation of person feature distances across several dimensions, while multi-loss approaches have potential, it is important to note that different losses have varying impacts on results, necessitating the incorporation of novel loss combinations. However, utilizing multiple losses increases training costs significantly as the computational complexity of the system is augmented.
To address the above-mentioned issues, this paper proposes a multi-scale interactive transformer structure that utilizes Vision Transformer as the Re-ID backbone to extract hierarchical information and fine-grained features. Additionally, a multi-scale feature fusion module is proposed to achieve bidirectional fusion of the global modeling ability of the Transformer and the local capture ability of the CNN, effectively improving the performance and stability of the model. The joint action of the two modules improves the network’s perception of crucial person information. This paper also incorporates a learnable attention mechanism to further enhance the model’s performance by allocating more attention to features. Experimental results on widely used datasets demonstrate that the method proposed in this paper achieves impressive performance and robustness in Re-ID tasks, significantly improving mAP and Top-1 accuracy compared to the current state-of-the-art methods. The primary contributions of this paper are as follows:
(1) We investigated the feasibility of replacing convolutional neural networks (CNNs) with vision transformers (ViT) in the person re-identification (Re-ID) domain, and we successfully integrated visual specificity induction bias into the standard ViT architecture, achieving comparable and enhanced performance comparable to recent transformer-based models.
(2) We introduce a spatial prior module and two feature interaction operations, seamlessly integrating them into the backbone network in a hybrid manner. This approach harmonizes the strengths of various modules, fostering efficient information exchange within the network. As a result, our model captures local details more precisely and reorganizes fine-grained multi-scale features, significantly improving matching accuracy and stability.
(3) We conducted rigorous experiments and ablation studies on benchmark single-person Re-ID datasets, Market1501 and DukeMTMC-reID, along with cross-dataset validation to assess our method’s generalization. Our approach’s effectiveness is evident from numerous experiments, demonstrating competitive performance against state-of-the-art methods.

2. Related Work

In this section, we will review previous relevant work to provide reference to our research and compare and analyze our work with these studies.

2.1. Transformer in Person Re-ID

In recent years, with the widespread use of Transformer [21] in natural processing, research on its application to computer vision tasks [22,23,24] to study image dependencies has gained momentum. In Re-ID [1,2,25], spatial alignment [13,26] and multi-scale features [27,28] are crucial for feature learning. However, implementing spatial alignment and multi-scale features in visual-based Transformer structure remains a challenging research area. In their work, He et al. [29] proposed a side information embedding and puzzle patch modules to discriminative features in a pure Transformer framework for improving performance. Chen et al. [30] constructed fully relational features for Re-ID with a fully relational high-order transformer structure (OH Transformer), acquiring attention matrices based on queries and isolated key pairs at each spatial position and further modeling high-order statistical information for non-local mechanisms. To enhance the robustness of feature extraction, Chen et al. [31] designed a hybrid backbone Res Transformer by combining ResNet-50 with Transformer blocks. Lai et al. [32] proposed an adaptive component partitioning model to extract local features more effectively in human Re-ID. In this paper, we propose a bidirectional fusion of Transformer’s global modeling ability and CNN’s local capture ability. This is achieved by combining Transformer with CNN, and we utilize a multi-scale feature fusion module for feature extraction at different levels. Our proposed approach achieves state-of-the-art performance on three datasets.

2.2. Multi-Scale Feature Method

In Re-ID tasks, different feature levels such as color, texture, shape, and posture contribute to feature learning. Therefore, introducing multi-scale features has shown potential in improving model performance. The method of multi-scale features involves adding various feature extraction branches, each with a different scale, to the backbone network. The features from different scales extracted by these branches are aggregated for training, as shown in Figure 1d. This approach aims to learn the depth features of different body layers, optimizing model performance. Cai et al. [33] used attention mechanisms to extract characteristics at three scales. Wang et al. [34] extracted multi-scale features based on different stages of the backbone network and added these to the subsequent coding task. Liu et al. [35] proposed a multi-scale feature enhancement model that aggregates shallow, medium, and deep-level features to extract feature maps with good spatial structures and rich semantic information. Wu et al. [27] proposed an attention depth framework with multi-scale deep supervision by inserting attention modules into the backbone network for efficient feature extraction. Zhou et al. [36] proposed a lightweight full-scale network model that extracts multi-scale features through full-scale residual blocks. Chen et al. used feature block processing and multi-layer fusion methods to extract multi-scale features. Hao Zhang et al. [37] proposed the multi-scale visual attribute co-attention model (mVACA), enhancing zero-shot image recognition performance and achieving outstanding results in benchmarks. Zhang et al. [38] developed an innovative region-based multi-scale network that significantly improves emotional image recognition by seamlessly integrating features extracted from focal regions along with their surrounding contexts. Zhou et al. [39] introduced a DNN model that integrates spatial pyramid pooling with feature pyramids to enhance distorted image quality assessment. This model leverages deep end-to-end supervision to efficiently utilize multi-scale features, thereby improving the accuracy of perceptual quality prediction. Chen et al. [40] introduced an innovative multi-scale SER parallel network, AMSNet, which integrates fine-grained frame-level manual features with coarse-grained utterance-level deep features. This network also incorporates an enhanced attention-based LSTM and a CNN with Squeeze-and-Excitation blocks (SCNN) to extract emotional information from speech signals more comprehensively. Hu et al. [41] innovatively introduced the Multi-scale Multi-angle Attention Network (MMAN), integrating 3D and 2D convolutional layers to extract spectral and spatial features, respectively, for a comprehensive understanding of image content. While these methods have their advantages and disadvantages, they effectively extract feature information at various scales and prove to be beneficial attempts and explorations for pedestrian recognition tasks.

3. The Proposed Method

This section provides a comprehensive exposition of our novel approach that addresses the research problem.

3.1. Framework Overview

In order to tackle the intricacies of the detection framework and the absence of dimensional correlation inherent in the domain of person re-identification, we present our network architecture. The proposed person re-ID architecture based on multi-scale hybrid features includes a ViT architecture, an information fusion architecture consisting of a spatial prior module and a feature fusion module, and a person re-identification classification layer.
As illustrated in Figure 1, we focus on optimizing both the network model architecture and the deep model. This paper employs a modified Vision Transformer as our backbone for generating person identity features. The network structure is shown in Figure 1a, consisting of a patch embedding layer and N Transformer encoder layers. We employ patch embedding technology to partition the image into 16 × 16 non-overlapping small blocks. These blocks are then flattened and mapped to the feature space, resulting in a reduction of feature resolution by a factor of 16 compared to the original image. Following this, we apply positional encoding to incorporate spatial relationships between the features before feeding them into N encoder layers for further processing. Within each encoder layer, every feature participates in a global attention mechanism that enables it to interact with all other features for information exchange.
Regarding information fusion architecture, this paper employs an efficient and innovative approach, as illustrated in Figure 1b. We process the input image by feeding it into the spatial prior module to obtain three spatial features at different scales. Following this, we flatten these feature maps as inputs for feature interaction. Specifically, we divide the ViT encoder into N blocks, where N is typically 12. For each block, we inject the spatial prior information into it through the spatial feature fusion module and then extract hierarchical features from the output, using a multi-scale feature extractor. Next, we perform N feature interactions to obtain high-quality multi-scale features for enhancing the model’s performance.
The person re-identification classification layer of the network includes a linear layer, a normalization layer, and a fully connected layer, as illustrated in Figure 1c. We transform the input feature vector into a 768-dimensional feature representation for person identity description. Subsequently, we determine the person’s identity in the re-identification task using a fully connected layer.

3.2. Space Prior Module

Recent studies [42,43] have shown that incorporating convolution operations into the Transformer architecture can better capture local spatial information. However, rather than modifying the original ViT architecture, we were inspired to introduce the Spatial Prior Module ( S P M ) to address this issue. The S P M module is designed to work parallel with the embedding layer while representing the local spatial context of the input image. Its precise definition is articulated as follows:
F 1 , F 2 , F 3 = S P M ( X )
where X R C × H × W represents the holistic depiction of the input image.
As illustrated in Figure 1d, the proposed module includes three convolutional layers and one maximum pooling layer to extract and downsample features from the input image. We then utilize a set of 1 × 1 strided convolutions to expand the number of channels four times without altering the feature map size. This approach yields a feature pyramid { F 1 , F 2 , F 3 } which is used as an input for feature interaction. The precise computational procedure for F i is as follows.
F i + 1 = f BN C o n v F i , w i , β i , γ i
where f represents the activation function (such as R e L U ), C o n v denotes the convolution operation, and B N denotes the batch operation. w represents the weight of convolution kernel, and β and γ represent the scale and factors in batch normalization operation.

3.3. Feature Fusion Module

In order to tackle the intricacies of the detection framework and the absence of dimensional correlation inherent, we introduce the Global Feature Fusion Module ( F F M ), as illustrated in Figure 2. This module enables the continuous coupling of local features and global representations via interaction.
It is important to note that CNN and Transformer models usually have different feature dimensions. In CNN, the feature map has dimensions of F R C × H × W , where C, H, and W represent the number of channels, height, and width, respectively. In Transformer, the shape of the patch embedding is P R ( K + 1 ) × E , where K, 1, and E represent the number of multiple patches, tokens, and embedded dimensions of the image, respectively.
To ensure the alignment of spatial dimensions between the feature map in the CNN branch and the patch embedding in the Transformer branch, an upsampling operation is performed on the CNN feature map F s R C s × H s × W s . The feature map is then adjusted to C s channels using a 14 × 1 convolution to obtain F c o n v s R C s × H s × W s . Finally, the upsampled feature map is added to F as F f i n a l = F + F c o n v s .
When propagating from the branch to the Transformer branch, a 1 × 1 convolution operation is applied to adjust the channel number to C t , resulting in P c o n v t R C t × ( K + 1 ) × E ) . The downsampling operation is then used to align the spatial dimension, obtaining P c o n v p o o l t R ( C t × ( K + p ) × E ) , where p represents the number of spatial dimensions that need to be expanded. Add P c o n v p o o l t to patch embedding P to obtain P f i n a l = P + P c o n v t R ( ( K + p ) × E ) . The final F f i n a l and P f i n a l are normalized by using LayerNorm and BatchNorm.
In contrast, feature maps and patch embeddings exhibit significant semantic differences. Local convolution operators identify features within a local region and encode them into distinct feature maps. Conversely, the global self-attention mechanism considers all information and captures global relationships between pixels. To bridge these semantic differences, we introduce F F M to each block (except the first) in order to progressively integrate semantic information across the network.

3.4. Loss Function

In person re-identification tasks, the loss function plays a crucial role in model training by adjusting the model’s discriminative ability and capturing correlations between similar samples, either individually or jointly. To comprehensively evaluate the proposed model under different loss constraints in person re-identification (Re-ID) tasks, incorporate several metric learning loss functions. These include the ID loss, triplet loss, circle loss, contrast loss, instance loss, and sphere loss, which can be used individually or in combination.

3.4.1. Cross-Entropy Loss

In person re-identification tasks, the cross-entropy loss is the most commonly used ID loss and is used to measure the difference between the results of the model and the actual labels. The smaller the cross-entropy loss, the closer the predicted results and actual labels are.
Specifically, for an input image x, the predicted output of the model for class i is denoted as p i (the prediction probability for class i), and the true label of x is denoted as y. The expression for the cross-entropy loss is given by Equation (3):
L C E = i N q i log p i q i = 0 , y i q i = 1 , y = i
In ReID tasks, noise, missing labels, or inaccurate labels can lead to overfitting of the model and limit its general ability when using a separate cross-entropy loss. To address this issue, we propose the use of label smoothing [44] in cross-entropy to prevent the loss of negative samples from being ignored. Unlike traditional cross-entropy loss, the class label q i is calculated using a certain probability instead of being strictly defined as either 0 or 1. This improves the model’s performance, and α is represented by Equation (4):
q i = 1 N 1 N ε , i f i = y ε N , o t h e r w i s e
where N represents the number of classes in the multi-classification problem, and ε is a small hyperparameter.

3.4.2. Triplet Loss

Person re-identification tasks present unique challenges in terms of training models to accurately measure distance and ensure high discriminability, making traditional loss functions (such as the cross-entropy loss) inadequate. To address this, we selected metric-based learning using the triplet loss function [45]. By using the distance between similar and dissimilar samples to optimize the feature embedding space, triplet loss compensates for the limitations of simple metric learning. The formula for calculating triplet loss is shown in Equation (5):
L T r i = d p d n α +
where d p and d n represents the distance between the feature representation pairs of positive and negative sample pairs, usually calculated using Euclidean distance or cosine distance. α is the marginal distance of loss.

3.4.3. Circle Loss

Referring to Equations (3) and (5), the circles of the same color show person features of the same identity. The identity loss calculates the cosine distance between features, while the triplet loss calculates the Euclidean distance between features. In person Re-ID tasks, using identity loss alone for metric learning generally achieves better results than triplet loss, and we also validate this claim. The Circle loss can be defined as follows, where a given pedestrian sample x in the feature space has K within-class similarity scores and L between-class similarity scores:
L C i r c l e = log 1 + j = 1 L exp γ α n j s n j Δ n i = 1 K exp γ α p i s p i Δ p
where s n j represent between-class similarity, j { 1 , 2 , , L } . s p i represent within-class similarity, i { 1 , 2 , , K } . γ is the scale factor. α n j and α p i are linear factors that constrain the learning rate of s n j and s p i . Δ n and Δ p denote the between-class and within-class thresholds, respectively.

4. Experiment

In this section, we conducted experiments to verify the proposed method’s effectiveness and compare it with the state-of-the-art techniques. We selected two benchmark Re-ID datasets and conducted comparative experiments using the suggested method, as presented in Table 1.

4.1. Datasets

Market1501 [46] dataset comprises 32,668 images of 1501 individuals captured by six cameras. Each individual is captured by at least two cameras, and the dataset includes multiple training image sets of 751 individuals, totaling 12,936 images and averaging 17.2 training images per person. The test set contains 19,732 images of 750 individuals, with 3368 of these images used for query purposes, and the remaining images used as a library.
DukeMTMC-ReID [36] dataset comprises 36,411 images and 1812 identities captured by eight different cameras. Of the identities, 1404 were captured by more than one camera, while 408 were captured by only one camera. The training set includes 702 images of identities, while the remaining identities were used for testing. For each identity in each camera, one image was selected as the query set in the test set, while the rest of the images were reserved as the image library.
Evaluation indicators: Two commonly used metrics for pedestrian recognition tasks are mean average precision (mAP) and cumulative characteristic (CMC). The traditional accuracy indicator, referred to as Top-1 accuracy, reflects the matching result between the identity with the highest probability prediction by the model and the ground truth. Conversely, Top-5 accuracy is determined by considering the identities with the top five highest probability predictions.

4.2. Implementation Details

The experiments in this article were conducted on a Windows 11 operating system with a 12th generation Intel Core i9-12900H processor and 30 GB of memory, and an Nvidia RTX 3090 graphics card (24 GB) using CUDA version 11.6. We used Python 3.8.17 and PyTorch 1.12.1 for data processing.
The model was trained for 500 epochs with images resized to 256 × 128. Stochastic Gradient Descent (SGD) was utilized as the optimization algorithm with the learning rate initialized to 0.08 and subsequently optimized using a simulated annealing algorithm. We set the drop rate to 0.1 to minimize overfitting.

4.3. Ablation Study

The effectiveness of SPM and FFM. To corroborate the performance advantages of SPM and FFM, we compare the performance of different backbones on MHDNet in Table 2 and Figure 3. We can observe that ViT series backbones bring better retrieval performance, but at the same time, their increased model complexity brings more inference time consumption. It is important to highlight that in the Market-1501, MHDNet achieves the mAP of 87.7, which is 10.2 higher than the value of 77.5 by ResNet-50. Similarly, in the DukeMTMCID, MHDNet obtains the mAP value of 79.7, which is 11.2 higher than the 68.5 mAP value of ResNet-50. This outcome underlines that MHDNet leveraging SPM and FFM is significantly superior to other methods on two datasets. Our experiments have demonstrated that the SPM and FFM components have a considerable impact on improving the baseline performance.
Effective Feature Selection. In order to demonstrate the effectiveness of each module in the proposed MHDNet, we experiment on Market-1501, DukeMTMC-ReID. We remove all the above modules and set the ViT (ViT with basic operation module) as the baseline. Starting with the baseline, three modules are added to the baseline model in turn. The experimental results are shown in Table 3 and Figure 4. In Table 3, we found using GAP in MHDNet performed better than baseline, while BN did not yield significant improvements until BN was introduced. In addition, we experimented with the effect of MHDNet with an FC layer and found that its accuracy is not as high as MHDNet with one-dimensional convolution.

4.4. Effectiveness of Different Loss Functions

Since different loss functions have other effects on person re-identificatin features, the model performance is also significantly different. So, we conducted ablation experiments based on loss selection to demonstrate the differential effects of different loss functions, as shown in Table 4 and Table 5.
Based on empirical evidence, the ID loss outperforms the Triplet loss when both losses are used separately. Moreover, when combined with the Triplet loss or Circle loss, the ID loss demonstrates superior performance compared to the Instance loss and Sphere loss. This indicates that sensible choice of loss function for metric learning can consistently enhance model performance. However, considering our primary focus on showcasing the performance of the FFM module, we limit our subsequent experiments to using only the ID loss and Triplet loss as the distance metrics.

4.5. Comparison with the State-of-the-Art Methods in Single-Domain Person Re-Identification

In this Section, we compare MHDNet with other methods on Market-1501 [46] and DukeMTMC-ReID [36], as shown in Table 6. For fair comparisons, no post-processing such as re-ranking strategies or multi-query fusion was used for our methods.
Table 6 shows that our MHDNet achieved 87.7% mAP and 94.6% Rank-1 on the Market-1501 [46]. Its Rank-1 accuracy is slightly, 0.8%, lower than PAT and the same as Rank1 accuracy in TransReID, yet MHDNet clearly surpasses all methods in terms of mAP. On the DukeMTMC-ReID [36] dataset, MHDNet achieved the Rank-1 accuracy of 89.1% and the mAP value of 79.7%, respectively, which are higher than most models. However, it still cannot reach the level of some recent excellent models in all indicators. This is because we only improved the model’s performance by modifying the backbone to make it work well on devices with limited performance, while strictly limiting the model’s resource consumption. In contrast, other models do not have this limitation.
As a result, our model achieves results comparable to most classical models, but there is still a gap compared to the best models. Compared with other models, we improve only on the backbone to improve the accuracy, which has better scalability and can be easily combined with other methods to further improve the model accuracy.

4.6. Comparison with the State-of-the-Art Methods in Cross-Domain Person Re-Identification

To assess and demonstrate the efficacy and superiority of MHDNet, we conducted a comparative evaluation of the proposed approach against several leading UDA re-ID methods on the Market-1501→DukeMTMC and DukeMTMC→Market-1501. The experimental outcomes are summarized in Table 7, providing quantitative evidence for the performance of MHDNet in UDA re-ID scenarios.
From Table 7, we can observe that MHDNet achieves 83.8% Rank-1 and 63.8% mAP on the DukeMTMC→Market-1501 and 76.4% and 58.8% on the Market-1501→DukeMTMC, respectively. Specifically, compared to PREST, which employs a progressive representation enhancement approach, MHDNet demonstrates improvements of 2.0% and 2.7% in Rank-1 and mAP on the Market-1501→DukeMTMC and 1.3% and 1.4% on the DukeMTMC→Market-1501, respectively. This suggests that MHDNet effectively reduces the domain gap between the source and target domains and gradually adapts to the data distributions of both domains compared to PREST. In comparison to Soft-mask, the top performer in Table 7, MHDNet demonstrates a minor decrement in performance, achieving a slightly lower mAP by 1.7% and a Rank-1 accuracy reduction of 2.1% on the DukeMTMC→Market-1501. Similarly, on the Market-1501→DukeMTMC, MHDNet exhibits a minor decline in Rank-1 accuracy by 0.5% and a 2.5% drop in mAP. Evidently, MHDNet does not possess a notable edge over Soft-mask in the DukeMTMC→Market-1501 and Market-1501→DukeMTMC. This could be due to the significant stylistic disparities between the Market-1501 and DukeMTMC datasets, originating from Asia and America, respectively, as well as the limitations in our network’s cross-domain fusion capabilities. In our future endeavors, we aim to investigate strategies that effectively integrate domains exhibiting substantial stylistic variations.

4.7. Visualized Attention Maps of the MHDNet

In this section, we conduct an exploratory analysis of the performance of the proposed network by visualizing the final output feature maps of MHDNet, ResNet-50, DeiT, and ViT, as shown in Figure 5. The purpose of this analysis is to gain a deeper understanding of the characteristics of the network and to evaluate whether our design has achieved the expected performance.
It is evident from the first and second rows on the left side of the attached image that ResNet-50 struggles in handling regional features, whereas MHDNet demonstrates profound attention to such features, excelling not only in balance but also in notable multi-scale attention capabilities. In contrast, the first and second rows on the right side reveal that while the traditional ViT structure model can, to a certain extent, focus on regional features, it falls short in extracting these features with precision. MHDNet achieves remarkable balance between local and global information, enabling accurate separation of individuals from their surroundings. These characteristics fully demonstrate that the use of MHDNet in person re-identification (ReID) tasks can more effectively capture human body features, enhance recognition accuracy, and better adapt to various environmental and scene processing requirements.

4.8. Rank-List Visualization Analysis

This section examines feature extraction and similarity matching for each probe person identification. We used the MHDNet, ViT, and ResNet-50 models to extract person features. Then, we applied a similarity matching algorithm to select the top 10 search results based on their similarity score. Correctly matched results are indicated in green boxes, while incorrectly matched results are highlighted in red boxes. The significant advantage of MHDNet in person recognition tasks is demonstrated by the in-depth analysis of query results for different test samples (numbered 1–4) presented in Figure 6.
The inference results presented in Figure 6 demonstrate that MHDNet retrieves targets with the same ID rapidly and achieves better matching results than ResNet-50 and ViT models for each query. Furthermore, in complex cases, such as case B, MHDNet can correctly match the target and ignore other possible interfering factors, indicating that it has stronger robustness and accuracy and can perform well even in complex person recognition tasks.

5. Conclusions

This paper proposes an innovative hybrid multi-scale neural network architecture aimed at addressing the challenges in person re-identification (ReID) tasks. Firstly, we introduce a carefully designed spatial prior module that effectively extracts multi-scale features from different levels of network branches, capturing the diverse information present in pedestrian images. Subsequently, we construct a fusion module that bridges the gap between CNN and ViT backbone networks, leveraging a self-attention mechanism to learn global and local information of different granularities. This design successfully introduces image-related inductive biases into the architecture without altering the inherent structure of the ViT model, enabling the reconstruction of fine-grained multi-scale features crucial for accurate predictions. Through a series of comprehensive experimental validations, our dual-backbone model demonstrates superior performance in Re-ID tasks. Compared to currently well-designed vision transformers, our approach not only achieves comparable or better results but also does not significantly increase the number of parameters and computational complexity, thus achieving a good balance between performance and efficiency. Looking ahead, we will continue to delve deeper into the pivotal role of multi-scale features in person re-identification tasks and strive to address the challenge of multi-scale alignment in multi-modal Re-ID. We anticipate contributing more innovative outcomes to the development of the person re-identification field through further research and optimization.

6. Privacy and Ethical Considerations

The proposed hybrid multi-scale neural network architecture, despite its promising performance in Re-ID tasks, necessitates a thorough examination of its privacy and ethical implications. The extraction of multi-scale features, albeit effective in enhancing ReID accuracy, introduces potential privacy risks associated with the handling of sensitive personal data. The integration of various networks further complicates these ethical considerations. Consequently, it is imperative to strike a balance between optimizing performance and safeguarding individual privacy, ensuring that the architecture adheres to ethical standards. Future research endeavors must prioritize the development of sophisticated privacy-enhancing techniques and ethical frameworks to facilitate the responsible and sustainable deployment of ReID technology within the academic community.

Author Contributions

Methodology, J.W. (Jinghui Wang); Software, J.W. (Jinghui Wang); Investigation, J.W. (Jinghui Wang); Supervision, J.W. (Jun Wang). All authors have read and agreed to the published version of the manuscript.

Funding

This research did not receive any external funding, and the article processing charges were borne by the author individually or by their affiliated institution.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Almasawa, M.O.; Elrefaei, L.A.; Moria, K. A Survey on Deep Learning-Based Person Re-Identification Systems. IEEE Access 2019, 7, 175228–175247. [Google Scholar] [CrossRef]
  2. Zahra, A.; Perwaiz, N.; Shahzad, M.; Fraz, M.M. Person re-identification: A retrospective on domain specific open challenges and future trends. Pattern Recognit. 2023, 142, 109669. [Google Scholar] [CrossRef]
  3. Huang, H.; Li, D.; Zhang, Z.; Chen, X.; Huang, K. Adversarially Occluded Samples for Person Re-identification. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5098–5107. [Google Scholar] [CrossRef]
  4. Hou, R.; Ma, B.; Chang, H.; Gu, X.; Shan, S.; Chen, X. VRSTC: Occlusion-Free Video Person Re-Identification. arXiv 2019, arXiv:1907.08427. [Google Scholar] [CrossRef]
  5. Zhao, H.; Tian, M.; Sun, S.; Shao, J.; Yan, J.; Yi, S.; Wang, X.; Tang, X. Spindle Net: Person Re-identification with Human Body Region Guided Feature Decomposition and Fusion. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 907–915. [Google Scholar] [CrossRef]
  6. Song, C.; Huang, Y.; Ouyang, W.; Wang, L. Mask-Guided Contrastive Attention Model for Person Re-identification. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1179–1188. [Google Scholar] [CrossRef]
  7. Xiong, F.; Gou, M.; Camps, O.; Sznaier, M. Person Re-Identification Using Kernel-Based Metric Learning Methods. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  8. Zhao, R.; Ouyang, W.; Wang, X. Person Re-identification by Salience Matching. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2528–2535. [Google Scholar] [CrossRef]
  9. Guillaumin, M.; Verbeek, J.; Schmid, C. Is that you? Metric learning approaches for face identification. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 498–505. [Google Scholar] [CrossRef]
  10. Chen, J.; Zhang, Z.; Wang, Y. Relevance Metric Learning for Person Re-identification by Exploiting Global Similarities. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 1657–1662. [Google Scholar] [CrossRef]
  11. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2012, 60, 84–90. [Google Scholar] [CrossRef]
  12. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  13. Sun, Y.; Zheng, L.; Yang, Y.; Tian, Q.; Wang, S. Beyond Part Models: Person Retrieval with Refined Part Pooling (and A Strong Convolutional Baseline). In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Luo, H.; Gu, Y.; Liao, X.; Lai, S.; Jiang, W. Bag of Tricks and a Strong Baseline for Deep Person Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1487–1495. [Google Scholar] [CrossRef]
  15. Zhang, Z.; Lan, C.; Zeng, W.; Jin, X.; Chen, Z. Relation-Aware Global Attention for Person Re-Identification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3183–3192. [Google Scholar] [CrossRef]
  16. Chen, T.; Ding, S.; Xie, J.; Yuan, Y.; Chen, W.; Yang, Y.; Ren, Z.; Wang, Z. ABD-Net: Attentive but Diverse Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8350–8360. [Google Scholar] [CrossRef]
  17. Wang, G.; Yuan, Y.; Chen, X.; Li, J.; Zhou, X. Learning Discriminative Features with Multiple Granularities for Person Re-Identification. In Proceedings of the 26th ACM international conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018. [Google Scholar]
  18. Yang, W.; Huang, H.; Zhang, Z.; Chen, X.; Huang, K.; Zhang, S. Towards Rich Feature Discovery With Class Activation Maps Augmentation for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1389–1398. [Google Scholar] [CrossRef]
  19. Zheng, F.; Deng, C.; Sun, X.; Jiang, X.; Guo, X.; Yu, Z.; Huang, F.; Ji, R. Pyramidal Person Re-IDentification via Multi-Loss Dynamic Training. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8506–8514. [Google Scholar] [CrossRef]
  20. Wang, P.; Zhao, Z.; Su, F.; Zu, X.; Boulgouris, N.V. HOReID: Deep High-Order Mapping Enhances Pose Alignment for Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 2908–2922. [Google Scholar] [CrossRef]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  22. Jiang, Y.; Chang, S.; Wang, Z. TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up. In Proceedings of the Neural Information Processing Systems, Online, 6–14 December 2021. [Google Scholar]
  23. Li, X.; Hou, Y.; Wang, P.; Gao, Z.; Xu, M.; Li, W. Trear: Transformer-Based RGB-D Egocentric Action Recognition. IEEE Trans. Cogn. Dev. Syst. 2022, 14, 246–252. [Google Scholar] [CrossRef]
  24. Zhang, Q.L.; Yang, Y. ResT: An Efficient Transformer for Visual Recognition. arXiv 2021, arXiv:2105.13677. [Google Scholar]
  25. Zheng, L.; Yang, Y.; Hauptmann, A. Person Re-identification: Past, Present and Future. arXiv 2016, arXiv:1610.02984. [Google Scholar]
  26. Wang, G.; Yang, S.; Liu, H.; Wang, Z.; Yang, Y.; Wang, S.; Yu, G.; Zhou, E.; Sun, J. High-Order Information Matters: Learning Relation and Topology for Occluded Person Re-Identification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 6448–6457. [Google Scholar] [CrossRef]
  27. Wu, D.; Wang, C.; Wu, Y.; Wang, Q.C.; Huang, D.S. Attention Deep Model With Multi-Scale Deep Supervision for Person Re-Identification. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 5, 70–78. [Google Scholar] [CrossRef]
  28. Jiao, S.; Pan, Z.; Hu, G.; Shen, Q.; Du, L.; Chen, Y.; Wang, J. Multi-scale and multi-branch feature representation for person re-identification. Neurocomputing 2020, 414, 120–130. [Google Scholar] [CrossRef]
  29. He, S.; Luo, H.; Wang, P.; Wang, F.; Li, H.; Jiang, W. TransReID: Transformer-based Object Re-Identification. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 14993–15002. [Google Scholar]
  30. Chen, X.; Xu, J.; Xu, J.; Gao, S. OH-Former: Omni-Relational High-Order Transformer for Person Re-Identification. arXiv 2021, arXiv:2109.11159. [Google Scholar]
  31. Chen, Y.; Xia, S.; Zhao, J.; Zhou, Y.; Niu, Q.; Yao, R.; Zhu, D.; Liu, D. ResT-ReID: Transformer block-based residual learning for person re-identification. Pattern Recognit. Lett. 2022, 157, 90–96. [Google Scholar] [CrossRef]
  32. Lai, S.; Chai, Z.; Wei, X. Transformer Meets Part Model: Adaptive Part Division for Person Re-Identification. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 4133–4140. [Google Scholar] [CrossRef]
  33. Cai, H.; Wang, Z.; Cheng, J. Multi-Scale Body-Part Mask Guided Attention for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 15–20 June 2019; pp. 1555–1564. [Google Scholar]
  34. Wang, C.; Song, L.; Wang, G.; Zhang, Q.; Wang, X. Multi-scale multi-patch person re-identification with exclusivity regularized softmax. Neurocomputing 2020, 382, 64–70. [Google Scholar] [CrossRef]
  35. Liu, X.; Tan, H.; Tong, X.; Cao, J.; Zhou, J. Feature preserving GAN and multi-scale feature enhancement for domain adaption person Re-identification. Neurocomputing 2019, 364, 108–118. [Google Scholar] [CrossRef]
  36. Zhou, K.; Yang, Y.; Cavallaro, A.; Xiang, T. Omni-Scale Feature Learning for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Long Beach, CA, USA, 15–20 June 2019; pp. 3701–3711. [Google Scholar]
  37. Zhang, H.; Tian, L.; Wang, Z.; Xu, Y.; Cheng, P.; Bai, K.; Chen, B. Multiscale Visual-Attribute Co-Attention for Zero-Shot Image Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 6003–6014. [Google Scholar] [CrossRef]
  38. Zhang, H.; Xu, M. Multiscale Emotion Representation Learning for Affective Image Recognition. IEEE Trans. Multimed. 2023, 25, 2203–2212. [Google Scholar] [CrossRef]
  39. Zhou, W.; Chen, Z. Deep Multi-Scale Features Learning for Distorted Image Quality Assessment. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems (ISCAS), Daegu, Republic of Korea, 22–28 May 2021; pp. 1–5. [Google Scholar]
  40. Chen, Z.; Li, J.; Liu, H.; Wang, X.; Wang, H.; Zheng, Q. Learning multi-scale features for speech emotion recognition with connection attention mechanism. Expert Syst. Appl. 2023, 214, 118943. [Google Scholar] [CrossRef]
  41. Hu, J.; Tu, B.; Ren, Q.; Liao, X.; Cao, Z.; Plaza, A. Hyperspectral Image Classification via Multiscale Multiangle Attention Network. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–18. [Google Scholar] [CrossRef]
  42. Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  43. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable Person Re-identification: A Benchmark. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar] [CrossRef]
  44. Ristani, E.; Solera, F.; Zou, R.S.; Cucchiara, R.; Tomasi, C. Performance Measures and a Data Set for Multi-target, Multi-camera Tracking. In Proceedings of the ECCV Workshops, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  45. Wang, Y.; Wang, L.; You, Y.; Zou, X.; Chen, V.; Li, S.; Huang, G.; Hariharan, B.; Weinberger, K.Q. Resource Aware Person Re-identification Across Multiple Resolutions. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8042–8051. [Google Scholar] [CrossRef]
  46. Chen, B.; Deng, W.; Hu, J. Mixed High-Order Attention Network for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 371–381. [Google Scholar] [CrossRef]
  47. Sun, Y.; Zheng, L.; Yang, Y.; Tian, Q.; Wang, S. Beyond Part Models: Person Retrieval with Refined Part Pooling. In Proceedings of the European Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  48. Zhang, A.; Gao, Y.; Niu, Y.; Liu, W.; Zhou, Y. Coarse-to-Fine Person Re-Identification with Auxiliary-Domain Classification and Second-Order Information Bottleneck. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 598–608. [Google Scholar] [CrossRef]
  49. Xu, Y.; Zhao, L.; Qin, F. Dual attention-based method for occluded person re-identification. Knowl.-Based Syst. 2021, 212, 106554. [Google Scholar] [CrossRef]
  50. Zhou, K.; Yang, Y.; Cavallaro, A.; Xiang, T. Learning Generalisable Omni-Scale Representations for Person Re-Identification. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5056–5069. [Google Scholar] [CrossRef]
  51. Jin, H.; Lai, S.; Qian, X. Occlusion-Sensitive Person Re-Identification via Attribute-Based Shift Attention. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2170–2185. [Google Scholar] [CrossRef]
  52. Li, H.; Wu, G.; Zheng, W. Combined Depth Space based Architecture Search For Person Re-identification. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6725–6734. [Google Scholar]
  53. Li, Y.; He, J.; Zhang, T.; Liu, X.; Zhang, Y.; Wu, F. Diverse Part Discovery: Occluded Person Re-identification with Part-Aware Transformer. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 2897–2906. [Google Scholar] [CrossRef]
  54. Chen, J.; Jiang, X.; Wang, F.; Zhang, J.; Zheng, F.; Sun, X.; Zheng, W.S. Learning 3D Shape Feature for Texture-insensitive Person Re-identification. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 8142–8151. [Google Scholar] [CrossRef]
  55. Huang, J.; Yu, X.; An, D.; Wei, Y.; Bai, X.; Zheng, J.; Wang, C.; Zhou, J. Learning consistent region features for lifelong person re-identification. Pattern Recognit. 2023, 144, 109837. [Google Scholar] [CrossRef]
  56. Zhu, H.; Ke, W.; Li, D.; Liu, J.; Tian, L.; Shan, Y. Dual Cross-Attention Learning for Fine-Grained Visual Categorization and Object Re-Identification. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4682–4692. [Google Scholar]
  57. Mamedov, T.; Kuplyakov, D.; Konushin, A. Approaches to Improve the Quality of Person Re-Identification for Practical Use. Sensors 2023, 23, 7382. [Google Scholar] [CrossRef] [PubMed]
  58. Li, G.; Liu, P.; Cao, X.; Liu, C. Dynamic Weighting Network for Person Re-Identification. Sensors 2023, 23, 5579. [Google Scholar] [CrossRef]
  59. Wang, M.; Ma, H.; Huang, Y. Information complementary attention-based multidimension feature learning for person re-identification. Eng. Appl. Artif. Intell. 2023, 123, 106348. [Google Scholar] [CrossRef]
  60. Chang, X.; Yang, Y.; Xiang, T.; Hospedales, T.M. Disjoint Label Space Transfer Learning with Common Factorised Space. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  61. Qi, L.; Wang, L.; Huo, J.; Zhou, L.; Shi, Y.; Gao, Y. A Novel Unsupervised Camera-Aware Domain Adaptation Framework for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8079–8088. [Google Scholar]
  62. Li, M.; Zhu, X.; Gong, S. Unsupervised Tracklet Person Re-Identification. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1770–1782. [Google Scholar] [CrossRef]
  63. Zhong, Z.; Zheng, L.; Luo, Z.; Li, S.; Yang, Y. Invariance Matters: Exemplar Memory for Domain Adaptive Person Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 598–607. [Google Scholar]
  64. Li, Y.J.; Lin, C.S.; Lin, Y.B.; Wang, Y. Cross-Dataset Person Re-Identification via Unsupervised Pose Disentanglement and Adaptation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7918–7928. [Google Scholar]
  65. Chen, Y.; Zhu, X.; Gong, S. Instance-Guided Context Rendering for Cross-Domain Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 232–242. [Google Scholar] [CrossRef]
  66. Mekhazni, D.; Bhuiyan, A.; Ekladious, G.S.E.; Granger, E. Unsupervised Domain Adaptation in the Dissimilarity Space for Person Re-identification. arXiv 2020, arXiv:2007.13890. [Google Scholar]
  67. Zhai, Y.; Lu, S.; Ye, Q.; Shan, X.; Chen, J.; Ji, R.; Tian, Y. AD-Cluster: Augmented Discriminative Clustering for Domain Adaptive Person Re-Identification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9018–9027. [Google Scholar]
  68. Tang, Y.; Yang, X.; Wang, N.; Song, B.; Gao, X. CGAN-TM: A Novel Domain-to-Domain Transferring Method for Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 5641–5651. [Google Scholar] [CrossRef]
  69. Verma, A.; Subramanyam, A.; Wang, Z.; Satoh, S.; Shah, R.R. Unsupervised Domain Adaptation for Person Re-Identification Via Individual-Preserving and Environmental-Switching Cyclic Generation. IEEE Trans. Multimed. 2023, 25, 364–377. [Google Scholar] [CrossRef]
  70. Zhang, H.; Cao, H.; Yang, X.; Deng, C.; Tao, D. Self-Training With Progressive Representation Enhancement for Unsupervised Cross-Domain Person Re-Identification. IEEE Trans. Image Process. 2021, 30, 5287–5298. [Google Scholar] [CrossRef]
  71. Li, H.; Pang, J.; Tao, D.; Yu, Z. Cross adversarial consistency self-prediction learning for unsupervised domain adaptation person re-identification. Inform. Sci. 2021, 559, 46–60. [Google Scholar] [CrossRef]
  72. Khatun, A.; Denman, S.; Sridharan, S.; Fookes, C. End-to-End Domain Adaptive Attention Network for Cross-Domain Person Re-Identification. IEEE Trans. Inform. Forensics Secur. 2021, 16, 3803–3813. [Google Scholar] [CrossRef]
  73. Zhang, H.; Li, Y.; Zhuang, Z.; Xie, L.; Tian, Q. 3D-GAT: 3D-Guided adversarial transform network for person re-identification in unseen domains. Pattern Recognit. 2021, 112, 107799. [Google Scholar] [CrossRef]
  74. Chong, Y.; Peng, C.; Zhang, J.; Pan, S. Style transfer for unsupervised domain-adaptive person re-identification. Neurocomputing 2021, 422, 314–321. [Google Scholar] [CrossRef]
  75. Tang, G.; Gao, X.; Chen, Z.; Zhong, H. Unsupervised adversarial domain adaptation with similarity diffusion for person re-identification. Neurocomputing 2021, 442, 337–347. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of MHDNet. (a) The ViT, whose encoder layers are divided into N (usually N = 12) equal blocks (c) for feature interaction. (b) Our Information Fusion Architecture, which contains two key designs, including (e) a spatial prior module for modeling local spatial contexts from the input image, (f) Feature Fusion Module for introducing spatial priors into ViT. (d) Block i.
Figure 1. Overall architecture of MHDNet. (a) The ViT, whose encoder layers are divided into N (usually N = 12) equal blocks (c) for feature interaction. (b) Our Information Fusion Architecture, which contains two key designs, including (e) a spatial prior module for modeling local spatial contexts from the input image, (f) Feature Fusion Module for introducing spatial priors into ViT. (d) Block i.
Electronics 13 01435 g001
Figure 2. Feature fusion module.
Figure 2. Feature fusion module.
Electronics 13 01435 g002
Figure 3. Evaluate the Rank-1 of effective placement of Block on two datasets.
Figure 3. Evaluate the Rank-1 of effective placement of Block on two datasets.
Electronics 13 01435 g003
Figure 4. CMC curve of Baseline, Baseline + GAP, Baseline + GAP + BN, MHDNet with FC Layer and MHDNet on the two datasets.
Figure 4. CMC curve of Baseline, Baseline + GAP, Baseline + GAP + BN, MHDNet with FC Layer and MHDNet on the two datasets.
Electronics 13 01435 g004
Figure 5. Visualization of feature maps based on different Re-ID models.
Figure 5. Visualization of feature maps based on different Re-ID models.
Electronics 13 01435 g005
Figure 6. Top 10 visualization results of the rank-list Market-1501.
Figure 6. Top 10 visualization results of the rank-list Market-1501.
Electronics 13 01435 g006
Table 1. Person Re-ID datasets introduction.
Table 1. Person Re-ID datasets introduction.
DatasetsCamerasTrainIDsTrainImgsTestIDsQueryImgsGalleryImgs
Market-1501 [46]675112,396750336819,732
DukeMTMC-reID [36]870216,522702222817,661
Table 2. Comparison of different backbones.
Table 2. Comparison of different backbones.
MethodsInference TimeMarket1501DukeMTMC-reID
mAPRank 1Rank 5mAPRank 1Rank 5
ResNet-185 ms77.591.295.568.582.292.8
ResNet-3411 ms82.492.596.273.484.793.4
ResNet-5016 ms85.794.297.675.986.294.7
ResNet-10124 ms86.894.498.377.287.395.2
ViT-B40 ms86.894.598.179.388.895.4
Deit-B36 ms86.694.497.878.988.395.1
MHDNet (Our)38 ms87.794.698.679.789.195.6
Table 3. Ablation studies of the proposed method on individual components.
Table 3. Ablation studies of the proposed method on individual components.
MethodsMarket1501DukeMTMC-reID
mAPRank 1Rank 5Rank 10mAPRank 1Rank 5Rank 10
Baseline84.092.495.496.676.485.493.294.4
+ GAP85.193.797.298.077.887.194.695.3
+ BN85.293.997.598.178.187.594.995.5
+ FC Layer86.894.598.198.979.388.895.496.2
MHDNet (Our)87.794.698.699.379.789.195.696.6
Table 4. The effectiveness of loss selections on Market1501.
Table 4. The effectiveness of loss selections on Market1501.
Loss FunctionmAPRank 1Rank 5Rank 10
ID81.893.197.298.5
Triplet80.391.896.298.1
ID + Spere78.590.997.198.2
ID + Contrast82.693.097.998.7
ID + Cricle83.493.798.198.9
ID + Instance82.292.997.798.6
ID + Triplet87.794.698.699.3
ID + Triplet + Cricle84.193.698.398.8
ID + Triplet + Instance83.893.497.598.9
Table 5. The effectiveness of loss selections on DukeMTMC-reID.
Table 5. The effectiveness of loss selections on DukeMTMC-reID.
Loss FunctionmAPRank 1Rank 5Rank 10
ID75.386.494.596.2
Triplet74.285.893.796.1
ID + Spere71.881.694.195.9
ID + Contrast76.987.895.296.3
ID + Cricle77.688.195.496.5
ID + Instance76.287.695.196.0
ID + Triplet79.789.195.696.6
ID + Triplet + Cricle78.588.795.596.6
ID + Triplet + Instance77.487.995.096.4
Table 6. Validation results on Market1501 and DukeMTMC-ReID. Bold indicates the best performance.
Table 6. Validation results on Market1501 and DukeMTMC-ReID. Bold indicates the best performance.
MethodVenueMarket1501DukeMTMC-reID
mAPRank 1mAPRank 1
PCB [47]ECCV 201881.693.869.283.3
BoT [14]CVPR 201985.994.576.486.4
HOReID [26]CVPR 202084.994.275.686.9
M-DEFNetMTA 202082.794.873.184.7
ADC-2OIB [48]CVPR 202187.794.874.987.4
DAReID [49]KBS 202187.094.678.488.9
OSNet [50]TPAMI 202186.794.876.388.7
ASAN [51]TCSVT 202185.394.676.388.7
CDNet [52]CVPR 202186.095.176.888.6
PAT [53]CVPR 202186.695.478.288.8
L3DS [54]CVPR 202187.395.076.188.2
TransReID [29]ICCV 202186.894.679.388.8
ConRFL [55]PR 202281.492.868.480.5
CALCVPR 202287.594.774.186.2
AOPS [51]TCSVT 202284.193.474.186.2
DeiT-Small + DCAL [56]CVPR 202285.394.077.487.9
IIANetMTA 202384.994.2--
With Res2Net50 [57]Sensors 202387.195.077.688.1
DWNet-R [58]Sensors 202387.594.979.188.4
UV-ReID-ABLMMVA 202375.089.961.881.4
ICAM [59]EAAI 202382.393.371.685.6
MHDNet (Ours) 87.794.679.789.1
Table 7. Validation results on M→D and D→M. Bold indicates the best performance.
Table 7. Validation results on M→D and D→M. Bold indicates the best performance.
MethodDukeMTMC-reID→Market1501Market1501→DukeMTMC-reID
mAPRank 1Rank 5Rank 10mAPRank 1Rank 5Rank 10
CFSM [60]28.361.2--27.349.8--
UCDA-CCE [61]30.960.4--31.047.7--
UTAL [62]46.269.2--44.662.3--
ECN [63]43.075.687.591.640.463.375.880.4
PDA-Net [64]47.675.286.390.245.163.277.082.5
CR-CAN+ [65]54.077.789.792.748.668.980.284.7
D-MMD [66]48.870.687.091.546.063.578.883.9
AD-Cluster [67]68.386.794.496.554.172.682.585.5
CGAN-TM [68]35.257.3--36.265.3--
Soft-mask [69]69.586.9--61.376.9--
PREST [70]62.482.592.194.956.174.483.785.9
CAC–CSP [71]36.969.482.8-37.057.571.2-
EDAAN [72]35.464.583.0-39.657.872.2-
3D-GAT [73]28.659.475.2-26.145.159.3-
STReID [74]31.662.379.1-29.252.365.9-
UADA-SD [75]30.257.472.430.345.357.8-
MHDNet (Ours)63.881.888.891.558.870.980.883.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Wang, J. MHDNet: A Multi-Scale Hybrid Deep Learning Model for Person Re-Identification. Electronics 2024, 13, 1435. https://doi.org/10.3390/electronics13081435

AMA Style

Wang J, Wang J. MHDNet: A Multi-Scale Hybrid Deep Learning Model for Person Re-Identification. Electronics. 2024; 13(8):1435. https://doi.org/10.3390/electronics13081435

Chicago/Turabian Style

Wang, Jinghui, and Jun Wang. 2024. "MHDNet: A Multi-Scale Hybrid Deep Learning Model for Person Re-Identification" Electronics 13, no. 8: 1435. https://doi.org/10.3390/electronics13081435

APA Style

Wang, J., & Wang, J. (2024). MHDNet: A Multi-Scale Hybrid Deep Learning Model for Person Re-Identification. Electronics, 13(8), 1435. https://doi.org/10.3390/electronics13081435

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop