Next Article in Journal
Loss Control-Based Key Distribution under Quantum Protection
Previous Article in Journal
Effect of Al Content on Microstructure and Properties of AlxCr0.2NbTiV Refractory High-Entropy Alloys
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Person Re-Identification through Attention-Driven Global Features and Angular Loss Optimization

1
School of Information and Cyber Security, People’s Public Security University of China, Beijing 100038, China
2
Key Laboratory of Security Prevention Technology and Risk Assessment of Ministry of Public Security, Beijing 100038, China
3
Beijing Public Security Bureau, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(6), 436; https://doi.org/10.3390/e26060436
Submission received: 23 April 2024 / Revised: 15 May 2024 / Accepted: 16 May 2024 / Published: 21 May 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
To address challenges related to the inadequate representation and inaccurate discrimination of pedestrian attributes, we propose a novel method for person re-identification, which leverages global feature learning and classification optimization. Specifically, this approach integrates a Normalization-based Channel Attention Module into the fundamental ResNet50 backbone, utilizing a scaling factor to prioritize and enhance key pedestrian feature information. Furthermore, dynamic activation functions are employed to adaptively modulate the parameters of ReLU based on the input convolutional feature maps, thereby bolstering the nonlinear expression capabilities of the network model. By incorporating Arcface loss into the cross-entropy loss, the supervised model is trained to learn pedestrian features that exhibit significant inter-class variance while maintaining tight intra-class coherence. The evaluation of the enhanced model on two popular datasets, Market1501 and DukeMTMC-ReID, reveals improvements in Rank-1 accuracy by 1.28% and 1.4%, respectively, along with corresponding gains in the mean average precision (mAP) of 1.93% and 1.84%. These findings indicate that the proposed model is capable of extracting more robust pedestrian features, enhancing feature discriminability, and ultimately achieving superior recognition accuracy.

1. Introduction

Person re-identification tasks aim to retrieve and match specific pedestrians from multiple nonoverlapping cameras [1] and have attracted widespread attention in academia in recent years, becoming a research hotspot. This technology can be combined with facial recognition, pedestrian detection, etc., to promote the development of intelligent security fields such as suspect tracking.
Traditional person re-identification consists of two sub-tasks, feature extraction and similarity measurement, which cannot achieve end-to-end processing from raw data to results. It is inefficient and time-consuming and generally faces the problem of low accuracy. With the rise of deep learning, person re-identification integrates feature extraction and similarity measurements into a unified framework. By using convolutional neural networks (CNNs) to extract high-level features of pedestrians and designing loss functions to control intra-class and inter-class distances, the performance of person re-identification is greatly improved.
The challenge of a person re-identification task lies in adequately expressing and discriminating pedestrian features. In practical situations, various factors, including camera angles, lighting conditions, occlusions, and intricate backgrounds, hinder the performance of person re-identification methods. These factors lead to an insufficient representation of pedestrian features, compromised discriminatory power, and reduced recognition accuracy [2]. Consequently, the extraction of more distinctive pedestrian features and the adoption of effective similarity measurement methods have become crucial for enhancing inter-class differences and minimizing intra-class variations in person re-identification.
In response to this situation, we optimized the person re-identification method based on global feature learning in both feature extraction and similarity measurement. We propose a person re-identification method based on global feature learning and classification optimization which fully extracts key pedestrian features, makes the network model output more discriminative regarding pedestrian features, and improves the accuracy of person re-identification. This method is based on the backbone network ResNet50, and our main contributions can be summarized as follows:
  • Incorporating the Normalization-based Channel Attention Module [3] into the residual module of the backbone network ResNet50 and using scaling factors to weight and strengthen key pedestrian feature information so that the network model selectively enhances key features and suppresses useless features;
  • Using the dynamic activation function Dynamic ReLU [4], which only shares spatial dimensions, as the activation layer of the backbone network, the parameters of ReLU are dynamically adjusted based on the input convolutional feature maps to enhance the nonlinear expression ability of the network model for pedestrian features;
  • Introducing Arcface loss [5] into cross-entropy loss, the supervised network model learns pedestrian features with large inter-class distance and small intra-class distance, making the network model output more discriminative pedestrian features.

2. Related Work

2.1. Overview of Person Re-Identification

A diverse array of person re-identification methods have been put forth, encompassing methods rooted in feature learning, metric learning [2], and those utilizing generative adversarial networks [6] (GANs). Depending on the varying representations of pedestrian features, feature learning-based methods can broadly be categorized into those focused on global feature extraction, those emphasizing local feature analysis, and those leveraging auxiliary feature learning [7].
Methods of person re-identification based on global feature learning involve representing an entire image as a single feature vector, followed by the measurement of image similarity. Wang et al. [8] proposed a person re-identification framework that combines single-image feature matching and cross-image feature matching and utilized convolutional neural networks (CNNs) to extract pedestrian features. Zhang et al. [9] proposed a person re-identification method based on global features which utilizes a multi-receptive field fusion module and an attention module to fully obtain pedestrian contextual information, as well as a multi-branched network structure to fuse features of different depths to predict pedestrian identity.
Methods of person re-identification based on local feature learning represent an entire image as a set of local features and then measure the similarity of the image. This can be achieved through methods such as dividing pedestrian components or simple vertical region partitioning. Zhao et al. [10] proposed a multi-stage feature decomposition and fusion method for pedestrian re-identification guided by human body regions, taking into account human body structural information. Wang et al. [11] proposed a pose-driven local feature alignment method for person re-identification, which uses a pose encoder to guide the network in extracting visible features of pedestrians and uses a pedestrian component alignment module to extract local features of pedestrians and align them to reduce interference from non-pedestrian features.
Methods of person re-identification based on auxiliary feature learning enhance the expression of pedestrian features through auxiliary information such as instance attributes and identity attributes. Lin et al. [12] optimized recognition performance on large-scale data by utilizing complementary information learned from attribute labels.

2.2. Methods Based on Global Feature Learning

In 2016, Zheng et al. [13] proposed the IDE model, which regarded person re-identification as a classification task and became a baseline model for many subsequent works. For example, in terms of feature extraction, Sun et al. [7] addressed the issue of insufficient pedestrian feature expression by embedding the channel attention mechanism’s SE module into the backbone network, enabling the network model to output more representative pedestrian expression features. Du et al. [14] introduced a spatial attention mechanism to improve the global feature extraction network of the person re-identification algorithm with topological relation occlusion, taking into account the information interference problems in scenes such as occlusion and complex backgrounds. Lin et al. [15] proposed a method called the Dual Descriptor Feature Enhancement (DDFE) network, which uses two independent sub-networks to extract descriptors from the same pedestrian and combine them to create a comprehensive multi-view representation, significantly improving person re-identification performance. In terms of similarity measurement, commonly used loss measures include contrast loss [16], triplet loss [17], quadruple loss [18], etc. Zhang et al. [19] improved the original triplet loss and optimized the intra-class distance.
Among the methods mentioned above, methods that focus on innovation in feature extraction [7,13,14,15] can easily cause problems such as a network model having insufficient generalization ability and misfitting deviations in the training data. Methods that focus on improving the loss function [16,17,18,19] can easily cause problems such as disappearing or exploding gradients and difficulty in capturing complex nonlinear relationships in feature spaces. In contrast, combining feature extraction and similarity measurement and selecting appropriate methods to make them complement each other can help effectively deal with the above problems and optimize the performance of the network model globally and harmoniously.
In recent years, in order to improve the accuracy of person re-identification on datasets, a large amount of work has adopted a combination of global feature learning and local feature learning methods. Wang et al. [20] designed and introduced PAN modules into the global feature branch and designed local feature branches to enable the model to better learn fine-grained features of local regions while preserving the correlation information between regions. Zhang et al. [21] proposed a person re-identification method that utilizes feature fusion and multi-scale information. After extracting the global feature map of a human body image, the global feature map is horizontally segmented to obtain local features. By combining adaptive spatial feature alignment, local feature fusion, and multi-scale feature extraction, pedestrian features are obtained.
Person re-identification methods that obtain local features of pedestrians through horizontal segmentation usually involve multiple branches, which increases the complexity of the network structure, and rough partitioning can easily cause misalignment problems between corresponding components. In contrast, person re-identification methods based on global feature learning have a clear design and a simpler network model process, which has great research significance.

3. Methods

In this section, the framework of the network model proposed in this paper is introduced in Section 3.1, followed by a detailed explanation of three specific optimizations in feature extraction and similarity measurement in the proposed method from Section 3.2, Section 3.3 and Section 3.4.

3.1. Framework of the Proposed Method

The process of person re-identification methods based on global feature learning is roughly as follows: After extracting features from the backbone network, a global average pool (GAP) is used to obtain a global feature vector, which is then processed through a fully connected layer to obtain low-dimensional output features. Then, the triplet loss and cross-entropy loss are used as the measurement loss and classification loss for the similarity measurement, respectively. This type of method has a simple structure but usually lower accuracy.
On this basis, we proposed a novel person re-identification method based on global feature learning. For the structural framework of the network model, we removed the global average pooling layer and fully connected layer at the end of the backbone network ResNet50 and added an adaptive maximum pooling layer to output 2048   ×   4   ×   1 feature vectors. For both feature extraction and similarity measurement, we made the following 3 specific optimizations. The network structure and overall process of our method are shown in Figure 1. Among them, BN represents batch normalization [22] and FC represents a fully connected layer. L triplet represents triplet loss, and L Arcface represents cross-entropy loss after introducing Arcface loss; they are used as the measurement loss and classification loss for the similarity measurement after obtaining output features. In the Dynamic ReLU (spatial-shared only), H represents the height of the feature graph, W represents the width of the feature graph, and C represents the number of channels in the feature graph. In the Normalization-based Channel Attention Module, F 1 represents the input features put into the attention module and M C represents the corresponding output feature.

3.2. Normalization-Based Channel Attention Module

An attention mechanism can help neural network models suppress less significant features in channels, reduce interference caused by irrelevant information, and thus improve discriminative power regarding pedestrian features. We added the Normalization-based Channel Attention Module as a lightweight attention mechanism in the channel dimension using a scaling factor for batch normalization [22] (BN), which uses the standard deviation to represent the importance of weights, avoiding the addition of fully connected and convolutional layers present in other attention models such as Squeeze and Excitation [23] (SE) modules and convolutional block attention modules [24] (CBAMs), which makes computation more efficient while maintaining performance.
For the backbone network ResNet50, the Normalization-based Channel Attention Module was embedded after the last block from Layer 1 to Layer 4, with each block embedding a Normalization-based Channel Attention Module, as shown in Figure 2. Among them, NAM is short for the Normalization-based Channel Attention Module, as is the case below.
The Normalization-based Channel Attention Module utilizes the scaling factor, γ , of batch normalization [22] to measure the variance of each channel and reflect the magnitude of the changes in each channel, thereby indicating the importance of that channel. The role of the scaling factor in BN is shown in Equation (1). In the equation, B in and B out represent the input and output of BN, respectively.
B o u t = B N B i n = γ B i n μ B σ B 2 + ε + β
In the equation, μ B and σ B are the mean and standard deviation of a small batch B , respectively, and γ and β are trainable affine transformation parameters (scaling and shifting). After obtaining the scaling factor, γ , for each channel, the weight, W γ , for each channel can be calculated using Equation (2).
W γ = γ i j = 0 γ j
After calculating the weight, W γ , of each channel, for input feature F 1 , we can obtain the corresponding output feature, M C , through Equation (3).
M C = s i g m o i d W γ B N F 1
The specific workflow of the Normalization-based Channel Attention Module is shown in Figure 3.
We incorporated the Normalization-based Channel Attention Module into the residual module of the backbone network ResNet50, which makes the pedestrian expression features extracted by the proposed person re-identification method more representative and generalized. Specifically, firstly, after the input was processed by the residual block, we used the Normalization-based Channel Attention Module to carry out the weighted adjustment of the feature channels, that is, to enhance the useful feature channels and weaken the redundant feature channels. Secondly, we added the basic feature information processed by the residual block to the key detail feature information weighted by the attention mechanism, aiming to improve the generalization and robustness of the model through a feature fusion strategy. Finally, we connected the input of the residual block to a deeper network layer through Identity Mapping [25], which adds neither extra parameters nor computational complexity. The residual structure of the network model after embedding the Normalization-based Channel Attention Module is shown in Figure 4.

3.3. Dynamic Activation Function

For network models, the activation function affects their ability to express nonlinear modeling of input features. Compared to ReLU [26], Leaky ReLU [27], PreLU [28], etc., Dynamic ReLU can better capture the nonlinear expression of input features, mainly because this function can dynamically adjust the parameters of the activation function based on input features. As shown in Figure 5, the dynamic activation function is a piecewise function whose core hyperfunction, θ ( x ) , is a function that dynamically adjusts the parameters of the activation function according to different inputs, and parameters θ 1 and θ 2 are obtained by the hyperfunction θ ( x ) on the input x . The hyperfunction θ ( x ) automatically adapts to the activation function by integrating the characteristics of various dimensions of the input, which can significantly improve the expressive ability of the network with a small amount of additional computation.
In the dynamic activation function, the implementation of the hyperfunction θ ( x ) is achieved using the Squeeze and Excitation [23] (SE) module. As shown in Figure 6, for the input features with a size of C × W × H , the global average pooling layer is first used for compression, then two fully connected layers (including ReLU) are used for processing, and finally the output result is constrained between ( 1 , 1 ) through the normalization layer, 2 σ ( x )     1 , where σ ( x ) is the sigmoid function.
There are three different forms of dynamic activation functions, specifically manifested as differences in sharing mechanisms in the spatial location and channel dimensions. We used the dynamic activation function, which only shares spatial positions, as the activation layer of the backbone network to enhance the ability of the network model to express pedestrian features.

3.4. Arcface Loss

In the context of feature learning, a suitable loss function plays a pivotal role in enabling the model to effectively extract valuable information from the available data. Consequently, this enhances the discriminative capabilities of the network model in extracting relevant features, thereby improving its overall performance.
The cross-entropy loss is widely used in image classification tasks. It calculates the probability of the extracted feature vectors belonging to each category through the softmax function and then calculates the loss. The network model is optimized by minimizing the difference between the true probability distribution and the predicted probability distribution. For the i -th learned feature f i , the cross-entropy loss is represented as Equation (4). In this equation, N represents the number of images in a batch, W k represents the weight vector of the corresponding category, and C represents the number of categories of pedestrians in the training set.
L 1 = 1 N i = 1 N log e W ( y i ) T f i k = 1 C e W k T f i
The cross-entropy loss function makes pedestrian images close to their respective categories, achieving classification results. However, the cross-entropy function only focuses on the accuracy of predicting the probability of correct labels while ignoring the differences in other incorrect labels, which leads to scattered pedestrian features being learned by the network model. To overcome the aforementioned shortcomings of cross-entropy loss, we introduce Arcface loss [5] into the supervised network model, which learns pedestrian features with large inter-class distances and small intra-class distances, thereby making the network model output more discriminative in terms of pedestrian features. We implemented the following three steps to improve the cross-entropy loss by increasing the inter-class distance while reducing the intra-class distance in the angle space.
The initial step was to standardize the feature vectors, making sure that the module length was one. This standardization ensures consistency between the model weights and the distribution of feature vectors, ensuring that the distance measurement between samples solely relies on their angles. Subsequently, the introduction of the additive angle interval, margin (m), defined the classification boundary between positive and negative samples. By establishing limits on the maximum angle between similar samples and the minimum angle between dissimilar samples, we narrowed the angle range for each classification, significantly enhancing the distance between classes. Lastly, we added the scaling factor, scale (s), which acts as the normalized coefficient for the feature vector. When multiplied by the embedded dimension, it facilitates a more rapid training process for the model and accelerates the convergence of the loss function.
The cross-entropy loss after introducing the Arcface loss is shown in Equation (5). In this equation, m represents the additive angle margin, and s is the scaling factor scale.
L 2 = 1 N i = 1 N log e s cos θ y i + m e s cos θ y i + m + k = 1 , k y i C e s cos θ k

4. Experiment and Analysis

4.1. Datasets and Evaluation Metric

We conducted experiments on the proposed person re-identification method on two mainstream datasets, Market1501 [29] and DukeMTMC-ReID [30], as shown in Table 1. The Market1501 dataset contains 1501 entity objects and 32,688 detected pedestrian rectangular boxes. Among them, the training set contains 751 entities and 12,936 images, while the test set contains 750 entities and 19,732 images; The DukeMTMC-ReID dataset contains 1404 entity objects and 36,411 pedestrian images (including 2228 query set images). The training set contains 702 entities and 16,522 images, while the test set contains 702 entities and 17,661 images. Image examples of the two datasets are shown in Figure 7. Among them, we selected eight images belonging to four identities (IDs) from each dataset, including four types of examples showing the front, back, side, and occlusions of pedestrians.
We evaluated the performance of the proposed person re-identification network model using the mean average precision (mAP) and cumulative match characteristic (CMC) as performance evaluation metrics. The CMC curve shows the probability of a correct match as the ranking of the query results increases. Here, Rank-m represents the hit rate of the search target in the first m bits of the sorted list returned by the algorithm. For example, Rank-1 refers to the first hit of the search target, and Rank-5 refers to the possibility of the search target appearing in the first five matching targets. The advantage of the CMC curve is that it provides an intuitive understanding of the performance of the model, especially when considering the first few matching cases. However, the CMC curve cannot be applied to image retrieval with multiple cameras. Therefore, Zheng et al. [31] proposed using the mAP to evaluate the retrieval performance with multiple cameras. The mAP considers the average recall rate of all queries by first calculating the accuracy of each query and then averaging these accuracies to comprehensively reflect the performance of the model.

4.2. Experimental Settings

The experiment was implemented on a machine with an NVIDIA TITAN V GPU using the PyTorch deep learning framework version 1.4.0. The backbone network used was ResNet50 [25], and the last fully connected layer and average pooling layer were removed. The input image underwent preprocessing operations such as random horizontal flipping and cropping, and the image size was adjusted to 256 × 128 . After normalization, it was input into the network model. The initial learning rate was 0.05, and after training for 30 epochs, the learning rate decayed to 1/10 of the original. A total of 60 epochs were trained; the final convolution stride was set to 2, and the batch size was set to 32. The hyperparameter m was set to 0.4, and s was set to 135. A Stochastic Gradient Descent (SGD) [32] optimizer was used for gradient updates, with weight decay set to 5 × 10 4 and momentum set to 0.9.

4.3. Comparison with Existing Methods

We compared the proposed person re-identification method with four existing person re-identification methods, and the results are shown in Table 2. Among them, the singular value decomposition network [33] (SVDNet) and pedestrian alignment network [34] (PAN) are person re-identification methods based on global feature learning which utilize dimensionality reduction, the orthogonalization of fully connected layers, and character alignment, respectively. The pose-driven deep convolutional (PDC) model [35] and attention-aware compositional network [36] (AACN) are person re-identification methods based on global and local feature learning which utilize pose estimation and alignment. Furthermore, DyReLU is short for Dynamic ReLU, as is the case below. It can be seen that the proposed person re-identification method has significantly improved performance on mainstream person re-identification datasets, indicating that the proposed method can obtain more robust pedestrian expression features in the feature extraction stage and can output more discriminative pedestrian features in the similarity measurement stage.

4.4. Ablation Study

4.4.1. Incorporating Normalization-Based Channel Attention Module

From Table 3, it can be seen that incorporating the Normalization-based Channel Attention Module into the ResNet50 network improved the performance of the network model on both datasets. Among them, the improvement is more significant on the Market1501 dataset, which may be related to differences in the dataset itself. On the Market1501 dataset, the Rank-1 hit rate increased by 1.07%, and the mAP increased by 1.18%. On the DukeMTMC-ReID dataset, the Rank-1 hit rate increased by 1.21%, and the mAP increased by 0.93%.

4.4.2. Using Dynamic ReLU

From Table 4, it can be seen that the use of dynamic activation functions slightly improved the performance of the network model on both datasets. Among them, on the Market1501 dataset, the Rank-1 hit rate increased by 0.04%, and the mAP increased by 0.73%. On the DukeMTMC-ReID dataset, the Rank-1 hit rate increased by 0.94%, and the mAP increased by 0.82%.

4.4.3. Introducing Arcface Loss

From Table 5, it can be seen that introducing Arcface loss into cross-entropy loss significantly improved the performance of the network model on both datasets. On the Market1501 dataset, the Rank-1 hit rate increased by 0.77%, and the mAP increased by 1.43%. On the DukeMTMC-ReID dataset, the Rank-1 hit rate increased by 0.78%, and the mAP increased by 1.49%.

4.4.4. Ablation Experiments to Verify the Effectiveness of Three Modules

To verify the effectiveness of various parts of the network model, we conducted ablation experiments on the Market1501 and DukeMTMC-ReID datasets. The Rank-1 and mAP obtained from each group of experiments are shown in Table 6. It can be seen that the proposed person re-identification method, which simultaneously introduces the NAM, dynamic activation function, and Arcface loss, performs the best on the Market1501 and DukeMTMC-ReID datasets, with Rank-1 improving by 1.28% and 1.4%, respectively; the MAP increased by 1.93% and 1.84%, respectively.

4.4.5. Ablation Experiments to Determine the Values of Hyperparameters m and s

Setting the margin (m) and scale (s) hyperparameters correctly is crucial for the effectiveness of Arcface loss. In order to achieve its best performance, we conducted ablation experiments on hyperparameters m and s on the Market1501 dataset. The Rank-1 and mAP obtained from each group of experiments are shown in Table 7 and Figure 8. It can be seen that for the proposed person re-identification method, the best performance is achieved when m and s are set to 0.4 and 135, with a 0.77% increase in the Rank-1 hit rate and a 1.43% increase in the mAP.

4.5. Visualization Analysis

The CMC evaluation curve of the proposed person re-identification method on the Market1501 dataset is shown in Figure 9. It can be seen that by incorporating Normalization-based Channel Attention Modules into the backbone network ResNet50, using dynamic activation functions for activation, and introducing Arcface loss into cross-entropy loss, the network model can perform better on the dataset. This indicates that the proposed method significantly improves expression ability in terms of pedestrian features and has stronger feature discrimination ability.
We tested the proposed person re-identification method on the Market1501 dataset. The query results returned by the method are shown in Figure 10. Among them, query is the image to be queried, and 1–10 is the query result returned from the gallery library. The image with a green superscript is a positive sample image, which means it belongs to the same pedestrian entity as the query image to be queried. The image with a red superscript is a negative sample image, which means it does not belong to the same pedestrian entity as the query image.
From the results in Figure 10, it can be seen that for simple frontal facial images (query 1 and query 2), the proposed person re-identification method successfully extracts rich information for feature description and recognition. For more challenging samples, such as images of people from the back (query 3 and query 4), the proposed person re-identification method can also distinguish and recognize them to a certain extent in all visually similar images. Moreover, for a few pedestrian images to be queried (query 5 and query 6), the proposed person re-identification method cannot achieve ideal recognition results, which may be due to its insufficient ability to capture small differences such as clothing changes and pedestrian postures by global features.

5. Conclusions

We propose a novel person re-identification method based on global feature learning. Firstly, we incorporated the Normalization-based Channel Attention Module into the backbone network ResNet50 to selectively obtain more critical image information and output more representative pedestrian expression features. Secondly, we applied the dynamic activation function to the activation layer of the backbone network ResNet50, dynamically adjusting the activation method based on input, which enhanced the nonlinear expression ability of the network model for pedestrian features. Finally, we introduced Arcface loss into cross-entropy loss, which increased the inter-class distance while reducing the intra-class distance in angle space, making the network model output more discriminative regarding pedestrian features. In the experimental stage, we evaluated the enhanced model on two popular datasets, Market1501 and DukeMTMC-ReID, and used the mean average precision (mAP) and cumulative match characteristic (CMC) as performance evaluation metrics to comprehensively evaluate the proposed method. The results show that compared to other existing methods, the Rank-1 of the proposed method increased by at least 2.79% and 3.61%, respectively, and the mAP correspondingly increased by at least 5.34% and 2.13%. These results indicate that the model can extract more robust pedestrian features, enhance the distinguishability of features, and ultimately achieve higher recognition accuracy. In the future, we will explore how to further optimize algorithms, improve recognition accuracy, and provide more effective solutions to practical problems by introducing advanced object detection technologies to improve the receptive field and understanding ability of algorithms, or by adopting rich data augmentation strategies and advanced training methods to enhance the generalization ability and robustness of network models.

Author Contributions

Conceptualization, Y.B.; methodology, Y.B.; formal analysis, Y.B. and R.W.; investigation, Y.B. and R.W.; writing—original draft preparation, Y.B.; writing—review and editing, Y.B. and Q.Z.; supervision, R.W., R.L. and M.W.; funding acquisition, R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Double First-Class Innovation Research Project for the People’s Public Security University of China, grant number 2023SYL08.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available online. These datasets were derived from the following public resources: [Market1501, DukeMTMC-ReID].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hu, M.; Zeng, K.; Wang, Y.; Guo, Y. Threshold-based hierarchical clustering for person re-identification. Entropy 2021, 23, 522. [Google Scholar] [CrossRef] [PubMed]
  2. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep learning for person re-identification: A survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 2872–2893. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, Y.; Shao, Z.; Teng, Y.; Hoffmann, N. NAM: Normalization-based attention module. arXiv 2021, arXiv:2111.12419. [Google Scholar]
  4. Chen, Y.; Dai, X.; Liu, M.; Chen, D.; Yuan, L.; Liu, Z. Dynamic relu. In European Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2020; pp. 351–367. [Google Scholar]
  5. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  6. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  7. Sun, Y.; Zhang, W.; Wang, R.; Li, C.; Zhang, Q. Person re-identification method based on channel attention mechanism. J. Beijing Univ. Aeronaut. Astronaut. 2022, 48, 881–889. [Google Scholar]
  8. Wang, F.; Zuo, W.; Lin, L.; Zhang, D.; Zhang, L. Joint learning of single-image and cross-image representations for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1288–1296. [Google Scholar]
  9. Zhang, X. Person re-identification based on global feature improvement. Comput. Syst. Appl. 2022, 31, 298–303. [Google Scholar]
  10. Zhao, H.; Tian, M.; Sun, S.; Shao, J.; Yan, J.; Yi, S.; Wang, X.; Tang, X. Spindle net: Person re-identification with human body region guided feature decomposition and fusion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1077–1085. [Google Scholar]
  11. Wang, Q.; Liu, Z.; Wang, M.; Zhao, X. Person re-recognition based on posture driven local feature alignment. Comput. Syst. Appl. 2023, 32, 268–273. [Google Scholar]
  12. Lin, Y.; Zheng, L.; Zheng, Z.; Wu, Y.; Hu, Z.; Yan, C.; Yang, Y. Improving person re-identification by attribute and identity learning. Pattern Recognit. 2019, 95, 151–161. [Google Scholar] [CrossRef]
  13. Zheng, L.; Yang, Y.; Hauptmann, A.G. Person re-identification: Past, present and future. arXiv 2016, arXiv:1610.02984. [Google Scholar]
  14. Du, Y.; Tang, M.; Zha, Y.; Song, H.; Liu, W. Research on person re-identification algorithm integrating attention mechanism and global feature network. J. Test Meas. Technol. 2023, 37, 37–42. [Google Scholar]
  15. Lin, R.; Wang, R.; Zhang, W.; Wu, A.; Sun, Y.; Bi, Y. Person Re-Identification Method Based on Dual Descriptor Feature Enhancement. Entropy 2023, 25, 1154. [Google Scholar] [CrossRef] [PubMed]
  16. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE computer society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  17. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  18. Chen, W.; Chen, X.; Zhang, J.; Huang, K. Beyond triplet loss: A deep quadruplet network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 403–412. [Google Scholar]
  19. Zhang, T.; Yi, Z.; Li, X. An improved algorithm for person re-identification based on global features. Prog. Laser Optoelectron. 2020, 57, 324–330. [Google Scholar]
  20. Wang, J.; Chen, H.; Zhang, D. Person re-identification lightweight network combining global and local features. In Proceedings of the 42nd Chinese Control Conference, Tianjin, China, 24–26 July 2023; pp. 863–867. [Google Scholar]
  21. Zhang, B.; Ma, J.; Zhang, S.; Li, C.; Zhong, Z. A person re-identification method using global and local correlation features. J. Electron. Meas. Instrum. 2022, 36, 205–212. [Google Scholar]
  22. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning; PMLR: New York, NY, USA, 2015; pp. 448–456. [Google Scholar]
  23. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  24. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics; JMLR Workshop and Conference Proceedings; PMLR: New York, NY, USA, 2011; pp. 315–323. [Google Scholar]
  27. Xu, J.; Li, Z.; Du, B.; Zhang, M.; Liu, J. Reluplex made more practical: Leaky ReLU. In Proceedings of the 2020 IEEE Symposium on Computers and Communications (ISCC), Rennes, France, 7–10 July 2020; pp. 1–7. [Google Scholar]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  29. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable person re-identification: A benchmark. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  30. Zheng, Z.; Zheng, L.; Yang, Y. Unlabeled samples generated by gan improve the person re-identification baseline in vitro. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3754–3762. [Google Scholar]
  31. Chen, Y.C.; Zheng, W.S.; Lai, J.H.; Yuen, P.C. An asymmetric distance model for cross-view feature mapping in person reidentification. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 1661–1675. [Google Scholar] [CrossRef]
  32. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  33. Sun, Y.; Zheng, L.; Deng, W.; Wang, S. Svdnet for pedestrian retrieval. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3800–3808. [Google Scholar]
  34. Zheng, Z.; Zheng, L.; Yang, Y. Pedestrian alignment network for large-scale person re-identification. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 3037–3045. [Google Scholar] [CrossRef]
  35. Su, C.; Li, J.; Zhang, S.; Xing, J.; Gao, W.; Tian, Q. Pose-driven deep convolutional model for person re-identification. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3960–3969. [Google Scholar]
  36. Xu, J.; Zhao, R.; Zhu, F.; Wang, H.; Ouyang, W. Attention-Aware Compositional Network for Person Re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
Figure 1. Overall architecture of the proposed method.
Figure 1. Overall architecture of the proposed method.
Entropy 26 00436 g001
Figure 2. Normalization-based Channel Attention Module in the proposed method.
Figure 2. Normalization-based Channel Attention Module in the proposed method.
Entropy 26 00436 g002
Figure 3. Specific workflow of the NAM.
Figure 3. Specific workflow of the NAM.
Entropy 26 00436 g003
Figure 4. Residual structure of NAM-ResNet. (a) ResNet; (b) NAM-ResNet.
Figure 4. Residual structure of NAM-ResNet. (a) ResNet; (b) NAM-ResNet.
Entropy 26 00436 g004
Figure 5. Schematic diagram of dynamic activation function.
Figure 5. Schematic diagram of dynamic activation function.
Entropy 26 00436 g005
Figure 6. Dynamic ReLU (spatial-shared only).
Figure 6. Dynamic ReLU (spatial-shared only).
Entropy 26 00436 g006
Figure 7. Image examples of the two datasets. (a) Market1501; (b) DukeMTMC-ReID.
Figure 7. Image examples of the two datasets. (a) Market1501; (b) DukeMTMC-ReID.
Entropy 26 00436 g007
Figure 8. Results of ablation experiments on Arcface loss hyperparameters m and s. (a) The variation in mAP with different values of m while keeping the value of s equal to 135; (b) the variation in mAP with different values of s while keeping the value of m equal to 0.4.
Figure 8. Results of ablation experiments on Arcface loss hyperparameters m and s. (a) The variation in mAP with different values of m while keeping the value of s equal to 135; (b) the variation in mAP with different values of s while keeping the value of m equal to 0.4.
Entropy 26 00436 g008
Figure 9. CMC evaluation curves of previous and proposed methods on Market1501 dataset.
Figure 9. CMC evaluation curves of previous and proposed methods on Market1501 dataset.
Entropy 26 00436 g009
Figure 10. Visualization of person re-identification results using the proposed method.
Figure 10. Visualization of person re-identification results using the proposed method.
Entropy 26 00436 g010
Table 1. Datasets’ conditions.
Table 1. Datasets’ conditions.
DatasetsYear of DisclosureNumber of CamerasNumber of IDsNumber of Images
Market150120156150132,668
DukeMTMC-ReID20178140436,411
Table 2. Performance comparison of different models on Market1501 and DukeMTMC-ReID datasets.
Table 2. Performance comparison of different models on Market1501 and DukeMTMC-ReID datasets.
MethodMarket1501DukeMTMC-ReID
Rank-1mAPRank-1mAP
SVDNet [33]82.3062.1076.7056.80
PAN [34]82.8163.3571.5951.51
PDC [35]84.1463.41--
AACN [36]85.9066.8776.8459.25
baseline87.4170.2879.0559.54
baseline + NAM + DyReLU + Arcface loss (ours)88.6972.2180.4561.38
Table 3. Performance comparison of models embedded with Normalization-based Channel Attention Module on Market1501 and DukeMTMC-ReID datasets.
Table 3. Performance comparison of models embedded with Normalization-based Channel Attention Module on Market1501 and DukeMTMC-ReID datasets.
MethodMarket1501DukeMTMC-ReID
Rank-1mAPRank-1mAP
baseline87.4170.2879.0559.54
baseline + NAM88.4871.4680.2660.47
Table 4. Performance comparison of models with Dynamic ReLU on Market1501 and DukeMTMC-ReID datasets.
Table 4. Performance comparison of models with Dynamic ReLU on Market1501 and DukeMTMC-ReID datasets.
MethodMarket1501DukeMTMC-ReID
Rank-1mAPRank-1mAP
baseline87.4170.2879.0559.54
baseline + DyReLU87.4571.0179.9960.36
Table 5. Performance comparison of models with Arcface loss on Market1501 and DukeMTMC-ReID datasets.
Table 5. Performance comparison of models with Arcface loss on Market1501 and DukeMTMC-ReID datasets.
MethodMarket1501DukeMTMC-ReID
Rank-1mAPRank-1mAP
baseline87.4170.2879.0559.54
baseline + Arcface loss88.1871.7179.8361.03
Table 6. Results of ablation experiments on Market1501 and DukeMTMC-ReID datasets.
Table 6. Results of ablation experiments on Market1501 and DukeMTMC-ReID datasets.
MethodMarket1501DukeMTMC-ReID
Rank-1mAPRank-1mAP
baseline87.4170.2879.0559.54
baseline + NAM88.4871.4680.2660.47
baseline + DyReLU87.4571.0179.9960.36
baseline + Arcface loss88.1871.7179.8361.03
baseline + NAM + DyReLU88.3671.6080.1260.78
baseline + NAM + Arcface loss88.3371.5079.9361.10
baseline + DyReLU + Arcface loss88.1271.8880.4460.86
baseline + NAM + DyReLU + Arcface loss88.6972.2180.4561.38
Table 7. Ablation experimental results for the hyperparameters m and s of Arcface loss on Market1501 dataset.
Table 7. Ablation experimental results for the hyperparameters m and s of Arcface loss on Market1501 dataset.
s = 135Rank-1mAP m = 0.4Rank-1mAP m = 0.4Rank-1mAP
m s s
0.187.4170.301587.2670.309088.1271.30
0.286.3769.433086.2868.6910586.7369.98
0.387.8071.504587.1470.5012088.1271.22
0.488.1871.716087.5070.1513588.1871.71
0.587.2368.907587.9570.8415087.8370.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bi, Y.; Wang, R.; Zhou, Q.; Lin, R.; Wang, M. Enhancing Person Re-Identification through Attention-Driven Global Features and Angular Loss Optimization. Entropy 2024, 26, 436. https://doi.org/10.3390/e26060436

AMA Style

Bi Y, Wang R, Zhou Q, Lin R, Wang M. Enhancing Person Re-Identification through Attention-Driven Global Features and Angular Loss Optimization. Entropy. 2024; 26(6):436. https://doi.org/10.3390/e26060436

Chicago/Turabian Style

Bi, Yihan, Rong Wang, Qianli Zhou, Ronghui Lin, and Mingjie Wang. 2024. "Enhancing Person Re-Identification through Attention-Driven Global Features and Angular Loss Optimization" Entropy 26, no. 6: 436. https://doi.org/10.3390/e26060436

APA Style

Bi, Y., Wang, R., Zhou, Q., Lin, R., & Wang, M. (2024). Enhancing Person Re-Identification through Attention-Driven Global Features and Angular Loss Optimization. Entropy, 26(6), 436. https://doi.org/10.3390/e26060436

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop