Next Article in Journal
The Use of Infrared Thermography to Develop and Assess a Wearable Sock and Monitor Foot Temperature in Diabetic Subjects
Previous Article in Journal
A Review of Recent Distributed Optical Fiber Sensors Applications for Civil Engineering Structural Health Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Feature Pyramid Network: A Heavily Occluded Pedestrian Detection Network Based on ResNet

1
School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing 100044, China
2
Shanghai Aerospace Control Technology Institute, Shanghai 201109, China
3
Beijing Xinghang Mechanical-Electrical Equipment Co., Ltd., Beijing 100074, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1820; https://doi.org/10.3390/s21051820
Submission received: 27 January 2021 / Revised: 22 February 2021 / Accepted: 2 March 2021 / Published: 5 March 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
The existing pedestrian detection algorithms cannot effectively extract features of heavily occluded targets which results in lower detection accuracy. To solve the heavy occlusion in crowds, we propose a multi-scale feature pyramid network based on ResNet (MFPN) to enhance the features of occluded targets and improve the detection accuracy. MFPN includes two modules, namely double feature pyramid network (FPN) integrated with ResNet (DFR) and repulsion loss of minimum (RLM). We propose the double FPN which improves the architecture to further enhance the semantic information and contours of occluded pedestrians, and provide a new way for feature extraction of occluded targets. The features extracted by our network can be more separated and clearer, especially those heavily occluded pedestrians. Repulsion loss is introduced to improve the loss function which can keep predicted boxes away from the ground truths of the unrelated targets. Experiments carried out on the public CrowdHuman dataset, we obtain 90.96% AP which yields the best performance, 5.16% AP gains compared to the FPN-ResNet50 baseline. Compared with the state-of-the-art works, the performance of the pedestrian detection system has been boosted with our method.

1. Introduction

Pedestrian detection has been a hot topic in the field of computer vision, widely applied in recent years in both research and industry applications such as automatic driving, video surveillance, human-machine interaction, etc. Many scholars have endeavored to solve the problem in these application fields with great results [1,2]. The problem of pedestrian detection can be simplified as a binary classification between foregrounds (pedestrians) and backgrounds. However, pedestrians are usually exposed to complex environments. For instance, images captured by intelligent vehicles driving in urban scenarios have intricate backgrounds, which will bring about difficulties for pedestrian detection. Meanwhile, pedestrian targets in natural environments are inevitably occluded by each other. Occlusion problems can be classified as inter-class occlusion and intra-class occlusion. The former refers to targets being occluded by objects of different classes, the latter occurs when targets of the same kind are occluded by each other. Both of them remain a huge challenge for pedestrian detection tasks in practice. Especially in the case of highly crowded scenes, it is difficult for pedestrian detectors to precisely detect targets. For example, the CrowdHuman dataset [3] has an average of 23 pedestrians in each image, with various levels of occlusion. It is difficult for general object detectors to solve this kind of problem.
Feature extraction is the key to pedestrian detection. Most existing feature extraction methods can be summarized as traditional hand-crafted features and convolutional neural network (CNN) features. The hand-crafted feature extraction methods such as HOG [4], Haar [5], LBP [6], etc., have to be artificially distinguished whether they are representative or not. Furthermore, it is generally difficult to design features with robustness in practical and complex backgrounds. The CNN features generally refer to features extracted by CNNs, such as AlexNet [7], VGGNet [8], STINet [9], etc., which can extract more complicated and semantic features and result in higher detection accuracy. Therefore, CNN has a wider range of applications in pedestrian detection. However, the features of heavily occluded pedestrians are too similar to distinguish for the detector. To better distinguish the features, Zhang et al. [10] proposed that pedestrians could be respectively divided into several parts, whose features were then extracted partially for further fusion. However, the method requires a much longer time to extract and fuse features which results in a cumbersome training process.
In this paper, we propose a novel scheme called MFPN for heavily occluded pedestrian detection. The main contributions of this paper are two-fold: (1) We propose a novel feature extraction network called double FPN integrated with ResNet (DFR) to enhance the semantic information and contours of occluded pedestrians, and to simplify the network structure and the parameters. (2) We introduce the concept of repulsion loss of minimum (RLM) to keep predicted boxes away from the ground truths of the other pedestrians, which can also monitor the learning process of predicted boxes. The repulsion loss [11] is introduced based on original loss, which can not only reduce the loss between predicted boxes and ground truths but also repel the predicted boxes from their surrounding ground truths which are not their targets. The proposed MFPN can solve the heavy occlusion issue existing in pedestrian detection. Furthermore, the object detector with our RLM can achieve higher detection accuracy.
The remainder of this paper is organized as follows: In Section 2, the methods of pedestrian detection in crowded scenes are introduced. Section 3 describes our crucial MFPN network. Section 4 introduces our experimental processes and discusses the results in detail. Conclusions are presented in Section 5.

2. Related Works

Deep learning has been increasingly utilized in computer vision tasks recently. Especially the emergence of CNN has resulted in a breakthrough in the research on object detection [12,13,14,15] and other computer vision tasks [16]. Object detection methods based on deep learning are classified into two categories: one-stage detection and two-stage detection.
The one-stage detectors [12,17,18,19,20,21,22,23], represented by YOLO and SSD, have broken through the detection speed bottleneck of two-stage detectors. The two-stage detectors [24,25,26,27] represented by R-CNN, Fast R-CNN, and Faster R-CNN and based on region proposal network (RPN) have higher detection accuracy but lower speed compared to the one-stage detectors. As the key of the two-stage algorithms, RPN is utilized to generate anchors of different sizes and scales on feature maps. Then, post-processing methods such as non-maximum suppression (NMS) are used to remove duplicate proposals. Although there are no RPNs in the one-stage algorithm, it uses predefined anchors, which can be regressed on feature maps directly. These detection networks can be applied to specific object detection tasks including pedestrian detection, vehicle detection, face detection, etc.
Although these methods are reported to achieve outstanding performance on some datasets, low detection accuracy or a high number of missed detections will occur in the heavily occluded pedestrian detection task. As mentioned in the previous paragraph, the predictions are likely to be mistakenly suppressed by NMS since pedestrians may heavily overlap with each other. Optimized NMS solutions to the problem have been found such as Soft-NMS [28], Softer-NMS [29], Adaptive-NMS [30], etc. Soft-NMS and Softer-NMS were proposed to reduce missed detection of traditional NMS in heavily occluded object detection. Adaptive NMS was proposed to set the threshold of confidence based on the density of targets and to adapt to different levels of occlusions. However, such heuristic variants of NMS cannot be applied flexibly under different circumstances. Furthermore, they cannot solve the heavy occlusion problem fundamentally.
Some works have addressed the occlusion problem from other different perspectives, such as loss function, part-based anchors, and improved network:
(1)
Loss Function. Wang et al. [11] proposed repulsion loss to optimize the original loss function, attempting to solve the problem of pedestrian detection in crowded scenes. Zhang et al. [10] proposed aggregation loss, which enabled the bounding boxes to be close to ground truths and locate compactly. Although these loss functions can help improve the detection accuracy to some degree, it is difficult to recall overlapped proposals due to the use of traditional NMS.
(2)
Part-based Anchors. Zhou et al. [31] presented a method of pedestrian detection: whereby the whole body and visible parts of pedestrians were respectively located by regressing two bounding boxes. Chi et al. [32] proposed Pedhunter which can handle occlusion in pedestrian detection. However, the methods are complicated and take a long time.
(3)
Improved Network. Pang et al. [33] proposed the mask-guided attention network (MGAN) which emphasized visible parts of pedestrians and adjusted overall features to suppress invisible areas and detect occluded pedestrians. Zhang et al. [10] proposed occlusion-aware R-CNN, which divided pedestrians into several parts and extracted features for further fusion. Wang et al. [34] used compositional convolutional neural networks to detect objects. Wu et al. [35] exploited the local temporal context of pedestrians in videos and proposed a tube feature aggregation network (TFAN) aiming to detect occluded pedestrians. However, these methods are too intricate to implement or not robust enough to heavily occluded pedestrians.
Apart from the methods mentioned above, Cao et al. [36] paid more attention to predicted bounding boxes with worse location precision and extracted more contextual information around objects to detect pedestrians. However, this method was only validated on the CityPersons and Caltech datasets. Therefore, experiments carried out on more crowded datasets such as CrowdHuman cannot necessarily be expected to produce the same results.

3. Materials and Methods

In this section, we will elaborate our approach for heavily occluded pedestrian detection at length. Our proposed network MFPN is motivated by the following considerations:
(1)
Intra-class occlusion. Intra-class occlusion and inter-class occlusion generally exist in pedestrian detection. For intra-class occlusion, the features of targets in the same class are equally important. Different levels of extracting similar features and expressing these similar features will influence the later detection. For example, there will be more false positives occur if the expression ability of features is low. The feature extraction network should not only extract and fuse high-level semantic features but also retain the contour information.
(2)
Loss function related to occluded pedestrians. The loss function also influences the detection accuracy of occluded targets in pedestrian detection systems. To repel predicted boxes from surrounding ground truths of other pedestrians in crowds, we introduce the repulsion loss [11] integrated with our minimum of two losses to separate irrelevant boxes.
We formulate the details of our approach as follows: In Section 3.1, we introduce the whole architecture network based on CNN. Section 3.2 describes our crucial DFR network. Finally, Section 3.3 depicts our novel loss in detail.

3.1. Architecture Network

The overview of our MFPN network is shown in Figure 1, which is mainly composed of four parts: DFR, RPN, RoI Align and RLM. In this paper, we use the two-stage detection method and choose ResNet50 as our backbone, followed by our DFR, shown in Figure 2, which is an improved network compared with the traditional FPN. As mentioned in the related works, the RPN of two-stage detection is responsible for generating various anchors on multi-scale feature maps. Some proposals regressed by anchors have been displayed in the RPN module (blue) in Figure 1. The RoI Align module can transform the feature maps of different sizes to the same size combined with the feature maps from fout_1 to fout_5 and information of scales. Then the fully-connected network can predict offsets towards one proposal. The last one is our RLM, composed of LossMinimum and LossRep. The loss function diagram will be shown in detail in Figure 3. The LossRep can be applied to the network flexibly.

3.2. DFR Network

Effectively extracting and fusing features are of great importance for pedestrian detection tasks. Therefore, feature maps extracted by feature extraction network directly affect the detection accuracy. To achieve the purpose of integrating high-level features with more abstract semantic information and low-level features with more contour information, Lin et al. [37] proposed a pioneering idea: a top-down pathway that can be combined with bottom-up networks to extract and fuse multi-scale features. Following the idea, PANet [38] added an extra bottom-up path on top of FPN. Tan et al. [17] proposed a weighted BiFPN structure that can be scaled to fuse features. The cross-scale connection is proposed in BiFPN to integrate the above network structure, as shown in Figure 2a. Although these works seem to optimize network extraction and fusion and get better accuracy, there are some defects to be overcome: (1) It is still a limit to extracting features of heavily occluded targets for these networks. (2) The overwhelming parameters and computations decrease the efficiency of networks. (3) The networks cannot be employed flexibly.
Inspired by previous works, we propose the DFR network, an effective multi-scale feature extraction and fusion network, shown in Figure 2b. The overall DFR is composed of two components, ResNet50 and the double FPN network, which is an improved network compared to ResNet and FPN. We choose ResNet50 as the bottom-up path to extract features because experiments reported in [39] reveal that ResNet50 has low error rates compared with ResNet18 and ResNet34 and requires half the number of parameters compared with ResNet101 and ResNet152. However, ResNet50, ResNet101, and ResNet152 can achieve almost the same accuracy. Our proposed double FPN with ResNet50 (DFR, Figure 2b) makes the following improvements based on the two layers of BiFPN (Figure 2a): First, the connection between the initial input and output of the same level is removed, and then the weight in the corresponding feature fusion should also be removed, which has the advantage of reducing initialization parameters. Second, the separable convolution of the original BiFPN initially introduced in [40], is replaced by a general convolution. Although the network architecture with separable convolution has fewer parameters, experimental results indicate our network with separable convolution cannot surpass the general convolution in all criteria. Meanwhile, in our experiments, the feature maps (Figure 4) extracted by the general-convolution network are clearer than those of separable convolution, which means our network with general convolution can better extract and fuse the features. Therefore, general convolution is better for our network. The modified network structure is shown in Figure 2b.
Suppose the feature map fin_i extracted by ResNet50 at layer i with dimension hi × ωi × ci, i . Let F 1 , 2 i (pink rectangle background) be the novel function that accomplishes the mapping from fin_i to fout_i, denoted as:
F 1 i = f i n _ i f i n _ i , F 2 i = f i n _ i f o u t _ i w h e r e , f i n _ i = h i × w i × 2 i + C 1 , f o u t _ i = h i × w i × 2 1 + C 1 ; a n d , ( h i , w i ) = ( H , W ) / 2 i + C 2 , i = 1 , 2 , 3 , 4 , 5
where the two same operations F 1 i and F 2 i are composed of two operations respectively. Parameters H and W are the height and width of the input image, generally H = 800 and W = 1400 for the CrowdHuman dataset. h i × w i × c i is the dimension of feature maps, where ci_in = 2i+C1, ci_out = 21+C1 are the numbers of channels, and hi × ωi is the resolution of feature maps. Here, we let C1 = 7, C2 = 1.
For brevity, we only demonstrate F 1 i in detail, which is made up of two parts. The first one is from fin_i (i = 1, 2, 3, 4, 5) to fup_i (i = 2, 3, 4), which can be defined as:
f u p _ i = F C o n v ( ω i n _ i f i n _ i + ω ( i + 1 ) i F U p s a m p l e ( f u p _ i + 1 ) ω i n _ i + ω ( i + 1 ) i ) , i = 2 , 3 f u p _ 4 = F C o n v ( ω i n _ 4 f i n _ 4 + ω 54 F U p s a m p l e ( f i n _ 5 ) ω i n _ 4 + ω 54 )
where FUpsample and FConv are the operations of upsample and convolution. Notably, the channel of the feature map fin_5 is the same as fin_4, but the resolution of the former is the quarter of the latter because fin_5 is obtained by max pooling fin_4. The channels of feature maps except for fin_2−fin_5 are all 256 for the sake of adding different layers.
The second part is from f u p _ i   ( i = 2 , 3 , 4 ) to f i n _ i   ( i = 1 , 2 , 3 , 4 , 5 ) , which can be defined as:
f i n _ i = F C o n v ( ω u p _ i f u p _ i + ω ( i 1 ) i F M a x p o o l i n g ( f i n _ i 1 ) ω u p _ i + ω ( i 1 ) i ) , i = 2 , 3 , 4 f i n _ 5 = F C o n v ( ω i n _ 5 f i n _ 5 + ω 45 F M a x p o o l i n g ( f i n _ 4 ) ω i n _ 5 + ω 45 ) , f i n _ 1 = F C o n v ( ω i n _ 1 f i n _ 1 + ω 21 F U p s a m p l e ( f u p _ 2 ) ω i n _ 1 + ω 21 )
where FMaxpooling is the max pooling operation. As shown in Figure 2b, the Equation (1)–(3) only demonstrate the operation F 1 i reaching out intermediate results of feature maps f i n _ 1 f i n _ 5 , and the final results fout_i will be obtained through the same operation F 2 i .
Previous works resize feature maps in different layers with a 1 × 1 convolution kernel and sum them up directly, resulting in different layers corresponding to the same importance. However, different layers contribute to the output feature unequally. The weighted feature fusion can learn the importance of feature maps in each layer. Each resized feature map (upsample or max pooling) will be summed up with the next layer’s feature map. The weighted feature fusion has been described in Equations (2) and (3), where ω is the weight corresponding to the feature map. Specifically, ωin_i and ω(i+1)i are on the top-down pathway; ωup_j and ω(j−1)j are on the bottom-up pathway, which are all the learned weights representing the importance of feature maps for feature fusion. After each feature fusion, Fconv (general convolution) is used to process the feature maps.

3.3. Repulsion Loss of Minimum

In this paper, we adopted the paradigm of [41]. That is, for one proposal, two or more bounding boxes can be predicted, which can avoid the situation when bounding boxes are heavily overlapped and only one box survives after the post-processing of NMS. The loss function in pedestrian detection is composed of regression loss and classification loss. Based on the paradigm, we propose the regression loss termed Repulsion Loss of Minimum (RLM), an improved method of loss function compared to the original in [41], which only uses the first part of our loss function. The classification loss follows the common definition in object detection. In this section, we will detail the RLM loss function to tackle the occlusion problem in pedestrian detection. The repulsion loss of minimum is made up of two components, defined as:
L o s s R L M = L o s s M i n i m u m + L o s s R e p
As shown in Figure 3a, image, which can not be detected easily and precisely. Let two people, for example PED1 and PED2, be the pedestrians we want to detect. The red and the green boxes are their ground truths, respectively. The blue box is the proposal responsible for locating PED1 and PED2. We can get two sets of offsets towards the proposal further calculate the locations of Pred box1 and Pred box2, described as:
( d x i , d y i , d w i , d h i ) , i ( 1 , 2 )
Which ground truth of the pedestrian do the purple Pred box1 and the yellow Pred box2 correspond to, respectively? There is the solution that for the two boxes predicted by one proposal, we can calculate a total of two losses corresponding to two pedestrians, respectively, and choose the minimum of the loss. In this case, the pedestrians corresponding to the two boxes are the instances to be predicted. As shown in Figure 3b, the purple Pred box1 and yellow Pred box2 are predicted from the same blue proposal. There are two cases of predictions matching specific instances. The first one is that the purple Pred box1 needs to predict PED1 which the red ground truth (gt1) represents, and the loss between the two boxes is denoted as l o s s 11 ; the yellow Pred box2 needs to predict PED2 which the green ground truth (gt2) represents, and the loss between the two boxes is denoted as l o s s 12 . Therefore the total loss of the first case is:
l o s s 1 = l o s s 11 + l o s s 12
The second one is that the purple Pred box1 needs to predict PED2 which the green ground truth (gt2) represents, and the loss between the two boxes is denoted as l o s s 21 ; the yellow Pred box2 needs to predict PED1 which the red ground truth (gt1) represents, and the loss between the two boxes is denoted as l o s s 22 . The total loss of the second case is:
l o s s 2 = l o s s 21 + l o s s 22
If loss1 > loss2, the second case is selected as the final prediction result, that is:
L o s s M i n i m u m = M i n ( ( l o s s 11 + l o s s 12 ) , ( l o s s 21 + l o s s 22 ) )
vice versa. The smaller the loss is, the better the object detector is. That is just where the term “loss of minimum” comes from.
Next is the second LossRep which is firstly proposed in [11]. The proposed LossRep is to keep the two predicted boxes (Pred box1 and Pred box2) away from ground truths of other surrounding pedestrians. In this paper, with the dataset CrowdHuman, there are only two classes: pedestrian and background, so the parameter K, the number of pedestrians one proposal predicts, should be set as 2. Let P + = {P1, P2, …, Pn} represents the positive proposals regressed from anchors. As shown in Figure 3b, the LossRep is defined as:
L o s s R e p = l o s s ( g t 3 , P r e d b o x 1 ) + l o s s ( g t 3 , P r e d b o x 2 ) )
Generally, for the n-th proposal Pn, its repulsion ground truth is defined as the ground truth pedestrian with which it has the largest IoU region except for its designated two targets:
G R e p P n = arg max G G   \   { G 1 , 2 P n } I o U ( G , P n )
where G = {G} is denoted as the set of all ground truths in the image, G 1 , 2 P n = arg max G G   I o U ( G , P n ) are the two ground truth pedestrians corresponding to the proposal Pn. The LossRep is between the ground truth G R e p P n and two predicted boxes B 1 , 2 P n regressed by Pn, denoted as:
L o s s R e p = P n P + Smooth l n ( I o G ( B 1 , 2 P n , G R e p P n ) ) 2 P +
where:
I o G ( B , G ) = a r e a ( B G ) a r e a ( G ) [ 0 , 1 ]
and:
Smooth l n = ln ( 1 x ) x σ x σ 1 σ ln ( 1 σ ) x > σ
Smoothln is a smoothed ln function, which is continuously differentiable in (0,1). The parameter σ is the smooth parameter to adjust the sensitiveness of the repulsion loss to outliers. From Equations (11) and (13), we can find that the more the proposal tends to overlap with a non-target ground truth pedestrian, the larger penalty the LossRep will add to the bounding box regressor. In this way, the LossRep can effectively prevent the predicted bounding boxes from moving to their neighboring pedestrians which are not their targets. Therefore, we expect a higher LossRep, which suggests that the predicted boxes are far from other ground truths, further lead to detect the heavily occluded pedestrians correctly.

4. Experiments

In this section, we will compare our method with existing methods in several aspects to verify the effectiveness of our method. The experiment section is organized as follows: (1) We will introduce some basic settings of our experiments: datasets, detailed settings, and experiment platform. (2) We compare the feature maps extracted by our proposed network DFR and previous network, ResNet50 and FPN. (3) The two-stage detector with our method is compared with some other methods on the CrowdHuman and CityPersons datasets.

4.1. Datasets

We use the CrowdHuman and CityPersons datasets to evaluate the effectiveness of our method. The CrowdHuman dataset [3] has a large scale, containing 15,000, 4370, and 5000 images for training, validation, and testing, respectively, rich annotations and high diversity. There are about 23 pedestrians in each image with various occlusion scenes. Compared with other datasets, like COCOPerson, Caltech, and CityPersons, the CrowdHuman dataset has more pedestrians in each image, which is a challenge for any pedestrian detector. The CityPersons dataset is widely used in pedestrian detection, which is built on the Cityscapes dataset, a dataset for the task of semantic segmentation in urban street scenes. It includes 2975, 500 and 1525 images for training, validation and testing, respectively. Compared with the CrowdHuman dataset, there are 7 pedestrians in each image. Therefore, experimental results on the two datasets can manifest the effectiveness of our method more persuasively.

4.2. Detailed Settings

We use the frequently-used two-stage algorithm of Faster RCNN for the baseline network. The ResNet50 pre-trained on ImageNet is adopted for all the experiments, followed by double FPN. We use RoI Align instead of RoI pooling in this network. For every experiment, we train in total 35 epochs. As for the learning rate, we adopt the method of warming up: every time we train from scratch, the learning rate is gradually and linearly increasing from 0.0001 to 0.001 from iteration 1 to iteration 800 and then remains unchanged. By the 24th epoch, the learning rate is changed to one-tenth of the original, and one percent in the 28th epoch, i.t. 0.00001. In our work, due to the multi-size of feature maps, we just need to generate three anchors with different scales such as {1:1, 1:2, 2:1} for both datasets. In the train and test process, K = 2. The sampling ratio of positive to negative proposals for the RoI branch is 1:1 for the CrowdHuman dataset and 1:3 for the CityPersons dataset. All box overlap IoU thresholds are set as 0.5.

4.3. Experiment Platform

We implement our network with the PyTorch framework running on a PC equipped with an Intel® CoreTM i7-6500U CPU @ 2.50 GHz and an NVIDIA GTX1080Ti GPU. The machine is running Linux Ubuntu 16.04 with NVIDIA CUDA 10.1 and cuDNN 7.0.

4.4. Comparison of Feature Maps

In this section, we compare in detail the feature maps extracted by our ResNet50 and double FPN with counterparts extracted by other network structures. For the sake of fairness, an image of the CrowdHuman dataset has been randomly selected for the experiment, and different network structures are used to compare the extracted feature maps with the maximum scale. Table 1 shows different combinations of network structures. Figure 4 shows the feature maps extracted by these networks. We determine the quality of feature maps according to their clarity. For the sake of our explanation, we divide the features of persons into two parts, boxed up by white and black rectangles respectively. The top (bottom) represents the smaller (bigger) targets in the back (front) row. We can find the clarity of contours of these targets are really different in various feature maps. Although it is much easier to extract obscure features of targets in the back row (heavily occluded, white marked), the features extracted by our network can be more separated and clearer, especially those occluded persons in the back row. The contours of those “bigger’’ persons in the front row (light occluded, black marked) are also separated rather than mixed, which can help the detector learn the location of persons.

4.5. Ablation Study

In this section, we carry out the ablation experiments on the CrowdHuman dataset. Table 2 shows the experimental results of the proposed approach in previous works, including BiFPN, Separable Convolution, and our double FPN and RLM. The method without “separable” uses general convolution. The baseline network is ResNet50 and FPN [37] which uses NMS for post-processing. The remaining networks in Table 2 all utilize Set-NMS [41]. Obviously, our method, which achieves 90.96% AP, 40.24% MR−2, 83.12% JI, yields the best performance on the CrowdHuman dataset. Compared to the baseline, our approach achieves 5.16%, 2.66%, and 3.32% improvements in AP, MR−2, and JI, respectively. The experimental results indicate that our proposed method without RLM is equipped with good performance in all criteria: 0.58%, 1.68%, and 0.92% improvements even compared with the state–of–the–art method [41]. The improvement indicates that our network can detect more objects accurately although there are heavily occluded targets. At the same time, the improvement of MR−2 demonstrates that more false predictions are not introduced. For the separable convolution, the experimental result manifests that the performance of our network with general convolution is all better in three criteria than using separable convolution. Moreover, the results show RLM affects little our network, which only obtains improvements less than 0.5% in all criteria. The performance of our network without RLM decreases by 0.08%, 0.28% and 0.2% respectively. Higher values of AP and JI indicate better performance, which is in contrast to the MR−2.

4.6. Comparison of Previous Works

We compare our method with some previous methods on the CrowdHuman and CityPersons datasets. We use three metrics to evaluate our method, AP, MR−2 and JI. AP is the abbreviation of average precision. AP represents both the precision and recall ratios of the detection results. A larger AP means better performance. MR−2, generally used in pedestrian detection, is short for log-average miss rate on false positive per image (FPPI) in [10−2, 100]. MR−2 is sensitive to false positives (FPs). A smaller MR−2 indicates better performance. JI evaluates how much the predicted boxes overlap with ground truths. A larger JI represents better performance.
The comparison results are listed in Table 3. For the sake of fairness, the methods listed above except ours are based on FPN. We set the IoU threshold of NMS as 0.5 for post-processing. As shown in Table 3, our method still obtains the best performance in all criteria, with 90.96% AP, 40.24% MR−2 and 83.12%. Compared with Cascade R-CNN, our method obtain 5.36% AP, 2.76% MR−2 and 2.52% JI gains on the validation set of CrowdHuman. We achieve higher improvements, 6.26% AP and 9.46% MR−2 compared with adaptive NMS. On the CityPersons dataset, our method also performs better than other methods. Although our method improves little, it indicates the good robustness of our method. As shown in Table 3, higher AP and JI mean that our method can detect more pedestrians correctly. Meanwhile, the improved MR−2 suggests that our method cannot introduce more false positives compared with other methods.
Figure 5 displays some detection results by our method. We can see that pedestrians in images can be detected correctly even though there is a heavy occlusion problem. Furthermore, we show the comparisons of the detection results between the ResNet50 and FPN baseline and our method in Figure 6.
We have boxed up the heavily occluded pedestrians correctly detected by our method with red solid boxes but not detected by the baseline with red dashed in contrast. That is, the dashed boxes are missed detections. Consequently, our method can deal with the heavy occlusion problem in pedestrian detection.

5. Conclusions

In this paper, we propose an effective network called double FPN based on ResNet, (abbreviated to DFR network), which can effectively extract and fuse features of images. It can not only extract rich and semantic information but also keep a complete contour, which improves the performance of the network towards heavily occluded targets. Besides, we put forward a new kind of loss function named repulsion loss of minimum, which can solve the occlusion from another perspective. Combining these two ideas, our network has achieved good performance on the CrowdHuman and CityPersons datasets.

Author Contributions

Funding acquisition, W.Y., Y.S. and Z.W.; Project administration, W.Y., Y.S. and Z.W.; Investigation, W.Y.; Supervision, X.S. and Y.S.; Software, Q.W. and Y.X.; Validation, X.S. and Z.W.; Writing—original draft preparation, Q.W., Y.C. and Y.X.; Writing-review and editing, Y.C., Y.S. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by National Natural Science Foundation of China under Grant (51874299, 61702032 and 61573057), in part by Pre-research Fund in the 13th-five Year Plan under Grant (61404130316).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets involved in this paper are all public datasets. CrowdHuman: https://www.crowdhuman.org/download.html (accessed on 5 March 2021); CityPersons: https://www.cityscapes-dataset.com/ (accessed on 5 March 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shen, Y.; Zhang, L.; Wang, Z.L.; Hao, X.L.; Hou, Y.L. Multi-Level Residual Up-Projection Activation Network for Image SuperResolution. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2841–2845. [Google Scholar] [CrossRef]
  2. Rasouli, A.; Tsotsos, J.K. Autonomous vehicles that interact with pedestrians: A survey of theory and practice. IEEE Trans. Intell. Transp. Syst. 2020, 21, 900–918. [Google Scholar] [CrossRef] [Green Version]
  3. Shao, S.; Zhao, Z.J.; Li, B.X.; Xiao, T.T.; Yu, G.; Zhang, X.Y.; Sun, J. Crowdhuman: A benchmark for detecting humans in a crowd. arXiv 2018, arXiv:1805.00123. preprint. [Google Scholar]
  4. Vimal, S.P.; Ajay, B.; Thiruvikiraman, P.K. Context pruned histogram of oriented gradients for pedestrian detection. In Proceedings of the 2013 International Mutli-Conference on Automation, Computing, Communication, Control and Compressed Sensing (iMac4s), Kottayam, India, 22–23 March 2013; pp. 718–722. [Google Scholar] [CrossRef]
  5. Zhuang, J. Compressive tracking based on HOG and extended Haar-like feature. In Proceedings of the 2016 2nd IEEE International Conference on Computer and Communications (ICCC), Chengdu, China, 14–17 October 2016; pp. 326–331. [Google Scholar] [CrossRef]
  6. Cosma, C.; Brehar, R.; Nedevschi, S. Pedestrians detection using a cascade of LBP and HOG classifiers. In Proceedings of the 2013 IEEE 9th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 5–7 September 2013; pp. 69–75. [Google Scholar] [CrossRef]
  7. Dominguez-Sanchez, A.; Cazorla, M.; OrtsEscolano, S. Pedestrian Movement Direction Recognition Using Convolutional Neural Networks. IEEE Trans. Intell. Transp. Syst. 2017, 18, 3540–3548. [Google Scholar] [CrossRef]
  8. Zhao, J.X.; Li, J.; Ma, Y.D. RPN+ fast boosted tree: Combining deep neural network with traditional classifier for pedestrian detection. In Proceedings of the 2018 4th International Conference on Computer and Technology Applications (ICCTA), Istanbul, Turkey, 3–5 May 2018; pp. 141–150. [Google Scholar] [CrossRef]
  9. Zhang, Z.S.; Gao, J.Y.; Mao, J.H.; Liu, Y.K. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11343–11352. [Google Scholar] [CrossRef]
  10. Zhang, S.F.; Wen, L.Y.; Bian, X.; Lei, Z.; Li, S.Z. Occlusion-Aware R-CNN: Detecting Pedestrians in a Crowd. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 657–674. [Google Scholar]
  11. Wang, X.L.; Xiao, T.T.; Jiang, Y.N.; Shao, S.; Sun, J.; Shen, C.H. Repulsion Loss: Detecting Pedestrians in a Crowd. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7774–7783. [Google Scholar] [CrossRef] [Green Version]
  12. Shao, X.T.; Wei, C.K.; Shen, Y.; Wang, Z.L. Feature Enhancement Based on CycleGAN for Nighttime Vehicle Detection. IEEE Access. 2021, 9, 849–859. [Google Scholar] [CrossRef]
  13. Ke, W.; Zhang, T.L.; Huang, Z.Y.; Ye, Q.X.; Liu, J.Z.; Huang, D. Multiple Anchor Learning for Visual Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10203–10212. [Google Scholar] [CrossRef]
  14. Chen, Y.H.; Cao, Y.; Hu, H.; Wang, L.W. Memory Enhanced Global-Local Aggregation for Video Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10334–10343. [Google Scholar] [CrossRef]
  15. Li, Y.; Wang, T.; Kang, B.Y.; Tang, S.; Wang, C.F.; Li, J.T.; Feng, J.S. Overcoming Classifier Imbalance for Long-Tail Object Detection with Balanced Group Softmax. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10988–10997. [Google Scholar] [CrossRef]
  16. Shen, Y.; Luo, M.; Chen, Y.; Shao, X.T.; Wang, Z.L.; Hao, X.L.; Hou, Y.L. Cross-View Image Translation Based on Local and Global Information Guidance. IEEE Access 2021, 9, 12955–12967. [Google Scholar] [CrossRef]
  17. Tan, M.X.; Pang, R.M.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
  18. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
  19. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  20. Bochkovskiy, A.; Wan, C.Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. preprint. [Google Scholar]
  21. Zhu, C.C.; He, Y.H.; Savvides, M. Feature Selective Anchor-Free Module for Single-Shot Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 840–849. [Google Scholar] [CrossRef] [Green Version]
  22. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.M.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans Pattern Anal. Mach. Intell. 2017, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Tan, M.X.; Le, Q.V. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 10691–10700. [Google Scholar]
  24. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  25. Ren, S.Q.; He, K.W.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. He, K.M.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  27. Cai, Z.W.; Vasconcelos, N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. IEEE Trans Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS-Improving Object Detection with One Line of Code. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5562–5570. [Google Scholar] [CrossRef] [Green Version]
  29. He, Y.H.; Zhu, C.C.; Wang, J.R.; Savvides, M.; Zhang, X.Y. Bounding Box Regression with Uncertainty for Accurate Object Detection. In Proceedings of the 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 2883–2892. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, S.T.; Huang, D.; Wang, Y.H. Adaptive NMS: Refining Pedestrian Detection in a Crowd. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 6452–6461. [Google Scholar] [CrossRef] [Green Version]
  31. Zhou, C.L.; Yuan, J.S. Bi-box Regression for Pedestrian Detection and Occlusion Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 138–154. [Google Scholar]
  32. Chi, C.; Zhang, S.F.; Xing, J.L.; Lei, Z.; Li, S.Z.; Zou, X.D. PedHunter: Occlusion Robust Pedestrian Detector in Crowded Scenes. In Proceedings of the 2020 AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 10639–10646. [Google Scholar] [CrossRef]
  33. Pang, Y.W.; Xie, J.; Khan, M.H.; Anwer, R.M.; Khan, F.S.; Shao, L. Mask-Guided Attention Network for Occluded Pedestrian Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 4966–4974. [Google Scholar] [CrossRef] [Green Version]
  34. Wang, A.T.; Sun, Y.H.; Kortylewski, A.; Yuille, A. Robust Object Detection under Occlusion with Context-Aware CompositionalNets. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12642–12651. [Google Scholar] [CrossRef]
  35. Wu, J.L.; Zhou, C.L.; Yang, M.; Zhang, Q.; Li, Y.; Yuan, J.S. Temporal-Context Enhanced Detection of Heavily Occluded Pedestrians. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13427–13436. [Google Scholar] [CrossRef]
  36. Cao, J.L.; Pang, Y.W.; Han, J.G.; Gao, B.L.; Li, X.L. Taking a Look at Small-Scale Pedestrians and Occluded Pedestrians. IEEE Trans. Image Process. 2020, 29, 3143–3152. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.M.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef] [Green Version]
  38. Liu, S.; Qi, L.; Qin, H.F.; Shi, J.P.; Jia, J.Y. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef] [Green Version]
  39. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  40. Sifre, L. Rigid-Motion Scattering for Image Classification. Ph.D. Thesis, Ecole Polytechnique, Palaiseau, France, 2014; pp. 111–114. [Google Scholar]
  41. Chu, X.G.; Zheng, A.L.; Zhang, X.Y.; Sun, J. Detection in Crowded Scenes: One Proposal, Multiple Predictions. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12211–12220. [Google Scholar] [CrossRef]
Figure 1. Overall network architecture of MFPN. Feature maps fout_1–fout_5 are extracted by ResNet50. Two modules with a yellow highlighted background are our DFR and RLM modules. DFR is the feature fusion module to fuse feature maps with different resolutions. We utilize the weighted feature fusion to sum up feature maps distinguishingly, where “⊕” represents the fusion process. The predictions box1 and box2 are two predictional boxes from the same proposal. Our new loss function RLM can keep predicted boxes away from other ground truths.
Figure 1. Overall network architecture of MFPN. Feature maps fout_1–fout_5 are extracted by ResNet50. Two modules with a yellow highlighted background are our DFR and RLM modules. DFR is the feature fusion module to fuse feature maps with different resolutions. We utilize the weighted feature fusion to sum up feature maps distinguishingly, where “⊕” represents the fusion process. The predictions box1 and box2 are two predictional boxes from the same proposal. Our new loss function RLM can keep predicted boxes away from other ground truths.
Sensors 21 01820 g001
Figure 2. BiFPN and our DFR network. (a) BiFPN. The structure of BiFPN is the optimization of FPN and PANet. It utilizes cross-scale connection and weighted feature fusion to extract and fuse features of different resolutions base on the EfficientNet backbone [23]. (b) DFR network. We stack two BiFPN layers of bottom-up and top-down structures and then combine the structure with ResNet50, which is widely used in the object detection process to achieve a sound performance of feature extraction and fusion. The blue circles on the left are feature maps extracted by ResNet50, and the remaining on the right is the Double FPN structure we proposed. “⊕” represents the fusion process. The sign of double lines with an arrow “⟹” represents the operation of upsampling which can resize the feature map twice as large as the original. The dashed arrow means the operation of max pooling, which can resize the feature map half of the original.
Figure 2. BiFPN and our DFR network. (a) BiFPN. The structure of BiFPN is the optimization of FPN and PANet. It utilizes cross-scale connection and weighted feature fusion to extract and fuse features of different resolutions base on the EfficientNet backbone [23]. (b) DFR network. We stack two BiFPN layers of bottom-up and top-down structures and then combine the structure with ResNet50, which is widely used in the object detection process to achieve a sound performance of feature extraction and fusion. The blue circles on the left are feature maps extracted by ResNet50, and the remaining on the right is the Double FPN structure we proposed. “⊕” represents the fusion process. The sign of double lines with an arrow “⟹” represents the operation of upsampling which can resize the feature map twice as large as the original. The dashed arrow means the operation of max pooling, which can resize the feature map half of the original.
Sensors 21 01820 g002
Figure 3. Repulsion loss of minimum. (a) Solid boxes(red, green and orange) except for the blue proposal represent ground truths. Dashed boxes (purple and yellow) indicate two predictions by the same blue proposal. (b) Dashed (purple and yellow) boxes are predicted boxes by the blue proposal B. Solid boxes, red, green and orange (gt1, gt2, gt3) are the ground truths that have the maximum IoU with the blue proposal. Calculate the loss between the orange box and yellow and purple boxes, resulting in LossRep.
Figure 3. Repulsion loss of minimum. (a) Solid boxes(red, green and orange) except for the blue proposal represent ground truths. Dashed boxes (purple and yellow) indicate two predictions by the same blue proposal. (b) Dashed (purple and yellow) boxes are predicted boxes by the blue proposal B. Solid boxes, red, green and orange (gt1, gt2, gt3) are the ground truths that have the maximum IoU with the blue proposal. Calculate the loss between the orange box and yellow and purple boxes, resulting in LossRep.
Sensors 21 01820 g003
Figure 4. Comparisons of feature maps extracted by different networks. As shown in (b), the feature map extracted by ResNet50 and FPN is the worst, with unclear key features and contours. After the combination of ResNet50 and BiFPN (c,d), the features in the front row are better than (b): contours of features in the front row are gradually clearer, and features of key parts can be displayed clearly. (e,f) are feature maps extracted by our network. The only difference is that (f) uses general convolution. It can be seen that the contours and key parts of the features can be clearly represented, especially the features of the targets in the front row. The contours of features extracted by our network in the back row are also gradually separated and clear. Furthermore, we can find that the effect of general convolution on feature extraction network is better than that of separable convolution, comparing (d) with (c), (f) with (e). Under the same conditions, separable convolution has fewer parameters than general convolution. Therefore, it is usually used in some lightweight structures. Due to the purpose of improving the performance of the pedestrian detection system, the general convolution is utilized with the factor of parameters ignored. (a) Input Image, (b) ResNet50 + FPN(GC), (c) ResNet50 + BiFPN(SC), (d) ResNet50 + BiFPN(GC), (e) ResNet50 + Double FPN(SC), (f) ResNet50 + Double FPN(GC).
Figure 4. Comparisons of feature maps extracted by different networks. As shown in (b), the feature map extracted by ResNet50 and FPN is the worst, with unclear key features and contours. After the combination of ResNet50 and BiFPN (c,d), the features in the front row are better than (b): contours of features in the front row are gradually clearer, and features of key parts can be displayed clearly. (e,f) are feature maps extracted by our network. The only difference is that (f) uses general convolution. It can be seen that the contours and key parts of the features can be clearly represented, especially the features of the targets in the front row. The contours of features extracted by our network in the back row are also gradually separated and clear. Furthermore, we can find that the effect of general convolution on feature extraction network is better than that of separable convolution, comparing (d) with (c), (f) with (e). Under the same conditions, separable convolution has fewer parameters than general convolution. Therefore, it is usually used in some lightweight structures. Due to the purpose of improving the performance of the pedestrian detection system, the general convolution is utilized with the factor of parameters ignored. (a) Input Image, (b) ResNet50 + FPN(GC), (c) ResNet50 + BiFPN(SC), (d) ResNet50 + BiFPN(GC), (e) ResNet50 + Double FPN(SC), (f) ResNet50 + Double FPN(GC).
Sensors 21 01820 g004
Figure 5. Detection results with our method. Although there are heavy occlusions in these images, the detector with our method can precisely detect the pedestrians.
Figure 5. Detection results with our method. Although there are heavy occlusions in these images, the detector with our method can precisely detect the pedestrians.
Sensors 21 01820 g005
Figure 6. Comparisons of detection results between the ResNet50 and FPN baseline and our method. The images on the first row are the detection results by our method. The last row is the results by the baseline. The solid red boxes are the pedestrians detected by our method but not detected by the baseline method. In contrast, the dashed red boxes are the missed detections.
Figure 6. Comparisons of detection results between the ResNet50 and FPN baseline and our method. The images on the first row are the detection results by our method. The last row is the results by the baseline. The solid red boxes are the pedestrians detected by our method but not detected by the baseline method. In contrast, the dashed red boxes are the missed detections.
Sensors 21 01820 g006
Table 1. Ablation study of feature maps. There are five different combinations.
Table 1. Ablation study of feature maps. There are five different combinations.
HeatmapsResNet50FPNBiFPNDouble FPNSeparable Convolution(SC)General Convolution(GC)
(a)
(b)
(c)
(d)
(e)
Table 2. Ablation study implemented on the validation set of CrowdHuman.
Table 2. Ablation study implemented on the validation set of CrowdHuman.
MethodAP/%MR−2/%JI/%
ResNet50 + FPN baseline (impl. by [41]) 85.842.979.8
ResNet50 + FPN (impl. by [41])90.342.282.0
ResNet50 + BiFPN (Separable)89.8142.8181.75
ResNet50 + BiFPN90.6440.8282.83
ResNet50 + DoubleFPN (Separable w/o RLM)90.7641.0582.83
ResNet50 + DoubleFPN (w/o RLM)90.8840.5282.92
ResNet50 + DoubleFPN (with RLM)90.9640.2483.12
Table 3. Comparisons with previous methods on the CrowdHuman and CityPersons datasets.
Table 3. Comparisons with previous methods on the CrowdHuman and CityPersons datasets.
DatasetMethodAP/%MR−2/%JI/%
CrowdHumanFPN baseline (impl. by [41])85.842.979.8
ResNet50 + FPN (impl. by [41])90.342.282.0
FPN + Soft-NMS [28]88.242.979.8
Adaptive NMS [30]84.749.7
Cascade R-CNN [27] (impl. by [41])85.643.080.6
Ours90.9640.2483.12
CityPersonsFPN baseline [41]95.211.7
FPN + Soft-NMS [28]95.311.8
ResNet50 + FPN [41]96.110.7
Ours96.2310.64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shao, X.; Wang, Q.; Yang, W.; Chen, Y.; Xie, Y.; Shen, Y.; Wang, Z. Multi-Scale Feature Pyramid Network: A Heavily Occluded Pedestrian Detection Network Based on ResNet. Sensors 2021, 21, 1820. https://doi.org/10.3390/s21051820

AMA Style

Shao X, Wang Q, Yang W, Chen Y, Xie Y, Shen Y, Wang Z. Multi-Scale Feature Pyramid Network: A Heavily Occluded Pedestrian Detection Network Based on ResNet. Sensors. 2021; 21(5):1820. https://doi.org/10.3390/s21051820

Chicago/Turabian Style

Shao, Xiaotao, Qing Wang, Wei Yang, Yun Chen, Yi Xie, Yan Shen, and Zhongli Wang. 2021. "Multi-Scale Feature Pyramid Network: A Heavily Occluded Pedestrian Detection Network Based on ResNet" Sensors 21, no. 5: 1820. https://doi.org/10.3390/s21051820

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop