Next Article in Journal
Virtual Reality with 360-Video Storytelling in Cultural Heritage: Study of Presence, Engagement, and Immersion
Previous Article in Journal
Rectification and Super-Resolution Enhancements for Forensic Text Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Seeing Pedestrian in the Dark via Multi-Task Feature Fusing-Sharing Learning for Imaging Sensors

1
Hubei Key Laboratory of Intelligent Robot, School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430073, China
2
School of Electronics and Information Engineering, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(20), 5852; https://doi.org/10.3390/s20205852
Submission received: 10 September 2020 / Revised: 1 October 2020 / Accepted: 13 October 2020 / Published: 16 October 2020
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Pedestrian detection is an essential problem of computer vision, which has achieved tremendous success under controllable conditions using visible light imaging sensors in recent years. However, most of them do not consider low-light environments which are very common in real-world applications. In this paper, we propose a novel pedestrian detection algorithm using multi-task learning to address this challenge in low-light environments. Specifically, the proposed multi-task learning method is different from the most commonly used multi-task learning method—the parameter sharing mechanism—in deep learning. We design a novel multi-task learning method with feature-level fusion and a sharing mechanism. The proposed approach contains three parts: an image relighting subnetwork, a pedestrian detection subnetwork, and a feature-level multi-task fusion learning module. The image relighting subnetwork adjusts the low-light image quality for detection, the pedestrian detection subnetwork learns enhanced features for prediction, and the feature-level multi-task fusion learning module fuses and shares features among component networks for boosting image relighting and detection performance simultaneously. Experimental results show that the proposed approach consistently and significantly improves the performance of pedestrian detection on low-light images obtained by visible light imaging sensor.

1. Introduction

Pedestrian detection is a vital problem in computer vision with significant impact on a number of applications, such as advanced driver assistance systems, robot navigation, and intelligent video surveillance systems. These applications use a large number of imaging sensors. The continuous development of deep learning has made a huge contribution to the performance of pedestrian detection. Generally speaking, pedestrian detection is a specific domain of object detection [1,2,3,4,5,6,7,8,9]. From this perspective, pedestrian detection algorithms in deep learning can be divided into two categories: anchor box-based and keypoint-based methods.
In pedestrian detection algorithms that involve designing handcrafted features, researchers usually need to design the complicated features to ensure the detection performance. Pierre et al. [10] proposed unsupervised multi-stage feature learning and designed low-level features to improve the performance of pedestrian detection. Shen et al. [11] proposed a novel Taylor feature transform (TAFT) feature for pedestrian detection, which is derived by treating an image by a 2D function and applying its Taylor series for approximation. Paolanti et al. [12] used multiple nearest neighbor classifiers and neighborhood component features selection to achieve pedestrian re-identification. Although the above methods have achieved good results, the heavy feature design brings a lot of resource consumption. Girshick et al. [1] first introduced Convolutional Neural Networks (CNNs) into object detection, named “R-CNN”, which enabled object detection without designing handcrafted features. Then, Girshick [2] optimized the training process for R-CNN and proposed a single-stage training algorithm that speeds up the training and improves the performance of the algorithm. Ren et al. [3] first proposed the anchor boxes-based detection algorithm and a Region Proposal Network (RPN) that shares fully image convolutional features with the detection network, which can achieve almost costless region proposals. Since then, anchor boxes have been widely used in SSD [4], YOLOv2 [5], YOLOv3 [6], and other excellent object detection algorithms. Liu et al. [13] presented an asymptotic localization fitting module to refine anchor boxes step by step into final detection results. Although the above anchor boxes-based methods achieved acceptable performance, they depended on a very large set of anchor boxes and heavy training burden.
Keypoint-based object detection algorithms generate object bounding boxes by detecting and grouping their key points. They greatly simplify the output of the network and eliminate the need for designing anchor boxes. Zhou et al. [14] modeled an object as a center point of bounding box, and this detector used keypoint estimation to find center points and regressed to all other object properties, such as size, 3D location and orientation. Liu et al. [15] proposed a Center and Scale Prediction (CSP) based detector as new perspective of treating pedestrian detection and predicting the scale of the central points. Law et al. [16] presented a novel approach, named “CornerNet”, to detect an object bounding box as a pair of keypoints, the top-left corner and the bottom-right corner. Law et al. [17] subsequently improved CornerNet and named as “CornerNet-Saccade”, which adopted the attention mechanism to further improve the performance of object detection and reduce parameters.
Although the above-mentioned pedestrian detection algorithms yield satisfactory performance under normal lighting conditions. Most of them do not consider pedestrian detection in low-light environments. As we know, in real-world applications, normal lighting condition is not always guaranteed. On the contrary, the low-light environment is very common. To address the low-light pedestrian detection problem, Kruthiventi et al. [18] proposed a deep convolutional network under low illumination conditions. However, it needs to be trained by thermal and RGB images, and the thermal images need to be obtained by infrared thermal imaging sensors, which incurs high cost. The main reason for the poor pedestrian detection performance in low-light environment is that low lighting causes serious distortion of color and texture information in the obtained inputs. However, color and texture information plays a vital role in pedestrian detection.
For both pedestrian detection and image relighting tasks, there is a basic discovery that these two tasks can share with some features during learning due to learning framework from CNNs. Inspired by multi-task learning [19], we propose a novel multi-task learning method with feature-level fusion and sharing mechanism, which will be described in Section 3. The proposed Seeing Pedestrian in the Dark via Multi-task Feature Fusing-sharing Learning, dubbed as “SPMF” in this paper, has three components: an image relighting subnetwork, a pedestrian detection subnetwork, and a feature-level multi-task fusion learning module. The first two subnetworks implement light enhancement and object detection, respectively. By cascading above two subnetworks, the feature-level multi-task fusion learning module shares fused features among different subnetworks for image relighting and detection. Experimental results show that SPMF consistently and significantly improves the accuracy of pedestrian detection on low-light images.
The contributions of this paper are summarized as follows.
(i)
Different from most commonly used multi-task learning methods with a parameter sharing mechanism for deep learning, we propose a novel multi-task learning method with feature-level fusion and sharing mechanism in the serial tasks. The proposed multi-task learning method is used in the feature-level multi-task fusion learning module to fuse the feature information of upstream and downstream tasks, and then the fused features are learned for each tasks to boost upstream and downstream tasks performance simultaneously.
(ii)
A novel Self-Calibrated Split Attention Block is proposed, named “SCSAB”, which combines Self-Calibrated convolution layer and Split Attention mechanism. SCSAB further improves the ability of SPMF to detect pedestrians in the dark.

2. Related Work

In this section, we first describe the present situation of multi-task learning and then focus on introducing two most common methods for multi-task learning in Deep Learning. Next, the self-calibrated convolution and Split Attention Networks are introduced in detail.

2.1. Multi-Task Learning in Deep Learning

Multi-task learning has been successful in the field of machine learning, from image processing [2] to language processing [20] and medical drug discovery [21]. There are many forms of multi-task learning: joint learning, learning to learn, and assisted task learning are only some names that have been used to refer to it. Caruana [22] summarized the merit of multi-task learning: “Multi-task learning improves generalization by leveraging the domain-specific information contained in the training signals of related tasks”.
In view of the great success of multi-task learning in machine learning, Ruder [19] summarized two methods of multi-task learning for Deep Learning: the hard parameter sharing method and soft parameter sharing method. Panels (A,B) in Figure 1 show the structure of these two methods. Hard parameter sharing was originally proposed by Caruana [23], but its use is still very common in recent years. Long et al. [24] proposed a Deep Relationship Networks, which was a typical representative of Hard parameter sharing, to learn multiple tasks. Although the hard parameter sharing method greatly reduces the risk of overfitting for each task, it collapse quickly if there is no close relationship between tasks or reasoning at different levels. In the soft parameter sharing method, each task has its own model and parameters. Then, the distance between model parameters is regularized to encourage the parameters to be similar. Misra et al. [25] proposed a Cross-stitch Network for multi-task learning, which used the soft parameter sharing mechanism. They used cross-stitch units allowing the model to determine how task-specific networks would use knowledge from other tasks by learning a parameterized linear combination of the output of the previous layers. Gao et al. [26] proposed a novel convolutional neural network structure for multi-task learning, named “NDDR-CNN”, which enables automatic feature fusing at every layer from different tasks. Then, the authors of [27] extended the works in [25,26], which proposed the neural architecture search (NAS) into multi-task learning. This gets rid of the way of feature fusion between different tasks of hand-crafted designing.
Although the above two multi-task learning methods are widely used in machine learning, these learning mechanisms can only be used for tasks with high relevance and parallel tasks. If these multi-task learning mechanisms are used between the upstream and downstream tasks, it may lead to poor performance, and model training may be difficult to converge. In this study, we propose a novel multi-task learning method with feature-level fusion and sharing mechanism to fuse the feature information of upstream and downstream tasks. In this way, the upstream and downstream tasks can boost performance simultaneously. Figure 1C shows the structure of this method.

2.2. Self-Calibrated Convolutions and Split Attention Network

In recent years, novel architecture designing has made world-renowned progress. As an early work, ResNet [28] introduces the residual connections and using batch normalization [29] to greatly improve the sequential structure, making it possible to build very deep neural networks and to enhance the performance of versatile tasks in computer vision [30,31,32,33]. ResNeXt [34] uses group convolutional layers or increases their width to extend ResNet. GoogLeNet [35] has been successful in multipath representation, in which each network block is composed of different convolutional kernels. NASNet [36] learns to automatically build model architecture by exploring a predefined search space. DenseNet [37] aggregates the features of different convolution layers through complex bottom-up skip connections. SENet [38] introduces the squeezing-and-excitation operation and channel-attention mechanism to explicitly model the inner relationship in channel-level.
The above methods focus too much on designing or adjusting the network architectures to generate rich and better feature representations, which requires a lot of manpower, material resources, and time. To solve this problem, Liu et al. [39] presents a novel Self-Calibrated Convolution that explicitly expands fields-of-view of each convolutional layer through internal Interactive and hence enriches the output features. Self-Calibrated Convolutions improve the basic convolutional feature representation process of CNNs without adjusting and designing the model architectures. Benefiting from the great contribution of ResNet and its variants in deep learning, Zhang et al. [40] proposed ResNeSt, which combines the attention mechanism of SENet [38], SKNet [41], and ResNeXt [34] into group-level. ResNeSt introduced the Split Attention Network, which enables feature-map attention across different feature-map groups. In this study, we propose a Self-Calibrated Split Attention Block, dubbed as “SCSAB” in this paper, which combines Self-Calibrated Convolutions and Split Attention Network. The details of SCSAB are explained in Section 3.

3. Proposed Methods

In this section, we present the architecture of the proposed network, including the main framework and details for individual modules.

3.1. Network Architecture

In order to boost the pedestrian detection performance in low-light environment, we use a novel multi-task learning mechanism to simultaneously learn two-task shared features for rendering rich texture images and for obtaining better pedestrian detection performance in low-light environment. The proposed approach consists of three parts: an image relighting subnetwork, a pedestrian detection subnetwork, and a feature-level multi-task fusion learning module. Figure 2 shows the architecture of SPMF, including two subnetworks and the detailed structure of feature-level multi-task fusion learning module.

3.2. Self-Calibrated Split Attention Block (SCSAB)

Previous CNN-based object/pedestrian detection methods [14,15,16,17] use large-scale residual blocks, which improve the performance of the detection algorithm. However, in the convolution feature transformation, the residual blocks cannot obtain a larger fields-of-view for each spatial position. To efficiently fuse informative contextual information for each spatial location and expand receptive field of each convolutional layer, we propose the Self-Calibrated Split Attention Block (SCSAB), which combines Self-Calibrated Convolutions [39] and Split Attention Network [40]. The detailed structure of SCSAB is shown in Figure 3. SCSAB can perform convolutional feature transformation in two different feature spaces: an original size feature space with the same resolution of the feature map and the input, and a small size feature space after downsampling. Where the SCConv is the Self-Calibrated Convolution layer, and the detailed structure of Split Attention is proposed in Section 3 in [40].

3.3. Multi-Task Learning with Feature-Level Fusion and Sharing

In this subsection, we introduce a novel multi-task learning method with a feature-level fusion and sharing mechanism and explain in detail how it works.
Different from most commonly used multi-task learning methods of parameter sharing mechanism in deep learning, the proposed multi-task learning method with feature-level fusion and sharing mechanism is used to fuse the feature information of upstream and downstream tasks, and then the fused features are learned for upstream and downstream tasks simultaneously.
The workflow of the proposed mechanism is shown in Figure 4. Suppose there are two tasks, A and B, task A is an upstream task and task B is a downstream task. The output feature of the convolutional layer C A 1 in the task A is O C A 1 , and the output feature of the convolutional layer C B 1 in the task B is O C B 1 , and C A 2 and C B 2 are the next convolutional layers of C A 1 and C B 1 , respectively. I C A 2 is the input of C A 2 convolutional layer, I C B 2 is the input of C B 2 convolutional layer, and F i is the fused feature obtained from the i t h end-to-end alternately iteration. The F i is formulated as
F i = ( O C A 1 + O C B 1 ) / 2 .
When i = 1 , there is no fused feature between the task A and task B. Here, the I C A 2 and I C B 2 come from the output of C A 1 and C B 1 . I C A 2 and I C B 2 are formulated as
I C A 2 = O C A 1 I C B 2 = O C B 1 .
When i > 1 , the i t h iteration for I C A 2 and I C B 2 are formulated as
I C A 2 = O C A 1 × s i g m o i d F i 1 I C B 2 = O C B 1 × s i g m o i d F i 1 ,
where F i 1 is produced in ( i 1 ) t h iteration, then is mapped by S i g m o i d function, and finally is multiplied by O C A 1 and O C B 1 , respectively, as the input of C A 2 and C B 2 .

3.4. Image Relighting Subnetwork

In the image relighting subnetwork, the proposed method directly uses RetinexNet [42] as the image relighting subnetwork. The classic Retinex theory models the human color perception which assumes that the observed images can be decomposed into two components: reflectance and illumination channels. Let S represent a source image, then it can be denoted by S = R I , where R represents reflectance component, I represents illumination component, and * represents element-wise multiplication. Here, the loss function of RetinexNet is written as L e n h . The loss function L e n h is formulated as
L e n h = L r e c o n + λ i r L i r + λ i s L i s ,
where λ i r and λ i s denote the coefficients to balance the reflectance and the illumination. Loss functions L r e c o n , L i r and L i s represent reconstruction, invariable reflectance, and illumination smoothness functions, respectively.

3.5. Pedestrian Detection Subnetwork

In the pedestrian detection subnetwork, we notice that the CornerNet-Saccade has made remarkable achievements in the key-point based object detection algorithm, which uses the Hourglass-54 proposed by [17] as the backbone. Therefore, we design a new Hourglass backbone network, named “SCSAB-Hourglass”, which uses the proposed SCSAB as basic block.
In addition, we propose a lightweight backbone network by pruning SCSAB-Hourglass and named it “SCSAB-HourglassLite”. We apply the focal loss with α = 2 and β = 4 . Here, let p a i j be the score at the position i , j for pedestrian in the input image, and let y a i j be the ground-truth. Then, we have
L d e t = 1 N c = 1 C i = 1 H j = 1 W 1 P α log P i f Y = 1 1 Y β P α log 1 P o t h e r w i s e ,
where N is the number of pedestrians in an image. C, H, and W represent the number of channels, height, and width, respectively, from input. α and β are the hyperparameters to control the contribution of each point. P and Y represent p a i j and y a i j , respectively.
When the input image passes through the convolutional layer, the size of the output is usually smaller than the input image. Therefore, a pedestrian position x , y in the image is mapped to the position x n , y n in the heatmaps, where n is the downsampling factor. When we remap the positions from the heatmaps to the original size input image, some accuracy may be lost. To solve this issue, we predict the position offsets to adjust the corner positions marginally before remapping them to the original input image.
o z = x z n x z n , y z n y z n ,
where o z is the value of offset, and x z and y z are the x and y coordinates for corner z. For training, we apply the smooth L 1 loss as the offset loss function, and mark this loss function as L o f f . The details of this function are elaborated in Section 2 of [2]. There exists multiple pedestrians in an image, and thus multiple top-left and bottom-right corners can be detected. We need to determine if a pair of the top-left corner and bottom-right corner is from the same pedestrian. Let e t m be the top-left corner of pedestrian m and e b m for the bottom-right corner. As in [43], we use the “pull” and “push” loss to group the corners and separate the corners:
L p u l l = 1 N m = 1 N e t m e m 2 + e b m e m 2 ,
L p u s h = 1 N N 1 m = 1 N j = 1 j m N max 0 , 1 e m e j ,
where e m is the average of e t m and e b m .

3.6. Feature-Level Multi-Task Fusion Learning Module

In this subsection, we introduce how the feature-level multi-task fusion learning module to fuse and share the features of image relighting subnetwork and pedestrian detection subnetwork. The proposed module is shown in Figure 2.
In the feature-level multi-task fusion learning module, we fuse the features a 1 i and a 2 i in image relighting subnetwork to the features b 1 i and b 2 i in pedestrian detection subnetwork respectively in the i t h iteration, and the two fused features are marked as F 1 i and F 2 i . The feature a 1 i and feature a 2 i are from the first 3 × 3 convolutional layer and the third last 3 × 3 convolutional layer in Enhance-Net, respectively. The feature b 1 i and feature b 2 i are from the second last SCSAB and the first SCSAB in SCSAB-Hourglass, respectively. Meanwhile, the sizes of these features are the same. The F 1 i and F 2 i are formulated as
F 1 i = ( a 1 i + b 1 i ) / 2 , F 2 i = ( a 2 i + b 2 i ) / 2 .
Then, F 1 i and F 2 i with a s i g m o i d activation function are marked as s i g m o i d ( F 1 i ) and s i g m o i d ( F 2 i ) . In the ( i + 1 ) t h iteration, s i g m o i d ( F 1 i ) is element-wise product by a 1 i + 1 and b 1 i + 1 , respectively, as the input of the second 3 × 3 convolution layer of Enhance-Net and the last SCSAB of the SCSAB-Hourglass, the inputted features are marked as I a 1 i + 1 and I b 1 i + 1 , which are formulated as
I a 1 i + 1 = [ a 1 i + 1 × s i g m o i d ( F 1 i ) ] ,
I b 1 i + 1 = [ b 1 i + 1 × s i g m o i d ( F 1 i ) ] .
The same fusion and sharing method is used as the input of the second last 3 × 3 convolution layer in Enhance-Net and the second SCSAB in SCSAB-Hourglass. The inputted features are marked as I a 2 i + 1 and I b 2 i + 1 , which are formulated as
I a 2 i + 1 = [ a 2 i + 1 × s i g m o i d ( F 2 i ) ] ,
I b 2 i + 1 = [ b 2 i + 1 × s i g m o i d ( F 2 i ) ] .
Finally, the detection network training loss function is formulated as
L = L d e t + δ L p u l l + η L p u s h + γ L o f f + ζ L e n h ,
where δ , η , and γ are the weights for the pull, push, and offset loss, respectively, and ζ is the weight for the light enhancement loss. We set δ and η to 0.1, γ to 1, and ζ to 0.05.

4. Experiments and Discussion

4.1. Dataset and Implementation Details

The performance of SPMF is evaluated on the CityPersons dataset [44]. The CityPersons is a publicly available large-scale pedestrian detection dataset that contains 2975 images and approximately 20,000 annotated pedestrian instances in training subset. We use the CityPersons validation dataset as our testing subset, which contains 500 images. Because SPMF requires paired low/normal illumination images for training, we need to render low-light images from normal illumination images in the existing CityPersons dataset. At the same time, same operation of rendering low-light images is used to generate testing subset. Our experiments use the RGB spatial brightness adjustment algorithm to generate low-light images. For example, if the current pixel is (50, 100, 200) and the coefficient of adjustment is 1.1, the adjusted pixel is (55, 110, 220). This paper uses low-light images with an adjustment coefficient of 0.8.
The standard Caltech evaluation: log average Miss Rate over False Positive Per Image (FPPI) range of 10 2 , 10 0 (denoted as M R 2 ) [45] is used as the evaluation metric for pedestrian detection algorithm. The lower value of M R 2 , the better performance of the algorithm. We use IoU (Intersection over Union) as the threshold for the evaluation metric, with values of 0.5, 0.75, and 0.5:0.95 (0.5:0.95 means average M R 2 over different IoU thresholds, from 0.5 to 0.95, step 0.05), respectively, to evaluate the performance of different algorithms.
SPMF is implemented in the Pytorch, with 3 RTX 2080Ti GPUs for training. A mini-batch contains 8 images per GPU. The Adam solver is applied. For the image relighting and pedestrian detection subnetworks, we use the end-to-end alternately training method. First of all, we train two subnetworks, respectively. Then the two trained models are used as pretrained models to train 100 epochs by feature-level multi-task fusion learning module.

4.2. Ablation Experiments

In this subsection, we conduct the ablation studies on the low-light CityPersons testing dataset to demonstrate the effectiveness of the proposed backbones.
Comparing with CornerNet-Saccade, which uses Hourglass-54 as the backbone network, we design a new hourglass backbone network in SPMF. The novel hourglass network uses proposed SCSAB as the basic block, which is named “SCSAB-Hourglass”. In addition, we propose a more lightweight backbone than “SCSAB-Hourglass” and name it “SCSAB-HourglassLite”, which is reduced the parameters of model greatly by pruning SCSAB-Hourglass. We name the method using SCSAB-Hourglass backbone as the SPMF, and the method using SCSAB-HourglassLite backbone as SPMF-Lite.
Table 1 shows the parameters of different backbone networks, we can find that comparing with the Hourglass-54, the number of parameters of SCSAB-Hourglass was reduced by about 14%. Meanwhile, the parameters of the proposed SCSAB-HourglassLite have been greatly reduced by about 50% comparing with SCSAB-Hourglass. From the experiment results shown in Table 2, the proposed SCSAB indeed brings an obvious gain in performance. Even we use the SCSAB-HourglassLite backbone network, the detection performance does not decrease much, especially when IoU = 0.5, the M R 2 of SCSAB-HourglassLite actually is lower than that of SCSAB-Hourglass. This confirms the role of new backbones of detecting network.

4.3. Comparison with State-of-the-Art

In this subsection, we select some excellent pedestrian detection or object detection methods as: CSP [15], ALFNet [13], CenterNet [14], CornerNet [16], and CornerNet-Saccade [17]. ALFNet is the most representative algorithm using anchor boxes in pedestrian detection, the CSP and CenterNet are the best algorithms using the central point based method in pedestrian detection and object detection, respectively. Meanwhile, CornerNet and CornerNet-Saccade are the representative algorithms for the point of corner-based method in object detection. The work in [18] takes into account the low light environment, but it learns to extract both RGB images feature and thermal images feature. Therefore, it is not fair to compare with this method.
For fair comparison, two different testing modes are used in testing: end-to-end testing mode and cascade testing mode. In end-to-end testing mode, we use same training dataset to train every selected state-of-the-art, from this point, because both low/normal light images are fed into networks, all these pedestrian detection networks has abilities to deal with low-light images.

4.3.1. End-to-End Testing Mode

Table 3 list the results for end-to-end testing mode, it is obviously that SPMF is superior to these state-of-the-art algorithms. In terms of visual performance, Figure 5 shows the detection results of different methods in five representative samples. Row (a) is the input images, and row (i) is the ground-truth as the benchmark for each detection algorithms. In these visual results, rows (b–d) are from the outputs of CSP, ALFNet, and CenterNet, respectively, in five test samples. Row (b) has missed detection in Test Sample (1)–(3) and has false detection in Test Sample (1)–(5). In the results of row (c), Test Sample (2) appears missed detection, Test Sample (4) has some false detection, and Test Samples (1) and (3) not only have missed detection, but also have false detection. Although all pedestrians are detected in row (d) except Test Sample (2), there are lots of false detections in Test Samples (2)–(5). Rows (e,f) are the test results for CornerNet and CornerNet-Saccade respectively. We can find that row (e) has many missed detection and false detection in these test sample, and row (f) has many missed detection. In the above five algorithms, we find that CornerNet achieves a very poor visual detection results. The main reason for this situation may be that the CornerNet needs 10 GPUs (10 Titan X (PASCAL), total of 120 GB) and uses a batch size of 49 to train, but we do not have such huge computing resources. Under the same experimental conditions (3 RTX 2080Ti GPUs, total of 33 GB), we train the CornerNet with a batch size of 9. The visual results of our models are shown in rows (g,h), intuitively, SPMF and SPMF-Lite achieve good detecting results and with high visual performance. In Test Sample (2), two pedestrians are severely obscured, and it is difficult to see the two obscured pedestrians in an intuitive visual experience. However, SPMF and SPMF-Lite can still detect partially severely obscured pedestrians, and other state-of-the-art cannot detect these, which also proves the proposed methods having high performance.

4.3.2. Cascade Testing Mode

In cascade testing mode, we separately train relighting and detecting of two tasks, where RetinexNet is used for relighting task and the other state-of-the-art pedestrian detection algorithms are used for detecting task. Thus, in this setting, it is more fair to compare SPMF with two cascaded tasks. Figure 6 shows the outputs from the RetinexNet for five selected test samples in low-light CityPersons testing set, we can find that comparing with the original images, the images output by RetinexNet has caused some distortion of texture information and loss of color information on the basis of restoring image brightness.
The results of cascade testing mode are shown in Table 4, and we can see that even if we use the images after relighting for testing these state-of-the-art algorithms, SPMF still achieves awesome performance, which proves the effectiveness of proposed multi-task learning method. The visual results of cascade testing mode are shown in Figure 7, intuitively, the detecting performance of these state-of-the-art algorithms in the cascade testing mode are still unsatisfactory. Especially in the test sample (3), these algorithms have achieved very poor detection results, even two algorithms (CornerNet and CornerNet-Saccade) can not detect pedestrians. The reason for this situation may be that comparing with other test samples, test sample (3) had more distortion of texture information and loss of color information when it is relighted by RetinexNet. In contrast, SPMF can relight image with high visual quality, which is also one of the key factors for SPMF to achieve high detection results.

4.3.3. Visual Results on Real-World Images

In this subsection, we use two different imaging sensors, the SLR (Single Lens Reflex) camera and the iPhone 8 Plus, to obtain real-world low-light images and test the proposed method. Figure 8 shows the visual results of the proposed method on real-world low-light images. The first line shows the visual results obtained on the iPhone 8 Plus imaging sensor, and the second line shows the visual results obtained on the SLR imaging sensor. The pedestrians in the benchmark images are manually labeled.
We can intuitively see that our method not only achieves good detection performance, but it is also satisfactory for the relighting image, and can better recover pedestrians in real-world low-light images. These results further show that even if the proposed method is trained on low-light dataset in non-real-world scenes, the proposed method still maintains high reliability on low-light images in real world scenes. It also provides the reference and inspiration for the feasibility research of real-world scenarios in the future.

5. Conclusions

In this paper, a novel multi-task learning method is proposed, which uses a feature-level fusion and sharing mechanism to fuse the features from the image relighting and pedestrian detection networks, and then the fused features are learned for each network to boost relighting and detecting performances together. Meanwhile, we introduce a novel Self-Calibrated Split Attention Block as the basic block of backbone network, which further improves detecting performance. Experimental results show that the proposed multi-task learning method can also effectively improve the performance of pedestrian detection algorithms in the dark. However, our proposed method also has some limitations. The first point is that the method used to obtain low-light dataset cannot simulate the real-world low-light image well. In the later stage, we will add noise and nonlinear pixel transformation in image preprocessing. The second point is that the proposed method does not have high adaptive generalization ability. If the input is regular-lighting images, the image will be seriously damaged after entering the image relighting subnetwork. In the later stage, we will design a switch network to determine whether the input image needs image relighting. In the near future, we will compare the proposed multi-task feature fusing-sharing learning method with other excellent multi-task learning methods in different tasks, and some extension of multi-task learning should be investigated for other vision-based applications, such as person re-identification, semantic segmentation, pose estimation, etc.

Author Contributions

Y.W. (Yuanzhi Wang) and T.L. conceived the idea. T.L. and T.Z. analyzed the performance and designed the experiments. Y.W. (Yuanzhi Wang) performed the experiments and analyzed the data. Y.W. (Yuanzhi Wang) and Y.W. (Yuntao Wu) wrote the paper, all of the authors actively participated in review, and revised the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China (2016YFB1200100), the National Natural Science Foundation of China (62072350, 61771353, 91638301), Hubei Technology Innovation Project (2019AAA045), the Central Government Guides Local Science and Technology Development Special Projects (2018ZYYD059), 2020 Hubei Province High-value Intellectual Property Cultivation Project, the Wuhan Enterprise Technology Innovation Project (202001602011971), Scientific Research Foundation of Wuhan Institute of Technology (CX2019244, CX2019243, CX2019245).

Acknowledgments

The authors would like to thank the editors and the anonymous reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  2. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  3. Ren, S.; He, K.; Ross, G.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems 28, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  4. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  5. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  6. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  7. GAO, F.; LIU, A.; LIU, K.; YANG, E.; HUSSAIN, A. A novel visual attention method for target detection from SAR images. Chin. J. Aeronaut. 2019, 32, 1946–1958. [Google Scholar] [CrossRef]
  8. Gao, F.; Ma, F.; Wang, J.; Sun, J.; Yang, E.; Zhou, H. Visual Saliency Modeling for River Detection in High-Resolution SAR Imagery. IEEE Access 2018, 6, 1000–1014. [Google Scholar] [CrossRef] [Green Version]
  9. Shao, Z.; Wu, W.; Wang, Z.; Du, W.; Li, C. SeaShips: A Large-Scale Precisely Annotated Dataset for Ship Detection. IEEE Trans. Multimed. 2018, 20, 2593–2604. [Google Scholar] [CrossRef]
  10. Sermanet, P.; Kavukcuoglu, K.; Chintala, S.; Lecun, Y. Pedestrian Detection with Unsupervised Multi-stage Feature Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, Oregon, 9 April 2013; pp. 3626–3633. [Google Scholar]
  11. Shen, J.; Zuo, X.; Yang, W.; Prokhorov, D.; Mei, X.; Ling, H. Differential Features for Pedestrian Detection: A Taylor Series Perspective. IEEE Trans. Intell. Transp. Syst. 2019, 20, 2913–2922. [Google Scholar] [CrossRef]
  12. Paolanti, M.; Romeo, L.; Liciotti, D.; Cenci, A.; Frontoni, E.; Zingaretti, P. Person Re-Identification with RGB-D Camera in Top-View Configuration through Multiple Nearest Neighbor Classifiers and Neighborhood Component Features Selection. Sensors 2018, 18, 3471. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Liu, W.; Liao, S.; Hu, W.; Liang, X.; Chen, X. Learning Efficient Single-stage Pedestrian Detectors by Asymptotic Localization Fitting. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 618–634. [Google Scholar]
  14. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  15. Liu, W.; Liao, S.; Ren, W.; Hu, W.; Yu, Y. High-Level Semantic Feature Detection: A New Perspective for Pedestrian Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5182–5191. [Google Scholar]
  16. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  17. Law, H.; Teng, Y.; Russakovsky, O.; Deng, J. CornerNet-Lite: Efficient Keypoint Based Object Detection. arXiv 2019, arXiv:1904.08900. [Google Scholar]
  18. Kruthiventi, S.S.S.; Sahay, P.; Biswal, R. Low-light pedestrian detection from RGB images using multi-modal knowledge distillation. In Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 17–20 Spetember 2017; pp. 4207–4211. [Google Scholar]
  19. Ruder, S. An Overview of Multi-Task Learning in Deep Neural Networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  20. Collobert, R.; Weston, J. A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 June 2008; pp. 160–167. [Google Scholar]
  21. Ramsundar, B.; Kearnes, S.M.; Riley, P.; Webster, D.; Konerding, D.E.; Pande, V.S. Massively Multitask Networks for Drug Discovery. arXiv 2015, arXiv:1502.02072. [Google Scholar]
  22. Caruana, R. Multitask Learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  23. Caruana, R. Multitask Learning: A Knowledge-Based Source of Inductive Bias. In Proceedings of the Tenth International Conference on Machine Learning, San Francisco, CA, USA, 27–29 June 1993; pp. 41–48. [Google Scholar]
  24. Long, M.; Wang, J. Learning Multiple Tasks with Deep Relationship Networks. arXiv 2015, arXiv:1506.02117. [Google Scholar]
  25. Misra, I.; Shrivastava, A.; Gupta, A.; Hebert, M. Cross-Stitch Networks for Multi-Task Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3994–4003. [Google Scholar]
  26. Gao, Y.; Ma, J.; Zhao, M.; Liu, W.; Yuille, A.L. NDDR-CNN: Layerwise Feature Fusing in Multi-Task CNNs by Neural Discriminative Dimensionality Reduction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 21–25 June 2019; pp. 3205–3214. [Google Scholar]
  27. Gao, Y.; Bai, H.; Jie, Z.; Ma, J.; Jia, K.; Liu, W. MTL-NAS: Task-Agnostic Neural Architecture Search Towards General-Purpose Multi-Task Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020; pp. 11543–11552. [Google Scholar]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  29. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  30. Jiang, K.; Wang, Z.; Yi, P.; Wang, G.; Lu, T.; Jiang, J. Edge-Enhanced GAN for Remote Sensing Image Superresolution. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5799–5812. [Google Scholar] [CrossRef]
  31. Wang, Z.; Yi, P.; Jiang, K.; Jiang, J.; Han, Z.; Lu, T.; Ma, J. Multi-Memory Convolutional Neural Network for Video Super-Resolution. IEEE Trans. Image Process. 2019, 28, 2530–2544. [Google Scholar] [CrossRef] [PubMed]
  32. Zhou, L.; Wang, Z.; Luo, Y.; Xiong, Z. Separability and Compactness Network for Image Recognition and Superresolution. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3275–3286. [Google Scholar] [CrossRef] [PubMed]
  33. Yi, P.; Wang, Z.; Jiang, K.; Shao, Z.; Ma, J. Multi-Temporal Ultra Dense Memory Network for Video Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2503–2516. [Google Scholar] [CrossRef]
  34. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  35. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  36. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar]
  37. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  38. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  39. Liu, J.J.; Hou, Q.; Cheng, M.M.; Wang, C.; Feng, J. Improving Convolutional Networks with Self-Calibrated Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10096–10105. [Google Scholar]
  40. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Zhang, Z.L.; Lin, H.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. ResNeSt: Split-Attention Networks. arXiv 2020, arXiv:2004.08955. [Google Scholar]
  41. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 510–519. [Google Scholar]
  42. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. In Proceedings of the British Machine Vision Conference. British Machine Vision Association, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  43. Alejandro, N.; Jia, D. Pixels to Graphs by Associative Embedding. In Advances in Neural Information Processing Systems 31; Curran Associates, Inc.: Long Beach, CA, USA, 2017; pp. 2171–2180. [Google Scholar]
  44. Zhang, S.; Benenson, R.; Schiele, B. CityPersons: A Diverse Dataset for Pedestrian Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4457–4465. [Google Scholar]
  45. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian Detection: An Evaluation of the State of the Art. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 743–761. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Three multi-task learning methods. (A,B) Two widely used multi-task learning methods. (C) The method proposed in this paper, which is a novel multi-task learning method with feature-level fusion and sharing.
Figure 1. Three multi-task learning methods. (A,B) Two widely used multi-task learning methods. (C) The method proposed in this paper, which is a novel multi-task learning method with feature-level fusion and sharing.
Sensors 20 05852 g001
Figure 2. The architecture of SPMF includes three parts: image relighting subnetwork, pedestrian detection subnetwork, and feature-level multi-task fusion learning module. The feature-level multi-task fusion learning module uses feature-level fusion and sharing mechanism to fuse features from two subnetworks, and then shares fused features to these two subnetworks.
Figure 2. The architecture of SPMF includes three parts: image relighting subnetwork, pedestrian detection subnetwork, and feature-level multi-task fusion learning module. The feature-level multi-task fusion learning module uses feature-level fusion and sharing mechanism to fuse features from two subnetworks, and then shares fused features to these two subnetworks.
Sensors 20 05852 g002
Figure 3. Architecture of SCSAB. AvgPool layer is used to produce a small scale space and Up is a bilinear interpolation operator that maps the small size feature space to the original size feature space.
Figure 3. Architecture of SCSAB. AvgPool layer is used to produce a small scale space and Up is a bilinear interpolation operator that maps the small size feature space to the original size feature space.
Sensors 20 05852 g003
Figure 4. The mechanism of multi-task learning method with feature-level fusion and sharing.
Figure 4. The mechanism of multi-task learning method with feature-level fusion and sharing.
Sensors 20 05852 g004
Figure 5. Visual results (end-to-end testing mode) from the different detection methods. Five images from the low-light CityPersons testing set is selected as the samples to show the detected results (with marked bounding box in different color).
Figure 5. Visual results (end-to-end testing mode) from the different detection methods. Five images from the low-light CityPersons testing set is selected as the samples to show the detected results (with marked bounding box in different color).
Sensors 20 05852 g005
Figure 6. The results of five selected test samples passing through the RetinexNet network.
Figure 6. The results of five selected test samples passing through the RetinexNet network.
Sensors 20 05852 g006
Figure 7. Visual results (cascade testing mode) from the different detection methods. Five images from the low-light CityPersons testing set is selected as the samples to show the detected results.
Figure 7. Visual results (cascade testing mode) from the different detection methods. Five images from the low-light CityPersons testing set is selected as the samples to show the detected results.
Sensors 20 05852 g007
Figure 8. The visual results on real-world images.
Figure 8. The visual results on real-world images.
Sensors 20 05852 g008
Table 1. The parameters of different backbone networks.
Table 1. The parameters of different backbone networks.
BackbonesParameters
Hourglass-54116.97 M
SCSAB-Hourglass100.74 M
SCSAB-HourglassLite50.18 M
Table 2. The experimental results of different backbone networks in SPMF.
Table 2. The experimental results of different backbone networks in SPMF.
Backbones MR 2 (%)
IoU = 0.5IoU = 0.75IoU = 0.5:0.95
Hourglass-5415.962.364.8
SCSAB-Hourglass12.556.658.8
SCSAB-HourglassLite11.757.260.4
Table 3. Comparison with the state-of-the-art on low-light CityPersons (End-to-end testing mode).
Table 3. Comparison with the state-of-the-art on low-light CityPersons (End-to-end testing mode).
Methods MR 2 (%)
IoU = 0.5IoU = 0.75IoU = 0.5:0.95
CSP [15]24.864.459.9
ALFNet [13]35.662.562.6
CenterNet [14]22.864.762.7
CornerNet [16]40.279.978.0
CornerNet-Saccade [17]35.376.677.8
SPMF (Ours)12.556.658.8
SPMF-Lite (Ours)11.757.260.4
Table 4. Comparison with the state of the arts on low-light CityPersons (Cascade testing mode).
Table 4. Comparison with the state of the arts on low-light CityPersons (Cascade testing mode).
Methods MR 2 (%)
IoU = 0.5IoU = 0.75IoU = 0.5:0.95
RetinexNet+CSP [15]31.556.759.3
RetinexNet+ALFNet [13]33.559.660.1
RetinexNet+CenterNet [14]24.165.964.3
RetinexNet+CornerNet [16]39.977.976.5
RetinexNet+CornerNet-Saccade [17]29.971.473.1
SPMF (Ours)12.556.658.8
SPMF-Lite (Ours)11.757.260.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Y.; Lu, T.; Zhang, T.; Wu, Y. Seeing Pedestrian in the Dark via Multi-Task Feature Fusing-Sharing Learning for Imaging Sensors. Sensors 2020, 20, 5852. https://doi.org/10.3390/s20205852

AMA Style

Wang Y, Lu T, Zhang T, Wu Y. Seeing Pedestrian in the Dark via Multi-Task Feature Fusing-Sharing Learning for Imaging Sensors. Sensors. 2020; 20(20):5852. https://doi.org/10.3390/s20205852

Chicago/Turabian Style

Wang, Yuanzhi, Tao Lu, Tao Zhang, and Yuntao Wu. 2020. "Seeing Pedestrian in the Dark via Multi-Task Feature Fusing-Sharing Learning for Imaging Sensors" Sensors 20, no. 20: 5852. https://doi.org/10.3390/s20205852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop