Next Article in Journal
ST-PN: A Spatial Transformed Prototypical Network for Few-Shot SAR Image Classification
Previous Article in Journal
Rapid Tsunami Potential Assessment Using GNSS Ionospheric Disturbance: Implications from Three Megathrusts
Previous Article in Special Issue
Self-Supervised Stereo Matching Method Based on SRWP and PCAM for Urban Satellite Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy

1
Yunnan Key Laboratory of Computer Technologies Application, Kunming University of Science and Technology, Kunming 650500, China
2
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(9), 2020; https://doi.org/10.3390/rs14092020
Submission received: 14 March 2022 / Revised: 15 April 2022 / Accepted: 20 April 2022 / Published: 22 April 2022

Abstract

:
In the field of remote sensing image applications, RGB and infrared image object detection is an important technology. The object detection performance can be improved and the robustness of the algorithm will be enhanced by making full use of their complementary information. Existing RGB-infrared detection methods do not explicitly encourage RGB and infrared images to achieve effective multimodal learning. We find that when fusing RGB and infrared images, cross-modal redundant information weakens the degree of complementary information fusion. Inspired by this observation, we propose a redundant information suppression network (RISNet) which suppresses cross-modal redundant information and facilitates the fusion of RGB-Infrared complementary information. Specifically, we design a novel mutual information minimization module to reduce the redundancy between RGB appearance features and infrared radiation features, which enables the network to take full advantage of the complementary advantages of multimodality and improve the object detection performance. In addition, in view of the drawbacks of the current artificial classification of lighting conditions, such as the subjectivity of artificial classification and the lack of comprehensiveness (divided into day and night only), we propose a method based on histogram statistics to classify lighting conditions in more detail. Experimental results on two public RGB-infrared object detection datasets demonstrate the superiorities of our proposed method over the state-of-the-art approaches, especially under challenging conditions such as poor illumination, complex background, and low contrast.

1. Introduction

Object detection technology is one of the main technologies in the computer vision field. According to the different detection steps, the object detection technology can be divided into two types. One is the two-stage object detection technology. The detection process of this technology is to first generate a proposal and then perform fine-grained object detection. Its representative algorithms include Fast- RCNN [1], Faster-RCNN [2], etc. The other is the one-stage object detection technology. The detection process of this technology is to directly extract the image features in the network to predict the object category and location, and its representative algorithms include RetinaNet [3], SSD [4], etc.
Currently, object detection methods mainly use single-mode images related to the object detection task as training data. For example, Liu et al. used the RGB dataset collected in harsh environments to train models to detect vehicles and pedestrians [5]; Chen et al. used RGB images of KITTI and other datasets for cross-domain adaptive learning, making the model robust in different scenarios [6]. However, when the model using RGB images as training data is applied in complex environments, such as night, cloudy and rainy days, its detection effect is often unsatisfactory and its robustness is poor. Therefore, in view of the robustness problem caused by using RGB as training data, some scholars have proposed the scheme of using the infrared image as training data. For example, Zhao et al. used the infrared image dataset to train the model for the task of pedestrian detection [7]; Dai et al. used the SIRST dataset to train the model for the task of infrared small target detection [8]. However, when the infrared image is used as the training data of the model, the detection accuracy of the algorithm is low since the infrared image lacks the texture information of the target. In addition, there are many small target problems in application scenarios such as remote sensing, and the detection performance of relevant algorithms needs to be improved, such as DLAO [9] and DRA [10]. It is worth mentioning that drawing on face detection technology is an effective way to improve the performance of related algorithms in detecting small targets [11].
In summary, only using single-modal images as the training data of the model will make it difficult to apply in the actual complex scene. Therefore, in order to overcome the above difficulties, many scholars have proposed the method of using multimodal image data as the training data for the detection model. Multimodal data have complementary advantages, such as RGB-infrared images. The advantage of infrared images is that they rely on heat imaging generated by the target and are not affected by external lighting conditions. The advantage of RGB images is that they can clearly capture the texture characteristics of the target, but are susceptible to lighting conditions. Therefore, research on multispectral object detection based on multimodal input has become a hot topic. In recent years, there’s a growing joint use of RGB and infrared images for related research in various fields. For example, in the field of remote sensing, Dang et al. jointly use RGB and UAV near-infrared images to detect radish blight [12], and Y et al. jointly uses RGB and infrared images for terrain classification [13]. In the field of assisted driving, Li et al. jointly use RGB and infrared images for roadside Pedestrian detection [14], and Tian et al. jointly uses RGB and far-infrared images for 3D pedestrian detection in farmland [15]. However, even though multimodal data has great advantages, how efficiently fusing complementary information of multimodal data has become the core and difficulty of current algorithms.
Current cross-modal object detection algorithms based on RGB and infrared images mostly fuse complementary information by simply concatenating input features or designing an attention module to generate robust target features [16,17,18,19,20,21,22,23]. In terms of concatenating fusion, based on the two-stage object detection algorithm Faster-RCNN, Li et al. directly spliced RGB and infrared image features to obtain fusion features, and then input them into the RPN module [16]. Wagner et al. designed two direct fusion models to analyze their effects [17]. The first is early fusion, which directly combines RGB and infrared images at the pixel level. The second is late fusion, which fuses RGB and infrared image features in the final fully connected layer. Their analysis shows that the late fusion effect is more effective. In terms of attention-based mechanisms, Zhang et al. established two separate feature hierarchies for each modality, and then, by obtaining global features, the correlation between the two modalities was encoded into the attention module for subsequent fusion operations [18]. Zhang et al. proposed a cross-spectral fusion network, which added a channel attention module (CAM) [24] and a spatial attention module (SAM) [25] along the channel and spatial dimensions, making the network learn strongly meaningful features while suppressing interference at both channel and spatial levels [19]. Fu et al. proposed an adaptive spatial pixel-level feature fusion network. Through SAM and position attention module (PAM) [24] respectively, their network obtains the fusion weights of different spatial positions and pixels in the two feature maps, and uses the fusion weights to re-calibrate the original features of the RGB and infrared images to obtain a multi-scale fusion feature layer [20]. Cao et al. proposed a new multi-spectral channel feature fusion module to fuse the RGB and infrared image features, which uses the squeeze-and-excitation (SE) [26] module to obtain the fusion weights, and uses the fusion weights to recalibrate the original features [21]. However, these methods do not make full use of the complement information between different modalities, and do not give full play to the advantages of multimodal complementarity. In addition, these methods do not consider the redundant information between multiple modalities when fusing complementary information, which leads to the fusion of redundant information when fusing complementary information.
The lighting condition of most RGB and infrared image datasets is artificially divided into two categories (day and night), such as the KAIST pedestrian dataset [27] and the DroneVehicle remote sensing dataset [28]. The criterion of classifying lighting conditions for the KAIST pedestrian dataset is that direct classification according to whether the shooting time is day or night. Similarly, the DroneVehicle remote sensing dataset divides the dataset into three lighting scenarios, daytime, nighttime, and dark nighttime, according to the time of the shooting. However, there are two obvious drawbacks in this way. The first is that the artificial classification is subjective, and it cannot accurately determine whether the light intensity of the image belongs to the day or night. For example, in the night scene category, there are also images with strong illumination caused by light sources such as street lights. The second is that only two lighting scenes are not comprehensive enough to fully reflect the actual lighting environment.
The above two problems restrict the further combined use of RGB and infrared images. Therefore, in order to solve the contradiction between complementary information and redundant information contained in different modalities, this paper proposes to design mutual information module to reduce redundant information between different modalities, and use an attention mechanism and illumination perception module to fuse complementary information of different modalities. In addition, given the subjectivity and incompleteness of artificial classification of lighting conditions, we propose a lighting condition classification method that uses histograms to statistics the gray values of RGB images and considers different gray value intervals as different lighting intensities.
In summary, the main contributions of this paper are listed as follows:
  • This paper designs a mutual information module, which weakens the influence of redundant information on complementary information fusion.
  • A lighting condition classification method based on histogram statistics is proposed to automatically classify lighting conditions in more detail, which facilitates the complementary fusion of infrared and RGB images.
The rest of this paper is structured as follows: In Section 2, we describe the network structure and methods in detail. Section 3 gives the details of our work and experimental results and related comparison to verify the effectiveness of our method. Finally, we summarize the research content in Section 4.

2. Methods

2.1. Overall Network Architecture

The overall architecture of the proposed RISNet is shown in Figure 1. RISNe is divided into four modules: feature extraction module, illumination perception module, feature fusion module, and detection module. Among them, the mutual information module proposed in this paper is embedded in the feature extraction module. The feature fusion module adds the CAM, and the detection module uses the light intensity information extracted by the illumination perception module, and combines the RGB and infrared image features to make the final detection.
Firstly, the input data to the network model is an RGB-infrared image pair. The feature map is obtained by extracting image features through the feature extraction module, and then the corresponding feature map is input into the mutual information module. Then the information entropy of the two feature maps is calculated, and the parameters of the feature extraction module are optimized by the information entropy, which reduces the redundant information of RGB-infrared image features. In the subsequent feature fusion process, the CAM captures the key regions of the RGB-infrared feature map, thus focusing on fusing the complementary information of the network attention region part. Based on the fact that the light intensity information is beneficial to improving the object detection effect, this paper adds the illumination perception module to extract the light intensity information from RGB images. Finally, the detection module combines the features obtained by the feature fusion module and the light intensity information captured by the illumination perception module to predict the location and category of the object.

2.2. Mutual Information Module

One of the difficulties in target detection using RGB-infrared images is that there is both complementary and redundant information. Ideally, we want the network to fuse only the complementary features of RGB-infrared images. However, in practice, when the network fuses complementary information of RGB-infrared images, it inevitably fuses redundant information, resulting in poor detection performance of multimodal algorithms. Therefore, we filter the redundant information of RGB-infrared images in the feature extraction stage to resolve the contradiction between complementary information and redundant information in the feature fusion stage.
Therefore, based on the above ideas and inspired by the concept of information entropy in “information theory”, we first treat RGB and infrared image features as two information distributions, and measure the amount of duplicate information between information distributions using mutual information [29,30]. Mutual information can be used to measure the amount of duplicate information between information distributions, which can optimize feature parameters, and finally reduce redundant information of infrared and RGB image features.
The mutual information module is shown in Figure 2. The input data of the module is the RGB-infrared image feature obtained by the feature extraction module. In order to reduce the parameters, computation of the network and facilitate the subsequent creation of the information distribution, we first map the RGB-infrared image feature to two low-dimensional feature vectors. In the module, we design two convolution layers and full connection layers to map the RGB-infrared image features to two 1 × 6 vectors Z r and Z i . Then based on the two low-dimensional feature vectors Z r and Z i , we create two information distributions (such as using Z r or Z i as the mean and Z i or Z r as the variance to create a Gaussian distribution in this paper), and then calculate the mutual information of the two information distributions:
M ( Z r , Z i ) = H ( Z r ) + H ( Z i ) H ( Z r , Z i )
where M ( Z r , Z i ) represents the mutual information of low-dimensional feature vectors Z r and Z i . H ( Z r ) and H ( Z i ) indicate the information entropy of low-dimensional feature vectors Z r and Z i , respectively. H ( Z r , Z i ) represents the feature joint entropy of low-dimensional feature vectors Z r and Z i . The relationship between information entropy, cross-entropy and relative entropy is as follows:
H ( Z i ) = C ( Z r , Z i ) D ( Z r | | Z i )
H ( Z r ) = C ( Z i , Z r ) D ( Z i | | Z r )
where C ( Z r , Z i ) is the cross-entropy of low-dimensional feature vectors Z i and Z r , D ( Z r | | Z i ) is the relative entropy of low-dimensional feature vectors Z r and Z i . We then sum Equations (1)–(3), and obtain:
M ( Z r , Z i ) = C ( Z r , Z i ) + C ( Z i , Z r ) D ( Z r | | Z i ) D ( Z i | | Z r ) H ( Z r , Z i )
In order to simplify Equation (4) and consider the non-negative property of joint entropy, we delete H ( Z r , Z i ) , and finally get the formula for optimizing mutual information as Equation (5):
L m = C ( Z r , Z i ) + C ( Z i , Z r ) D ( Z r | | Z i ) D ( Z i | | Z r )
L m represents the amount of duplicate information of RGB-infrared image features. By optimizing L m , the feature extraction module parameters are optimized, and then the redundant information in RGB-infrared image features is filtered to obtain rich complementary information. Resolves the contradiction between the inability to distinguish and fusion complementary information and redundant information in the subsequent network fusion stage.

2.3. Lighting Condition Classification Based on Histogram Statistics

The subjectivity and incompleteness of artificially classified lighting conditions is another important reason that restricts the fusion of RGB-infrared images for object detection. Therefore, we propose a method based on histogram statistics to automatically classify lighting conditions in more detail. In order to further solve the incompleteness of classifying lighting scenes, we have made a detailed division of the gray value interval, covering as many lighting scenes as possible, to reflect lighting conditions in more detail.
First, we use the histogram to statistic the gray values of RGB images under different light intensities, and then divide the lighting scene according to the corresponding gray value intervals. As shown in Figure 3, we can clearly see from the histogram corresponding to the RGB image that the gray value of the RGB image in full daytime is concentrated in (100, 255), the gray value of the RGB image at night but with strong light is concentrated in (34, 100), the gray value of the RGB image with weak light at night is concentrated in (10, 34), and the gray value of the completely dark RGB image is concentrated in (0, 10). Therefore, based on the above four gray value intervals, we divide the scene lighting condition into four categories: daytime, weak light at night, strong light at night and dark night.

2.4. Illumination Perception Module

The illumination perception module is designed to extract the information of light intensity from RGB images, and to judge the reliability and importance of the RGB-infrared image features. If the RGB image has strong light intensity information, it means that the lighting condition is good. Visually, when the lighting conditions are good, the RGB image features can be used for detection. Otherwise, under poor lighting conditions, the infrared image features are mainly used to detect, taking advantage of infrared images.
As shown in Figure 1, the input data of the illumination perception module is the RGB image resized to 56 × 56. After the convolution layer and the full connection layer, four predicted values are generated: W d , W s , W w , and W n , which represent the confidence of the daytime, strong light at night, weak light at night, and dark night, respectively. Then, W d and W s are added, W w and W n are added, respectively. Then, pass the sigmoid function:
W l _ s = sigmoid ( W d + W s )
W l _ w = sigmoid ( W w + W n )
where W l _ s represents the weight of strong light intensity information and W l _ w represents the weight of light intensity information.

2.5. Feature Fusion Module and Detection Module

The core of the multimodal detection algorithm is how to effectively fuse the features between different modalities. Therefore, in order to effectively fuse the complementary information of RGB-infrared images, a channel attention mechanism is specially added to the feature fusion module, as shown in Figure 1. Firstly, the RGB-infrared image features are subtracted element by element, and the subtracted features represent the different features of RGB infrared image features, which are regarded as complementary features of the RGB-infrared image. Then the different feature is input into the CAM to obtain the channel weight. Finally, the channel weight is multiplied by the corresponding RGB-infrared image feature, and added with another modal feature to obtain the final fused feature, as shown in the formula:
F i = CAM ( F r F i ) F r + F i
F r = CAM ( F i F r ) F i + F r
Among them, the represents the multiplication operation, F r and F i are the feature of RGB and infrared image before fusion, F r and F i are the fused RGB and infrared image features.
The input data of the detection module is the light intensity information W l _ s , W l _ w and the final fused feature F r and F i , Through the additional illumination information, more robust detection can be achieved. A more detailed design of the feature fusion module and prediction module can refer to MBNet [31].
Therefore, the objective function of the network is divided into four parts: illumination loss L i , mutual information loss L m , classification loss L c l s and regression loss L r e g , where the illumination loss L i uses cross-entropy loss, the classification loss L c l s uses Focal loss, and the regression loss L r e g uses smooth L1 loss. Our final objective function is:
L = λ L m + L i + L c l s + L r e g
Among them, in order to balance the learning, the loss weight is set to 0.1.

3. Experiment and Analysis

In order to verify the effectiveness of the mutual information optimization module for reducing redundant information between modalities, and proves that the new lighting condition classification method is helpful to optimize the illumination perception module, we conduct experiments on the KAIST pedestrian dataset and DroneVehicle remote sensing dataset.

3.1. Dataset Introduction

KAIST pedestrian dataset contains 95,328 RGB-infrared image pairs, which contain three categories: person, people, and cyclist. Among them, the better-distinguished individuals are labeled as person. Not distinguishable individuals were labeled as people, and riding a two-wheeled vehicle were labeled as cyclists. Since the dataset is taken from consecutive frames of video, the adjacent pictures are not much different, so the dataset is cleaned to a certain extent, and pedestrians with only half or less than 50 pixels are eliminated. Therefore, the final training set contains 8963 pairs of images, of which 5622 were captured during the day, and 3341 were captured at night. The test set contains 2252 image pairs, of which 1455 were captured during the day and 797 were captured at night.
DroneVehicle remote sensing dataset contains 28,439 RGB-infrared image pairs, which are classified as vehicles and captured by UAVs equipped with cameras. The regional scenes are divided into urban roads, residential areas and highways. The lighting conditions during shooting are divided into dark night, night and day. The height of shooting is divided into 80 m, 100 m, and 120 m. The shooting angle is divided into 15°, 35°, and 45°. During the dataset calibration phase, RGB-infrared image pairs are cropped by affine transformation and region cropping to ensure that most of the cross-modality image pairs are aligned.
Some samples in the above dataset are shown in Figure 4.

3.2. Implementation Details

All experiments in this paper were implemented in Tensorflow on NVIDIA Tesla V100, and its operating system is ubuntu16.04, and the CUDA version is 10.2. For the processing of the dataset, this paper performs random color distortion and flip processing on the input image to achieve the purpose of increasing the diversity of data. For parameter settings during training, the whole network is trained by an adam optimizer for 30 epochs with an initial learning rate of 0.00001, and the learning rate is reduced by 10 times every 6 epochs. The batch size is set to 8. Followed by [32], the width of initial anchors are set to (25.84, 29.39), (33.81, 38.99), (44.47, 52.54), (65.80, 131.40) for four feature fusion stage with a single anchor ratio of 0.41. The dimension of feature vectors Z r and Z i is set to 6 through experimental analysis.

3.3. Evaluation Metrics

Precision, recall, average precision (AP), miss rate (MR), log-average miss rate (denoted as MR−2) etc., are general performance evaluation indicators of object detection algorithms. In general, the better the classifier, the higher the AP and the lower the MR. In this paper, we adopt AP and MR−2 as evaluation metrics. Among them, the calculation formula of precision and recall are as follows:
P r e c i s i o n = T P T P + F P
Recall = T P T P + F N
TP (true positive) represents the positive samples that were correctly detected; TN (true negative) represents the negative samples that were correctly detected; FP (false positive) represents the positive samples that were wrongly detected; FN (false negative) represents the negative samples that were wrongly detected. Among them, the standard for judging whether it is a positive or negative sample is IOU, and its calculation formula is as follows:
IOU = D G D G
where D represents the predicted object bounding box area, and G represents the real object bounding box area. The IOU threshold in this paper is set to 0.5. If the IOU is greater than 0.5, the sample is considered to be a positive sample.
Taking Precision as the horizontal axis and Recall as the vertical axis, the area enclosed by the curve and the coordinate axis is the AP.
The MR reflects the ability of the model to correctly predict the purity of negative samples, and is calculated as follows:
MR = 1 R e c a l l
The calculation method of MR−2: in the logarithmic coordinate system, evenly take 9 FPPI (False Positives Per Image) values from 10−2 to 100, and average the 9 MR values corresponding to these 9 FPPI values. Among them, FPPI is calculated as follows:
FPPI = F P the   number   of   image

3.4. Analysis of Results

Firstly, comparison, and ablation experiments are carried out on the DroneVehicle remote sensing dataset, which proves that the method proposed in this paper is competitive. In order to further verify that the mutual information module is beneficial to RGB-infrared image complementary information fusion, a large number of comparative experiments were carried out on the KAIST pedestrian dataset, which verified that the mutual information module has a great effect on reducing redundant information between modalities.

3.4.1. Experiments on the DroneVehicle Remote Sensing Dataset

We compare with DroneVehicle remote sensing dataset benchmarks [28], which include CMDet, UA-CMDet. We also use Yolov3 [33], Faster-RCNN [2] and MBNet [31] as the baselines to verify the effectiveness of the proposed method. Experiments show that the RISNet proposed in this paper has obvious advantages. The specific experimental results are shown in Table 1.
Compared with the benchmark indicators of this dataset, the experimental results show that the AP value of our proposed RISNet is 66.40%, which has achieved the optimal effect, proving the effectiveness of RISNet for multimodal data detection.
In order to more clearly see the effectiveness of the mutual information module proposed in this paper, we show test image samples under different lighting scenarios, as shown in Figure 5.
From the comparative detection effect images of different lighting scenes in Figure 5, It is clearly seen that compared with the baseline, our proposed network is more robust, and still has a certain detection ability under various lighting conditions. In conclusion, it is proved that the mutual information module proposed in this paper and the histogram-based classification of lighting scenes are advanced, and provide a new idea for the joint utilization of RGB-infrared images.

3.4.2. Experiments on the KAIST Pedestrian Dataset

In order to further verify the effect of the mutual information module, we compare it with the current advanced multispectral pedestrian detection algorithms, such as MSDS-RCNN [32], AR-CNN [34], CIAN [18], MBNet [31]. We also use Yolov3 [33] as the baseline. The experimental results show that the proposed RISNet has obvious advantages over other algorithms. The experimental results are shown in Table 2.
Experimental results show that our proposed RISNet achieves the best MR−2 in 8 subsets in the evaluation MR−2 metric of nine subsets., Among, we obtain 7.89 MR−2, 7.61 MR−2, 7.08 MR−2 in all day, daytime and night, respectively. In terms of pedestrian distance subsets, we get 0.00 MR−2, 14.01 MR−2, 52.67 MR−2 on the near, medium, and far subsets, respectively. In terms of pedestrian occlusion degree subsets, we obtained 25.23 MR−2, 34.25 MR−2 on no occlusion, partial occlusion subsets, respectively. Obviously, the most MR−2 value of our proposed RISNet is lower than the best competitive algorithm MBNet. The reason for this is that our proposed mutual information module can effectively reduce the redundant information of RGB-infrared features. For example, in dark conditions, the target is covered and the small target detection, the mutual information module reduces the redundant information of RGB-infrared features, so that the contour information of the target in the infrared image and the texture information of the target in the RGB image can fully play a role, and finally detected the target effectively.
In order to clearly compare the detection effects of different algorithms, we took out test image samples of different scenarios. As shown in Figure 6.
From the comparative detection effect images of different scenes in Figure 6, we can clearly see that MBNet is prone to miss detection in different scenes, but in the detection effect images of RISNet, it can be accurately detected. Therefore, it is proved that the mutual information module proposed in this paper can improve the missed detection of the algorithm by reducing the redundant information of different modalities.
In order to compare the detection speed of various algorithms, we tested their FPS on different computing platforms and the results are shown in Table 3.
Comparing the FPS of each algorithm, it is obvious that the Yolov3 of the one-stage algorithm is the fastest, which effectively improves the detection speed at the expense of certain accuracy. Compared with the multimodal object detection algorithms, the fastest one is to reach 14 FPS, and the slowest one is only 5 FPS. Our proposed RISNet achieves 10 FPS, surpassing most multi-modal object detection algorithms, and can be more easily deployed to mobile terminals. In fact, FPS is not only related to the complexity of the model, but also related to the hardware platform tested. Since the computing performance of GTX 2060 is the worst, the proposed RISNet will achieve higher FPS on a better computing platform.

3.5. Ablation Experiments

In order to prove the effectiveness of the mutual information module and the new lighting condition classification method for the detection algorithm, we conducted ablation experiments in different lighting conditions and datasets, and the experimental results are shown in Table 4 and Table 5.
As shown in Table 4 and Table 5, when the mutual information module is added to the baseline model, the AP values of the four lighting scenes are improved by about 0.98% to 2.6%, which indicates that adding the mutual information module helps to eliminate redundant information of RGB-infrared image information and gives full play to the complementary advantages of RGB-infrared image information. But for different lighting scenarios, it cannot be improved to the same extent and equally. For example, lighting scenarios of weak light at night increased by 2.6%, while lighting scenarios of daytime only increased by 0.98%, which shows that the mutual information module can improve the algorithm detection performance, but it is still easily affected by the lighting scene, and cannot achieve the same degree and balance improvement of algorithm detection ability under different lighting conditions. When the new lighting scene classification method is introduced in the baseline model, the AP values of the four lighting scenes have been improved to the same extent and equally, which shows that the introduction of the new lighting scene classification method can enhance the robustness of the algorithm to different lighting conditions, but the improvement effect is not outstanding. On the basis of the baseline model, the mutual information module is added and the new lighting scene classification method is introduced, that is, RISNet proposed in this paper. The results show that under different lighting scenes, the improvement can be equal and excellent.
In conclusion, these results show that the mutual information module can improve the detection effect of the algorithm by reducing the redundant information of different modalities; it proves the scientificity of classifying lighting scenes based on the histogram, which makes the algorithm robust under different illumination conditions.

4. Conclusions and Discussion

In this work, we propose a new network named RISNet. It alleviates the problem of insufficient complementary information fusion caused by redundant information by designing the mutual information optimizer. Specifically, the mutual information module optimizes the feature extraction network parameters by calculating the mutual information of infrared and RGB image features, thereby reducing redundant information between modalities. Secondly, in view of the subjectivity and incompleteness problems faced by the current manual classification of lighting conditions, a method based on histogram statistics is proposed to solve the subjectivity of manual classification of lighting conditions, and through the detailed division of gray value interval, it covers as many lighting conditions as possible and solves the problem of the incompleteness of lighting conditions. The feature fusion module adds an attention mechanism to fully fuse the complementary information of infrared and RGB image features. In addition, an illumination perception module is added to the model to complement adaptive lighting perception features, making the detector robust to different lighting. From the experimental results, the proposed RISNet can reduce the redundant information and fully integrate the complementary information between multimodalities. In different lighting environments, it has better detection ability, reflecting the characteristics of high precision and strong robustness. Compared with the multimodal detection methods in recent years, our proposed RISNet demonstrates the superiority on most of the indicators.
There are a large number of small targets in remote sensing images, and the problem of small targets greatly affects the detection effect. Therefore, in future work, we will focus on how to improve the small target detection accuracy and further improve the detection effect of multimodal object detection methods. In addition, we will explore more effective ways to fuse multimodal features. Moreover, we continue to explore the relationship between complementary and redundant information of multimodal fusion in other computer vision tasks, such as RGB-infrared image semantic segmentation, RGB-Depth image semantic segmentation and so on.

Author Contributions

Conceptualization, Q.W. and T.S.; methodology, Q.W.; software, Y.C.; validation, Z.Z., J.S. and Y.Z.; formal analysis, Q.W. and Y.C.; investigation, J.S.; resources, T.S.; data curation, Y.C.; writing—original draft preparation, Y.C.; writing—review and editing, Q.W.; visualization, Y.C.; supervision, T.S. and Y.Z.; project administration, T.S.; funding acquisition, Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Opening Foundation of Yunnan Key Laboratory of Computer Technologies Application (Grant No. 2021CTA08), the Yunnan Fundamental Research Projects (Grant No. 202101BE070001-008), the National Natural Science Foundation of China (Grant No. 61971208), the Yunnan Reserve Talents of Young and Middle-aged Academic and Technical Leaders (Grant No. 2019HB005), the Yunnan Young Top Talents of Ten Thousands Plan (Grant No. 201873), and the Major Science and Technology Projects in Yunnan Province (Grant No. 202002AB080001-8).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The DroneVehicle remote sensing dataset is obtained from https://github.com/VisDrone/DroneVehicle, accessed on 29 December 2021. The KAIST pedestrian dataset is obtained from https://github.com/SoonminHwang/rgbt-ped-detection/tree/master/data, accessed on 12 November 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Washington, DC, USA, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  2. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Cambridge, MA, USA, 7–12 December 2015; Volume 28, pp. 91–99. [Google Scholar]
  3. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  4. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  5. Liu, W.; Ren, G.; Yu, R.; Guo, S.; Zhu, J.; Zhang, L. Image-Adaptive YOLO for Object Detection in Adverse Weather Conditions. arXiv 2021, arXiv:2112.08088. [Google Scholar]
  6. Chen, Y.; Li, W.; Sakaridis, C.; Dai, D.; Gool, L.V. Domain adaptive faster r-cnn for object detection in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 3339–3348. [Google Scholar]
  7. Zhao, C.; Wang, J.; Su, N.; Yan, Y.; Xing, X. Low Contrast Infrared Target Detection Method Based on Residual Thermal Backbone Network and Weighting Loss Function. Remote Sens. 2022, 14, 177. [Google Scholar] [CrossRef]
  8. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  9. Su, N.; Huang, Z.; Yan, Y.; Zhao, C.; Zhou, S. Detect Larger at Once: Large-Area Remote-Sensing Image Arbitrary-Oriented Ship Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  10. Pan, X.; Ren, Y.; Sheng, K.; Dong, W.; Yuan, H.; Guo, X.; Ma, C.; Xu, X. Dynamic refinement network for oriented and densely packed object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–18 June 2020; pp. 11207–11216. [Google Scholar]
  11. Minaee, S.; Luo, P.; Lin, Z.; Bowyer, K. Going deeper into face detection: A survey. arXiv 2021, arXiv:2103.14983. [Google Scholar]
  12. Dang, L.M.; Wang, H.; Li, Y.; Min, K.; Kwak, J.T.; Lee, O.N.; Park, H.; Moon, H. Fusarium Wilt of Radish Detection Using RGB and Near Infrared Images from Unmanned Aerial Vehicles. Remote Sens. 2020, 12, 2863. [Google Scholar] [CrossRef]
  13. Iwashita, Y.; Nakashima, K.; Stoica, A.; Kurazume, R. TU-Net and TDeepLab: Deep Learning-Based Terrain Classification Robust to Illumination Changes, Combining Visible and Thermal Imagery. In Proceedings of the IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28–30 March 2019; pp. 280–285. [Google Scholar]
  14. Li, X.; Li, X.; Fabian, F.; Wang, J.; Xiong, H.; Bernhard, M.; Pan, S.; Dariu, M.G.; Li, K. A unified framework for concurrent pedestrian and cyclist detection. IEEE Trans. Intell. Transp. Syst. 2016, 18, 269–281. [Google Scholar] [CrossRef]
  15. Tian, W.; Deng, Z.; Yin, D.; Zheng, Z.; Huang, Y.; Bi, X. 3D Pedestrian Detection in Farmland by Monocular RGB Image and Far-Infrared Sensing. Remote Sens. 2021, 13, 2896. [Google Scholar] [CrossRef]
  16. Li, C.; Song, D.; Tong, R.; Tang, M. Illumination-aware faster R-CNN for robust multispectral pedestrian detection. Pattern Recognit. 2019, 85, 161–171. [Google Scholar] [CrossRef] [Green Version]
  17. Wagner, J.; Fischer, V.; Herman, M.; Behnke, S. Multispectral Pedestrian Detection using Deep Fusion Convolutional Neural Networks. ESANN 2016, 587, 509–514. [Google Scholar]
  18. Zhang, L.; Liu, Z.; Zhang, S.; Yang, X.; Qiao, H.; Huang, K.; Hussain, A. Cross-modality interactive attention network for multispectral pedestrian detection. Inf. Fusion 2019, 50, 20–29. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Yin, Z.; Nie, L.; Huang, S. Attention Based Multi-Layer Fusion of Multispectral Images for Pedestrian Detection. IEEE Access 2020, 8, 165071–165084. [Google Scholar] [CrossRef]
  20. Fu, L.; Gu, W.; Ai, Y.; Li, W.; Wang, D. Adaptive spatial pixel-level feature fusion network for multispectral pedestrian detection. Infrared Phys. Technol. 2021, 116, 103770–103781. [Google Scholar] [CrossRef]
  21. Cao, Z.; Yang, H.; Zhao, J.; Guo, S.; Li, L. Attention Fusion for One-Stage Multispectral Pedestrian Detection. Sensors 2021, 21, 4184. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, L.; Liu, Z.; Chen, X.; Yang, X. The cross-modality disparity problem in multispectral pedestrian detection. arXiv 2019, arXiv:1901.02645. [Google Scholar]
  23. Zhang, H.; Fromont, E.; Lefevre, S.; Avignon, B. Guided attentive feature fusion for multispectral pedestrian detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 72–80. [Google Scholar]
  24. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  25. Zhu, X.; Cheng, D.; Zhang, Z.; Lin, S.; Dai, J. An empirical study of spatial attention mechanisms in deep networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6688–6697. [Google Scholar]
  26. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  27. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; So Kweon, I. Multispectral Pedestrian Detection: Benchmark Dataset and Baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar]
  28. Sun, Y.; Cao, B.; Zhu, P.; Hu, Q. Drone-based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning. arXiv 2021, arXiv:2003.02437v2. [Google Scholar] [CrossRef]
  29. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  30. Zhang, J.; Fan, D.P.; Dai, Y.; Yu, X.; Zhong, Y.; Barnes, N.; Shao, L. RGB-D Saliency Detection via Cascaded Mutual Information Minimization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 4338–4347. [Google Scholar]
  31. Zhou, K.; Chen, L.; Cao, X. Improving multispectral pedestrian detection by addressing modality imbalance problems. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 787–803. [Google Scholar]
  32. Li, C.; Song, D.; Tong, R.; Tang, M. Multispectral pedestrian detection via simultaneous detection and segmentation. arXiv 2018, arXiv:1808.04818. [Google Scholar]
  33. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  34. Zhang, L.; Zhu, X.; Chen, X.; Yang, X.; Lei, Z.; Liu, Z. Weakly aligned cross-modal learning for multispectral pedestrian detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019; pp. 5127–5137. [Google Scholar]
Figure 1. Overview framework of the proposed redundant information suppression network (RISNet). The RISNet consists of four parts: feature extraction stage, mutual information measurement module, feature fusion module and detection stage.
Figure 1. Overview framework of the proposed redundant information suppression network (RISNet). The RISNet consists of four parts: feature extraction stage, mutual information measurement module, feature fusion module and detection stage.
Remotesensing 14 02020 g001
Figure 2. Mutual information module.
Figure 2. Mutual information module.
Remotesensing 14 02020 g002
Figure 3. The first row is the RGB images of different lighting scenes, from (ad) are daytime, weak light at night, strong light at night and dark night, and the second row is the corresponding histogram.
Figure 3. The first row is the RGB images of different lighting scenes, from (ad) are daytime, weak light at night, strong light at night and dark night, and the second row is the corresponding histogram.
Remotesensing 14 02020 g003
Figure 4. Sample visualization. (a) some images of DroneVehicle remote sensing dataset; (b) some images of KAIST pedestrian dataset. The first row is the RGB image, the second row is the corresponding infrared image.
Figure 4. Sample visualization. (a) some images of DroneVehicle remote sensing dataset; (b) some images of KAIST pedestrian dataset. The first row is the RGB image, the second row is the corresponding infrared image.
Remotesensing 14 02020 g004
Figure 5. Visualization of baseline (a) and proposed RISNet (b) detection results on DroneVehicle remote sensing dataset. The first row shows the detection results in the daytime scenarios. Targets circled in red represent targets not detected by baseline but detected by RISNet.
Figure 5. Visualization of baseline (a) and proposed RISNet (b) detection results on DroneVehicle remote sensing dataset. The first row shows the detection results in the daytime scenarios. Targets circled in red represent targets not detected by baseline but detected by RISNet.
Remotesensing 14 02020 g005
Figure 6. Visualization of MBNet (a) and the proposed RISNet (b) detection results on the KAIST dataset. The first row shows the detection results in the night scenarios. The second row shows the detection results in the daytime scenarios. The third row shows the detection results in the near scenarios. The fourth row shows the detection results in the far scenarios. Targets circled in red represent targets not detected by MBNet but detected by RISNet.
Figure 6. Visualization of MBNet (a) and the proposed RISNet (b) detection results on the KAIST dataset. The first row shows the detection results in the night scenarios. The second row shows the detection results in the daytime scenarios. The third row shows the detection results in the near scenarios. The fourth row shows the detection results in the far scenarios. Targets circled in red represent targets not detected by MBNet but detected by RISNet.
Remotesensing 14 02020 g006
Table 1. Comparison results of different algorithms, the best AP indicators are generally black.
Table 1. Comparison results of different algorithms, the best AP indicators are generally black.
MethodsModalityAP
Yolov3 [33]R56.40
Faster-RCNN [2]R44.59
Yolov3 [33]I59.32
Faster-RCNN [2]I54.55
CMDet [28]R + I62.58
UA-CMDet [28]R + I64.01
MBNet [31]R + I62.83
RISNet(ours)R + I66.40
Table 2. Evaluations on the KAIST dataset under all nine test subsets (MR−2). The best MR−2 indicators are generally black.
Table 2. Evaluations on the KAIST dataset under all nine test subsets (MR−2). The best MR−2 indicators are generally black.
MethodsModalityAllDayNightNearMediumFarNonePartialHeavy
Yolov3 [33]R52.8543.2572.8833.9461.2891.4368.5581.5887.39
Yolov3 [33]I32.6737.6223.386.2943.5978.0452.1562.9178.72
AR-CNN [34]R + I9.349.948.380.0016.0869.0031.4038.6355.73
CIAN [18]R + I14.1214.7711.133.7119.0455.8230.3141.5762.48
MSDS-RCNN [32]R + I11.6310.6013.731.2916.1963.7329.8638.7163.37
MBNet [31]R + I8.138.287.860.0016.0755.9927.7436.4359.14
RISNet (ours)R + I7.897.617.080.0014.0152.6725.2334.25 56.14
Table 3. FPS of different algorithms tested on different platforms. Ranking of computing performance (TFLOPs) of three computing platforms: GTX 1080 Ti (11.3 TFLOPs) > ITAN X (11 TFLOPs) > GTX 2060 (6.45 TFLOPs).
Table 3. FPS of different algorithms tested on different platforms. Ranking of computing performance (TFLOPs) of three computing platforms: GTX 1080 Ti (11.3 TFLOPs) > ITAN X (11 TFLOPs) > GTX 2060 (6.45 TFLOPs).
MethodsPlatformFPS
Yolov3 [33]GTX 206063
Yolov3 [33]GTX 206063
AR-CNN [34]GTX 1080 Ti8
CIAN [18]GTX 1080 Ti14
MSDS-RCNN [32]TITAN X5
MBNet [31]GTX 1080 Ti14
RISNet (ours)GTX 206010
Table 4. Ablation experiments on the DroneVehicle remote sensing dataset under different lighting conditions, where the baseline is MBNet [31], MI represents the mutual information module, and NLSC represents the new lighting condition classification method. The best AP indicators are generally black.
Table 4. Ablation experiments on the DroneVehicle remote sensing dataset under different lighting conditions, where the baseline is MBNet [31], MI represents the mutual information module, and NLSC represents the new lighting condition classification method. The best AP indicators are generally black.
MethodsDaytimeStrong Light at NightWeak Light at NightDark Night
baseline70.4670.6052.4143.98
baseline + MI71.4472.3955.0145.34
baseline + NLSC70.8171.0554.6745.00
RISNet (ours)72.0471.9154.9446.20
Table 5. Ablation experiments on the KAIST dataset under all nine test subsets, where the baseline is MBNet [31], MI represents the mutual information module, and NLSC represents the new lighting condition classification method. The best MR−2 indicators are generally black.
Table 5. Ablation experiments on the KAIST dataset under all nine test subsets, where the baseline is MBNet [31], MI represents the mutual information module, and NLSC represents the new lighting condition classification method. The best MR−2 indicators are generally black.
MethodsAllDayNightNearMediumFarNonePartialHeavy
baseline8.138.287.860.0016.0755.9927.7436.4359.14
baseline + MI9.029.907.280.0014.8253.1225.8935.0760.93
baseline + NLSC8.979.178.230.0016.8355.2027.1436.9761.57
RISNet (ours)7.897.617.080.0014.0152.6725.2334.2556.14
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Q.; Chi, Y.; Shen, T.; Song, J.; Zhang, Z.; Zhu, Y. Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy. Remote Sens. 2022, 14, 2020. https://doi.org/10.3390/rs14092020

AMA Style

Wang Q, Chi Y, Shen T, Song J, Zhang Z, Zhu Y. Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy. Remote Sensing. 2022; 14(9):2020. https://doi.org/10.3390/rs14092020

Chicago/Turabian Style

Wang, Qingwang, Yongke Chi, Tao Shen, Jian Song, Zifeng Zhang, and Yan Zhu. 2022. "Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy" Remote Sensing 14, no. 9: 2020. https://doi.org/10.3390/rs14092020

APA Style

Wang, Q., Chi, Y., Shen, T., Song, J., Zhang, Z., & Zhu, Y. (2022). Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy. Remote Sensing, 14(9), 2020. https://doi.org/10.3390/rs14092020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop