Next Article in Journal
A LAPS-Based Differential Sensor for Parallelized Metabolism Monitoring of Various Bacteria
Previous Article in Journal
A Survey of Collaborative UAV–WSN Systems for Efficient Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-NMS: A Method for Autonomously Removing False Detection Boxes from Aerial Image Object Detection Results

1
Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
2
Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110169, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang 110016, China
5
The Key Lab of Image Understanding and Computer Vision, Shenyang 110016, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(21), 4691; https://doi.org/10.3390/s19214691
Submission received: 24 September 2019 / Revised: 17 October 2019 / Accepted: 25 October 2019 / Published: 28 October 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In the field of aerial image object detection based on deep learning, it’s difficult to extract features because the images are obtained from a top-down perspective. Therefore, there are numerous false detection boxes. The existing post-processing methods mainly remove overlapped detection boxes, but it’s hard to eliminate false detection boxes. The proposed dual non-maximum suppression (dual-NMS) combines the density of detection boxes that are generated for each detected object with the corresponding classification confidence to autonomously remove the false detection boxes. With the dual-NMS as a post-processing method, the precision is greatly improved under the premise of keeping recall unchanged. In vehicle detection in aerial imagery (VEDAI) and dataset for object detection in aerial images (DOTA) datasets, the removal rate of false detection boxes is over 50%. Additionally, according to the characteristics of aerial images, the correlation calculation layer for feature channel separation and the dilated convolution guidance structure are proposed to enhance the feature extraction ability of the network, and these structures constitute the correlation network (CorrNet). Compared with you only look once (YOLOv3), the mean average precision (mAP) of the CorrNet for DOTA increased by 9.78%. Commingled with dual-NMS, the detection effect in aerial images is significantly improved.

1. Introduction

Object detection is a fundamental task in the area of computer vision. Its purpose is to identify the categories of objects in an image and obtain the locations of each object. In recent years, with the rapid development of deep learning, various types of high-performance image segmentation and object detection networks have been proposed. The commonality of the abovementioned networks is that they all employ the powerful feature extraction capabilities of convolutional neural networks. The regions with convolutional neural network (R-CNN) was a pioneering work on applying neural networks in the field of object detection, R-CNN’s phased design idea essentially reflected its utilization of the convolutional neural network’s powerful feature extraction ability [1]. The network structure can be simplified by integrating different parts of the R-CNN [2,3]. The reason for simplifying the network structure is that the convolutional neural network can autonomously extract features, and also has strong fitting ability. Neural networks can map highly abstract features to object categories, locations and even depth information [4,5]. The one-stage object detection networks that were proposed in [6,7] validate the above viewpoint. The anchor-free network structure that was proposed in [8,9,10] makes full use of the neural network’s fitting ability.
Although the convolutional neural network already has strong feature extraction capabilities, it is difficult to achieve a further improvement in its detection precision, especially in specific scenes such as aerial images. Because objects are relatively small in aerial images, their features cannot be fully extracted, which makes it difficult for the classical object detection network to effectively distinguish the objects to be detected from the background. In addition, a large number of detection boxes are incorrect. For the above problems, two methods can be used to improve the detection precision of the object detection network in specific scenarios. (1) The network’s ability to extract object features can be enhanced by optimizing the network structure. It fundamentally improves the performance of the object detection network in specific scenarios. This correspond to the feature extraction process on the upper side of Figure 1. (2) The decoding process from the feature maps to the detection results and the post-processing method could be optimized. This corresponds to the post-processing process in the blue box at the bottom of Figure 1. The existing algorithms mainly optimize these two stages in the object detection process. The ability of the network to extract object features is enhanced by fusing different levels of feature maps [11,12,13,14]. By deepening and widening the network, more abstract features can be obtained to improve the classification accuracy [15,16,17]. References [18,19,20] mainly improve the accuracy by optimizing the decoding process of the network. References [21,22,23,24] improved the NMS to adapt to more complex detection scenarios. In the small object detection of aerial images, the features of the objects that are to be identified are difficult to extract, and objects and backgrounds are indistinguishable; therefore, there are many false detection boxes in the detection results, which directly affect the precision of the detection network. The existing methods do not optimize the detection network from the perspective of removing false detection boxes. If the false detection boxes can be accurately deleted, the detection precision will naturally be improved.
To solve the problems that were mentioned above, the dual-NMS is proposed in this paper, which corresponds to the blue dotted box in Figure 1. Considering the density of the detection boxes for each detected object and the classification confidence of the corresponding detection boxes, the false detection boxes are adaptively removed, which greatly improves the detection precision without changing the detection network structure. In addition, the detection network is optimized to fundamentally enhance its ability to extract object features. The detection network: you only look once (YOLOv3) accomplishes object classification and location simultaneously in the process of convolution [25]. Unlike two-stage detection network, YOLOv3 completes object detection in forward convolution process, so it has faster detection speed. Compared with other one-stage object detection networks, YOLOv3 has a better detection effect. Therefore, the research work of this paper is based on YOLOv3. The dilated convolution guidance structure is added to the initial part of YOLOv3, which allows the network to better extract the overall features of the objects. The loss of small objects in the feature maps is avoided by reducing the downsampling times of the detection network. References [26,27,28] calculate the correlation of the image contents, which improves the segmentation accuracy of the network. However, they use the entire feature tensor to calculate the correlation. These methods recombine the separated foreground and background features and reduce the ability of the network to distinguish individual objects. In this paper, the correlation calculation layer is introduced in the detection network, and the feature channels are separated from each other. By calculating the correlations between similar objects, the correlation calculation layer allows the network to more easily to detect small objects in aerial images. These optimization methods that were mentioned above make up the CorrNet. The process of extracting the features of the input image by using the CorrNet and removing the false and overlapped detection boxes by using the dual-NMS are shown in Figure 1. The structure that is shown in Figure 1 has achieved better results for the VEDAI, RSOD [29], UCAS-AOD and DOTA [30] aerial image object detection datasets. Our main contributions can be summarized as follows:
1)
The dual-NMS is proposed. As an optimization algorithm for detection results, the precision of the object detection results is greatly improved under the premise of keeping the recall unchanged.
2)
We proposed the CorrNet. The feature extraction capability of the convolutional network is enhanced by introducing the correlation calculation layer of the feature channel separation and the dilated convolution guidance structure. This enhancement makes the CorrNet more suitable for the detection of small objects in aerial images.
3)
Experiments were carried out by using several aerial image object detection datasets to verify the effects of the dual-NMS and CorrNet, and the results were compared with those of YOLOv3. The experimental results show that the CorrNet has better detection effects for the small objects in aerial images.
The remainder of this paper is organized as follows: Section 2 briefly reviews the related works, and Section 3 details the dual-NMS and the optimized detection network: CorrNet. The experimental results and methods are given in Section 4, which is followed by the conclusions in Section 5.

2. Related Works

In the existing object detection networks, after the convolution and feature extraction of the input images, the categories and positions of each object in the images are obtained by decoding the feature maps, and a series of detection boxes are generated in the meantime. In the Faster-RCNN [3], YOLOv3 [25], SSD [7] and some other detection networks, the detection boxes with higher classification confidence by default have higher positioning accuracies. This hypothesis is discussed in [18], the classification confidence and positioning accuracy of the detection boxes are not strictly linear, and the positioning score of the detection boxes is taken as the optimization condition. When the detected objects are dense, removing false detection boxes by using the NMS will lead to missed detections. By reducing the classification confidence of the highly overlapped detection boxes, missed detections can be avoided to a certain extent [21]. Reference [22] used the Kullback–Leibler divergence loss to measure the distribution loss between the prediction boxes and the ground truth, which improves the localization accuracy of the detection boxes. Reference [23] increased the flexibility of the NMS by training a network to set different intersection-over-union (IoU) thresholds for detection boxes with different IoUs. The NMS is realized through the trained network, which makes it more suitable for dense scenes with high occlusion [24]. The NMS is optimized for text detection in natural scenes [31,32,33]. The abovementioned post-processing methods optimized the detection results by removing the overlapped detection boxes. Similarly, the detection results can be optimized by removing false detection boxes.
In addition to optimizing the post-processing of detection results, improving the feature extraction capability of the convolutional network is the fundamental solution to improving object detection. There are many optimization methods for the convolution process. Under the same number of parameters and computational complexity, better overall features can be obtained by using the dilated convolution [34,35]. References [36,37,38,39] apply different weights to the feature map through an attention mechanism so that the convolution networks could autonomously focus on the key objects in images. In aerial image object detection, because the objects are small, it is difficult to extract the features during object detection. A continuous convolution and pooling process will lead to the disappearance of small objects in high-level feature maps. However, the dilated convolution can be used to extract features in low-level convolution layers to obtain the overall features of a single object. To avoid the loss of small objects in the feature map due to excessive downsampling, the downsampling times need to be adjusted. The correlations of the same contents in an image can be calculated, which improves the segmentation accuracy of the network [26,27,28]. Similarly, the correlation at different positions in a single feature map can be calculated, and similar objects can be integrated through the correlation. When combining with the dual-NMS, the false detection boxes are removed, and the detection efficiency of small objects in aerial images is improved by the object detection network.

3. Our Methods

The main idea of our method is to autonomously remove false detection boxes from the detection results according to the density of the detected boxes and optimize the object detection network according to the characteristics of aerial images. The corresponding methods are the dual-NMS and the CorrNet respectively. In this section, the proposed dual-NMS algorithm and the CorrNet are analyzed in detail.

3.1. Dual-NMS

3.1.1. The False Detection Boxes

In aerial images, the images are obtained from the top-down perspective, there are fewer features that can be used for object recognition, and the directions of the objects are random; therefore, false detection boxes are more likely to appear. The false detection boxes mainly include two types as shown in Figure 2a: (1) the background is misidentified as an object, such as the Tractor (red box); and (2) the real objects are misclassified, such as the Boat (blue box).

3.1.2. The Differences between True and False Detection Boxes

To autonomously remove the false detection boxes, the density of the detection boxes is defined as the condition that judges whether an object is falsely detected. The density of the detection boxes can be explained in detail as follows. As shown in Figure 2a, there are dense boxes around each aircraft (yellow boxes). In other words, if the object is correctly detected, more detection boxes will be generated around an object. If the detected object is a false detection, the detection boxes that are generated around the detected object are sparse, and the classification confidence of the detection boxes will be smaller, as shown in Figure 2a with the Tractor (red box) and Boat (blue box). These sparse boxes are likely to be false detection unless they have high classification confidences. Therefore, the density of the detection boxes is defined as the number of detection boxes around each detected object. The average densities of the detection boxes that are generated around the true and false detected objects in the output results of the detection network are shown in Table 1. By comparison, the average density of true detection boxes is approximately 3 times that of false detection boxes. Therefore, the density of the detection boxes can be used as a condition to suppress the false detection boxes before the NMS is executed.

3.1.3. Removing the False Detection Boxes by Using the Dual-NMS

The false detection boxes are difficult to effectively eliminate by using post-processing algorithms such as the NMS, and the method still leaves the highest classification confidence false detection boxes, such as the Tractor (red box) and the Boat (blue box) in Figure 2c. If these false detection boxes can be autonomously removed, the precision of the detection results will be greatly improved.
The proposed dual-NMS takes the density of the detection boxes as the condition to remove some detection boxes. Especially in aerial image object detection, it can adaptively identify and remove the false detection boxes. The dual-NMS is divided into two parts: (1) judging whether the current boxes are detected by errors according to the density of the detection boxes and, if so, deleting them; and (2) removing the highly overlapped detection boxes. The results of the dual-NMS adaptively identifying and removing the false detection boxes is shown in Figure 2b, where the false detection boxes have been removed, such as the Tractor (red box) and the Boat (blue box). The dual-NMS is shown in Algorithm 1.
Algorithm 1: Dual-NMS
Input: B = {b1, …, bN}, S = {s1, …, sN}, Nt1, Nt2, α, β, γ
B is a set of predicted bounding boxes for class A
S is a set of corresponding classification confidences
Nt1 and Nt2 are the NMS thresholds
α, β and γ are the parameters in function f that are used to suppress false predicted bounding boxes
Output: D the final predicted bounding boxes
1: D = ∅
2: while B ≠ ∅ do
3:     bt ← the first element in B; T = ∅; Sp = ∅
4:     TT ∪ {bt}
5:     SpSp ∪ {st}
6:     BB \ {bt}
7:     SS \ {st}
8:     for biB do
9:      if IoU (bt, bi) ≥ Nt1 then
10:       TT ∪ {bi}
11:       SpSp ∪ {si}
12:       BB \ {bi}
13:       SS \ {si}
14:      end if
15:     end for
16:     Ssumsum (Sp)
17:     dlen(T)
18:     s_margmax (Sp)
19:     if Ssum > f (d, s_m, α, β, γ) then
20:      while T ≠ ∅ do
21:       smargmax (Sp)
22:       TT \ {tm}
23:       SpSp \ {sm}
24:       DD ∪ {tm}
25:       for tiT do
26:        if IoU (tm, ti) ≥ Nt2 then
27:         TT \ {ti}
28:         SpSp \ {Si}
29:        end if
30:       end for
31:      end while
32:     end if
33: end while
34: return D
In Algorithm 1, the classical NMS corresponds to lines 20~34, which remove the highly overlapped detection boxes. The dual-NMS corresponds to lines 2~19, which determine whether the detection boxes are false detections according to the density of the detection boxes. The implementation can be divided into three steps:
1)
All the detection boxes are grouped according to the IoU between them (lines 3~15). The detection boxes with a certain degree of overlap are divided into a group.
2)
The density of the detection boxes in each group is counted. The sum and the maximum of the classification confidences in each group of the detection boxes are calculated.
3)
The dynamic threshold that is calculated by function f and the above statistics are used to determine whether the detection boxes in this group are false or not. If they are false detection boxes, delete them.
The dynamic threshold calculation function f has the following form:
f   ( d ,   s _ m ,   α ,   β , γ ) = β × e s _ m × d 1 α × s _ m γ × d ,
where d is the density of the detection boxes of the current group, s _ m is the maximum value of the classification confidence in the current group, and α ,   β and γ are the corresponding scaling factors. If the density of the detection boxes is larger, the possibility of real objects appearing at this position is greatly increased. When there are detection boxes with high classification confidence in the current group, even if the density of the detection boxes is lower, the possibility of detecting real objects is also greatly increased. In these cases, dynamic thresholds should be rapidly lowered to avoid blindly deleting true detection boxes. The low density of the detection boxes indicates that the possibility of a false detection is larger, and the dynamic threshold should be kept at a larger value to autonomously remove the false detection boxes. Based on the above analysis,   e s _ m ensures that the threshold decreases as the maximum classification confidence increases. d 1 α × s _ m γ × d guarantees that the threshold decreases as the density of the detection boxes increases, and it also controls the action intervals of different thresholds. The distributions of f with respect to the density of the detection boxes and the maximum value of the classification confidence are shown in Figure 3. The dynamic threshold decreases as the density of the detection boxes and the maximum classification confidence increase. The sum of the classification confidence of all detection boxes reflects the density of the detection boxes and the reliability of a correct detection. By comparing the dynamic threshold with the sum of the classification confidences of all the detection boxes, the process of judging whether the current group of boxes is a false detection is realized according to the density of the detection boxes. After removing highly overlapped detection boxes using the classical NMS, the whole dual-NMS process is completed. The above procedure can adaptively identify and remove the false detections and avoid incorrect removals of true detection boxes. The corresponding experimental validation is described in detail in Section 4.2. The values of α ,   β and   γ are experimentally determined in Section 4.3. The classical NMS algorithm can be replaced with the soft-NMS in the subsequent part of the dual-NMS to make the algorithm more adaptable.

3.2. CorrNet

3.2.1. The Characteristics of Aerial Images

Most of the existing object detection networks are trained on datasets such as VOC and COCO. The structure of the detection network should match the characteristics of the datasets. Compared with VOC and other datasets, the scale of the objects in aerial images is much smaller. In the VEDAI, RSOD, UCAS-AOD and DOTA datasets, the proportion of objects with different scales to all targets is shown in Table 2. The scales can be divided into four levels: S1 ≤ Min (H, W)/32, S2 ≤ Min (H, W)/16, S3 ≤ Min (H, W)/8 and S4 ≤ Min (H, W)/4. H and W are the length and width of the input image, respectively. The scale (S) is the larger value of the object’s length or width. In the datasets, more than 90% of the objects have scales below 1/8 of the input image’s length and width. In addition, in VEDAI, 67.85% of the object scales are less than 1/32 of the input image’s length and width. The statistical results show that small objects account for the vast majority in the four datasets. It makes detection network difficult for the existing object detection models to achieve the ideal detection results in aerial image datasets.
In the object detection network, the input image is continuously convoluted and subsampled to get the feature map of the input image. Then, the feature maps are decoded to get the categories and the box coordinates of the objects in the image [1,2,3,6,7]. The max-pooling operation reduces the resolution of the image and increases the receptive field of the convolution kernel. Continuous max-pooling results in a rapid decrease of the image resolution, which makes small-scale objects rapidly disappear in feature maps, and leads to the lack of small object information in highly abstract feature maps. For the above reasons, the key to improve the detection network’s ability to detect small objects in aerial images is to avoid losing small objects in the convolution and pooling processes, and to extract the overall and highly abstract features of the small objects at the bottom of the detection network.

3.2.2. The Optimization Methods

By improving the resolution of the feature maps that are used to calculate the losses, the problem of small objects disappearing in these feature maps can be solved. The resolutions of the feature maps that are used to calculate the losses in the CorrNet are twice those of YOLOv3. Therefore, the downsampling times in the CorrNet are half those of YOLOv3. The overall structure of the detection network is shown in Figure 4. In the CorrNet, feature maps with lower resolutions are upsampled and fused with high-resolution feature maps to compensate for the lack of high abstract features. Through the above improvements, the disappearance of small objects in high-level feature maps is avoided. However, the receptive field of the convolution kernel in the high-level convolutional layers is reduced, and the detection network’s ability to extract the overall features of objects is weakened, which indirectly affects the detection effect of the network.
To enhance the ability of the convolution network to extract the overall features of the object in the image, in the CorrNet, the input image first passes through the dilated convolution guidance layer, which is the red dashed box in Figure 4. The implementation is shown in Figure 5. The input image is subjected to the conventional convolution and two parallel dilated convolutions, and the results of the three convolution operations are concatenated for the subsequent feature extraction. The receptive field of the convolution kernel is increased by the dilated convolution guidance structure to enhance the network’s ability to extract the overall features of the objects. These two methods greatly improve the detection network’s ability to extract image features, thereby making the detection network more suitable for small object detection in aerial images. The experimental effects and parameters are detailed in Section 4.4.
In aerial images, only the top of the object can be seen; therefore, the foregrounds and the backgrounds are indistinguishable. In this case, the fine-grained features of the object and the correlation between objects of the same category are the key to detecting objects. Hence, the correlation calculation layer in which the feature channels are separated from each other is added to the CorrNet. The correlation calculation layer is represented by the green box in Figure 4, and its implementation is shown in Figure 6. The different channels of the feature map contain the features of different positions in the image [36], so the internal relevance of each feature map is calculated in the correlation calculation layer. To reduce the computational complexity, the correlation calculation layer is added to the feature maps with the lowest resolution. For the feature map f with C input channels, first, the number of channels is reduced to 1/8, and the feature map f is obtained. Then, the R, S and T operations are performed on f in turn. R, S and T correspond to the order modification, mean normalization, and transposition operations of the matrix, respectively:
R S ( f ) =   R ( f f ¯ ) w × h = ( f f ¯ ) 1 × w h ,
R S T ( f ) =   R S ( f ) T = ( f f ¯ ) 1 × w h T ,
R ( f ) =   R ( f ) w × h = f 1 × w h ,
After the above transformation, the weight of each feature map is obtained by the following formula:
W w h × w h = s o f t m a x ( R S T ( f ) × R S ( f ) ) w h × w h   ,
Then, the corresponding weight is applied to each feature map to obtain the weighted feature map f :
f = R ( R ( f ) × W w h × w h ) w × h   ,
Finally, the number of f channels is expanded to C and connected to the original feature map f as the final feature output. The above process calculates the correlation of similar image contents in a single feature map, and it generates corresponding weights for each channel, which improves the ability of the network to detect small objects in aerial images. Section 4.4 will give the experimental results of network structure optimization methods and compare the detection performance of the CorrNet with other detection networks.

4. Experiments and Results

To verify the effectiveness of the dual-NMS and the network optimization methods, experiments were carried out on four aerial image object detection datasets, including VEDAI, UCAS-AOD, RSOD and DOTA. The detection results of the original YOLOv3 were used as the baseline, and several evaluation metrics were analyzed. All the experiments were accomplished on a computer with a GTX 1080Ti GPU and a 4.0 GHz CPU. All programs are performed on Ubuntu 14.04 via Tensorflow1.0 and Python3.5. The initial learning rate is set to 0.0001. The optimizer of the network is Adaptive Moment Estimation (Adam), which can automatically adjust the learning rate for each parameter to be learned. Because of the limitation of memory, the batch size is set to 2. When the network achieves the optimal detection effect, the number of epochs is 70.
In the four aerial image datasets, the images in VEDAI are the clearest, with a total of 9 categories to be detected. There are 1087 images as training set and 123 images as test set. On UCAS-AOD and RSOD, the number of images in the training set is 1354 and 841, and the number of images in the test set is 156 and 95. The two datasets contain two and four categories of objects to be detected, respectively. The DOTA dataset contains the largest number of object categories, totaling 15 categories. The size of the image in the DOTA dataset is different, and the high-resolution images make the network difficult to train. We cropped the image to 1024 × 1024 and remove the images containing incomplete targets. Finally, in the DOTA dataset, the training set contains 2287 pictures and the test set contains 254 pictures. In the training process of the detection network, the resolution of all the input images was resized to 512 × 512. Different datasets have different training time. The training time in DOTA datasets is 9 h.

4.1. The Metrics that Were Used in the Experiments

Appropriate metrics can reflect the pros and cons of the algorithm. In this paper, the precision, recall, F1 score, removal rate ( R r ), frames per second (FPS) and the mean Average Precision (mAP) are selected as the metrics to evaluate the effectiveness of our methods. The mAP is the average precision at different recalls (0.0~1.0) in all categories. The FPS is the number of frames that the program can process per second. It is an important standard to measure the efficiency of the algorithm.
The precision and recall are defined as follows:
P r e c i s i o n =   T P T P   +   F P ,
R e c a l l   =   T P T P   +   F N ,
where T P : the number of true objects that are correctly detected in the detection results, F P : the number of objects that are incorrectly identified as objects and falsely classified in the detection results, and F N : the number of objects that failed to be correctly detected.
The F1 score combines the precision and recall, and it is defined as follows:
F 1 = ( 1 + β 2 ) · p r e c i s i o n   ·   r e c a l l β 2   ·   p r e c i s i o n   +   r e c a l l   ,   β =   1 ,
Except for the precision, the recall and F1 score are used to evaluate the dual-NMS, and the effect of the dual-NMS on the true and false detection boxes was measured by the R r of the detection boxes in the experiment. The removal rate is defined as follows:
R r =   1 C i = 1 C N i R i N i ,
where C is the number of object categories in the dataset, and N i and R i are the numbers of detection boxes of class i in the results that are obtained by the classical NMS and the dual-NMS, respectively.

4.2. The Effectiveness of the Dual-NMS

As the post-processing algorithm of detection results, the dual-NMS can adaptively remove the false detection boxes, while avoiding any impact on true detection. The precision, recall, and F1 score take into account the number of true and false detection boxes in the detection results. Because the dual-NMS can autonomously remove the false detection boxes, consequently, the F P is reduced. In addition, attention should be paid to the impact on the correct detection objects, that is to say, the recall should remain unchanged. The precision and recall of the dual-NMS and classical NMS in the VEDAI are shown in Figure 7 and Figure 8, respectively. It can be seen that the precision of the detection results is greatly improved by the dual-NMS. Only the recall of CampingCar, Car, and Truck are slightly lower.
The F1 score can more objectively and comprehensively reflect the quality of the detection results. The F1 score of the dual-NMS and classic NMS in different datasets are compared as shown in Table 3. The detection results that are optimized by the dual-NMS are significantly better. The experimental results show that the dual-NMS can adaptively remove the false detection boxes and minimize the impact on the correct results, which proves the effectiveness of the dual-NMS.
The R r synthetically reflects the ability of the dual-NMS to remove the detection boxes. The max/min removal rates of the true and false detection boxes in each dataset of YOLOv3 and the CorrNet are shown in Table 4. In the two detection networks, although there are errors in removing the true detection boxes, the effect of removing false detection boxes is more significant. The removal rate of false detection boxes (RrF) in each dataset is obviously higher than that of true detection boxes (RrT). It is still possible to remove more than 10% of the false detection boxes, with a minimal impact on the correct detections. Therefore, the dual-NMS can be used as a general-purpose post-processing algorithm in existing detection networks.

4.3. Ablation Experiment

In the dual-NMS, different control parameters affect the dynamic threshold. The control parameters indirectly determine the ability of the dual-NMS to suppress false detection boxes and impact true detections. A good control factor should allow the dual-NMS to remove the false detection boxes to the greatest extent, while avoiding impacting true detections. In this section, VEDAI is selected as the test dataset. To reduce the amount of experimental calculations, the correlation calculation layer is not added to the model. The changing trend of the true and false detection box removal rates with β, α and γ are shown in Figure 9. The removal rate increases as β increases and decreases as α and γ increase. The overall trend of the removal rates accords with the distribution form of the dynamic threshold. The changing trends of the mAP and F1 score under different α and γ values are shown in Figure 10. The removal rate of the false detection boxes is the most obvious when γ is less than 0.3; therefore, the value of γ in Figure 10 is set from 0.1 to 0.3. The overall value of the dynamic threshold is affected by β. According to the removal rate of false detection boxes, β is directly set as 1.1. In Figure 10, the mAP increases α increases, while the F1 score shows a downward trend. It indicates that the removal rate of the false detection boxes is larger when α is small, and the dual-NMS has a better inhibitory effect on false detection boxes, but excessive suppression of the true detections leads to a decreased mAP. Finally, when the control coefficients are selected as α = 2.6 and γ = 0.1, the negative effect of the dual-NMS on true detections is the smallest and the false detection boxes are eliminated to some extent.

4.4. Analysis of the Effectiveness of the Network Optimization Methods

The role of each optimization method in different aerial image object detection datasets is shown in Table 5. The experimental results are compared with the results of YOLOv3. The resolutions of the feature maps that used to calculate the loss in YOLOv3 are 16 × 16, 32 × 32 and 64 × 64. Conversely the resolutions are set as 32 × 32, 64 × 64 and 128 × 128 in the CorrNet, and the downsample rates are changed from 8, 16, and 32 to 4, 8, and 16. If the down-sample rate is set 32, the objects with a scale less than 32 pixels will disappear in the feature map that with the lowest resolution. It can be seen from the S2 in Table 2, if the resolution of the input image is 512 × 512, the length and the width of most objects are less than 32 pixels. In this case, it is difficult to effectively detect the objects on the lowest resolution feature map. Therefore, the downsample rate in the CorrNet is changed to half that of YOLOv3. By this method, the disappearing of small objects in the feature maps is avoided effectively, and the computational complexity does not increase significantly. The effect of reducing downsample rate is most obvious in the DOTA datasets, the mAP is increased by 7.4%.
The addition of the dilated convolution guidance structure solves the problem that the receptive field of convolution kernel decreases due to the reduction of downsample rate, and enhances the ability of the convolutional network to extract the overall features of objects. When the dilated rate is set 3, the detection network has the best detection effect. In the VEDAI and DOTA datasets, the mAP improves by 4.09% and 8.8%, respectively. The experimental results show that the overall features play important roles in identifying different objects in images.
The correlation calculation layer of the feature channel separation uses the correlation between the objects of the same category as the weights of the corresponding feature layers. Through the correlation between the same category of objects, the same category of objects is fused into one, which reduces the difficulty of small object detection in aerial images. In particular, the VEDAI and DOTA datasets have more object categories to be detected, and the mAP is improved by 5.19% and 9.78%, respectively. The correlation calculation layer in the UCAS-AOD dataset results in an inconspicuous improvement. The reason is that there are only two categories in the dataset, and there is only a single object category in an image, therefore, the effect of the correlation calculation layer of the feature channel separation is limited.
In Table 6, the detection performance of the CorrNet and other detection networks in four aerial image datasets is compared. The CorrNet has better detection effect in VEDAI, UCAS-AOD and RSOD. In the dataset of DOTA, the performance of two-stage detection network is obviously better than that of one-stage. Especially the Faster-RCNN, the mAP of its detection results exceeds more than 10% of the one-stage detection network. The reason for this phenomenon is that there are more categories in the DOTA dataset, and the classification of targets in the two-stage detection network is accomplished by independent branches. But in other datasets, the CorrNet achieves better detection results with more concise network structure.
The dual-NMS is applied in the YOLOv3 and the CorrNet, respectively, and the mAP and FPS of the detection results in each dataset are shown in Table 7. Compared with the NMS, the application of the dual-NMS makes the mAP slightly lower. This is because the dual-NMS will erroneously remove a very small number of true detection boxes when removing false detection boxes. According to the mAP calculation method, the reduction of the recall will have a serious impact on the mAP. Therefore, the max/min removal rates of the dual-NMS for false and true detection boxes are calculated in Table 4. The effect of the dual-NMS on false detection boxes is much greater than that on true detection boxes. This result has great practical significance in applications requiring strict precision. Further research on how to reduce the impact of the dual-NMS on true detections will be the focus of subsequent work. In this paper, the computational efficiency of the dual-NMS and the CorrNet is evaluated by FPS. From Table 7, it can be seen that the dual-NMS has little effect on the computing time of the detection network, which is less than 0.5 FPS. CorrNet’s FPS is slightly lower than YOLOv3, and the loss of computational efficiency is due to the increase in computational complexity caused by the correlation layer. Although FPS decreased slightly, the mAP of the CorrNet increased significantly.

4.5. Comparison of the Detection Results

In the previous experimental part, the effectiveness of the dual-NMS and the optimization methods of the network structure are analyzed through specific experimental data. The results of the dual-NMS and different network structure optimization methods are intuitively presented in Figure 11. The comparison of the 1st and 2nd columns in Figure 11 shows that the detection result of the optimized detection network is obviously superior to that of the original detection network. For example, the pink box (Truck) at the top of the 1st row and the blue box (aircraft) in the 3th row are significantly smaller than the results of YOLOv3. In the 4th row, YOLOv3 did not detect any objects. Comparing the 2nd and 3rd columns, the false detection boxes have been removed by the dual-NMS to some extent, such as the pink box (Truck) in the 1st row, the blue box (car) in the 2nd row, the blue box (aircraft) in the 3rd row, and the blue box (soccer-ball-field) in the 4th row. The comparison results fully indicate that the CorrNet can better detect small objects in aerial images, and the dual-NMS can autonomously identify and remove the false detection boxes.

5. Conclusions

In this paper, the proposed dual-NMS considers the density of the detection boxes and the corresponding classification confidence. The dual-NMS can adaptively identify and remove the false detection boxes. The precision of object detection results is greatly improved under the premise of keeping the recall unchanged. For the VEDAI and DOTA datasets, the removal rate of false detection boxes is more than 50%. The proposed CorrNet combines the dilated convolution guidance structure with the feature channel separation correlation calculation layer, which fundamentally improves the ability of the network to detect small objects in aerial images. In the CorrNet, the mAP is increased by 9.78% for the DOTA dataset. In further research, the recognition results of false detection boxes can be used as feedback information for network training so that the detection network can automatically identify and remove the false detection boxes in the convolution process.

Author Contributions

Z.L. and Q.W. conceived and designed the algorithm; Z.L. performed the experiments, analyzed the data and wrote the main body of the paper. S.F. partially designed the experiments and supervised the work. S.W. gave writing advice. Z.Z. and Y.K. revised the paper.

Acknowledgments

This project is supported by National Natural Science Foundation of China (U1713216).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OA, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  2. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  3. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  4. Gordon, A.; Li, H.; Jonschkowski, R.; Angelova, A. Depth from Videos in the Wild: Unsupervised Monocular Depth Learning from Unknown Cameras. arXiv 2019, arXiv:1904.04998. [Google Scholar]
  5. Li, Y.; Wang, H.; Zhang, Y.; Lu, H. Monocular image depth estimation based on structured depth learning. Robot 2017, 6, 812–819. [Google Scholar]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  7. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  8. Law, H.; Teng, Y.; Russakovsky, O.; Deng, J. CornerNet-Lite: Efficient Keypoint Based Object Detection. arXiv 2019, arXiv:1904.08900. [Google Scholar]
  9. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Object Detection with Keypoint Triplets. arXiv 2019, arXiv:1904.08189. [Google Scholar]
  10. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point Set Representation for Object Detection. arXiv 2019, arXiv:1904.11490. [Google Scholar] [Green Version]
  11. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, Hawaii, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  12. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  13. Tang, T.; Zhou, S.; Deng, Z.; Zou, H.; Lei, L. Vehicle Detection in Aerial Images Based on Region Convolutional Neural Networks and Hard Negative Example Mining. Sensors 2017, 17, 336. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, P.; Ke, Y.; Zhang, Z.; Wang, M.; Li, P.; Zhang, S. Urban Land Use and Land Cover Classification Using Novel Deep Learning Models Based on High Spatial Resolution Satellite Imagery. Sensors 2018, 18, 3717. [Google Scholar] [CrossRef] [PubMed]
  15. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  16. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  18. Jiang, B.; Luo, R.; Mao, J.; Xiao, T.; Jiang, Y. Acquisition of Localization Confidence for Accurate Object Detection. In Proceedings of the Computer Vision—ECCV 2012, Florence, Italy, 7–13 October 2012; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2018; pp. 816–832. [Google Scholar] [Green Version]
  19. He, Y.; Zhu, C.; Wang, J.; Savvides, M.; Zhang, X. Bounding box regression with uncertainty for accurate object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 2888–2897. [Google Scholar]
  20. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 658–666. [Google Scholar]
  21. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS — Improving Object Detection with One Line of Code. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5562–5570. [Google Scholar]
  22. He, Y.; Zhang, X.; Savvides, M.; Kitani, K. Softer-NMS: Rethinking Bounding Box Regression for Accurate Object Detection. arXiv 2018, arXiv:1809.08545. [Google Scholar]
  23. Hosang, J.; Benenson, R.; Schiele, B. A convnet for non-maximum suppression. In Proceedings of the German Conference on Pattern Recognition, Hannover, Germany, 12–15 September 2016; pp. 192–204. [Google Scholar]
  24. Hosang, J.; Benenson, R.; Schiele, B. Learning non-maximum suppression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, 21–26 July 2017; pp. 4507–4515. [Google Scholar]
  25. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  26. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 3146–3154. [Google Scholar]
  27. Mou, L.; Hua, Y.; Zhu, X.X. A Relation-Augmented Fully Convolutional Network for Semantic Segmentation in Aerial Scenes. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 12416–12425. [Google Scholar]
  28. Napoletano, P.; Piccoli, F.; Schettini, R. Anomaly Detection in Nanofibrous Materials by CNN-Based Self-Similarity. Sensors 2018, 18, 209. [Google Scholar] [CrossRef] [PubMed]
  29. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate Object Localization in Remote Sensing Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote. Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  30. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, AL, USA, 18–22 June 2018; pp. 3974–3983. [Google Scholar]
  31. Yuliang, L.; Lianwen, J.; Shuaitao, Z.; Sheng, Z. Detecting Curve Text in the Wild: New Dataset and New Solution. arXiv 2017, arXiv:1712.02170. [Google Scholar]
  32. Dai, Y.; Huang, Z.; Gao, Y.; Xu, Y.; Chen, K.; Guo, J.; Qiu, W. Fused text segmentation networks for multi-oriented scene text detection. In Proceedings of the 2018 24th International Conference on Pattern Recognition, Beijing, China, 20–24 August 2018; pp. 3604–3609. [Google Scholar]
  33. Jiang, Y.; Zhu, X.; Wang, X.; Yang, S.; Li, W.; Wang, H.; Fu, P.; Luo, Z. R2CNN: Rotational Region CNN for Orientation Robust Scene Text Detection. arXiv 2017, arXiv:1706.09579. [Google Scholar]
  34. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  35. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE winter conference on applications of computer vision (WACV), Lake Tahoe, NV/CA, USA, 12–15 Mrach 2018; pp. 1451–1460. [Google Scholar]
  36. Zhang, X.; Wang, T.; Qi, J.; Lu, H.; Wang, G. Progressive Attention Guided Recurrent Network for Salient Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, AL, USA, 18–22 June 2018; pp. 714–722. [Google Scholar]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, AL, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  38. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Genmany, 10–13 September 2018; pp. 3–19. [Google Scholar]
  39. Shao, J.; Qu, C.; Li, J.; Peng, S. A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification. Sensors 2018, 18, 3039. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The detection network structure. The upper side is the CorrNet for extracting the features of the input images. The lower side is the post-processing part of the detection results. The blue dotted box is the dual-NMS.
Figure 1. The detection network structure. The upper side is the CorrNet for extracting the features of the input images. The lower side is the post-processing part of the detection results. The blue dotted box is the dual-NMS.
Sensors 19 04691 g001
Figure 2. Comparison of the results of the classical NMS and dual-NMS. (a) Outputs of the detection network with no post-processing operations. (b) Detection results where the false detection boxes have been removed by the dual-NMS. (c) The classical NMS algorithm removes the overlapped detection boxes, but fails to remove the false detection boxes. (d) The detection results that were processed by the dual-NMS, where the false detection boxes have been correctly removed.
Figure 2. Comparison of the results of the classical NMS and dual-NMS. (a) Outputs of the detection network with no post-processing operations. (b) Detection results where the false detection boxes have been removed by the dual-NMS. (c) The classical NMS algorithm removes the overlapped detection boxes, but fails to remove the false detection boxes. (d) The detection results that were processed by the dual-NMS, where the false detection boxes have been correctly removed.
Sensors 19 04691 g002
Figure 3. (a) Dynamic threshold distribution of the classification confidence and density of the detection boxes. (b,c) are the univariate relationships of the thresholds for the maximum classification confidence and the density of the detection boxes, respectively.
Figure 3. (a) Dynamic threshold distribution of the classification confidence and density of the detection boxes. (b,c) are the univariate relationships of the thresholds for the maximum classification confidence and the density of the detection boxes, respectively.
Sensors 19 04691 g003
Figure 4. CorrNet structure. The red dashed box indicates the dilated convolution guidance structure. The green box represents the correlation calculation layer of the feature channel separation. The downsampling rates of the feature map of the original YOLOv3 for decoding the calculation loss is 8, 16 and 32, which makes it easy to cause small object losses.
Figure 4. CorrNet structure. The red dashed box indicates the dilated convolution guidance structure. The green box represents the correlation calculation layer of the feature channel separation. The downsampling rates of the feature map of the original YOLOv3 for decoding the calculation loss is 8, 16 and 32, which makes it easy to cause small object losses.
Sensors 19 04691 g004
Figure 5. Dilated convolution guidance structure.
Figure 5. Dilated convolution guidance structure.
Sensors 19 04691 g005
Figure 6. Correlation calculation layer in which the feature channels are separated.
Figure 6. Correlation calculation layer in which the feature channels are separated.
Sensors 19 04691 g006
Figure 7. Precision comparison of the detection results after processing by the dual-NMS and NMS, respectively.
Figure 7. Precision comparison of the detection results after processing by the dual-NMS and NMS, respectively.
Sensors 19 04691 g007
Figure 8. Recall comparison of the detection results after processing by the dual-NMS and NMS, respectively.
Figure 8. Recall comparison of the detection results after processing by the dual-NMS and NMS, respectively.
Sensors 19 04691 g008
Figure 9. Trend of the removal rate with different β, α and γ values.
Figure 9. Trend of the removal rate with different β, α and γ values.
Sensors 19 04691 g009
Figure 10. (a) Trend of the mAP with different α and γ values. (b) Trend of the F1 score with different α and γ values.
Figure 10. (a) Trend of the mAP with different α and γ values. (b) Trend of the F1 score with different α and γ values.
Sensors 19 04691 g010
Figure 11. Comparison of the detection results of different network structures. From top to bottom, the results are for the VEDAI, UCAS-AOD, RSOD and DOTA datasets.
Figure 11. Comparison of the detection results of different network structures. From top to bottom, the results are for the VEDAI, UCAS-AOD, RSOD and DOTA datasets.
Sensors 19 04691 g011
Table 1. The average density of the true and false detection boxes in different datasets.
Table 1. The average density of the true and false detection boxes in different datasets.
DatasetDensity (True Detection Boxes)Density (False Detection Boxes)
VEDAI13.435.58
UCAS-AOD20.086.29
RSOD21.617.86
DOTA12.183.85
Table 2. The proportion of objects with different scales to all targets in each dataset.
Table 2. The proportion of objects with different scales to all targets in each dataset.
DatasetS1S2S3S4
VEDAI67.85%97.56%99.76%99.97%
UCAS-AOD15.78%54.67%90.87%96.60%
RSOD16.01%77.73%99.32%100.00%
DOTA14.31%70.07%100.00%100.00%
Table 3. Comparison of the F1 score for the dual-NMS in different data sets (The best result in each dataset is marked in bold).
Table 3. Comparison of the F1 score for the dual-NMS in different data sets (The best result in each dataset is marked in bold).
ModelVEDAIUCAS-AODRSODDOTA
NMS0.53920.92000.78610.3356
Dual-NMS0.58420.94950.79660.3842
Table 4. Max/min removal rates of the true and false detection boxes for each dataset in different detection networks.
Table 4. Max/min removal rates of the true and false detection boxes for each dataset in different detection networks.
ModelRrVEDAIUCAS-AODRSODDOTA
YOLOv3RrT11.54%/1.81%2.88%/0.86%2.81%/0.18%25.53%/8.50%
RrF56.92%/15.42%57.94%/30.05%51.02%/8.89%52.89%/32.19%
CorrNetRrT6.42%/2.09%1.01%/0.15%0.79%/0.08%11.06%/2.66%
RrF52.22%/16.22%40.44%/28.05%23.47%/14.37%61.56%/22.10%
Table 5. Effectiveness of different optimization methods in each dataset (The best results are marked in bold).
Table 5. Effectiveness of different optimization methods in each dataset (The best results are marked in bold).
ModelReduced Downsampling RatesDilated Convolution Guidance LayerCorrelation Calculation LayermAP (%)
VEDAIUCAS-AODRSODDOTA
YOLOv3 62.8695.9983.2338.91
CorrNet 65.5796.4785.8646.31
66.9597.1486.2147.71
68.0596.8288.0248.69
Table 6. Comparative detection performance of the CorrNet and other detection network (The best results are marked in bold).
Table 6. Comparative detection performance of the CorrNet and other detection network (The best results are marked in bold).
ModelmAP (%)
VEDAIUCAS-AODRSODDOTA
Faster-RCNN64.7089.0084.4760.46
R-FCN61.8093.5084.0752.58
SSD46.1078.8887.8529.86
YOLOv250.3089.4187.3539.20
YOLOv362.8695.9983.2338.91
CorrNet68.0596.8288.0248.69
Table 7. The effects of the dual-NMS on the mAP and FPS in different detection networks.
Table 7. The effects of the dual-NMS on the mAP and FPS in different detection networks.
ModelNMSDual-NMSmAP (%)FPS
VEDAIUCAS-AODRSODDOTA
YOLOv3 62.8695.9983.2338.9111.25
62.1995.3282.8037.5411.04
CorrNet 68.0596.8288.0248.699.99
67.7096.7188.0048.149.81

Share and Cite

MDPI and ACS Style

Lin, Z.; Wu, Q.; Fu, S.; Wang, S.; Zhang, Z.; Kong, Y. Dual-NMS: A Method for Autonomously Removing False Detection Boxes from Aerial Image Object Detection Results. Sensors 2019, 19, 4691. https://doi.org/10.3390/s19214691

AMA Style

Lin Z, Wu Q, Fu S, Wang S, Zhang Z, Kong Y. Dual-NMS: A Method for Autonomously Removing False Detection Boxes from Aerial Image Object Detection Results. Sensors. 2019; 19(21):4691. https://doi.org/10.3390/s19214691

Chicago/Turabian Style

Lin, Zhiyuan, Qingxiao Wu, Shuangfei Fu, Sikui Wang, Zhongyu Zhang, and Yanzi Kong. 2019. "Dual-NMS: A Method for Autonomously Removing False Detection Boxes from Aerial Image Object Detection Results" Sensors 19, no. 21: 4691. https://doi.org/10.3390/s19214691

APA Style

Lin, Z., Wu, Q., Fu, S., Wang, S., Zhang, Z., & Kong, Y. (2019). Dual-NMS: A Method for Autonomously Removing False Detection Boxes from Aerial Image Object Detection Results. Sensors, 19(21), 4691. https://doi.org/10.3390/s19214691

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop