Next Article in Journal
The Differences and Influence Factors in Extracting Urban Green Space from Various Resolutions of Data: The Perspective of Blocks
Next Article in Special Issue
A CNN-Transformer Network Combining CBAM for Change Detection in High-Resolution Remote Sensing Images
Previous Article in Journal
Infrared Small Marine Target Detection Based on Spatiotemporal Dynamics Analysis
Previous Article in Special Issue
Application of a Novel Multiscale Global Graph Convolutional Neural Network to Improve the Accuracy of Forest Type Classification Using Aerial Photographs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Smooth GIoU Loss for Oriented Object Detection in Remote Sensing Images

School of Electrical and Information Engineering, Zhengzhou University of Light Industry, Zhengzhou 450002, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(5), 1259; https://doi.org/10.3390/rs15051259
Submission received: 26 December 2022 / Revised: 10 February 2023 / Accepted: 19 February 2023 / Published: 24 February 2023

Abstract

:
Oriented object detection (OOD) can more accurately locate objects with an arbitrary direction in remote sensing images (RSIs) compared to horizontal object detection. The most commonly used bounding box regression (BBR) loss in OOD is smooth L1 loss, which requires the precondition that spatial parameters are independent of one another. This independence is an ideal that is not achievable in practice. To avoid this problem, various kinds of IoU-based BBR losses have been widely used in OOD; however, their relationships with IoUs are approximately linear. Consequently, the gradient value, i.e., the learning intensity, cannot be dynamically adjusted with the IoU in these cases, which restricts the accuracy of object location. To handle this problem, a novel BBR loss, named smooth generalized intersection over union (GIoU) loss, is proposed. The contributions it makes include two aspects. First of all, smooth GIoU loss can employ more appropriate learning intensities in the different ranges of GIoU values to address the above problem and the design scheme of smooth GIoU loss can be generalized to other IoU-based BBR losses. Secondly, the existing computational scheme of GIoU loss can be modified to fit OOD. The ablation study of smooth GIoU loss validates the effectiveness of its design scheme. Comprehensive comparisons performed on two RSI datasets demonstrate that the proposed smooth GIoU loss is superior to other BBR losses adopted by existing OOD methods and can be generalized for various kinds of OOD methods. Furthermore, the core idea of smooth GIoU loss can be generalized to other IoU-based BBR losses.

1. Introduction

Oriented object detection (OOD) methods can locate objects with arbitrary orientation more accurately than horizontal object detection (HOD) in remote sensing images (RSIs) [1,2,3,4]. OOD has attracted much attention due to its wide applications in military and civilian fields [5,6,7,8,9,10], such as intelligent transportation, disaster rescue, military base reconnaissance, etc. [11,12,13,14,15,16,17].
With the rapid development of deep learning technology [18,19,20,21], OOD methods based on deep learning have made great progress in recent years [22,23,24,25]. According to whether OOD methods include a stage for generating region proposals or not, state-of-the-art (SOTA) OOD methods can be divided into two categories: one-stage [26,27,28,29,30,31,32,33] and two-stage methods [34,35,36,37,38,39,40,41,42,43]. Two-stage methods generate region proposals on the basis of many anchors and classify and regress them, whereas one-stage methods directly classify and regress the anchors. Two-stage methods usually have higher detection accuracy than one-stage methods, whereas one-stage methods usually have faster inference speed than two-stage methods. To the best of our knowledge, the mAP of one-stage and two-stage methods can reach 75.93% [44] and 82.40% [45] on the DOTA dataset (single scale training), respectively, and their inference speeds can reach 200 FPS [46] and 33 FPS [47], respectively. They are briefly described as follows.
The classical one-stage method RetinaNet [40] uses a fully convolutional network to directly predict the category and spatial location of each anchor; furthermore, a focal loss was proposed for assigning greater weight to hard samples. To solve the problem of mismatch between presetting anchors and objects with large aspect ratios, an alignment convolution was introduced in S2ANet [41], which could adaptively align the convolutional kernels with oriented bounding boxes. Label assignment strategies were improved in DAL [42] and Dynamic R-CNN [43]. With DAL, researchers proposed a novel quantitative metric of matching degrees to dynamically assign positive labels to high-quality anchors. This solved the suboptimal localization problem caused by traditional label assignment strategies, wherein locations are solely determined by intersection over union (IoU), resulting in the problem of an imbalanced capability regarding classification and regression. The IoU threshold between the anchors and ground truth bounding boxes was adaptively adjusted in Dynamic R-CNN.
The basic framework of many two-stage OOD methods is inherited from Faster R-CNN [34]. Generally speaking, the region proposal network (RPN) is used to first generate a region of interest (RoI), and then the RoI pooling operation is employed to obtain the deep features with a fixed size. Finally, the classifications and bounding box regressions (BBRs) of RoIs are implemented based on their deep features. To generate these oriented RoIs, the early two-stage method [48] presets 54 anchors with different scales, aspect ratios and angles at each location of the feature map. However, the computational and memory costs are obviously increased. To address this problem, Ding et al. [49] proposed an RoI transformer to generate oriented RoIs from horizontal RoIs, which significantly reduced the number of preset anchors. To further reduce the computational costs and improve detection accuracy, the Oriented R-CNN [50] introduced an oriented RPN to directly generate oriented RoIs from the presetting anchors.
Regardless of whether a one-stage or two-stage method has been employed, smooth L1 loss [51], which can optimize spatial parameters, is usually adopted for BBR. A precondition for the use of smooth L1 loss is that the spatial parameters are independent of one another; however, this ideal cannot be achieved in practice. A number of IoU-based BBR losses have been proposed, such as IoU loss [52] and PIoU loss [53], since the IoU was invariant to scale change and implicitly encoded the relationship of each spatial parameter during area calculation. After this, GIoU loss [54], DIoU loss [55], EIoU loss [56], etc., were proposed, because the IoU cannot handle the situation of two bounding boxes with no intersection. However, the relationship between the existing IoU-based BBR losses and the IoU is approximately linear; consequently, the gradient value, i.e., the learning intensity, cannot be dynamically adjusted with the IoU, which restricts the accuracy of object location.
To handle the aforementioned problems, a novel BBR loss, named smooth generalized intersection over union (GIoU) loss, is proposed for OOD in RSIs. First of all, the computational scheme of GIoU used for HOD is modified to fit OOD. Afterwards, a piecewise function, i.e., smooth GIoU loss, is proposed to adaptively adjust the gradient value with the GIoU value. Specifically, the gradient value of smooth GIoU loss is relatively constant if the GIoU value is small, and they can achieve stable and fast convergence at this stage. However, the absolute value of the gradient will gradually decrease with the increase in the GIoU value if the GIoU value is greater than the threshold; thus, fine-tuning of the learning intensity is required when the GIoU approaches one to avoid training oscillation.
The contributions of this paper are as follows:
(1)
A novel BBR loss, named smooth GIoU loss, is proposed for OOD in RSIs. The smooth GIoU loss can employ more appropriate learning intensities in the different ranges of the GIoU value. It is worth noting that the design scheme of smooth GIoU loss can be generalized to other IoU-based BBR losses, such as IoU loss [52], DIoU loss [55], EIoU loss [56], etc., regardless of whether it is OOD or HOD.
(2)
The existing computational scheme of GIoU is modified to fit OOD.
(3)
The experimental results on two RSI datasets demonstrate that the smooth GIoU loss is superior to other BBR losses and has good generalization capability for various kinds of OOD methods.

2. Materials and Methods

The proposed smooth GIoU loss is closely related to GIoU; therefore, the original computational scheme of GIoU and existing GIoU-based BBR losses are introduced in this section.

2.1. Related Works

2.1.1. Computational Scheme of GIoU for Horizontal Object Detection

The I o U of bounding boxes A and B is formulated as:
I o U = A r e a A B A r e a A B ,   I o U 0 , 1
where A r e a A B and A r e a A B denote the intersection and the union area of A and B, respectively. As shown in Figure 1b, the IoU is always 0 when boxes A and B are non-overlapping, in other words, the IoU cannot measure the distance between them. The GIoU is proposed to address this problem, and its definition is as follows:
G I o U = I o U A r e a C A r e a A B A r e a C ,   G I o U 1 , 1
where C denotes the smallest enclosing rectangle (SER) of boxes A and B, as shown in Figure 1, and A r e a C denotes the area of C. As shown in Equation (2), the GIoU is inversely proportional to the distance between A and B even if they are non-overlapping.

2.1.2. Existing GIoU-Based Bounding Box Regression Losses

Up to now, there are two types of GIoU-based BBR loss, one is the GIoU loss ( L G I o U ) in the form of a linear function [54]:
L G I o U = 1 G I o U
Another one is the IGIoU loss [57] ( L I G I o U ) in the form of a logarithmic function:
L I G I o U = 2 × ln 2 2 × ln ( 1 + G I o U )
Some experiments have demonstrated that the performance of L I G I o U is superior to L G I o U , and the L I G I o U is adopted as the baseline BBR loss of the proposed smooth GIoU loss (see Equation (8)).

2.1.3. Other Important Works

The existing OOD methods are usually divided into two types, i.e., one-stage and two-stage OOD methods. Significant works (except those described in the Introduction) in the two kinds of methods and their limitations that motivate our method are given as follows.
(1) One-stage OOD methods. To improve the localization accuracy of objects with a high aspect ratio, Chen et al. [53] proposed a pixel level computational scheme of the IoU, named PIoU, and designed a PIoU loss for BBR. CFC-Net [58] proposed a polarization attention module to obtain the discriminative feature representation corresponding to classification and BBR tasks. Afterwards, a rotation anchor refinement module was used to refine the horizontal anchors for obtaining oriented anchors, and then a dynamic anchor learning scheme is employed to select the final anchors. DRN [59] proposed a feature selection module, in which the receptive field of neurons could be adaptively adjusted according to the shape and orientation of object; furthermore, the dynamic refinement head is proposed to refine the features in an object-aware manner. Oriented RepPoints [44] used adaptive points to represent the oriented objects, and proposed an adaptive point learning scheme to select representative samples for better training. Other important one-stage OOD methods include DARNet [60], CG-Net [61], FCOS-O [62], Double-Heads [63], etc.
(2) Two-stage OOD methods. AOPG [64] proposed a coarse location module to generate initial proposals, and then a RPN based on AlignConv was used to refine the initial proposals. Finally, the refined proposals were used for OOD. Eight spatial parameters, which are related to OBB and its SER, and an area ratio between OBB and its SER were jointly used to represent the OBB by a gliding vertex [65], and then the BBR of the above nine parameters was employed to precisely localize the OBB. DODet [66] proposed a novel five parameter representation of the OBB to generate the oriented proposals, and then the spatial location output from the BBR branch was used to extract more accurate RoI features for classification, which can handle the feature shifting between classification and BBR branches. Instead of an anchor-based RPN and NMS operation, RSADet [67] directly generated the candidate proposals from the attention map by taking into account the spatial distribution of remote sensing objects; furthermore, deformable convolution was adopted to obtain the features of OBB more accurately. Other important two-stage OOD methods include ReDet [68], CenterMap-Net [69], MPFP-Net [70], PVR-FD [71], DPGN [72], QPDet [73], etc.
(3) Limitations that motivate our method. Regardless of whether a one-stage or two-stage OOD method is used, the smooth L1 loss is usually adopted for BBR. The precondition of using the smooth L1 loss is that the spatial parameters are independent from each other; however, this is not always true in a practical situation. IoU-based BBR losses can avoid the above problem. However, the relationship between the existing IoU-based BBR losses and the IoU is approximately linear; therefore, the learning intensity cannot be dynamically adjusted with the IoU, which restricts the accuracy of object location. To address the aforementioned problems, a novel smooth GIoU loss is proposed, which can adaptively adjust the learning intensity with the GIoU value. Furthermore, the design of smooth GIoU loss can be generalized to other IoU-based BBR losses.

2.2. Proposed Method

The flowchart of our method is shown in Figure 2. First of all, the oriented bounding boxes (OBBs) in RSI are predicted by the OBB detector. Secondly, all possible predicted and ground truth OBB pairs are matched if their IoUs exceed 0.5. Thirdly, the existing computational scheme of the GIoU of HBBs is modified for the computation of the GIoU of OBBs. Finally, the proposed smooth GIoU loss is calculated for the BBR of OBBs.

2.2.1. Computational Scheme of GIoU for Oriented Object Detection

Equation (2) is still used for the computation of the GIoU in OOD; however, the computation of A r e a C in Equation (2) is quite different from the one in HOD. The details can be found in Figure 3 and Algorithm 1.
Algorithm 1: Computational scheme of Area(C)
Input:
        Vertex of OBB A:VA = {A1, A2, A3, A4}
        Vertex of OBB B:VB = {B1, B2, B3, B4}
Output:
        Area(C)
1. Initialize V =
2. for i = 1 to 4 do
3.    if Ai is not enclosed by OBB B then
4.     V = append(V, Ai)
5.    end if
6.    if Bi is not enclosed by OBB A then
7.     V = append(V, Bi)
8.    end if
9. end for
10. Sort all the vertices in V in a counter/clockwise direction, V = {v1, ,vn}, n ≤ 8.
 Connect the sorted vertices to obtain P, its n edges is E = {e1,…,en}.
11. for j = 1 to n do
12.  for k = 1 to n do
13.     d( j, k) = dist_p2l(vj,ek) //smallest distance between vj and ek
14.  end for
15. end for
16. w = max d(j, k),
17. (jm, km) = argmax d(j, k),
18. Draw n parallel lines L = {l1, ,ln} through v1∼vn which are all perpendicular to ekm.
19. for j = 1 to n do
20.  for k = 1 to n do
21.     ld(j, k) = dist_l2l(lj, lk) //distance between lj and lk
22.  end for
23. end for
24. h = max ld(j, k),
25. Area(C) = w × h
26. return Area(C)

2.2.2. Smooth GIoU Loss

Formulation of Smooth GIoU Loss

The core idea of smooth L1 loss is that the convergence is stable and fast when the deviations between the predicted and ground truth values are greater than the threshold. On the contrary, the convergence speed gradually slows down. Inspired by above idea, the absolute value of the gradient of the smooth GIoU loss should be relatively large and constant when the G I o U is small, i.e., the predicted box can quickly approach the ground truth box at this stage. Accordingly, the absolute value of the gradient of the smooth GIoU loss should be gradually decreased when the G I o U is large, i.e., the speed with which the predicted box approaches the ground truth box gradually decreases. In summary, the purpose of smooth GIoU loss is that it does not only avoid training oscil-lation, but also has a moderate learningspeed, so that the localization accuracy can be improved without sacrificing the learning speed. Following the above scheme, the smooth GIoU loss is formulated as follows:
S m o o t h G I o U = G I o U _ L 1 G I o U         G I o U < β G I o U _ L 2 G I o U         G I o U β
G I o U _ L 1 G I o U = B G G I o U 0.5 × B G β
G I o U _ L 2 G I o U = 0.5 × B G 2 G I o U B G β
B G x = k × ln 2 k × ln 1 + x ,   x ( 1 , 1 ]
where S m o o t h G I o U denotes the smooth GIoU loss, G I o U _ L 1 G I o U and G I o U _ L 2 G I o U denote GIoU_L1 loss and GIoU_L2 loss, respectively, B G x denotes the baseline GIoU loss and its formulation refers to IGIoU loss [57] and k and β are hyper-parameters and are quantitatively analyzed in Section 4.2.
As shown in Figure 4a, the gradient value of GIoU_L1 loss is relatively constant, the absolute value of the gradient of GIoU_L2 loss is large when the GIoU is small; however, it rapidly decreases with the increase in the GIoU. Consequently, as shown in Figure 4b, the smooth GIoU loss selects GIoU_L1 loss for the BBR when G I o U < β ; accordingly, the predicted box is close to the ground truth box at a relatively constant rate, the smooth GIoU loss switches to GIoU_L2 loss when G I o U β and the predicted box slowly approaches the ground truth box at this stage.

Proof of the Smooth Characteristic of Smooth GIoU Loss

The smooth characteristic of smooth GIoU loss requires that the smooth GIoU loss should be differentiable and continuous [34], and the key of related proof includes two aspects. On the one hand, the GIoU_L1 and GIoU_L2 losses should be differentiable. On the other hand, the GIoU_L1 and GIoU_L2 losses should be differentiable and continuous at G I o U = β . The details of related proof are as follows:
(1)
Proof that GIoU_L1 and GIoU_L2 losses are differentiable:
lim x 0 f ( x + Δ x ) f ( x ) Δ x = f ( x )
lim Δ G I o U 0 G I o U _ L 1 ( G I o U + Δ G I o U ) G I o U _ L 1 ( G I o U ) Δ G I o U = lim Δ G I o U 0 d G I o U _ L 1 ( G I o U + Δ G I o U ) G I o U _ L 1 ( G I o U ) d Δ G I o U d Δ G I o U d Δ G I o U = lim Δ G I o U 0 d G I o U _ L 1 ( G I o U + Δ G I o U ) d Δ G I o U = lim Δ G I o U 0 k 1 + G I o U + Δ G I o U = k 1 + G I o U = d G I o U _ L 1 G I o U d Δ G I o U
According to Equation (9) (definition of derivative), GIoU_L1 loss is proven to be differentiable.
The proof that GIoU_L2 loss is differentiable can be obtained in a similar way.
(2)
Proof that GIoU_L1 and GIoU_L2 losses are differentiable and continuous at G I o U = β .
G I o U = β is substituted into Equations (6) and (7):
G I o U _ L 1 β = B G β 0.5 × B G β = 0.5 × B G β G I o U _ L 2 β = 0.5 × B G 2 β B G β = 0.5 × B G β
Equation (11) demonstrates that the function values of the GIoU_L1 and GIoU_L2 losses are continuous at G I o U = β . Subsequently, the GIoU_L1 and GIoU_L2 losses take derivatives of GIoU, respectively, and G I o U = β is substituted into their derivative formulation:
d G I o U _ L 1 G I o U d G I o U G I o U = β = k 1 + G I o U G I o U = β = k 1 + β d G I o U _ L 2 G I o U d G I o U G I o U = β = B G G I o U B G β × k 1 + G I o U G I o U = β = B G β B G β × k 1 + β = k 1 + β
Equation (12) demonstrates that the gradient value of GIoU_L1 and GIoU_L2 are continuous at G I o U = β .

3. Results

3.1. Experimental Setup

3.1.1. Datasets

The DOTA dataset [45] includes 2806 RSIs with a size of 800 × 800~4000 × 4000, it contains 188,282 instances of 15 categories: bridge (BR), harbor (HA), ship (SH), plane (PL), helicopter (HC), small vehicle (SV), large vehicle (LV), baseball diamond (BD), ground track field (GTF), tennis court (TC), basketball court (BC), soccer-ball field (SBF), roundabout (RA), swimming pool (SP) and storage tank (ST). Both training and validation sets were used for training; the testing results are obtained from DOTA’s evaluation server. In the single-scale experiment, each image is split into a number of sub-images with 1024 × 1024 pixels by using a step of 824 pixels. In the multi-scale experiment, each image is resized using three scales {0.5,1.0,1.5}, and then split into a number of sub-images with 1024 × 1024 pixels by using a step of 512 pixels.
The DIOR-R dataset [64] includes 23,463 RSIs with a size of 800 × 800 pixels, it contains 192,518 instances of 20 categories: airplane (APL), airport (APO), baseball field (BF), basketball court (BC), bridge (BR), chimney (CH), expressway service area (ESA), expressway toll station (ETS), dam (DAM), golf field (GF), ground track field (GTF), harbor (HA), overpass (OP), ship (SH), stadium (STA), storage tank (STO), tennis court (TC), train station (TS), vehicle (VE) and windmill (WM). Both training and validation sets were used for training, and the rest is for testing.

3.1.2. Implementation Details

Training Details: Stochastic gradient descent (SGD) was used for optimization, where the momentum decay and weight decay were set to 0.9 and 0.0001 [64], respectively. The batch size and initial learning rate were the same as the baseline method, the learning rate decays at the 8th (24th) and 11th (33th) epoch if the total number of epochs is 12 (36) [44], the learning rate becomes 1/10 of the last stage after each decay [64]. All of the training samples were augmented by random horizontal and vertical flipping [64].
Evaluation Metric: The mean average precision (mAP) was adopted as the evaluation metric in this paper, which is consistent with PASCAL VOC [74].
Environment Setting: The experiments were implemented on the MMDetection platform [75] in the Pytorch framework and were run on a workstation with Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz (a total of 10 × 2 cores), 256G memory and eight NVIDIA GeForce RTX 2080Ti (a total of 11 × 8 GB memory).

3.2. Parameters Analysis

As shown in Equation (5), the smooth GIoU loss contains two hyper-parameters, k and β. Quantitative analyses of k and β were conducted on the DOTA dataset to evaluate their impacts on the overall performance, where the S2ANet is adopted as the baseline method of which the backbone is R-50-FPN (denotes ResNet-50-FPN [76]). First of all, k was quantitatively analyzed with an a value of β of 0.5. As shown in Figure 5a, the mAP is at a maximum of 75.18% when k = 5. Subsequently, β was quantitatively analyzed when k = 5. As shown in Figure 5b, mAP is at a maximum of 75.87% when β = 0.75. Finally, k and β were set to 5 and 0.75, respectively.

3.3. Comparisons with Other Bounding Box Regression Losses and Ablation Study

To validate the superiority of smooth GIoU loss, it was quantitatively compared with other BBR losses used by OOD methods on the DOTA and DIOR-R datasets, where the S2ANet [41] is adopted as the baseline method. The BBR losses for comparison include PIoU loss [53], Smooth L1 loss [51], IoU loss [52], Smooth IoU Loss, GioU loss [54], Baseline GioU loss [57], GioU_L1 loss and GioU_L2 loss, where the smooth IoU loss is a modified version of the IoU loss following the definition of a smooth GIoU loss, i.e., the B G x defined in Equation (8) is changed to IoU loss. The formulation of baseline GIoU loss, GIoU_L1 loss and GIoU_L2 loss are presented in Equations (6)–(8), respectively. The comparisons with baseline GIoU loss, GIoU_L1 and GIoU_L2 losses indicate the ablation study of smooth GIoU loss.
As shown in Table 1, smooth GIoU loss exhibits the best performance on both the DOTA and DIOR-R datasets, which validates its superiority. The smooth IoU loss exhibits a suboptimal performance, but is superior to IoU loss, which demonstrates that the core idea of the proposed smooth GIoU loss is also valid for IoU loss. In other words, our core idea has good generalization capability and can be applied to other IoU-based BBR losses. In addition, the superiority over baseline GIoU loss, GIoU_L1 Loss and GIoU_L2 Loss validates the effectiveness of our scheme design.

3.4. Evaluation of Generalization Capability

To evaluate the generalization capability of smooth GIoU loss, some quantitative comparisons were conducted on the DOTA and DIOR-R datasets, where six representative OOD methods (RoI Transformer [49], AOPG [64], Oriented RCNN [50], S2ANet [41], RetinaNet-O [40] and Gliding Vertex [65]) were selected as baseline methods, which include one-stage and two-stage methods and anchor-based and anchor-free methods. The original BBR losses of six OOD methods were replaced with smooth GIoU loss, and then the mAP of original and replaced versions were compared with each other. As shown in Table 2, the replaced version increased by 0.62%~1.75% on the DOTA dataset, and increased by 0.46%~2.10% on the DIOR-R dataset, which demonstrates that the smooth GIoU loss is effective for the six OOD methods, i.e., the smooth GIoU loss has good generalization capability for various kinds of OOD methods.

3.5. Comprehensive Comparisons

To fully validate the effectiveness of smooth GIoU loss, it was separately embedded into a representative one-stage and two-stage baseline method, and then quantitatively compared with other one-stage and two-stage methods on the DOTA and DIOR-R datasets. S2ANet [41] and AOPG [64] were adopted as the one-stage and two-stage baseline methods, respectively, and the embedded methods are denoted as ‘S2ANet + Smooth GIoU’ and ‘AOPG + Smooth GIoU’, respectively.
Results on the DOTA dataset: As shown in Table 3, the S2ANet + Smooth GIoU (epochs = 36) achieved the highest mAP among the one-stage methods with single scale training; it was increased by 0.22% compared with Oriented Reppoint (epochs = 40), which achieved the best results among the one-stage OOD methods. The S2ANet + Smooth GioU (epochs = 12) obtained an mAP of 75.87%; an increase of 1.75% compared with S2ANet, which achieved the best results among the one-stage methods with 12 epochs. The AOPG + Smooth GIoU (epochs = 12) achieved an mAP of 76.19%; an increase of 0.32% compared with Oriented R-CNN (epochs = 12), which achieved the best results among the two-stage OOD methods. The S2ANet + Smooth GIoU and AOPG + Smooth GIoU (epochs = 36) obtained mAPs of 80.61% and 81.03%, respectively, in the multi-scale experiments, which shows a quite competitive performance compared to the DOTA dataset. Furthermore, our method is the best in 14 of the 15 categories in both comparisons of one-stage and two-stage methods. In summary, the comprehensive performance of our method is optimal on the DOTA dataset.
Results on the DIOR-R dataset: As shown in Table 4, the S2ANet + Smooth GIoU (epochs = 36) achieved the highest mAP among the one-stage methods, it increased by 0.64% compared with Oriented Reppoint (epochs = 40), which achieved the best results among the one-stage methods. The S2ANet + Smooth GIoU (epochs = 12) obtained an mAP of 63.90%; an increase of 1.00% compared with S2ANet, which achieved the best results among the one-stage methods with 12 epochs. The AOPG + Smooth GIoU (epochs = 12) obtained an mAP of 66.51%, it increased by 1.41% compared with DODet (epochs = 12), which achieved the best results among the two-stage methods. The AOPG + Smooth GIoU (epochs = 36) obtained an mAP of 69.37%, which shows a quite competitive performance on the DIOR-R dataset. Furthermore, our methods are the best in 15 (16) of the 20 categories in one-stage (two-stage) comparisons. In summary, the comprehensive performance of our methods is optimal on the DIOR-R dataset.
The visualizations of some detection results of S2ANet + Smooth GIoU (epochs = 12) on the DOTA and DIOR-R datasets are shown in Figure 6 and Figure 7, respectively.
In summary, the above comparisons on the DOTA and DIOR-R datasets can fully demonstrate the effectiveness of smooth GIoU loss.

4. Discussion

4.1. Analysis of False Positive Results

As shown in Figure 8, our method gives false positive results in some cases. The main reasons are as follows:
(1)
Similar background. The background regions have similar characteristics to the foreground objects in some cases; consequently, the object detection will be disturbed by the background. As shown in Figure 8a, the harbor has a similar color to the water and trees; therefore, the harbor cannot be completely enclosed by the OBBs.
(2)
Dense objects. Foreground objects in the same category are usually similar in appearance; thus, the object detection will be disturbed by adjacent objects in the same category. As shown in Figure 8b, three vehicles are very close to each other and have similar appearances; consequently, two vehicles are enclosed by a single OBB.
(3)
Shadow. The shadow of a foreground object usually has similar characteristics to object itself; therefore, the shadow will be falsely identified as the part of object. As shown in Figure 8c,d, the shadows of the helicopter and the windmill are also enclosed by the OBBs.
Figure 8. Visualization of some false positive results on two RSI datasets. (a,c) are selected from the DOTA dataset, (b,d) are selected from the DIOR-R dataset.
Figure 8. Visualization of some false positive results on two RSI datasets. (a,c) are selected from the DOTA dataset, (b,d) are selected from the DIOR-R dataset.
Remotesensing 15 01259 g008

4.2. Analysis of Running Time and Detection Accuracy after Expanding GPU Memory

Most of the training settings of our method, such as the number of GPUs, the batch size, optimization algorithm, etc., are the same as the baseline methods for fair comparison. Obviously, using more GPU, i.e., expanding the GPU memory, can accelerate the training and testing speed; therefore, it is necessary to analyze the changes in running time and the detection accuracy after expanding the GPU memory. To this end, a comparison between baseline methods with different GPU memories are conducted in terms of training time, testing time and mAP.
As shown in Table 5, the training and testing time of two baseline methods are apparently reduced through expanding the GPU memory because the parallel computation capability is improved by using more GPU. However, the mAP slightly decreased after expanding the GPU memory, a possible reason for this is as follows. The randomness of the gradient descent direction is weakened by using more GPU (i.e., increasing the batch size); therefore, the deep learning model more easily converges to the local optimal solution [77].

5. Conclusions

Most of the IoU-based BBR losses for OOD in RSIs cannot dynamically adjust the learning intensity with the IoU, which restricts the accuracy of object location; therefore, a novel BBR loss, named smooth GIoU loss, is proposed in this paper to handle this problem. First of all, the existing computational scheme of GIoU is modified to fit OOD. Afterwards, a smooth GIoU loss in the form of a piecewise function is proposed to adaptively adjust the gradient value with the GIoU value. The ablation study of smooth GIoU loss validates the effectiveness of its design. Quantitative comparisons with other BBR losses demonstrate that the performance of the smooth GIoU loss is superior to other BBR losses adopted by existing OOD methods. Quantitative comparisons of four representative OOD methods, which include one-stage and two-stage methods and anchor-based and anchor-free methods, show that replacing their original BBR losses with smooth GIoU loss can improve their performance, which indicates that the smooth GIoU loss can be generalized for various kinds of OOD methods. Comprehensive comparisons conducted on the DOTA and DIOR-R datasets further validate the effectiveness of smooth GIoU loss.
The core idea of smooth GIoU loss is also generalized to the existing IoU loss, named smooth IoU loss. The experimental results demonstrate that it also exhibits a good performance. As a matter of fact, the core idea of smooth GIoU loss can also be generalized to other IoU-based BBR losses in the future.

Author Contributions

Conceptualization, X.Q.; formal analysis, X.Q.; funding acquisition, X.Q; methodology, X.Q. and N.Z.; project administration, W.W.; resources, W.W.; software, N.Z.; supervision, W.W.; validation, W.W.; writing—original draft, N.Z.; writing—review and editing, X.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant no. 62076223.

Data Availability Statement

The DOTA and DIOR-R datasets are available at following URLs: https://captain-whu.github.io/DOTA/dataset.html (accessed on 25 December 2022) and https://gcheng-nwpu.github.io/#Datasets (accessed on 25 December 2022), respectively.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yao, X.; Shen, H.; Feng, X.; Cheng, G.; Han, J. R2IPoints: Pursuing Rotation-Insensitive Point Representation for Aerial Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  2. Cheng, G.; Cai, L.; Lang, C.; Yao, X.; Chen, J.; Guo, L.; Han, J. SPNet: Siamese-prototype network for few-shot remote sensing image scene classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–11. [Google Scholar] [CrossRef]
  3. Wu, X.; Hong, D.; Chanussot, J. Convolutional neural networks for multimodal remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–10. [Google Scholar] [CrossRef]
  4. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. An augmented linear mixing model to address spectral variability for hyperspectral unmixing. IEEE Trans. Image Process. 2018, 28, 1923–1938. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Zhang, L.; Zhang, J.; Wei, W.; Zhang, Y. Learning discriminative compact representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8276–8289. [Google Scholar] [CrossRef]
  6. Cheng, G.; Han, J.; Zhou, P.; Xu, D. Learning rotation-invariant and fisher discriminative convolutional neural networks for object detection. IEEE Trans. Image Process. 2018, 28, 265–278. [Google Scholar] [CrossRef]
  7. Zhu, Q.; Guo, X.; Deng, W.; Guan, Q.; Zhong, Y.; Zhang, L.; Li, D. Land-use/land-cover change detection based on a Siamese global learning framework for high spatial resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 63–78. [Google Scholar] [CrossRef]
  8. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph convolutional networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5966–5978. [Google Scholar] [CrossRef]
  9. Wu, X.; Hong, D.; Chanussot, J. UIU-Net: U-Net in U-Net for Infrared Small Object Detection. IEEE Trans. Image Process. 2022, 32, 364–376. [Google Scholar] [CrossRef]
  10. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More diverse means better: Multimodal deep learning meets remote-sensing imagery classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4340–4354. [Google Scholar] [CrossRef]
  11. Zhang, L.; Zhang, Y.; Yan, H.; Gao, Y.; Wei, W. Salient object detection in hyperspectral imagery using multi-scale spectral-spatial gradient. Neurocomputing 2018, 291, 215–225. [Google Scholar] [CrossRef]
  12. Liao, W.; Bellens, R.; Pizurica, A.; Philips, W.; Pi, Y. Classification of hyperspectral data over urban areas using directional morphological profiles and semi-supervised feature extraction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1177–1190. [Google Scholar] [CrossRef]
  13. Gao, L.; Zhao, B.; Jia, X.; Liao, W.; Zhang, B. Optimized kernel minimum noise fraction transformation for hyperspectral image classification. Remote Sens. 2017, 9, 548. [Google Scholar] [CrossRef] [Green Version]
  14. Du, L.; You, X.; Li, K.; Meng, L.; Cheng, G.; Xiong, L.; Wang, G. Multi-modal deep learning for landform recognition. ISPRS J. Photogramm. Remote Sens. 2019, 158, 63–75. [Google Scholar] [CrossRef]
  15. Zhang, L.; Shi, Z.; Wu, J. A hierarchical oil tank detector with deep surrounding features for high-resolution optical satellite imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 4895–4909. [Google Scholar] [CrossRef]
  16. Stankov, K.; He, D.-C. Detection of buildings in multispectral very high spatial resolution images using the percentage occupancy hit-or-miss transform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4069–4080. [Google Scholar] [CrossRef]
  17. Han, X.; Zhong, Y.; Zhang, L. An efficient and robust integrated geospatial object detection framework for high spatial resolution remote sensing imagery. Remote Sens. 2017, 9, 666. [Google Scholar] [CrossRef] [Green Version]
  18. Qian, X.; Zeng, Y.; Wang, W.; Zhang, Q. Co-saliency Detection Guided by Group Weakly Supervised Learning. IEEE Trans. Multimed. 2022. [CrossRef]
  19. Qian, X.; Huo, Y.; Cheng, G.; Yao, X.; Li, K.; Ren, H.; Wang, W. Incorporating the Completeness and Difficulty of Proposals Into Weakly Supervised Object Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1902–1911. [Google Scholar] [CrossRef]
  20. Qian, X.; Li, J.; Cao, J.; Wu, Y.; Wang, W. Micro-cracks detection of solar cells surface via combining short-term and long-term deep features. Neural Netw. 2020, 127, 132–140. [Google Scholar] [CrossRef]
  21. Zhu, Q.; Deng, W.; Zheng, Z.; Zhong, Y.; Guan, Q.; Lin, W.; Zhang, L.; Li, D. A spectral-spatial-dependent global learning framework for insufficient and imbalanced hyperspectral image classification. IEEE Trans. Cybern. 2021, 52, 11709–11723. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, D.; Meng, D.; Han, J. Co-saliency detection via a self-paced multiple-instance learning framework. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 865–878. [Google Scholar] [CrossRef] [PubMed]
  23. Han, J.; Zhang, D.; Hu, X.; Guo, L.; Ren, J.; Wu, F. Background prior-based salient object detection via deep reconstruction residual. IEEE Trans. Circuits Syst. Video Technol. 2014, 25, 1309–1321. [Google Scholar] [CrossRef] [Green Version]
  24. Han, J.; Yao, X.; Cheng, G.; Feng, X.; Xu, D. P-CNN: Part-based convolutional neural networks for fine-grained visual categorization. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 44, 579–590. [Google Scholar] [CrossRef] [PubMed]
  25. Han, J.; Ji, X.; Hu, X.; Zhu, D.; Li, K.; Jiang, X.; Cui, G.; Guo, L.; Liu, T. Representing and retrieving video shots in human-centric brain imaging space. IEEE Trans. Image Process. 2013, 22, 2723–2736. [Google Scholar] [CrossRef] [Green Version]
  26. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef] [Green Version]
  27. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  28. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  29. Liu, W.; Ma, L.; Wang, J. Detection of multiclass objects in optical remote sensing images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 791–795. [Google Scholar] [CrossRef]
  30. Chen, S.; Zhan, R.; Zhang, J. Geospatial object detection in remote sensing imagery based on multiscale single-shot detector with activated semantics. Remote Sens. 2018, 10, 820. [Google Scholar] [CrossRef] [Green Version]
  31. Tang, T.; Zhou, S.; Deng, Z.; Lei, L.; Zou, H. Arbitrary-oriented vehicle detection in aerial imagery with single convolutional neural networks. Remote Sens. 2017, 9, 1170. [Google Scholar] [CrossRef] [Green Version]
  32. Tayara, H.; Chong, K.T. Object detection in very high-resolution aerial images using one-stage densely connected feature pyramid network. Sensors 2018, 18, 3341. [Google Scholar] [CrossRef] [Green Version]
  33. Chen, Z.; Zhang, T.; Ouyang, C. End-to-end airplane detection using transfer learning in remote sensing images. Remote Sens. 2018, 10, 139. [Google Scholar] [CrossRef] [Green Version]
  34. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  35. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. Adv. Neural Inf. Process. Syst. 2016, 29, 379–387. [Google Scholar]
  36. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  37. Li, K.; Cheng, G.; Bu, S.; You, X. Rotation-insensitive and context-augmented object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2017, 56, 2337–2348. [Google Scholar] [CrossRef]
  38. Zhong, Y.; Han, X.; Zhang, L. Multi-class geospatial object detection based on a position-sensitive balancing framework for high spatial resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2018, 138, 281–294. [Google Scholar] [CrossRef]
  39. Chen, C.; Gong, W.; Chen, Y.; Li, W. Object detection in remote sensing images based on a scene-contextual feature pyramid network. Remote Sens. 2019, 11, 339. [Google Scholar] [CrossRef] [Green Version]
  40. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  41. Han, J.; Ding, J.; Li, J.; Xia, G.-S. Align deep features for oriented object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–11. [Google Scholar] [CrossRef]
  42. Ming, Q.; Zhou, Z.; Miao, L.; Zhang, H.; Li, L. Dynamic anchor learning for arbitrary-oriented object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 2355–2363. [Google Scholar] [CrossRef]
  43. Zhang, H.; Chang, H.; Ma, B. Dynamic R-CNN: Towards high quality object detection via dynamic training. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  44. Li, W.; Chen, Y.; Hu, K. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  45. Xia, G.-S.; Bai, X.; Ding, J. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  46. Mehta, R.; Ozturk, C. Object detection at 200 frames per second. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  47. Zhou, X.; Koltun, V.; Krähenbühl, P. Probabilistic two-stage detection. arXiv 2021, arXiv:2103.07461. [Google Scholar]
  48. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-oriented scene text detection via rotation proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef] [Green Version]
  49. Ding, J.; Xue, N.; Long, Y. Learning roi transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  50. Xie, X.; Cheng, G.; Wang, J. Oriented R-CNN for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  51. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  52. Zhou, D.; Fang, J.; Song, X.; Guan, C.; Yin, J.; Dai, Y.; Yang, R. Iou loss for 2d/3d object detection. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019. [Google Scholar]
  53. Chen, Z.; Chen, K.; Lin, W. Piou loss: Towards accurate oriented object detection in complex environments. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  54. Rezatofighi, H.; Tsoi, N.; Gwak, J. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  55. Zheng, Z.; Wang, P.; Liu, W. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  56. Zhang, Y.-F.; Ren, W.; Zhang, Z. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  57. Qian, X.; Lin, S.; Cheng, G. Object detection in remote sensing images based on improved bounding box regression and multi-level features fusion. Remote Sens. 2020, 12, 143. [Google Scholar] [CrossRef] [Green Version]
  58. Ming, Q.; Miao, L.; Zhou, Z. CFC-Net: A critical feature capturing network for arbitrary-oriented object detection in remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar] [CrossRef]
  59. Pan, X.; Ren, Y.; Sheng, K. Dynamic refinement network for oriented and densely packed object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  60. Zhang, F.; Wang, X.; Zhou, S. DARDet: A dense anchor-free rotated object detector in aerial images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  61. Wei, Z.; Liang, D.; Zhang, D. Learning calibrated-guidance for object detection in aerial images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2721–2733. [Google Scholar] [CrossRef]
  62. Tian, Z.; Chu, X.; Wang, X.; Wei, X.; Shen, C. Fully Convolutional One-Stage 3D Object Detection on LiDAR Range Images. arXiv 2022, arXiv:2205.13764. [Google Scholar]
  63. Wu, Y.; Chen, Y.; Yuan, L. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  64. Cheng, G.; Wang, J.; Li, K. Anchor-free oriented proposal generator for object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  65. Xu, Y.; Fu, M.; Wang, Q. Gliding vertex on the horizontal bounding box for multi-oriented object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 1452–1459. [Google Scholar] [CrossRef] [Green Version]
  66. Cheng, G.; Yao, Y.; Li, S. Dual-aligned oriented detector. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  67. Yu, D.; Ji, S. A new spatial-oriented object detection framework for remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  68. Han, J.; Ding, J.; Xue, N. Redet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 202.
  69. Wang, J.; Yang, W.; Li, H.-C. Learning center probability map for detecting objects in aerial images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4307–4323. [Google Scholar] [CrossRef]
  70. Shamsolmoali, P.; Chanussot, J.; Zareapoor, M. Multipatch feature pyramid network for weakly supervised object detection in optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
  71. Miao, S.; Cheng, G.; Li, Q. Precise Vertex Regression and Feature Decoupling for Oriented Object Detection. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022. [Google Scholar] [CrossRef]
  72. Li, Q.; Cheng, G.; Miao, S. Dynamic Proposal Generation for Oriented Object Detection in Aerial Images. IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022. [Google Scholar]
  73. Yao, Y.; Cheng, G.; Wang, G. On Improving Bounding Box Representations for Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2022. [Google Scholar] [CrossRef]
  74. Everingham, M.; Van Gool, L.; Williams, C.K. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  75. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  76. Shafiq, M.; Gu, Z. Deep residual learning for image recognition: A survey. Appl. Sci. 2022, 12, 8972. [Google Scholar] [CrossRef]
  77. Li, Z.; Wang, W.; Li, H. Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
Figure 1. Illustrations of (a) two intersecting and (b) non-overlapping bounding boxes. The box C denotes the SER of box A and B.
Figure 1. Illustrations of (a) two intersecting and (b) non-overlapping bounding boxes. The box C denotes the SER of box A and B.
Remotesensing 15 01259 g001
Figure 2. Flowchart of our method. (a) Predicting the OBBs of an input RSI via an OBB detector, (b) all possible predicted (red) and ground truth (green) OBB pairs are matched according to their IoU, (c) calculating the GIoU of each OBB pair via the proposed computational scheme of GIoU, (d) calculating the proposed smooth GIoU loss for BBR.
Figure 2. Flowchart of our method. (a) Predicting the OBBs of an input RSI via an OBB detector, (b) all possible predicted (red) and ground truth (green) OBB pairs are matched according to their IoU, (c) calculating the GIoU of each OBB pair via the proposed computational scheme of GIoU, (d) calculating the proposed smooth GIoU loss for BBR.
Remotesensing 15 01259 g002
Figure 3. Illustration of the SER of two OBBs. P and C denote the smallest enclosing polygon and SER of OBB A and B, respectively. w and h denote the width and height of C, respectively. A1, A2, A3 and A4 are the vertexes of OBB A. B1, B2, B3 and B4 are the vertexes of OBB B.
Figure 3. Illustration of the SER of two OBBs. P and C denote the smallest enclosing polygon and SER of OBB A and B, respectively. w and h denote the width and height of C, respectively. A1, A2, A3 and A4 are the vertexes of OBB A. B1, B2, B3 and B4 are the vertexes of OBB B.
Remotesensing 15 01259 g003
Figure 4. The curves of GIoU_L1 loss (red solid line), GIoU_L2 loss (blue solid line) and smooth GIoU loss (yellow dashed line). (a) The curves of three losses within [−0.1, 1.0]. (b) Enlarging curves of three losses within [0.5, 1.0]. k and β in the three losses are set to 5 and 0.75, respectively (see Section 4.2).
Figure 4. The curves of GIoU_L1 loss (red solid line), GIoU_L2 loss (blue solid line) and smooth GIoU loss (yellow dashed line). (a) The curves of three losses within [−0.1, 1.0]. (b) Enlarging curves of three losses within [0.5, 1.0]. k and β in the three losses are set to 5 and 0.75, respectively (see Section 4.2).
Remotesensing 15 01259 g004
Figure 5. Quantitative parameter analyses of β and k on the DOTA dataset. (a) The mAP of our method with different values of β, (b) the mAP of our method with different values of k.
Figure 5. Quantitative parameter analyses of β and k on the DOTA dataset. (a) The mAP of our method with different values of β, (b) the mAP of our method with different values of k.
Remotesensing 15 01259 g005aRemotesensing 15 01259 g005b
Figure 6. Visualizations of some of the detection results of S2ANet + Smooth GIoU (epochs = 12) on the DOTA dataset. OBBs with different colors indicate different categories.
Figure 6. Visualizations of some of the detection results of S2ANet + Smooth GIoU (epochs = 12) on the DOTA dataset. OBBs with different colors indicate different categories.
Remotesensing 15 01259 g006
Figure 7. Visualizations of some of the detection results of S2ANet + Smooth GIoU (epoch = 12) on the DIOR-R dataset. OBBs with different colors indicate different categories.
Figure 7. Visualizations of some of the detection results of S2ANet + Smooth GIoU (epoch = 12) on the DIOR-R dataset. OBBs with different colors indicate different categories.
Remotesensing 15 01259 g007
Table 1. Quantitative comparisons between smooth GIoU loss and other BBR losses used by OOD methods on the DOTA and DIOR-R datasets, where S2ANet [41] is adopted as the baseline method. The comparisons of GIoU_L1 and GIoU_L2 losses indicate the ablation study of smooth GIoU loss. Bold denotes the best performance.
Table 1. Quantitative comparisons between smooth GIoU loss and other BBR losses used by OOD methods on the DOTA and DIOR-R datasets, where S2ANet [41] is adopted as the baseline method. The comparisons of GIoU_L1 and GIoU_L2 losses indicate the ablation study of smooth GIoU loss. Bold denotes the best performance.
MethodEpochBackboneLossmAP
DOTADIOR-R
S2ANet [41]12R-50-FPNPIoU Loss72.4260.02
Smooth L1 Loss74.1262.90
IoU Loss74.6462.94
Smooth IoU Loss75.6963.47
GIoU Loss74.8062.99
Baseline GIoU loss74.1262.58
GIoU_L1 Loss74.4262.78
GIoU_L2 Loss74.9363.13
Smooth GIoU Loss75.8763.90
Table 2. Quantitative comparisons between the six representative OOD methods and their revised version in which the original BBR losses are replaced with the smooth GIoU loss.
Table 2. Quantitative comparisons between the six representative OOD methods and their revised version in which the original BBR losses are replaced with the smooth GIoU loss.
MethodsBackboneEpochSmooth GIoU LossmAP
DOTADIOR-R
RoI Transformer ☨ [49]R-50-FPN1274.61
76.30 (+1.69)
63.87
64.84 (+0.97)
AOPG [64]75.24
76.19 (+0.95)
64.41
66.51 (+2.10)
Oriented RCNN [50]75.87
76.49 (+0.62)
64.63
65.09 (+0.46)
S2ANet [41]74.12
75.87 (+1.75)
62.90
63.90 (+1.00)
RetinaNet-O ☨ [40]68.43
69.23 (+0.80)
57.55
57.89 (+0.34)
Gliding Vertex [65]75.02
75.37 (+0.35)
60.06
61.02 (+0.96)
☨ denotes the results from AerialDetection.
Table 3. Comparison of existing OOD methods on the DOTA dataset. The proposed smooth GIoU loss is embedded in a representative one-stage method (S2ANet [41]) and two-stage method (AOPG [64]) for comparison. The detection results of each category are reported to better demonstrate where the performance gains come from. Blue and red bold fonts denote the best results for each category in one-stage and two-stage methods, respectively (the same applies to below).
Table 3. Comparison of existing OOD methods on the DOTA dataset. The proposed smooth GIoU loss is embedded in a representative one-stage method (S2ANet [41]) and two-stage method (AOPG [64]) for comparison. The detection results of each category are reported to better demonstrate where the performance gains come from. Blue and red bold fonts denote the best results for each category in one-stage and two-stage methods, respectively (the same applies to below).
MethodsBackboneEpochPLBDBRGTFSVLVSHTC
One-stage:
PIoU [53] DLA-34 [36]1280.9069.7024.1060.2038.3064.4064.8090.90
RetinaNet-O ☨ [40]R-50-FPN1288.6777.6241.8158.1774.5871.6479.1190.29
DRN [59] H-104 [45]12088.9180.2243.5263.3573.4870.6984.9490.14
DAL [42]R-50-FPN1288.6876.5545.0866.8067.0076.7679.7490.84
CFC-Net [58] R-50-FPN2589.0880.4152.4170.0276.2878.1187.2190.89
S2ANet [41]R-50-FPN1289.1182.8448.3771.1178.1178.3987.2590.83
Oriented RepPoints [44] R-50-FPN4087.0283.1754.1371.1680.1878.4087.2890.90
Ours:
S2ANet + SmoothGIoUR-50-FPN1289.3681.9553.4074.6078.0381.5087.5890.90
S2ANet + SmoothGIoUR-50-FPN3688.7880.1553.8873.5779.2582.1487.8790.87
S2ANet + SmoothGIoU *R-50-FPN1290.0883.9958.2880.9279.6783.3288.1590.87
S2ANet + SmoothGIoU *R-50-FPN3689.9184.9759.5081.2478.9385.2688.7090.90
Two-stage:
Faster R-CNN ☨ [34]R-50-FPN1288.4473.0644.8659.0973.2571.4977.1190.84
CenterMap-Net [69] R-50-FPN1288.8881.2453.1560.6578.6266.5578.1088.83
PVR-FD [71]R-50-FPN1288.7082.2050.1070.4078.7074.2086.2090.80
DPGN [72]R-101-FPN1289.1579.5751.2677.6176.2981.6779.9590.90
RoI Transformer ☨ [49]R-50-FPN1288.6582.6052.5370.8777.9376.6786.8790.71
AOPG [64] R-50-FPN1289.2783.4952.5069.9773.5182.3187.9590.89
DODet [66] R-50-FPN1289.3484.3151.3971.0479.0482.8688.1590.90
Oriented R-CNN [50]R-50-FPN1289.4682.1254.7870.8678.9383.0088.2090.90
Ours:
AOPG + Smooth GIoUR-50-FPN1289.6282.8952.2472.3478.2783.0888.1890.89
AOPG + Smooth GIoUR-50-FPN3688.9383.8755.0470.6178.7684.4088.2190.87
AOPG + Smooth GIoU *R-50-FPN1290.2584.8059.4679.8078.9985.1588.5690.90
AOPG + Smooth GIoU *R-50-FPN3689.7785.4059.2979.1679.0484.7888.6290.90
MethodsBackboneEpochBCSTSBFRAHASPHCmAP
One-stage:
PIoU [53] DLA-34 [36]1277.2070.4046.5037.1057.1061.9064.0060.50
RetinaNet-O ☨ [40]R-50-FPN1282.1874.3254.7560.6062.5769.5760.6468.43
DRN [59] H-104 [45]12083.8584.1150.1258.4167.6268.6052.5070.70
DAL [46]R-50-FPN1279.5478.4557.7162.2769.0573.1460.1171.44
CFC-Net [58] R-50-FPN2584.4785.6460.5161.5267.8268.0250.0973.50
S2ANet [41]R-50-FPN1284.9085.6460.3662.6065.2669.1357.9474.12
Oriented RepPoints [44] R-50-FPN4085.9786.2559.9070.4973.5372.2758.9775.97
Ours:
S2ANet + SmoothGIoUR-50-FPN1285.0884.5963.2364.8473.6269.2560.0675.87
S2ANet + SmoothGIoUR-50-FPN3685.1784.4362.7670.0875.5971.8556.5376.19
S2ANet + SmoothGIoU *R-50-FPN1284.5287.1672.3372.0880.7780.7165.2179.87
S2ANet + SmoothGIoU *R-50-FPN3688.1987.6970.1369.2482.5478.2473.7180.61
Two-stage:
Faster R-CNN ☨ [34]R-50-FPN1278.9483.9048.5962.9562.1864.9156.1869.05
CenterMap-Net [69] R-50-FPN1277.8083.6149.3666.1972.1072.3658.7071.74
PVR-FD [71]R-50-FPN1283.1085.1058.2063.2064.8066.0054.8073.10
DPGN [72]R-101-FPN1285.5284.7161.7262.8275.3166.3450.8674.25
RoI Transformer ☨ [49]R-50-FPN1283.8382.5153.9567.6174.6768.7561.0374.61
AOPG [64] R-50-FPN1287.6484.7160.0166.1274.1968.3057.8075.24
DODet [66] R-50-FPN1286.8884.9162.6967.6375.4772.2245.5475.49
Oriented R-CNN [50]R-50-FPN1287.5084.6863.9767.6974.9468.8452.2875.87
Ours:
AOPG + Smooth GIoUR-50-FPN1287.1084.4863.4966.5572.6469.1062.0376.19
AOPG + Smooth GIoUR-50-FPN3687.0684.1863.6568.4276.7572.1659.4876.83
AOPG + Smooth GIoU *R-50-FPN1287.1087.0970.9672.4982.3679.2773.8580.74
AOPG + Smooth GIoU *R-50-FPN3687.5087.7371.1672.7983.2681.6874.3781.03
☨ denotes the results from AerialDetection; * denotes multiscale training and testing (the same applies to below).
Table 4. Comparisons with the existing OOD methods on the DIOR-R dataset.
Table 4. Comparisons with the existing OOD methods on the DIOR-R dataset.
MethodBackboneEpoch APLAPOBFBCBRCHDAMETSESAGF
One-stage:
FCOS-O [62]R-50-FPN12 48.7024.8863.5780.9718.4168.9923.2642.3760.2564.83
RetinaNet-O [40]R-50-FPN12 61.4928.5273.5781.1723.9872.5419.9472.3958.2069.25
Double-Heads [63]R-50-FPN12 62.1319.5371.5087.0928.0172.1720.3561.1964.5673.37
S2ANet [41]R-50-FPN12 65.4042.0475.1583.9136.0172.6128.0165.0975.1175.56
Oriented RepPoints [44]R-50-FPN40 ----------
Ours:
S2ANet + SmoothGIoUR-50-FPN12 65.0146.2176.1385.0338.4472.6131.7865.7375.8674.94
S2ANet + SmoothGIoUR-50-FPN36 66.9154.1675.5285.8245.3376.7135.2570.4378.8377.86
Two-stage:
PVR-FD [71]R-50-FPN12 62.9025.2071.8081.4033.2072.2072.5065.3018.5066.10
Faster R-CNN ☨ [34]R-50-FPN12 62.7926.8071.7280.9134.2072.5718.9566.4565.7566.63
Gliding Vertex [65]R-50-FPN12 65.3528.8774.9681.3333.8874.3119.5870.7264.772.30
RoI Transformer [49]R-50-FPN12 63.3437.8871.7887.5340.6872.6026.8678.7168.0968.96
QPDet [73]R-50-FPN12 63.2241.3971.9788.5541.2372.6328.8278.9069.0070.07
AOPG [64]R-50-FPN12 62.3937.7971.6287.6340.9072.4731.0865.4277.9973.20
Oriented R-CNN [50]R-50-FPN12 62.0044.9271.7887.9343.8472.6435.4666.3981.3574.10
DODet [66]R-50-FPN12 63.4043.3572.1181.3243.1272.5933.3278.7770.8474.15
Ours:
AOPG + Smooth GIoUR-50-FPN12 69.8740.6677.2887.7741.1372.6432.0763.6574.4973.57
AOPG + Smooth GIoUR-50-FPN36 69.6452.9871.6888.7446.5877.3836.0772.9482.7579.53
MethodBackboneEpochGTFHAOPSHSTASTOTCTSVEVMmAP
One-stage:
FCOS-O [62]R-50-FPN1260.6631.8440.8073.0966.3256.6177.5538.1030.6955.8751.39
RetinaNet-O [40]R-50-FPN1279.5432.1444.8777.7165.5761.0981.4647.3338.0160.2457.55
Double-Heads [63]R-50-FPN1281.9740.6842.4080.3673.1262.3787.0954.9441.3264.8659.45
S2ANet [41]R-50-FPN1280.4735.9152.1082.3365.8966.0884.6154.1348.0069.6762.90
Oriented RepPoints [44]R-50-FPN40----------66.71
Ours:
S2ANet + SmoothGIoUR-50-FPN1280.3638.2853.4983.6964.5968.2485.2055.4849.5267.4263.90
S2ANet + SmoothGIoUR-50-FPN3680.7648.6958.6786.4068.8569.1286.2058.6253.2069.5967.35
Two-stage:
PVR-FD [71]R-50-FPN1281.3037.0048.3081.0070.1062.2081.5051.8042.5064.1059.50
Faster R-CNN ☨ [34]R-50-FPN1279.2434.9548.7981.1464.3471.2181.4447.3150.4665.2159.54
Gliding Vertex [65]R-50-FPN1278.6837.2249.6480.2269.2661.1381.4944.7647.7165.0460.06
RoI Transformer [49]R-50-FPN1282.7447.7155.6181.2178.2370.2681.6154.8643.2765.5263.87
QPDet [73]R-50-FPN1283.0147.8355.5481.2372.1562.6689.0558.0943.3865.3664.20
AOPG [64]R-50-FPN1281.9442.3254.4581.1772.6971.3181.4960.0452.3869.9964.41
Oriented R-CNN [50]R-50-FPN1280.9543.5258.4281.2568.0162.5288.6259.3143.2766.3164.63
DODet [66]R-50-FPN1275.4748.0059.3185.4174.0471.5681.5255.4751.8666.4065.10
Ours:
AOPG + Smooth GIoUR-50-FPN1281.7346.9154.5889.0771.6877.3188.4559.2352.9970.1566.51
AOPG + Smooth GIoUR-50-FPN3682.4652.5659.6089.7175.5578.2681.4962.5955.9470.9969.37
☨ denotes the results from AerialDetection.
Table 5. Comparisons between baseline methods with different GPU running memory in terms of training time, testing time and mAP (%). The single GPU (NVIDIA GeForce RTX 2080Ti) memory in our workstation is 11 GB. The number of GPUs used by S2ANet and AOPG is four [41] and one [64], respectiely, i.e., the GPU memories used by S2ANet and AOPG are 44 GB and 11 GB, respectively. The total GPU memory of our workstation is 88 GB.
Table 5. Comparisons between baseline methods with different GPU running memory in terms of training time, testing time and mAP (%). The single GPU (NVIDIA GeForce RTX 2080Ti) memory in our workstation is 11 GB. The number of GPUs used by S2ANet and AOPG is four [41] and one [64], respectiely, i.e., the GPU memories used by S2ANet and AOPG are 44 GB and 11 GB, respectively. The total GPU memory of our workstation is 88 GB.
MethodMemory
(GB)
Training Time
(min)
Testing Time
(s)
mAP
(%)
S2ANet
(One-stage baseline)
44
88
206
139
235
132
74.04
72.43
AOPG
(Two-stage baseline)
11
88
1045
209
90
31
75.10
74.62
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qian, X.; Zhang, N.; Wang, W. Smooth GIoU Loss for Oriented Object Detection in Remote Sensing Images. Remote Sens. 2023, 15, 1259. https://doi.org/10.3390/rs15051259

AMA Style

Qian X, Zhang N, Wang W. Smooth GIoU Loss for Oriented Object Detection in Remote Sensing Images. Remote Sensing. 2023; 15(5):1259. https://doi.org/10.3390/rs15051259

Chicago/Turabian Style

Qian, Xiaoliang, Niannian Zhang, and Wei Wang. 2023. "Smooth GIoU Loss for Oriented Object Detection in Remote Sensing Images" Remote Sensing 15, no. 5: 1259. https://doi.org/10.3390/rs15051259

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop