Next Article in Journal
An Anti-Forensics Video Forgery Detection Method Based on Noise Transfer Matrix Analysis
Previous Article in Journal
U-TAG: Electromagnetic Wireless Sensing System for Robotic Hand Pre-Grasping
Previous Article in Special Issue
Lizard Body Temperature Acquisition and Lizard Recognition Using Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Elevating Detection Performance in Optical Remote Sensing Image Object Detection: A Dual Strategy with Spatially Adaptive Angle-Aware Networks and Edge-Aware Skewed Bounding Box Loss Function

Institute of System Engineering, Academy of Military Sciences, Beijing 100141, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(16), 5342; https://doi.org/10.3390/s24165342 (registering DOI)
Submission received: 11 July 2024 / Revised: 4 August 2024 / Accepted: 13 August 2024 / Published: 18 August 2024
(This article belongs to the Special Issue Object Detection Based on Vision Sensors and Neural Network)

Abstract

:
In optical remote sensing image object detection, discontinuous boundaries often limit detection accuracy, particularly at high Intersection over Union (IoU) thresholds. This paper addresses this issue by proposing the Spatial Adaptive Angle-Aware ( SA 3 ) Network. The SA 3 Network employs a hierarchical refinement approach, consisting of coarse regression, fine regression, and precise tuning, to optimize the angle parameters of rotated bounding boxes. It adapts to specific task scenarios using either class-aware or class-agnostic strategies. Experimental results demonstrate its effectiveness in significantly improving detection accuracy at high IoU thresholds. Additionally, we introduce a Gaussian transform-based IoU factor during angle regression loss calculation, leading to the development of Edge-aware Skewed Bounding Box Loss (EAS Loss). The EAS loss enhances the loss gradient at the final stage of angle regression for bounding boxes, addressing the challenge of further learning when the predicted box angle closely aligns with the real target box angle. This results in increased training efficiency and better alignment between training and evaluation metrics. Experimental results show that the proposed method substantially enhances the detection accuracy of ReDet and ReBiDet models. The SA 3 Network and EAS loss not only elevate the mAP of the ReBiDet model on DOTA-v1.5 to 78.85% but also effectively improve the model’s mAP under high IoU threshold conditions.

1. Introduction

In the field of optical remote sensing image target detection, the random orientation of almost all objects in the image often leads to horizontal bounding boxes encompassing a significant amount of irrelevant background when annotating detected objects. In comparison, rotated bounding boxes can greatly improve the pixel area ratio between objects and background within the selected region, better accommodating the annotation of targets in arbitrary directions. They provide accurate information about the direction, position, and size of the targets, enabling the model to perform better in target detection tasks. Existing methods for rotated object detection are mainly built upon generic object detection approaches. They redefine the representation of detection boxes, introduce additional angular dimensions for rotated detection boxes, and optimize them through distance loss. Rotated object detection methods have played a crucial role in various fields, including text detection [1,2,3,4,5,6], face recognition [7,8,9], and remote sensing image target detection [10,11,12].
However, introducing angle parameters also introduces uncertainty into the detection task, with the most prominent issue being the discontinuous boundary problem. Additionally, because the evaluation metric for the final model is based on IoU values rather than angular differences, optimizing the predicted box angles through distance loss introduces a mismatch between loss calculation and evaluation metrics. These two issues have led to suboptimal performance of existing models under high IoU threshold conditions (IoU > 0.5). Achieving higher accuracy in rotated object detection has become a significant research focus.
This section first introduces the discontinuous boundary problem, describing the origins of the discontinuity in parameterized regression. It then discusses the mismatch between the calculation of angles in the loss function and evaluation metrics. Following that, it surveys the progress of domestic and international research and concludes by introducing the research objectives of this paper.

1.1. Discontinuous Boundary Issue

Rotated bounding boxes are commonly represented in two ways: the five-parameter representation [1,5,11,12,13,14,15] and the eight-parameter representation [2,16,17,18]. Both representations can describe the position, size, and rotation angle of a rotated bounding box. The five-parameter representation is often preferred due to its simplicity, intuitive nature, and the reduced number of parameters for computation and storage. In this representation, scholars have devised various methods for representing rotated boxes, establishing a range for the angle in radians, binding it closely with the height and width of the detection box to avoid the confusion of multiple angle values caused by the angular periodicity.
Due to different definitions of the θ range, three major categories of rotated box definitions have gradually emerged in rotated object detection. The OpenCV definition [5,12,13,14,15], abbreviated as OC, sets the angle range as angle ( 0 , 90 ] , θ ( 0 , 1 2 π ] , where the acute angle between the positive x-axis and the edge defined as the width is positive. The long edge definition [5,11] defines the long edge as the width and the short edge as the height. The angle θ is the angle between the long edge width and the x-axis. Based on the angle range, it is further divided into le135 and le90 definitions. The angle ranges are set as angle [ 45 , 135 ) or θ [ π 4 , 3 π 4 ) for le135, and angle [ 90 , 90 ) or θ [ π 2 , π 2 ) for le90.
In parameterized regression-based rotational detection methods, as illustrated in Figure 1, significant differences exist between the ideal regression path and the actual regression path under the le90 definition and the counter-clockwise (CCW) representation. When regressing candidate boxes, if the ideal regression path (indicated by the dark red dashed arrow in Figure 1a) is followed, the candidate box rotates clockwise by 1 8 π to approximate the true target box. Although the predicted box has an I o U 1 with the true target box, the loss value is significantly greater than 0. This is because the angle of the predicted box changes from 1 2 π to 5 8 , increasing the discrepancy with the true target box’s angle of 3 8 π . Due to the periodic nature of angles, although the boxes are almost coincident visually, the angle difference leads to an increase in the loss value. In this situation, the model needs to regress along a more distant path, namely, counter-clockwise rotation by 7 8 π (indicated by the dashed red arrow in Figure 1c), achieving an I o U 1 and a loss value L o s s 0 to generate the final predicted box.
In summary, the discontinuous boundary problem is primarily caused by the angular periodicity and the exchangeability of the long and short sides in the five-parameter representation. The ideal regression path exceeds the predefined range, encountering boundary issues, leading to an increase in regression loss.
Faced with this challenge, the transformation of predicted boxes under the OpenCV definition involves both the exchange of width and height, as well as angle transformation, resulting in an overall complexity that may lead to less accurate predictions in such conditions [19]. In contrast, the long edge definition only necessitates considering a broad range of angle transformations when addressing the issue of discontinuous boundaries. To align with the module design in the subsequent sections of this article, the le90 definition is adopted as the angle range definition for rotated bounding boxes. Even in the case of the four-point representation method, discontinuous boundary problems may arise due to the order of corner points, a matter which will not be further elaborated upon here.

1.2. Mismatch between Loss and Evaluation Metrics

Over the years, the primary approach to improving the detection accuracy of rotated object detection models has been to propose better network architectures and more effective strategies for extracting reliable local features. However, there is a largely overlooked optimization path that has received limited research attention in the field: designing a novel loss function to replace the commonly used regression loss function, aiming to enhance the model’s detection performance. In object detection, the Intersection over Union (IoU) between the detection boxes generated by the model and the ground truth boxes is a crucial evaluation metric used to calculate the mAP. Therefore, in rotated object detection, special consideration needs to be given to minimizing angle errors and maximizing IoU values.
Most existing state-of-the-art rotated object detection models largely adopt the widely used Smooth L1 loss [20,21] as the loss function for the regression branch [22,23], following common practice in general object detection. This loss calculates the value based on the numerical differences between two bounding boxes to minimize coordinate, aspect ratio, and angle deviations. As depicted in Figure 2, the curve illustrates that the angle difference between two rectangles with an aspect ratio of 1:6 and overlapping centroids does not show a linear relationship with IoU values. When the angle difference is less than 20°, a smaller angle difference results in a larger change in IoU. Specifically, when the angle difference is within 12.96°, further reduction in the angle difference leads to a sharp increase in IoU values. However, at this point, the loss value of the model’s angle difference diminishes linearly. This scenario causes the model to lack sufficient loss gradient when the predicted box’s angle is close to the angle of the real target box. As a result, the model fails to learn how to generate more accurately angled detection boxes, preventing the detection boxes from further aligning accurately, which is clearly not an ideal outcome.
In summary, the Smooth L1 loss is not sufficiently sensitive to small angle errors, and a good local optimum based on the Smooth L1 loss may not necessarily be a local optimum for IoU. This makes it challenging to effectively guide the model to maximize the IoU values between the detection boxes and the real target boxes.

1.3. Related Work

Due to the complexity of scenes in remote sensing images, including small sizes, large aspect ratios, randomly and densely distributed objects, and arbitrarily oriented objects, rotation object detection models based on bounding box regression continue to dominate due to their higher accuracy and robustness. Parameterized regression methods are mainly divided into two branches.
Single-stage: DAL [24] proposes dynamic adaptive learning to dynamically select high-quality anchor boxes for accurate object detection, significantly reducing the number of predefined anchor boxes. R3Det [12] introduces a feature refinement module based on RetinaNet, using a progressive regression approach for fast and accurate object detection. It also suggests a skewed IoU (SkewIoU) loss to mitigate sensitivity to angle transformations in targets with large aspect ratios. RSDet [17] regresses rotated bounding boxes using four points and introduces a modulation rotation loss to address the discontinuity issue in existing losses when facing discontinuous boundaries. S2A-Net [25] introduces a Feature Alignment Module (FAM) for adaptive alignment with high-quality anchor boxes and an Orientation-aware Detection Module (ODM) for encoding orientation information, alleviating inconsistencies between the classification and regression branches. CSL [26] transforms angle prediction from a regression problem into a classification task to address the issue of discontinuous boundaries.
Two-stage: ICN [10] designs a joint image cascade network to extract multiscale semantic features and optimize regression losses. RoI Transformer [11] predicts a coarse rotated RoI based on RPN in its first stage and refines the prediction using RoI Align in the second stage, resulting in more accurate rotated RoIs and improving both efficiency and accuracy compared to rotated RPN. SCRDet [13] uses a sampling fusion network to enhance sensitivity to small objects and proposes supervised pixel attention networks and channel attention networks for joint detection of small and cluttered objects. It combines Smooth L1 loss with IoU factors to address discontinuous boundary issues in rotated bounding boxes. Gliding Vertex [18] accurately describes target orientation by sliding along each corresponding edge of a horizontal bounding box, introducing an area ratio factor between the target and its horizontal bounding box to guide the model in predicting quadrilateral detection boxes accurately. Oriented R-CNN [22] introduces a midpoint offset representation on Faster R-CNN, modifying the output parameters of the RPN regression branch from 4 to 6 to achieve rotated candidate boxes, significantly improving detection accuracy and computational efficiency. These are mainstream state-of-the-art two-stage detection models proposed in the last three years, inheriting the architecture of Faster R-CNN [21]. ReDet [23] introduces the concept of E(2)-Equivariant Steerable CNNs [27] (E(2)-CNNs) into object detection based on RoI Transformer, utilizing E2-CNNs to rewrite ResNet50 as ReResNet50. It redesigns the Rotation-invariant Region of Interest Align (RiRoIAlign) module, aligning in both channel and spatial dimensions to obtain rotation-invariant features. ReBiDet [28], based on ReDet, enhances feature fusion through the ReBiFPN module, balances the difficulty and proportion of positive samples during training using the DPRL module, and optimizes anchor box sizes and aspect ratios to further improve the model’s detection performance.
Existing object detection methods more or less involve the study of discontinuous boundary issues and mismatches between loss calculation and evaluation metrics. However, researchers have primarily concentrated on improving detection accuracy at IoU = 0.5, with limited attention to enhancing accuracy at IoU = 0.55 and above. These two issues result in suboptimal performance of current models at high IoU thresholds (greater than 0.5). As detection accuracy at IoU = 0.5 approaches a bottleneck, achieving higher precision in rotational object detection is poised to become a key focus of future research.

1.4. Goal of the Research

This paper addresses two primary issues: the discontinuous boundary problem and the mismatch between loss calculation and evaluation metrics. Solutions are proposed based on the ReBiDet model, and our work is organized into the following two aspects:

1.4.1. Discontinuous Boundary Issue

We conduct a thorough analysis of the discontinuous boundary problem, focusing on its origins in parameterized regression. Our analysis reveals that the core issue is the inaccurate localization of detection boxes due to large-angle variations during the regression of rotated bounding boxes. To address this problem, we propose the Spatially Adaptive Angle-Aware ( SA 3 ) Network, a cascaded structure designed to handle the large-angle variations caused by discontinuous boundaries. This network enhances the model’s ability to adapt to complex scenarios, thereby improving the accuracy of object detection.

1.4.2. Mismatch between Loss Calculation and Evaluation Metrics

To tackle the issue of traditional loss functions being insensitive to Intersection over Union (IoU) values, we propose the Edge-aware Skewed Bounding Box Loss (EAS Loss). This novel loss function addresses the nonlinear decay of the loss value when the predicted box’s angle closely matches the target box’s angle. By incorporating this loss function, we aim to improve the alignment between the loss calculation and IoU values, enhancing overall model performance.
Additionally, we perform ablation studies and extensive comparative experiments with state-of-the-art models on two datasets to validate the effectiveness of the proposed modules and approaches.

2. Methods

2.1. Spatially Adaptive Angle-Aware Network

2.1.1. Spatially Adaptive Angle-Aware Network Structure

The discontinuous boundary issue fundamentally results in imprecise localization of the final detection box due to the wide range of angle variations during rotated bounding box regression. Initial rotated object detection models, such as those based on Faster R-CNN [21], incorporated an angle parameter during the bounding box regression stage. However, the precision of the prediction boxes generated by a single regression was unsatisfactory. Even in general object detection, studies have shown that when dealing with candidate boxes of varying quality generated by the RPN module, a single regression function struggles to align all predicted bounding boxes accurately with the ground truth. Subsequently, the RoI Transformer [11] proposed a two-stage regression approach to produce more accurate prediction boxes. This strategy performed well under the condition of IoU = 0.5 and has been widely accepted by researchers, continuing to be adopted by state-of-the-art rotated object detection models. However, in rotated object detection, the introduction of the parameter θ exacerbates the discontinuous boundary issue. As a result, even with two regression functions, it remains challenging to achieve precise regression tasks in rotated object detection when IoU > 0.5.
Inspired by the Cascade R-CNN [29] approach, this subsubsection decomposes the regression task for rotated bounding boxes into three stages. Specifically, three dedicated regression functions { f 1 , f 2 , f 3 } are designed for each stage. These regression functions optimize the bounding boxes generated at their respective stages. The overall regression equation is as follows:
F ( x , b ) = f 3 f 2 f 1 ( x , b ) ,
Here, F ( x , b ) represents the regression equation, x denotes the image features corresponding to the candidate bounding boxes, and f 1 , f 2 , f 3 represent the regression functions for the three stages. It is important to note that each regression function f t ( t = 1 , 2 , 3 ) in the cascade is specifically trained for the optimization of the corresponding bounding box b t ( t = 1 , 2 , 3 ) at its respective stage. This approach gradually improves the generated predicted bounding boxes through coarse regression, fine-tuning, and precision refinement. The entire model learns automatically based on the training sample set { G i , B i } , requiring no further manual intervention, achieving the goal of adaptive angle regression. The regression process is illustrated in Figure 3. In an ideal scenario, the true target box has an angle of θ = 7 16 π . After being filtered by the RPN module, a proposal box is generated that overlaps the centroid of the ground truth box, matching its width and height, but with an angle of θ = 1 2 π . Initially, Stage 1 performs a large-angle transformation, coarsely regressing the proposal box to the light blue Coarse Regression box. Subsequently, Stages 2 and 3 refine the regression further, resulting in the Refined Predicted box that closely aligns with the ground truth box.
Based on the concept of adaptive angle-regression bounding boxes mentioned above, we have designed the SA 3 Network, as depicted in Figure 4. The input to the SA 3 Network is the candidate bounding boxes generated by RPN, and the output consists of the predicted bounding boxes and the classification of those boxes. The processing flow of the SA 3 Network is as follows: Stage One: Rotation Bounding Box Regression. The primary function is to regress horizontal candidate bounding boxes generated by RPN into rotated bounding boxes. The process involves filtering horizontal candidate bounding boxes through the DPRL sampler. Subsequently, RoI Align extracts the feature maps of RoIs. Following the practices of many scholars in the field [11,20,21,23,30], these feature maps are resized to 7 × 7 (a compromise between computational complexity, feature expressiveness, and simplicity of network design) to facilitate subsequent classification and bounding box regression operations. These fixed-size feature map blocks are fed into the Rotated BBox Head for convolution and fully connected operations, resulting in the regression-generated rotated predicted bounding boxes and their respective classifications; Stages Two and Three: Rotation Bounding Box Angle Fine-tuning. The main function is to fine-tune various parameters of the rotated bounding boxes generated in the first stage. The process is similar to the first stage, where rotated bounding boxes are filtered through the Rotated DPRL sampler. Next, the RiRoI Align extracts feature maps corresponding to the rotated candidate bounding boxes. These feature maps are processed through the Rotated BBox Head for convolution and fully connected operations, producing refined rotated predicted bounding boxes and their respective classifications.

2.1.2. Boundary Box Regression Class Strategy Selection

Boundary box regression class strategies include class-agnostic and class-aware strategies. Both have their advantages and disadvantages, requiring consideration of specific application scenarios.
Class-Agnostic Boundary Box Regression Strategy. From a global perspective, in remote sensing images, the orientation of objects of each category is randomly distributed. Given the widely varying aspect ratios of the bounding boxes corresponding to all objects, it is not meaningful to consider the category of objects in the feature map during the coarse regression stage. Therefore, for the regression function f 1 ( x , b ) in the first stage, a class-agnostic strategy is adopted for training the bounding box regression function. The network structure is depicted in Figure 5. The class-agnostic [31] strategy regresses bounding boxes without distinguishing specific categories during the process. This approach enhances recall rates, particularly in cases where there is incomplete annotation in the dataset, improving the model’s robustness. Importantly, it significantly reduces the computational parameter volume of the bounding box regression branch. However, the drawback is that the class-agnostic strategy regresses all objects that the network deems possibly foreground, without determining their specific category. This can lead to multiple detection boxes corresponding to a single complex-patterned object and inaccurate classification. To mitigate the impact on the final detection accuracy, class-aware strategies are employed in the subsequent two stages of regression functions.
Some studies in the general object detection field have demonstrated the effective enhancement of model detection performance using class-agnostic strategies [31]. Experimental findings indicate that, in cases of severe dataset category imbalance, the class-agnostic strategy effectively mitigates the impact of dataset incompleteness on the performance of the object detection model.
Class-Aware Boundary Box Regression Strategy. Unlike images captured from a horizontal perspective, objects of different categories in remote sensing images exhibit significant differences in the aspect ratios of their true target boxes. For instance, the aspect ratio of bridges and docks may exceed 1:10, while objects like planes and helicopters may have a ratio of 1:1. Directly applying the research experience of general object detection without considering the actual situation of optical remote sensing images may not be wise. Therefore, in the design of the regression functions f 2 , f 3 ( x , b ) for the fine-tuning and precision refinement stages, a class-aware strategy is employed. The network structure is shown in Figure 6.
When using a class-aware strategy to regress bounding boxes, the regression network individually infers each category by traversing all categories, computing bounding boxes for each category in the feature map block, and outputting the coordinates of bounding boxes for all categories. Then, based on the output results of the classification network branch, the bounding boxes corresponding to the category of the classification network branch are indexed, labeled as detection boxes, and other irrelevant category bounding boxes are discarded, obtaining the final detection results. The advantage of this approach is that it allows the model to undergo fine training for each category, serving as the basis for fine regression and refinement learning, ultimately enabling the model to generate accurate detection boxes for each category’s object. The drawback is that for a dataset with annotated N categories of objects, the computational complexity of the regression network branch is N times that of the class-agnostic strategy, making it less efficient. However, for this project, improving detection accuracy through algorithm design takes precedence, while also considering a reduction in computational complexity. As per the empirical knowledge of generative artificial intelligence, the size of the computed parameter quantity is positively correlated with the actual performance of the model. In experiments, it was observed that although the class-aware strategy incurs a higher computational cost, it does not necessarily achieve better detection performance when the number of categories in the dataset is imbalanced or the total number of targets is small.
In summary, when using the SA 3 Network, one should consider whether the number of targets in each category in the dataset is balanced, the total number of targets, and other practical issues. A comprehensive judgment should be made regarding whether to adopt a class-aware or class-agnostic strategy during the fine regression and refinement stages of the SA 3 Network. Inappropriate strategy selection may result in a significant reduction in the model’s detection performance.

2.2. Edge-Aware Skewed Bounding Box Loss

A persistent issue in the design of regression losses for rotated object detection is the inconsistency between model training metrics and final evaluation metrics. This inconsistency acts as a bottleneck, necessitating the design of a loss function specifically tailored for rotated object detection. Such a loss function must consider the unique characteristics of this task, especially in terms of the Intersection over Union (IoU), to enhance the model’s sensitivity to changes in angles.
To address this challenge, we introduce a novel regression loss, named the EAS loss. This loss function takes into account the IoU value between the predicted bounding box and the ground truth. The EAS loss effectively mitigates the inconsistency between training and evaluation metrics, enabling the model to accurately regress angles, particularly when the predicted box is close to the ground truth.

2.2.1. EAS Loss Design

To enhance the sensitivity to small angle deviations and improve metric consistency, a natural idea is to use IoU when regressing the angle θ . In this case, the loss calculation is defined as in Equation (2):
Loss I o U = log ( eps + IoU ) ,
However, this introduces a new problem: when the centroid of the predicted box is far from the centroid of the ground truth box, the IoU value will be very small, leading to slow angle regression by the model. Assuming the aspect ratio of the box is 1:6, as shown in Figure 7, even when the centroid of the predicted box and the ground truth box coincide, the slope of the Loss I o U value becomes flat when the angle difference θ exceeds 60°. This reduces the regression efficiency of the model for larger angle differences.
To address this, we introduce an IoU factor into the angle regression loss L θ * calculation, as shown in Equation (3):
L θ = L θ * log ( eps + IoU ) ,
Here, L θ * adopts the calculation method of the Smooth L1 loss, leading to the ideal EAS loss calculation Equation (4):
Loss EAS ( d θ , t θ , β , γ ) = 0.5 · | d θ t θ | 2 β γ · log ( eps + IoU ) , if | d θ t θ | < β | d θ t θ | 0.5 · β γ · log ( eps + IoU ) , otherwise ,
Here, eps = 10 6 , and β and γ are adjustable variables. When β = 1 and γ = 1 / 9 , the relationship between Loss EAS and the angle difference θ is depicted in Figure 8.
When the centroids of the predicted box and the ground truth box do not coincide, the EAS loss, incorporating the IoU factor in the angle regression loss component, prevents ineffective rotations of the predicted box. When the centroids of the predicted box and the ground truth box do coincide, the IoU factor allows the model to quickly and accurately learn to generate more precise predicted boxes as the predicted and ground truth angles approach each other. This makes it an ideal angle regression loss function.
The next issue is how to calculate the IoU value. Here, the IoU between the rotated bounding boxes is termed SkewIoU to distinguish it from horizontal IoU. The definition of SkewIoU is the same as IoU, as shown in Equation (5):
SkewIoU ( P , G ) = Area ( P G ) Area ( P G ) ,
where P and G are two rotated bounding boxes. To compute their intersection and union areas accurately, skewed bounding boxes are treated as polygons, and the convex polygon intersection calculation method [5] is employed to precisely calculate SkewIoU. This process involves sorting and combining the coordinates of the vertices and intersection points of the rotated boxes, followed by the application of the triangulation method to calculate the area of the intersection region. The overall computation is relatively complex, with sorting and triangulation being the primary time-consuming factors. In practical applications, libraries such as OpenCV and Shapely are often employed to calculate SkewIoU. The specific implementation and optimizations in these libraries can impact performance, but the fundamental principles of computational complexity remain unchanged.
In general object detection, the IoU loss has long been a focus for effectively mitigating the inconsistency between evaluation metrics (dominated by IoU) and regression loss calculation methods. However, in rotated object detection, due to the computationally expensive nature of SkewIoU calculation, it is primarily used for validation and evaluation and has not been widely adopted in loss functions.

2.2.2. Gaussian Transformation for Approximate IoU Calculation

Recent studies, such as PIoU [32], projection IoU [33], GWD [34], and KLD [35], have explored methods to simulate the approximate SkewIoU loss. Among these, GWD and KLD introduced a Gaussian modeling approach, simplifying the complex SkewIoU calculation into a more efficient process. However, their methods involve nonlinear transformations and hyperparameters in the final loss function design using Gaussian distribution distance metrics, making them not fundamentally SkewIoU. KFIoU [36] is a loss function based on the approximation of SkewIoU and center-point distance, but its performance in the proposed model in this paper is not ideal. As shown in Table 1, after incorporating the KFIoU loss, the model’s AP50, AP75, and mAP all experienced varying degrees of decline. In the original paper, KFIoU achieved satisfactory results with backbone networks like ResNet-152 with large parameter sizes. In contrast, the model in this paper utilizes the ReResNet-50 backbone network with relatively fewer computational parameters, which might be one of the reasons for the performance degradation.
Despite KFIoU’s poor performance in ReBiDet, a simpler and more efficient method for approximating SkewIoU is worth considering. This method converts two rotated bounding boxes into Gaussian distributions, representing the overlapping region with a simpler mathematical formula and calculating the area of the minimum enclosing rectangle of this region. This approach significantly simplifies the original geometric computation while maintaining a high approximation to the exact IoU value [36]. This method adheres to the SkewIoU calculation process, is mathematically rigorous, and does not introduce additional hyperparameters. Although the accuracy of the SkewIoU values obtained is relatively low, precise SkewIoU calculation is not strictly necessary in loss functions; minor errors are tolerable as long as the trend of the approximated SkewIoU matches the true value. Below are the basic steps and derivations for approximating SkewIoU using the Gaussian transformation method [36]:
Transforming Rotated Rectangular Boxes into Gaussian Distributions. To begin, the rotated rectangular boxes are transformed into Gaussian distributions G ( μ , Σ ) . Each rectangular box is represented by two parameters: the covariance matrix Σ and the center coordinates μ . For a rotated rectangular box B ( x , y , w , h , θ ) , μ = x y , and the covariance matrix is calculated as shown in Equation (6):
Σ = R Λ R T ,
where R = cos θ sin θ sin θ cos θ ,   Λ = 1 4 w 2 0 0 h 2 2 . The final form of the covariance matrix Σ is given in Equation (7):
Σ = w 2 4 cos 2 θ + h 2 4 sin 2 θ ( w 2 h 2 ) 4 cos θ sin θ ( w 2 h 2 ) 4 cos θ sin θ w 2 4 sin 2 θ + h 2 4 cos 2 θ ,
The transformed Gaussian distribution is illustrated in Figure 9.
Calculating Overlapping Area of Two Gaussian Distributions. The overlapping area of two Gaussian distributions is computed by multiplying the Gaussian distributions of the predicted box P and the true target box G using the Kalman filter’s multiplication rule. Specifically, the predicted box’s Gaussian distribution G P ( μ P , Σ P ) is treated as the predicted value, and the true target box’s Gaussian distribution G G ( μ G , Σ G ) is treated as the observed value. This yields the approximate Gaussian distribution G I ( μ I , Σ I ) for the overlapping region I, as shown in Equation (8):
α G I ( μ I , Σ I ) = G P ( μ P , Σ P ) · G G ( μ G , Σ G ) ,
Here, α G I ( μ I , Σ I ) does not have a probability sum of 1 and is not a standard Gaussian distribution. The coefficient α can be expressed by Equation (9):
α = G α ( μ G , Σ P + Σ G ) = 1 det ( 2 π ( Σ P + Σ G ) ) e 1 2 ( μ P μ G ) T ( Σ P + Σ G ) 1 ( μ P μ G ) ,
When μ P μ G 0 , α can be approximated as a constant. Since the EAS loss is computed separately for the centroids of the bounding boxes, we can consider the scenario where the centroids of the predicted box P and the ground truth box G coincide, i.e., μ P = μ G . In this case, the parameters μ I and Σ I of the Gaussian distribution G I ( μ I , Σ I ) for the overlap region I are calculated using Equation (10):
μ I = μ P + K ( μ G μ P ) , Σ I = Σ P K Σ P ,
where K = Σ P ( Σ P + Σ G ) 1 . When the central points of the predicted box P and the true target box G are close to overlapping ( μ P = μ G ), the center point μ I of the Gaussian distribution for the overlap region I coincides with the central points of the predicted box P and the true target box G, as shown in Figure 10. The covariance matrix Σ I of the overlapping area I can be approximated using Equation (11):
Σ I = Σ P 1 Σ P Σ P + Σ G = Σ P Σ G Σ P + Σ G ,
It can be observed that the covariance Σ I of the overlapping area I is influenced by the covariance Σ P and Σ G of the Gaussian distributions of the predicted box P and the true target box G.
Calculating Areas of Externally Circumscribed Rectangles. The areas of the minimum enclosing rectangles for the predicted box P, the ground truth box G, and the overlap region I are calculated using their respective Gaussian distributions. Since the dimensions of the predicted box P and the ground truth box G are known, S P ( Σ P ) and S G ( Σ G ) can be directly calculated. The area S I ( Σ I ) can be conveniently derived from the covariance matrix Σ I , as shown in Equation (12):
S I ( Σ I ) = 2 2 i eig ( Σ I ) = 4 · Σ I 1 / 2 = 4 · | Σ I | ,
Calculating the Approximate SkewIoU. The approximate SkewIoU ( SkewIoU approx ) is calculated as follows:
SkewIoU approx = S I ( Σ I ) S P ( Σ P ) + S G ( Σ G ) S I ( Σ I ) ,
However, S I ( Σ I ) is not the exact area of the overlap of the predicted box P and the true target box G. Since α G I ( μ I , Σ I ) is not a standard Gaussian distribution and α is treated as a constant in the calculation, Xue Yang et al. derived that the upper bound of · S I ( Σ I ) S P ( Σ P ) + S G ( Σ G ) S I ( Σ I ) is 1 3 [36]. Using this upper bound, a linear transformation is applied to expand its value range to [0, 1], resulting in the approximate SkewIoU:
SkewIoU approx = 3 · S I ( Σ I ) S P ( Σ P ) + S G ( Σ G ) S I ( Σ I ) ,
This approximation method for SkewIoU approx demonstrates high consistency with the true SkewIoU plain while being an ideal method for practical applications [36]. Despite the complexity of the derivation, the implementation is straightforward in the EAS loss function, where SkewIoU approx is used as a factor for angle regression without introducing additional hyperparameters and not participating in the backpropagation process.
The EAS loss function is given by:
Loss EAS ( d θ , t θ , β , γ ) = 0.5 · | d θ t θ | 2 β γ · log ( ε + SkewIoU approx ) if | d θ t θ | < β | d θ t θ | 0.5 · β γ · log ( ε + SkewIoU approx ) otherwise ,
where ε = 10 6 , and β and γ are adjustable parameters.

3. Experimental Results

3.1. Datasets

3.1.1. DOTA-v1.5 Dataset

DOTA [37,38], released by Wuhan University in January 2018, currently comprises three versions: DOTA-v1.0, DOTA-v1.5, and DOTA-v2.0.
DOTA-v1.0 [37] includes 2806 images sourced from various platforms, such as Google Earth, JL-1, and GF-2. This dataset contains 188,282 annotated objects across 15 different categories: Plane (PL), Baseball Diamond (BD), Bridge (BR), Ground Track Field (GTF), Small Vehicle (SV), Large Vehicle (LV), Ship (SH), Tennis Court (TC), Basketball Court (BC), Storage Tank (ST), Soccer-Ball Field (SBF), Roundabout (RA), Harbor (HA), Swimming Pool (SP), and Helicopter (HC). The annotations are quadrilaterals defined by four points { x 1 , y 1 ; x 2 , y 2 ; x 3 , y 3 ; x 4 , y 4 } . However, DOTA-v1.0 misses many small-sized objects (approximately 10 pixels or smaller), leading to incomplete annotations. This limitation may result in inaccurate evaluations of object detection model performance [28].
DOTA-v1.5 [38], built upon DOTA-v1.0, expands the number of annotations to 402,089. It also includes additional annotations for the small objects omitted in DOTA-v1.0 and introduces a new category, Container Crane (CC), making it a more challenging dataset. The dataset presents the following difficulties: (1) the pixel size of images varies significantly, ranging from 800 × 800 to 4000 × 4000 , complicating training on certain GPUs; (2) a high proportion of objects are small, with 98% of objects being smaller than 300 pixels and 57% smaller than 50 pixels [37], resulting in a substantial scale variation between tiny and large objects, complicating detection.
Due to the limited research published using DOTA-v2.0, cross-performance comparisons are challenging; thus, this paper does not utilize DOTA-v2.0. Instead, DOTA-v1.5 is employed for testing and analyzing the proposed methods, as it addresses the shortcomings of DOTA-v1.0 and remains the most commonly used public object detection dataset in the remote sensing field. It is noteworthy that the creators of the DOTA dataset have not publicly released the annotations for the test set, requiring researchers to upload their detection results for evaluation, thereby restricting comprehensive analysis of experimental results.

3.1.2. DFShip Dataset

The DFShip dataset is a fine-grained optical remote sensing ship dataset released by the Big Data and Decision-Making (National) Laboratory for the 2023 National Big Data and Computational Intelligence Challenge [39]. It comprises 41,495 images, all annotated with ship targets. The training set includes 30,285 images with corresponding target annotation files, totaling 120,605 annotated targets across 133 fine-grained categories. This makes DFShip the most detailed and extensive fine-grained ship dataset currently available, posing a significant challenge to the performance of object detection models. Both the preliminary and final test sets consist of 11,210 images, each presenting varying levels of detection difficulty. The organizers did not provide annotation files for the test sets; participants must test locally and upload the packaged test results to a designated server for validation. Since the organizers did not provide a specific name for this dataset, we refer to it as DFShip in this paper.

3.2. Setup

The experimental setup closely follows the approach described in the ReBiDet paper, with several key upgrades. The GPU configuration has been enhanced from 2 NVIDIA GTX 3090 Ti to 2 NVIDIA GTX 4090 GPUs. Additionally, the software stack has been updated: CUDA has been upgraded from version 11.8 to 12.0, PyTorch from version 1.11.0 to 1.13.1, torchvision from version 0.12.0 to 0.14.1, and the MMRotate framework from version 0.3.2 to 0.3.4. These updates ensure improved compatibility and facilitate the reproducibility of experiments across different environments.
During training, data augmentation techniques, such as horizontal and vertical flips, were applied. The batch size per GPU was set to 2, resulting in a total batch size of 4. The network was optimized using the Stochastic Gradient Descent (SGD) algorithm with a momentum of 0.9 and a weight decay of 0.0001. For the Region Proposal Network (RPN), the IoU threshold for positive samples was set to 0.7. The horizontal box Non-Maximum Suppression (NMS) threshold was set to 0.7, while the rotation box NMS threshold was set to 0.1.
The DOTA-v1.5 dataset consists of a total of 2806 images, with 1411 images in the training set, 458 images in the validation set, and 937 images in the test set. To ensure fairness in comparison experiments, we followed the practices of many scholars in the field and processed the data accordingly. Since the images in the dataset have varying sizes, the original images from the DOTA-v1.5 dataset were cropped to a size of 1024 × 1024 pixels with a stride of 824 pixels. After cropping, the DOTA-v1.5 dataset contains 15,749 images in the training set, 5297 images in the validation set, and 10,833 images in the test set. Both the training and validation sets were used for model training. Additionally, we performed multiple scale augmentation on the dataset, resizing the original 2806 images to three scales: 0.5, 1.0, and 1.5. These resized images were then cropped to 1024 × 1024 pixels with a stride of 524 pixels, resulting in a final dataset of 416,651 images for training and 71,888 images for testing. The model was trained for 12 epochs, with an initial learning rate of 0.01, and it was divided by 10 at epochs 9 and 11.
The DOTA-v1.5 dataset consists of a total of 2806 images, with 1411 images in the training set, 458 images in the validation set, and 937 images in the test set. To ensure fairness in comparative experiments, we followed established practices and processed the data accordingly. Given the varying sizes of images in the dataset, the original images were cropped to a size of 1024 × 1024 pixels with a stride of 824 pixels. After cropping, the DOTA-v1.5 dataset was expanded to include 15,749 images in the training set, 5297 images in the validation set, and 10,833 images in the test set. Both the training and validation sets were utilized for model training. Additionally, multiple scale augmentations were performed, resizing the original 2806 images to three scales: 0.5, 1.0, and 1.5. These resized images were cropped to 1024 × 1024 pixels with a stride of 524 pixels, resulting in a final dataset of 416,651 images for training and 71,888 images for testing. The model was trained for 12 epochs, with an initial learning rate of 0.01, which was reduced by a factor of 10 at epochs 9 and 11.
For the DFShip dataset, the original images are typically 1024 × 1024 in size and do not require additional cropping. Since the competition organizer’s validation server only provides mAP at an IoU threshold of 0.5, the original training set was randomly split in a 4:1 ratio for this study. This resulted in a training set with 24,228 images and 96,644 annotated targets, and a test set with 6057 images and 23,961 annotated targets. The model was trained for 36 epochs, with an initial learning rate of 0.01, which was reduced by a factor of 10 at epochs 24 and 33.

3.3. Results and Analysis

3.3.1. Ablation Experiments of SA 3 Network

To evaluate the effectiveness of the proposed SA 3 Network, ablation experiments were conducted using ReDet and ReBiDet as baseline models. These experiments aimed to assess the impact of incorporating class-aware and class-agnostic bounding box regression strategies.
In these ablation experiments, the SA 3 Network was integrated into both ReDet and ReBiDet models to objectively evaluate its contribution. Due to the absence of ground truth annotations for the test set of the DOTA dataset, the trained models’ results were packaged and uploaded. The detection metrics were then verified by the DOTA-v1.5 dataset authors using their designated servers.
Table 2 presents a performance comparison of the ReDet and ReBiDet models with and without the SA 3 Network. The integration of the SA 3 Network results in improved detection performance for both models. The mean average precision (mAP) reported in the table is computed using the COCO evaluation method, which averages the average precision (AP) values across all possible IoU thresholds. AP50 and AP75 refer to the AP at IoU thresholds of 0.5 and 0.75, respectively.
The results indicate that integrating the SA 3 Network leads to improvements in mAP, AP50, and AP75 for both the ReDet and ReBiDet models. For ReDet, the class-aware strategy increases AP50 by 0.92% and AP75 by 0.84%. The class-agnostic strategy, however, shows even more significant improvements, with AP50 increasing by 1.28% and AP75 by 2.38%. Similarly, for ReBiDet, the class-aware strategy results in AP50 increasing by 0.68% and AP75 by 0.4%, while the class-agnostic strategy yields an increase of 1.33% in AP50 and 1.59% in AP75.
These results demonstrate that the SA 3 Network significantly enhances detection accuracy, particularly at higher IoU thresholds. This improvement underscores the network’s ability to optimize model precision for predicting rotated bounding boxes.
Despite the expectation that the class-aware strategy would provide more accurate results due to its tailored approach for each object class, the experimental outcomes show the opposite. The class-agnostic strategy performs better, which can be attributed to the imbalanced class distribution in the DOTA-v1.5 dataset. Some classes, such as Container Crane (CC), have very few instances compared to others like Small Vehicle (SV), which has a large number of instances. Figure 11 illustrates this imbalance, with CC having only 283 instances and SV having 295,272 instances. In such cases, the class-agnostic strategy is more suitable.
To further investigate the effectiveness of the SA 3 Network and the rationale behind strategy selection, we performed random rotation and offline multi-scale augmentations on the DOTA-v1.5 dataset. The class distribution after augmentation, shown in Figure 12, remains uneven but with an increased minimum instance count for Container Crane (CC), which rises to 1719. This theoretically enhances training effectiveness.
Table 3 displays the ablation experiment results using the augmented DOTA-v1.5 dataset. With the SA 3 Network, the class-aware strategy shows significant improvements over the ReDet baseline, with AP50 increasing by 1.00% and AP75 by 1.67%. The class-agnostic strategy results in AP50 increasing by 0.68% and AP75 by 1.64%. For ReBiDet, the class-aware SA 3 Network leads to an increase of 0.54% in AP50 and 0.85% in AP75, contrasting with the previous results where the class-agnostic strategy showed higher effectiveness.
The augmented dataset, despite its imbalances, now provides sufficient instances for each class, making the class-aware strategy more effective. Therefore, while the class-agnostic strategy of the SA 3 Network is more suitable for datasets with significant class imbalances, the class-aware strategy proves advantageous for datasets with adequate instances for each class, even if imbalances exist.

3.3.2. Ablation Experiments of EAS Loss

Initially, we performed ablation experiments using ReDet as the baseline model to assess the effectiveness of the EAS loss function. Table 4 compares the performance of the ReDet and ReDet + SA 3 models, both with and without the EAS loss. This preliminary verification indicates that the proposed EAS loss function improves the model’s detection performance. The evaluation metrics are defined as previously described. The results reveal that both baseline models, ReDet and ReDet + SA 3 , exhibit varying degrees of improvement in mAP, AP50, and AP75 with the adoption of the EAS loss. Specifically, ReDet shows a 0.60% increase in AP50 and a 0.18% increase in AP75. For ReDet + SA 3 , the class-aware strategy significantly enhances performance, with AP50 increasing by 1.10% and AP75 by 1.60%.
Further ablation experiments were conducted using ReBiDet as the baseline model to corroborate the effectiveness of the EAS loss function. Table 5 presents the performance comparison of the ReBiDet + SA 3 models with and without EAS loss. Similar to the ReDet results, the ReBiDet + SA 3 model exhibits notable improvements under the class-aware strategy when utilizing EAS loss: AP50 increases by 1.11%, AP75 by 1.38%, and mAP by 1.63%. The class-agnostic strategy shows smaller improvements, with AP75 and mAP increasing by 0.80% and 0.35%, respectively, while AP50 decreases by 0.62%, consistent with the results obtained with the ReDet model. This observed phenomenon is not coincidental. The EAS loss function facilitates more comprehensive training of the model’s bounding box regression branch, whereas the class-agnostic strategy treats all categories as a single class, which can be less effective when training samples are insufficient. This fundamental conflict between the mechanisms explains the reduced generalization capability of the SA 3 Network under the class-agnostic strategy when influenced by EAS loss.
These experimental results indicate that the EAS loss function enhances the training effectiveness of the model’s bounding box regression branch and optimizes the SA 3 Network, particularly improving the learning efficiency of the class-aware strategy. This ensures that the model can be effectively trained even with imbalanced categories and fewer samples.
We then evaluated the performance of the EAS loss function on a dataset with a larger number of training samples by applying random rotation and offline multi-scale augmentation to the DOTA-v1.5 dataset.
Initially, ablation experiments were conducted using ReDet + SA 3 as the baseline model. Given that the class-agnostic strategy performs relatively poorly on the unaugmented DOTA-v1.5 dataset, we employed the class-aware strategy for SA 3 in this case. Table 6 compares the performance of the ReDet + SA 3 models with and without EAS loss. The results show slight improvements across all performance indicators: AP50 increased by 0.25%, AP75 by 0.16%, and mAP by 0.36%. These improvements are significantly less than those observed with the unaugmented DOTA-v1.5 dataset.
Similarly, ablation experiments using ReBiDet + SA 3 as the baseline model were conducted, maintaining the class-aware strategy for SA 3 as explained above. Table 7 presents a performance comparison of the ReBiDet + SA 3 models with and without EAS loss. As with the ReDet + SA 3 model, the EAS loss yields minor improvements: AP50 increased by 0.35%, AP75 decreased by 0.12%, and mAP increased by 0.55%. These improvements are again notably smaller compared to those seen with the unaugmented DOTA-v1.5 dataset.
The diminished advantage of the EAS loss function with an increased number of effective training samples is consistent with our previous observations. This effect is evident in both the ReDet + SA 3 and ReBiDet + SA 3 models, with the latter exhibiting stronger feature extraction capabilities. The results suggest that enhanced feature extraction and sample selection strategies cannot maintain the significant advantage of EAS loss. Nevertheless, the performance of both ReDet/ReBiDet + SA 3 models still benefits from the use of EAS loss.
In summary, the EAS loss function effectively improves the learning efficiency of the model’s regression branch for bounding box angle regression. It ensures adequate training even with imbalanced categories and fewer samples. Specifically, the class-aware strategy of the SA 3 Network benefits significantly from the EAS loss, enhancing the performance metrics, such as AP50, AP75, and mAP. In scenarios with insufficient training samples for certain categories, the SA 3 Network does not need to choose between class-aware and class-agnostic strategies. As the number of effective training samples increases and each category has sufficient instances, the advantage of the EAS loss diminishes but remains relative to the Smooth L1 loss. These results validate the effectiveness of the EAS loss function in various scenarios.

3.3.3. Multi-IoU Threshold Comparison Experiment on the DFShip Dataset

The DOTA series datasets, depending on the version, include 15 or 16 categories, such as airplanes and cars, with ships being one of the categories. To further demonstrate the effectiveness of the proposed solutions, design concepts, and research methods in addressing ship detection issues, we conducted validation experiments on the DFShip dataset, which specializes in ship detection. This subsection provides a horizontal comparison of the ReBiDet + SA 3 + EAS model proposed in this paper with other state-of-the-art models on the DFShip dataset. We selected and reproduced several leading models from the DOTA series datasets for this comparison, including CSL [26], R3Det [12], S2A-Net [25], Oriented R-CNN [22], and our baseline model ReDet [23]. These models are among the most advanced in the field. Due to the unavailability of annotated test data from the official source, we split the original training set into training and test sets in a 4:1 ratio for the experiments, focusing on the precision of generated bounding boxes, specifically mAP at high IoU thresholds.
Table 8 presents the experimental results. At low threshold conditions, such as IoU = 0.5, the performance advantage of our proposed model over other advanced models is not significant. This is partly due to the dataset characteristics, which include clear images, accurate annotations, and a large number of training samples, all of which generally enhance the detection performance of the models. However, under high threshold conditions, such as IoU = 0.75, our model demonstrates a clear advantage with a detection accuracy of 98.28%, outperforming all other models in the comparison and surpassing the baseline model ReDet by 2.22%. The detection accuracy at IoU = 0.75 is only 0.56% lower than at IoU = 0.5.
At an even higher threshold of IoU = 0.85, the detection accuracy of ReBiDet + SA 3 + EAS remains robust at 93.18%, while the performance of all other models significantly declines. At IoU = 0.95, ReBiDet + SA 3 + EAS maintains a detection accuracy of 19.18%, whereas the other models’ accuracies drop to single digits. Figure 13 visually presents the comparison results from Table 8. These results confirm that our model not only achieves high detection accuracy but also generates more precise bounding boxes. This validates our research approach and demonstrates the effectiveness of the design methods and strategies employed.

3.3.4. Results

This subsubsection presents a comprehensive comparison of the ReDet and ReBiDet models, enhanced with the SA 3 Network and EAS loss, against several state-of-the-art models on the DOTA-v1.5 dataset. The objective is to evaluate the effectiveness of the proposed methods by performing a horizontal comparison with other advanced models published in recent years. All comparison models are sourced from reputable journals and conferences, and their performance metrics are cited directly.
Table 9 provides a detailed comparison of the ReDet and ReBiDet models integrated with the SA 3 Network and EAS loss against other state-of-the-art models on the DOTA-v1.5 dataset. On the unaugmented DOTA-v1.5 dataset, the ReBiDet + SA 3 + EAS model, using the class-aware strategy, surpasses all compared models with a mean average precision (mAP) of 71.28%. This is slightly higher than the ReBiDet + SA 3 model utilizing the class-agnostic strategy. Notably, ReBiDet + SA 3 + EAS ranks within the top three for average precision (AP) values across 15 subcategories and achieves first place in 10 subcategories. It excels in categories with a high aspect ratios, such as bridges (BR) and harbors (HA), as well as in categories with low aspect ratios, such as small vehicles (SV) and large vehicles (LV), which share some characteristics with ships.
When evaluated on the augmented DOTA-v1.5 dataset, which includes random rotation and offline multi-scale enhancements, the ReBiDet + SA 3 + EAS model demonstrates superior performance, achieving a mAP of 78.85%. It ranks in the top three for accuracy in 10 subcategories, including ships (SH). These results underscore the effectiveness of the proposed approach in significantly enhancing the target detection performance of existing models in optical remote sensing images.

4. Discussion

The experimental results confirm that the proposed SA 3 Network demonstrates excellent precision and localization performance in detecting rotated objects. The EAS loss function, designed to enhance edge perception of inclined bounding boxes, significantly improves the model’s learning efficiency in rotated object detection tasks. This leads to more accurate localization of detection boxes, highlighting the effectiveness of the design approach and methodology presented in this paper.
The SA 3 Network incorporates a cascaded regression branch. Initially, a coarse regression branch converts horizontal bounding boxes generated by the Region Proposal Network (RPN) into rotated bounding boxes based on the features of the detected objects. Depending on the task scenario, the network employs either a class-aware or class-agnostic strategy. Fine regression and refinement branches further optimize the angle parameters of the rotated bounding boxes, thereby enhancing the fitting accuracy of the final detection boxes.
The inclusion of EAS loss in the angle regression branch introduces an Intersection over Union (IoU) factor, which mitigates the mismatch between traditional loss functions and evaluation metrics. This results in overall improvements across various detection accuracy metrics. The EAS loss adjusts the function gradient of the angle regression branch, addressing the issue of limited learning when the predicted box angle is close to the true target angle. Consequently, the model becomes more sensitive towards the end of the angle regression branch, producing more accurate predicted box angles and significantly improving the AP75 accuracy metric.
Figure 14 illustrates the detection results for the same port remote sensing image. In Figure 14a, the detection results of the ReBiDet model are shown. The ReBiDet model struggles with accurate bounding box localization for elongated objects, such as ships, due to issues with discontinuous boundaries. Figure 14b displays the results after integrating the SA 3 Network. It is evident that ReBiDet + SA 3 generates more precise bounding boxes for ships and docks, and improves detection confidence for the ships in the upper-right corner and the dock in the middle-lower part of the image. Additionally, the improved localization accuracy leads to a higher detection rate and a lower false detection rate. After incorporating the SA 3 Network, the harbor in the middle-upper part of the image is correctly detected, and previously misidentified objects on the harbor are now accurately classified as irrelevant.
However, the proposed method has some limitations concerning computational parameters and inference speed. The ReBiDet + SA 3 + EAS model increases the number of computational parameters by 13.99 M compared to the baseline ReBiDet model. While the EAS loss itself does not significantly contribute to this increase, the additional parameters primarily result from the SA 3 Network. This increase in computational parameters leads to an additional 3.5 ms per image on an RTX4090 platform, corresponding to a decrease of 1.8 FPS. Although these computational trade-offs are acceptable in many scenarios given the improvements in detection accuracy, reducing computational parameters and enhancing inference speed while maintaining high detection accuracy remains a critical area for future research.
In conclusion, the methods proposed in this paper are highly effective in improving the detection accuracy of randomly oriented objects, particularly those with elongated contour features, in remote sensing applications. These advancements provide valuable insights for enhancing ship detection capabilities in optical remote sensing images.

5. Conclusions

This paper addresses the challenges associated with significant deviations in final detection boxes due to a wide range of angles, as well as the imprecision inherent in traditional angle regression losses. To tackle these issues, we propose the SA 3 Network and EAS loss. The SA 3 Network employs a hierarchical regression structure that includes coarse, fine, and refinement stages to progressively optimize the angle parameters of rotated bounding boxes. The EAS loss introduces the SkewIoU factor, calculated using Gaussian transformation, to enhance the precision of angle regression losses. This approach improves both training efficiency and model performance, particularly under high IoU threshold conditions.
Experimental results validate the effectiveness of the SA 3 Network and EAS loss. The proposed methods significantly improve detection accuracy, especially for rotated objects in optical remote sensing images.
Future work will focus on enhancing the interpretability of rotation-equivariant convolutional neural networks. Understanding the operational mechanisms of these networks presents a significant research opportunity. Our goal is to gain a more intuitive understanding of the features extracted by these networks. We aim to explore techniques for visualizing rotation-equivariant features, which will facilitate a deeper analysis of their limitations and potential improvement strategies.

Author Contributions

Methodology, Z.Y. and J.F.; software, Z.Y.; supervision, Z.L. and Y.X.; writing—original draft, Z.Y.; writing—review and editing, Z.Y.; visualization, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The DOTA dataset was publicly released by Wuhan University in November 2017 and includes complete sets of training, validation, and testing images, as well as annotation files for the training and validation sets. It can be downloaded at this address: https://captain-whu.github.io/DOTA/dataset.html (accessed on 29 May 2024). Since the downloaded package does not include annotation files for the testing set, users need to submit their model’s inference results to the Wuhan University official Evaluation Server to obtain the accuracy of the inference results, at this address: https://captain-whu.github.io/DOTA/evaluation.html (accessed on 29 May 2024). After this article is published, our code and other related materials will also be released.

Acknowledgments

This work was supported by the Institute of Systems Engineering, Academy of Military Sciences, and we would like to express our gratitude to Yongqiang Xie and Zhongbo Li, for their support and guidance throughout the project. Their contributions have been invaluable to the success of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoUIntersection over Union
SA 3 Spatial Adaptive Angle-Aware
EAS LossEdge-aware Skewed Bounding Box Loss
RPNRegion Proposal Network

References

  1. Jiang, Y.; Zhu, X.; Wang, X.; Yang, S.; Li, W.; Wang, H.; Fu, P.; Luo, Z. R2CNN: Rotational Region CNN for Orientation Robust Scene Text Detection. arXiv 2017, arXiv:abs/1706.09579. [Google Scholar]
  2. Liao, M.; Shi, B.; Bai, X. TextBoxes++: A Single-Shot Oriented Scene Text Detector. IEEE Trans. Image Process. 2018, 27, 3676–3690. [Google Scholar] [CrossRef]
  3. Liao, M.; Zhu, Z.; Shi, B.; Xia, G.s.; Bai, X. Rotation-Sensitive Regression for Oriented Scene Text Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5909–5918. [Google Scholar] [CrossRef]
  4. Liu, X.; Liang, D.; Yan, S.; Chen, D.; Qiao, Y.; Yan, J. FOTS: Fast Oriented Text Spotting with a Unified Network. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5676–5685. [Google Scholar] [CrossRef]
  5. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-Oriented Scene Text Detection via Rotation Proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef]
  6. Zhou, X.; Yao, C.; Wen, H.; Wang, Y.; Zhou, S.; He, W.; Liang, J. EAST: An Efficient and Accurate Scene Text Detector. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2642–2651. [Google Scholar] [CrossRef]
  7. Huang, C.; Ai, H.; Li, Y.; Lao, S. High-Performance Rotation Invariant Multiview Face Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 671–686. [Google Scholar] [CrossRef] [PubMed]
  8. Rowley, H.A.; Baluja, S.; Kanade, T. Rotation invariant neural network-based face detection. In Proceedings of the 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231), Santa Barbara, CA, USA, 25 June 1998; pp. 38–44. [Google Scholar] [CrossRef]
  9. Zhou, L.F.; Gu, Y.; Wang, P.S.P.; Liu, F.Y.; Liu, J.; Xu, T.Y. Rotation-Invariant Face Detection with Multi-task Progressive Calibration Networks. In Pattern Recognition and Artificial Intelligence; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 513–524. [Google Scholar]
  10. Azimi, S.M.; Vig, E.; Bahmanyar, R.; Körner, M.; Reinartz, P. Towards Multi-class Object Detection in Unconstrained Remote Sensing Imagery. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  11. Ding, J.; Xue, N.; Long, Y.; Xia, G.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2844–2853. [Google Scholar] [CrossRef]
  12. Yang, X.; Liu, Q.; Yan, J.; Li, A. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. Proc. Aaai Conf. Artif. Intell. 2021, 35, 3163–3171. [Google Scholar] [CrossRef]
  13. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8231–8240. [Google Scholar]
  14. Yang, X.; Sun, H.; Fu, K.; Yang, J.; Sun, X.; Yan, M.; Guo, Z. Automatic Ship Detection in Remote Sensing Images from Google Earth of Complex Scenes Based on Multiscale Rotation Dense Feature Pyramid Networks. Remote. Sens. 2018, 10, 132. [Google Scholar] [CrossRef]
  15. Yang, X.; Yan, J.; Yang, X.; Tang, J.; Liao, W.; He, T. SCRDet++: Detecting Small, Cluttered and Rotated Objects via Instance-Level Feature Denoising and Rotation Loss Smoothing. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2384–2399. [Google Scholar] [CrossRef] [PubMed]
  16. Liu, Y.; Zhang, S.; Jin, L.; Xie, L.; Wu, Y.; Wang, Z. Omnidirectional scene text detection with sequential-free box discretization. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; AAAI Press: Washington, DC, USA, 2019. [Google Scholar]
  17. Qian, W.; Yang, X.; Peng, S.; Yan, J.; Guo, Y. Learning Modulated Loss for Rotated Object Detection. Proc. Aaai Conf. Artif. Intell. 2021, 35, 2458–2466. [Google Scholar] [CrossRef]
  18. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.S.; Bai, X. Gliding Vertex on the Horizontal Bounding Box for Multi-Oriented Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1452–1459. [Google Scholar] [CrossRef]
  19. Yang, X.; Zhang, G.; Yang, X.; Zhou, Y.; Wang, W.; Tang, J.; He, T.; Yan, J. Detecting Rotated Objects as Gaussian Distributions and its 3-D Generalization. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4335–4354. [Google Scholar] [CrossRef]
  20. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  22. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3500–3509. [Google Scholar] [CrossRef]
  23. Han, J.; Ding, J.; Xue, N.; Xia, G. ReDet: A Rotation-equivariant Detector for Aerial Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2785–2794. [Google Scholar]
  24. Ming, Q.; Zhou, Z.; Miao, L.; Zhang, H.; Li, L. Dynamic Anchor Learning for Arbitrary-Oriented Object Detection. Proc. Aaai Conf. Artif. Intell. 2021, 35, 2355–2363. [Google Scholar] [CrossRef]
  25. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align Deep Features for Oriented Object Detection. IEEE Trans. Geosci. Remote. Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  26. Yang, X.; Yan, J.; He, T. On the Arbitrary-Oriented Object Detection: Classification Based Approaches Revisited. Int. J. Comput. Vis. 2022, 130, 1873–1874. [Google Scholar] [CrossRef]
  27. Weiler, M.; Cesa, G. General E(2)-Equivariant Steerable CNNs. arXiv 2019, arXiv:1911.08251. [Google Scholar] [CrossRef]
  28. Yan, Z.X.; Li, Z.B.; Xie, Y.Q.; Li, C.Y.; Li, S.A.; Sun, F.W. ReBiDet: An Enhanced Ship Detection Model Utilizing ReDet and Bi-Directional Feature Fusion. Appl. Sci. 2023, 13, 7080. [Google Scholar] [CrossRef]
  29. Chen, K.; Pang, J.; Wang, J.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Shi, J.; Ouyang, W.; et al. Hybrid Task Cascade for Instance Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4969–4978. [Google Scholar] [CrossRef]
  30. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef]
  31. Jaiswal, A.; Wu, Y.; Natarajan, P.; Natarajan, P. Class-agnostic Object Detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 918–927. [Google Scholar] [CrossRef]
  32. Chen, Z.; Chen, K.; Lin, W.; See, J.; Yu, H.; Ke, Y.; Yang, C. PIoU Loss: Towards Accurate Oriented Object Detection in Complex Environments. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany; pp. 195–211. [Google Scholar]
  33. Zheng, Y.; Zhang, D.; Xie, S.; Lu, J.; Zhou, J. Rotation-Robust Intersection over Union for 3D Object Detection. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Berlin/Heidelberg, Germany; pp. 464–480. [Google Scholar]
  34. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking Rotated Object Detection with Gaussian Wasserstein Distance Loss. In Proceedings of the 2021 International Conference on Machine Learning, Virtual, 18– 24 July 2021. [Google Scholar]
  35. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning high-precision bounding box for rotated object detection via kullback-leibler divergence. Adv. Neural Inf. Process. Syst. 2021, 34, 18381–18394. [Google Scholar]
  36. Yang, X.; Zhou, Y.; Zhang, G.; Yang, J.; Wang, W.; Yan, J.; Zhang, X.; Tian, Q. The KFIoU Loss for Rotated Object Detection. arXiv 2022, arXiv:abs/2201.12558. [Google Scholar]
  37. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar] [CrossRef]
  38. Ding, J.; Xue, N.; Xia, G.S.; Bai, X.; Yang, W.; Yang, M.Y.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; et al. Object Detection in Aerial Images: A Large-Scale Benchmark and Challenges. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7778–7796. [Google Scholar] [CrossRef]
  39. Data, B.; Laboratory, D.M.N. Fine-Grained Dense Ship Detection Task Based on High-Resolution Remote Sensing Visible Light Data. 2023. Available online: https://www.datafountain.cn/competitions/635/datasets (accessed on 31 July 2024).
  40. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  41. Lang, S.; Ventola, F.G.; Kersting, K. DAFNe: A One-Stage Anchor-Free Deep Model for Oriented Object Detection. arXiv 2021, arXiv:abs/2109.06148. [Google Scholar]
  42. Li, C.; Xu, C.; Cui, Z.; Wang, D.; Jie, Z.; Zhang, T.; Yang, J. Learning Object-Wise Semantic Representation for Detection in Remote Sensing Imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  43. Lyu, C.; Zhang, W.; Huang, H.; Zhou, Y.; Wang, Y.; Liu, Y.; Zhang, S.; Chen, K. RTMDet: An Empirical Study of Designing Real-Time Object Detectors. arXiv 2022, arXiv:abs/2212.07784. [Google Scholar]
Figure 1. Prediction box generation method under le90 definition and CCW representation conditions. (a) Ideal regression path for generating prediction boxes. (b) Initial position of the proposal. (c) Actual regression path for generating prediction boxes.
Figure 1. Prediction box generation method under le90 definition and CCW representation conditions. (a) Ideal regression path for generating prediction boxes. (b) Initial position of the proposal. (c) Actual regression path for generating prediction boxes.
Sensors 24 05342 g001
Figure 2. Relationship curve between angular difference and IoU for rectangles with a 1:6 aspect ratio and overlapping centroids.
Figure 2. Relationship curve between angular difference and IoU for rectangles with a 1:6 aspect ratio and overlapping centroids.
Sensors 24 05342 g002
Figure 3. Process of generating predicted boxes by the Spatially Adaptive Angle-aware Network.
Figure 3. Process of generating predicted boxes by the Spatially Adaptive Angle-aware Network.
Sensors 24 05342 g003
Figure 4. Spatially Adaptive Angle-aware Network structure.
Figure 4. Spatially Adaptive Angle-aware Network structure.
Sensors 24 05342 g004
Figure 5. Structure diagram of the class-agnostic strategy regression function in the coarse regression stage.
Figure 5. Structure diagram of the class-agnostic strategy regression function in the coarse regression stage.
Sensors 24 05342 g005
Figure 6. Structure diagram of class-aware strategy regression function in the fine-tuning and precision refinement stages.
Figure 6. Structure diagram of class-aware strategy regression function in the fine-tuning and precision refinement stages.
Sensors 24 05342 g006
Figure 7. Curve depicting the relationship between IoU loss and angle difference.
Figure 7. Curve depicting the relationship between IoU loss and angle difference.
Sensors 24 05342 g007
Figure 8. Curve depicting the relationship between EAS loss and angle.
Figure 8. Curve depicting the relationship between EAS loss and angle.
Sensors 24 05342 g008
Figure 9. Transformed Gaussian distribution of rotated rectangular boxes.
Figure 9. Transformed Gaussian distribution of rotated rectangular boxes.
Sensors 24 05342 g009
Figure 10. The overlapping area I of two Gaussian distributions.
Figure 10. The overlapping area I of two Gaussian distributions.
Sensors 24 05342 g010
Figure 11. Distribution of various object classes in the DOTA-v1.5 training and validation sets.
Figure 11. Distribution of various object classes in the DOTA-v1.5 training and validation sets.
Sensors 24 05342 g011
Figure 12. Distribution of various object classes in the augmented DOTA-v1.5 training and validation sets.
Figure 12. Distribution of various object classes in the augmented DOTA-v1.5 training and validation sets.
Sensors 24 05342 g012
Figure 13. Horizontal comparison of model performance on the DFShip dataset at IoU thresholds from 0.5 to 0.95. The mAP is calculated using the all-point interpolation method.
Figure 13. Horizontal comparison of model performance on the DFShip dataset at IoU thresholds from 0.5 to 0.95. The mAP is calculated using the all-point interpolation method.
Sensors 24 05342 g013
Figure 14. Detection results on an image from the DOTA-v1.5 dataset. (a) Detection result of ReBiDet. (b) Detection result of the proposed ReBiDet + SA 3 .
Figure 14. Detection results on an image from the DOTA-v1.5 dataset. (a) Detection result of ReBiDet. (b) Detection result of the proposed ReBiDet + SA 3 .
Sensors 24 05342 g014
Table 1. Experimental results of the ReBiDet model and KFIoU loss on the DOTA-v1.5 dataset.
Table 1. Experimental results of the ReBiDet model and KFIoU loss on the DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReBiDet [28]41.1569.4842.54
ReBiDet + KFIoU 36.70 68.40 33.62
ReBiDet * [28]49.2677.9652.53
ReBiDet * + KFIoU 46.09 77.06 48.14
* Indicates training and testing on DOTA-v1.5 with random rotation augmentation and offline multi-scale augmentation.
Table 2. Experimental results of the SA 3 Network with two strategies on the DOTA-v1.5 dataset.
Table 2. Experimental results of the SA 3 Network with two strategies on the DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReDet  [23]-66.86-
ReDet *41.0068.0242.10
ReDet * + SA 3 Class-aware41.2368.9442.94
ReDet * + SA 3 Class-agnostic 41.91 69.30 44.48
ReBiDet [28]41.1569.4842.54
ReBiDet + SA 3 Class-aware41.3270.1642.94
ReBiDet + SA 3 Class-agnostic 42.48 71.21 44.13
* Denotes the detection results reproduced on the experimental platform in this paper. Note: The training and inference of models in this table were directly conducted using DOTA-v1.5 for both training and inference, without applying random rotation augmentation and offline multi-scale augmentation to the dataset.
Table 3. Experimental results of the SA 3 Network with two strategies on the augmented DOTA-v1.5 dataset.
Table 3. Experimental results of the SA 3 Network with two strategies on the augmented DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReDet [23]-76.80-
ReDet *48.1376.9550.99
ReDet * + SA 3 Class-aware 49.3877.9552.66
ReDet * + SA 3 Class-agnostic49.0277.6352.63
ReBiDet [28]49.2677.9652.53
ReBiDet + SA 3 Class-aware 49.5878.5053.38
ReBiDet + SA 3 Class-agnostic49.0777.8452.11
* Denotes the detection results reproduced on the experimental platform in this paper. Note: The training and inference of models in this table were conducted using DOTA-v1.5, with both random rotation augmentation and offline multi-scale augmentation applied during training and inference.
Table 4. EAS loss experiment results based on the ReDet model in the DOTA-v1.5 dataset.
Table 4. EAS loss experiment results based on the ReDet model in the DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReDet [23]-66.86-
ReDet *41.0068.0242.10
ReDet * + EAS41.2768.6242.28
ReDet * + SA 3 Class-aware41.2368.9442.94
ReDet * + SA 3 Class-aware + EAS 42.34 70.04 44.54
ReDet * + SA 3 Class-agnostic41.9169.3044.48
ReDet * + SA 3 Class-agnostic + EAS 42.3268.9144.81
* Indicates the detection results reproduced on the experimental platform in this paper. Note: The training and inference of the models in this table were conducted directly using the DOTA-v1.5 dataset, without applying random rotation augmentation or offline multi-scale enhancement. The mAP in the table follows the COCO calculation method, representing the average AP across all possible IoU thresholds. AP50 refers to the AP when the IoU threshold is 0.5, while AP75 denotes the average precision at an IoU threshold of 0.75.
Table 5. Experimental results of EAS loss on the ReBiDet model in the DOTA-v1.5 dataset.
Table 5. Experimental results of EAS loss on the ReBiDet model in the DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReBiDet + SA 3 Class-aware41.3270.1642.94
ReBiDet + SA 3 Class-aware + EAS 42.95 71.28 44.33
ReBiDet + SA 3 Class-agnostic42.4871.2144.13
ReBiDet + SA 3 Class-agnostic + EAS 42.8370.5944.93
Note: The training and inference of the models in this table were conducted directly using DOTA-v1.5, without applying random rotation augmentation or offline multi-scale enhancement.
Table 6. Experimental results of EAS loss on the ReDet model in the augmented DOTA-v1.5 dataset.
Table 6. Experimental results of EAS loss on the ReDet model in the augmented DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReDet [23]-76.80-
ReDet *48.1376.9550.99
ReDet * + SA 3 49.3877.9552.66
ReDet * + SA 3 + EAS 49.74 78.20 52.82
* indicates the reproduction of detection results on the experimental platform in this paper. Note: the SA 3 Networks in this table adopt the class-aware strategy, and the model is trained and tested using random rotation augmentation and offline multi-scale enhancement on the DOTA-v1.5 dataset.
Table 7. Experimental results of EAS loss on the ReBiDet model in the augmented DOTA-v1.5 dataset.
Table 7. Experimental results of EAS loss on the ReBiDet model in the augmented DOTA-v1.5 dataset.
MethodmAPAP50AP75
ReBiDet [28]49.2677.9652.53
ReBiDet + SA 3 49.5878.5053.38
ReBiDet + SA 3 + EAS 50.13 78.85 53.26
Note: the SA 3 Network in this table adopts the class-aware strategy, and the model is trained and tested using random rotation augmentation and offline multi-scale enhancement on the DOTA-v1.5 dataset.
Table 8. Horizontal comparison of model performance on the DFShip dataset at IoU thresholds from 0.5 to 0.95.
Table 8. Horizontal comparison of model performance on the DFShip dataset at IoU thresholds from 0.5 to 0.95.
IoUCSLR3DetS2A-NetReDetOriented R-CNNReBiDet + SA3 + EAS
0.5088.1495.9497.7298.7698.5298.81
0.5587.3395.6197.5998.7398.5298.81
0.6086.0695.1797.3898.6298.4198.76
0.6583.6394.3596.9698.3298.3198.68
0.7078.9692.2296.0797.8998.0498.56
0.7570.9987.6094.2796.0697.5698.22
0.8054.9278.2589.0191.4595.7997.20
0.8533.8157.7871.8878.5887.3093.18
0.9013.1025.6235.3744.4355.5277.05
0.951.891.733.185.444.6621.30
mAP59.8872.4377.9480.8383.2688.06
Note: The mAP is calculated using the all-point interpolation method. The red bold font represents the highest detection accuracy value at the same threshold in the horizontal comparison.
Table 9. Comparisons with state-of-the-art methods on DOTA-v1.5 OBB task.
Table 9. Comparisons with state-of-the-art methods on DOTA-v1.5 OBB task.
MethodPLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCCCmAP
single-scale:
RetinaNet-O [40]71.4377.6442.1264.6544.5356.7973.3190.8476.0259.9646.9569.2459.6564.5248.060.8359.16
FR-O [37]71.8974.4744.4559.8751.2868.9879.3790.7877.3867.5047.7569.7261.2265.2860.471.5462.00
Mask R-CNN [30]76.8473.5149.9057.8051.3171.3479.7590.4674.2166.0746.2170.6163.0764.4657.819.4262.67
HTC [29]77.8073.6751.4063.9951.5473.3180.3190.4875.1267.3448.5170.6364.8464.4855.875.1563.40
CMR [29]77.7774.6251.0963.4451.6472.9079.9990.3574.9067.5849.5472.8564.1964.8855.873.0263.41
DAFNe [41]----------------64.76
FR OBB [37] + RT [11]71.9276.0751.8769.2452.0575.1880.7290.5378.5868.2649.1871.7467.5165.5362.169.9965.03
ReDet [23]79.2082.8151.9271.4152.3875.7380.9290.8375.8168.6449.2972.0373.3670.5563.3311.5366.86
ReDet + SA 3 80.1283.5454.0872.5652.7677.1887.6390.8784.0574.2562.0773.7775.7165.6167.317.2669.30
ReBiDet [28]80.5482.9053.6274.5552.5579.6587.5390.8484.5772.9365.0273.0575.8765.5665.187.3269.48
ReDet + SA 3 + EAS80.2184.2553.5072.5352.7477.0487.8190.8883.9169.3464.2673.3375.8466.2464.809.0970.04
ReBiDet + SA 3 80.7083.6754.8974.5857.9979.9388.3890.8785.0374.2666.0673.1576.7970.0967.3215.7071.21
ReBiDet + SA 3 +EAS80.1283.7956.2173.1058.2780.5288.0990.8984.1274.0067.2075.0477.0470.7267.9813.3371.28
multi-scale:
DAFNe [41]----------------71.99
OWSR * [42]----------------74.90
RTMDet-R-tiny * [43]88.1483.0951.8077.5465.9982.2289.8190.8880.5481.3464.6471.5177.1376.3272.1146.6774.98
RTMDet-R-s * [43]88.1485.8252.9082.0965.5881.8389.7890.8283.3182.4768.5170.9378.0075.7773.0947.3276.02
RTMDet-R-m * [43]89.0786.7152.5782.4766.1382.5589.7790.8884.3983.3469.5173.0377.8275.9880.2142.0076.65
ReDet * [23]88.5186.4561.2381.2067.6083.6590.0090.8684.3075.3371.4972.0678.3274.7376.1046.9876.80
ReBiDet * [28]86.2385.8961.9982.4167.8683.9489.7890.8886.3783.7072.1277.5878.3873.2475.0152.0577.96
RTMDet-R-l * [43]89.3186.3855.0983.1766.1182.4489.8590.8486.9583.7668.3574.3677.6077.3977.8760.3778.12
ReBiDet * + SA 3 88.7686.3560.9781.9373.3984.2690.0590.8887.2083.3072.1977.0778.6772.6272.9455.3478.50
ReBiDet * + SA 3 + EAS86.8485.5862.2382.6068.0583.9289.8390.9086.9183.2373.6276.6478.5472.1277.5363.0578.85
* indicates multi-scale training and testing. Note. The RetinaNet OBB (RetinaNet-O) [40], Faster R-CNN OBB (FR-O) [37], Mask R-CNN [30], and Hybrid Task Cascade (HTC) [29] results are based on a reproduced version of DOTA-v1.5 [38] and have been used by some scholars [23,43]. “Single-scale” indicates the model is directly trained and tested on the DOTA-v1.5 dataset, while “multi-scale” indicates the model is trained and tested using random rotation and offline multi-scale enhancement on DOTA-v1.5. For ease of reading and comparison, the first, second, and third highest values in each column are marked in red, yellow, and green, respectively, and are bolded.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, Z.; Fan, J.; Li, Z.; Xie, Y. Elevating Detection Performance in Optical Remote Sensing Image Object Detection: A Dual Strategy with Spatially Adaptive Angle-Aware Networks and Edge-Aware Skewed Bounding Box Loss Function. Sensors 2024, 24, 5342. https://doi.org/10.3390/s24165342

AMA Style

Yan Z, Fan J, Li Z, Xie Y. Elevating Detection Performance in Optical Remote Sensing Image Object Detection: A Dual Strategy with Spatially Adaptive Angle-Aware Networks and Edge-Aware Skewed Bounding Box Loss Function. Sensors. 2024; 24(16):5342. https://doi.org/10.3390/s24165342

Chicago/Turabian Style

Yan, Zexin, Jie Fan, Zhongbo Li, and Yongqiang Xie. 2024. "Elevating Detection Performance in Optical Remote Sensing Image Object Detection: A Dual Strategy with Spatially Adaptive Angle-Aware Networks and Edge-Aware Skewed Bounding Box Loss Function" Sensors 24, no. 16: 5342. https://doi.org/10.3390/s24165342

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop