Next Article in Journal
Distributed Semi-Supervised Partial Multi-Label Learning over Networks
Previous Article in Journal
TCAD Simulation of an E-Mode Heterojunction Bipolar p-FET with Imax > 240 mA/mm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Quality Instance Mining and Weight Re-Assigning for Weakly Supervised Object Detection in Remote Sensing Images

1
Zhengzhou University of Light Industry, Zhengzhou 450002, China
2
School of Information and Communication Engineering, Hainan University, Haikou 570228, China
3
School of Aerospace Science and Technology, Xidian University, Xi’an 710126, China
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(23), 4753; https://doi.org/10.3390/electronics13234753
Submission received: 25 October 2024 / Revised: 27 November 2024 / Accepted: 29 November 2024 / Published: 1 December 2024

Abstract

:
Weakly supervised object detection (WSOD) in remote sensing images (RSIs) achieves high-value object classification and localization by only using image-level labels. However, two problems limit its performance. Firstly, adjacent instances are often misclassified because their pseudo-labels are determined solely based on the spatial distances between them and their corresponding seed instances. Secondly, most WSOD methods assign the highest weight to the instance that covers the discriminative part of an object, thereby urging WSOD models to focus on the discriminative part rather than the whole object. To handle the first problem, a high-quality instance mining (HQIM) module, which incorporates the feature similarities between instances into the label propagation process, enabling some misclassified adjacent instances to be removed, is proposed. To tackle the second problem, a weight re-assigning (WRA) strategy, which redistributes the loss weights of instances, is proposed. Specifically, the loss weights of instances focusing on the discriminative part are exchanged with those of instances that broadly cover the whole object. Ablation studies demonstrate the effectiveness of HQIM and WRA, while comparisons with popular models on two RSI benchmarks further verify the effectiveness of our model.

1. Introduction

Remote sensing images exhibit a wide range of applications across diverse fields, such as object detection [1,2], classification [3,4], and anomaly detection [5,6]. Among these, object detection in remote sensing images (RSIs) stands out as particularly compelling. This technology boasts a diverse array of applications in both military and civilian domains, including but not limited to surveillance and monitoring, disaster response, urban planning, and agricultural management. With advances in remote sensing technology and deep learning [7,8], fully supervised object detection (FSOD) models [9,10] for RSIs have achieved remarkable outcomes by leveraging instance-level labels. However, in RSIs with sophisticated backgrounds, annotating all objects with instance-level labels is an exceptionally labor-intensive and time-consuming task, and some objects may even be impossible to annotate accurately. Conversely, annotating the image-level labels is straightforward. Consequently, weakly supervised object detection (WSOD) in RSIs holds significant practical value and has garnered considerable attention.
The weakly supervised deep detection networks (WSDDNs) [2], as a path-breaking work, initially integrated multiple-instance learning with deep learning to train an object detector that solely uses image-level labels. The trained detector predicts the class scores (CSs) of proposals to accomplish object localization and classification.
The Online Instance Classifier Refinement (OICR) [11] model, which builds upon the WSDDN framework, incorporates several instance classifier refinement (ICR) branches to iteratively refine the CSs of proposals. The instance-level labels of the current ICR branch, which are generated by using the CSs of the upper ICR branch, are used to train the ICR branch. The pseudo-instance-level labels include positive instances (seed instances and neighboring instances) and negative instances (background instances). Specifically, the instance with the highest CS is regarded as the seed instance of each class. The remaining instances are subsequently divided into neighboring instances and background instances based on spatial distances between the seed instance and other instances.
A majority of WSOD [12,13,14,15,16,17,18,19,20,21,22,23] methods adopt OICR as the baseline to improve their performance. These methods can be broadly divided into two classes. The first class is designed to alleviate the problem of missing instances by mining more seed instances. For example, the proposal cluster learning (PCL) model [13] utilizes the k-means algorithm to divide the proposals into multiple clusters; then, the clusters centers are selected as seed instances. The neighboring instances are determined solely based on the spatial distances between the seed instances and other instances. Other similar models include the complete and invariant instance classifier refinement model [20], the multiple-instance self-training (MIST) model [24], and the weakly supervised contrastive loss (WSCL) [21]. The second class aims to guide the model in recognizing the whole object, rather than just its distinguishing parts. For example, the multi-scale image splitting-based feature enhancement model [17] enforces the spatial attention maps of positive instances to approximate the maximum spatial attention maps, thereby encouraging the WSOD model to pay more attention to the whole object. Other similar works include the semantic segmentation-guided pseudo-label mining [12] model, the instance re-detection comprehensive attention self-distillation model [14], and the pseudo-instance soft labels mining [25] model. Although the aforementioned methods have obtained notable performance in WSOD, there are still two problems that require attention and resolution.
For the first problem, during the label propagation process, the pseudo-labels of adjacent instances are determined solely based on the spatial distances between them and their corresponding seed instance. Consequently, some adjacent instances are often misclassified. As depicted in Figure 1a, the basketball court is incorrectly identified as the tennis court by the baseline model, as the basketball court is an adjacent instance of tennis court.
For the second problem, most WSOD methods tend to assign a higher loss weight to the instance that only covers the discriminative part of an object, whereas instances that broadly cover the whole object obtain a lower loss weight. Consequently, the inappropriate loss weights encourage the WSOD model to detect the discriminative part of an object rather than the whole object. As illustrated in Figure 1b, the baseline model only detects the most discriminative part of an airplane rather than the whole airplane.
To overcome the first problem, a high-quality instance mining (HQIM) strategy is proposed. Firstly, the seed instances are identified by selecting the top p percent proposals from the sorted proposals. Then, these seed instances are refined through the local Non-Maximum Suppression (NMS) operation. Secondly, the initial neighboring instances are determined based on the spatial distances between them and their corresponding seed instances; furthermore, these initial neighboring instances are refined by using the feature similarities between them and the seed instances to remove some misidentified adjacent instances. As illustrated in Figure 1a, the combination of the baseline and the HQIM module can avoid the misclassification of adjacent instances.
To overcome the second problem, we propose a weight re-assigning (WRA) strategy. First, the instance that fully covers the object is identified by using the spatial distances between positive instances. Then, we re-assign the loss weight of this comprehensive instance based on the CS of the instance covering the object’s discriminative part. Conversely, the loss weight of the instance focusing on the discriminative part is adjusted by using the CS of the comprehensive instance. Consequently, as illustrated in Figure 1b, the baseline model with the proposed WRA strategy tends to capture the whole object.
Overall, as shown in Figure 1c, the baseline model with HQIM + WRA not only avoids the misclassification of adjacent instances but also encourages the model to focus on the whole object. The summarized contributions are outlined as follows:
  • During the label propagation process, the label of a neighboring instance is determined solely based on the spatial distance between it and its corresponding seed instance. Inevitably, this leads to some neighboring instances being misclassified. To address this issue, the HQIM module is proposed. This module utilizes feature similarity between seed instances and their neighboring instances to further refine the neighboring instances, thereby removing the misclassified neighboring instances.
  • Most WSOD models often assign higher loss weights to instances focusing on the discriminative part of an object, compared with those covering the entire object. Consequently, they tend to detect the discriminative part of an object. To address this issue, we propose the WRA strategy, which exchanges the loss weights between these two types of instances.

2. Related Works

OICR [11] is used as the baseline of our model. To enhance the localization capacity of the WSOD model, many researchers incorporate bounding box regression (BBR) branches into the OICR model. Therefore, the WSDDN, OICR, and BBR branches are introduced in turn.

2.1. Weakly Supervised Deep Detection Network

Initially, as illustrated in Figure 2, a variety of proposals P R = { p r 1 , p r m , p r M } are generated by using the selective search algorithm [26] for the input image I, where M denotes the quantity of proposals. Secondly, the feature vectors of all proposals, denoted by { f m } m = 1 M , are obtained by importing I and its proposals into backbone networks, region of interest (RoI) pooling layer, and two fully connected (Fc) layers. Thirdly, the { f m } m = 1 M are imported into classification branch to obtain a classification score vector of class c, denoted by X c c l s R 1 × M . Similarity, the detection score vector of class c, denoted by X c d e c R 1 × M , is obtained by importing the { f m } m = 1 M into the detection branch. The classification and detection branches both include an FC layer and a softmax classifier. Then, the instance-level predicted CS vector of class c, denoted by X c R 1 × M , is acquired as follows:
X c = X c c l s X c d e c ,
where ⊙ represents the Hadamard product. The image-level predicted CS of class c, denoted by ϕ c , is obtained as follows:
ϕ c = m = 1 M x c m ,
where x c m X c denotes the CS of the mth proposal in class c. Finally, the loss of the WSDDN, denoted by L W S D D N , is obtained as follows:
L W S D D N = c = 1 C ( y c log ϕ c + ( 1 y c ) log ( 1 ϕ c ) ) ,
where C denotes the quantity of classes and y c = 1 or 0 represents that I contains or does not contain category c.

2.2. Online Instance Classifier Refinement

To improve the performance of the WSDDN, the K ICR branches are incorporated into the WSDDNs, where the ICR branch contains an FC layer and a softmax classifier. The feature vectors of the proposals are imported into the kth ICR branch to obtain the predicted CS vectors of all classes, denoted by { S c k R 1 × M } c = 1 C + 1 , k 1 , 2 , K , where the (C+1)th dimension denotes the background class. The pseudo-instance-level labels of the kth ICR branch, denoted by y k , are generated by using { S c k 1 R 1 × M } c = 1 C + 1 , and the pseudo-labels of the first ICR branch are generated by using { X c } c = 1 C . Specifically, the instance with the highest prediction CS is considered the seed instance, and the neighboring instances are determined by using the spatial distances between the seed instances and other instances. The loss of the kth ICR branch, denoted by L I C R k , is obtained by the following equation:
L I C R k = m = 1 M c = 1 C + 1 w m k y c , m k s c , m k ,
where y c , m k { 0 , 1 } represents the element in the mth column and cth row of y k , s c , m k denotes the CS of the kth ICR branch of the mth proposal in class c. The w m k = s c , m k 1 denotes the loss weight of the mth proposal in the kth ICR branch, and s c , m k 1 denotes the CS of ( k 1 )th ICR branch of the mth proposal in class c.

2.3. Incorporating the Regression Branches into OICR

To improve the localization performance of OICR, K bounding box regression (BBR) branches are incorporated into OICR, where each BBR branch includes an FC layer. The loss of the kth BBR branch, denoted by L B B R k , is obtained as follows:
L B B R k = 1 P k j = 1 P k ( s m o o t h L 1 ( g j , g ^ j ) ) ,
where P k denotes the set of positive instances, P k denotes the number of positive instances, and g ^ r and g r denote the target offset and prediction offset of the jth positive instance, respectively.

3. Proposed Method

3.1. Overview

As shown in Figure 2, our proposed method integrates the HQIM module and WRA strategy into the OICR model. Firstly, the HQIM module utilizes the sorted CSs of the proposals to mine seed instances. The initial neighboring instances are selected by using the spatial distances between seed instances and other instances, which are further refined by using the feature similarities between them and seed instances to remove some misclassified neighboring instances. Secondly, the WRA strategy employs spatial distances between positive instances to identify the one that broadly covers the whole object. The loss weight of this comprehensive instance is then re-assigned based on the CS of an instance that covers the object’s discriminative part. Conversely, the loss weight of the instance focusing solely on the discriminative part is adjusted by using the CS of the comprehensive instance.

3.2. High-Quality Instance Mining Module

Popular WSOD models propagate the labels from the real image level to the pseudo-instance level for supervising the ICR branch. The performance of the WSOD model relies on the quality of pseudo-instance-level labels. The pseudo-instance-level labels include positive instances (seed instances and neighboring instances) and negative instances (background instances). Once the positive instances are identified, the rest of the instances are automatically classified as negative instances. Consequently, the quality of the positive instances ultimately determines the overall quality of the pseudo-instance-level labels.
In most WSOD models, the neighboring instances are determined by using the spatial distances between seed instances and other instances. This often results in the misclassification of neighboring instances, thereby limiting the quality of positive instances. To address this issue, the HQIM strategy is proposed to eliminate the misclassified neighboring instances. Firstly, the seed instances are mined from the sorted CSs of proposals. Then, to remove small and redundant seed instances, the NMS operation is applied to refine the seed instances. Secondly, the spatial distances between the seed instance and other instances are used to determine the initial neighboring instances. Subsequently, as shown in Figure 2, the feature distribution curves of the seed instance (represented by pink lines) and the neighboring instances (represented by orange lines) are illustrated. The feature distributions of misclassified neighboring instances are not similar to those of their corresponding seed instances. Therefore, the feature similarities between the seed instances and their neighboring instances are used to further refine the initial neighboring instances, removing the misclassified neighboring instances. The details of the HQIM module are as follows.
Considering the multiple objects of the same category in RSIs, firstly, we assume that the input image I contains category c; the initial set of seed instances of the kth ICR branch in class c, denoted by I S I c k , is determined by picking the top p percent proposals from the sorted S c k 1 . The set of seed instances of class c in the kth ICR branch, denoted by S I c k , is obtained by implementing NMS among I S I c k .
Secondly, as shown in Figure 2, the initial set of neighboring instances of the ith seed instance s i c , i k S I c k , denoted by I N I s i c , i k , is obtained by selecting the instances that have high spatial distances (e.g., IoU > 0.5) between them and s i c , i k . However, some neighboring instances barely cover the object. To remove these low-quality neighboring instances, feature similarity is proposed to evaluate the similarity between each seed instance and its neighboring instances, allowing us to eliminate those low-quality neighboring instances.
The definition of the feature similarity f s s i c , i k i n i j between seed instance s i c , i k and initial neighboring instance i n i j is as follows:
f s s i c , i k i n i j = s i m ( f s i c , i k , f i n i j ) ,
where s i m ( · , · ) denotes the dot product between inputs, f s i c , i k denotes the feature vector of seed instances s i c , i k , i n i j I N I s i c , i k denotes the jth element of I N I s i c , i k , and f i n i j denotes the feature vector of i n i j . The initial neighboring instance i n i j is regarded as a neighboring instance if f s s i c , i k i n i j > δ , where δ denotes the threshold of feature similarity, which determines whether or not an initial neighboring instance belongs to the neighboring instances. Therefore, the set of neighboring instances of seed instance s i i k , c , denoted by N I s i c , i k , is obtained. Finally, the neighboring instances and seed instances form the high-quality positive instances.

3.3. Weight Re-Assigning Strategy

The loss weights of instances not only ensure the stability of training but also encourage the model to focus on the instances with higher loss weights. Therefore, the reasonable assignment of instance loss weights greatly impacts the WSOD model’s performance. However, the instance that covers the discriminative part of object usually has a higher CS, and its loss weight is obtained by copying its CS. Conversely, the instance broadly covers the whole object, and both its CS and loss weight tend to be relatively small. Therefore, the trained object detector tends to detect the discriminative part of the object. To handle the problem, the WRA module is proposed. As shown in Figure 2, firstly, the spatial distance matrix between positive instances is calculated to select the instance that broadly covers the whole object. Secondly, the WRA strategy switches the loss weights of the instance that covers the distinguishing region of an object and the instance that broadly covers the whole object, to encourage the WSOD model to pay more attention to the whole object. The details of the WRA strategy are as follows.
As shown in Figure 2, firstly, the seed instance and its neighboring instance set are constructed as a positive instance cluster S N = r 1 , r n , , r N , where r n denotes the nth instance in the cluster, and N denotes the number of elements in the cluster. Secondly, the spatial distance matrix, denoted by G R N × N , is constructed for the instances in the cluster according to their spatial distance. g n , n G denotes the spatial distance between the r n th and r n th instances in the cluster S N , which is obtained as follows:
g n , n = r n r n r n r n .
Then, the spatial distance vector, denoted by V R N × 1 , is obtained as follows:
V = n = 1 N G .
Finally, the index of the instance that broadly covers the whole object is selected as follows:
z = arg max V .
The index of the instance that has the highest CS in the cluster S N is obtained as follows:
p = arg max { s c , r n k 1 } n = 1 N ,
where s c , r n k 1 denotes the CS of (k-1)th ICR branch of the r n th positive instance in class c. The loss weights of instances r z and r p are assigned as follows:
w r z = s c , r p k 1 ,
w r p = s c , r z k 1 ,
where w r z ( w r p ) denotes the loss weight of the r z th ( r p th) positive instance and s c , r p k 1 ( s c , r z k 1 ) denotes the CS of the (k-1)th ICR branch of the r z th ( r p th) positive instance in class c. Therefore, the WRA strategy encourages the WSOD model to detect the whole object.

3.4. Overall Training Loss

The overall loss of the proposed method, represented as L A L L , is obtained as follows:
L A L L = L W S D D N + k = 1 K ( L I C R k + L B B R k ) .
During the inference stage, the feature vectors of all proposals are imported into the K trained ICR and BBR branches to obtain prediction scores and prediction offsets. Subsequently, the prediction scores and prediction offsets are averaged separately to jointly infer the initial detection results. Finally, the detection results are derived by applying NMS operation to the initial detection results.

4. Materials, Data, and Experiments

4.1. Materials and Data

The source codes of our model are designed to run on the Ubuntu 16.04 platform, leveraging eight GPUs (Titan RTX). These codes have been implemented within the PyTorch framework, ensuring compatibility and efficiency for deep learning tasks.
The NWPU VHR-10.v2 dataset [27,28] contains 1172 RGB images, each with a uniform size of 400 × 400 pixels. The spatial resolutions of these images vary from 0.08 m to 2 m. The images are sourced from Google Earth and the Vaihingen dataset [29]. Additionally, the dataset is divided into three subsets: training, validation, and test. Specifically, the training subset comprises 679 images, the validation subset includes 200 images, and these are used to train our model. Meanwhile, the test subset, which consists of 293 images, is utilized to validate the performance of our trained model. Furthermore, the NWPU VHR-10.v2 dataset encompasses 10 classes, totaling 2775 object instances.
The DIOR dataset [30] contains 23,463 RGB images, each with a uniform size of 800 × 800 pixels. The spatial resolutions of these images vary from 0.5 m to 30 m. The images are sourced from Google Earth. Additionally, the dataset encompasses 20 classes, totaling 192,472 object instances. Furthermore, the dataset is divided into three subsets: training, validation, and test. Specifically, the training subset comprises 5862 images, the validation subset includes 5863 images, and these are used to train our model. Meanwhile, the test subset, which consists of 11,738 images, is utilized to validate the performance of our trained model.
The PASCAL VOC 2007 dataset [31] contains 9963 RGB images. The dataset encompasses 20 classes, totaling 24,640 object instances. The PASCAL VOC 2007 dataset is also divided into a training subset (2501 images), a validation subset (2510 images), and a test subset (4952 images). The training and validation subsets are used to train our model, while the test subset is used to test the trained model.

4.2. Experiments

4.2.1. Implementation Details

Our method builds on the OICR [11] model and adapts the VGG16 [32] pre-trained on the ImageNet dataset [33] as the background. The count of ICR branches is configured to be 3, i.e., K = 3; the top p percent is set to 0.15, i.e., p = 0.15; the threshold of feature similarity is set to 0.3, i.e., δ = 0.3; and the threshold of NMS during the refinement of seed instances is set to 0.1. In the inference stage, the NMS [34] threshold is established at 0.3.
The NWPU VHR-10.v2, DIOR, and PASCAL VOC 2007 datasets were augmented by applying horizontal flipping as well as rotations of 90° and 180°. Similar to almost all WSOD models [2,11,17,21,35], all images were resized to one of the following dimensions: {480, 576, 688, 864, 1200}. The model was optimized by using the stochastic gradient descent algorithm, and the momentum, weight decay, and batchsize were 0.9, 0.005, and 8, respectively. The number of iterations was 30 K for the NWPU VHR-10.v2 and PASCAL VOC 2007 datasets. The initial learning rate was 0.1, and it was reduced to 10% of the previous value at 20 K and 26 K for the NWPU VHR-10.v2 and PASCAL VOC 2007 dataset, respectively. Additionally, for the DIOR dataset, the number of iterations was 60 K. The initial learning rate was 0.1 and was decreased to 10% of its previous value at 50 K and 56 K iterations for the DIOR dataset.
The detection accuracy of the trained model in the test subset was assessed by the mean average precision (mAP), and the localization capability of our model was assessed by correct localization (CorLoc) in the training and validation subsets. CorLoc focuses more on the accuracy of object localization, while mAP considers a broader range of model performance indicators, including precision, recall, and confidence scores.

4.2.2. Parameter Analysis

The parameter p in the HQIM module determines the extent of seed mining, which was analyzed in terms of mAP on both the NWPU VHR-10.v2 and DIOR datasets. As shown in Figure 3a,b, our model attained the maximum mAP at p = 0.15, while the mAP experiences a slight decline at p = 0.16. Consequently, this article adopted p = 0.15.
The parameter δ in the HQIM module determines whether or not an initial neighboring instance belongs to a neighboring instance, which was analyzed in terms of mAP on both the NWPU VHR-10.v2 and DIOR datasets. As shown in Figure 4a,b, our model attained the maximum mAP at δ = 0.3, while the mAP experiences a slight decline at δ = 0.4. Consequently, this article adopted δ = 0.3.
The parameter K in our model determines the number of ICR branches, which was analyzed in terms of mAP on both the NWPU VHR-10.v2 and DIOR datasets. As illustrated in Figure 5a,b, our model attained the maximum mAP at K = 3, while the mAP experiences a slight decline at K = 4. Consequently, this article adopted K = 3.

4.2.3. Ablation Study

The ablation studies, conducted on the NWPU VHR-10.v2 dataset and more challenging DIOR dataset, were used to validate the efficacy of HQIM, WRA, and their combination. Particularly, to enhance the localization ability of the WSOD model, like numerous WSOD models [12,16,20,21,24], the combination of the OICR and BBR branches was employed as the baseline model.
Quantitative ablation study: As shown in Table 1, compared with the baseline, the mAP (CorLoc) of baseline + HQIM increased by 9.46% (7.97%) and 4.81% (5.75%) on the NWPU VHR-10.v2 and DIOR datasets, which validates the effectiveness of the HQIM module. Furthermore, compared with the baseline, the mAP(CorLoc) of baseline + WRA increased by 12.55% (11.40%) and 7.13% (8.10%) on the NWPU VHR-10.v2 and DIOR datasets, which validates the effectiveness of the WRA strategy. In addition, compared with the baseline, the mAP (CorLoc) of baseline + HQIM + WRA increased by 17.28% (15.35%) and 8.81% (11.13%) on the NWPU VHR-10.v2 and DIOR datasets, which validates the effectiveness of our method.
Subjective ablation study: To further validate the effectiveness of the HQIM module, the WRA strategy, and their combination, subjective ablation studies were conducted on the NWPU VHR-10.v2 and DIOR datasets. Firstly, as illustrated in Figure 6a, subjective comparisons were conducted between the detection results of the baseline and those of baseline + HQIM. Obviously, baseline + HQIM demonstrated the ability to reduce the misclassification of neighboring instances, effectively addressing the first problem outlined in the Introduction. Secondly, as shown in Figure 6b, subjective comparisons were conducted between the detection results of the baseline and those of baseline + WRA. Apparently, compared with the baseline, baseline + WRA captured the whole object, rather than just its discriminative parts, effectively tackling the second problem introduced in the Introduction. Thirdly, as shown in Figure 6c, subjective comparisons were conducted between the detection results of the baseline and those of baseline + HQIM + WRA. Obviously, compared with the baseline, baseline + HQIM + WRA model not only avoided the misclassification of neighboring instances but also encouraged the model to focus on the whole object.

4.2.4. Quantitative Comparison with Popular Models

To verify the performance of our model, two classical FSOD models (Fast R-CNN [1] and Faster R-CNN [9]) and eleven popular WSOD models, which included the WSDDN [2], OICR [11], MIST [24], the dynamic curriculum learning (DCL) model [16], the progressive contextual instance refinement (PCIR) model [19], the multiple-instance graph (MIG) model [36], the triple context-aware (TCA) model [35], the self-supervised adversarial and equivariant (SAE) model [37], the self-guided proposal generation (SPG) model [22], the PISLM model [25], and the SGPLM model [12], were quantitatively compared with our model on the NWPU VHR-10.v2 and more challenging DIOR datasets. Particularly, the results of the compared methods are cited from previous publications, and the DCL model without available source codes lacks class-by-class CorLoc metrics in the NWPU VHR-10.v2 and DIOR datasets.
As shown in Table 2, the table lists different methods (columns) and the corresponding average precision (AP) for various object categories (rows) in the NWPU VHR-10.v2 dataset. The mAP of each model is provided in the last column of Table 2, which represents the mean AP across all categories for each model. Our method achieves the mAP of 66.24% on the NWPU VHR-10.v2 dataset, outperforming the benchmark model OICR by 31.72%, thereby demonstrating its overall effectiveness. Furthermore, our model’s mAP surpasses the following models by the respective percentages: WSDDN by 31.12%, MIST by 14.72%, DCL by 14.13%, PCIR by 11.27%, MIG by 10.29%, TCA by 7.42%, SAE by 5.52%, SPG by 3.44%, PISLM by 2.44%, and SGPLM by 1.04%.
As shown in Table 3, the table lists different methods (columns) and the corresponding CorLoc for various object categories (rows) in the NWPU VHR-10.v2 dataset. The CorLoc of each model is provided in the last column of Table 3, which represents the mean CorLoc across all categories for each model. The CorLoc of our model achieves 76.89% on the NWPU VHR-10.v2 dataset. Furthermore, our model’s CorLoc surpasses the following models by the respective percentages: WSDDN by 41.65%, OICR by 36.88%, MIST by 6.59%, DCL by 7.19%, PCIR by 5.02%, MIG by 6.73%, TCA by 4.13%, SAE by 3.43%, SPG by 3.48%, PISLM by 2.59%, and SGPLM by 1.49%.
As shown in Table 4, the table lists different methods (columns) and the corresponding AP for various object categories (rows) in the DIOR dataset. The mAP of each model is provided in the last column of Table 4, which represents the mean AP across all categories for each model. Our methodology achieves the mAP of 28.91% on the more challenging DIOR dataset, surpassing the benchmark model OICR by 12.41%. Moreover, our model’s mAP outperforms the following models by the respective percentages: WSDDN by 15.65%, MIST by 6.73%, DCL by 8.72%, PCIR by 3.99%, MIG by 3.80%, TCA by 3.09%, SAE by 1.81%, SPG by 3.14%, PISLM by 0.31%, and SGPLM by 0.41%.
As shown in Table 5, the table lists different methods (columns) and the corresponding CorLoc for various object categories (rows) in the DIOR dataset. The CorLoc of each model is provided in the last column of Table 5, which represents the mean CorLoc across all categories for each model. Notably, our model achieves a CorLoc of 53.92% on the challenging DIOR dataset. Furthermore, our model’s CorLoc surpasses the following models by significant margins: WSDDN by 21.48%, OICR by 19.15%, MIST by 10.32%, DCL by 11.72%, PCIR by 7.80%, MIG by 7.12%, TCA by 5.51%, SAE by 4.50%, SPG by 5.62%, PISLM by a narrow margin of 0.72%, and SGPLM by 0.72%.
In addition, our model significantly reduces the disparity with respect to the FSOD models on the NWPU VHR-10.v2 and DIOR datasets. Notably, our model’s performance in specific classes, e.g., airplane, ground track field, and baseball field, approaches that of the two classical FSOD models.
Furthermore, to validate the generalization capability of our method, we also conducted experiments on the PASCAL VOC 2007 dataset. Specifically, our model achieved the mAP of 57.25% on the PASCAL VOC 2007 dataset, surpassing the benchmark model OICR by 16.05%. Additionally, our model’s CorLoc reached 71.56%, outperforming OICR by 10.96%.

4.2.5. Subjective Evaluation

Figure 7 and Figure 8 give the visualization of some detection results of our model on the NWPU VHR-10.v2 and DIOR datasets, respectively. Obviously, our method intuitively demonstrates excellent performance on the 24 test images.
In summary, the better performance of our method mainly contributes to two aspects. Firstly, the HQIM module removes the misclassified neighboring instances; thus, the quality of positive instances is improved. Secondly, the WRA strategy re-assigns a high loss weight to the instances that broadly cover the whole object, thereby urging our model to detect the whole object.

4.2.6. Evaluation of Computational Cost

As shown in Table 6, in order to evaluate the computational cost, our method was compared with two classical WSOD models (WSDDN and OICR) and five advanced WSOD models with available source codes (PCL, MELM, MIST, PISLM, and SPGLM) on the NWPU VHR-10.v2 dataset in terms of training time, frames per second (FPS), giga floating point operations (GFLOPs), and mAP.
The proposed method exhibits a marginally longer training time and slightly higher GFLOPs compared with OICR. However, it significantly outperforms OICR in terms of mAP. Particularly, in practical application, the trained model is deployed within a workstation to infer the test image. Obviously, during the inference stage, our method attains an impressive 4.93 fps while maintaining a high mAP of 66.24%. In summary, these results underscore better overall efficiency of our method among the four evaluated approaches.

5. Conclusions

The primary contributions of this paper include two significant parts. On the one hand, the HQIM module is proposed to handle the problem of misclassification of neighboring instances. During the label propagating process, the labels of neighboring instances are assigned solely based on spatial distances. The HQIM module further refines the selected neighboring instances by utilizing feature similarities between them and their corresponding seed instances to remove some misclassified neighboring instances. On the other hand, the WRA strategy is proposed to handle the problem that the instance covering the discriminative part of object is assigned the highest loss weight. The loss weights of instances typically guide the model’s attention; thus, WSOD models pay more attention to the instances that cover the discriminative part of an object rather than those that cover the whole object. The WRA strategy switches the loss weights between instances focused on the discriminative part of the object and instances that cover the whole object, encouraging the WSOD model to pay more attention to the whole object. The ablation studies demonstrate the efficacy of the HQIM module, the WRA strategy, and their combination. The quantitative comparisons with popular WSOD models show that our method achieves the mAP (CorLoc) of 66.24% (76.89%) on the NWPU VHR-10.v2 dataset, outperforming the benchmark model OICR by 31.72% (36.88%), further illustrating the superiority of our model.
The localization accuracy of our method currently depends on the proposals produced by the conventional SS algorithm, which has inherent limitations. Hence, the next step in refining our method involves improving the localization precision by adopting advanced segmentation techniques.

Author Contributions

Conceptualization, C.W.; formal analysis, P.X.; funding acquisition, C.W.; methodology, P.X. and C.W.; project administration, M.H.; resources, M.H.; software, C.W.; supervision, P.X.; validation, Y.C. and M.H.; writing—original draft, C.W.; writing—review and editing, C.W. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (grant No. 62076223).

Data Availability Statement

The DIOR and NWPU VHR-10.v2 datasets are available at the following URLs: https://drive.google.com/drive/folders/1UdlgHk49iu6WpcJ5467iT-UqNPpx__CC (accessed on 27 March 2023) and https://drive.google.com/file/d/15xd4TASVAC2irRf02GA4LqYFbH7QITR-/view (accessed on 27 March 2023), respectively.

Acknowledgments

We are deeply grateful to the authors of the WSDDN and OICR for generously sharing their source codes, which have significantly facilitated our experiment. Additionally, the Vaihingen dataset was provided by the German Society for Photogrammetry, Remote Sensing and Geoinformation (DGPF) [Cramer, 2010]: https://www.ifp.uni-stuttgart.de/dgpf/DKEP-Allg.html (accessed on 28 March 2023).

Conflicts of Interest

The authors explicitly state that they have no conflicts of interest to declare.

Abbreviations

The following abbreviations are used in this manuscript:
WSODweak supervised object detection
RSIremote sensing image
CSclass score
HQIMhigh-quality instance mining
WRAweight re-assigning

References

  1. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV 2015), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  2. Bilen, H.; Vedaldi, A. Weakly supervised deep detection networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2846–2854. [Google Scholar]
  3. Cheng, G.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  4. Li, M.; Li, W.; Liu, Y.; Huang, Y.; Yang, G. Adaptive Mask Sampling and Manifold to Euclidean Subspace Learning with Distance Covariance Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5508518. [Google Scholar] [CrossRef]
  5. Huo, Y.; Cheng, X.; Lin, S.; Zhang, M.; Wang, H. Memory-Augmented Autoencoder with Adaptive Reconstruction and Sample Attribution Mining for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5518118. [Google Scholar] [CrossRef]
  6. Cheng, X.; Zhang, M.; Lin, S.; Zhou, K.; Zhao, S.; Wang, H. Two-Stream Isolation Forest Based on Deep Features for Hyperspectral Anomaly Detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5504205. [Google Scholar] [CrossRef]
  7. Qian, X.; Zeng, Y.; Wang, W.; Zhang, Q. Co-Saliency Detection Guided by Group Weakly Supervised Learning. IEEE Trans. Multimed. 2023, 25, 1810–1818. [Google Scholar] [CrossRef]
  8. Cheng, X.; Huo, Y.; Lin, S.; Dong, Y.; Zhao, S.; Zhang, M.; Wang, H. Deep Feature Aggregation Network for Hyperspectral Anomaly Detection. IEEE Trans. Instrum. Meas. 2024, 73, 5033016. [Google Scholar] [CrossRef]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  10. Qian, X.; Wu, B.; Cheng, G.; Yao, X.; Wang, W.; Han, J. Building a Bridge of Bounding Box Regression Between Oriented and Horizontal Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605209. [Google Scholar] [CrossRef]
  11. Tang, P.; Wang, X.; Bai, X.; Liu, W. Multiple instance detection network with online instance classifier refinement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21 July–26 July 2017; pp. 2843–2851. [Google Scholar]
  12. Qian, X.; Li, C.; Wang, W.; Yao, X.; Cheng, G. Semantic segmentation guided pseudo label mining and instance re-detection for weakly supervised object detection in remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2023, 119, 103301. [Google Scholar] [CrossRef]
  13. Tang, P.; Wang, X.; Bai, S.; Shen, W.; Bai, X.; Liu, W.; Yuille, A. PCL: Proposal Cluster Learning for Weakly Supervised Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 176–191. [Google Scholar] [CrossRef] [PubMed]
  14. Huang, Z.; Zou, Y.; Kumar, B.V.K.V.; Huang, D. Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection. In Proceedings of the Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Sydney, Australia, 2020; Volume 33, pp. 16797–16807. [Google Scholar]
  15. Wan, F.; Wei, P.; Jiao, J.; Han, Z.; Ye, Q. Min-entropy latent model for weakly supervised object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1297–1306. [Google Scholar]
  16. Yao, X.; Feng, X.; Han, J.; Cheng, G.; Guo, L. Automatic weakly supervised object detection from high spatial resolution remote sensing images via dynamic curriculum learning. IEEE Trans. Geosci. Remote Sens. 2020, 59, 675–685. [Google Scholar] [CrossRef]
  17. Qian, X.; Wang, C.; Li, C.; Li, Z.; Zeng, L.; Wang, W.; Wu, Q. Multiscale Image Splitting Based Feature Enhancement and Instance Difficulty Aware Training for Weakly Supervised Object Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 7497–7506. [Google Scholar] [CrossRef]
  18. Huo, Y.; Qian, X.; Li, C.; Wang, W. Multiple Instance Complementary Detection and Difficulty Evaluation for Weakly Supervised Object Detection in Remote Sensing Image. IEEE Geosci. Remote Sens. Lett. 2023, 20, 6006505. [Google Scholar] [CrossRef]
  19. Feng, X.; Han, J.; Yao, X.; Cheng, G. Progressive contextual instance refinement for weakly supervised object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8002–8012. [Google Scholar] [CrossRef]
  20. Qian, X.; Wang, C.; Wang, W.; Yao, X.; Cheng, G. Complete and Invariant Instance Classifier Refinement for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5627713. [Google Scholar] [CrossRef]
  21. Seo, J.; Bae, W.; Sutherland, D.J.; Noh, J.; Kim, D. Object Discovery via Contrastive Learning for Weakly Supervised Object Detection. In Proceedings of the Computer Vision—ECCV 2022; Springer Nature: Cham, Switzerland, 2022; pp. 312–329. [Google Scholar]
  22. Cheng, G.; Xie, X.; Chen, W.; Feng, X.; Yao, X.; Han, J. Self-Guided Proposal Generation for Weakly Supervised Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5625311. [Google Scholar] [CrossRef]
  23. Xie, X.; Cheng, G.; Feng, X.; Yao, X.; Qian, X.; Han, J. Attention Erasing and Instance Sampling for Weakly Supervised Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5600910. [Google Scholar] [CrossRef]
  24. Ren, Z.; Yu, Z.; Yang, X.; Liu, M.Y.; Lee, Y.J.; Schwing, A.G.; Kautz, J. Instance-aware, context-focused, and memory-efficient weakly supervised object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10598–10607. [Google Scholar]
  25. Qian, X.; Huo, Y.; Cheng, G.; Gao, C.; Yao, X.; Wang, W. Mining High-Quality Pseudoinstance Soft Labels for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5607615. [Google Scholar] [CrossRef]
  26. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
  27. Cheng, G.; Zhou, P.; Han, J. Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  28. Li, K.; Cheng, G.; Bu, S.; You, X. Rotation-Insensitive and Context-Augmented Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2337–2348. [Google Scholar] [CrossRef]
  29. Cramer, M. The DGPF-Test on Digital Airborne Camera Evaluation Overview and Test Design. Photogramm.-Fernerkund.-Geoinf. 2010, 2010, 73–82. [Google Scholar] [CrossRef]
  30. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  31. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  34. Hosang, J.; Benenson, R.; Schiele, B. Learning non-maximum suppression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21 July–26 July 2017; pp. 4507–4515. [Google Scholar]
  35. Feng, X.; Han, J.; Yao, X.; Cheng, G. TCANet: Triple Context-Aware Network for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6946–6955. [Google Scholar] [CrossRef]
  36. Wang, B.; Zhao, Y.; Li, X. Multiple Instance Graph Learning for Weakly Supervised Remote Sensing Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5613112. [Google Scholar] [CrossRef]
  37. Feng, X.; Yao, X.; Cheng, G.; Han, J.; Han, J. SAENet: Self-Supervised Adversarial and Equivariant Network for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5610411. [Google Scholar] [CrossRef]
Figure 1. Illustration of problems and contributions of (a) HQIM, (b) WRA, and (c) HQIM + WRA.
Figure 1. Illustration of problems and contributions of (a) HQIM, (b) WRA, and (c) HQIM + WRA.
Electronics 13 04753 g001
Figure 2. The overview of our model. The RSI and its proposals are imported into a ConvNet, followed by the RoI pooling layer and two FC layers, to extract feature vectors for all proposals. The CSs of these proposals, obtained from the WSDDN, are then imported into the HQIM module. This module effectively mines positive instances by leveraging spatial distances and feature similarities among instances. Furthermore, the WRA strategy redistributes the loss weights of positive instances to encourage the model to focus on the whole object.
Figure 2. The overview of our model. The RSI and its proposals are imported into a ConvNet, followed by the RoI pooling layer and two FC layers, to extract feature vectors for all proposals. The CSs of these proposals, obtained from the WSDDN, are then imported into the HQIM module. This module effectively mines positive instances by leveraging spatial distances and feature similarities among instances. Furthermore, the WRA strategy redistributes the loss weights of positive instances to encourage the model to focus on the whole object.
Electronics 13 04753 g002
Figure 3. Parameter analysis of p in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Figure 3. Parameter analysis of p in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Electronics 13 04753 g003
Figure 4. Parameter analysis of δ in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Figure 4. Parameter analysis of δ in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Electronics 13 04753 g004
Figure 5. Parameter analysis of K in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Figure 5. Parameter analysis of K in terms of mAP on the DIOR (a) and NWPU VHR-10.v2 (b) datasets.
Electronics 13 04753 g005
Figure 6. A subjective validation of the HQIM module (a), the WRA strategy (b), and their combination (c) was conducted on the NWPU VHR-10.V2 and DIOR datasets. Specifically, the RSIs enclosed within green rectangles were selected from the DIOR dataset, and the remaining RSIs were selected from the NWPU VHR-10.V2 dataset.
Figure 6. A subjective validation of the HQIM module (a), the WRA strategy (b), and their combination (c) was conducted on the NWPU VHR-10.V2 and DIOR datasets. Specifically, the RSIs enclosed within green rectangles were selected from the DIOR dataset, and the remaining RSIs were selected from the NWPU VHR-10.V2 dataset.
Electronics 13 04753 g006
Figure 7. Visualization of some detection results of our model on the NWPU VHR-10.v2 dataset.
Figure 7. Visualization of some detection results of our model on the NWPU VHR-10.v2 dataset.
Electronics 13 04753 g007
Figure 8. Visualization of some detection results of our model on the DIOR dataset.
Figure 8. Visualization of some detection results of our model on the DIOR dataset.
Electronics 13 04753 g008
Table 1. Ablation study of HQIM module and WRA strategy on NWPU VHR-10.v2 and DIOR datasets.
Table 1. Ablation study of HQIM module and WRA strategy on NWPU VHR-10.v2 and DIOR datasets.
BaselineHQIMWRANWPU VHR-10.v2DIOR
mAP CorLoc mAP CorLoc
🗸 48.9661.5420.1042.79
🗸🗸 58.7969.5124.9148.54
🗸 🗸61.5172.9427.2350.89
🗸🗸🗸66.2476.8928.9153.92
Table 2. Comparisons with popular models on the NWPU VHR-10.v2 dataset in terms of mAP (%).
Table 2. Comparisons with popular models on the NWPU VHR-10.v2 dataset in terms of mAP (%).
MethodAirplaneShipStorage
Tank
Baseball
Diamond
Tennis
Court
Basketball
Court
Ground
Track Field
HarborBridgeVehiclemAP
Fast R-CNN [1]90.9190.6089.2947.32100.0085.8584.8688.2280.2969.8482.72
Faster R-CNN [9]90.9086.3090.5398.2489.7269.64100.0080.1161.4978.1484.51
WSDDN [2]30.0841.7235.9888.9012.8623.8599.4313.941.923.6035.12
OICR [11]13.6667.3557.1655.1613.6439.6692.800.231.843.7334.52
MIST [24]69.6949.1648.5580.9127.0879.8591.3446.998.2913.3651.52
DCL [16]72.7074.2537.0582.6436.8842.2783.9539.5716.8235.0052.11
PCIR [19]90.7878.8136.4090.8022.6452.1688.5142.3611.7435.4954.97
MIG [36]88.6971.6175.1794.1937.4547.68100.0027.278.339.0655.95
TCA [35]89.4378.1878.4290.8035.2750.3690.9142.444.1128.3058.82
SAE [37]82.9174.4750.2096.7455.6672.94100.0036.466.3331.8960.76
SPG [22]90.4281.0059.5392.3135.6451.4499.9258.7116.9942.9962.89
PISLM [25]87.6081.0057.3094.0036.4080.40100.0056.909.8035.6063.80
SGPLM [12]90.7079.9069.3097.5041.6077.50100.0044.4017.2033.5065.20
Ours90.8180.5273.4296.5947.3578.94100.0043.8918.3533.5166.24
Table 3. Comparisons with popular models on the NWPU VHR-10.v2 dataset in terms of CorLoc (%).
Table 3. Comparisons with popular models on the NWPU VHR-10.v2 dataset in terms of CorLoc (%).
MethodAirplaneShipStorage
Tank
Baseball
Diamond
Tennis
Court
Basketball
Court
Ground
Track Field
HarborBridgeVehiclemAP
WSDDN [2]22.3236.8139.9592.4817.9624.2499.2614.831.692.8935.24
OICR [11]29.4183.3320.5181.7640.8532.0886.607.413.7014.4440.01
MIST [24]90.2082.5080.3098.6048.5087.4098.3066.5014.6035.8070.30
DCL [16]----------69.70
PCIR [19]100.0093.0664.1099.3264.7979.2589.6962.9613.2652.2271.87
MIG [36]97.7990.2687.1898.6554.9364.15100.0074.0712.9621.5770.16
TCA [35]96.9191.7895.1388.6566.9062.8395.9854.1819.6355.5072.76
SAE [37]97.0691.6787.8198.6540.8681.13100.0070.3714.8152.2273.46
SPG [22]98.0692.6770.0899.6551.8680.1296.2072.4412.9960.0273.41
PISLM [25]94.4086.6068.5097.8069.8087.50100.0068.6016.0056.6074.60
SGPLM [12]98.2093.8089.3099.1050.2088.90100.0071.0012.3051.2075.40
Ours99.2994.7991.6897.9558.4391.67100.0068.7813.6456.6876.89
Table 4. Comparisons with popular models on the DIOR dataset in terms of mAP (%).
Table 4. Comparisons with popular models on the DIOR dataset in terms of mAP (%).
MethodAirplaneAirportBaseball
Field
Basketball
Court
BridgeChimneyDamExpressway
Service Area
Expressway
Toll Station
Golf Field
Fast R-CNN [1]44.1766.7966.9660.4915.5672.2851.9565.8744.7672.11
Faster R-CNN [9]50.2862.6066.0480.8828.8068.1747.2658.5148.0660.44
WSDDN [2]9.0639.6837.8120.160.2512.280.570.6511.884.90
OICR [11]8.7028.2644.0518.221.3020.150.090.6529.8913.80
MIST [24]32.0139.8762.7128.977.4612.870.315.1417.3851.02
DCL [16]20.8922.7054.2111.506.0361.010.091.0731.0130.87
PCIR [19]30.3736.0654.2226.609.0958.590.229.6536.1832.59
MIG [36]22.2052.5762.7625.788.4767.420.668.8528.7157.28
TCA [35]25.1330.8462.9240.004.1367.788.0723.8029.8922.34
SAE [37]20.5762.4162.6523.547.5964.620.2234.5230.6255.38
SPG [22]31.3236.6662.7929.106.0862.660.3115.0030.1035.00
PISLM [25]29.1049.8070.9041.407.2045.500.2035.4036.8060.80
SGPLM [12]39.1064.6064.4026.906.3062.300.9012.2026.3055.30
Ours42.1965.0166.1525.746.7060.151.2913.4825.3157.81
MethodGround
Track Field
HarborOverpassShipStadiumStorage
Tank
Tennis
Court
Train
Station
VehicleWindmillmAP
Fast R-CNN [1]62.9346.1838.0332.1370.9835.0458.2737.9119.2038.1049.98
Faster R-CNN [9]67.0043.8646.8758.4852.3742.3579.5248.0234.7765.4455.49
WSDDN [2]42.534.661.060.7063.033.956.060.514.551.1413.27
OICR [11]57.3910.6611.069.0959.297.100.680.149.090.4116.50
PCL [13]56.3616.7611.059.0957.629.092.470.124.554.518.19
MELM [15]41.0526.120.439.098.2815.0220.579.810.040.5318.65
MIST [24]49.485.3612.2429.4335.5325.360.814.5922.220.8022.18
DCL [16]56.455.052.659.0963.659.0910.360.027.270.7920.19
PCIR [19]58.518.6021.6312.0964.289.0913.620.309.097.5224.92
MIG [36]47.7323.770.776.4254.1313.154.1214.760.232.4325.11
TCA [35]53.8524.8411.069.0946.4013.7430.981.479.091.0025.82
SAE [37]52.7017.576.859.0951.5915.431.6914.441.419.1627.10
SPG [22]48.0227.1112.0010.0260.0415.1021.009.923.150.0625.77
PISLM [25]48.5014.0025.1018.5048.9011.7011.903.5011.301.7028.60
SGPLM [12]60.609.4023.1013.4057.4017.701.5014.0011.503.5028.50
Ours61.4210.4120.4014.1458.6218.912.1613.6110.024.6928.91
Table 5. Comparisons with popular models on the DIOR dataset in terms of CorLoc (%).
Table 5. Comparisons with popular models on the DIOR dataset in terms of CorLoc (%).
MethodAirplaneAirportBaseball
Field
Basketball
Court
BridgeChimneyDamExpressway
Service Area
Expressway
Toll Station
Golf Field
WSDDN [2]5.7259.8894.2455.944.9223.401.036.7944.5212.75
OICR [11]15.9851.4594.7755.792.6323.890.004.8256.6822.42
MIST [24]91.6053.2093.5066.3010.8030.701.5014.0335.2047.50
DCL [16]----------
PCIR [19]93.1045.6095.5068.303.6092.100.205.4058.4047.50
MIG [36]76.9846.8695.3963.6123.0095.070.2116.9657.8850.77
TCA [35]81.5851.3396.1773.455.0394.6915.8932.7945.9548.56
SAE [37]91.2069.3795.4867.5218.8897.780.2170.5454.3251.43
SPG [22]80.4832.0498.6865.0015.2096.0822.5216.9946.0850.96
PISLM [25]85.5068.9096.8075.8011.6094.700.8067.5060.5046.50
SGPLM [12]92.2058.3097.8074.2016.2095.200.3051.3056.2052.30
Ours994.1459.3498.1270.4617.5994.614.8052.1954.2753.47
MethodGround
Track Field
HarborOverpassShipStadiumStorage
Tank
Tennis
Court
Train
Station
VehicleWindmillCorLoc
WSDDN [2]89.905.4510.0022.9698.5479.6115.063.4511.563.2232.44
OICR [11]91.4118.1818.7031.8098.2881.297.451.2215.831.9834.77
MIST [24]87.1038.6023.4050.7080.5089.2022.4011.5022.202.4043.60
DCL [16]----------42.20
PCIR [19]88.6015.805.2039.5098.1085.6013.4056.509.700.6046.10
MIG [36]89.3942.1219.7837.9497.9380.6513.7710.3410.506.9446.80
TCA [35]85.2638.9120.1730.6384.5991.4656.283.7910.451.2548.41
SAE [37]88.2848.032.2833.5614.1183.3565.5919.8816.412.8549.42
SPG [22]89.1849.4522.0035.1698.6190.0432.5612.739.982.3448.30
PISLM [25]75.2050.5028.3039.7092.6077.0055.1010.1020.905.6053.20
SGPLM [12]91.7048.6023.0032.7098.8089.3043.5019.5018.304.0053.20
Ours93.0449.7122.0434.8199.0190.4244.1318.3117.5210.5153.92
Table 6. Comparisons with popular models on the NWPU VHR-10.V2 dataset in terms of mAP and computation cost.
Table 6. Comparisons with popular models on the NWPU VHR-10.V2 dataset in terms of mAP and computation cost.
MethodTraining Time (H)Inference Speed (fps)GFLOPS (G)mAP (%)
WSDDN [2]2.894.99287.6135.12
OICR [11]3.254.15286.2334.52
PCL [13]4.854.71287.1439.41
MELM [15]5.124.79290.3442.29
MIST [24]6.504.01290.4351.52
PISLM [25]9.833.58296.4363.80
SGPLM [12]10.504.21334.2665.20
Ours4.124.93289.9366.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xing, P.; Huang, M.; Wang, C.; Cao, Y. High-Quality Instance Mining and Weight Re-Assigning for Weakly Supervised Object Detection in Remote Sensing Images. Electronics 2024, 13, 4753. https://doi.org/10.3390/electronics13234753

AMA Style

Xing P, Huang M, Wang C, Cao Y. High-Quality Instance Mining and Weight Re-Assigning for Weakly Supervised Object Detection in Remote Sensing Images. Electronics. 2024; 13(23):4753. https://doi.org/10.3390/electronics13234753

Chicago/Turabian Style

Xing, Peixu, Mengxing Huang, Chenhao Wang, and Yang Cao. 2024. "High-Quality Instance Mining and Weight Re-Assigning for Weakly Supervised Object Detection in Remote Sensing Images" Electronics 13, no. 23: 4753. https://doi.org/10.3390/electronics13234753

APA Style

Xing, P., Huang, M., Wang, C., & Cao, Y. (2024). High-Quality Instance Mining and Weight Re-Assigning for Weakly Supervised Object Detection in Remote Sensing Images. Electronics, 13(23), 4753. https://doi.org/10.3390/electronics13234753

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop