Next Article in Journal
R-LRBPNet: A Lightweight SAR Image Oriented Ship Detection and Classification Method
Previous Article in Journal
Overview of High-Power and Wideband Radar Technology Development at MIT Lincoln Laboratory
Previous Article in Special Issue
MFIL-FCOS: A Multi-Scale Fusion and Interactive Learning Method for 2D Object Detection and Remote Sensing Image Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAM-Induced Pseudo Fully Supervised Learning for Weakly Supervised Object Detection in Remote Sensing Images

College of Electrical and Information Engineering, Zhengzhou University of Light Industry, Zhengzhou 450002, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(9), 1532; https://doi.org/10.3390/rs16091532
Submission received: 12 March 2024 / Revised: 19 April 2024 / Accepted: 23 April 2024 / Published: 26 April 2024

Abstract

:
Weakly supervised object detection (WSOD) in remote sensing images (RSIs) aims to detect high-value targets by solely utilizing image-level category labels; however, two problems have not been well addressed by existing methods. Firstly, the seed instances (SIs) are mined solely relying on the category score (CS) of each proposal, which is inclined to concentrate on the most salient parts of the object; furthermore, they are unreliable because the robustness of the CS is not sufficient due to the fact that the inter-category similarity and intra-category diversity are more serious in RSIs. Secondly, the localization accuracy is limited by the proposals generated by the selective search or edge box algorithm. To address the first problem, a segment anything model (SAM)-induced seed instance-mining (SSIM) module is proposed, which mines the SIs according to the object quality score, which indicates the comprehensive characteristic of the category and the completeness of the object. To handle the second problem, a SAM-based pseudo-ground truth-mining (SPGTM) module is proposed to mine the pseudo-ground truth (PGT) instances, for which the localization is more accurate than traditional proposals by fully making use of the advantages of SAM, and the object-detection heads are trained by the PGT instances in a fully supervised manner. The ablation studies show the effectiveness of the SSIM and SPGTM modules. Comprehensive comparisons with 15 WSOD methods demonstrate the superiority of our method on two RSI datasets.

1. Introduction

Compared with fully supervised object detection (FSOD) [1,2,3,4,5,6,7,8], the major advantage of weakly supervised object detection (WSOD) is that only image-level category annotations are necessary for training the WSOD model. Considering the low cost of data labeling, WSOD has been widely researched in recent years [9,10,11,12,13,14,15,16,17] and has been applied in scene classification [18,19], disaster detection [20,21], military [22,23], and other applications [24,25,26,27,28,29].
Weakly supervised deep detection networks (WSDDNs) [30] firstly combined deep learning with multiple instance learning (MIL) [30,31,32,33,34], and online instance classifier refinement (OICR) [31] introduced the instance classifier refinement (ICR) branch based on the WSDDN. OICR has been adopted as the baseline framework by many WSOD methods, and its process is briefly described as follows. Firstly, generate proposals through selective search (SS) [35], and import them into the backbone network to attain their features. Then, the instance-level category score (CS) of each proposal is obtained through MIL learning. Finally, the CSs are continuously optimized by several ICR branches, where the positive samples of each ICR branch include the seed instance (SI) and its neighbor instance, and the proposal with the highest CS predicted by the previous ICR branch is defined as the SI.
So far, there are still two problems that have not been well solved in the above baseline framework. For the first problem, the SIs that are mined solely relying on the CS are unreliable. On the one hand, existing methods usually select instances with high CSs as the SIs [31,32,36,37,38,39]; however, these instances usually focus on the most salient parts, rather than the whole object. On the other hand, compared with natural scene images (NSIs), remote sensing images (RSIs) have a more complex background; consequently, the reliability of the predicted CSs and SIs mined by the CSs is insufficient.
For the second problem, the localization capability of existing methods [34,40,41,42] is restricted by the proposals generated by the SS or edge boxes (EBs) [43]. As a matter of fact, most of the WSOD methods focus on improving the CS of each proposal, and the localization of each proposal is not changed. Although recent works added a regression branch for each ICR branch, where the SIs are used as the pseudo-ground truth (PGT) of each regression branch, the SIs are selected from the proposals; in other words, the localization of the SIs is still determined by the SS (EB). Considering that the SS (EB) is an early method in which the deep learning technique is not applied, its localization capability is limited, which has become a bottleneck of object localization for WSOD.
In order to address the aforementioned problems, a novel segment anything model (SAM)-induced pseudo-fully supervised learning (SPFS) model is proposed for WSOD in RSIs. Specifically, to overcome the first problem, as shown in Figure 1, a SAM-induced seed instance-mining (SSIM) module is proposed. First of all, the SAM-induced bounding boxes (SAMBs) are generated from the segmentation map inferred by SAM, where each SAMB tightly encloses one segment. Afterwards, the SAMBs are utilized to calculate the object completeness score (OCS) of each proposal. Finally, the OCS is integrated with the CS to obtain the object quality score (OQS), which is more reliable than the CS and can indicate the comprehensive characteristic of the object category and the object completeness, and the OQS is used to mine the SIs.
To overcome the second issue, a SAM-based pseudo-ground truth-mining (SPGTM) strategy is proposed for training the pseudo-fully supervised object-detection (PFSOD) head. The PGT instances of each PFSOD head are mined from the SAMBs according to the spatial relationship between the SAMBs and SIs of each ICR branch; consequently, the localization of mined PGT instances is more accurate than traditional proposals by fully making use of the advantages of SAM.
The main contributions of our method are as follows:
  • An SSIM module is proposed to address the issue of the SIs that are mined solely depending on the CS being unreliable. The SSIM module mines the SIs according to the OQS, which can indicate the comprehensive characteristic of the object category and the object completeness;
  • An SPGTM strategy is proposed to break the bottleneck of object localization brought by the SS or EB. The SPGTM strategy is utilized to mine PGT instances, for which the localization is more accurate than traditional proposals by fully making use of the advantages of SAM, and then, the PFSOD head is trained by using the PGT instances;
  • To our best knowledge, this is the first attempt to build a WSOD model by using the vision foundation model. It is worth noting that our SPFS model gives a unified solution of how to improve the localization capability of the WSOD model by using the segmentation technique; in other words, SAM is not the only choice for segmentation, and it can be replaced by better segmentation models in the future.

2. Related Works

To handle the problem that a single image contains multiple objects, many methods have changed from selecting the highest score of proposals to the local highest score for mining more SIs. On the one hand, some classical methods select the highest score of proposals as the SI. For example, Tang et al. [31] proposed an OICR model, which selected the highest prediction CS of the proposals as the SI, and then, the SI with its neighboring instances were used to train the WSOD model. Wu et al. [40] proposed to combine bottom-up aggregated attention (BUAA) and a phase-aware loss to select the highest CS of the proposals as the SI. Other similar works include [34,44], etc.
On the other hand, some improved methods select the local highest score of the proposals as the SIs. For example, Tang et al. [37] proposed a proposal-cluster-learning (PCL) scheme, which divided the prediction CS of the proposals into multiple clusters to mine the SIs for training the WSOD model. Ren et al. [32] proposed a multiple instance self-training (MIST) strategy, which selected the top prediction CSs as the candidate SIs, and the final SIs were determined by using the local NMS strategy among the candidate SIs. The mining high-quality pseudo-instance soft labels (MHQ-PSL) scheme [9] uses the proposal quality score to mine PGT instances and assign soft labels to each instance. Qian et al. [10] proposed an objectness score to mine SIs and incorporated the difficulty evaluation score into the training loss. The semantic segmentation-guided pseudo-label mining and instance re-detection (SGPLM-IR) method [11] proposes a class-specific object confidence score to mine PGT instances and an instance re-detection module to improve localization accuracy.

3. Basic WSOD Framework

3.1. Weakly Supervised Deep Detection Network

Currently, most of the WSOD methods use the OICR as the basic framework, which is constructed on the basis of the WSDDN, which is used to infer the initial CS of each proposal. Then, the OICR uses K ICR branches to continuously refine the CSs. The final detection results are obtained through non-maximum suppression (NMS) [45,46,47,48,49] in terms of the CSs.
Specifically, as shown in Figure 1, firstly, a series of proposals of the input image, denoted as R = { r 1 , r 2 , , r R } , is generated through the SS, where R denotes the amount of proposals. Secondly, the corresponding feature map is obtained by importing the image into the backbone, and then, the proposal is imported to the region of interest (RoI) pooling and two fully connected (FC) layers in the proper sequence to acquire the corresponding feature vector. Thirdly, the classification and detection score matrices are acquired by inputting the feature vectors into two parallel streams. The final score matrix, denoted as Z R V × R , is obtained through the elementwise product of two matrices. The final image-level prediction score of category v, denoted as I v , can be attained by the following equation:
I v = j = 1 R z v , j , v 1 , V
where V denotes the quantity of the category and  z v , j Z denotes the element in the vth row and jth column of Z. At this point, the WSDDN can be optimized by the following loss:
L B = v = 1 V y v log I v + ( 1 y v ) log ( 1 I v )
where L B represents the loss function of the WSDDN and y v 0 , 1 represents the image-level label of category v; y v = 1 if the image contains category v, and otherwise, y v = 0 .

3.2. Online Instance Classifier Refinement

Fourthly, the CS matrices, denoted as X k R V + 1 × R , are attained by inputting the feature vectors of all proposals into the kth ICR stream, where k 1 , 2 , , K and the ( V + 1 ) th category represents the background. The index of the proposal with the highest CS of category v in the ( k 1 ) th ICR stream is denoted as j v k 1 , and then, r j v k 1 is defined as the SI of category v in the kth ICR stream. The assemble of positive samples of category v in the kth ICR stream, denoted as P v k , consists of r j v k 1 and its neighbor proposals. Finally, the loss function L I C R k of the ICR branch is formulated as follows:
L I C R k = 1 R j = 1 R v = 1 V + 1 ω j k y v , j k log x v , j k
where x v , j k X k represents the CS of r j for category v in the kth ICR stream, ω j k = x v , j v k 1 denotes the weight of r j , and x v , j v k 1 X k 1 represents the CS of r j v k 1 for category v in the ( k 1 )th ICR stream; y v , j k = 1 if r j P v k , and otherwise, y v , j k = 0 .

3.3. Bounding Box Regression

Similar to recent WSOD methods [10,50], a bounding box regression (BBR) branch is attached to each ICR branch to improve the localization capability. The loss function L B B R k of the kth BBR stream is defined as follows:
L B B R k = 1 v = 1 V P v k v = 1 V i = 1 P v k s m o o t h L 1 t v , i k , t ^ v , i k
where P v k denotes the cardinality of P v k , s m o o t h L 1 · , · denotes the smooth L1 loss function [1], and t v , i k and t ^ v , i k denote the predicted and target localization offsets of the ith positive sample of category v in the kth BBR branch, respectively.

4. Proposed Method

4.1. Overview

As shown in Figure 1 and Section 3, our SPFS model utilizes the “OICR+BBR” as the basic WSOD framework, and the proposed SSIM and SPGTM modules are incorporated into it. Firstly, the RSI is fed into the backbone to obtain its feature maps, and the SS algorithm is used to extract the proposals from the RSI. Secondly, the above proposals are projected onto the feature maps, and their feature maps with a fixed size are generated through the RoI pooling operation, then the feature maps of the proposals are fed into two FC layers to extract the feature vector of each proposal. Thirdly, the feature vectors of all proposals are fed into the WSDDN trained by image-level labels to acquire the CS of all proposals, and the related details can be seen in Section 3.1. Fourthly, the segmentation map of the RSI is generated by using SAM (the related details can be seen in Section 4.2), and then, the rectangles are marked on it to obtain the SAMBs. Fifthly, the CSs of all proposals and SAMBs are fed into the SSIM module to mine high-quality SIs, and the related details can be seen in Section 4.3. Sixthly, the SIs and SAMBs are fed into the SPGTM module to obtain the PGT instances of the RSI, and the related details can be seen in Section 4.4. Seventhly, the feature vectors of all proposals are fed into the first PFSOD head trained by the above PGT instances to refine the CSs of all proposals. Steps 5∼step 7 are repeated K-times to accomplish the training of K PFSOD heads, and the related details can be seen in Section 4.5. Finally, the K trained PFSOD heads are jointly used to infer the detection results.

4.2. Segment Anything Model

SAM [51] is a proposed vision foundation model based on deep learning technology. Over eleven million images with over one billion masks are used to train SAM through a multi-stage learning scheme; thus, SAM can learn more complex image features and the morphology of foreground objects. Compared to traditional image-segmentation methods [52,53,54], SAM can precisely segment various objects with any shape and category in more complex scenarios, such as RSIs. Furthermore, SAM has an excellent capability in terms of being real time and in terms of accuracy, whether on static images or dynamic videos, which provides powerful support for various practical applications.

4.3. SAM-Induced Seed Instance Mining

First of all, the segmentation map of the input RSI is inferred by SAM, which is a powerful universal segmentation model, as shown in the top-left corner of Figure 1; each segment is visualized with different colors. Secondly, the SAMB of each segment is marked according to the coordinates of the four vertices of each segment; consequently, each segment is tightly enclosed by its SAMB.
Thirdly, the OCS of proposal r j , denoted as O C S j , is calculated through the following equation:   
O C S j = max I O U ( r j , b n ) n = 1 B , b n B
where B denotes the assemble of all SAMBs, B denotes the number of SAMBs, b n denotes the nth SAMB in B, I O U ( r j , b n ) denotes the IoU between r j and b n , and  max · denotes the operation of taking the maximum value. Equation (5) is used to measure the completeness by which each proposal covers the foreground object by using the advantages of SAM, which can precisely and fully cover each object.
Fourthly, the OQS of r j for category v in the kth ICR branch, denoted as O Q S v , j k , is calculated through the following equation:
O Q S v , j k = α x v , j k + 1 α O C S j
where α 0 , 1 is used to adjust the relative weight between x v , j k and O C S j As shown in Equation (6), the OQS of each proposal is composed of the CS and OCS; consequently, the SIs mined through the OQS have accurate category information and can cover the objects as much as possible. Inspired by MHQ-PSLs [9] and SGPLM-IR [11], α is defined as follows:
α = c i T I , c i < T T T I , c i T
where c i , T I , and T denote the current iteration steps, the total iteration steps, and the threshold of the iteration steps, respectively. The rationale of using Equation (7) is as follows. Our model was not well optimized at the beginning of the training; therefore, the reliability of x v , j k is not sufficient. At this point,  α should be given a small value. As training continues, the dependability of O C S j increases; thus, the value of α gradually increases. Considering the importance of O C S j , the upper bound of α should be restricted.
Finally, the assemble of all SIs of category v in the kth ICR branch, denoted as S v k = s v , l k l = 1 S v k , is mined by using O Q S v , j k and the mining strategy proposed by MIST [32], where s v , l k denotes the lth SI in S v k and S v k denotes the number of SIs in S v k .

4.4. SAM-Based Pseudo-Ground Truth Mining

Guided by the aforementioned SIs, the SPGTM module is used to mine the PGT instances from the SAMBs. On the one hand, the SIs usually contain accurate category information, although they only focus on the most salient part of the object. On the other hand, the foreground objects can be accurately localized by the SAMBs; however, the category information of the SAMBs is lacking. Consequently, the proposed SPGTM module mines the PGT instances of each PFSOD head by integrating the advantages of the SI and SAMB. The core idea of the SPGTM module is as follows. Some SAMBs that are considered as PGT instances are selected from the assemble of all SAMBs according to the spatial relationship between the SAMBs and SIs, and the categories of the selected SAMBs are copied from the matched SIs. The details of the SPGTM module can be seen in Algorithm 1.

4.5. Pseudo-Fully Supervised Training of Object-Detection Head

The aforementioned G k is used to train the kth PFSOD head in a fully supervised manner. The assemble of positive samples of category v in the kth PFSOD head, denoted as Q v k , is mined from the proposals according to the spatial distance between the proposals and G k . Specifically, the proposal r j is considered as a member of Q v k if I O U r j , g v , m k 0.5 , g v , m k g v k G k . At this point, Equations (3) and (4) can be reformulated as follows:  
L I C R k = 1 R j = 1 R v = 1 V + 1 ω j k h v , j k log x v , j k
L B B R k = 1 v = 1 V Q v k v = 1 V i = 1 Q v k s m o o t h L 1 t v , i k , t ^ v , i k
where h v , j k = 1 if r j Q v k , and otherwise, h v , j k = 0 . The training loss of the kth PFSOD head, denoted as L P k , is attained as follows:
L P k = L I C R k + L B B R k
Algorithm 1 SPGTM.
Input: B and S v k v = 1 V // B denotes the
  assemble of all SAMBs, and  S v k v = 1 V denotes the assemble of the SIs of all categories in the kth ICR branch.
Output: Assemble the PGT instance in the kth PFSOD head ( G k )
 1:
  Initialize ψ = ϕ , G k = ϕ
 2:
  for n=1 to B  do
 3:
    for v=1 to V do
 4:
        for l=1 to S v k  do
 5:
             if  I o U b n , s v , l k 0.5  then
 6:
                 Append ( ψ , s v , l k ) // For each SAMB, the SI for which the IoU between the
                 SAMB and it is greater than 0.5 is considered as the candidate matching SI of
                 the SAMB.
 7:
             end if
 8:
        end for
 9:
    end for
 10:
     if  ψ ϕ  then
 11:
          v * = argmax v I o U ( b n , s v , l k ) , s v , l k ψ , v * [ 1 , V ] // The SI for which the IoU
         between SAMB and it is the highest among all candidate matching SIs of the SAMB
         is selected as the final matching SI of the SAMB, and the category information of
         the matching SI is assigned to the SAMB.
 12:
      A PGT instance of category v * in the kth PFSOD head, denoted as g v * , m k g v * k , is
      obtained by assigning the category v * to b n , where m 1 , g v * k and g v * k G k
      denotes the assemble of PGT instances of category v * in the kth PFSOD head.
 13:
      Append ( g v * k , g v * , m k )
 14:
       ψ = ϕ
 15:
     end if
 16:
end for
 17:
G k = g v * k v * = 1 V // G k denotes the assemble of PGT instances in the kth PFSOD head.

4.6. Overall Training Loss and Inference

The total training loss of our SPFS model, denoted as L, is defined as follows:
L = L B + k = 1 K L P k
In the inference stage, SAM is not involved, and the average of the output of K PFSOD heads is used to infer the final detection results.

5. Experiments

5.1. Experiment Setup

5.1.1. Datasets

Our method was assessed on two RSI benchmarks, i.e., NWPU VHR-10.v2 [55,56] and DIOR [57]. The NWPU VHR-10.v2 dataset includes 1172 RSIs of 400 × 400 pixels, containing 10 categories and 2775 instances. The DIOR dataset includes 23,463 RSIs of 800 × 800 pixels, containing 20 categories and 192,472 instances. The 879 and 11,725 RSIs are used for training on the NWPU VHR-10.v2 and DIOR datasets, respectively, and the rest of the RSIs are used for testing. The detailed information of the two datasets is shown in Figure 2.

5.1.2. Metrics

This article utilizes the mean average precision (mAP) and correct localization (CorLoc) [58] to assess the overall detection and localization accuracy, respectively.

5.1.3. Implementation Details

The VGG-16 [59] pre-trained on ImageNet [60] was adopted as the backbone of our SPFS model. The number of PFSOD heads was set to 3, i.e., K = 3. The IoU threshold of NMS was set to 0.3 [41,61,62,63,64]. The stochastic gradient descent (SGD) algorithm [65,66,67,68] was used to optimize our model. The initial learning rate, batch size, momentum, and weight decay were set to 0.1, 8, 0.9, and 0.005, respectively. The total training iterations were set to 30K and 60K on the NWPU VHR-10.v2 and DIOR datasets, respectively. A step learning rate decay strategy was adopted, where the learning rate was reduced by 10% at iterations 20K and 26K on the NWPU VHR-10.v2 dataset, and the learning rate was reduced by 10% at iterations 50K and 56K on the DIOR dataset. All of training samples were augmented through rotation with 90° and 180° and horizontally flipping. The input images were stochastically resized to six scales { 480 , 576 , 688 , 864 , 1000 , 1200 } for multi-scale training and testing. Our experiments were based on the PyTorch framework running on 8 GPUs (8 × 24-GB memory, Titan RTX, NVIDIA, Santa Clara, CA, USA).

5.2. Parameter Analysis

5.2.1. Parameter α

As shown in Section 4.3, α was used to adjust the weight between two scores, and it was analyzed on the DIOR dataset, as shown in Figure 3; both the mAP and CorLoc achieved the highest scores when α was set to 0.5.

5.2.2. Parameter K

K (the number of PFSOD Heads) was analyzed on the DIOR dataset, as shown in Figure 4; both the mAP and CorLoc achieved the highest scores when K was set to 3.

5.3. Ablation Study

The ablation studies were conducted on the more challenging DIOR dataset to verify the effectiveness of the SSIM and SPGTM modules. As mentioned before, the “OICR+BBR” was adopted as the baseline [10,32,50]. As shown in Table 1, the mAP (CorLoc) increased by 5.63% and 6.71% (5.82% and 7.10%) when the SSIM and SPGTM modules were separately added to the baseline, respectively, which validates the effectiveness of the two modules. The mAP (CorLoc) improved by 10.80% (13.62%) when both the SSIM and SPGTM modules were added to the baseline, which proves the validity of the combination of the two modules.

5.4. Comparisons with Other Methods

5.4.1. Comparisons in Terms of mAP

As shown in Table 2 and Table 3, the mAP of our method achieved 73.4% (30.9%) on the NWPU VHR-10.v2 (DIOR) dataset, which outperformed the WSDDN [30], OICR [31], PCL [37], MELM [69], MIST [32], DCL [70], PCIR [41], MIG [34], TCA [42], SAE [64], SPG [38], MHQ-PSL [9], SGPLM-IR [11], RINet [62], and AE-IS [13]. Obviously, our method gave the optimal performance in terms of the mAP on the two RSI datasets.

5.4.2. Comparisons in Terms of CorLoc

As shown in Table 4 and Table 5, the CorLoc of our method achieved 78.7% (56.4%) on the NWPU VHR-10.v2 (DIOR) dataset, which outperformed the WSDDN, OICR, PCL, MELM, MIST, PCIR, MIG, TCA, SAE, SPG, MHQ-PSL, SGPLM-IR, RINet, and AE-IS. Obviously, our method also gave the best performance in terms of CorLoc on the two RSI datasets.
In conclusion, the overall capability of our method was superior to the other WSOD methods.
In addition, as shown in Table 2 and Table 3, our model was compared with two FSOD methods, i.e., fast R-CNN [1] and faster R-CNN [2]; the comparison results showed that our model significantly narrowed the gap between the WSOD and FSOD methods. Note that the capability of our method was almost comparable with the 2 FSOD methods on the NWPU VHR-10.v2 dataset.

5.5. Subjective Evaluation

Some detection results of our SPFS model on two RSI datasets are visualized in Figure 5 and Figure 6, respectively. Obviously, our model gave excellent results for most of the categories, which intuitively verified the effectiveness of our method.

6. Conclusions

A novel SPFS model that includes the SSIM and SPGTM modules is proposed in this article. First of all, the SIs mined by the CS tend to concentrate on the most significant parts of the target; moreover, they are unreliable because the reliability of the CS is not sufficient for RSIs with complex backgrounds. The SSIM module was proposed to address the aforementioned problem, which mines the SIs by using the OQS, which can represent the comprehensive characteristic of the category and the completeness of the target. Secondly, the localization capability of current methods is restricted because it solely depends on the proposals generated by the SS or EB, which is an early method and does not use the deep learning techniques. The SPGTM module was proposed to address the aforementioned problem, which can make full use of the advantages of SAM to obtain good-quality PGT instances, for which the localization was more accurate than the traditional proposals, and the PFSOD heads were trained by the PGT instances in a fully supervised manner. The ablation experiments verified the validity of the SSIM, SPGTM, and their integration. The quantitative comparisons with 15 WSOD methods demonstrated the excellent capability of our method.

Author Contributions

Conceptualization, X.Q.; formal analysis, X.Q.; funding acquisition, X.Q.; methodology, X.Q. and C.L.; project administration, W.W.; resources, W.W.; software, C.L.; supervision, W.W.; validation, W.W. and Z.C.; writing—original draft, C.L.; writing—review and editing, X.Q. and Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No. 62076223), the Key Research Project of Henan Province Universities (Grant No. 24ZX005), and the Key Science and Technology Program of Henan Province (Grant No. 232102211018).

Data Availability Statement

The NWPU VHR-10.v2 and DIOR datasets are available at the following URLs: https://drive.google.com/file/d/15xd4TASVAC2irRf02GA4LqYFbH7QITR-/view (accessed on 15 March 2023) and https://drive.google.com/drive/folders/1UdlgHk49iu6WpcJ5467iT-UqNPpx__CC (accessed on 15 March 2023), respectively.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
WSODweakly supervised object detection
RSIremote sensing image
SIseed instance
PGTpseudo-ground truth
SSIMSAM-induced seed instance mining
SPGTMSAM-based pseudo-ground truth mining
OCSobject completeness score
OQSobject quality score
PFSODpseudo-fully supervised object detection

References

  1. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  2. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  3. Qian, X.; Wu, B.; Cheng, G.; Yao, X.; Wang, W.; Han, J. Building a Bridge of Bounding Box Regression Between Oriented and Horizontal Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–9. [Google Scholar] [CrossRef]
  4. Li, L.; Yao, X.; Wang, X.; Hong, D.; Cheng, G.; Han, J. Robust Few-Shot Aerial Image Object Detection via Unbiased Proposals Filtration. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
  5. Cheng, G.; Li, Q.; Wang, G.; Xie, X.; Min, L.; Han, J. SFRNet: Fine-Grained Oriented Object Recognition via Separate Feature Refinement. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–10. [Google Scholar] [CrossRef]
  6. Xie, X.; Lang, C.; Miao, S.; Cheng, G.; Li, K.; Han, J. Mutual-Assistance Learning for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15171–15184. [Google Scholar] [CrossRef]
  7. Xie, X.; Cheng, G.; Li, Q.; Miao, S.; Li, K.; Han, J. Fewer is more: Efficient object detection in large aerial images. Sci. China Inf. Sci. 2024, 67, 112106. [Google Scholar] [CrossRef]
  8. Liang, Y.; Feng, J.; Zhang, X.; Zhang, J.; Jiao, L. MidNet: An anchor-and-angle-free detector for oriented ship detection in aerial images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  9. Qian, X.; Huo, Y.; Cheng, G.; Gao, C.; Yao, X.; Wang, W. Mining High-Quality Pseudoinstance Soft Labels for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  10. Qian, X.; Huo, Y.; Cheng, G.; Yao, X.; Li, K.; Ren, H.; Wang, W. Incorporating the completeness and difficulty of proposals into weakly supervised object detection in remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1902–1911. [Google Scholar] [CrossRef]
  11. Qian, X.; Li, C.; Wang, W.; Yao, X.; Cheng, G. Semantic segmentation guided pseudo label mining and instance re-detection for weakly supervised object detection in remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2023, 119, 103301. [Google Scholar] [CrossRef]
  12. Qian, X.; Wang, C.; Li, C.; Li, Z.; Zeng, L.; Wang, W.; Wu, Q. Multiscale Image Splitting Based Feature Enhancement and Instance Difficulty Aware Training for Weakly Supervised Object Detection in Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 7497–7506. [Google Scholar] [CrossRef]
  13. Xie, X.; Cheng, G.; Feng, X.; Yao, X.; Qian, X.; Han, J. Attention Erasing and Instance Sampling for Weakly Supervised Object Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–10. [Google Scholar] [CrossRef]
  14. Wu, Z.; Wen, J.; Xu, Y.; Yang, J.; Li, X.; Zhang, D. Enhanced spatial feature learning for weakly supervised object detection. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 961–972. [Google Scholar] [CrossRef]
  15. Wu, Z.; Wen, J.; Xu, Y.; Yang, J.; Zhang, D. Multiple instance detection networks with adaptive instance refinement. IEEE Trans. Multimed. 2021, 25, 267–279. [Google Scholar] [CrossRef]
  16. Zhang, D.; Li, H.; Zeng, W.; Fang, C.; Cheng, L.; Cheng, M.M.; Han, J. Weakly Supervised Semantic Segmentation via Alternate Self-Dual Teaching. IEEE Trans. Image Process. 2023, 72, 1. [Google Scholar] [CrossRef]
  17. Zhang, D.; Guo, G.; Zeng, W.; Li, L.; Han, J. Generalized weakly supervised object localization. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 5395–5406. [Google Scholar] [CrossRef]
  18. Tong, W.; Chen, W.; Han, W.; Li, X.; Wang, L. Channel-attention-based DenseNet network for remote sensing image scene classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 60, 4121–4132. [Google Scholar] [CrossRef]
  19. Chen, W.; Ouyang, S.; Tong, W.; Li, X.; Zheng, X.; Wang, L. GCSANet: A global context spatial attention deep learning network for remote sensing scene classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 60, 1150–1162. [Google Scholar] [CrossRef]
  20. Tekumalla, R.; Banda, J.M. TweetDIS: A large twitter dataset for natural disasters built using weak supervision. In Proceedings of the 2022 IEEE International Conference on Big Data (Big Data), Osaka, Japan, 17–20 December 2022; Volume 60, pp. 4816–4823. [Google Scholar] [CrossRef]
  21. Presa-Reyes, M.; Tao, Y.; Chen, S.C.; Shyu, M.L. Deep learning with weak supervision for disaster scene description in low-altitude imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–10. [Google Scholar] [CrossRef]
  22. Tang, W.; Deng, C.; Han, Y.; Huang, Y.; Zhao, B. SRARNet: A unified framework for joint superresolution and aircraft recognition. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 327–336. [Google Scholar] [CrossRef]
  23. He, Q.; Sun, X.; Yan, Z.; Li, B.; Fu, K. Multi-object tracking in satellite videos with graph-based multitask modeling. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  24. Lin, S.; Zhang, M.; Cheng, X.; Shi, L.; Gamba, P.; Wang, H. Dynamic Low-Rank and Sparse Priors Constrained Deep Autoencoders for Hyperspectral Anomaly Detection. IEEE Trans. Instrum. Meas. 2023, 73, 2500518. [Google Scholar] [CrossRef]
  25. Lin, S.; Zhang, M.; Cheng, X.; Zhou, K.; Zhao, S.; Wang, H. Hyperspectral Anomaly Detection via Sparse Representation and Collaborative Representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 946–961. [Google Scholar] [CrossRef]
  26. Lin, S.; Zhang, M.; Cheng, X.; Zhou, K.; Zhao, S.; Wang, H. Dual Collaborative Constraints Regularized Low-Rank and Sparse Representation via Robust Dictionaries Construction for Hyperspectral Anomaly Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 2009–2024. [Google Scholar] [CrossRef]
  27. Cheng, X.; Zhang, M.; Lin, S.; Li, Y.; Wang, H. Deep Self-Representation Learning Framework for Hyperspectral Anomaly Detection. IEEE Trans. Instrum. Meas. 2023, 73, 5002016. [Google Scholar] [CrossRef]
  28. Cheng, X.; Zhang, M.; Lin, S.; Zhou, K.; Zhao, S.; Wang, H. Two-Stream Isolation Forest Based on Deep Features for Hyperspectral Anomaly Detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  29. Huo, Y.; Qian, X.; Li, C.; Wang, W. Multiple Instance Complementary Detection and Difficulty Evaluation for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  30. Bilen, H.; Vedaldi, A. Weakly supervised deep detection networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2846–2854. [Google Scholar]
  31. Tang, P.; Wang, X.; Bai, X.; Liu, W. Multiple instance detection network with online instance classifier refinement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2843–2851. [Google Scholar]
  32. Ren, Z.; Yu, Z.; Yang, X.; Liu, M.Y.; Lee, Y.J.; Schwing, A.G.; Kautz, J. Instance-aware, context-focused, and memory-efficient weakly supervised object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10598–10607. [Google Scholar]
  33. Yin, Y.; Deng, J.; Zhou, W.; Li, L.; Li, H. Fi-wsod: Foreground information guided weakly supervised object detection. IEEE Trans. Multimed. 2022, 25, 1890–1902. [Google Scholar] [CrossRef]
  34. Wang, B.; Zhao, Y.; Li, X. Multiple instance graph learning for weakly supervised remote sensing object detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  35. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
  36. Wei, Y.; Shen, Z.; Cheng, B.; Shi, H.; Xiong, J.; Feng, J.; Huang, T. Ts2c: Tight box mining with surrounding segmentation context for weakly supervised object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 434–450. [Google Scholar] [CrossRef]
  37. Tang, P.; Wang, X.; Bai, S.; Shen, W.; Bai, X.; Liu, W.; Yuille, A. PCL: Proposal Cluster Learning for Weakly Supervised Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 176–191. [Google Scholar] [CrossRef]
  38. Cheng, G.; Xie, X.; Chen, W.; Feng, X.; Yao, X.; Han, J. Self-guided proposal generation for weakly supervised object detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  39. Xia, R.; Li, G.; Huang, Z.; Meng, H.; Pang, Y. CBASH: Combined backbone and advanced selection heads with object semantic proposals for weakly supervised object detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6502–6514. [Google Scholar] [CrossRef]
  40. Wu, Z.; Liu, C.; Wen, J.; Xu, Y.; Yang, J.; Li, X. Selecting high-quality proposals for weakly supervised object detection with bottom-up aggregated attention and phase-aware loss. IEEE Trans. Image Process. 2022, 32, 682–693. [Google Scholar] [CrossRef]
  41. Feng, X.; Han, J.; Yao, X.; Cheng, G. Progressive contextual instance refinement for weakly supervised object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8002–8012. [Google Scholar] [CrossRef]
  42. Feng, X.; Han, J.; Yao, X.; Cheng, G. TCANet: Triple context-aware network for weakly supervised object detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6946–6955. [Google Scholar] [CrossRef]
  43. Zitnick, C.L.; Dollár, P. Edge boxes: Locating object proposals from edges. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13, 2014. pp. 391–405. [Google Scholar] [CrossRef]
  44. Lin, C.; Wang, S.; Xu, D.; Lu, Y.; Zhang, W. Object instance mining for weakly supervised object detection. Proc. AAAI Conf. Artif. Intell. 2020, 34, 11482–11489. [Google Scholar] [CrossRef]
  45. Feng, J.; Liang, Y.; Zhang, X.; Zhang, J.; Jiao, L. SDANet: Semantic-embedded density adaptive network for moving vehicle detection in satellite videos. IEEE Trans. Image Process. 2023, 32, 1788–1801. [Google Scholar] [CrossRef]
  46. Feng, J.; Bai, G.; Li, D.; Zhang, X.; Shang, R.; Jiao, L. MR-selection: A meta-reinforcement learning approach for zero-shot hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–20. [Google Scholar] [CrossRef]
  47. Qian, X.; Zeng, Y.; Wang, W.; Zhang, Q. Co-Saliency Detection Guided by Group Weakly Supervised Learning. IEEE Trans. Multimed. 2023, 25, 1810–1818. [Google Scholar] [CrossRef]
  48. Feng, J.; Gao, Z.; Shang, R.; Zhang, X.; Jiao, L. Multi-complementary generative adversarial networks with contrastive learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–18. [Google Scholar] [CrossRef]
  49. Qian, X.; Zhang, N.; Wang, W. Smooth giou loss for oriented object detection in remote sensing images. Remote Sens. 2023, 15, 1259. [Google Scholar] [CrossRef]
  50. Seo, J.; Bae, W.; Sutherland, D.J.; Noh, J.; Kim, D. Object Discovery via Contrastive Learning for Weakly Supervised Object Detection. In Proceedings of the Computer Vision—ECCV 2022; Springer Nature: Cham, Switzerland, 2022; pp. 312–329. [Google Scholar]
  51. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. arXiv 2023, arXiv:2304.02643. [Google Scholar] [CrossRef]
  52. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 15–17 June 2016; pp. 2921–2929. [Google Scholar] [CrossRef]
  53. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  54. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar] [CrossRef]
  55. Cheng, G.; Zhou, P.; Han, J. Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  56. Li, K.; Cheng, G.; Bu, S.; You, X. Rotation-Insensitive and Context-Augmented Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2337–2348. [Google Scholar] [CrossRef]
  57. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  58. Deselaers, T.; Alexe, B.; Ferrari, V. Weakly supervised localization and learning with generic knowledge. Int. J. Comput. Vis. 2012, 100, 275–293. [Google Scholar] [CrossRef]
  59. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  60. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  61. Hosang, J.; Benenson, R.; Schiele, B. Learning non-maximum suppression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4507–4515. [Google Scholar]
  62. Feng, X.; Yao, X.; Cheng, G.; Han, J. Weakly Supervised Rotation-Invariant Aerial Object Detection Network. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 14126–14135. [Google Scholar] [CrossRef]
  63. Chen, Z.; Fu, Z.; Jiang, R.; Chen, Y.; Hua, X.S. Slv: Spatial likelihood voting for weakly supervised object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12995–13004. [Google Scholar] [CrossRef]
  64. Feng, X.; Yao, X.; Cheng, G.; Han, J.; Han, J. SAENet: Self-Supervised Adversarial and Equivariant Network for Weakly Supervised Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  65. Yang, K.; Zhang, P.; Qiao, P.; Wang, Z.; Dai, H.; Shen, T.; Li, D.; Dou, Y. Rethinking Segmentation Guidance for Weakly Supervised Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 4069–4073. [Google Scholar] [CrossRef]
  66. Wang, G.; Zhang, X.; Peng, Z.; Jia, X.; Tang, X.; Jiao, L. MOL: Towards accurate weakly supervised remote sensing object detection via Multi-view nOisy Learning. ISPRS J. Photogramm. Remote Sens. 2023, 196, 457–470. [Google Scholar] [CrossRef]
  67. Chen, M.; Tian, Y.; Li, Z.; Li, E.; Liang, Z. Online Progressive Instance-Balanced Sampling for Weakly Supervised Vibration Damper Detection. IEEE Trans. Instrum. Meas. 2023, 72, 1–14. [Google Scholar] [CrossRef]
  68. Wang, G.; Zhang, X.; Peng, Z.; Tang, X.; Zhou, H.; Jiao, L. Absolute wrong makes better: Boosting weakly supervised object detection via negative deterministic information. arXiv 2022, arXiv:2204.10068. [Google Scholar] [CrossRef]
  69. Wan, F.; Wei, P.; Jiao, J.; Han, Z.; Ye, Q. Min-entropy latent model for weakly supervised object detection. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1297–1306. [Google Scholar]
  70. Yao, X.; Feng, X.; Han, J.; Cheng, G.; Guo, L. Automatic weakly supervised object detection from high spatial resolution remote sensing images via dynamic curriculum learning. IEEE Trans. Geosci. Remote Sens. 2020, 59, 675–685. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed method. The WSDDN, SSIM, and SPGTM modules and PFSOD heads are introduced in Section 3.1, Section 4.2, Section 4.3, and Section 4.4, respectively.
Figure 1. The framework of the proposed method. The WSDDN, SSIM, and SPGTM modules and PFSOD heads are introduced in Section 3.1, Section 4.2, Section 4.3, and Section 4.4, respectively.
Remotesensing 16 01532 g001
Figure 2. The number of object instances for each category in the two datasets.
Figure 2. The number of object instances for each category in the two datasets.
Remotesensing 16 01532 g002
Figure 3. The values of (a) mAP and (b) CorLoc with respect to different α on the DIOR dataset.
Figure 3. The values of (a) mAP and (b) CorLoc with respect to different α on the DIOR dataset.
Remotesensing 16 01532 g003
Figure 4. The values of (a) mAP and (b) CorLoc with respect to different K on the DIOR dataset.
Figure 4. The values of (a) mAP and (b) CorLoc with respect to different K on the DIOR dataset.
Remotesensing 16 01532 g004
Figure 5. Visualization of some results of the SPFS model on the NWPU VHR-10.v2 dataset.
Figure 5. Visualization of some results of the SPFS model on the NWPU VHR-10.v2 dataset.
Remotesensing 16 01532 g005
Figure 6. Visualization of some results of the SPFS model on the DIOR dataset.
Figure 6. Visualization of some results of the SPFS model on the DIOR dataset.
Remotesensing 16 01532 g006
Table 1. Ablation studies of SSIM and SPGTM on the DIOR dataset.
Table 1. Ablation studies of SSIM and SPGTM on the DIOR dataset.
BaselineSSIMSPGTMmAPCorLoc
20.1042.79
25.7348.61
26.8149.89
30.9056.41
Table 2. The comparisons of average precision (%) with other methods on the NWPU VHR-10.v2 dataset. The bold font denotes the best result. The same below.
Table 2. The comparisons of average precision (%) with other methods on the NWPU VHR-10.v2 dataset. The bold font denotes the best result. The same below.
MethodAirplaneShipStorage
Tank
Baseball
Diamond
Tennis
Court
Basketball
Court
Ground
Track Field
HarborBridgeVehiclemAP
Fast R-CNN [1]90.9190.6089.2947.32100.0085.8584.8688.2280.2969.8482.72
Faster R-CNN [2]90.9086.3090.5398.2489.7269.64100.0080.1161.4978.1484.51
WSDDN [30]30.0841.7235.9888.9012.8623.8599.4313.941.923.6035.12
OICR [31]13.6667.3557.1655.1613.6439.6692.800.231.843.7334.52
PCL [37]26.0063.762.5089.8064.4576.0777.940.001.3015.6739.41
MELM [69]80.8669.3010.4890.1712.8420.1499.1717.1014.178.6842.29
MIST [32]69.6949.1648.5580.9127.0879.8591.3446.998.2913.3651.52
DCL [70]72.7074.2537.0582.6436.8842.2783.9539.5716.8235.0052.11
PCIR [41]90.7878.8136.4090.8022.6452.1688.5142.3611.7435.4954.97
MIG [34]88.6971.6175.1794.1937.4547.68100.0027.278.339.0655.95
TCA [42]89.4378.1878.4290.8035.2750.3690.9142.444.1128.3058.82
SAE [64]82.9174.4750.2096.7455.6672.94100.0036.466.3331.8960.76
SPG [38]90.4281.0059.5392.3135.6451.4499.9258.7116.9942.9962.89
MHQ-PSL [9]87.6081.0057.3094.0036.4080.40100.0056.909.8035.6063.80
SGPLM-IR [11]90.7079.9069.3097.5041.6077.50100.0044.4017.2033.5065.20
RINet [62]90.3086.3079.6090.7058.2080.40100.0057.7018.9041.6070.40
AE-IS [13]91.0088.2078.3093.2060.6082.40100.0060.4019.6045.8072.00
SPFS (ours)91.2383.3273.6490.5673.1085.28100.0063.5910.2463.5273.45
Table 3. The comparisons of average precision (%) with other methods on the DIOR dataset.
Table 3. The comparisons of average precision (%) with other methods on the DIOR dataset.
MethodAirplaneAirportBaseball
Field
Basketball
Court
BridgeChimneyDamExpressway
Service Area
Expressway
Toll Station
Golf Field
Fast R-CNN [1]44.1766.7966.9660.4915.5672.2851.9565.8744.7672.11
Faster R-CNN [2]50.2862.6066.0480.8828.8068.1747.2658.5148.0660.44
WSDDN [30]9.0639.6837.8120.160.2512.280.570.6511.884.90
OICR [31]8.7028.2644.0518.221.3020.150.090.6529.8913.80
PCL [37]21.5235.1959.8023.492.9543.710.120.901.492.88
MELM [69]28.143.2362.5128.720.0662.510.2128.3913.0915.15
MIST [32]32.0139.8762.7128.977.4612.870.315.1417.3851.02
DCL [70]20.8922.7054.2111.506.0361.010.091.0731.0130.87
PCIR [41]30.3736.0654.2226.609.0958.590.229.6536.1832.59
MIG [34]22.2052.5762.7625.788.4767.420.668.8528.7157.28
TCA [42]25.1330.8462.9240.004.1367.788.0723.8029.8922.34
SAE [64]20.5762.4162.6523.547.5964.620.2234.5230.6255.38
SPG [38]31.3236.6662.7929.106.0862.660.3115.0030.1035.00
MHQ-PSL [9]29.1049.8070.9041.407.2045.500.2035.4036.8060.80
SGPLM-IR [11]39.1064.6064.4026.906.3062.300.9012.2026.3055.30
RINet [62]26.2057.4062.7025.109.9069.201.4013.3036.2051.40
AE-IS [13]31.8050.9063.2029.408.9068.701.3015.1035.5051.60
SPFS (ours)35.9462.8966.0830.539.7169.771.9312.8834.9050.49
MethodGround
Track Field
HarborOverpassShipStadiumStorage
Tank
Tennis
Court
Train
Station
VehicleWindmillmAP
Fast R-CNN [1]62.9346.1838.0332.1370.9835.0458.2737.9119.2038.1049.98
Faster R-CNN [2]67.0043.8646.8758.4852.3742.3579.5248.0234.7765.4455.49
WSDDN [30]42.534.661.060.7063.033.956.060.514.551.1413.27
OICR [31]57.3910.6611.069.0959.297.100.680.149.090.4116.50
PCL [37]56.3616.7611.059.0957.629.092.470.124.554.518.19
MELM [69]41.0526.120.439.098.2815.0220.579.810.040.5318.65
MIST [32]49.485.3612.2429.4335.5325.360.814.5922.220.8022.18
DCL [70]56.455.052.659.0963.659.0910.360.027.270.7920.19
PCIR [41]58.518.6021.6312.0964.289.0913.620.309.097.5224.92
MIG [34]47.7323.770.776.4254.1313.154.1214.760.232.4325.11
TCA [42]53.8524.8411.069.0946.4013.7430.981.479.091.0025.82
SAE [64]52.7017.576.859.0951.5915.431.6914.441.419.1627.10
SPG [38]48.0227.1112.0010.0260.0415.1021.009.923.150.0625.77
MHQ-PSL [9]48.5014.0025.1018.5048.9011.7011.903.5011.301.7028.60
SGPLM-IR [11]60.609.4023.1013.4057.4017.701.5014.0011.503.5028.50
RINet [62]53.9028.604.809.1052.7015.8020.6012.909.104.7028.30
AE-IS [13]52.3028.8013.3011.2056.9016.3022.4014.008.002.6029.10
SPFS (ours)56.9225.2926.3015.1852.2912.6225.8113.8811.613.1030.90
Table 4. The comparisons of correct localization (%) with other methods on the NWPU VHR-10.v2 dataset.
Table 4. The comparisons of correct localization (%) with other methods on the NWPU VHR-10.v2 dataset.
MethodAirplaneShipStorage
Tank
Baseball
Diamond
Tennis
Court
Basketball
Court
Ground
Track Field
HarborBridgeVehiclemAP
WSDDN [30]22.3236.8139.9592.4817.9624.2499.2614.831.692.8935.24
OICR [31]29.4183.3320.5181.7640.8532.0886.607.413.7014.4440.01
PCL [37]11.7650.0012.8298.6584.5177.3690.720.009.2615.5645.06
MELM [69]85.9677.4221.4398.3310.7143.4895.0040.0011.7614.6349.87
MIST [32]90.2082.5080.3098.6048.5087.4098.3066.5014.6035.8070.30
PCIR [41]100.0093.0664.1099.3264.7979.2589.6962.9613.2652.2271.87
MIG [34]97.7990.2687.1898.6554.9364.15100.0074.0712.9621.5770.16
TCA [42]96.9191.7895.1388.6566.9062.8395.9854.1819.6355.5072.76
SAE [64]97.0691.6787.8198.6540.8681.13100.0070.3714.8152.2273.46
SPG [38]98.0692.6770.0899.6551.8680.1296.2072.4412.9960.0273.41
MHQ-PSL [9]94.4086.6068.5097.8069.8087.50100.0068.6016.0056.6074.60
SGPLM-IR [11]98.2093.8089.3099.1050.2088.90100.0071.0012.3051.2075.40
AE-IS [13]98.3094.2072.40100.0056.8083.6098.4076.8018.2062.4076.10
SPFS (ours)95.4990.3281.5390.1870.1089.94100.078.9019.3870.8178.67
Table 5. The comparisons of correct localization (%) with other methods on the DIOR dataset.
Table 5. The comparisons of correct localization (%) with other methods on the DIOR dataset.
MethodAirplaneAirportBaseball
Field
Basketball
Court
BridgeChimneyDamExpressway
Service Area
Expressway
Toll Station
Golf Field
WSDDN [30]5.7259.8894.2455.944.9223.401.036.7944.5212.75
OICR [31]15.9851.4594.7755.792.6323.890.004.8256.6822.42
PCL [37]61.1446.8695.3963.617.3295.070.215.715.1450.77
MELM [69]76.9828.9492.6663.0113.0090.090.2137.8816.9644.62
MIST [32]91.6053.2093.5066.3010.8030.701.5014.0335.2047.50
PCIR [41]93.1045.6095.5068.303.6092.100.205.4058.4047.50
MIG [34]76.9846.8695.3963.6123.0095.070.2116.9657.8850.77
TCA [42]81.5851.3396.1773.455.0394.6915.8932.7945.9548.56
SAE [64]91.2069.3795.4867.5218.8897.780.2170.5454.3251.43
SPG [38]80.4832.0498.6865.0015.2096.0822.5216.9946.0850.96
MHQ-PSL [9]85.5068.9096.8075.8011.6094.700.8067.5060.5046.50
SGPLM-IR [11]92.2058.3097.8074.2016.2095.200.3051.3056.2052.30
RINet [62]92.7080.9092.7069.508.6090.100.2071.3062.0065.50
AE-IS [13]91.4078.6096.1068.8016.0092.3022.8068.9060.6062.70
SPFS (ours)90.8280.1398.5965.5019.4590.382.3570.6162.2466.51
MethodGround
Track Field
HarborOverpassShipStadiumStorage
Tank
Tennis
Court
Train
Station
VehicleWindmillmAP
WSDDN [30]89.905.4510.0022.9698.5479.6115.063.4511.563.2232.44
OICR [31]91.4118.1818.7031.8098.2881.297.451.2215.831.9834.77
PCL [37]89.3942.1219.7837.9497.9380.6513.770.2010.506.9441.52
MELM [69]88.0849.3915.6528.1998.2882.9722.7510.344.622.2343.34
MIST [32]87.1038.6023.4050.7080.5089.2022.4011.5022.202.4043.60
PCIR [41]88.6015.805.2039.5098.1085.6013.4056.509.700.6046.10
MIG [34]89.3942.1219.7837.9497.9380.6513.7710.3410.506.9446.80
TCA [42]85.2638.9120.1730.6384.5991.4656.283.7910.451.2548.41
SAE [64]88.2848.032.2833.5614.1183.3565.5919.8816.412.8549.42
SPG [38]89.1849.4522.0035.1698.6190.0432.5612.739.982.3448.30
MHQ-PSL [9]75.2050.5028.3039.7092.6077.0055.1010.1020.905.6053.20
SGPLM-IR [11]91.7048.6023.0032.7098.8089.3043.5019.5018.304.0053.20
RINet [62]85.1051.4015.7044.6098.6080.3014.8022.706.902.6052.80
AE-IS [13]88.2050.9023.4040.2098.8091.5033.2018.4012.302.5055.90
SPFS (ours)93.2850.8125.5642.6195.7883.5549.5415.6918.895.8856.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qian, X.; Lin, C.; Chen, Z.; Wang, W. SAM-Induced Pseudo Fully Supervised Learning for Weakly Supervised Object Detection in Remote Sensing Images. Remote Sens. 2024, 16, 1532. https://doi.org/10.3390/rs16091532

AMA Style

Qian X, Lin C, Chen Z, Wang W. SAM-Induced Pseudo Fully Supervised Learning for Weakly Supervised Object Detection in Remote Sensing Images. Remote Sensing. 2024; 16(9):1532. https://doi.org/10.3390/rs16091532

Chicago/Turabian Style

Qian, Xiaoliang, Chenyang Lin, Zhiwu Chen, and Wei Wang. 2024. "SAM-Induced Pseudo Fully Supervised Learning for Weakly Supervised Object Detection in Remote Sensing Images" Remote Sensing 16, no. 9: 1532. https://doi.org/10.3390/rs16091532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop