Next Article in Journal
Pixel-Level Recognition of Trace Mycotoxins in Red Ginseng Based on Hyperspectral Imaging Combined with 1DCNN-Residual-BiLSTM-Attention Model
Next Article in Special Issue
Proposal-Free Fully Convolutional Network: Object Detection Based on a Box Map
Previous Article in Journal
Surfing Time–Motion Characteristics Possible to Gain Using Global Navigation Satellite Systems: A Systematic Review
Previous Article in Special Issue
Semi-TSGAN: Semi-Supervised Learning for Highlight Removal Based on Teacher-Student Generative Adversarial Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Stabilized Few-Shot Object Detection with Less Forgetting via Sample Normalization

School of Aeronautics and Astronautics, Sichuan University, Chengdu 610207, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3456; https://doi.org/10.3390/s24113456
Submission received: 2 March 2024 / Revised: 15 May 2024 / Accepted: 24 May 2024 / Published: 27 May 2024
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection: 2nd Edition)

Abstract

:
Few-shot object detection is a challenging task aimed at recognizing novel classes and localizing with limited labeled data. Although substantial achievements have been obtained, existing methods mostly struggle with forgetting and lack stability across various few-shot training samples. In this paper, we reveal two gaps affecting meta-knowledge transfer, leading to unstable performance and forgetting in meta-learning-based frameworks. To this end, we propose sample normalization, a simple yet effective method that enhances performance stability and decreases forgetting. Additionally, we apply Z-score normalization to mitigate the hubness problem in high-dimensional feature space. Experimental results on the PASCAL VOC data set demonstrate that our approach outperforms existing methods in both accuracy and stability, achieving up to +4.4 [email protected] and +5.3 mAR in a single run, with +4.8 [email protected] and +5.1 mAR over 10 random experiments on average. Furthermore, our method alleviates the drop in performance of base classes. The code will be released to facilitate future research.

1. Introduction

Object detection is a fundamental task in computer vision and has achieved particularly remarkable development with the help of large-scale annotated data sets [1,2,3] and well-designed powerful detectors [4,5,6,7,8,9,10]. However, in practical applications, the frequent occurrence of low-data scenarios limits the capability of detectors, making few-shot object detection (FSOD), which aims to solve such tasks with a small number of training samples, a critical approach in bringing object detection from theoretical research to real-world applications and deployments.
Many FSOD approaches have been developed to improve the generalization ability of neural networks, which can be mainly divided into transfer-learning-based [11,12,13,14,15,16,17] and meta-learning-based [18,19,20,21,22,23,24,25] methods. The former type aims to train suitable network parameters for invariant representation across domains and focuses on how to freeze fewer components of the detector without performance degradation [26,27]. Consistent with the general object detection framework, transfer-learning-based approaches present streamlined training pipelines without complex episodic procedures. However, such methods are limited by the few-shot setting to gain a map from the source domain to the target domain, which makes them lack generalization and robustness, facing serious overfitting and forgetting. In contrast, meta-learning usually consists of two branches, called support and query, and pays more attention to the aggregate information of these two branches to acquire the ability of ‘learning to learn’, which brings better robustness and generalization ability compared with transfer-learning-based methods.
However, we observe that meta-learning-based methods are generally unable to maintain consistent performance when trained on different data and may encounter the issue of forgetting. The detection results of meta methods vary widely, especially for lower-shot scenarios. As illustrated in Figure 1b, the classical meta-learning-based method Meta-DETR [18] exhibits instability when detecting a novel class, such as ‘cat’, failing to detect anything with the first two-shot training objects, and cannot distinguish between ‘dog’ and ‘cat’ with the third training samples. Most FSOD methods prioritize the performance for novel classes over a single run or average results over multiple experiments. Researchers of some methods [12,28] have noted the large sample variance due to few-shot samples and have focused on the worst performance under different samples, but they have not proposed an effective way to balance stability and accuracy. Additionally, meta-learning-based methods also suffer from the problem of forgetting, leading to poor performance in base classes especially under one-shot scenes, although several works [29] have attempted to address this problem with extremely complex architectures. We also note that for the object detection task, high-dimensional feature space makes the vectors suffer from the hubness problem [30,31], which refers to the K-nearest point of many other points in the high-latitude feature space and can damage detection performance.
In this paper, we explore two gaps in the few-shot setting for the support branch in meta-learning-based methods, which lead to instability and forgetting. To tackle these issues, we propose a simple but effective method named sample normalization to stabilize the meta-learning-based network and maintain consistent detection performance for both novel and base classes. We demonstrate the effectiveness of our method using Meta-DETR [18] as the backbone and apply sample normalization in the support branch. Additionally, we mitigate the negative impact of hubness by applying Z-score normalization [30] followed by a projection layer. Caused by the high-dimensional feature space, the hubness problem refers to the K-nearest point of many other points in the high-latitude feature space and can damage the performance. As shown in Figure 1c, with sample normalization, the network obtains more stable detection results for the novel class under different two-shot training samples. Experimental results show that our method presents competitive and stable performance and also reduces the problem of forgetting.
In summary, our main contributions are as follows:
  • We reveal two gaps caused by few-shot settings in the meta-learning framework between different data, which hurt both performance stability for novel classes and the reduction of the results concerning the base classes.
  • We propose a method named sample normalization under the meta-learning framework [18] to improve the ability of the network to migrate meta-knowledge to mitigate forgetting and enhance the stability of the network for different small samples.
  • Experimental results demonstrate that our method achieves competitive performance for the detection of both novel classes and base classes over a single run and multiple experiments and can stabilize the network when facing various few-shot training objects.

2. Related Work

2.1. Object Detection

In recent years, as a fundamental computer vision task, plenty of work has been undertaken for object detection, which can be mainly divided into one-stage, two-stage, and transformer-based methods. Based on RCNN [32], two-stage algorithms first generate proposals and revise them to accomplish classification and bounding box regression. Then, the following detectors [7,8,33,34,35] gradually improve the detection speed and accuracy. In contrast, one-stage methods like YOLO [4] and SSD [36] directly encode the feature map to predict bounding boxes and categories, and they are applicable for real-time detection but obtain lower accuracy. The following algorithms [5,6,37,38,39] achieve a huge boost by adding different modules and training tricks. In addition, the transformer has aroused great attention in computer vision since ViT [40] first utilized a transformer in classification, and the following methods [41,42,43] demonstrated its superior performance in object detection. DETR [9] predicts the set containing all objectives at once and remains a simple architecture, achieving state-of-the-art performance while meeting convergence slowly. Deformable DETR [10] samples keys near the reference point to compute attention and solve the problem of DETR [9]. In contrast to one-stage and two-stage methods, transformer-based methods obtain a more concise structure and more representative features with better performance.

2.2. Few-Shot Object Detection

Few-shot object detection has many applications in practical scenarios. Existing work on few-shot object detection can be divided into four paradigms: data augmentation, distance metric learning, transfer learning, and meta-learning. Data augmentation [44] and distance metric learning [45] are less considered as solutions for FSOD. The former may not learn reliable hypotheses with a few data available. Distance metric learning is useful for classification but not for location. The mainstream methods are largely based on meta-learning and transfer learning. The transferability of features has been demonstrated in [26], which establishes the foundation for both approaches. A classical transfer-learning method, LSTD [11] is based on SSD [36] and Faster-RCNN [7] and trained under detection loss and regression loss to fine-tune the network. Furthermore, the following algorithms [15,16,17,44] further improve the accuracy with approaches like multiscale structure, contrastive learning, and so on.
Transfer-learning-based methods, by directly applying object detection knowledge from large data sets to a few-shot setting, can result in issues like class imbalance, overfitting, and complexity mismatch, thereby compromising generalization performance. Meta-learning approaches [19,20,21] have introduced the concept of ‘learning to learn’ into FSOD, which is achieved by extracting generalized knowledge across different meta-tasks. These approaches [24,25] utilize relation distillation and contrastive learning. By using prototypes [46] as the support branch, some meta-learning methods [47] thoroughly explore the relationship between the query branch and prototypes to address the FSOD problem. Meta-DETR [18] solves few-shot object detection at the image level, bypassing inaccurate region proposals and effectively handling multiple support categories simultaneously. Despite the FSOD tasks, some research [30,31] shows that high-dimensional features meet hubness, damaging detection performance. They propose to normalize the hyperspherical representations with zero mean, avoiding this problem.
The work published in recent years has primarily focused on improving detection accuracy through complex networks, with less emphasis on the stability of performance when dealing with various few-shot training samples. Our goal is to strike a balance between detection precision and stability within a meta-learning framework. We thoroughly examined the gaps that exist in the feature space for meta-learning-based methods under the few-shot setting, which have been overlooked in current research. Instead of adding complex modules, we propose a simple yet effective method. Regardless, we endeavor to alleviate the issue of forgetting while maintaining performance on novel classes, an aspect that has been neglected in most existing methods.

3. Materials and Methods

3.1. Overview and Motivation

In this section, we present the mathematical definition of few-shot object detection, the overview of the meta-learning framework, and our motivation.
Problem definition: Generally, there is a set of source domain D s containing base classes C base with abundant labeled data for each class, while another set of target domain D t contains novel classes C novel , each of which has only a few annotated samples. We note that C base C novel = . If there are C classes in C novel and each class only owns K training data, it is called a C-way K-shot object detection task.
Overview: The overall pipeline of the proposed framework is illustrated in Figure 2, where we use the well-known Meta-DETR [18] as the backbone, given a query image I query = { x query } and a set of support images I support = { x k s } , k [ 1 , N ] with instance annotations, where N is the number of support samples once input to the net. I support is from various randomly sampled classes, and only ground-truth objects belonging to the support classes are kept as input of the support branch. A weight-shared feature extractor first encodes them into the same feature space and obtains corresponding feature maps F q = { f q } and F s = { f k s } . A transformer encoder obtains the feature vector S = { s k } of each support sample, taking task encoding T s = { t k s } of each support object into account. Subsequently, we use another transformer with a correlation aggregation module to polymerize the support feature and the query feature, exploiting the interclass correlation and obtaining the enhanced query feature f eq . Finally, the proposed sample normalization is applied in the support branch to obtain a relatively consistent input and weaken the gaps, with aggregates Z-score normalization before final detection.
Motivation: The instability of the meta-learning-based network under different few-shot data limits its practical applicability. We deeply consider two gaps in selecting support images, which lead to instability and negative transfer of meta-knowledge, resulting in the forgetting problem.
Gap between the sizes of training and test data: FSOD faces a challenge due to the distribution gap between training and test data sets in the target domain. The training samples are mostly fixed without artificial choice, which are randomly oriented toward the test data. Meta-DETR randomly samples support images from the few-shot data set during meta-fine-tuning. As shown in Figure 3a, few-shot samples used for training must be randomly combined in the feature space. Sample A selects images close between classes but far from the center of the category within the class, located at the class edge in different directions. Sample B chooses objects far away between classes, with large intercategory distances and relatively close intraclass distances to the class center. These sampling combinations exhibit significant differences in the feature space. With limited combinations in the few-shot data sets, the weights of neural networks are biased towards labeled data, which may cause mismatching for test data. When facing various few-shot data sets, the network performance obtains poor stability.
Gap between the sizes of the source and target domain: In meta-learning frameworks, the randomness in selecting support data from the source domain, due to the difference in data volume between the source and target domains, hinders the migration of meta-knowledge and leads to forgetting. Figure 3a illustrates various randomly sampled support object combinations in the feature space, such as sampling A, sampling B, sampling C, etc. These combinations in the source domain often do not align with the fixed few combinations in the target domain’s feature space, causing the migration of meta-knowledge. This discrepancy leads to the target domain’s detection not fully utilizing the meta-knowledge learned in the source domain, which is detrimental to detection. Subsequently, after fine-tuning, the network tends to favor the few-shot combination in the support branch, leading to the forgetting of various combinations in the source domain.
These two types of gaps, arising from both large and small sample sizes, constitute the core challenge in few-shot learning. One randomness arises from the transition from small to large data, while the other results from the transition in the opposite direction. To address these dual randomness issues, we introduce sample normalization within the meta-learning framework for support objects. This normalization aims to minimize the disparities between the source and target domains, as well as to harmonize the disparities among training and testing samples within the target domain. As shown in Figure 3b, we standardize the support data onto a unit hypersphere, ensuring consistency in the combinations of input support samples. Through this normalization process, we aim to generalize support features during meta-training and meta-fine-tuning, thereby enhancing the transfer of meta-knowledge and mitigating the forgetting of base classes. Additionally, this approach helps narrow the divide between training and test data, further enhancing the model’s performance.
To address these gaps, we propose sample normalization (SN) for the meta-learning framework, specifically targeting support objects. This normalization aims to mitigate the challenges posed by the large and small sample sizes, thereby reducing the discrepancies between the source and target domains and the differences between training and testing samples in the target domain.

3.2. Sample Normalization

We implement the sample normalization (SN) on S to obtain S = { s k } as
s k = s k μ σ ,
where k [ 1 , N ] , μ = 1 N k = 1 N s k , and σ 2 = k = 1 N s k μ 2 N 1 . This sample normalization is taken at each forward process, not at each episode, so this occurs for the distribution of the SN-passed support feature S i in the feature space in the i-th forward process. As shown in Figure 3b, we normalize the support data into a unit hypersphere and make combinations of input support samples standardized. The support features following SN S reside on the identical unit hypersphere, ensuring both the relative positioning of support samples and the aggregation of samples of the same class in limited-data sets. This configuration offers dual benefits:
(1)
It mitigates the disparity in data volume between base training and fine-tuning stages. Specifically, meta-knowledge acquisition entails progressively learning to discern query images using randomly sampled support objects, as depicted in Figure 2. Due to data volume disparities between source and target domains, the support feature combinations encountered during base training often mismatch with the limited support features of few-shot samples. Regardless of the training stage, SN selects support samples that reside on the same hypersphere. This alignment reduces representation gaps across training stages, enhancing meta-knowledge transfer and maintaining base class detection performance.
(2)
SN mitigates inconsistencies within random few-shot training data within a fixed target domain. In practical applications, target domain data remain constant for a single experiment but may vary across multiple experiments. When training support samples for the same category are randomly selected from different images, network weights can diverge, affecting stability. By implementing SN during fine-tuning, the transformer’s input reduces intraclass disparities while maintaining relative sample characteristics beyond mere categorical centroids.

3.3. Z-Score Normalization

In FSOD, the similarity of input features in the feature space due to the small number of samples can lead to the hubness problem. To address it, we apply Z-Score normalization (ZN) followed by a linear projection before final detection. Concretely, let f i , i [ 1 , 2 , . . . , D ] denote the i -th component of each enhanced query feature f eq R D . Furthermore, we first apply Z-score normalization to obtain f zn
f z n = f e q μ f 1 σ f R D ,
where μ f = 1 D i = 1 D f i , σ f 2 = 1 D i = 1 D f i μ f 2 , and 1 = 1 , 1 , 1 . . . , 1 T is a D -dimensional vector with its components. Subsequently, we apply a linear layer to obtain the final feature f :
f = W × f z n + B
where W and B are parameters which we train in each episode.
ZN projects the original feature vectors along 1 to a hyperplane which contains the original and is perpendicular to 1 . These vectors are then scaled to the same length of D , which can be calculated by f zn .

3.4. Network Training

In this section, we provide the detailed loss functions utilized to optimize our network and training procedure for the meta-learning framework.
Loss function: Assume that our meta framework infers N predictions as the same as the fixed number of object queries within a single feed-forward process. Furthermore, we denote the query image by x query and represent the ground-truth set of objects within x query by y = { y i } i = 1 N . When y i indicates an object, y i = c i , b i , where c i denotes the class label and b i denotes the bounding box of the object. When y i indicates no object, y i = ( , ) . Our meta-framework makes N predictions for target class, which are y ^ = { y i ^ } i = 1 N = { ( c i ^ , b i ^ ) } i = 1 N . Our method follows deformable DETR [10], which adopts a set-based Hungarian loss that forces unique prediction for each object via bipartite matching. We adopt a pair-wise matching loss L m a t c h ( y i , y σ ( i ) ^ ) to search for a bipartite matching between y ^ and y with the lowest cost:
σ ^ = a r g m i n σ i = 1 N L m a t c h ( y i , y σ ( i ) ^ ) ,
where σ denotes a permutation of N elements and σ ^ denotes the optimal assignment between predictions and targets. Since the matching should consider both classification and localization, the matching loss is defined as
L m a t c h ( y i , y σ ( i ) ^ ) = 1 { c i } L c l s ( c i , c ^ σ ( i ) ) + 1 { c i } L b o x ( b i , b ^ σ ( i ) ) .
With the optimal assignment of σ ^ obtained by Equations (4) and (5), we optimize the network using the following loss function:
L ( y , y ^ ) = i = 1 N [ L c l s ( c i , c ^ σ ( i ) ) + 1 { c i } L b o x ( b i , b ^ σ ( i ) ) ] ,
where we use sigmoid focal loss [48] for L c l s and apply a linear combination of 1 loss and GIoU loss [49] for L b o x . Furthermore, L ( y , y ^ ) is employed to each layer of the transformer decoder. According to Meta-DETR [18], we utilize a cosine similarity cross-entropy loss [50] to classify the class prototypes of various classes to be distinguished from each other.
Two-stage training strategy: Our method follows the training strategy made by Meta-DETR [18], consisting of meta-training and meta-fine-tuning. In meta-training, the support image and the query image are selected from D s as the input of the two branches with abundant samples from each base class. When fine-tuning, only K objects (K shots) are available for every novel category. Several objects from base classes are selected with K-shots novel objects to form a data set as the input of the support and query branches.

4. Experiment

4.1. Experimental Settings

Dataset: Pascal VOC [2,3] is a widely used FSOD data set containing a large number of annotated images across 20 categories. The model is trained on the Pascal VOC 07+12 train-evaluation data and evaluated on the test data in VOC 07 [2]. We employ the settings of Meta-DETR [18] for both the source and target domain categories, a common configuration in FSOD. To evaluate our approach, we take three types of target and source domain categories into account. For instance, we use (‘bird’, ‘bus’, ‘cow’, ‘motorbike’, ‘sofa’/others), (‘aeroplane’, ‘bottle’, ‘cow’, ‘horse’, ‘sofa’/others), and (‘boat’, ‘cat’, ‘motorbike’, ‘sheep’, ‘sofa’/others) as our target domain/source domain categories. We varied the number of few-shot sample shots from 1 to 10. Mean average precision (mAP) and mean average recall (mAR) values at an IoU threshold of 0.5 are used to measure the performance. We use different samples for training, obtaining individual and average performance results across 10 groups of randomly sampled few-shot data sets.
Implementation details: We adopt Resnet-101 [51] as the feature extractor and maintain the same hyperparameters and network architecture as [10]. For fair comparisons, the model is implemented in a single-scale version and aggregated following [52]. Our model is trained on one Nvidia A6000 GPU, and the AdamW optimizer with initial parameters is adopted. The batch size and the number of support objects are set to 16 and 5, respectively. For Pascal VOC, the initial learning rate of 2 × 10 4 and a weight decay of 1 × 10 4 are performed for 50 epochs of meta-training. Subsequently, we use the same settings as [18] to fine-tune the model on the few-shot data set until convergence and retain the best results.

4.2. Results and Discussion

Detection performance for novel classes: Table 1 presents the FSOD performance ([email protected]) for novel classes of Pascal VOC test07, which includes results from both single-run and multiple-run experiments. We conducted 10 runs with different random seeds, each utilizing different few-shot samples from the three class splits. For fair comparisons, we retrain Meta-DETR with the original settings denoted by *. This adjustment achieved better performance than results in [18]. As shown in Table 1, our method outperforms others in most K-shot settings across class splits in a single experiment. With sample normalization and Z-score normalization, we observe noticeable performance improvements with an increase of up to +0.3 [email protected] in 1 shot, +4.4 [email protected] in 2 shots, +3.8 [email protected] in 3 shots, +1.7 [email protected] in 5 shots, and +3.0 [email protected] in 10 shots. It proves our method can effectively enhance the detection capability of the meta-learning framework, particularly for larger K.
Across multiple runs with randomly sampled support objects, our method outperforms the original framework in average [email protected], particularly in class splits 1 and 3, with a notable margin of +3.6 [email protected] in 2 shots, +4.8 [email protected] in 3 shots, +1.5 [email protected] in 5 shots, and +2.8 [email protected] in 10 shots. Although our method may show slightly lower performance in split 2 and for lower shots in split 1 and 3, it still maintains strong detection ability compared to Meta-DETR. Table 1 demonstrates the superiority of our method in FSOD tasks, even with various few-shot training samples. For mAR, our method consistently outperforms the baseline, with improvements of up to +5.1 mAR in 1 shot, +3.9 mAR in 2 shots, +4.7 mAR in 3 shots, +3.2 mAR in 5 shots, and +3.7 mAR in 10 shots in a single run. Moreover, over multiple runs, our method achieves the following improvements: up to +4.8 mAR, +4.4 mAR, +5.1 mAR, +1.8 mAR, and +3.2 mAR, respectively, in 1, 2, 3, 5, and 10 shots, as shown in Table 2.
Detection stability: We analyze the stability of our method via variance of detection performance, including [email protected] and mAR, as shown in Table 3. Our method exhibits smaller variance in most settings, indicating improved stability. We can see that our method is particularly effective in stabilizing mAR.
We plot the scatter of results ([email protected] and mAR) in Figure 4 for 3 class splits. We can conclude that our method makes the performance more clustered facing various few-shot objects from (a)–(j), although there exist several outliers causing the increase of variance, specifically [email protected] in one shot and two shots of the first class split and in five shots of the second class split and so on.
In Figure 5 and Figure 6, we show our proposed sample normalization demonstrates superior detection performance across different two-shot training samples in class split 3. Additionally, even when faced with various changes in few-shot training samples, our method exhibits greater stability compared to the baseline [18], fully showcasing the enhancement of sample normalization on the stability of the meta-learning framework.
Detection performance for base classes: Our method demonstrates superior performance in both base and novel classes compared to other methods, as indicated in Table 4. Additionally, we show significant progress in reducing forgetting compared to the baseline, Meta-DETR. Specifically, we obtain +11.2 [email protected] in 1 shot, +13.7 [email protected] in 3 shots, +13.8 [email protected] in 5 shots, and +15.0 [email protected] in 10 shots, while we remain mostly the best performance in novel classes. Although there is a limited decrease in [email protected] after fine-tuning, our method retains effective base class distinction. Our proposed sample normalization with Z-score normalization consistently improves FSOD performance both in single runs and across multiple experiments. It enhances the stability of the meta-learning network when trained on diverse few-shot samples while maintaining superior performance for both base and novel classes.
Dense object detection: We were pleasantly surprised to find that our method performs exceptionally well in detecting dense targets, as demonstrated in Figure 7. Specifically, in challenging scenarios involving dense clustering, variable target sizes, or occlusions, our approach consistently outperforms the baseline, accurately identifying individual targets without missing any or falsely detecting entire clusters.

4.3. Ablation Study

In Table 5, we present the results ([email protected] and mAR) of the network with and without sample normalization and Z-score normalization for novel classes in the third class split of Pascal VOC. Sample normalization maintains [email protected] with only a slight decrease in most settings and demonstrates improved performance in three shots over multiple runs. It effectively enhances mAR for novel classes in both single and multiple experiments. Additionally, Z-score normalization improves [email protected] in lower shots and shows a slight enhancement in mAR performance for novel classes. When both SN and ZN are used, our method often achieves a balanced performance across all aspects.
Furthermore, as shown in Table 6, we present the detection results for base classes in a class split 3 of Pascal VOC. We can draw the conclusion that SN can effectively solve the problem of forgetting for Meta-DETR both in mAP and in mAR and can nearly retain the ability to detect base classes when the number of shots is greater than one. For instance, before meta-fine-tuning, the [email protected] for base classes is 92.6 and the mAR is 77.1.
We plot the results of [email protected] and mAR for the third class split in Pascal VOC [2,3], respectively, in Figure 4k,l. As illustrated in Figure 4k,l, we can draw the conclusion that our proposed sample normalization does affect the stability of the network and cluster the results in each shot.
In Figure 4k, there exist two outliers in one shot, representing the best and the worst results. Furthermore, we show the on3-shot training samples corresponding to these results in Figure 8. These two randomly selected targets exhibit extreme differences, with the first row achieving the best [email protected] of 43.5, while the second row only achieves 20.6. We analyze that the reason for this, which is that the sample normalization method causes the network to normalize relatively uniform samples in the feature space, without disrupting the relative positions.

5. Conclusions

Addressing the issue of instability and forgetting in the meta-learning framework, we deeply analyze two gaps: the first one is the gap between the sizes of training data and test data, which leads to instability, and the second is the gap between the sizes of the source domain and target domain, causing forgetting. We propose sample normalization applied in the support branch, fusing Z-score normalization to solve the problem of hubness, which obtains better performance especially compared with the baseline, i.e., Meta-DETR. Using two universal metrics, mAP and mAR, we prove that our method can not only stabilize Meta-DETR when facing various few-shot samples but also partly maintain the detection ability for base classes.
Limitations: Although our method has strong capabilities in stabilizing meta-learning frameworks and reducing source domain forgetting, in some extreme cases, such as the significant differences in samples under different random numbers mentioned above, the network always tends to favor one extreme, unable to balance the detection performance in both situations. Additionally, compared to some algorithms [47] that do not require fine-tuning and can be directly used for detecting new categories, the two-stage training strategy is quite complex and time-consuming. In the future, we will continue to explore more effective training methods and research on sample data distribution in order to eliminate the gap from few-shot to large-sample data.

Author Contributions

Conceptualization, Y.R. and M.Y.; Methodology, Y.R. and M.Y.; Software, Y.R. and Y.H.; Validation, Y.R., Y.H. and W.L.; Data curation, Y.R. and W.L.; Writing—original draft, Y.R.; Writing—review and editing, M.Y., Y.H. and W.L.; Visualization, Y.R.; Supervision, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China number 62271334 and the 173 key project number 2019-JCJQ-ZD-342-00.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  2. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes Challenge 2007 (VOC2007). Available online: http://www.pascal-network.org/challenges/VOC/voc2007/workshop/index.html (accessed on 7 April 2007).
  3. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes Challenge 2012 (VOC2012). Available online: http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html (accessed on 20 February 2012).
  4. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  5. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  6. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  8. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  9. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  10. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  11. Chen, H.; Wang, Y.; Wang, G.; Qiao, Y. Lstd: A low-shot transfer detector for object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  12. Wang, X.; Huang, T.E.; Darrell, T.; Gonzalez, J.E.; Yu, F. Frustratingly Simple Few-Shot Object Detection. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020. [Google Scholar]
  13. Zhu, C.; Chen, F.; Ahmed, U.; Shen, Z.; Savvides, M. Semantic relation reasoning for shot-stable few-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8782–8791. [Google Scholar]
  14. Wu, A.; Han, Y.; Zhu, L.; Yang, Y. Universal-prototype enhancing for few-shot object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9567–9576. [Google Scholar]
  15. Sun, B.; Li, B.; Cai, S.; Yuan, Y.; Zhang, C. Fsce: Few-shot object detection via contrastive proposal encoding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7352–7362. [Google Scholar]
  16. Gao, B.B.; Chen, X.; Huang, Z.; Nie, C.; Liu, J.; Lai, J.; Jiang, G.; Wang, X.; Wang, C. Decoupling Classifier for Boosting Few-shot Object Detection and Instance Segmentation. Adv. Neural Inf. Process. Syst. 2022, 35, 18640–18652. [Google Scholar]
  17. Zhang, X.; Liu, F.; Peng, Z.; Guo, Z.; Wan, F.; Ji, X.; Ye, Q. Integral migrating pre-trained transformer encoder-decoders for visual object detection. arXiv 2022, arXiv:2205.09613. [Google Scholar]
  18. Zhang, G.; Luo, Z.; Cui, K.; Lu, S.; Xing, E.P. Meta-DETR: Image-Level Few-Shot Detection with Inter-Class Correlation Exploitation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 12832–12843. [Google Scholar] [CrossRef] [PubMed]
  19. Wang, Y.X.; Ramanan, D.; Hebert, M. Meta-learning to detect rare objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9925–9934. [Google Scholar]
  20. Kang, B.; Liu, Z.; Wang, X.; Yu, F.; Feng, J.; Darrell, T. Few-shot object detection via feature reweighting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8420–8429. [Google Scholar]
  21. Yan, X.; Chen, Z.; Xu, A.; Wang, X.; Liang, X.; Lin, L. Meta r-cnn: Towards general solver for instance-level low-shot learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9577–9586. [Google Scholar]
  22. Fan, Q.; Zhuo, W.; Tang, C.K.; Tai, Y.W. Few-shot object detection with attention-RPN and multi-relation detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4013–4022. [Google Scholar]
  23. Li, B.; Yang, B.; Liu, C.; Liu, F.; Ji, R.; Ye, Q. Beyond max-margin: Class margin equilibrium for few-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7363–7372. [Google Scholar]
  24. Hu, H.; Bai, S.; Li, A.; Cui, J.; Wang, L. Dense relation distillation with context-aware aggregation for few-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10185–10194. [Google Scholar]
  25. Park, D.; Lee, J.M. Hierarchical attention network for few-shot object detection via meta-contrastive learning. arXiv 2022, arXiv:2208.07039. [Google Scholar]
  26. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar]
  27. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. Decaf: A deep convolutional activation feature for generic visual recognition. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 647–655. [Google Scholar]
  28. Fu, M.; Cao, Y.H.; Wu, J. Worst Case Matters for Few-Shot Recognition. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 99–115. [Google Scholar]
  29. Fan, Z.; Ma, Y.; Li, Z.; Sun, J. Generalized Few-Shot Object Detection Without Forgetting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4527–4536. [Google Scholar]
  30. Fei, N.; Gao, Y.; Lu, Z.; Xiang, T. Z-score normalization, hubness, and few-shot learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 142–151. [Google Scholar]
  31. Trosten, D.J.; Chakraborty, R.; Løkse, S.; Wickstrøm, K.K.; Jenssen, R.; Kampffmeyer, M.C. Hubs and Hyperspheres: Reducing Hubness and Improving Transductive Few-shot Learning with Hyperspherical Embeddings. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7527–7536. [Google Scholar]
  32. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  33. Girshick, R. Fast r-cnn. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  34. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  35. Qiao, S.; Chen, L.C.; Yuille, A. Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10213–10224. [Google Scholar]
  36. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision—ECCV 2016; Springer: Cham, Switzerland, 2016. [Google Scholar]
  37. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  38. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  39. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  40. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  41. Beal, J.; Kim, E.; Tzeng, E.; Park, D.H.; Zhai, A.; Kislyuk, D. Toward transformer-based object detection. arXiv 2020, arXiv:2012.09958. [Google Scholar]
  42. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  43. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. Deepvit: Towards deeper vision transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  44. Wu, J.; Liu, S.; Huang, D.; Wang, Y. Multi-scale positive sample refinement for few-shot object detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 456–472. [Google Scholar]
  45. Karlinsky, L.; Shtok, J.; Harary, S.; Schwartz, E.; Aides, A.; Feris, R.; Giryes, R.; Bronstein, A.M. Repmet: Representative-based metric learning for classification and few-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 5197–5206. [Google Scholar]
  46. Snell, J.; Swersky, K.; Zemel, R.S. Prototypical networks for few-shot learning. arXiv 2017, arXiv:1703.05175. [Google Scholar]
  47. Zhang, X.; Wang, Y.; Boularias, A. Detect Every Thing with Few Examples. arXiv 2023, arXiv:2309.12969. [Google Scholar]
  48. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  49. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  50. Chen, W.Y.; Liu, Y.C.; Kira, Z.; Wang, Y.C.F.; Huang, J.B. A Closer Look at Few-shot Classification. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  52. Xiao, Y.; Lepetit, V.; Marlet, R. Few-shot object detection and viewpoint estimation for objects in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3090–3106. [Google Scholar] [CrossRef] [PubMed]
  53. Zhang, W.; Wang, Y.X. Hallucination improves few-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13008–13017. [Google Scholar]
Figure 1. Few-shot training samples and corresponding object detection results. Each row is under a seed to select two-shot training samples: (a) presents two-shot training samples of a novel class ‘cat’; (b,c) reveal the results of detecting ‘cat’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR recognizes nothing in row 1 and misidentifies the ‘cat’ as ‘dog’ in row 3, while our method can achieve stable performance, as shown in (c).
Figure 1. Few-shot training samples and corresponding object detection results. Each row is under a seed to select two-shot training samples: (a) presents two-shot training samples of a novel class ‘cat’; (b,c) reveal the results of detecting ‘cat’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR recognizes nothing in row 1 and misidentifies the ‘cat’ as ‘dog’ in row 3, while our method can achieve stable performance, as shown in (c).
Sensors 24 03456 g001
Figure 2. Overall pipeline of our framework, where we use Meta-DETR [18] as the backbone. Sample normalization is applied after translating each support image into an embedding vector, and Z-score normalization is used before the final detection.
Figure 2. Overall pipeline of our framework, where we use Meta-DETR [18] as the backbone. Sample normalization is applied after translating each support image into an embedding vector, and Z-score normalization is used before the final detection.
Sensors 24 03456 g002
Figure 3. Multiple sampling strategies in feature space for support branch in the meta-learning framework. A hypersphere in the feature space represents a category, and different points in one red circle represent support images sampled in each episode for each category. We only display three sample combinations of A, B, and C. As shown in (a), without sample normalization, each support object is randomly sampled for its category and combined into random shapes in the feature space like sample A, sample B, and sample C. The images in (b) represent combinations of support objects after sample normalization, which are distributed on a standard hypersphere.
Figure 3. Multiple sampling strategies in feature space for support branch in the meta-learning framework. A hypersphere in the feature space represents a category, and different points in one red circle represent support images sampled in each episode for each category. We only display three sample combinations of A, B, and C. As shown in (a), without sample normalization, each support object is randomly sampled for its category and combined into random shapes in the feature space like sample A, sample B, and sample C. The images in (b) represent combinations of support objects after sample normalization, which are distributed on a standard hypersphere.
Sensors 24 03456 g003
Figure 4. Results dispersion for different shots under 10 seeds in Pascal VOC [2]. (ae) plot the [email protected] in 1, 2, 3, 5, and 10 shots, respectively. (fj) present scatter for mAR in 1, 2, 3, 5, and 10 shots, respectively, which are both for 3 class splits. In (k,l), we show the dispersion contrast between the Meta-DETR [18] with/without our proposed sample normalization.
Figure 4. Results dispersion for different shots under 10 seeds in Pascal VOC [2]. (ae) plot the [email protected] in 1, 2, 3, 5, and 10 shots, respectively. (fj) present scatter for mAR in 1, 2, 3, 5, and 10 shots, respectively, which are both for 3 class splits. In (k,l), we show the dispersion contrast between the Meta-DETR [18] with/without our proposed sample normalization.
Sensors 24 03456 g004
Figure 5. Few-shot training samples in class split 3 and corresponding object detection results for the novel class ‘boat’. Each row is under a seed to select 2-shot training samples. (a) presents 2-shot training samples of a novel class ‘boat’. (b,c) reveal the results of detecting ‘boat’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR recognizes nothing in rows 1–3 and provides a somewhat inaccurate bounding box in row 4, while our method can obtain stable performance as shown in (c).
Figure 5. Few-shot training samples in class split 3 and corresponding object detection results for the novel class ‘boat’. Each row is under a seed to select 2-shot training samples. (a) presents 2-shot training samples of a novel class ‘boat’. (b,c) reveal the results of detecting ‘boat’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR recognizes nothing in rows 1–3 and provides a somewhat inaccurate bounding box in row 4, while our method can obtain stable performance as shown in (c).
Sensors 24 03456 g005
Figure 6. Few-shot training samples in class split 3 and corresponding object detection results for the novel class ‘motorbike’. Each row is under a seed to select 2-shot training samples. (a) presents 2-shot training samples of a novel class ‘motorbike’. (b,c) reveal the results of detecting ‘motorbike’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR shows significant variance, while our algorithm demonstrates excellent stability and improved detection performance across different training samples as shown in (c).
Figure 6. Few-shot training samples in class split 3 and corresponding object detection results for the novel class ‘motorbike’. Each row is under a seed to select 2-shot training samples. (a) presents 2-shot training samples of a novel class ‘motorbike’. (b,c) reveal the results of detecting ‘motorbike’ using Meta-DETR [18] and our method, respectively. In (b), under various few-shot samples, Meta-DETR shows significant variance, while our algorithm demonstrates excellent stability and improved detection performance across different training samples as shown in (c).
Sensors 24 03456 g006
Figure 7. Comparison of detection results for occluded or densely packed objects. (a,c) show the detection results of Meta-DETR, while (b,d) correspond to the detection results of our method for the left images.
Figure 7. Comparison of detection results for occluded or densely packed objects. (a,c) show the detection results of Meta-DETR, while (b,d) correspond to the detection results of our method for the left images.
Sensors 24 03456 g007
Figure 8. One-shot training objects for class split 3 in Pascal VOC [2,3]. The images in row 1 and row 2 are randomly sampled from the same new classes using different random seeds. With vastly different training samples, Meta-DETR with the proposed sample normalization yields two extreme performances, achieving the best results ([email protected] 43.5) in row 1 and the inferior results ([email protected] 20.6) in row 2. We mark the bounding boxes and categories of novel classes, while only the bounding boxes of the base classes are marked. Some images are clipped for visualization.
Figure 8. One-shot training objects for class split 3 in Pascal VOC [2,3]. The images in row 1 and row 2 are randomly sampled from the same new classes using different random seeds. With vastly different training samples, Meta-DETR with the proposed sample normalization yields two extreme performances, achieving the best results ([email protected] 43.5) in row 1 and the inferior results ([email protected] 20.6) in row 2. We mark the bounding boxes and categories of novel classes, while only the bounding boxes of the base classes are marked. Some images are clipped for visualization.
Sensors 24 03456 g008
Table 1. Quantitative comparisons ([email protected]) on Pascal VOC test07 [2] for novel classes. We tested the results of a single experiment of 1, 2, 3, 5, and 10 shots in 3 class splits. ‘Multiple’ represents that results are averaged over 10 random runs, while ‘Single’ is over a single run. The samples for these 10 experiments were sampled based on 10 random numbers. The best results are highlighted in bold, and the second-best results are underlined. * denotes our reimplemented results.
Table 1. Quantitative comparisons ([email protected]) on Pascal VOC test07 [2] for novel classes. We tested the results of a single experiment of 1, 2, 3, 5, and 10 shots in 3 class splits. ‘Multiple’ represents that results are averaged over 10 random runs, while ‘Single’ is over a single run. The samples for these 10 experiments were sampled based on 10 random numbers. The best results are highlighted in bold, and the second-best results are underlined. * denotes our reimplemented results.
Method/ShotsClass Split 1Class Split 2Class Split 3
1 2 3 5 10 1 2 3 5 10 1 2 3 5 10
SingleLSTD [11]8.21.012.429.138.511.43.85.015.731.012.68.515.027.336.3
TFA w/cos [12]39.836.144.755.756.023.526.934.135.139.130.834.842.849.549.8
MPSR [44]41.743.151.455.261.824.429.539.239.347.835.640.642.348.049.7
Retentive R-CNN [29]42.445.845.953.756.121.727.835.237.040.330.237.643.049.750.1
TFA w/cos+Halluc [53]45.144.044.755.055.923.227.535.134.939.030.535.141.449.049.3
CME [23]41.547.550.458.260.927.230.241.442.546.834.439.645.148.351.5
SRR-FSD [13]47.850.551.355.256.832.535.339.140.843.840.141.544.346.946.4
FSCE [15]44.243.851.461.963.427.329.543.544.250.237.241.947.554.658.5
Meta-DETR [18] *47.757.163.363.167.137.141.243.649.154.644.348.661.265.368.6
Ours43.957.364.064.669.835.737.446.749.057.644.653.065.067.069.1
MultipleDeformable-DETR-ft-full [10]5.613.321.734.245.010.913.018.427.339.47.316.620.832.241.8
TFA w/cos [12]25.336.442.147.952.818.327.530.934.139.517.927.234.340.845.6
FsDetView [52]24.235.342.249.157.421.624.631.937.045.721.230.037.243.849.6
MPSR [44]34.742.646.149.456.722.630.531.036.743.327.532.538.244.650.5
DCNet [24]33.937.443.751.159.623.224.830.636.746.632.334.939.742.650.7
FSCE [15]32.944.046.852.959.723.730.638.443.048.522.633.439.547.354.0
Meta-DETR [18] *37.852.356.160.765.528.535.340.244.153.736.145.252.660.466.0
Ours36.352.859.762.268.327.132.239.743.353.334.548.857.461.267.4
Table 2. Quantitative comparisons (mAR) for the Meta-DETR [18] and our method on Pascal VOC [2]. The best results are highlighted in bold. * denotes our reimplemented results.
Table 2. Quantitative comparisons (mAR) for the Meta-DETR [18] and our method on Pascal VOC [2]. The best results are highlighted in bold. * denotes our reimplemented results.
Method/ShotsClass Split 1Class Split 2Class Split 3
1 2 3 5 10 1 2 3 5 10 1 2 3 5 10
SingleMeta-DETR *47.655.757.358.760.339.942.845.047.050.849.455.358.260.864.7
Ours48.956.258.559.262.441.444.446.149.154.554.559.263.564.067.5
MultipleMeta-DETR *43.752.753.457.058.933.938.541.444.349.443.450.854.859.961.9
Ours44.652.956.958.261.636.940.744.946.751.748.255.259.961.765.1
Table 3. The variance of the mean [email protected] of 10 seeds on Pascal test07 [2] for Meta-DETR [18] and our method. Lower is better, and the best results are highlighted in bold.
Table 3. The variance of the mean [email protected] of 10 seeds on Pascal test07 [2] for Meta-DETR [18] and our method. Lower is better, and the best results are highlighted in bold.
Method/ShotsClass Split 1Class Split 2Class Split 3
1 2 3 5 10 1 2 3 5 10 1 2 3 5 10
Meta-DETR33.78.211.92.71.324.513.724.810.310.026.07.520.813.32.1
Ours19.416.13.81.81.119.610.112.29.96.125.113.915.712.11.3
Table 4. Quantitative comparisons ([email protected]) for both base and novel classes on the first split of Pascal VOC [2]. The best results are highlighted in bold, and the second-best results are underlined. indicates results averaged on multiple runs.
Table 4. Quantitative comparisons ([email protected]) for both base and novel classes on the first split of Pascal VOC [2]. The best results are highlighted in bold, and the second-best results are underlined. indicates results averaged on multiple runs.
Method/ShotsBase ClassesNovel Classes
01 03 05 10 01 03 05 10
Meta-YOLO [20]66.464.863.463.614.826.733.947.2
FsDetView [52] 64.269.269.871.124.242.249.157.4
TFA w/cos [12] 77.677.377.477.525.342.147.952.9
MPSR [44] 60.665.968.269.834.746.149.456.7
FSCE [15] 75.573.775.075.232.946.852.959.7
Meta-DETR 67.270.073.073.537.856.160.765.5
Ours 78.483.786.888.536.359.762.268.3
Table 5. Ablation results ([email protected]) for novel classes in class split 3 on Pascal VOC 07 [2]. ‘Multiple’ represents that results are averaged over multiple random runs, while ‘Single’ is over a single run.
Table 5. Ablation results ([email protected]) for novel classes in class split 3 on Pascal VOC 07 [2]. ‘Multiple’ represents that results are averaged over multiple random runs, while ‘Single’ is over a single run.
SNZN[email protected]mAR
01 02 03 05 10 01 02 03 05 10
Single 44.348.661.265.368.649.455.358.260.864.7
53.649.856.564.367.055.159.058.061.064.7
43.550.261.559.168.251.358.960.661.665.2
44.653.065.067.069.154.559.263.564.067.5
Multiple 36.145.252.660.466.043.450.854.859.961.9
41.845.751.158.765.848.853.155.159.562.8
34.044.755.955.065.248.854.258.060.163.7
34.548.857.461.267.448.255.259.961.765.1
Table 6. Ablation results ([email protected] and mAR) in a single run for base classes in class split 3 of Pascal VOC 07 [2].
Table 6. Ablation results ([email protected] and mAR) in a single run for base classes in class split 3 of Pascal VOC 07 [2].
SNZN[email protected]mAR
01 02 03 05 10 01 02 03 05 10
63.175.977.185.985.455.263.362.870.470.6
82.388.387.090.690.967.773.372.475.675.1
78.887.988.089.490.467.973.072.474.474.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, Y.; Yang, M.; Han, Y.; Li, W. Towards Stabilized Few-Shot Object Detection with Less Forgetting via Sample Normalization. Sensors 2024, 24, 3456. https://doi.org/10.3390/s24113456

AMA Style

Ren Y, Yang M, Han Y, Li W. Towards Stabilized Few-Shot Object Detection with Less Forgetting via Sample Normalization. Sensors. 2024; 24(11):3456. https://doi.org/10.3390/s24113456

Chicago/Turabian Style

Ren, Yang, Menglong Yang, Yanqiao Han, and Weizheng Li. 2024. "Towards Stabilized Few-Shot Object Detection with Less Forgetting via Sample Normalization" Sensors 24, no. 11: 3456. https://doi.org/10.3390/s24113456

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop