Next Article in Journal
Soil Data Cube and Artificial Intelligence Techniques for Generating National-Scale Topsoil Thematic Maps: A Case Study in Lithuanian Croplands
Next Article in Special Issue
Frequency Agile Anti-Interference Technology Based on Reinforcement Learning Using Long Short-Term Memory and Multi-Layer Historical Information Observation
Previous Article in Journal
Influence of Climate, Topography, and Hydrology on Vegetation Distribution Patterns—Oasis in the Taklamakan Desert Hinterland
Previous Article in Special Issue
Faster and Better: A Lightweight Transformer Network for Remote Sensing Scene Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Few-Shot High-Resolution Range Profile Ship Target Recognition Based on Task-Specific Meta-Learning with Mixed Training and Meta Embedding

1
College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
2
Nanjing Research Institute of Electronics Engineering, Nanjing 210007, China
3
Department of Electronic and Computer Engineering, University of Calgary, Calgary, AB T2P 2M5, Canada
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(22), 5301; https://doi.org/10.3390/rs15225301
Submission received: 13 September 2023 / Revised: 4 November 2023 / Accepted: 7 November 2023 / Published: 9 November 2023
(This article belongs to the Special Issue Remote Sensing and Machine Learning of Signal and Image Processing)

Abstract

:
High-resolution range profile (HRRP), characterized by its high availability and rich target structural information, has been extensively studied. However, HRRP-based target recognition methods using closed datasets exhibit limitations when it comes to identifying new classes of targets. The scarcity of samples for new classes leads to overfitting during the deep learning process, and the similarity in the scattering structures of different ships, combined with the significant structural differences among samples of the same ship, contribute to a high level of confusion among targets. To address these challenges, this paper proposed Task-Specific Mate-learning (TSML) for few-shot HRRP. Firstly, a Task-Adaptive Mixed Transfer (TAMT) strategy is proposed, which combines basic learning with meta-learning, to reduce the likelihood of overfitting and enhance adaptability for recognizing new classes of ships. Secondly, a Prototype Network is introduced to enable the recognition of new classes of targets with limited samples. Additionally, a Space-Adjusted Meta Embedding (SAME) is proposed based on the Prototype Network. This embedding function, designed for HRRP data, modifies the distances between samples in meta-tasks by increasing the distances between samples from different ships and decreasing the distances between samples from the same ship. The proposed method is evaluated based on an actual measured HRRP dataset and the experimental results prove that the proposed method can more accurately recognize the unknown ship classes with a small number of labels by learning the known classes of ships. In addition, the method has a degree of robustness to the number of training samples and a certain generalization ability, which can produce improved results when applied to other backbones.

Graphical Abstract

1. Introduction

Maritime control and situational awareness have garnered increasing significance, especially in the context of classifying and identifying maritime targets. In this regard, considerable attention has been devoted by researchers to radar automatic target recognition (RATR), which has led to the development of various classification approaches. Based on the type of radar data, target classification methods can be broadly categorized as radar cross section (RCS)-based methods [1,2], synthetic aperture radar (SAR)-image-based methods [3,4,5,6,7,8] and high-resolution range profiles (HRRP)-based methods [9,10,11]. HRRP, which is the projection of the spatial scattering structure of the target in the direction of the radar line of sight (LOS) [12], provides essential information such as target size, scattering structure, and position. In comparison to SAR images, HRRP offers the benefits of simple acquisition, easy storage, and straightforward processing. Consequently, HRRP-based target recognition methods have gained popularity in the field of RATR.
Traditional target recognition techniques have the advantage of requiring less data. For instance, Ning et al. [13] proposed a non-equidistant margin plane (NM) in high-dimensional feature space to enhance object recognition accuracy for HRRP data. Wang et al. [14] incorporated a defined separability measure based on the distribution of training samples into the construction of a decision tree, and proposed an improved support vector machine (SVM) decision tree. Furthermore, Chai et al. [15] developed a weighted k-nearest neighbors (KNN) classifier to mitigate the effects of outliers before cooperative target identification. However, the majority of feature extraction methods are unsupervised and lossy, as well as highly dependent on the knowledge and experience with HRRP data of researchers. This means that these methods do not focus well on the maximum separability features, and it can be challenging to obtain the expected results. Compared with the above methods, deep learning can produce better results when applied to HRRP-based RATR [16,17,18,19]. It is essential to note, however, that these methods are primarily trained and tested on specific datasets, and the classifiers used during testing have been exposed to a substantial number of samples from known categories. This category-complete, azimuthally comprehensive scenario is referred to as the closed dataset setting for HRRP target recognition. Networks trained under this paradigm are limited to recognizing specific target types.
In reality, a more intricate situation arises in which radar target recognition systems are required to identify novel classes of targets. However, typically, samples are scarce for these novel classes, giving rise to a class of open-set recognition problems which we refer to as the few-shot problem. When the number of labeled HRRP samples is limited, existing research has predominantly employed two approaches to address overfitting during deep learning. The first approach involves expanding existing data to generate additional samples [20,21,22], while the second strategy combines semi-supervised learning with transfer learning to facilitate knowledge-sharing [23,24]. Nevertheless, when data are seriously lacking, the effectiveness of these techniques is limited. Sufficient training data may not be generated through data expansion to cover diverse data distributions, and transfer learning may only transfer general prior knowledge among related domains, which may prove insufficient for efficient model learning when significant amounts of data are missing. Therefore, these methods are insufficient to address the problem of insufficient data and overfitting in deep learning.
Few-shot learning (FSL) [25] is a type of meta-learning that addresses the challenge of target classification when only a limited number of labeled instances are available. FSL involves two sets of visual concepts: seen classes that have sufficient labeled instances and unseen classes that have few labeled instances. The objective of FSL is to develop a visual classifier that can identify unseen classes using seen classes, making it task-agnostic. This means that the trained model can be applied to the recognition task of novel classes. However, FSL is still in the developmental stage in the field of optical images, and there are fewer studies on HRRP-based few-shot target recognition [24,26]. Moreover, existing FSL methods are not entirely applicable to HRRP due to two primary reasons. Firstly, the scattering structures of different ship classes are still very similar, which leads to small differences between HRRP samples of different ship classes. Secondly, high-resolution radar, being a large integrated system, is inevitably affected by various environmental factors when acquiring the HRRPs of a target. When the attitude of the target changes relative to the LOS, the position of each scatter on the target changes relative to the position of the radar. This change causes relatively substantial differences between HRRP samples of the same target. Applying current FSL methods for optical images to the field of HRRP is challenging due to the difficulty in accurately classifying ship samples that belong to the same class but have relatively large differences. Therefore, it is essential to emphasize the distinctions between classes and reduce the variation within each class when working with HRRP data.
To address the challenge of recognizing new classes of HRRP targets with limited samples and high target confusability, a method for HRRP target recognition is introduced in this paper. Firstly, base-learning and meta-learning methods are employed to mitigate the issue of overfitting in few-shot HRRP recognition. Few-shot recognition of HRRP is addressed during the meta-learning phase. And in the base-training phase, the feature extractor, through modeling all HRRP samples, acquires robust feature extraction capabilities, thereby providing a solid foundation for initializing the weights in the meta-learning phase. During the weight transfer process, the training parameters are reduced by freezing shallow-level blocks and retaining feature weights acquired during the base learning phase to extract envelope features and local physical characteristics from HRRP samples. Several fine-tuning strategies are applied to the deep-level block weights to adapt them to new classification tasks. This transfer strategy alleviates issues related to feature loss and poor adaptability to new tasks during the transfer process, reducing data redundancy and, consequently, minimizing overfitting risk in meta-training. Furthermore, the proposed method enhances discriminative embeddings through an additional embedding function based on prototype networks during meta-training. Specifically, Graph Convolutional Networks (GCNs) are utilized in the meta-training phase to embed support set samples. GCNs capture sample relationships and similarities, aligning samples of the same class in the embedding space, which results in prototypes, defined as the mean embedding vectors for each class. Multi-head Attention is applied to prototype embeddings, assigning feature weights to accentuate subtle feature distinctions among categories, thereby improving prototype distinguishability in the embedding space. Samples in the query set, once embedded by GCN, are classified based on their proximity to the nearest class prototypes. This part, based on the prototype network, not only addresses the problem of recognizing novel classes, but also, by embedding samples and prototypes differently, reduces classification confusion, thereby enhancing recognition accuracy.
Our overall contribution is thus four-fold:
(1)
To address the limitations of recognizing specific target types in the HRRP target identification task on a closed dataset, the prototype network is introduced in this paper. This enables the identification of novel class targets with limited samples.
(2)
To mitigate the overfitting issues that can arise when dealing with a high number of network parameters during meta-training, the Task-Adaptive Mixed Transfer is proposed, which reduces the training parameters, avoiding the redundancy in parameter fine-tuning and computational resource wastage, without introducing information loss during the transfer process.
(3)
To address the challenge of high target confusion during the identification process, a Space-Adjusted Meta Embedding for HRRP based on the prototype network is proposed, which generates discriminative embedding vectors and prototypes, thus improving the accuracy of target classification in HRRP.
(4)
The proposed method in this paper is evaluated based on the actual measured HRRP data. Experiments, both in terms of quantitative results and visualizations, demonstrate that the proposed method can more accurately identify unknown ship classes with a small number of labels by learning the known classes of ships. Furthermore, by varying the number of training samples, we find that the method exhibits a degree of robustness to the number of training samples.

2. Background

2.1. HRRP

HRRP [12] is a technique used for projecting the spatial scattering structure of a target in the direction of LOS, as shown in Figure 1. The scattering point model employed by this technique partitions the target along the LOS into several distance units. Assuming the absence of noise and interference, let the i distance unit contain N i scattering points of the target, i = 1 , 2 , 3 , , n , and n is the total number of distance units. The sum of all target scattering point echoes is superimposed in the i distance unit, and the expression can be represented as follows:
x ( i ) = j = 1 N i a i j e j 2 π f τ i j
where a i j represents the intensity coefficient of the j scattering point; τ i j is the echo delay caused by the j scattering point. x ( i ) contains the amplitude information and phase information of the target, in which the amplitude information is generally used in radar target identification. Thus, each ship sample can be expressed as:
x = x 1 , x 2 , , x n

2.2. Few-Shot Learning

FSL is an application of meta-learning within the domain of supervised learning. Meta-learning, also known as “learning to learn” [25], involves converting the training set into a large number of few-shot tasks during the training phase, with the goal of learning the generalization ability of the model in the presence of class changes. In the testing phase, the model can perform classification tasks using existing learned knowledge, without the need for further changes. Most of the few-shot classification studies are based on optical images and can be broadly classified into the following categories:
(1)
Meta-learning based on data expansion (Refs. [27,28,29]). This approach involves generating additional positive or negative data based on the labeled data provided, which can then be incorporated as auxiliary information to the meta-learner. The aim of this technique is to provide deeper insights and enhance the information fed to the deep neural network.
(2)
Metric-based learning methods, also called comparison learning [30,31,32,33]. This approach centers on the construction of a suitable embedding space that facilitates the generation of corresponding data feature representations. Subsequently, these representations are classified by comparing their similarity with labeled instances and test data features.
(3)
Optimization-based meta-learning approaches, also referred to as learning optimization [34,35], seek to develop a well-initialized optimizer that can rapidly adapt to unseen classes over multiple training iterations.
(4)
External memory-based meta-learning methods [36] replicate the processes of storage, retrieval and knowledge application in the human brain by introducing a memory module. The method stores acquired experiential information in the memory module and retrieves stored knowledge as required to perform classification tasks.
This paper is centered on the construction of an appropriate embedding space.

3. Proposed Method

Acquiring a large number of labeled samples of non-cooperative targets for HRRPs is a challenging task. Additionally, the scattering structures of different classes of ships are similar, and environmental factors significantly affect the HRRP samples of the same class, resulting in a considerable challenge for few-shot HRRP target recognition. Therefore, in this paper, a twofold objective is pursued by the proposed few-shot HRRP target recognition method: firstly, to fully extract the available information from the HRRPs of cooperative ship targets and employ appropriate transfer strategies for non-cooperative target recognition; and secondly, to enhance recognition accuracy by highlighting commonalities among samples of the same class and differences among samples of different classes to mitigate strong confusability during HRRP classification. To achieve this, transfer learning and meta-learning are employed to construct a suitable embedding space Φ · using cooperative target data with a large number of labeled samples.
As depicted in Figure 2, the method in this paper can be delineated into three primary stages: the base-training, meta-training, and meta-testing phases. The first stage, termed base-training, defines the task as HRRP target recognition on a closed dataset. In this stage, the feature extractor f θ ( · ) and classifier C b a s e ( · ) are trained using the complete dataset of cooperative HRRP targets. Subsequently, the weights of the feature extractor with superior classification performance are selected as the initial weights for the feature extractor in the meta-training phase. The second stage is designed to tackle the open-set recognition problem, specifically the challenge of dealing with few-shot HRRP target recognition. In this phase, a substantial number of tasks are sampled from the cooperative HRRP dataset to train and adapt the parameters θ , the embedding space function g ( · ) , and φ ( · ) . Each task includes a support set X k for network training and a query set X q for evaluating network performance. Firstly, the feature extractor f θ ( · ) and the embedding space function g ( · ) transform the samples from the support set X k into feature representations conducive to classification, facilitating the closer proximity of samples from the same class in the metric space. Secondly, the prototype network creates class prototypes by calculating the mean embedding vectors for each class. Finally, φ ( · ) accentuates the inter-class differences, enabling better differentiation of class prototypes in the embedding space. An in-depth explanation of the transfer strategy for the parameter θ during the training phase and the embedding space functions will be provided in Section 3.2. The third stage is the meta-testing phase. Based on the well-trained feature extractor f θ ( · ) , and the embedding space functions g ( · ) and φ ( · ) , a small number of samples is employed to establish prototype representations in the metric space for new classes. The class of a given sample can then be ascertained by calculating the distance between the embedding vector of the sample and the prototype representation. It is noteworthy that our research is based on Resent12, a widely used convolutional neural network.
In the standard formulation of FSL, a task is represented as an N-way K-shot classification problem with N classes sampled from a set of visual concepts and K labeled examples per class. The goal of FSL is to find a function that classifies unlabeled examples into one of the N classes. However, in FSL, K is often small (e.g., K = 1 or K = 5), making it challenging to construct complex functions. To this end, the learning algorithm is also supplied with additional data consisting of ample labeled instances (SEEN classes). The original few-shot task is referred to as the target task, which discerns N UNSEEN classes that do not overlap with SEEN classes. To avoid confusion, the dataset from the SEEN classes used for training are denoted as D t r a i n , comprising a total of N t ship classes. In practical applications, cooperative targets with ample labeled samples are considered as SEEN classes, while non-cooperative targets with limited labeled samples are regarded as UNSEEN classes.

3.1. Base-Training

The aim of this study is to extract as much information as possible from the cooperative HRRP data available for non-cooperative target identification. To achieve this, a feature extractor is trained utilizing all available cooperative target HRRP data. This phase is base-learning, and is intended to model the HRRP data distribution for each class of ships to accomplish the single task of classification referred to as the source task. Base-training facilitates the complete learning of cooperative ship target features based on the entire HRRP training data, leading to a well-initialized network for the meta-training phase. Consequently, meta-learning can converge quickly with fewer tasks. It should be noted that this phase solely relies on the SEEN class dataset D t r a i n and does not consider adaptation from other datasets.
In the initial learning phase, a classifier is trained for the N t class, consisting of a linear classifier C b a s e · , as shown in Figure 3, and a convolutional feature extractor f θ ( · ) , both of which are initialized randomly and updated via gradient descent:
f θ ( · ) , C b a s e · = : f θ ( · ) , C b a s e · γ L D t r a i n f θ ( · ) , C b a s e ·
where L D t r a i n · is the cross-entropy loss function and γ is the learning rate.
The trained feature extractor and classifier effectively extract HRRP features to improve the ship classification task performance, based on the complete dataset D t r a i n . The feature extractor acquires enhanced feature extraction capability by learning the separability features of large-scale data, and its weights are utilized as the initialization weights of the network in the meta-training phase. However, the classifier C b a s e is not used during the meta-learning phase due to modifications made to the classification task.

3.2. Task-Specific Meta-Training

The problem of a limited number of labeled instances and high confusion in HRRP ship classification can be addressed through meta-learning. In the initial stage, the learning of each class of cooperative ship is completed by the feature extractor, and improved feature extraction weights are obtained. However, the features extracted by the current feature extractor may be insufficient or redundant for the new task. Therefore, the meta-training phase involves not only learning how to classify N-way K-shot tasks across a large number of similar tasks, but also accurately and efficiently transferring the trained feature weights to the new task and adapting them to the new task.
There are a series of episodes in the meta-training set. Each episode in this study is randomly sampled from dataset D t r a i n , and comprises two subsets: a support set and a query set. The support set is similar to a traditional training set in machine learning, and includes N classes, each with K training samples. The query set, on the other hand, resembles a test set and comprises Q samples of the same classes as the support set, but without any overlapping samples. Additionally, the meta-learning process is divided into two phases: the meta-training and the meta-testing. For a task T i within the meta-training phase, its dataset can be represented as D t r a i n i = {( X k , Y k ), ( X q , X q )}, where ( X k , Y k ) constitutes the support set, and ( X q , X q ) forms the query set. The number of classes contained in X k and X q is denoted as N. Let X n , k represent the subset of classes in X k for n = 1, 2, 3…N, with corresponding labels Y n , k .
(1) Task-Adaptive Mixed Transfer.
The training objectives diverge between the base-training phase and the meta-training phase. Consequently, freezing all the weights θ of the feature extractor would render the network unable to adapt to new tasks. Nonetheless, whether dealing with SEEN or UNSEEN classes, the structural differences among ships are marginal, and the discriminative feature types necessary for ship HRRP data classification remain consistent. Completely relearning all the initial weights θ could potentially introduce redundancy and squander computational resources. To address these challenges, a Task-Adaptive Mixed Transfer (TAMT) method is proposed, which fine-tunes the parameters in different ways based on the distinct roles of various network blocks in feature extraction. Specifically, the shallow blocks are frozen to preserve the memory of the pre-trained model in extracting ship data envelope features and local physical structure features, while adjusting the deep blocks to adapt to the new few-shot tasks.
The fine-tuning process involves two methods: Neuron-level Scaling and Shifting (SS), and Parameter-level Fine-Tuning (FT). SS involves freezing neurons and then performing scaling and shifting based on their original weights and biases, as shown in Figure 4b. Given the trained f θ ( · ) , for its l-th layer containing M neurons, there are M pairs of parameters, respectively, as weight and bias, denoted as {( W l , m , b l , m )}. Note that the neuron location l, m will be omitted for readability. Based on the methods in this paper, we learn M pairs of scalars. Assuming X k is input, we apply S { 1 , 2 } to (W, b) as
S S ( X k ; W , b ; S { 1 , 2 } ) = ( W S 1 ) X k + ( b + S 2 )
where ⊙ denotes the element-wise multiplication. Taking Figure 4b as an example of a single 1 × 6 filter, after SS operations, this filter is scaled by S 1 then the feature maps after convolutions are shifted by S 2 in addition to the original bias b.
S S ( X k ; W , b ; S { 1 , 2 } ) = ( W S 1 ) X k + ( b + S 2 )
Figure 4b illustrates FT, which involves making small adjustments to all the neural network weight parameters of the model to make it better suited to specific tasks or data. It is evident that FT updates the entire values of W and b, encompassing a large number of parameters, making it susceptible to overfitting when the sample size is limited. Furthermore, performing FT on the entire network may result in the complete forgetting of pre-training memory, known as “catastrophic forgetting”. By contrast, as seen in the examples in the Figure 4, SS reduces the number of training parameters, significantly reducing the computational and memory overhead. However, when SS is applied to all neurons, the adaptation of the network to few-shot tasks may be somewhat lacking.
Assume that the feature extractor f θ ( · ) , initialized with weights, acquires new network weights denoted as θ i } i = 1 N b after the base-training, and considering that the number of network blocks N b is 4 due to our use of ResNet12 for feature extraction. During the meta-training phase, in addition to learning the parameters for the introduced meta embedding, there is a need to fine-tune the existing feature extractor network weights θ i } i = 1 N b to adapt to the changing recognition task. Based on FT and SS, the transfer strategy designed in this paper is depicted in Figure 5, which includes the mapping of network parameters before and after meta-training.
The frozen blocks are referred to as shallow static blocks, while the ones that need adjustment to adapt to the new task are referred to as deep dynamic blocks. Initially, the neurons in shallow blocks, including block 1 and block 2, are frozen to preserve their capability to extract HRRP sample envelope features and initial separability features. Thus, the network parameters θ i } i = 1 N b = 2 remain unchanged before and after training. Subsequently, to adapt to new few-shot tasks and extract relevant features based on task characteristics, block 3 and block 4 are fine-tuned using different fine-tuning methods. Block 3, serving as the deep dynamic block that bridges the two training modes, adopts FT, meaning that the neuron weights of the entire block can be learned. Its main task is to flexibly transform feature vectors from the shallow static module, thoroughly sift, and further extract relevant information to obtain useful features for the new task, initiating the initial adaptation learning for the new task. Block 4 undergoes scaling and shifting based on the initialized weights. Given that its channel count is eight times that of block 1, FT would significantly increase the number of parameters to be learned. In this paper, SS is employed to fine-tune Block 4, a method that substantially reduces the computational and memory overhead while allowing for further adaptation to new classification tasks built on the foundation of Block 3.
This transfer strategy accomplishes several key objectives. Firstly, it preserves the feature extraction capabilities of the feature extractor during the base-training phase for effective HRRP sample extraction, thereby preventing the occurrence of “catastrophic forgetting” that may arise when fine-tuning the entire network. Secondly, it ensures network flexibility and adaptability to new tasks. Finally, by freezing certain neurons and applying SS to others, the number of training parameters in the meta-training phase is reduced, thereby preventing overfitting when dealing with a limited number of labeled ship HRRP samples.
(2) Space-Adjusted Meta Embedding based on GCN and Multi-head Attention.
Achieving the goal of reducing the diversity within the same ships while increasing the differences between different classes is a challenging task when using only a uniform feature extraction network. To address the problem of high confusion about ship HRRP targets in the classification process, this paper introduces a task adaptation step that employs an embedding function focusing on instances and relationships between classes. The embedding function, based on the GCN with a good clustering effect and Multi-head Attention that can emphasize critical classification features, involves all instances entering a new embedding space. This space comprises a sample-based public embedding space and an independent embedding space based on prototypes, as illustrated in Figure 6, designed to perform opposite and complementary functions. The “public” embedding space aims to resolve the problem of significant sample differences within the same type of ships in HRRP, while the “independent” embedding space addresses the issue of slight variations between different ship classes.
Fully connected layers, such as classifier C b a s e · during base-training, are less flexible, since they cannot adapt to changes in the number of classifications. To enhance model flexibility and avoid learning complex recognition models from a limited number of labeled examples of HRRP ships, this study proposes using a prototype network to dynamically construct prototypes and perform the nearest prototype classification. In TAMT, initial adaptation to the new task from a feature extraction perspective is performed by adjusting the weights of the deep blocks. In contrast, the proposed embedding function facilitates further adaptation to the new task from the perspective of class classification and optimal prototype acquisition. The support set samples X k are processed through the feature extractor f θ ( · ) to obtain F k , which represents the features of the input samples. Correspondingly, within F k , the subset for categories where n = 1 , 2 , 3 N is denoted as F n , k , with their respective class labels Y n , k .
The GCN, initially utilized for irregular data structures, particularly graph structures, focuses on the relationships between nodes and exhibits a good clustering effect on the embedded data points. To address the relatively large differences between HRRP samples of the same ship target, GCN serves as a public embedding space g ( · ) based on samples to decrease the distance between samples of the same ships.
Each HRRP instance is considered a data point. For each data point, the GCN can consider both the adjacent data points and the feature information contained in itself. The input to the GCN is denoted as F k , where feature representation vector of each sample is considered as a separate data point. The similarity between each instance is represented by an adjacency matrix A. Specifically, if two ship instances belong to the same class, the corresponding element in A is set to 1, otherwise to 0. F k and A are fed as inputs to the GCN. The adjacency matrix A is symmetrized and normalized to obtain the matrix:
S = D 1 2 A + I D 1 2
where I is the identity matrix and D is the diagonal matrix whose elements are equal to the sum of the elements in the corresponding rows of A + I . The relationship between instances could be propagated based on S:
H t + 1 = R e L U S H t W , t = 1 , 2 , , T 1
where W is a projection matrix for feature transformation. H t is the input feature matrix of each layer, and H 0 is the output of the feature extractor F k . In GCN, the embedding in the set is transformed based on Equation (7) multiple times, which eventually produces the embedding result. The output result of GCN is denoted by G k :
G k = g F k
After the instances are embedded into the public embedding space using GCN, the data points belonging to the same ship are more tightly aggregated, exhibiting less variability. This reduces the probability of misclassifying instances that belong to the same ship but with large differences to different classes.
Prototype c n ( n = 1 , 2 , , N ) is obtained by calculating the average of all vectors belonging to a class of ships in G k :
c n = 1 X n , s x i X n , s f θ ( x i ) , n = 1 , 2 , 3 , , N
Therefore, the prototypes for the task T i are recorded as C i = c n } n = 1 N .
As ships are the only classification objects considered in this paper, their target structures exhibit a significant degree of similarity. Furthermore, the HRRP reflects the projection of the target scatterer in the direction of the LOS, causing the HRRP data returned by ships with similar structures to be highly similar as well. As a result, treating the prototype as a data point means that the distance between prototypes of different ships would be relatively small. To overcome this issue, an independent embedding space φ ( · ) is required for the prototype to accentuate the distinguishable differences between different ship classes.
The Multi-head Attention has the ability to focus on context, extracting important information from an extensive amount of data and highlighting it, while disregarding the rest of the information which is mostly irrelevant. By employing this mechanism in the embedding of the prototype, the Multi-head Attention highlights the vital information from the prototype set. For the classification task, this means enhancing the discernible information between prototypes of different ship classes by assigning weights to them and reducing their commonality. This process increases the distance between different class prototypes and reduces the probability of classifying different classes of ships into the same class. To accomplish this, we first apply the self-attention mechanism to the matrix C i = c n } n = 1 N :
Q = C i · W Q
K = C i · W K
V = C i · W V
O = A t t e n t i o n Q , K , V = s o f t m a x ( Q K T d k ) V
where d k represents the vector length.
Then, the output matrix of each self-attention operation is concatenated along the feature dimension to form a new matrix, which is subsequently fed into a fully connected layer to generate the final output:
C i = M u l t i H e a d ( C i , C i , C i ) = C o n c a t ( O 1 , O 2 , , O m u l t i ) W O
where W is a projection matrix for feature transformation, and C i = c n } n = 1 N is the result of prototype processed by the Multi-head Attention.
For all the embedding points x X q , the Euclidean distance d to a certain class prototype c n can be represented as:
d ( Φ q ( x ) , c n ) = ( Φ q ( x ) c n ) 2
where Φ q · = g f · .
Normalizing the distances from embedded samples to all prototypes using the Softmax can transform the distances into probabilities for the corresponding classes:
p y = n | x = exp d Φ q x , c n n = 1 N exp d Φ q x , c n
So, the loss function for this network on a single training sample is the negative natural logarithm of the true class probability:
L = l o g p ( y = n | x X n , q )
The presented process describes one task in the meta-training procedure. And the overarching objective is to learn the meta-function Φ · , which comprises the functions f · , g · and φ · .

4. Results

In this section, the proposed method was evaluated based on the actual measured HRRP data. To initiate the analysis, the measured data were introduced and pre-processed in preparation for the experiment. Subsequently, the proposed method was compared with the state-of-the-art few-shot learning methods, and the performance of both was analyzed. Moreover, the effects of embedding function and transfer strategy change on the experimental results were investigated. All experiments were conducted using PyTorch on an NVIDIA GeForce RTX 2080 Ti graphics card.

4.1. Data and Preprocessing

Figure 7 displays the distribution of data samples and the normalized representation of certain three types of ships, as utilized in this paper. All statistics were measured by a shore-based radar with a bandwidth of 600 MHz, containing 137,702 measured HRRP samples with 95 different ship targets. Each HRRP is a 4096-dimensional vector, encapsulating one-dimensional structural information of various ships. As observed in Figure 7b–d, the position of the ship relative to the radar at various instances contributes to the incorporation of diverse information about the ship across a range of azimuth and pitch angles within the sample.
The goal of the few-shot meta-learning is to train a meta-learner with the ability to train a learner to converge quickly and accurately for new tasks. Specifically, 60 classes were randomly selected as the training set D t r a i n among the classes with ample labeled samples, to investigate the impact of the number of training samples on recognition accuracy. The remaining classes were divided into 15 and 20 for validation set D v a l and testing set D t e s t , respectively. It is worth noting that the classes of the dataset do not intersect with each other. As shown in Figure 8, according to the meta-learning formulation settings, D t r a i n and D t e s t contain different episodes, which are divided into a support set and a query set. We defined the support set and query set of the meta-training set as D t r a i n s and D t r a i n q ; similarly, we defined the support set and query set of the meta-testing dataset as D t e s t s and D t e s t q . Different classes of HRRP samples are treated as meta-training set and meta-testing set as above described. The classification tasks of different azimuth angle HRRP samples are divided into different tasks in the field of meta-learning formulation, as indicated by the markers in Figure 8.
As a complex and integrated system, the intensity of the HRRP data obtained through high-resolution radar can be influenced by several factors, including radar transmit power, target distance, radar antenna gain, and radar receiver gain. The collected data will have an inconsistent intensity problem. This problem is also called intensity sensitivity. Before using HRRP for target recognition, the original HRRP echo is processed by normalization to improve the intensity sensitivity problem.
If the raw HRRP sample can be expressed as x = [ x 1 , x 2 , , x m ] , where m is the total number of range cells contained in the HRRP, the intensity normalized HRRP sample x n o r m can be expressed as
x n o r m = x x m i n x m a x x m i n
where the intensity is scaled to [0, 1]. Since the convolution, as well as pooling operations, are robust to sample shifts, there is no need for additional processing of translation sensitivity, making the recognition process more complete and reducing human manipulation.

4.2. Evaluation Metrics

This paper evaluates the classification accuracy of five-way one-shot and five-way five-shot, using Q instances from the same categories as test data. The final classification results are obtained by calculating the average accuracy of classification for 10,000 tasks based on the test set, and the calculation is as follows:
a c c = E T E c x X c , q p ( y = = c ) C A R D ( X c , q )
where X c , q is the set of samples in the query set with the class label c, C A R D ( · ) represents the number of elements in a set, p ( · ) represents the probability, y is the predicted class for sample x, c is the true class, and T represents a test task. The formula first calculates the average category recognition rate within a task and then computes the expected value of the average category recognition rate across all tasks.

4.3. Results and Analysis

4.3.1. Compared with State-of-the-Art Methods

The comparison results of the proposed method with several state-of-the-art few-shot methods are presented in Table 1 and Figure 9. The results indicate that although these methods exhibit good performance in the field of optical images, their performance is average in the field of HRRP-based RATA. The possible reason for this discrepancy is that these methods have good performance in recognizing 2D optical images, but are not suitable for 1D HRRP data. Notably, both MTL and FEAT classify ships based on ResNet12. However, the recognition accuracy of MTL is 5.28% and 6.36% lower than the method proposed in this paper, primarily because the linear classifier used in MTL requires learning with a small number of samples, which can limit its effectiveness due to insufficient learning or fewer parameters. In contrast, the proposed method achieves 1.34% and 2.07% higher accuracy than FEAT, because the clustering effect of the GCN reduces the differences between samples within the same class, thereby decreasing the probability of misclassification. Furthermore, the Multi-head Attention embeds prototypes, increasing the distance between prototypes of different classes, thereby enhancing their representativeness, which is another contributing factor to the reduction of misclassifications. The visualization results are shown in Figure 10.

4.3.2. Comparison with Different Transfer Strategies

When a network comprises multiple blocks, it is common to utilize shallow blocks to extract envelope features and local physical structure features of the HRRP, while the deep blocks are leveraged to adapt to a specific task. The base-trained feature extractor is already equipped with good initialization weights and can effectively extract features. However, it is not well-suited to the new task. Therefore, this study proposes a transfer strategy, where the shallow blocks are frozen and the deep blocks are fine-tuned to adapt to the new task. The effectiveness of this transfer strategy is demonstrated through two experimental scenarios. The first scenario involves no common embedding space GCN, while the second scenario employs a common embedding space GCN. The results of these experiments are presented in Table 2, and the contrast chart is presented in Figure 11.
In the absence of GCN embedding, the current study finds that the proposed method is comparable to the results obtained through full parameter fine-tuning. This suggests that the proposed transfer strategy effectively extracts features and adapts to the task, despite freezing some network blocks. Moreover, the proposed method is more computationally efficient and does not compromise accuracy. In contrast, the results indicate that freezing the deep blocks and training only the shallow blocks yield the worst performance, with recognition accuracies of 70.69% and 82.36%, respectively. These accuracies are 3.48% and 2.33% lower than those obtained through the proposed method. This inferior performance can be attributed to the attempt to extract new features, by training only the shallow blocks but keeping the initialized weights unchanged in the deep blocks. This approach not only diminishes the feature extraction capability established during base-training, but also fails to adapt to the new task. Furthermore, the full SS yields recognition accuracies of 73.72% and 83.74%, respectively, which are lower than the proposed method with differences of 0.45% and 1.95%. The results suggest that adapting to new tasks solely through SS is still insufficient.
In the presence of GCN embedding, the gap between all methods decreases when identifying tasks of one-shot, five-way, but differs from the approach proposed in this paper by only +0.29%, −0.1%, −0.24%, and −1.3%. This is because the GCN is responsible for adapting to the new task, masking inadequate task adaptation. However, freezing the deep blocks and training only the shallow blocks still performs the worst in this case, with 1.3% lower than the proposed method in this paper. When the number of labeled samples per class is increased to 5, the proposed methods show better performance, with accuracies of 0.79%, 1.81%, 1.16%, and 2.02% higher than the other methods, respectively. This increased gap between methods is due to the fact that, as the number of labeled samples increases, the remaining methods are unable to take full advantage of the labeled samples to learn the adjustable parameters, although they ensure that the base-training memory is fully utilized. It is worth noting that the proposed method is also 0.79% higher than the full fine-tuning of the parameters. This indicates that full fine-tuning of the parameters may suffer from “forgetting the base-training information” as the number of labeled samples increases.
The number of parameters. Based on the presented table, it is evident that the Full FT, 23Frozen, and 34Frozen transfer methods yield unsatisfactory recognition results when compared to the transfer strategy proposed in this study. This outcome can be attributed to incomplete adaptation to the new task or insufficient feature extraction. Furthermore, only the full FT yields similar recognition results to the transfer strategy in this paper. However, it requires 1.23 M training parameters for complete fine-tuning, which is more than twice the number of training parameters required for the proposed method in this study, which is 495.57 K. Thus, the transfer strategy proposed in this paper achieved improved recognition accuracy by fully extracting target features and adapting to the new task with a smaller number of parameters.

4.3.3. Comparison with Different Embedding Functions

The present experiment investigates the effect of different embedding functions on recognition accuracy in terms of independent embedding space only, public embedding space only, and a combination of different embedding functions. As shown in Table 3 and Figure 12, all three types of embedding methods improve recognition accuracy compared to no embedding at all.
The first part of the experiment focuses on the independent embedding space only. For the recognition of a one-shot, five-way task, the embedding space with Multi-head Attention demonstrates significantly higher recognition accuracy compared to other embedding functions. Specifically, the recognition accuracy of the attention mechanism is 1.4% higher than that of GCN, 1.86% higher than BiLSTM, and 1.5% higher than Deepset. This result can be attributed to the embedding space’s purpose of embedding prototypes of different classes, with the main task being to highlight differences between different classes. While GCN primarily concerns the relationship between data points, it is effective in clustering similar samples. BiLSTM is concerned with sequence correlation, while Deepset lacks the feature of highlighting the weak differences between class prototypes of different classes of ships. The experimental results demonstrate that the recognition accuracy improvement is not significant, even with the addition of other embedding functions on top of the Multi-head Attention. For the five-shot, five-way task recognition, the recognition accuracy of the embedding functions excluding attention decreases rather than increases when compared to no embedding. This is due to the fact that the added embedding functions not only fail to highlight differences between class prototypes but also increase network complexity, making network training more difficult.
The second part of the experiment focuses on the public embedding space only. Compared to Deepset, GCN improves recognition accuracy by 0.27% and 0.23% with fewer parameters. In the third part, different combinations of embedding functions are examined, demonstrating that the combination in this study does not rely on increasing network complexity to improve recognition accuracy. As shown in Table 3, the recognition accuracy of the embedding space combination with GCN as the public embedding function is higher than that of the embedding space combination with Deepset as the public embedding function for the five-shot, five-way task by 1.21%, 1.07%, and 2.51%. This result is due to the excellent clustering function of GCN, which weakens the variability between similar samples, making the samples more clustered, and reducing the chance of class prototypes. The GCN-Attention embeddings have higher recognition accuracy than the combined effect of other functions because GCN and attention play different and irreplaceable roles for HRRP characteristics, respectively. GCN weakens the differences between HRRP samples of the same class by clustering, while attention highlights the weak difference of the prototypes between different classes by calculating the weights.
The Number of parameters. Table 4 shows the parameter count for various embedding functions. The results indicate that the GCN-Attention embedding function has the lowest number of parameters in comparison to other dual-task embedding functions while also achieving superior accuracy. This suggests that the GCN-Attention embedding function has higher parameter efficiency.

4.3.4. Ablation Studies

This section examines the method proposed in this paper and its ablation variants. As presented in Table 5, with full parameters of the feature extractor trained, the recognition accuracy attained during the meta-training phase is 75.52% and 85.67%. Subsequently, after applying the transfer strategy proposed in this study, the accuracy is 75.2% and 86.46%. Notably, there is no reduction in accuracy and a gain of 0.79% when the number of labeled samples is only 5. These results indicate that our transfer strategy ensures no loss of information when adapting to new tasks and prevents “information forgetting” to the full fine-tuning method. Moreover, as presented in Figure 13, compared with the fully FT, the convergence speed of the method in this paper is faster.
Compared to no embedding function, attention as independent embedding space only and GCN as public embedding space only, respectively, enhance recognition accuracy, with Attention improving by 3.34% and 0.08%, and GCN improving by 1.44% and 0.03%. These findings highlight the crucial roles of Attention and GCN in improving accuracy in the classification process.
Moreover, the dual-task embedding function that utilizes Attention and GCN for instance embedding and prototype embedding, respectively, boosts accuracy by 4.4% and 1.85% compared to without embedding. These results demonstrate that Attention and GCN play different roles in the classification process and are critical to the success of the proposed method.

4.3.5. Effect of Network Parameters

The effect of the network parameters on the recognition accuracy was discussed in this subsection. Specifically, the variety of the recognition accuracy of the proposed method in different learning rates was studied on one-shot five-way and five-shot five-way tasks, as well as the effect of the number of heads of the Multi-head Attention of the five-shot, five-way task.
a. Learning rate. The influence of learning rate on recognition accuracy was discussed and the results were shown in Figure 14. As seen from Figure 14, a too large or too small learning rate will make the recognition accuracy unsatisfactory. A large learning rate will make the instability of the recognition accuracy and a small learning rate will lead to slow learning. Table 6 shows the accuracy curves of different learning rates. It can be seen from Table 6 that the performance for classification of one-shot, five-way of the proposed method is best when the learning rate is set to 10 6 . And the learning rate is set to 10 5 for the five-shot, five-way tasks.
b. Number of heads of the Multi-head Attention. Table 7 presents the impact of the number of heads in Multi-head Attention on the recognition accuracy of five-shot five-way tasks. The results indicate that there is no improvement in accuracy as the number of heads increases. On the contrary, increasing the number of heads leads to a rise in the number of training parameters and can result in overfitting when the number of samples is limited. As a consequence, the number of heads is set to 1 in this paper.

4.3.6. Effect of Number of Training Samples

In this experiment, the effect of the number of samples in the training set on the recognition accuracy is investigated. To this end, the training curves for different numbers of HRRP samples are obtained. Specifically, we perform meta-training based on initialized weights obtained from base-training with different sample sizes by keeping the number of samples in the test set constant but varying the number of HRRP samples in the training set.
Figure 15 shows the training curve with the different numbers of training samples. As expected, the accuracy increases with the increase in the number of samples in the training set. Table 8 presents the accuracy of the test set for different training sample sizes. We observe that the recognition accuracy decreases by only 0.74% when the number of samples is reduced from 600 to 200. This result indicates that our method is robust to changes in the sample size. Notably, the proposed method outperforms MatchNet, Protonet, and MTL even when the number of samples is only 100. Moreover, when the number of training samples is 200, our method achieves higher recognition accuracy than FEAT.
In conclusion, our experiments demonstrate the effectiveness and robustness of our proposed method in recognizing HRRP patterns. These findings are valuable for future research in this area.

4.3.7. Generalization Verification

This experiment employed the method of this paper in Conv4, Resnet12 (the network of this paper), and Resnet18, respectively, as shown in Table 9. Due to the simple structure of Conv4, no base-training and transfer operations were performed, and only the embedding function was assessed using this network.
It can be seen from the table that increasing the depth of the network does not necessarily enhance the recognition accuracy, due to the limited number of samples. Moreover, the dual-task embedding function in this paper performs well in both Resnet12 and Resnet18, surpassing the state-of-the-art few-shot method FEAT by 1.96% and 2.64%, respectively. Additionally, compared with Deepset-attention, which has the highest number of parameters, the accuracy achieved by the method in this paper is 2.4% and 0.56% higher for Resnet12 and Resnet18, respectively. In Conv4, Deepset-attention yields a slightly better accuracy of 0.16% than the method proposed in this paper, but with twice as many parameters. Notably, in Resnet18, the transfer methods used during the meta-training stage, namely fully SS, 23frozen, and 34frozen, show significantly lower accuracy than the method proposed in this paper, by 1.83%, 1.73%, and 2.12%, respectively. This could be attributed to inadequate transferring, memory loss, and insufficient adaptation to the task. In conclusion, the method presented in this paper outperforms other methods across different network skeletons.

5. Conclusions

In this study, a novel few-shot HRRP target recognition approach, denoted as Task-Specific Mate-learning, is introduced to address the challenge posed by limited labeled samples and high target variability. Initially, the base learning and meta-learning are employed to mitigate the prevalent overfitting issues within the few-shot HRRP recognition domain. Subsequently, the Task-Adaptive Mixed Transfer technique is proposed, which achieves a classification accuracy no less than that of complete fine-tuning, even when training with only half the number of parameters as compared to full fine-tuning. This highlights the effectiveness of this strategy in preserving valuable information while minimizing information loss during the transfer process, as supported by the recognition accuracy curve. Furthermore, upon visualizing the classification results, it is evident that Space-Adjusted Meta Embedding reduces intra-class sample variations, enhances the separability among distinct ship classes, and alleviates the challenges associated with high target variability during the target recognition process. This adjustment enables instance embeddings to better align with class prototypes in the HRRP target classification task.The actual measured HRRP data are also employed in the proposed method. In comparison to state-of-the-art few-shot learning methods, superior performance in the identification of unknown ship classes with limited labels is achieved by our approach, attributed to its ability to learn from known ship classes. Furthermore, our method demonstrates a degree of generalization and robustness concerning variations in the number of training samples. Specifically, the recognition accuracy of our approach experiences only a marginal decrease of 0.74%, even when the training sample size is reduced from 600 to 200. However, in reality, a more complex scenario exists where radar target recognition systems must identify new class targets with zero samples, a challenge known as zero-shot recognition. Our paper has not addressed this scenario, and as such, we will explore zero-shot learning to enable the zero-shot recognition of new class HRRP targets our future work.

Author Contributions

All the authors made significant contribution to the work. Conceptualization, Y.K. and Y.Z.; methodology, Y.K. and Y.Z.; software, Y.Z.; validation, Y.K. and Y.Z.; formal analysis, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.K.; supervision, H.L.; project administration, X.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Program of Remote Sensing Intelligent Monitoring and Emergency Services for Regional Security Elements; the National Natural Science Foundation of China (No. 62171220); Natural Science Foundation of Jiangsu (No.BK20140825); Aeronautical Science Foundation of China (No. 20182052012); Basic Research (No. NS2021030); and National Science and Technology Major Project (2017-II-0001-0017).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Xu, S.; Chen, Z. Convolutional neural network for classifying space target of the same shape by using RCS time series. IET Radar Sonar Navig. 2018, 12, 1268–1275. [Google Scholar] [CrossRef]
  2. Xiang, X.; Xu, X. Feature extraction for radar target recognition using time sequences of radar cross section measurements. In Proceedings of the 2013 6th International Congress on Image and Signal Processing (CISP), Hangzhou, China, 16–18 December 2013; Volume 3, pp. 1583–1587. [Google Scholar] [CrossRef]
  3. Ding, B.; Wen, G.; Huang, X.; Ma, C.; Yang, X. Data Augmentation by Multilevel Reconstruction Using Attributed Scattering Center for SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2017, 14, 979–983. [Google Scholar] [CrossRef]
  4. Li, C.; Du, L.; Li, Y.; Song, J. A Novel SAR Target Recognition Method Combining Electromagnetic Scattering Information and GCN. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4508705. [Google Scholar] [CrossRef]
  5. Feng, S.; Ji, K.; Zhang, L.; Ma, X.; Kuang, G. ASC-Parts Model Guided Multi-Level Fusion Network for SAR Target Classification. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 5240–5243. [Google Scholar] [CrossRef]
  6. Li, H.; Wang, T.; Wang, S. Few-Shot SAR Target Classification Combining Both Spatial and Frequency Information. In Proceedings of the GLOBECOM 2022—2022 IEEE Global Communications Conference, Rio de Janeiro, Brazil, 4–8 December 2022; pp. 480–485. [Google Scholar] [CrossRef]
  7. Sadjadi, F. Improved target classification using optimum polarimetric SAR signatures. IEEE Trans. Aerosp. Electron. Syst. 2002, 38, 38–49. [Google Scholar] [CrossRef]
  8. Zhou, Y.; Jiang, X.; Li, Z.; Liu, X. SAR Target Classification with Limited Data via Data Driven Active Learning. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2475–2478. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Zhang, L.; Kang, L.; Wang, H.; Luo, Y.; Zhang, Q. Space Target Classification With Corrupted HRRP Sequences Based on Temporal–Spatial Feature Aggregation Network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5100618. [Google Scholar] [CrossRef]
  10. Xia, Z.; Wang, P.; Dong, G.; Liu, H. Radar HRRP Open Set Recognition Based on Extreme Value Distribution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  11. Pan, M.; Liu, A.; Yu, Y.; Wang, P.; Li, J.; Liu, Y.; Lv, S.; Zhu, H. Radar HRRP Target Recognition Model Based on a Stacked CNN–Bi-RNN With Attention Mechanism. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5100814. [Google Scholar] [CrossRef]
  12. Xing, M.; Bao, Z.; Pei, B. Properties of high-resolution range profiles. Opt. Eng. 2002, 41, 493–504. [Google Scholar] [CrossRef]
  13. Ning, F.; Tao, F. Research of Radar Range Profile’s Recognition Based on an Improved C-SVM Algorithm. In Proceedings of the 2012 IEEE 12th International Conference on Computer and Information Technology, Chengdu, China, 27–29 October 2012; pp. 801–804. [Google Scholar] [CrossRef]
  14. Wang, X.; Wu, C. Using improved SVM decision tree to classify HRRP. In Proceedings of the 2005 International Conference on Machine Learning and Cybernetics, Guangzhou, China, 18–21 August 2005; Volume 7, pp. 4432–4436. [Google Scholar]
  15. Chai, J.; Liu, H.; Bao, Z. A W-KNN classifier to improve radar outlier rejection performance. In Proceedings of the 2009 IET International Radar Conference, Guilin, China, 20–22 April 2009; pp. 1–4. [Google Scholar] [CrossRef]
  16. Yu, S.; Xie, Y. Application of a Convolutional Autoencoder to Half Space Radar Hrrp Recognition. In Proceedings of the 2018 International Conference on Wavelet Analysis and Pattern Recognition (ICWAPR), Chengdu, China, 15–18 July 2018; pp. 48–53. [Google Scholar] [CrossRef]
  17. Jithesh, V.; Sagayaraj, M.J.; Srinivasa, K.G. LSTM recurrent neural networks for high resolution range profile based radar target classification. In Proceedings of the 2017 3rd International Conference on Computational Intelligence and Communication Technology (CICT), Ghaziabad, India, 9–10 February 2017; pp. 1–6. [Google Scholar] [CrossRef]
  18. Yifan, Z.; Fengchen, Q.; Fei, X. GS-RNN: A Novel RNN Optimization Method based on Vanishing Gradient Mitigation for HRRP Sequence Estimation and Recognition. In Proceedings of the 2020 IEEE 3rd International Conference on Electronics Technology (ICET), Chengdu, China, 8–12 May 2020; pp. 840–844. [Google Scholar] [CrossRef]
  19. Ding, B.; Chen, P. HRRP feature extraction and recognition method of radar ground target using convolutional neural network. In Proceedings of the 2019 International Conference on Electromagnetics in Advanced Applications (ICEAA), Granada, Spain, 9–13 September 2019; pp. 658–661. [Google Scholar] [CrossRef]
  20. Zhou, Q.; Wang, Y.; Song, Y.; Li, Y. Data augmentation for hrrp based on generative adversarial network. In Proceedings of the IET International Radar Conference (IET IRC 2020), Online, 4–6 November 2020; Volume 2020, pp. 305–308. [Google Scholar] [CrossRef]
  21. Zhang, W.; Lin, Y.; Zhuang, L.; Guo, J. Radar HRRP Data Augmentation Using CVAE with Extended Latent Space Distribution. In Proceedings of the 2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 18–20 June 2021; Volume 4, pp. 1346–1354. [Google Scholar] [CrossRef]
  22. Song, Y.; Zhou, Q.; Yang, W.; Wang, Y.; Hu, C.; Hu, X. Multi-View HRRP Generation With Aspect-Directed Attention GAN. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7643–7656. [Google Scholar] [CrossRef]
  23. Guo, C.; Wang, H.; Xia, X.; Chen, L.; Liu, C.; Zhang, G. Deep Transfer Learning Based Method for Radar Automatic Recognition with Small Data Size. In Proceedings of the 2021 IEEE International Conference on Unmanned Systems (ICUS), Beijing, China, 15–17 October 2021; pp. 995–999. [Google Scholar] [CrossRef]
  24. Yan, Y.; Sun, J.; Yu, J.; Yang, Y.; Jin, L. Few-Shot Radar Target Recognition based on Transferring Meta Knowledge. In Proceedings of the 2022 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xi’an, China, 25–27 October 2022; pp. 1–5. [Google Scholar] [CrossRef]
  25. Sung, F.; Zhang, L.; Xiang, T.; Hospedales, T.; Yang, Y. Learning to learn: Meta-critic networks for sample efficient learning. arXiv 2017, arXiv:1706.09529. [Google Scholar]
  26. Jia, Y.; Chen, B.; Tian, L.; Chen, W.; Liu, H. Memory-Based Neural Network for Radar HRRP Noncooperative Target Recognition. In Proceedings of the 2020 IEEE 11th Sensor Array and Multichannel Signal Processing Workshop (SAM), Hangzhou, China, 8–11 June 2020; pp. 1–5. [Google Scholar] [CrossRef]
  27. Hariharan, B.; Girshick, R. Low-Shot Visual Recognition by Shrinking and Hallucinating Features. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 3037–3046. [Google Scholar] [CrossRef]
  28. Wang, Y.; Girshick, R.; Hebert, M.; Hariharan, B. Low-Shot Learning from Imaginary Data. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7278–7286. [Google Scholar] [CrossRef]
  29. Schwartz, E.; Karlinsky, L.; Shtok, J.; Harary, S.; Marder, M.; Kumar, A.; Feris, R.; Giryes, R.; Bronstein, A. Delta-encoder: An effective sample synthesis method for few-shot object recognition. arXiv 2018, arXiv:1806.04734. [Google Scholar]
  30. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. arXiv 2017, arXiv:1703.05175. [Google Scholar]
  31. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to Compare: Relation Network for Few-Shot Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar] [CrossRef]
  32. Mensink, T.; Verbeek, J.; Perronnin, F.; Csurka, G. Distance-Based Image Classification: Generalizing to New Classes at Near-Zero Cost. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2624–2637. [Google Scholar] [CrossRef] [PubMed]
  33. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. arXiv 2016, arXiv:1606.04080. [Google Scholar]
  34. Ravi, S.; Larochelle, H. Optimization as a model for few-shot learning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  35. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
  36. Ramalho, T.; Garnelo, M. Adaptive posterior learning: Few-shot learning with a surprise-based memory module. arXiv 2019, arXiv:1902.02527. [Google Scholar]
  37. Sun, Q.; Liu, Y.; Chua, T.S.; Schiele, B. Meta-transfer learning for few-shot learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 403–412. [Google Scholar]
  38. Ye, H.; Hu, H.; Zhan, D.; Sha, F. Few-shot learning via embedding adaptation with set-to-set functions. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 8808–8817. [Google Scholar]
Figure 1. Schematic of HRRP of radar target.
Figure 1. Schematic of HRRP of radar target.
Remotesensing 15 05301 g001
Figure 2. Overall structure of Task-Specific Mate-learning (TSML) for few-shot HRRP.
Figure 2. Overall structure of Task-Specific Mate-learning (TSML) for few-shot HRRP.
Remotesensing 15 05301 g002
Figure 3. The linear classifier of base-training.
Figure 3. The linear classifier of base-training.
Remotesensing 15 05301 g003
Figure 4. Parameter-level Fine-Tuning (FT) and Scaling-Shifting (SS).
Figure 4. Parameter-level Fine-Tuning (FT) and Scaling-Shifting (SS).
Remotesensing 15 05301 g004
Figure 5. Task-Adaptive Mixed Transfer (TAMT).
Figure 5. Task-Adaptive Mixed Transfer (TAMT).
Remotesensing 15 05301 g005
Figure 6. The structure of the Space-Adjusted Meta Embedding (SAME).
Figure 6. The structure of the Space-Adjusted Meta Embedding (SAME).
Remotesensing 15 05301 g006
Figure 7. Dataset presentation. (a) Sample size for each class of ship. (bd) All samples of three classes of ships.
Figure 7. Dataset presentation. (a) Sample size for each class of ship. (bd) All samples of three classes of ships.
Remotesensing 15 05301 g007
Figure 8. The dataset setting of the proposed method.
Figure 8. The dataset setting of the proposed method.
Remotesensing 15 05301 g008
Figure 9. The recognition accuracy curves of different methods in different few-shot conditions.
Figure 9. The recognition accuracy curves of different methods in different few-shot conditions.
Remotesensing 15 05301 g009
Figure 10. Visual results of the TSML.
Figure 10. Visual results of the TSML.
Remotesensing 15 05301 g010
Figure 11. The contrast chart of different transfer strategies in different few-shot conditions.
Figure 11. The contrast chart of different transfer strategies in different few-shot conditions.
Remotesensing 15 05301 g011
Figure 12. The contrast chart of different embedding functions in different few-shot conditions.
Figure 12. The contrast chart of different embedding functions in different few-shot conditions.
Remotesensing 15 05301 g012
Figure 13. The recognition accuracy curves of ablation variants in different few-shot conditions.
Figure 13. The recognition accuracy curves of ablation variants in different few-shot conditions.
Remotesensing 15 05301 g013
Figure 14. The recognition accuracy curves of different learning rates in different few-shot conditions.
Figure 14. The recognition accuracy curves of different learning rates in different few-shot conditions.
Remotesensing 15 05301 g014
Figure 15. The recognition accuracy curves of different number of training samples in five-shot conditions.
Figure 15. The recognition accuracy curves of different number of training samples in five-shot conditions.
Remotesensing 15 05301 g015
Table 1. Recognition accuracy of different methods.
Table 1. Recognition accuracy of different methods.
MethodAccuracy%
One-Shot Five-WayFive-Shot Five-Way
Matchnet [33]64.5475.41
Protonet [30]65.5176.04
MTL [37]70.5280.10
FEAT [38]73.8984.39
Ours75.2386.46
The bolded parts are the methods with the best recognition results.
Table 2. Recognition accuracy of different transfer strategies.
Table 2. Recognition accuracy of different transfer strategies.
Accuracy %
Transfer StrategyEmbedding without GCNEmbedding with GCN
One-Shot Five-WayFive-Shot Five-WayOne-Shot Five-WayFive-Shot Five-Way
Fully FT73.8984.3975.5285.67
Fully SS73.7282.7475.1384.65
23Frozen73.2083.9974.9985.30
34Frozen70.6982.3673.9384.43
Ours74.1784.6975.2386.46
The bolded parts are the methods with the best recognition results.
Table 3. Recognition accuracy of different embedding functions.
Table 3. Recognition accuracy of different embedding functions.
Embedding FunctionAccuracy%
One-Shot Five-WayFive-Shot Five-Way
NONE-NONE70.8384.61
Attention74.1784.69
IndependentBiLstm72.3183.10
Attention+BiLstm72.9984.40
GCN72.2784.64
PublicDeepset72.0084.41
GCN-BiLstm72.3085.16
Deepset-Attention73.4283.95
Dual-taskGCN-Deepset71.9885.02
GCN-Attention75.2386.46
The bolded parts are the methods with the best recognition results.
Table 4. Number of parameters.
Table 4. Number of parameters.
Embedding FunctionGCN-BiLstmDeepset-AttentionGCN-DeepsetGCN-Attention
params (M)3.67 M6.30 M7.35 M3.15 M
The bolded parts are the methods with the best recognition results.
Table 5. Recognition accuracy of ablation variants.
Table 5. Recognition accuracy of ablation variants.
Accuracy%
Fully FTTATSGCNAttentionOne-Shot Five-WayFive-Shot Five-Way
1 73.8984.39
2 75.5285.67
3 70.8384.61
4 72.2784.64
5 74.1784.69
6 75.2386.46
The bolded parts are the methods with the best recognition results.
Table 6. Recognition accuracy of different learning rates.
Table 6. Recognition accuracy of different learning rates.
Learning RateAccuracy%
One-Shot Five-WayFive-Shot Five-Way
10 3 53.9056.17
10 4 75.0285.51
10 5 74.8986.46
10 6 75.2385.86
10 7 71.0480.22
The bolded parts are the methods with the best recognition results.
Table 7. Recognition accuracy of different numbers of heads.
Table 7. Recognition accuracy of different numbers of heads.
Number of HeadAccuracy %
Five-Shot Five-Way
1 86.46
285.99
386.61
486.10
The bolded parts are the methods with the best recognition results.
Table 8. Recognition accuracy of different numbers of training samples.
Table 8. Recognition accuracy of different numbers of training samples.
Number of SamplesAccuracy %
Five-Shot Five-Way
100 83.36
200 85.72
400 85.78
500 86.35
60086.46
The bolded parts are the methods with the best recognition results.
Table 9. Recognition accuracy of different backbones in five-shot conditions.
Table 9. Recognition accuracy of different backbones in five-shot conditions.
Accuracy %
Conv4Resnet12Resnet18
Ours77.9886.4683.60
FEAT75.1284.3980.96
Change the Embedding Function
GCN-Bilstm77.6785.1682.61
Deepset-attention78.1483.9583.04
Change Transfer Strategy
Fully FT-85.6783.53
Fully SS-84.6581.77
23Frozen-85.3081.87
34Frozen-84.4381.48
The bolded parts are the methods with the best recognition results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kong, Y.; Zhang, Y.; Peng, X.; Leung, H. Few-Shot High-Resolution Range Profile Ship Target Recognition Based on Task-Specific Meta-Learning with Mixed Training and Meta Embedding. Remote Sens. 2023, 15, 5301. https://doi.org/10.3390/rs15225301

AMA Style

Kong Y, Zhang Y, Peng X, Leung H. Few-Shot High-Resolution Range Profile Ship Target Recognition Based on Task-Specific Meta-Learning with Mixed Training and Meta Embedding. Remote Sensing. 2023; 15(22):5301. https://doi.org/10.3390/rs15225301

Chicago/Turabian Style

Kong, Yingying, Yuxuan Zhang, Xiangyang Peng, and Henry Leung. 2023. "Few-Shot High-Resolution Range Profile Ship Target Recognition Based on Task-Specific Meta-Learning with Mixed Training and Meta Embedding" Remote Sensing 15, no. 22: 5301. https://doi.org/10.3390/rs15225301

APA Style

Kong, Y., Zhang, Y., Peng, X., & Leung, H. (2023). Few-Shot High-Resolution Range Profile Ship Target Recognition Based on Task-Specific Meta-Learning with Mixed Training and Meta Embedding. Remote Sensing, 15(22), 5301. https://doi.org/10.3390/rs15225301

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop