Next Article in Journal
Efficient Deep Semantic Segmentation for Land Cover Classification Using Sentinel Imagery
Next Article in Special Issue
Hybrid Domain Efficient Modulation-Based Deceptive Jamming Algorithm for Nonlinear-Trajectory Synthetic Aperture Radar
Previous Article in Journal
Efficiency Study of Combined UAS Photogrammetry and Terrestrial LiDAR in 3D Modeling for Maintenance and Management of Fill Dams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PolSAR Image Classification Based on Relation Network with SWANet

Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, School of Computer Science and Technology, Xi’an University of Posts and Telecommunications, Xi’an 710061, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(8), 2025; https://doi.org/10.3390/rs15082025
Submission received: 16 March 2023 / Revised: 5 April 2023 / Accepted: 10 April 2023 / Published: 11 April 2023
(This article belongs to the Special Issue Spaceborne High-Resolution SAR Imaging)

Abstract

:
Deep learning and convolutional neural networks (CNN) have been widely applied in polarimetric synthetic aperture radar (PolSAR) image classification, and satisfactory results have been obtained. However, there is one crucial issue that still has not been solved. These methods require abundant labeled samples and obtaining the labeled samples of PolSAR images is usually time-consuming and labor-intensive. To obtain better classification results with fewer labeled samples, a new attention-based 3D residual relation network (3D-ARRN) is proposed for PolSAR image. Firstly, a multilayer CNN with residual structure is used to extract depth polarimetric features. Secondly, to extract more important feature information and improve the classification results, a spatial weighted attention network (SWANet) is introduced to concentrate the feature information, which is more favorable for a classification task. Then, the features of training and test samples are integrated and CNN is utilized to compute the score of similarity between training and test samples. Finally, the similarity score is used to determine the category of test samples. Studies on four different PolSAR datasets illustrate that the proposed 3D-ARRN model can achieve higher classification results than other comparison methods with few labeled data.

Graphical Abstract

1. Introduction

Polarimetric synthetic aperture radar (PolSAR) is an active microwave imaging technology. Compared to traditional SAR technology, PolSAR adopts the multi-channel and multi-polarization working mode, which can obtain rich target information through the transmission and reception of polarimetric electromagnetic waves. Because of these advantages, PolSAR image has made outstanding achievements in remote sensing applications in a variety of fields [1,2,3]. PolSAR image terrain classification is a very important fundamental project in these applications, which aims to classify the pixels of the whole map into the corresponding categories through the polarimetric information. With the continuous evolution of machine learning technology, various PolSAR classification methods have been presented [4,5,6]. Based on the availability of labeled samples, these methods can be classified as unsupervised, semi-supervised, and supervised methods.
In general, unsupervised classification methods of PolSAR images do not require labeled samples. They mainly study the polarimetric scattering mechanism [7,8,9,10,11,12] and statistical distributions of polarimetric features [13], such as Pauli decomposition, Freeman three-component decomposition, Cloude decomposition, Huynen decomposition, and Krogager decomposition. In addition, researchers also studied other polarimetric features, such as the roll-invariant features [14]. Based on these polarimetric features, many unsupervised classification methods were proposed to improve the classification accuracy. Although the unsupervised method is simple and fast, the classification accuracy is not good due to the lack of samples with labels. Unlike the unsupervised classification, the supervised method obtains better classification performance by utilizing labeled samples. For PolSAR images, some representative supervised methods have been proposed, such as k-nearest neighbor (KNN) [15], support vector machine (SVM) [16], and neural networks [17].
Recently years, deep learning, as a main neural network method, has achieved remarkable results on many problems, such as object detection [18], image classification [19], and target recognition [20]. With the continuous progress of deep learning, deep learning-based classification algorithms have been proposed for PolSAR images, such as the deep belief network (DBN) [21], the stacked autoencoder (SAE) [22], and convolutional neural networks (CNN) [23]. However, the classification accuracy of PolSAR images is significantly enhanced by spatial information [22]. CNN can well obtain the spatial feature of an image with its unique network structure. Shang et al. [24] proposed a spatial feature-based convolutional neural network (SF-CNN) to solve PolSAR classification problem in a limited labeled dataset. Cui et al. [25] optimized topologies of basic CNNs by adding multipath. Therefore, these CNN-based PolSAR classification methods are proposed to achieve higher classification precision. However, deep learning methods require sufficient labeled samples and collecting labeled samples usually requires a lot of laborers and materials for PolSAR images.
In contrast, it is easier to obtain unlabeled samples than labeled samples, which can provide more statistical information to compensate for the limited labeled samples. This fact promoted the concept of semi-supervised learning, which enables training with both labeled and unlabeled samples to obtain potential performance improvement. Therefore, some semi-supervised methods [26,27] have been put forth to figure out the issue of PolSAR image classification with limited data, such as the co-training method, tri-training method, self-training method, and graph-based method. Recently, in order to use the capabilities of deep learning to tackle the categorization challenge of PolSAR images with limited data, researchers proposed several semi-supervised methods based on deep learning. Fang et al. [28] proposed a semi-supervised 3D-CNN model using pseudo labels. Guo et al. [29] combined a memory mechanism and a semi-supervised learning method to construct a semi-supervised method based on a memory convolutional network for PolSAR classification. Although these methods have made great advances, it is still challenging to classify PolSAR images with deep learning methods especially for the limited labeled samples. Accordingly, this paper mainly studies the classification of PolSAR images with only a few labeled samples.
The main purpose of few-shot learning is to identify new categories from a small number of labeled data [30]. Although, the availability of few labeled samples hampers the typical deep learning fine-tuning process, some methods have made considerable advances in few-shot learning, such as matching networks [31], prototypical networks [32], and meta-learning [33]. The matching network identifies the unlabeled data (query set) using attention embedding matching measurements between labeled data (support set). The prototype network learns metric space and realizes classification by calculating the distance to the prototypical representation of each class. Meta-learning attempts to acquire a collection of projection functions so that when an image is represented in such an embedding, the image can be easily identified using a simple nearest neighbor or linear classifier.
In recent years, few-shot learning has attracted numerous interests in the subject of remote sensing. Liu et al. [34] suggested a residual network-based few-shot learning method for hyperspectral image classification that learns a metric space based on episode training samples of the source domain, then combines it with a basic classifier to improve the classification results. Gao et al. [35] developed a related network for hyperspectral image classification with few labeled samples. Tong et al. [36] introduced a few-shot hyperspectral image classification method based on attention-weighted graph convolutional networks, which use the GNN to few-shot hyperspectral image classification. Zuo et al. [37] suggested a hyperspectral image classification algorithm based on edge-labeling graph neural networks, which use a graph neural network to explicitly quantify the relationships between pixels. Zhang et al. [38] proposed a few-shot unsupervised deep representation learning based on contrastive learning for PolSAR classification, which learns transferrable representations from unlabeled data.
Although few-shot learning has achieved remarkable achievements in other remote sensing image classifications, the application in PolSAR images is still less common. Relation networks (RN) [30] is one of the few-shot learning methods based on meta-learning-based models, which is simple, flexible, efficient, and has obtained some excellent performances in image recognition [35,39]. However, the RN was initially designed for a natural image rather than a PolSAR image without considering the difference between them. For a natural image, an image is a sample, while a pixel is a sample in a PolSAR image. For a PolSAR image, each pixel has its spatial neighborhood, which is very important in deciding the category of the pixel, but natural images have no such neighborhood. Moreover, compared with natural images, PolSAR images have a more complex intraclass similarity and interclass difference, which requires more sufficient and valid features.
Based on the above analysis, this paper proposes a new few-shot learning classification method for PolSAR images based on RN. However, it should be noted that an RN requires a sufficient number of labeled samples from the source domain for training. For PolSAR images, it is very difficult to gather a sufficient number of labeled data with the same band and imaging system for training relational networks. Therefore, we proposed a training set construction method using few labeled samples to train the RN model based on the superpixels algorithm [40].
In addition, to fully extract spatial information and channel information for a PolSAR image, 3D-CNN [41,42] is used for network training and classification in the presented method. Moreover, the depth features extracted by the deep network inevitably have redundant information, and the traditional feature selection approaches are unaffected by classification. Accordingly, a spatial weighted attention network (SWANet) is proposed to automatically obtain the importance of spatial features through learning, avoiding the complex feature selection process. Then, according to the value of features, the useful features are promoted and the less useful ones are suppressed to improve classification accuracy.
Therefore, the important contributions of this study are stated below:
  • A new deep network 3D-ARRN is proposed for few-shot PolSAR image classification. This method can automatically select and extract features to achieve end-to-end final classification and effectively improve the classification results with fewer labeled samples.
  • According to the properties of a PolSAR image, a spatial weighted attention network (SWANet) is proposed to select important spatial features to improve the network performance.
  • We proposed a superpixels-based pseudo-labeled sample generation method, and used the pseudo-labeled sample to learn the transferrable representations. Then, the collected representations are transferred with limited labeled data to perform the few-shot PolSAR classification.
The remainder of this article is structured as follows. The specifics of the proposed method are presented in Section 2. Section 3 depicts the experimental design using four PolSAR data. The final experimental results are presented in Section 4. Section 5 concludes the paper.

2. The Proposed Method

2.1. Polarimetric Representation

Each resolution cell in PolSAR data is represented as a complicated coherency matrix T, which reflects the transformation relation between incident wave and scattered wave of target information. Its form is defined as:
T = [ T 11 T 12 T 13 T 21 T 22 T 23 T 31 T 32 T 33 ]
The coherency matrix T is a complex conjugate symmetric matrix. Its primary diagonal elements are real numbers and the off-diagonal elements are complex numbers that satisfy conjugation. In general, the coherency matrix T can be used to generate a true 6-dimensional feature vector, which is defined as:
A T = 10 log 10 ( T 11 + T 22 + T 33 ) B T = T 22 / ( T 11 + T 22 + T 33 ) C T = T 33 / ( T 11 + T 22 + T 33 ) D = | T 12 | / T 11 T 22 E = | T 13 | / T 11 T 33 F = | T 23 | / T 33 T 22
where A T indicates the total scattering power. B T and C T represent the normalized power ratios of T22 and T33, respectively. D T , E T , and F T denote the relative correlation coefficients.
Apart from the coherency matrix, other scattering features needed to be involved. In our study, scattering features obtained from target decomposition methods and roll-invariant polarimetric features are further used, as list in Table 1, because different features can reflect specifics of the scattering mechanism details from various perspectives. Therefore, we combine together all these polarimetric features to construct a 32-dimensional feature in this study.

2.2. Feature Selection Based on SWANet

In this paper, the 32-dimensional feature vector are extracted by coherence matrix T, which is used as the feature information of each pixel. After a multi-layer convolutional network, multi-dimensional deep features are extracted from 32-dimensional feature information. However, CNN-based classification methods of PolSAR images usually select pixel blocks to represent the features of central pixel points. Although these methods consider the spatial neighborhood information between pixels, the direct use of pixel blocks may affect the judgment of the center pixel category because the neighborhood pixels and the center pixels in the irregular area may belong to different categories. Therefore, to reduce the impact of different categories of pixels in the same neighborhood on the determination of the category of central pixels, this paper proposes a spatial weighted attention network (SWANet) to calculate the contribution of the neighborhood pixels in the pixel block to the center pixel. Then, the original input features are weighted by the weight after learning and the importance of the original features in the spatial range are re-calibrated. The specific process is shown in Figure 1.
As shown in Figure 1, the input feature information can be represented as a tensor h × w × c , where h represents the height, w denotes the width, and c represents the dimension of the input feature. First, compute the spatial similarity between the central pixel and other pixels in the pixel block:
d i = i = 1 c ( x i x j ) 2
where x i represents pixel feature information in the pixel block and x j represents the feature information of the center pixel in the pixel block.
The larger the value of d i , the greater the difference between the neighborhood pixels x i and center pixel x j inside the pixel block. In this case, the feature contribution of x i to the center pixel x j should be smaller when judging the category of the center pixel. Therefore, a kernel function is introduced as follows:
ε i = exp ( d i )
Then, the weights of all the points in the pixel block are normalized as follows:
δ i = ε i i = 1 h × w ε i
Finally, the scaling operation multiplies the input information Λ by weight parameter W and outputs a new feature map U , U R h × w × c ,
U = F s c a l e ( Λ , W ) = Λ W
where W represents the matrix of h row and w column composed of weights δ i .

2.3. Relation Network

The relation network is a classical method for few-shot learning, which is a simpler and efficient classification method and has made outstanding achievements in remote sensing applications in a variety of fields. The design idea of an RN is to constrain the functional structure of a neural network to capture the common characteristics of relational reasoning. That is to say, the ability to calculate relationships is implanted into the RN model without learning just as the ability to reason about space and translation invariant attributes is built-in in CNN.
In RN methods, there are two datasets: target domain data and source domain data. The target domain data is divided into the support set D t s u p p o r t and query set D t q u e r y , and the source domain data is also called the training set D s t r a i n . In the M-way K-shot classification task, the support set D t s u p p o r t contains I = M × K samples of M classes with K samples for each class. An RN utilizes training sets through episode-based learning. An episode consists of a randomly selected query sample D s q u e r y and several support samples D s s u p p o r t in source domain. After training, the trained parameters are stored in the RN model by executing the source domain. Then, this trained model is used to perform a new classification task. The labeled target data (support set D t s u p p o r t ) and trained model are intended to predict the labels of all unlabeled target class data (query set D t q u e r y ).
As depicted in Figure 2, the RN model contains two parts: the embedding module and the relation module. For each episode, randomly select class M from the source domain and randomly select K samples from each class M as the sample set to form a support set D s s u p p o r t = { ( x i , y i ) } i = 1 m   ( m = M × K ) and select a part of these class M samples as the query set D s q u e r y = { ( x j , y j ) } j = 1 n . For each sample pair x i and x j , they are first fed into the embedding module f ϕ to obtain deep features f ϕ ( x i ) and f ϕ ( x j ) , respectively. Then, these feature maps ( f ϕ ( x i ) and f ϕ ( x j ) ) are combined together with operator C ( f φ ( x i ) , f φ ( x j ) ) , where C is a concatenation of feature maps in depth. Finally, the combined feature maps are input into the relation module so that the relation score r i , j between the query sample x j and support sample x i can be generated to indicate their similarity. The value of relation score r i , j is 0 to 1, and the higher value indicates a greater similarity. It is defined as:
r i , j = g φ ( C ( f φ ( x i ) , f φ ( x j ) ) ) i = 1 , 2 , , m
Considering that similarity prediction is a regression problem in the RN model. In the training process, the RN model uses mean square error loss for network training, and the object function of an RN is defined as:
φ , ϕ arg min φ , ϕ j = 1 Q i = 1 M ( r i , j 1 ( y i = y j ) ) 2
where Q represents the quantity of the query sample, y i denotes the label for the support sample, and y j represents the label for the query sample.

2.4. Pseudo-Labels Generation Algorithm

In few-sample problems, the classifier trained only with support set samples has poor classification performance due to a scarcity of labeled samples. According to this problem, the RN carries out meta-learning algorithm on the source domain in order to derive transferable information so that we can carry out better feed efficiency from shot learning of the support set and more successfully classify the test set. Therefore, source domain data are essential to an RN. However, for PolSAR data, it is difficult to collect a relevant labeled sample with the same band and imaging system. As a result, a superpixel-based training set construction method is proposed that uses both limited labeled samples and significant unlabeled samples to generate a source domain data with a large number of pseudo-labels. Figure 3 shows the flowchart of this process, and the following are the specific steps:
  • Divide the Pauli image of PolSAR into Ns superpixels by turbopixels algorithm [42], and calculate the clustering center of each superpixel.
  • Select the superpixel containing the support sample and give it the same label as the support sample (Figure 3b). Then, calculate the clustering center of these labeled superpixels.
  • Among all superpixels, select the superpixel most similar to each class of labeled superpixel by Equation (9) and mark it.
d i , j ( V i , V j ) = ( N i + N j ) ln | V ~ | N i ln | V i ~ | N j ln | V j ~ |
V i ~ = 1 / N i t = 1 N i T t
V j ~ = 1 / N j t = 1 N j T t
V ~ = 1 / ( N i + N j ) t = 1 N i + N j T t
where N represents the number of pixels in the superpixels. T t represents the polarimetric coherence matrix. d i , j denotes the similarity between i-th and j-th superpixels; the smaller value of d i , j , the i-th and j-th superpixels are more similar.
4.
Repeat step 3 until sufficient labeled superpixels are obtained (Figure 3c).
5.
Calculate the distance from all pixels in each labeled superpixel to the center of the superpixel by Equation (13), and calculate the average value of these distances by Equation (14).
w t ( T t , V s ) = ln | V i | + T r ( V s 1 T t )
w c = 1 N s t = 1 N s w t
where V s = 1 / N s t = 1 N s T t denotes the center of superpixel.
6.
If the distance from any pixel in the labeled pixel to the superpixel center is less than the average distance w c , the label of the pixel is reserved; otherwise, the label of the pixel is removed (Figure 3d).
7.
Output these pseudo-labels sample as the training data.

2.5. Architecture of the 3D-ARRN

In this section, a novel few-shot PolSAR image classification method is proposed based on 3D-ARRN. As depicted in Figure 4, the structure of the proposed method mainly consists of four modules: the data input module, the embedding module, the relation module, and the final image classification. For the M-way K-shot classification problem of the PolSAR image, the proposed method first extracts polarimetric features from the coherency matrix T to represent the support samples and query samples. Secondly, construct a 3D embedding module based on the residual network and SWANet for feature extraction and feature selection. Then, use the 3D residual relation module to calculate the similarity of the query sample and support sample, and then calculate the objective function according to this similarity. Next, the objective function is minimized to train the proposed model in each episode. Finally, the category of the test sample is determined by relational score. The structure of the suggested method for training is summarized in Algorithm 1.
Algorithm 1. Framework of the 3D-ARRN for a training episode
Input: Support set D s s u p p o r t = { ( x i , y i ) } i = 1 m ( m = M × K ) , query set D s q u e r y = { ( x j , y j ) } j = 1 n .
Output: The objective function of the RN in an episode.
1: Obtain the vector representation of each support samples and query samples.
2: The residual network and SWANet are used for feature extraction and selection of the support set and query set, and extracted features are expressed as f ϕ ( x i ) and f ϕ ( x j ) , respectively.
3: Connect the feature maps of support set and query set by operator C ( f φ ( x i ) , f φ ( x j ) ) .
4: The combined features are fed into the relational network g ϕ to calculate the relational score r i , j = g φ ( C ( f φ ( x i ) , f φ ( x j ) ) ) between the query sample and the support samples and calculate the objective function in Equation (8).
5: Return the objective function to be minimized to train the model.

3. Experimental Design

In this section, four real PolSAR datasets are used to verify our proposed method in this paper, and all experiments were conducted on a computer with 2080Ti GPUs. The parameters are set as follows: adaptive moment estimation (Adam) is used for optimization, the learning rate has been set at 0.001, and the filter method adopts the Lee filter [43]. The specific parameters of the proposed classification network are shown in Table 2. In addition, the overall classification accuracy (OA) is taken to verify the effectiveness of our proposed method.

3.1. Data Sets

The first PolSAR dataset is L-band PolSAR data of Flevoland I, the Netherlands, acquired by the AIRSAR system in 1989, which has 750 × 1024 pixels. The resolution of this image is 6.6 m × 12.1 m . It contains fifteen different types of crops and each type of crop is identified by one color. Figure 5a is the Pauli RGB map of this image and Figure 5b is the corresponding ground-truth map of this image.
The second PolSAR dataset is C-band PolSAR data of Flevoland II, the Netherlands, acquired by RADARSAT-2 in 2008, which has 1400 × 1200 pixels. The resolution of this image is 10 m × 5 m . It contains four different terrain types. The Pauli RGB map of the image is shown in Figure 6a and the corresponding ground-truth map of this image is depicted in Figure 6b.
The third PolSAR dataset is C-band PolSAR data acquired by RADARSAT-2 in San Francisco in April 2008 with 1300 × 1300 pixels and a resolution of 10 m × 5 m . It contains 5 terrain objects: high-density, low-density, water, developed, and vegetation, and Figure 7a denotes the Pauli RGB map of this image. Figure 7b shows the appropriate ground-truth map of this image, and each type of crop is identified by one color.
The fourth dataset is a C-band PolSAR data of China, Shaanxi, Xi’an, acquired by RADARSAT-2 in 2010, which has 512 × 512 pixels. The resolution of this image is 8 m × 8 m . It contains three different categories: bench land, urban, and river. The Pauli RGB map of the image is shown in Figure 8a, and the corresponding ground-truth map of this image is shown in Figure 8b.

3.2. Experimental Design Discussion of Key Parameters

In this subsection, the Flevoland I dataset is utilized as an example to examine the three key parameters of our proposed method: the window size of samples between their neighborhood pixels, the number of support samples, and the batch size of training process.

3.2.1. Influence of Sample Window Size on Proposed Method

In our proposed method, we take the whole neighborhood pixel of a center pixel as the input data. Therefore, the window size w × w of the neighborhood pixels is an important parameter which affects the classification accuracy of this approach. In order to only analyze the effect of window size, we fix the other parameters. The number of support samples for each category is set to 10 and the batch size is set to 75. In this part, the range of w × w is set from 5 × 5 to 17 × 17.
Figure 9 shows the OA of the Flevoland I dataset for the proposed method with different window sizes. From Figure 9, it has been observed that OA gradually grows with increasing window size in the beginning. This is primarily due to the fact that when the window is relatively small, as the window size w × w increases, it contains more neighborhood pixels, provides more spatial information, and is more helpful for the classification of the center pixels. However, when the window reaches a certain range, some pixels in the neighborhood are farther away from the center pixel. The addition of this neighborhood pixel information will affect the decision of the center pixel category and thus will affect the classification accuracy. For the aforementioned reason, the window size is set to 15 × 15 throughout the following experiments.

3.2.2. Influence of Labeled Samples on Proposed Method

In this section, we only consider the impact of the number of labeled samples of each class on the experimental method. The number of labeled samples is important for the relation network, whose value K will affect the classification results of our proposed. In order to analyze the effect of different numbers of labeled samples and to select the optimal value for our proposed method, we test the classification performance of K support samples of each class from 1 to 10.
Figure 10 represents the classification accuracy of the proposed method with varying numbers of labeled samples on the Flevoland I dataset. From Figure 10, we can observe that the highest OA of the proposed method is 96.24% when K = 7 and is larger than other values. In addition, it can also find that when K is larger than 1, OA is larger than K = 1. This shows that the K-shot classification results are better than the one-shot classification results for the proposed method. Moreover, in Figure 10, it can be found that the extracted features will be diluted or averaged, and the recognition ability will be reduced when the value of K is too large. Therefore, in the following experiment, the value of K is set to 7.

3.2.3. Influence of Batch Size on Proposed Method

In this section, we only consider the effect of the batch size on classification accuracy. Batch size is a parameter that has a great influence on network training. In theory, if the batch size is small, the difference between adjacent mini-batches is large, causing serious loss curve oscillation and slowing down network convergence. If the batch size is too large, the difference between adjacent mini-batches is small, the loss curve is smooth, and the convergence speed is relatively fast. However, due to the smooth gradient descent, the training only moves in one direction and is prone to falling into the local minimum value. Therefore, in order to choose the appropriate batch size, we use batch sizes ranging from 15 to 150 to train the network and record the classification results of the proposed method with different batch sizes.
Figure 11 depicts the OA of our proposed method for different batch sizes on the Flevoland I dataset. From Figure 11, it can be found that when the batch size is 15, the classification accuracy of the proposed method is the lowest. When the value of the batch size increases gradually, the value of OA also increases gradually and when the value of the batch size is larger than 75, the value of OA tends to be stable. Moreover, in our proposed method, the batch size is usually a multiple of the number of categories. Therefore, on the basis of the above analysis, the batch size is set to 75/15 = five times the category M throughout the following experiments.

4. Experimental Results

The four above-mentioned real PolSAR datasets are used to validate the effectiveness of the approach proposed in this section. For these datasets, the size of the input sample window is 15 × 15. The number of labeled samples is seven and the batch size is set to five times the number of categories M. The overall accuracy (OA) and classification result maps are used to evaluate the property of the proposed method. In order to evaluate the function of each part, the proposed method (3D-ARRN) was compared with five methods: RN, RN-SWANet, RRN, RRN-SWANet, and 3D-RRN. Where RN-SWANet represents the RN combined with SWANet, RRN represents the RN combined with the residual structure, RRN-SWANet represents the RRN combined with SWANet, and 3D-RRN represents the RRN combined with 3D-CNN and residual structure. All these comparison methods use almost similar parameter settings, including the RN structure, SWANet, 3D-CNN structure, residual structure, network layers, the size of filter, and the size of input patches.

4.1. Classification Results of the Flevoland I Dataset

Figure 12 depicts the classification results, while Table 3 lists the classification accuracies. The best results are shown in bold. As depicted in Table 3, the OA of the RN, RN-SWANet, RRN, RRN-SWANet, 3D-RRN, and 3D-ARRN are 87.05%, 89.70%, 92.57%, 94.22%, 95.23%, and 96.22%, respectively. Obviously, the proposed method achieved the highest OA among these five comparison methods, which is 9.17% higher than the RN model under the same condition. This result shows that the proposed network structure can significantly enhance accuracy of the RN network in PolSAR image classification. Compared to the RN-SWANet and RN, the OA of RN-SWANet is 2.65% higher than the RN model. This shows that the proposed SWANet module with feature selection in an RN can significantly enhance the classification accuracy of the RN method. Similarly, through comparing the RRN-SWANet, RRN, and RN, it can be found that by introducing the residual structure, the depth of the network is increased and the obtained feature information is more helpful to image classification. In comparing 3D-RRN and RRN, the OA of 3D-RRN is 1.01% higher than RRN. This shows that 3D-CNN increases the mutual relations between different channel features and effectively improves the classification accuracy of the network. Through these comparative experiments, it can be found that each module of the proposed network structure plays the positive effect in the final classification performance. These comparative results also show the effectiveness of our proposed method.
Moreover, compared with Figure 12c–h, the classification map of Figure 12h is obviously superior to Figure 12c–g. This result also proves that our proposed method can achieve effective classification with few labeled samples.

4.2. Classification Results of the Flevoland II Dataset

For this dataset, each category randomly selects seven labeled samples as support samples. Figure 13c–h depicts the classification results and Table 4 shows the classification accuracies.
As shown in Table 4, 3D-ARRN achieved the best accuracy at 95.03% among the comparison methods, and it exhibited significant performance improvements in the majority of categories, such as urban, water, and cropland area. In comparing 3D-ARRN with RN, RN-SWANet, RRN, RRN-SWANet, and 3D-RRN, 3D-ARRN obtained the highest OA. Through the comparison between these methods, it can also be discovered that each module of our proposed method contributes significantly to enhancing the classification accuracy of PolSAR images.
Compared with Figure 13c–h, the classification result shown in Figure 13h is closest to the ground truth. As depicted in Figure 13c, many regions are misclassified. In Figure 13d,e, there are a lot of isolated points in homogeneous regions, which are misclassification points. In Figure 13f,g, these isolated points have been greatly reduced, particularly in the urban and cropland categories. Furthermore, isolated points in Figure 13h are significantly fewer than in Figure 13c–g. This result also indicates the validity of the proposed method again and demonstrates that the method can achieve effective classification with few labeled samples.

4.3. Classification Results of the San Francisco Datasets

We randomly picked seven labeled samples for each category as support samples in this dataset. It contains five different terrain types, which means that a total of 35 labeled samples were picked as training samples. The visual classification results are depicted in Figure 14. Table 5 shows the classification accuracy corresponding to each category.
As depicted in Table 5, the classification accuracy of the RN, RN-SWANet, RN-Res, RRN-SWANet, 3D-RRN, and 3D-ARRN are 89.62%, 92.22%, 93.78%, 94.30%, 94.39%, and 96.09%, respectively. This indicates that 3D-ARRN obtained the highest classification accuracy in these comparison methods. These comparisons show that all parts of the suggested method can significantly improve the classification accuracy. From the final classification results, we can observe the ability of our proposed method to solve few-sample problems.
Figure 14 indicates the classification result maps of different methods on the San Francisco area. Compared with Figure 14c–h, our proposed classification map of Figure 14h is better than other methods. As is shown in Figure 14, some water areas of Figure 14c–e are mistaken for the type of vegetation area. In Figure 14f,g, these isolated points are greatly reduced, especially in vegetation, high-density urban, and water areas. In Figure 14h, the classification results of water areas, high-density cities, and developed areas are significantly better than those of other methods in these three areas. The experimental results also demonstrate that the suggested method is efficient.

4.4. Classification Results of the Xi’an Datasets

The dataset consists of three main categories: bench land, urban, and river. We randomly selected seven labeled samples from each class as the support set, which means that a total of 21 labeled samples were selected as training samples. The classification results are shown in Figure 15a–h and the classification accuracy is listed in Table 6.
As shown in Table 6, in comparing 3D-ARRN with RN, RN-SWANet, RRN, RRN-SWANet, and 3D-RRN methods, it can be found that the OA and Kappa of 3D-ARRN are higher than the other comparison methods. This is because when the number of labeled samples is few, these network models cannot be fully trained, and the fitting capability is poor. Moreover, these comparisons indicate that each part of the proposed method can effectively improve its classification accuracy. This also indicates the ability of the proposed method to solve few-sample problems and proves the superiority of the proposed method.
Figure 15 shows the classification result maps of different methods in the Xi’an area. Compared with Figure 15a–h, the proposed classification map of Figure 15h is better than the other methods. In Figure 15c–h, the isolated points are greatly reduced, especially in bench land and urban areas. As shown in Figure 15h, the proposed method has significant advantages over the other methods in bench land and urban areas. These experiment results also prove the effectiveness of the proposed method.

4.5. Comparisons with Other PolSAR Classification Methods

To verify the efficiency of our suggested method, we compared it to other recent PolSAR classification methods. Table 7 shows the classification accuracy of these methods on the first dataset: the Flevoland I area. In the 3D-ARRN, SVM [15], Wishart [13], 3D-CNN [42], SF-CNN [26], CNN-PL, and 3D-CNN + PL classification methods, seven labeled samples from each class were chosen at random as training samples. CNN + PL represents the CNN combined with the pseudo-label generated by the superpixel algorithm. 3D-CNN + PL represents the CNN combined with the pseudo-label generated by the superpixel algorithm. In the STLLE [44], MCCNN [45], MAE [46], and CV-3D-CNN [47] classification methods, 1% of the labeled samples are used as the training samples for the other four methods.
As can be observed from Table 7, the OA of the 3D-ARRN proposed is much higher than that of the SVM, Wishart, 3D-CNN, SF-CNN, CNN + PL, and 3D-CNN + PL methods. When there are seven labeled samples in each class, the accuracy rate of the proposed method is higher than that of the comparison method in most categories. Moreover, in comparing the proposed method with the STLLE, MCCNN, MAE, and CV-3D-CNN methods, the OA of the proposed method with 105 training samples is still higher than that of the STLLE, MCCNN, MAE, and CV-3D-CNN methods with 1682 training samples. These comparisons demonstrate that our proposed method can effectively solve the problem of limited samples and that it is significantly better than the other methods under the same conditions.

5. Conclusions

In this study, a new 3D-ARRN network was proposed for PolSAR image classification when labeled samples are few. Our method avoids the cost of manually labeling a large number of samples and effectively improves the classification accuracy under few-samples. The experiment results on four typical datasets from different radar systems indicate that the proposed method was superior to the comparative classification method in these four tested cases. The advantages of our proposed method have been demonstrated by these experiments. That is, (1) This method can solve the classification problem of PolSAR images with few samples effectively and it can obtain better classification results when there are only a few labeled samples in each category; (2) the training samples construction method based on characteristics of PolSAR data provides a new idea for PolSAR terrain classification with a limited number of samples and can also be applied to the PolSAR image classification method based on meta-learning; (3) the proposed attention mechanism of a SWANet in this paper can effectively play the role of feature selection and improve classification accuracy. Compared with other RN-based methods (RN, RN-SWANet, RRN, RRN-SWANet, and 3D-RRN), our proposed method obtained the higher classification accuracy under the same conditions, which shows that each part of an RN model can effectively improve the proposed method. Moreover, compared with several typical PolSAR image classification methods, our proposed method has significant advantages, especially when labeled samples are few. In the future, we will focus on the physical scattering mechanism and study PolSAR classification with few samples driven by the physical scattering mechanism.

Author Contributions

W.H. implemented the main ideas, carried out the evaluation, and wrote the manuscript. Y.Z. and C.Z. provided suggestions on presenting the results and writing the paper. X.J. provided some suggestions on experimental design. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China [No. 61901368], the Special scientific research plan project of the Shaanxi Provincial Department of Education [19JK0798], and the Natural science foundation of Shaanxi Province [2019JQ-377].

Data Availability Statement

The original PolSAR data are publicly available and can be found: https://ietr-lab.univ-rennes1.fr/polsarpro-bio/sample_datasets/. The data presented in this study are available on reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Parikh, H.; Patel, S.; Patel, V. Modeling PolSAR classification using convolutional neural network with homogeneity based kernel selection. Model. Earth Syst. Environ. 2023, 1–13. [Google Scholar] [CrossRef]
  2. Chen, J.; Xing, M.; Yu, H.; Liang, B.; Peng, J.; Sun, G.C. Motion Compensation/Autofocus in Airborne Synthetic Aperture Radar: A Review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 185–206. [Google Scholar] [CrossRef]
  3. Jianlai, C.; Hanwen, Y. Wide-beam SAR autofocus based on blind resampling. Sci. China Inf. Sci. 2022, 66, 140304. [Google Scholar]
  4. Fang, Z.; Zhang, G.; Dai, Q.; Xue, B.; Wang, P. Hybrid Attention-Based Encoder–Decoder Fully Convolutional Network for PolSAR Image Classification. Remote Sens. 2023, 15, 526. [Google Scholar] [CrossRef]
  5. Yang, R.; Hu, Z.; Liu, Y.; Xu, Z. A Novel Polarimetric SAR Classification Method Integrating Pixel-Based and Patch-Based Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 431–435. [Google Scholar] [CrossRef]
  6. Jamali, A.; Roy, S.K.; Bhattacharya, A.; Ghamisi, P. Local Window Attention Transformer for Polarimetric SAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2023, 1. [Google Scholar] [CrossRef]
  7. Chen, S.W.; Li, Y.Z.; Wang, X.S.; Xiao, S.P.; Sato, M. Modeling and Interpretation of Scattering Mechanisms in Polarimetric Synthetic Aperture Radar: Advances and perspectives. IEEE Signal Process. Mag. 2014, 31, 79–89. [Google Scholar] [CrossRef]
  8. Cloude, S.R.; Pottier, E. A review of target decomposition theorems in radar polarimetry. IEEE Trans. Geosci. Remote Sens. 1996, 34, 498–518. [Google Scholar] [CrossRef]
  9. Freeman, A.; Durden, S.L. A three-component scattering model for polarimetric SAR data. IEEE Trans. Geosci. Remote Sens. 1998, 36, 963–973. [Google Scholar] [CrossRef] [Green Version]
  10. Huynen, J.R. Phenomenological Theory of Radar Targets. Ph.D. Thesis, Faculty of Electrical Engineering, Mathematics and Computer Science, Delft, The Netherlands, 1970. [Google Scholar]
  11. Cloude, S.R.; Pottier, E. An entropy based classification scheme for land applications of polarimetric SAR. IEEE Trans. Geosci. Remote Sens. 1997, 35, 68–78. [Google Scholar] [CrossRef]
  12. Krogager, E. New decomposition of the radar target scattering matrix. Electron. Lett. 1990, 18, 1525–1527. [Google Scholar] [CrossRef]
  13. Jong-Sen, L.; Grunes, M.R.; Ainsworth, T.L.; Li-Jen, D.; Schuler, D.L.; Cloude, S.R. Unsupervised classification using polarimetric decomposition and the complex Wishart classifier. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2249–2258. [Google Scholar] [CrossRef]
  14. Tao, C.; Chen, S.; Li, Y.; Xiao, S. PolSAR Land Cover Classification Based on Roll-Invariant and Selected Hidden Polarimetric Features in the Rotation Domain. Remote Sens. 2017, 9, 660. [Google Scholar] [CrossRef] [Green Version]
  15. Tao, M.; Zhou, F.; Liu, Y.; Zhang, Z. Tensorial Independent Component Analysis-Based Feature Extraction for Polarimetric SAR Data Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2481–2495. [Google Scholar] [CrossRef]
  16. Long, Y.; Liu, X. SVM lithological classification of PolSAR image in Yushigou Area Qilian Mountain. Sci. J. Earth Sci. 2013, 3, 128–132. [Google Scholar]
  17. Ma, F.; Gao, F.; Sun, J.; Zhou, H.; Hussain, A. Attention Graph Convolution Network for Image Segmentation in Big SAR Imagery Data. Remote Sens. 2019, 11, 2586. [Google Scholar] [CrossRef] [Green Version]
  18. Xu, H.; Lv, X.; Wang, X.; Ren, Z.; Bodla, N.; Chellappa, R. Deep Regionlets: Blended Representation and Deep Learning for Generic Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1914–1927. [Google Scholar] [CrossRef] [Green Version]
  19. Geng, J.; Wang, H.; Fan, J.; Ma, X. SAR Image Classification via Deep Recurrent Encoding Neural Networks. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2255–2269. [Google Scholar] [CrossRef]
  20. Wang, L.; Yang, X.; Tan, H.; Bai, X.; Zhou, F. Few-Shot Class-Incremental SAR Target Recognition Based on Hierarchical Embedding and Incremental Evolutionary Network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
  21. Liu, F.; Jiao, L.; Hou, B.; Yang, S. POL-SAR Image Classification Based on Wishart DBN and Local Spatial Information. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3292–3308. [Google Scholar] [CrossRef]
  22. Hu, Y.; Fan, J.; Wang, J. Classification of PolSAR Images Based on Adaptive Nonlocal Stacked Sparse Autoencoder. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1050–1054. [Google Scholar] [CrossRef]
  23. Zhou, Y.; Wang, H.; Xu, F.; Jin, Y.Q. Polarimetric SAR Image Classification Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1935–1939. [Google Scholar] [CrossRef]
  24. Shang, R.; Wang, J.; Jiao, L.; Yang, X.; Li, Y. Spatial feature-based convolutional neural network for PolSAR image classification. Appl. Soft Comput. 2022, 123, 108922. [Google Scholar] [CrossRef]
  25. Cui, Y.; Liu, F.; Jiao, L.; Guo, Y.; Liang, X.; Li, L.; Yang, S.; Qian, X. Polarimetric Multipath Convolutional Neural Network for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  26. Hua, W.; Wang, S.; Liu, H.; Liu, K.; Guo, Y.; Jiao, L. Semisupervised PolSAR Image Classification Based on Improved Cotraining. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4971–4986. [Google Scholar] [CrossRef]
  27. Wang, S.; Guo, Y.; Hua, W.; Liu, X.; Song, G.; Hou, B.; Jiao, L. Semi-Supervised PolSAR Image Classification Based on Improved Tri-Training with a Minimum Spanning Tree. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8583–8597. [Google Scholar] [CrossRef]
  28. Fang, Z.; Zhang, G.; Dai, Q.; Kong, Y.; Wang, P. Semisupervised Deep Convolutional Neural Networks Using Pseudo Labels for PolSAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  29. Guo, J.; Wang, L.; Zhu, D.; Zhang, G. Semisupervised Classification of PolSAR Images Using a Novel Memory Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  30. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
  31. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. Adv. Neural Inf. Process. Syst. 2016, 29, 1–12. [Google Scholar]
  32. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  33. Chen, Z.; Fu, Y.; Wang, Y.-X.; Ma, L.; Liu, W.; Hebert, M. Image deformation meta-networks for one-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 8680–8689. [Google Scholar]
  34. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  35. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep Relation Network for Hyperspectral Image Few-Shot Classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef] [Green Version]
  36. Tong, X.; Yin, J.; Han, B.; Qv, H. Few-Shot Learning with Attention-Weighted Graph Convolutional Networks for Hyperspectral Image Classification. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 25–28 October 2020; pp. 1686–1690. [Google Scholar]
  37. Zuo, X.; Yu, X.; Liu, B.; Zhang, P.; Tan, X. FSL-EGNN: Edge-Labeling Graph Neural Network for Hyperspectral Image Few-Shot Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  38. Zhang, L.; Zhang, S.; Zou, B.; Dong, H. Unsupervised Deep Representation Learning and Few-Shot Classification of PolSAR Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  39. Wang, Z.; Chen, B.; Lu, R.; Zhang, H.; Liu, H.; Varshney, P.K. FusionNet: An Unsupervised Convolutional Variational Network for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Image Process. 2020, 29, 7565–7577. [Google Scholar] [CrossRef]
  40. Levinshtein, A.; Stere, A.; Kutulakos, K.N.; Fleet, D.J.; Dickinson, S.J.; Siddiqi, K. TurboPixels: Fast Superpixels Using Geometric Flows. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2290–2297. [Google Scholar] [CrossRef] [Green Version]
  41. Liu, L.; Hu, F.; Zhao, J. Action Recognition Based on Features Fusion and 3D Convolutional Neural Networks. In Proceedings of the 2016 9th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 10–11 December 2016; pp. 178–181. [Google Scholar]
  42. Zhang, L.; Chen, Z.; Zou, B.; Gao, Y. Polarimetric SAR Terrain Classification Using 3D Convolutional Neural Network. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4551–4554. [Google Scholar]
  43. Jong-Sen, L.; Grunes, M.R.; Grandi, G.d. Polarimetric SAR speckle filtering and its implication for classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2363–2373. [Google Scholar] [CrossRef]
  44. Liu, H.; Wang, Z.; Shang, F.; Yang, S.; Gou, S.; Jiao, L. Semi-Supervised Tensorial Locally Linear Embedding for Feature Extraction Using PolSAR Data. IEEE J. Sel. Top. Signal Process. 2018, 12, 1476–1490. [Google Scholar] [CrossRef]
  45. Wang, Y.; Cheng, J.; Zhou, Y.; Zhang, F.; Yin, Q. A Multichannel Fusion Convolutional Neural Network Based on Scattering Mechanism for PolSAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  46. Wang, J.; Hou, B.; Jiao, L.; Wang, S. POL-SAR Image Classification Based on Modified Stacked Autoencoder Network and Data Distribution. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1678–1695. [Google Scholar] [CrossRef]
  47. Zhang, P.; Tan, X.; Li, B.; Jiang, Y.; Song, W.; Li, M.; Wu, Y. PolSAR Image Classification Using Hybrid Conditional Random Fields Model Based on Complex-Valued 3-D CNN. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 1713–1730. [Google Scholar] [CrossRef]
Figure 1. The structure of the spatial weighted attention network.
Figure 1. The structure of the spatial weighted attention network.
Remotesensing 15 02025 g001
Figure 2. The structure of relation network.
Figure 2. The structure of relation network.
Remotesensing 15 02025 g002
Figure 3. The flowchart of constructing training samples. (a) denotes the labeled samples; (b) denotes the labeled superpixels; (c) denotes the similar labeled superpixels; (d) denotes the labeled training samples.
Figure 3. The flowchart of constructing training samples. (a) denotes the labeled samples; (b) denotes the labeled superpixels; (c) denotes the similar labeled superpixels; (d) denotes the labeled training samples.
Remotesensing 15 02025 g003
Figure 4. The overall network structure of the 3D-ARRN.
Figure 4. The overall network structure of the 3D-ARRN.
Remotesensing 15 02025 g004
Figure 5. AIRSAR dataset in Flevoland I: (a) Pauli image. (b) Ground-truth image.
Figure 5. AIRSAR dataset in Flevoland I: (a) Pauli image. (b) Ground-truth image.
Remotesensing 15 02025 g005
Figure 6. Flevoland II dataset: (a) Pauli image. (b) Ground-truth image.
Figure 6. Flevoland II dataset: (a) Pauli image. (b) Ground-truth image.
Remotesensing 15 02025 g006
Figure 7. RADARSAT-2 data in San Francisco: (a) Pauli image. (b) Ground-truth image.
Figure 7. RADARSAT-2 data in San Francisco: (a) Pauli image. (b) Ground-truth image.
Remotesensing 15 02025 g007
Figure 8. RADARSAT-2 data in Xi’an: (a) Pauli image. (b) Ground-truth image.
Figure 8. RADARSAT-2 data in Xi’an: (a) Pauli image. (b) Ground-truth image.
Remotesensing 15 02025 g008
Figure 9. Size of the input sample window and OA.
Figure 9. Size of the input sample window and OA.
Remotesensing 15 02025 g009
Figure 10. Overall accuracy (OA) with different labeled samples.
Figure 10. Overall accuracy (OA) with different labeled samples.
Remotesensing 15 02025 g010
Figure 11. The classification accuracy varies with batch size.
Figure 11. The classification accuracy varies with batch size.
Remotesensing 15 02025 g011
Figure 12. Classification results in the Flevoland I area. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Figure 12. Classification results in the Flevoland I area. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Remotesensing 15 02025 g012
Figure 13. Classification results in the Flevoland II area. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Figure 13. Classification results in the Flevoland II area. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Remotesensing 15 02025 g013
Figure 14. Classification results of the San Francisco dataset. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Figure 14. Classification results of the San Francisco dataset. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Remotesensing 15 02025 g014
Figure 15. Classification results of the Xi’an dataset. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Figure 15. Classification results of the Xi’an dataset. (a) Pauli image. (b) Ground-truth image. (c) RN. (d) RN-SWANet. (e) RRN. (f) RRN-SWANet. (g) 3D-RRN. (h) 3D-ARRN.
Remotesensing 15 02025 g015
Table 1. PolSAR features used in this study.
Table 1. PolSAR features used in this study.
FeatureDescription
H, a, A, λ 1 , λ 2 , λ 3 Cloude decomposition [9]
P s , P d , P v Freeman–Durden decomposition [10]
| k s | 2 , | k v | 2 , | k d | 2 Krogager decomposition [12]
A 0 , B + B 0 , B B 0 , C , D , E , F , G , H Huynen decomposition [11]
| a | 2 , | b | 2 , | c | 2 Pauli decomposition [8]
θ n u l l _ Re [ T 12 ] , θ n u l l _ Im [ T 12 ] Roll-Invariant Polarimetric Features [7]
A T , B T , C T , D T , E T , F T Elements from the coherency matrix
Table 2. The specific parameters of the proposed classification network.
Table 2. The specific parameters of the proposed classification network.
Proposed Method
Layer NameEmbedding Module
Support SamplesQuery Samples
Output ShapeFilter SizePaddingOutput ShapeFilter SizePadding
InputK × 1 × 9 × 15 × 15N/ANQ × 1 × 9 × 15 × 15N/AN
Convolution1K × 9 × 9 × 15 × 151 × 9 × 3 × 3 × 3YQ × 9 × 9 × 15 × 151 × 9 × 3 × 3 × 3Y
Convolution2K × 9 × 9 × 15 × 159 × 9 × 3 × 3 × 3YQ × 9 × 9 × 15 × 159 × 9 × 3 × 3 × 3Y
Convolution3K × 9 × 9 × 15 × 159 × 9 × 3 × 3 × 3YQ × 9 × 9 × 15 × 159 × 9 × 3 × 3 × 3Y
ShortcutConvolution1 + Convolution3Convolution1 + Convolution3
Spatial WeightK × 9 × 9 × 15 × 15N/ANQ × 9 × 9 × 15 × 15N/AN
Max pooling1K × 9 × 9 × 7 × 71 × 2 × 2NQ × 9 × 9 × 7 × 71 × 2 × 2N
Convolution4K × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3YQ × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3Y
Convolution5K × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3YQ × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3Y
Convolution6K × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3YQ × 32 × 9 × 7 × 79 × 32 × 3 × 3 × 3Y
Short cutConvolution4 + Convolution6Convolution4 + Convolution6
Spatial WeightK × 32 × 9 × 7 × 7N/ANQ × 32 × 9 × 7 × 7N/AN
Max pooling2K × 32 × 9 × 3 × 31 × 2 × 2NQ × 32 × 9 × 3 × 31 × 2 × 2N
Convolution7K × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3YQ × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3Y
Convolution8K × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3YQ × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3Y
Convolution9K × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3YQ × 64 × 9 × 3 × 332 × 64 × 3 × 3 × 3Y
ShortcutConvolution7 + Convolution9Convolution7 + Convolution9
Spatial WeightK × 64 × 9 × 3 × 3N/ANQ × 64 × 9 × 3 × 3N/A N
Relation module
Output ShapeFilter SizePadding
Convolution10Nm × 128 × 9 × 3 × 3128 × 64 × 1 × 3 × 3Y
Max pooling3Nm × 64 × 11 × 1 × 11 × 2 × 2N
Convolution11Nm × 64 × 13 × 1 × 164 × 64 × 1 × 3 × 3Y
FlattenNm × 832N/AN
FlattenNm × 8N/AN
OutputNm × 1N/AN
Note: K and Q denote the number of support samples and query samples respectively; N m = Q × M , where M represents the number of categories of the classification.
Table 3. The classification accuracy (%) of different methods with 7 labeled samples per class on Flevoland I.
Table 3. The classification accuracy (%) of different methods with 7 labeled samples per class on Flevoland I.
MethodRNRN-SWANetRRNRRN-SWANet3D-RRN3D-ARRN
Region
Stem beans97.1996.0597.2798.1096.7396.54
Rapeseed79.2586.1490.2492.7588.9493.20
Bare soil99.7199.9810099.4999.9699.71
Potatoes74.6491.1586.0295.2795.8199.16
Beet88.3889.4295.2197.7793.1293.76
Wheat 284.2488.2496.0292.4095.0691.58
Peas91.2892.8898.5999.8299.7098.68
Wheat 390.9288.8297.5894.7898.4398.86
Lucerne97.3993.6196.3594.0496.9498.24
Barley93.5796.2896.6397.5199.1498.22
Wheat81.8377.3281.9688.7486.7397.07
Grasses87.7291.0991.3695.0294.5892.00
Forest93.8796.2096.1694.9597.1297.71
Water77.5283.9783.8387.9196.5290.45
Building88.6091.0793.5497.6695.7498.35
OA87.0589.7092.5794.2295.2396.22
Table 4. The classification accuracy of different methods with 7 labeled samples per class on Flevoland II.
Table 4. The classification accuracy of different methods with 7 labeled samples per class on Flevoland II.
MethodRNRN-SWANetRRNRRN-SWANet3D-RRN3D-ARRN
Region
Urban88.2295.0393.1789.0195.4696.19
Water99.2198.1198.6598.5697.5098.86
Forest92.7891.6994.0492.2691.0992.78
Cropland80.2384.1684.1389.9592.4193.20
OA89.9891.5592.0892.7993.8495.03
Table 5. The classification accuracy of all methods with 7 labeled samples per class on the San Francisco dataset.
Table 5. The classification accuracy of all methods with 7 labeled samples per class on the San Francisco dataset.
MethodRNRN-SWANetRRNRRN-SWANet3D-RRN3D-ARRN
Region
Water93.0798.8099.7999.4399.7599.98
Vegetation94.1487.5075.3372.9483.3790.28
Low-Density Urban73.6767.4395.5293.4879.9990.00
High-Density Urban84.8689.9790.7896.6195.1592.73
Developed87.1789.2196.0596.9292.9997.53
OA89.6292.2293.7894.3094.3996.09
Table 6. The classification accuracy (%) of different methods with 7 labeled samples per class on Xi’an dataset.
Table 6. The classification accuracy (%) of different methods with 7 labeled samples per class on Xi’an dataset.
MethodRNRN-SWANetRRNRRN-SWANet3D-RRN3D-ARRN
Region
Bench land77.1680.8280.6982.2682.3684.19
Urban79.6676.9284.9382.4385.0892.41
River 93.9494.1482.5191.2693.0488.97
OA80.5681.4482.8684.2485.2788.12
Table 7. The classification accuracy (%) of various methods on the Flevoland I dataset.
Table 7. The classification accuracy (%) of various methods on the Flevoland I dataset.
Method3D-
ARRN
SVMWishart3D-
CNN
SF-
CNN
CNN
+PL
3D-CNN
+PL
STLLEMC-
CNN
MAECV-3D-
CNN
Region
Sample
Number
Labeled Sample Per Class (The Total Number Is 105)Training Ratio
(The Total Number Is 1682)
71%
Stem beans96.5466.2389.7281.1187.5299.1698.2897.3597.0195.9198.63
Rapeseed93.2056.9171.1311.4291.2691.2286.4786.8791.9484.1197.48
Bare soil99.7179.5198.0494.0310010099.8697.4691.7092.6292.74
Potatoes99.1648.0574.9964.9293.5385.8991.5193.3596.0489.6493.60
Beet93.7668.6691.3285.2596.2596.8296.9596.9993.2695.7795.21
Wheat 291.5831.6364.1773.1389.8488.2992.8183.8598.1481.0295.73
Peas98.6870.4194.6438.3897.7895.2598.3397.6197.7696.4287.65
Wheat 398.8659.9776.7769.9184.8496.9897.3295.0697.6095.0699.44
Lucerne98.2465.5793.6981.2294.7994.3790.1894.6597.7795.3484.81
Barley98.2255.1091.8982.1494.1897.9594.3989.3799.3995.9884.14
Wheat97.0746.8579.8553.0292.5489.3383.3588.3286.2491.5798.79
Grasses92.0061.2164.7255.6884.6188.4395.2981.6897.8286.4172.39
Forest97.7182.5151.1058.1795.2496.7795.8390.1999.1491.1399.85
Water90.4569.4981.3174.4290.8083.8192.8998.8798.1598.0299.95
Building98.3561.7783.9458.7788.7198.0898.6386.8198.3884.0996.22
OA96.2261.5980.4965.4492.1392.6393.1492.3395.8392.0193.42
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hua, W.; Zhang, Y.; Zhang, C.; Jin, X. PolSAR Image Classification Based on Relation Network with SWANet. Remote Sens. 2023, 15, 2025. https://doi.org/10.3390/rs15082025

AMA Style

Hua W, Zhang Y, Zhang C, Jin X. PolSAR Image Classification Based on Relation Network with SWANet. Remote Sensing. 2023; 15(8):2025. https://doi.org/10.3390/rs15082025

Chicago/Turabian Style

Hua, Wenqiang, Yurong Zhang, Cong Zhang, and Xiaomin Jin. 2023. "PolSAR Image Classification Based on Relation Network with SWANet" Remote Sensing 15, no. 8: 2025. https://doi.org/10.3390/rs15082025

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop