Next Article in Journal
On the Importance of Non-Gaussianity in Chlorophyll Fluorescence Imaging
Next Article in Special Issue
A Novel Dual-Encoder Model for Hyperspectral and LiDAR Joint Classification via Contrastive Learning
Previous Article in Journal
A Single-Site Vehicle Positioning Method in the Rectangular Tunnel Environment
Previous Article in Special Issue
Hyperspectral Image Classification Based on a 3D Octave Convolution and 3D Multiscale Spatial Attention Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Attention-Based Encoder–Decoder Fully Convolutional Network for PolSAR Image Classification

1
Key Lab of Radar Imaging and Microwave Photonics, Ministry of Education, Nanjing University of Aeronautics and Astronautics, Nanjing 211100, China
2
Nanjing University of Aeronautics and Astronautics Shenzhen Research Institute, Shenzhen 518000, China
3
Shanghai Key Lab of Intelligent Information Processing, Fudan University, Shanghai 200433, China
4
Fujian Provincial Key Lab of Coastal Basin Environment, Fujian Polytechnic Normal University, Fuqing 350300, China
5
FKey Laboratory of Southeast Coast Marine Information Intelligent Perception and Application, Ministry of Natural Resources, Zhangzhou Institute of Surveying and Mapping, Zhangzhou 363001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(2), 526; https://doi.org/10.3390/rs15020526
Submission received: 15 November 2022 / Revised: 5 January 2023 / Accepted: 11 January 2023 / Published: 16 January 2023
(This article belongs to the Special Issue Recent Advances in Processing Mixed Pixels for Hyperspectral Image)

Abstract

:
Recently, methods based on convolutional neural networks (CNNs) achieve superior performance in polarimetric synthetic aperture radar (PolSAR) image classification. However, the current CNN-based classifiers follow patch-based frameworks, which need input images to be divided into overlapping patches. Consequently, these classification approaches have the drawback of requiring repeated calculations and only relying on local information. In addition, the receptive field size in conventional CNN-based methods is fixed, which limits the potential to extract features. In this paper, a hybrid attention-based encoder–decoder fully convolutional network (HA-EDNet) is presented for PolSAR classification. Unlike traditional CNN-based approaches, the encoder–decoder fully convolutional network (EDNet) can use an arbitrary-size image as input without dividing. Then, the output is the whole image classification result. Meanwhile, the self-attention module is used to establish global spatial dependence and extract context characteristics, which can improve the performance of classification. Moreover, an attention-based selective kernel module (SK module) is included in the network. In the module, softmax attention is employed to fuse several branches with different receptive field sizes. Consequently, the module can capture features with different scales and further boost classification accuracy. The experiment results demonstrate that the HA-EDNet achieves superior performance compared to CNN-based and traditional fully convolutional network methods.

1. Introduction

Remote sensing is a significant component of earth observation since it can detect and identify scenes based on physical characteristics. Remote-sensing detectors measure reflected and emitted radiation of objects without establishing direct touch. Polarimetric synthetic aperture radar (PolSAR), an efficient microwave detector, has attracted great attention [1,2,3]. PolSAR employs different polarimetric channels to obtain the polarimetric scattering characteristics of objects, which offers conveniences for subsequent information extraction of geoscience applications. In particular, it can provide more structural information than single-polarized SAR systems. Moreover, PolSAR can be used at any time and in any weather, so several successful applications have been made in environmental monitoring [4], resource management [5], urban planning [6], military [7], and so on. Among these applications, classification is a critical and difficult process that entails categorizing polarimetric scattering points into some predefined categories according to their scattering properties [8].
In recent years, copious classification methods for PolSAR image classification have been developed. Scattering-mechanism-based methods are frequently used, which employ the scattering information and imaging mechanism to increase the classification accuracy [9,10]. These techniques extract different scattering characteristics from the coherency matrix or covariance matrix of PolSAR images, such as Krogager decomposition [11], Huynen decomposition [12], Cameron decomposition [13], Freeman decomposition [14], H/alpha decomposition [15], Pauli decomposition [16], and so on. The methods based on scattering mechanisms are straightforward and efficient. Statistical-distribution-based methods have achieved considerable interest in the past few years. These approaches use different distributions to represent PolSAR data. Lee et al. [17] utilized the H/alpha decomposition and the complex Wishart classifier to process unsupervised PolSAR classification. Liu et al. [18] used a Wishart deep belief network (W-DBN) and local spatial information to classify. Jiao and Liu [19] combined Wishart distribution with a deep stacking network (W-DSN) for PolSAR classification. Xie et al. [20] proposed a PolSAR classification approach based on the complex Wishart distribution and convolutional autoencoder. Recently, with the advancement of machine learning, some classifiers such as support vector machines (SVMs) [21], decision trees (DTs) [22], and K-nearest neighbor (KNN) [23] are used for PolSAR classification and achieve superior performance compared with target decomposition approaches. However, these approaches require manual feature extraction and use only pixel-based polarimetric features. Deep-learning-based algorithms have been developed in several fields, including natural language processing [24] and computer vision [25], because of their exceptional performance. The deep structure of deep learning enables the model to learn discriminative, invariant, and high-dimensional data features autonomously. Several deep-learning-based frameworks have been introduced into PolSAR classification. For example, deep transfer learning [26], deep reinforcement learning [27], the sparse autoencoder [28], the convolutional neural network (CNN) [29], and the long short-term memory (LSTM) network [30] are frequently used. In these methods, CNN-based methods are commonly employed and have achieved tremendous success.
Due to its advantages in local contextual perception and feature transformation with parameter sharing, CNN-based approaches have gained popularity. Nevertheless, within the local learning framework, these CNN-based approaches often have repetitive calculations [31] for PolSAR image classification. The framework process consists of two parts: generating overlapping image patches and assigning labels to the corresponding central pixels. Patches generated by adjacent pixels overlap with each other, resulting in redundant computation. For this reason, approaches under the patch-based framework have difficulty running quickly. Moreover, the finite patch size only constrains some local features. Hence, it is hard for CNN-based methods to build long-range dependency. Additionally, traditional CNN-based methods usually adopt the fixed kernel size for feature extraction, which restricts the capability of context information extraction [32]. In general, the fixed kernel size cannot capture fine-grained and coarse-grained terrain structures simultaneously, which influences the performance of PolSAR classification. Recent research indicates that varying spatial kernel sizes are helpful to classification. Unfortunately, it is not easy to choose the weights of various kernel sizes.
In the recent years, attention-based methods have also been applied in PolSAR image classification. The purpose of the attention mechanism is to allow a neural network to focus on specific parts of its input when processing it, rather than having to consider the entire input equally. The existing attention-based methods for PolSAR image classification are generally based on two kinds of attention mechanisms: channel-driven attention and spatial-driven attention. Dong et al. [33] proposed an attention-based polarimetric feature selection module for a CNN network, which captures the relationship between input polarimetric features and ensures the validity of high-dimensional data classification. Hua et al. [34] introduced a feature selection method based on spatial attention to enhance the relationship between pixel spatial information. Yang et al. [35] introduced a convolutional block attention module to achieve better classification performance and accelerate network convergence. Ren et al. [36] proposed a residual attention module to enhance discriminate features in multiple resolutions. The existing attention-based methods are focused on channel and spatial features, but multi-scale features obtained by using different receptive fields are often ignored.
In light of the challenges mentioned above, we proposed a hybrid attention-based encoder–decoder fully convolutional network (EDNet) called HA-EDNet for PolSAR classification. In HA-EDNet, the EDNet is constructed as the patch-free backbone network. The network accepts arbitrary-size input data without any pretreatment. Similar to the human visual system, the attention mechanism is commonly used in computer vision since it can inhibit irrelevant features and enhance important features. Self-attention is one the most popular attention mechanisms and has become the dominant paradigm in NLP [37]. In this paper, self-attention is designed to build the long-range dependency between pixels of a PolSAR image. Moreover, the attention-based selective kernel module (SK-module) [38] is utilized to replace traditional convolution operations. This module can adjust the kernel sizes automatically for different terrain sizes. Compared with the conventional CNN models, the HA-EDNet framework can deal with repeated calculation and observe objects from multi-scale and long-distance perspectives.
The main contributions are summarized as follows:
(1)
An end-to-end encoder–decoder fully convolution network called EDNet is proposed to classify PolSAR images. The approach follows a patch-free architecture and accepts arbitrary-size input images. Then, the output is the whole image classification result.
(2)
A self-attention module is embedded into EDNet for global information extraction, where long-distance dependencies are modeled. Moreover, the self-attention module makes the classification results more refined and discriminative.
(3)
To further boost the performance, the SK module is used to extract multi-scale features, where different kernel sizes are fused by softmax attention. In this module, more discriminating features are extracted for better PolSAR classification.
(4)
Four widely known datasets are employed to test the effectiveness of the proposed approach. The experimental results show that the approach has better visual performance and classification accuracy than state-of-the-art methods.
The remainder of this paper is structured as follows. In Section 2, the related works on the fully convolutional network, attention mechanism, and self-attention are shown. In Section 3, we formulate the proposed methods in detail. Section 4 exhibits experimental results and discussions of four widely used PolSAR images. Finally, Section 5 depicts the conclusion and future work.

2. Related Works

This section introduces the fundamental principles of fully convolutional networks, attention mechanisms, and self-attention.

2.1. Fully Convolutional Network

A fully convolutional network (FCN) is built on the foundation of a traditional CNN and was originally intended for pixel-by-pixel image semantic segmentation [39]. A FCN model comprises three basic layers: convolution layers, pooling layers, and deconvolution layers. The convolution layer based on the shared-weight structure is used as a feature extraction layer, which can extract abstract and advanced information. The convolutional process in a CNN reduces the feature map’s size and resolution. The pooling layer aids in the transformation of high-dimensional characteristics into low-dimensional representative features, resulting in a reduction in spatial size and computation parameters. The deconvolution layer is the inverse operation of the convolution layer and pooling layer. The FCN changes the final fully connected layers with a 1 × 1 kernel size convolution operation to recover the size of the input data. The deconvolution operations restore the feature map created by the convolution and pooling layers to the original size using bilinear interpolation. This continuous upsampling operation assigns each predicted result to a pixel of input image, achieving end-to-end and dense classification. Moreover, the skip connection operations are used by FCN to reduce the loss of detailed information during the downsampling procedure by merging local and global feature information. Skip connection operations use the shallow convolutional layers’ spatial detail characteristics to augment the semantic features of the higher convolutional layers.

2.2. Attention Mechanism

The attention mechanism is a sophisticated cognition that is vital for human beings [40]. The key aspect of perception is that humans typically do not process all information simultaneously. Instead, people have a tendency to focus only on the information that is relevant at the time and place when it is required while ignoring other perceptible information at the same time. For example, while visually observing objects, individuals often do not see all of the scenery from beginning to finish but instead notice and pay attention to particular areas as required. People will learn to concentrate on it when comparable scenarios recur and pay greater attention to the advantageous feature if they discover that a scene often includes something they wish to notice in a certain portion. The mechanism allows people to swiftly choose high-value information from huge amounts of data while utilizing restricted processing capabilities. The attention mechanism substantially improves the speed and accuracy of processing perceptual information [41].

2.3. Self-Attention

Self-attention, often referred to as intra-attention, is a special attention mechanism that connects different points and models long-range dependency of features to calculate a representation of the same feature [42]. It is very beneficial in video classification, semantic segmentation, or image description generation. In an image or a sentence, the context at one point is calculated by self-attention as the total weight of all points. Self-attention was first used in machine translation to gauge the inputs’ overall interdependencies. Following that, several self-attention-based strategies in the area of computer vision have been presented [43]. These methods use contextual data to improve feature representation by various self-attention processes.

3. Proposed Method

In this section, the HA-EDNet is discussed in depth. First, the representation of PolSAR data is shown. Secondly, the SK module is introduced. Thirdly, the self-attention module is presented. Finally, the structure of the proposed network is shown.

3.1. Representation of PolSAR Data

To identify the scattering characteristics of targets, the scattering matrix is employed. The scattering matrix represents the horizontal and vertical polarization states of the sent and received signals. The following is the representation of the scattering matrix:
S = S H H S H V S V H S V V
Complex-valued scattering coefficients are denoted by S H H , S H V , S V H , and S V V in this formula, where S H V represents horizontal transmitting and vertical receiving. The other coefficients are defined similarly.
The scattering features of PolSAR were previously represented using a statistical coherence matrix due to speckle noise. In the condition of reciprocity ( S H V = S V H ), every pixel’s coherence matrix is expressed as a complex value matrix:
T = k p k p H = T 11 T 12 T 13 T 21 T 22 T 23 T 31 T 32 T 33
where k p = 1 2 S H H + S V V S H H S V V 2 S H V T is the Pauli scattering vector, and the superscript H is the conjugate transpose. Because the coherence matrix T is a Hermitian matrix, it is equivalent to its conjugate transpose. As a result, the polarimetric characteristics are represented by a 9-dimensional real vector, denoted as:
v = T 11 , T 22 , T 33 , Re T 12 , Im T 12 Re T 13 , Im T 13 , Re T 23 , Im T 23
where Re · and Im · represent the real and imaginary components of a complex value, respectively. For subsequent processing, every pixel of a PolSAR image is represented as the 9-dimensional real vector.

3.2. Attention-Based Selective Kernel Module

In PolSAR classification, approaches based on CNN have produced satisfactory results. However, in the above ways, the fixed kernel size is used for feature extraction, which limits the capability of multi-scale feature extraction. Therefore, it is necessary to automatically alter the kernel sizes of the network to improve the efficiency of PolSAR classification, which can be achieved by the SK module [44].
Figure 1 shows the illustration of the SK module, which is made up of three operations: split, fusion, and selection. The module utilizes feature maps X R H × W × C as input and produces the output feature maps O R H × W × C . Therefore, the module can be presented as:
O = f sk ( X ; θ )
where θ denotes parameters in the module. H, W, and C are the height, width, and channel of feature maps. In the split operation, two transformations: X U ˜ R H × W × C and X U ^ R H × W × C are utilized to illustrate, and 3 × 3 and 5 × 5 kernel sizes are used in the two transformations, respectively. Two output feature maps U ˜ and U ^ can be formulated as:
U ˜ = f ˜ ( X ) = X × W 3 × 3 + b
U ^ = f ^ ( X ) = X × W 5 × 5 + b
where W and b are convolutional kernels and biases, respectively. Different kernel sizes are employed to extract multi-scale information. Moreover, the Batch Normalization (BN) and activation function are included in the two transformations.
In the SK module, the fusion operation aims to allow neurons to learn multi-scale features by automatically adjusting the kernel sizes jointly. Firstly, an element-wise summation combines feature maps from the split operation. The output U R H × W × C is given by:
U = U ˜ U ^
where ⊕ represents element-wise addition operation. Then, a squeeze-and-excitation block (SE block) is utilized to extract the global information via global average pooling [45]. Finally, the output S R C contains channel-wise statistics and is calculated by reducing U through spatial dimensions using global average pooling (GAP):
S c = f G A P U c = 1 H × W i = 1 H j = 1 W U c ( i , j )
where S c denotes the cth element of output S , and U c is the cth channel of the fusion feature map U . Then, a fully connected layer with a ReLU function is used to generate compact features that can guide precise and adaptive selections. The compact feature Z R C r × 1 can be formulated as:
Z = ReLU ( W · S ) ,
where W R C r × C is a weight matrix, and r denotes a reduction ratio. In this paper, the reduction ratio r is fixed at 16.
In the selection operation, selective kernel attention across channels is utilized to select different scales of features adaptively. The compact feature Z is applied to compute the selective kernel attention vectors a and b by the fully connected layer and softmax function:
a c = e A c Z e A c Z + e B c Z
b c = e B c Z e A c Z + e B c Z
where a c and b c denote the cth element of attention vector a and b . Here, A , B R C × C r and A c and B c are the cth row of A , B . Moreover, the elements a c and b c have the following relationship:
a c + b c = 1
Finally, the output feature map O of the SK module with two kernel sizes is calculated by the attention vector as follows:
O = ( a U ˜ ) ( b U ^ )
Each neuron can modify the size of its receptive field using the SK module depending on multiple scales of input features. In the module, softmax attention is used to fuse different kernel sizes with the information in corresponding branches. Moreover, the module can capture target objects with multi-scales information, which is important for classification.

3.3. Self-Attention Module

In order to capture more contextual and spatial information in the learning network, we use a self-attention module to build long-range dependency and extract the complex land cover areas effectively. The self-attention module is illustrated in Figure 2, where X R H × W × C is the input feature map. H, W, and C denote the width, height, and channel, respectively. Then, three convolutions with 1 × 1 kernel are employed to transform the input feature map into three diverse embeddings.
α = W α ( X ) β = W β ( X ) γ = W γ ( X )
where α R H × W × N , β R H × W × N , γ R H × W × N , and N indicate the channel of reshaped feature map. Then, α , β , and γ are reshaped to ( H × W ) × N . To obtain the spatial self-attention map A R H W × H W , the matrix multiplication is applied between α and the transpose of β with the softmax function:
A ( i , j ) = exp α i × β j T k = 1 H W exp α k × β j T
Then, the spatial self-attention map A is multiplied by γ , and the result is B = A × γ .
The result B is reshaped to H × W × N . The final self-attention enhanced feature map P is formulated as:
P = F ( B ) + X
where F ( · ) is the nolinear transformation implemented by a convolutional layer with 1 × 1 kernel. It can be seen from Equation (16) that the attention feature map P is the sum of the the global feature map and input feature map. The global feature map contains relationships across all positions in the PolSAR image. This property enables the network to build the global spatial dependency for pixels belonging to the same category. Moreover, the global information contained in the PolSAR image can significantly improve the robustness of deep neural networks when confronted with blur [46].

3.4. PolSAR Classification with HA-EDNet

Patch-based CNN methods need to divide the input image into overlapping patches, which results in high computational complexity. This paper proposes a patch-free network architecture called HA-EDNet for full image classification. In patch-free networks, the explicit patching is replaced with the implicit receptive field of the model. The patch-free networks can avoid redundant computation on the overlapping areas and obtain a wider latent spatial context. The network accepts arbitrary-size images as input without pretreatment, and the output is the classification results of the whole image. The proposed model is described in the following.
As shown in Figure 3, the HA-EDNet is comprised of two basic subnetworks: (1) the encoder subnetwork and (2) the decoder subnetwork. As input, the coherence matrices of the whole PolSAR image are employed. Then, the encoder subnetwork is used to compute the hierarchical convolutional feature maps of the input PolSAR image, which is started with a 3 × 3 convolutional layer, a BN layer, and a ReLU function. The remaining part is composed of three SK modules and three downsampling layers. Here, 2 × 2 average pooling layers are utilized as the downsampling layers. In the top layer of the encoder subnetwork, the self-attention module is used to build long-range dependency. The decoder subnetwork is used to recover the spatial dimension of the coarsest convolutional feature map, which is a sample composed of three 3 × 3 convolutional layers and three upsample layers. Here, the upsampling layer is the bilinear interpolate function with a factor of 2. To effectively combine the spatial detail features in the encoder subnetwork and the semantic features in the decoder subnetwork, the feature maps of every SK module are added to the same size output of the decoder subnetwork. Finally, the softmax function is utilized to classify.
In the training process, the model is optimized by the cross-entropy loss function, which is written as:
L = 1 n i = 1 n y × log σ ( z ) i
y = 1 , if z belongs to class i 0 , otherwise
where z is the predicted result of the network. When z belongs to the ith class, y = 1 ; otherwise, y = 0 . σ ( z ) i is the output of the softmax function, which is the probability of belonging to a certain category.

4. Experiments and Results

In this part, we introduce four widely used PolSAR datasets employed in our experiments and experimental settings of the proposed approach. Additionally, classification results based on the proposed network and comparison approaches are presented.

4.1. Datasets

To evaluate the effectiveness of the proposed network, four PolSAR datasets are employed as follows:
(1)
Flevoland-15 dataset: The Flevoland-15 dataset is the most widely used PolSAR dataset and is L-band polarimetric data obtained by the Airborne Synthetic Aperture Radar (AIRSAR) of Flevoland, Netherlands, in 1989. This dataset is regarded as a benchmark dataset for PolSAR classification. The pseudocolor image, ground truth, and legends of classes are shown in Figure 4.
(2)
Flevoland-14 dataset: The Flevoland-14 dataset is also L-band data collected by the AIRSAR in 1991 over Flevoland. The pseudocolor image, ground truth, and legends of classes are shown in Figure 5. This dataset includes fourteen types of objects.
(3)
Oberpfaffenhofen dataset: The Oberpfaffenhofen dataset is acquired from the L-band ESAR sensor that covers Oberpfaffenhofen, Germany. The pseudocolor image, ground truth, and legends of classes are shown in Figure 6. The ground truth contain three land cover types.
(4)
San Francisco dataset: The San Francisco dataset is obtained from the C-band RADARSAT-2, which covered San Francisco in 2008. The pseudocolor image, the ground truth, and legends of classes are shown in Figure 7. The ground truth of the dataset includes five types of objects.

4.2. Experimental Setting

In this paper, training samples adopt a randomly selected strategy from each class, while the remaining samples serve as the testing set. Three assessment measures, including overall accuracy (OA), average accuracy (AA), and the kappa coefficient (Kappa), are used to assess classification performance. The experiments of four datasets are implemented in Python 3.7 and Pytorch, with Intel Silver 4210v4 2.2GHz CPU, NVIDIA RTX 2080Ti GPU, and 64G RAM.
The input layer of the proposed approach has a size of H × W × C , where H and W are the height and width of the input, respectively. C denotes the channel of a PolSAR dataset, which is the nine-dimensional real vector. The initial weights are chosen randomly for all methods. Two improtant training parameters: learning rate and the number of iterations are set as 0.005 and 300, respectively. Every method is tested ten times using different training samples. The averaged results are employed to compare.
In the experiments, the training set is selected randomly from the ground truth. To solve the imbalance issue, we randomly choose labeled samples of each annotated class for the training set instead of dividing the labeled samples by an average percentage. Finally, the training and testing samples of four datasets are listed in Table 1, Table 2, Table 3 and Table 4.

4.3. Result on the Flevoland-15 Dataset

The first dataset in Flevoland is employed to evaluate the efficiency of the proposed HA-EDNet method. Several state-of-the-art models are chosen for comparison, including CNN, 3D-CNN [47], multi-scale CNN (MS-CNN) [48], complex-valued CNN (CV-CNN) [49], FCN [50], and U-Net [51]. The classification accuracies of seven different approaches are shown in Table 5. For the PolSAR classification task, the visual performance of the classification network is essential. Figure 8 shows the classification results of the Flevoland-15 dataset. The pseudocolor image, ground truth map, CNN, 3D-CNN, MS-CNN, CV-CNN, FCN, U-Net, and HA-EDNet classification map are shown in Figure 8a–i, respectively. Consequently, several conclusions can be achieved from Figure 8 and Table 5. As is shown in Figure 8c,e, the MS-CNN classifier has a better performance than the CNN classifier because the MS-CNN classifier can extract multi-scale features that are strongly discriminative. HA-EDNet methods are superior to the FCN and U-Net classifier in spatial uniformity when comparing Figure 8g–i. It can be attributed to the fusion of spatial detail features, and the introduction of the hybrid attention mechanism can capture more discriminative features. In addition, the proposed HA-EDNet approach produces classification results with less noise and more precision than prior methods. Table 5 lists the results of the seven methods. From Table 5, we can find that the MS-CNN method wins nine categories and achieves higher accuracies (OA of 92.82%, AA of 91.32%, and Kappa of 0.9217) than CNN. It indicates that using multi-scale features improve the classification accuracy. Comparing the results of patch-wise with patch-free methods, patch-free methods obtain more accurate results than patch-based methods. It demonstrates that patch-free models can effectively extract spatial features. From Table 5, it is shown that the proposed HA-EDNet method obtained a higher OA of 99.39%, exceeding CNN by about 8.3%. These results demonstrate the effectiveness of the proposed module. In Table 6, we also compare the proposed method with the other four state-of-the-art methods, and the results are shown in Table 6. The proposed method achieves the highest accuracy with less than 1% training samples.

4.4. Result on the Flevoland-14 Dataset

In the experiment on the Flevoland-14 dataset, Table 7 provides the comparisons of OA, AA, Kappa, and the accuracy of each terrain. As shown in Table 7, the classification results of the HA-EDNet method are superior to those of the other comparison methods. The OA value of the proposed method is about 1.62%, 1.23%, 1.27%, 0.87%, 0.69%, and 0.53% higher than other methods. The proposed method achieves the highest accuracies in ten areas and achieved 100% accuracies in six areas. Figure 9 presents the classification maps for each approach. Figure 9c–f gives the classification results of CNN, 3D-CNN, MS-CNN, and CV-CNN, respectively. It is clear that the discrete misclassifications of patch-wise methods exist. Figure 9g,f show the classification results of FCN and U-Net. It can be seen that patch-free methods have better uniformity than patch-wise methods. However, they still have some misclassification in border regions. The classification map for the proposed method is shown in Figure 9i. We can notice that the proposed HA-attention has the best uniformity and performs better in border regions. The primary reason is the extraction capacity of multi-scale features and contextual information by the SK module and self-attention. In conclusion, the classification results can demonstrate the efficacy of the HA-attention model.

4.5. Result on the Oberpfaffenhofen Dataset

Table 8 shows the classification results of each model for three terrain classes in the Oberpfaffenhofen dataset. It can be seen from Table 8 that the performance of the proposed approach is superior to other comparison methods. The OA value of the proposed approach is 14.4%, 12.92%, 12.36%, 7.13%, 4.44%, and 1.79% higher than other comparison methods. It is obvious that patch-wise methods have poor performance. The Kappa coefficients of the patch-wise methods are no more than 0.83. It indicates that these methods do not have good consistency. The patch-free methods perform better in the dataset. The OA values of FCN and U-Net are over 90%. Moreover, the consistency is enhanced due to the spatial information. The proposed method makes further improvements compared to FCN and v. The HA-attention method presents an appealing classification performance, attaining a 96.57% OA value, a 96.59% AA value, and a Kappa coefficient of 0.9418. The experimental results demonstrate that the overall performance of the proposed approach is better than other comparison approaches. Figure 10 depicts classification maps of each method. It can be seen that CNN, 3D-CNN, MS-CNN, and CV-CNN have poor accuracies and present more misclassifications. The FCN and U-Net classifications have been improved, although border areas still include several misclassifications. Compared with the other approaches, the visual performance of the proposed approaches on each land cover area shows a better consistency. In addition, the border regions of the proposed approach on the classification map are much more uniform than those of existing comparison approaches.

4.6. Result on the San Francisco Dataset

In order to evaluate the effectivity of the HA-EDNet approach, we also performed experiments on the RADARSAT-2 San Francisco dataset. The dataset has 1300 × 1300 pixels, and the ground truth comprises five classes. The classification results of various approaches on the San Francisco dataset are shown in Table 9. From Table 9, the results show that the proposed approach obtains the optimal classification performance. The OA of the proposed network is 7.21%, 6.49%, 6.59%, 6.4%, 3.5%, and 2.75% higher than CNN, 3D-CNN, MS-CNN, CV-CNN, FCN, and U-Net, respectively. The proposed approach reaches a classification of 98.85% OA, 98.34% AA, and a Kappa coefficient of 0.9827. The classification maps of each model are shown in Figure 11. It is obvious that the classification results of CNN, 3D-CNN, MS-CNN, and CV-CNN models are poor, with more noise and speckles. The classification results of the FCN and U-Net model have less noise, which improves the classification performance of these models. The proposed model achieves significant performance gains with a combination of the hybrid attention blocks. The results demonstrate the effectiveness of the HA-EDNet approach.

5. Analysis and Discussion

5.1. Ablation Study

In this part, we evaluate how the classification performance is affected by each module of the proposed method. The effectiveness of different modules are presented in Table 10. From Table 10, the results indicate that both selective kernel attention and self-attention can significantly improve classification performance. It can be seen that the selective kernel attention increases the OA values by 2.39%, 0.49%, 0.2%, and 1.02% for four datasets, owing to its capability of attention-based multi-scale feature extraction. The attention-based multi-scale features can highlight and combine different scale information in the network. The self-attention module increases the OA values by 2.72%, 0.2%, 0.65%, and 0.86%, respectively. It is useful for exploring long-range dependency over local spatial features. Combining the SK module and self-attention module for all four datasets can achieve the best OA values.

5.2. Effect of Training Samples

The experimental results have demonstrated that the HA-EDNet network achieves excellent performance for PolSAR image classification, especially in cases of fewer training samples. In this part, we would like to further explore the scenarios of extremely rare training samples. The results of OA, AA, and Kappa with respect to a changed number of training samples are shown in Figure 12. For four datasets, training samples of per class are varied from 10 to 190 with an interval of 30. As expected, the classification accuracies increase with the training number increasing. It is obvious that the proposed approach performs well (up to 80%) when it only has 10 training samples of per class. Therefore, the proposed approach is suitable in the environment when training samples are rare.

5.3. Effect of Different Kernels

In the proposed method, the main parameter that affects the model performance is the kernel number of SK modules. The kernel size determines the range of the receptive field that is used to extract features. We verify the model performance when the kernel groups are (1, 3), (3, 5), (5, 7), (1, 3, 5), (3, 5, 7), and (1, 3, 5, 7). In Table 11 and Figure 13, the classification results show that using 3 × 3, 5 × 5, and 7 × 7 kernels can achieve better classification performance. When the kernel group is (3, 5), the model obtains the best classification accuracies. It demonstrates that 3 × 3 and 5 × 5 kernel groups are beneficial for capturing multi-scale features and significantly improve the classification performance.

6. Conclusions

A novel encoder–decoder method is proposed to extract multi-scale features and long-range dependency based on a hybrid attention mechanism for PolSAR classification in this paper. Inspired by the way humans mimic cognitive attention, the network devotes more focus to the small but important parts of the data, known as the attention mechanism. This mechanism is incorporated into our model. In this method, the patch-free framework, SK module, and self-attention module are utilized. First, an encoder–decoder network is built for patch-free classification, allowing an entire PolSAR image as input and not requiring dividing the image into overlapping patches. Then, the SK module is embedded into the EDNet, which can capture multi-scale features by automatically adjusting the kernel size. Finally, self-attention is employed to extract long-range dependency, which can improve classification performance. In the experiments, four PolSAR datasets are employed to test the effectiveness of the HA-EDNet architecture. The experimental results show that the proposed approach has effective and superior performance compared with some state-of-the-art approaches.
Other attention mechanisms will be introduced into the model for better feature extraction performance in future work. Moreover, a more effective patch-free model will also be investigated in our future works.

Author Contributions

Conceptualization, Z.F. and G.Z.; methodology, Z.F., Q.D. and B.X.; software, Z.F., Q.D. and B.X.; validation, P.W.; formal analysis, Z.F., G.Z. and Q.D.; investigation, Z.F. and G.Z.; resources, Z.F. and P.W.; data curation, Z.F., G.Z. and Q.D.; writing—original draft preparation, Z.F., G.Z., Q.D. and B.X.; writing—review and editing, Z.F., P.W. and G.Z.; visualization, Z.F., G.Z. and Q.D.; supervision, Z.F., G.Z. and Q.D.; project administration, Z.F., G.Z. and Q.D.; funding acquisition, G.Z. and P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This article was supported in part by the National Key Research and Development Program of China under Grant 2019YFB2102005; the National Natural Science Foundation of China under Grant 62271255; the Postgraduate Research and Practice Innovation Program of Jiangsu Province under Grant KYCX21_0216 and KYCX22_0363; the Aeronautical Science Foundation of China under Grant ASFC-201920007002; the Science and Technology Program of Shenzhen, China, under Grant JCYJ20210324134807019; the Open Research Program of Shanghai Key Lab of Intelligent Information Processing under Grant IIPL201908; and the Fujian Provincial Key Lab of Coastal Basin Environment (Grant No. S1-KF2103); Key Laboratory of Southeast Coast Marine Information Intelligent Perception and Application, MNR, (Grant No. 22101); Program of Remote Sensing Intelligent Monitoring and Emergency Services for Regional Security Elements.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duan, D.; Wang, Y. Reflection of and vision for the decomposition algorithm development and application in earth observation studies using PolSAR technique and data. Remote Sens. Environ. 2021, 261, 112498. [Google Scholar] [CrossRef]
  2. Zhai, W.; Zhang, J.; Xiao, X.; Wang, J.; Zhang, H.; Yin, X.; Wu, Z. Damaged building extraction from post-earthquake PolSAR data based on the Fourier transform. Remote Sens. Lett. 2021, 12, 594–603. [Google Scholar] [CrossRef]
  3. Parrella, G.; Hajnsek, I.; Papathanassiou, K. Model-Based Interpretation of PolSAR Data for the Characterization of Glacier Zones in Greenland. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 11593–11607. [Google Scholar] [CrossRef]
  4. Mahdavi, S.; Salehi, B.; Huang, W.; Amani, M.; Brisco, B. A PolSAR Change Detection Index Based on Neighborhood Information for Flood Mapping. Remote Sens. 2019, 11, 1854. [Google Scholar] [CrossRef] [Green Version]
  5. Fan, J.; Zhao, J.; An, W.; Hu, Y. Marine Floating Raft Aquaculture Detection of GF-3 PolSAR Images Based on Collective Multikernel Fuzzy Clustering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2741–2754. [Google Scholar] [CrossRef]
  6. Zhai, W.; Shen, H.; Huang, C.; Pei, W. Fusion of Polarimetric and Texture Information for Urban Building Extraction from Fully Polarimetric SAR Imagery. Remote Sens. Lett. 2016, 7, 31–40. [Google Scholar] [CrossRef]
  7. Han, P.; Han, B.; Shi, Q.; Song, T.; Lu, X.; Zhang, Z. Aircraft Detection Based on Eigen Decomposition and Scattering Similarity for PolSAR Image. In Proceedings of the 2018 IEEE/AIAA 37th Digital Avionics Systems Conference (DASC), London, UK, 23–27 September 2018; pp. 1–6. [Google Scholar] [CrossRef]
  8. Wang, H.; Xing, C.; Yin, J.; Yang, J. Land Cover Classification for Polarimetric SAR Images Based on Vision Transformer. Remote Sens. 2022, 14, 4656. [Google Scholar] [CrossRef]
  9. Cloude, S.R.; Pottier, E. A Review of Target Decomposition Theorems in Radar Polarimetry. IEEE Trans. Geosci. Remote Sens. 1996, 34, 498–518. [Google Scholar] [CrossRef]
  10. Chiang, C.Y.; Chen, K.S.; Chu, C.Y.; Chang, Y.L.; Fan, K.C. Color Enhancement for Four-Component Decomposed Polarimetric SAR Image Based on a CIE-Lab Encoding. Remote Sens. 2018, 10, 545. [Google Scholar] [CrossRef]
  11. Krogager, E. New Decomposition of the Radar Target Scattering Matrix. Electron. Lett. 1990, 18, 1525–1527. [Google Scholar] [CrossRef]
  12. Lee, J.S.; Schuler, D.L.; Ainsworth, T.L. Polarimetric SAR Data Compensation for Terrain Azimuth Slope Variation. IEEE Trans. Geosci. Remote Sens. 2000, 38, 2153–2163. [Google Scholar] [CrossRef]
  13. Cameron, W.L.; Rais, H. Conservative Polarimetric Scatterers and Their Role in Incorrect Extensions of the Cameron Decomposition. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3506–3516. [Google Scholar] [CrossRef]
  14. Ballester-Berman, J.D.; Lopez-Sanchez, J.M. Applying the Freeman–Durden Decomposition Concept to Polarimetric SAR Interferometry. IEEE Trans. Geosci. Remote Sens. 2009, 48, 466–479. [Google Scholar] [CrossRef]
  15. Guo, S.; Tian, Y.; Li, Y.; Chen, S.; Hong, W. Unsupervised Classification Based on H/alpha Decomposition and Wishart Classifier for Compact Polarimetric SAR. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1614–1617. [Google Scholar] [CrossRef]
  16. Erith, M.; Alfonso, Z.; Erik, L. A Multi-Sensor Approach to Separate Palm Oil Plantations from Forest Cover Using NDFI and a Modified Pauli Decomposition Technique. In Proceedings of the IGARSS 2020-2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 4481–4484. [Google Scholar] [CrossRef]
  17. Lee, J.S.; Grunes, M.R.; Ainsworth, T.L.; Du, L.J.; Schuler, D.L.; Cloude, S.R. Unsupervised Classification Using Polarimetric Decomposition and the Complex Wishart Classifier. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2249–2258. [Google Scholar] [CrossRef]
  18. Liu, F.; Jiao, L.; Hou, B.; Yang, S. POL-SAR Image Classification Based on Wishart DBN and Local Spatial Information. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3292–3308. [Google Scholar] [CrossRef]
  19. Jiao, L.; Liu, F. Wishart Deep Stacking Network for Fast POLSAR Image Classification. IEEE Trans. Image Process. 2016, 25, 3273–3286. [Google Scholar] [CrossRef]
  20. Xie, W.; Jiao, L.; Hou, B.; Ma, W.; Zhao, J.; Zhang, S.; Liu, F. POLSAR Image Classification via Wishart-AE Model or Wishart-CAE Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3604–3615. [Google Scholar] [CrossRef]
  21. Okwuashi, O.; Ndehedehe, C.E.; Olayinka, D.N.; Eyoh, A.; Attai, H. Deep Support Vector Machine for PolSAR Image Classification. Int. J. Remote Sens. 2021, 42, 6498–6536. [Google Scholar] [CrossRef]
  22. Yin, Q.; Cheng, J.; Zhang, F.; Zhou, Y.; Shao, L.; Hong, W. Interpretable POLSAR Image Classification Based on Adaptive-Dimension Feature Space Decision Tree. IEEE Access 2020, 8, 173826–173837. [Google Scholar] [CrossRef]
  23. Richardson, A.; Goodenough, D.G.; Chen, H.; Moa, B.; Hobart, G.; Myrvold, W. Unsupervised Nonparametric Classification of Polarimetric SAR Data Using the K-nearest Neighbor Graph. In Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium, Honolulu, HI, USA, 25–30 July 2010; pp. 1867–1870. [Google Scholar] [CrossRef]
  24. Otter, D.W.; Medina, J.R.; Kalita, J.K. A Survey of the Usages of Deep Learning for Natural Language Processing. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 604–624. [Google Scholar] [CrossRef] [Green Version]
  25. Ouhami, M.; Hafiane, A.; Es-Saady, Y.; El Hajji, M.; Canals, R. Computer Vision, IoT and Data Fusion for Crop Disease Detection Using Machine Learning: A Survey and Ongoing Research. Remote Sens. 2021, 13, 2486. [Google Scholar] [CrossRef]
  26. Wu, W.; Li, H.; Li, X.; Guo, H.; Zhang, L. PolSAR Image Semantic Segmentation Based on Deep Transfer Learning—Realizing Smooth Classification With Small Training Sets. IEEE Geosci. Remote Sens. Lett. 2019, 16, 977–981. [Google Scholar] [CrossRef]
  27. Nie, W.; Huang, K.; Yang, J.; Li, P. A Deep Reinforcement Learning-Based Framework for PolSAR Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  28. Zhang, L.; Ma, W.; Zhang, D. Stacked Sparse Autoencoder in PolSAR Data Classification Using Local Spatial Information. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1359–1363. [Google Scholar] [CrossRef]
  29. Zhou, Y.; Wang, H.; Xu, F.; Jin, Y.Q. Polarimetric SAR Image Classification Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1935–1939. [Google Scholar] [CrossRef]
  30. Wang, L.; Xu, X.; Dong, H.; Gui, R.; Yang, R.; Pu, F. Exploring Convolutional LSTM for PolSAR Image Classification. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8452–8455. [Google Scholar] [CrossRef]
  31. Zheng, Z.; Zhong, Y.; Ma, A.; Zhang, L. FPGA: Fast Patch-Free Global Learning Framework for Fully End-to-End Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5612–5626. [Google Scholar] [CrossRef]
  32. Chen, Z.; Tong, L.; Qian, B.; Yu, J.; Xiao, C. Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification. Remote Sens. 2021, 13, 3316. [Google Scholar] [CrossRef]
  33. Dong, H.; Zhang, L.; Lu, D.; Zou, B. Attention-Based Polarimetric Feature Selection Convolutional Network for PolSAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  34. Hua, W.; Wang, X.; Zhang, C.; Jin, X. Attention-Based Multiscale Sequential Network for PolSAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  35. Yang, Z.; Zhang, Q.; Chen, W.; Chen, C. PolSAR Image Classification Based on Resblock Combined with Attention Model. In Proceedings of the 2021 IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; pp. 340–344. [Google Scholar] [CrossRef]
  36. Ren, S.; Zhou, F. Polsar Image Classification with Complex-Valued Residual Attention Enhanced U-NET. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 3045–3048. [Google Scholar] [CrossRef]
  37. Yang, B.; Wang, L.; Wong, D.F.; Shi, S.; Tu, Z. Context-Aware Self-Attention Networks for Natural Language Processing. Neurocomputing 2021, 458, 157–169. [Google Scholar] [CrossRef]
  38. Gao, W.; Zhang, L.; Huang, W.; Min, F.; He, J.; Song, A. Deep Neural Networks for Sensor-Based Human Activity Recognition Using Selective Kernel Convolution. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar] [CrossRef]
  39. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  40. Xu, R.; Tao, Y.; Lu, Z.; Zhong, Y. Attention-Mechanism-Containing Neural Networks for High-Resolution Remote Sensing Image Classification. Remote Sens. 2018, 10, 1602. [Google Scholar] [CrossRef] [Green Version]
  41. Ghaffarian, S.; Valente, J.; van der Voort, M.; Tekinerdogan, B. Effect of Attention Mechanism in Deep Learning-Based Remote Sensing Image Processing: A Systematic Literature Review. Remote Sens. 2021, 13, 2965. [Google Scholar] [CrossRef]
  42. Shaw, P.; Uszkoreit, J.; Vaswani, A. Self-Attention with Relative Position Representations. arXiv 2018, arXiv:1803.02155. [Google Scholar] [CrossRef]
  43. Zhao, H.; Jia, J.; Koltun, V. Exploring Self-Attention for Image Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10076–10085. [Google Scholar] [CrossRef]
  44. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  45. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  46. Pu, W.; Bao, Y. RPCA-AENet: Clutter Suppression and Simultaneous Stationary Scene and Moving Targets Imaging in the Presence of Motion Errors. IEEE Trans. Neural Networks Learn. Syst. 2022, 1–14. [Google Scholar] [CrossRef]
  47. Zhang, L.; Chen, Z.; Zou, B.; Gao, Y. Polarimetric SAR Terrain Classification Using 3D Convolutional Neural Network. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4551–4554. [Google Scholar] [CrossRef]
  48. Gong, Z.; Zhong, P.; Yu, Y.; Hu, W.; Li, S. A CNN with Multiscale Convolution and Diversified Metric for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3599–3618. [Google Scholar] [CrossRef]
  49. Zhang, Z.; Wang, H.; Xu, F.; Jin, Y.Q. Complex-valued Convolutional Neural Network and Its Application in Polarimetric SAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
  50. Liu, X.; Jiao, L.; Tang, X.; Sun, Q.; Zhang, D. Polarimetric Convolutional Network for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 3040–3054. [Google Scholar] [CrossRef] [Green Version]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  52. Yang, C.; Hou, B.; Chanussot, J.; Hu, Y.; Ren, B.; Wang, S.; Jiao, L. N-Cluster Loss and Hard Sample Generative Deep Metric Learning for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  53. Cui, Y.; Liu, F.; Jiao, L.; Guo, Y.; Liang, X.; Li, L.; Yang, S.; Qian, X. Polarimetric Multipath Convolutional Neural Network for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  54. Cheng, J.; Zhang, F.; Xiang, D.; Yin, Q.; Zhou, Y.; Wang, W. PolSAR Image Land Cover Classification Based on Hierarchical Capsule Network. Remote Sens. 2021, 13, 3132. [Google Scholar] [CrossRef]
Figure 1. Illustration of the SK module.
Figure 1. Illustration of the SK module.
Remotesensing 15 00526 g001
Figure 2. Illustration of the self-attention module.
Figure 2. Illustration of the self-attention module.
Remotesensing 15 00526 g002
Figure 3. Illustration of the architecture of the HA-EDNet.
Figure 3. Illustration of the architecture of the HA-EDNet.
Remotesensing 15 00526 g003
Figure 4. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Flevoland-15 dataset.
Figure 4. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Flevoland-15 dataset.
Remotesensing 15 00526 g004
Figure 5. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Flevoland-14 dataset.
Figure 5. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Flevoland-14 dataset.
Remotesensing 15 00526 g005
Figure 6. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Oberpfaffenhofen dataset.
Figure 6. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for Oberpfaffenhofen dataset.
Remotesensing 15 00526 g006
Figure 7. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for San Francisco dataset.
Figure 7. The pseudocolor image (a), the ground truth (b), and legends of classes (c) for San Francisco dataset.
Remotesensing 15 00526 g007
Figure 8. Results of the Flevoland-15 dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Figure 8. Results of the Flevoland-15 dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Remotesensing 15 00526 g008
Figure 9. Result of the Flevoland-14 dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Figure 9. Result of the Flevoland-14 dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Remotesensing 15 00526 g009
Figure 10. Results of the Oberpfaffenhofen dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Figure 10. Results of the Oberpfaffenhofen dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Remotesensing 15 00526 g010
Figure 11. Results of the San Francisco dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Figure 11. Results of the San Francisco dataset: (a) pseudocolor image; (b) ground truth; (c) CNN; (d) 3D-CNN; (e) MS-CNN; (f) CV-CNN; (g) FCN; (h) U-Net; (i) proposed method.
Remotesensing 15 00526 g011
Figure 12. Classification accuracy changes with the number of training samples: (a) Flevoland-15 dataset; (b) Flevoland-14 dataset; (c) Oberpfaffenhofen dataset; (d) San Francisco dataset.
Figure 12. Classification accuracy changes with the number of training samples: (a) Flevoland-15 dataset; (b) Flevoland-14 dataset; (c) Oberpfaffenhofen dataset; (d) San Francisco dataset.
Remotesensing 15 00526 g012
Figure 13. Classification accuracy changes with different kernel groups: (a) Flevoland-15 dataset; (b) Flevoland-14 dataset; (c) Oberpfaffenhofen dataset; (d) San Francisco dataset.
Figure 13. Classification accuracy changes with different kernel groups: (a) Flevoland-15 dataset; (b) Flevoland-14 dataset; (c) Oberpfaffenhofen dataset; (d) San Francisco dataset.
Remotesensing 15 00526 g013
Table 1. Trainingand testing samples of the Flevoland-15 dataset.
Table 1. Trainingand testing samples of the Flevoland-15 dataset.
Class NumberRegionTraining NumberTesting NumberTotal Number
1Stem beam10060036103
2Peas10090119111
3Forest10014,84414,944
4Lucerne10093779477
5Wheat 210017,18317,283
6Beet100995010,050
7Potato10015,19215,292
8Bare soil10029783078
9Grass10061696269
10Rapeseed10012,59012,690
11Barley10070597159
12Wheat 110010,49110,591
13Wheat 310021,20021,300
14Water10013,37613,476
15Building100376476
Total 1500155,799157,299
Table 2. Training and testing samples of the Flevoland-14 dataset.
Table 2. Training and testing samples of the Flevoland-14 dataset.
Class NumberRegionTraining NumberTesting NumberTotal Number
1Potato10021,51321,613
2Fruit10042524352
3Oats10012941394
4Beet10010,71710,817
5Barley10024,44324,543
6Onions10020302130
7Wheat10026,17726,277
8Beans1009821082
9Peas10020602160
10Maize10011901290
11Flax10042014301
12Rapeseed10028,13528,235
13Grass10041044204
14Lucerne10028522952
Total 1400133,950135,350
Table 3. Training and testing samples of the Oberpfaffenhofen dataset.
Table 3. Training and testing samples of the Oberpfaffenhofen dataset.
Class NumberRegionTraining NumberTesting NumberTotal Number
1Built-up100327,951328,051
2Wood land100246,573246,673
3Open areas100736,794736,894
Total 3001,311,3181,311,618
Table 4. Training and testing samples of the San Francisco dataset.
Table 4. Training and testing samples of the San Francisco dataset.
Class NumberRegionTraining NumberTesting NumberTotal Number
1Water100689,707689,807
2Vegetation100198,502198,602
3High-Density Urban100112,161112,261
4Low-Density Urban100275,576275,674
5Developed10065,47365,573
Total 5001,497,2181,497,718
Table 5. Classification results of different methods in the Flevoland-15 dataset.
Table 5. Classification results of different methods in the Flevoland-15 dataset.
RegionCNN3D-CNNMS-CNNCV-CNNFCNU-NetProposed
Stem beam68.0269.2174.0698.4895.4597.9099.35
Peas96.8196.6696.4497.5099.7399.8999.89
Forest96.2495.5199.3292.0999.8398.9099.73
Lucerne97.2997.1196.0996.6395.8192.9697.61
Wheat 287.6687.6083.9993.3999.0198.1099.58
Beet98.7793.7397.7896.6299.5798.1499.78
Potato91.6295.1696.2995.3599.6798.1899.26
Bare soil99.5710099.8099.83100100100
Grass86.8286.1989.4890.1499.8299.4799.92
Rapeseed88.8289.7293.7078.0566.5764.8398.19
Barley99.1797.9399.6895.0199.8699.97100
Wheat 192.5792.5494.2680.8196.8195.8399.98
Wheat 398.4798.4599.0895.5299.9399.8299.04
Water71.5287.0678.2498.8592.8084.29100
Building56.6998.9471.5598.6710097.61100
OA91.0592.4892.8293.0595.8194.3599.39
AA88.6792.3991.3293.8096.3295.0699.49
Kappa0.90240.91800.92170.92410.95430.93840.9933
Table 6. Classification of state-of-the-art methods on the Flevoland-15 dataset.
Table 6. Classification of state-of-the-art methods on the Flevoland-15 dataset.
MethodTraining RatioClass NumberOA
N-cluster GAN [52]5%1599.10%
PolMPCNN [53]1%1599.14%
HCapsNet [54]1%1599.04%
AMSE-LSTM [34]1%1597.09%
Proposed method<1%1599.39%
Table 7. Classification results of different methods in the Flevoland-14 dataset.
Table 7. Classification results of different methods in the Flevoland-14 dataset.
RegionCNN3D-CNNMS-CNNCV-CNNFCNU-NetProposed
Potato99.5499.2399.3999.2999.8399.8798.56
Fruit99.6998.8998.6899.98100100100
Oats99.0799.2398.3099.31100100100
Beet91.4494.6795.9895.2490.3391.2899.73
Barley98.0298.1496.9698.7299.7499.7999.28
Onions92.2294.7891.2393.1598.2897.5499.80
Wheat98.3398.5099.1998.9499.7899.7999.81
Beans98.6898.4798.5799.0895.5197.6698.98
Peas99.3299.8199.9510010010099.95
Maize97.3196.0596.6499.1698.9999.92100
Flax99.5799.3897.2499.91100100100
Rapeseed98.4399.0399.0799.0299.2399.2799.48
Grass96.1796.9895.5797.4796.6697.95100
Lucerne96.8495.1698.8896.8499.7699.86100
OA97.8498.2398.1998.5998.7798.9399.46
AA97.4797.7497.5598.2998.4498.7899.69
Kappa0.97460.97910.97870.98330.98550.98730.9936
Table 8. Classification results of different methods in the Oberpfaffenhofen dataset.
Table 8. Classification results of different methods in the Oberpfaffenhofen dataset.
RegionCNN3D-CNNMS-CNNCV-CNNFCNU-NetProposed
Built-up areas86.2385.4986.5679.1290.8992.8793.99
Wood land45.0752.3256.3094.3197.6597.9598.82
Open areas92.7793.3192.5092.4092.6194.5796.97
OA82.1783.6584.2189.4493.1394.7896.57
AA74.6977.0478.4588.6193.7295.1396.59
Kappa0.69370.71910.73000.82250.88470.91200.9418
Table 9. Classification results of different methods in the San Francisco dataset.
Table 9. Classification results of different methods in the San Francisco dataset.
RegionCNN3D-CNNMS-CNNCV-CNNFCNU-NetProposed
Water99.5399.8799.8299.9799.9999.9799.93
Vegetation88.3288.8189.1386.6995.3696.6395.53
High-Density Urban79.1679.8085.2578.5994.7896.9499.21
Low-Density Urban81.1282.5579.7185.1383.8085.8298.47
Developed84.2286.8286.8985.3895.9195.3998.55
OA91.6492.3692.2692.4595.3596.1098.85
AA86.4787.5788.1687.1593.9794.9598.34
Kappa0.87440.88510.88380.88620.93020.94130.9827
Table 10. Performance contribution of each module in HA-EDNet.
Table 10. Performance contribution of each module in HA-EDNet.
MethodBased Method+SK+SA+SA+SK
Patch-free model
Selective kernel attention (SK)  
Self-attention (SA)  
Flevoland-1596.2798.6698.8999.39
Flevoland-1498.7699.2598.9699.46
Oberpfaffenhofen95.6195.8196.2696.57
San Francisco97.4498.4698.3098.85
Table 11. Performance contribution of different kernels in the SK module.
Table 11. Performance contribution of different kernels in the SK module.
Kernel
Number
(1, 3)(3, 5)(5, 7)(1, 3, 5)(3, 5, 7)(1, 3, 5, 7)
Flevoland-1598.2499.3999.3198.8999.2699.22
Flevoland-1498.0199.4699.3999.1899.3699.33
Oberpfaffenhofen96.0296.5796.2296.1696.0895.89
San Francisco98.4898.8598.5198.7398.8198.69
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fang, Z.; Zhang, G.; Dai, Q.; Xue, B.; Wang, P. Hybrid Attention-Based Encoder–Decoder Fully Convolutional Network for PolSAR Image Classification. Remote Sens. 2023, 15, 526. https://doi.org/10.3390/rs15020526

AMA Style

Fang Z, Zhang G, Dai Q, Xue B, Wang P. Hybrid Attention-Based Encoder–Decoder Fully Convolutional Network for PolSAR Image Classification. Remote Sensing. 2023; 15(2):526. https://doi.org/10.3390/rs15020526

Chicago/Turabian Style

Fang, Zheng, Gong Zhang, Qijun Dai, Biao Xue, and Peng Wang. 2023. "Hybrid Attention-Based Encoder–Decoder Fully Convolutional Network for PolSAR Image Classification" Remote Sensing 15, no. 2: 526. https://doi.org/10.3390/rs15020526

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop