Next Article in Journal
Testing of a Polarimetric Current Sensor in the Frequency Domain
Previous Article in Journal
Infrequent Pattern Detection for Reliable Network Traffic Analysis Using Robust Evolutionary Computation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple Classifiers Based Semi-Supervised Polarimetric SAR Image Classification Method

1
School of Resources and Environmental Engineering/Anhui Province Key Laboratory of Wetland Ecosystem Protection and Restoration, Anhui University, Hefei 230601, China
2
Department of Resource and Environmental Sciences, Wuhan University, Wuhan 430072, China
3
Information Materials and Intelligent Sensing Laboratory of Anhui Province, Anhui University, Hefei 230601, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(9), 3006; https://doi.org/10.3390/s21093006
Submission received: 20 March 2021 / Revised: 14 April 2021 / Accepted: 16 April 2021 / Published: 25 April 2021
(This article belongs to the Section Radar Sensors)

Abstract

:
Polarimetric synthetic aperture radar (PolSAR) image classification has played an important role in PolSAR data application. Deep learning has achieved great success in PolSAR image classification over the past years. However, when the labeled training dataset is insufficient, the classification results are usually unsatisfactory. Furthermore, the deep learning approach is based on hierarchical features, which is an approach that cannot take full advantage of the scattering characteristics in PolSAR data. Hence, it is worthwhile to make full use of scattering characteristics to obtain a high classification accuracy based on limited labeled samples. In this paper, we propose a novel semi-supervised classification method for PolSAR images, which combines the deep learning technique with the traditional scattering trait-based classifiers. Firstly, based on only a small number of training samples, the classification results of the Wishart classifier, support vector machine (SVM) classifier, and a complex-valued convolutional neural network (CV-CNN) are used to conduct majority voting, thus generating a strong dataset and a weak dataset. The strong training set are then used as pseudo-labels to reclassify the weak dataset by CV-CNN. The final classification results are obtained by combining the strong training set and the reclassification results. Experiments on two real PolSAR images on agricultural and forest areas indicate that, in most cases, significant improvements can be achieved with the proposed method, compared to the base classifiers, and the improvement is approximately 3–5%. When the number of labeled samples was small, the superiority of the proposed method is even more apparent. The improvement for built-up areas or infrastructure objects is not as significant as forests.

1. Introduction

Synthetic aperture radar (SAR) systems are active microwave imaging systems, which can obtain high-resolution images in both daytime and nighttime, and under all weather conditions [1]. As an advanced form of SAR, polarimetric SAR (PolSAR) systems can work in different polarization modes to characterize the observed land-cover types, and thus have a strong ability to obtain scattering information, resulting in their widespread use. As a result, PolSAR image classification has played an important role in many fields [2,3,4,5,6,7,8].
In the last three decades, a large number of algorithms have been developed for PolSAR image classification [9,10,11,12,13,14,15,16,17,18,19]. The traditional classifiers can be divided into three major categories. The first type is based on the statistical characteristics of PolSAR data. For example, Kong et al. [9] derived a maximum likelihood (ML) classifier for single-look complex (SLC) PolSAR imagery, which is based on a complex Gaussian distribution; for multilook PolSAR data, Lee et al. [10] applied the principle of ML and derived a Wishart distance measure; Liu et al. [11] proposed a superpixel-based classification method with an adaptive number of classes, which can take the spatial relations between pixels into consideration and make full use of the inherent statistical characteristics of PolSAR data; and Xie et al. [12] proposed a clustering-Wishart-auto-encoder classification model, which simultaneously considers the compactness and statistical distribution of the data by combining the k-means clustering algorithm with the objective function of a prior model. The second approach is focused on the inherent characteristics of the polarimetric scattering mechanisms and can take full advantage of the prior information about class type recognition. For example, Freeman et al. [13] proposed a hierarchical classification approach based on the scattering characteristics; based on the maximum entropy method, Kouskoulas et al. [14] proposed the Bayesian hierarchical classifier to classify low vegetation; Cheng et al. [15] proposed an unsupervised classifier to classify PolSAR pixels into eight combinations using three metrics extracted from the observed coherency matrix; and based on Freeman decomposition and H/Alpha decomposition, Zhao et al. [16] proposed a framework for the iterative classification of PolSAR data, which classifies the data into nine initial classes and obtains the final classification result through the use of a Wishart classifier. The third strategy uses a combination of polarimetric scattering characteristics and statistical properties. For example, Lee et al. [17] proposed an approach which uses both H/α target decomposition [18] and the ML classifier based on a complex Wishart distribution; and Chang et al. [19] proposed a PolSAR image classification method based on the degree of polarization and co-polarized phase-difference statistics, which has the ability to classify PolSAR data into four major classes. Hao et al. [20] proposed a classification framework for PolSAR image, which combined XGBoost, superpixels generation, and majority voting.
In the last few years, deep learning-based classifiers have become more and more widely used in the field of PolSAR image classification [21,22,23,24,25,26,27,28,29]. As a subset of machine learning, deep learning can process complex data efficiently, and has a strong feature extraction capability. For example, Hou et al. [21] applied a multilayer autoencoder and super-pixels for PolSAR image classification, which can make good use of the scattering characteristics and spatial relationships between pixels; and Liu et al. [22] utilized a task-oriented generative adversarial network (GAN) for PolSAR image classification and clustering. Li et al. [23] used the Deeplabv3 to segment fruit and twigs from the background based on the color and depth information in a RGB picture. Among the deep discriminative networks, the real-valued convolutional neural network (RV-CNN) model is one of the most popular models [25]. On account of the special multidimensional convolution operations, the RV-CNN model has significant advantages in image data classification [26]. However, the RV-CNN model only uses the amplitude of the PolSAR imagery, while neglecting the phase information. For PolSAR data represented by a covariance or coherency matrix, the phase of off-diagonal elements plays an important role in classifying different types of scatterers. Hirose et al. [27] first processed complex-valued data by a complex-valued neural network, in which weight and neural activation functions are all complex-valued. Soon afterwards, a novel complex-valued CNN (CV-CNN) classification method was proposed by Zhang et al. [28], which not only takes the amplitude and phase information of the PolSAR imagery as input, but also propagates the phase information through all the processes of the network.
Deep learning has achieved enormous success in PolSAR data interpretation [30], and generally obtains better performances than the traditional classifiers. However, the classification result of a deep learning network is dependent, to a large extent, on a large number of labeled samples [31,32]. Limited training samples can lead to an overfitting phenomenon, which refers to the fact that the model only learns the characteristics of the training samples in the training process, and the classification result on the test set is poor because of the low generalization performance of the trained model. The lack of labeled samples can lead to a poor classification result. Therefore, semi-supervised classification has become prevalent in deep learning because it avoids the labor-intensive task of acquiring a large number of labeled samples [33,34], and it can fully exploit the available labeled training samples. In the field of optical image classification, many semi-supervised algorithms have been developed in the past few decades [35,36,37]. However, the semi-supervised algorithms for PolSAR data classification are relatively deficient. In order to solve the lack of labeled datasets and minimize the speckle noise, Li et al. [38] proposed a semi-supervised algorithm based on self-training and superpixel, which used the segmentation and stacked sparse auto-encoder to expand the training set. Sun et al. [39] proposed an impartial semi-supervised learning strategy based on extreme gradient boosting to classify very high-resolution images with imbalanced data. Liu et al. [40] proposed a novel semi-supervised algorithm with neighborhood constraints to reduce the cost of labeled samples, which uses a number of PolSAR features of every pixel, including its neighbors, to construct a spatial group; and Hou et al. [41] proposed a robust semi-supervised probability graphic-based classification framework to solve the problem of the weak quantity and quality of the labeled training samples.
Another weakness of deep learning is that it may have difficulties in discriminating objects with similar textures, but with different scattering patterns [42]. In contrast, the traditional classifiers, such as the Wishart classifier, are based on a complex Wishart distribution, and can employ the scattering traits of PolSAR data. Hence, every method has its own advantages. Fusion strategies based on multiple classifiers have been developed, the main purpose of which is to make full use of the respective advantages of each classifier [43]. Theoretically, the combination of multiple classifiers is generally able to obtain a better classification result [44]. Classification methods can be integrated according to the way in which they are built. Majority voting is one of the most widely used multiple-classifier integration strategies, which involves selecting the predicted class with the most votes [45].
In the proposed method, motivated by the above viewpoints, we combine the deep learning technique with the traditional classifiers, the aim of which is to take advantage of each classifier and improve the overall accuracy (OA) of classification, even in the case of limited training samples. Firstly, the classification results of the Wishart classifier, SVM classifier, and CV-CNN model are used to conduct majority voting. By doing so, the labeled training samples are expanded by dividing the classification result into a strong dataset and a weak dataset. The strong dataset are used as pseudo-labels to reclassify the weak dataset by CV-CNN. In order to make full use of the strong dataset, the weak dataset is reclassified three times using the pseudo-labels derived by the strong dataset, and then followed by a majority voting manner to integrate the three classification results.
The main contributions of this present paper are as follows: A novel sample selected method is maintained by a voting strategy, which is used to expand labeled training samples based on three classifiers classification results; a novel semi-supervised classification method is presented, which can make full use of scattering characteristics and obtain a high classification accuracy based on limited labeled samples.
The rest of this paper is organized as follows. The characteristics of PolSAR data and the traditional classifiers are described in Section 2. In Section 3, the CV-CNN model and the proposed method are introduced. The experimental results obtained with the benchmark PolSAR images are described in Section 4. Finally, our conclusions are drawn in Section 5.

2. Related Works

2.1. PolSAR Data

Fully polarimetric SAR systems measure the complex scattering characteristics of an observed target with quad-polarizations. Each pixel of a PolSAR image in the SLC format can be expressed by a complex scattering matrix:
S = [ S h h S h v S v h S v v ]
where each element in the complex scattering matrix represents the amplitude and phase information. Svh denotes the scattering coefficient of the horizontal transmitting and vertical receiving polarizations, which can be characterized by the phase ϕ v h and the amplitude | S v h | : S v h = | S v h | e j ϕ v h = | S v h | cos ϕ v h + j | S v h | sin ϕ v h , where j is the imaginary unit. The other elements are similarly defined. Under a monostatic case and the reciprocity theorem, Shv and Svh are equivalent. Thus, under Pauli decomposition [46], the complex scattering vector h can be obtained as:
h = 1 2 [ S h h + S v v , S h v + S v v , 2 S h v ] T
where the superscript T is the transpose operation. The coherency matrix of the PolSAR imagery for the multilook case can be expressed as:
T = 1 n i = 1 n h i h i H = [ T 11 T 12 T 13 T 21 T 22 T 23 T 31 T 32 T 33 ]
where n is the number of looks, and the superscript H represents the conjugate transpose.

2.2. Traditional Classifiers

For the traditional classifiers applied in PolSAR image classification, Wishart and SVM are the classical methods. Wishart and SVM are both supervised classifiers which play an important role in the field of PolSAR data interpretation.

2.2.1. Wishart Classification

For multilook PolSAR data represented in coherency matrix T, Lee et al. [10] applied the principle of ML and derived a Wishart distance:
d ( T , V m ) = T r ( V m 1 T ) + ln | V m |
where d(T, Vm) means a distance between the sample pixel covariance matrix T and cluster center Vm. For each target pixel, the Wishart distance between it and each cluster center is calculated, and the pixel is classified to the class with the smallest Wishart distance. The Wishart distance is a measurement that is commonly used to measure the similarity of PolSAR imagery. For example, Jiao et al. proposed a Wishart deep stacking network method to improve the precision of PolSAR image classification, based on the rapid implementation of the Wishart distance [47].

2.2.2. Support Vector Machine Classification

SVM is a classification algorithm which minimizes the structural risk to improve the generalization ability of machine learning. SVM also minimizes the empirical risk and confidence range to obtain good classification results, even in the case of limited labeled samples [48]. Due to the superiority of SVM, a lot of algorithms have been developed based on SVM. For example, Maghsoudi et al. [49] proposed a system consisting of a feature selector based on a non-parametric evaluation function and SVM; SVM was combined with a radial basis kernel function and stochastic distance to assess the robustness of in region-based classification by Negri et al. [50].

2.2.3. Deep Learning Method for PolSAR

Deep Learning has demonstrated great advantages in the PolSAR image processing task. The deep learning method is a series of algorithms based on a complex structure, using nonlinear transformation to abstract the multiple high-level characteristics, whose essence is acquiring high-level characteristics by a multi-layer network. This is based on a large amount of data, so as to eventually improve the accuracy of classification and provide a new idea for high-level characteristics. Auto encoders (AE), deep belief network (DBN), and CNN are the main methods that have been widely used in PolSAR image application. PolSAR image classification based on the deep learning method generally consists of image preprocessing, dataset production, parameter testing, and result analysis.
Deep learning has achieved great success in the field of image classification [51], and the CV-CNN model can be considered as a representative method. As shown in Figure 1, as in the traditional RV-CNN model, the CV-CNN model consists of an input layer, an output layer, convolutional layers, pooling layers, and fully connected layers [28]. However, all elements of the CV-CNN model are complex-valued, including the input layer, convolutional layer, activation function, pooling layer, and output layer. Compared with the RV-CNN model, there are three main advantages of the CV-CNN model [28]: (1) the CV-CNN model not only uses the amplitude of the PolSAR imagery, but it also propagates the phase information through all the process; (2) all the mathematical operations of the whole network are extended under complex analysis theory, and at the same time, both the data and parameters are extended into the complex field; and (3) a complex back-propagation algorithm based on stochastic gradient descent is used for the model training. In the following, the details of the CV-CNN model are presented.
The function of the convolutional layer is to convolve the filters to extract the different features of the previous layer. For the input image, each filter bank will detect specific regional features, which means that every feature map represents a specific feature at different zones of the previous layer. The output of the (l + 1)th convolutional layer can be represented as:
O m ( l + 1 ) = f ( ξ ( V m ( l + 1 ) ) ) + j f ( ψ ( V m ( l + 1 ) ) )
p ( l + 1 ) = k K w p K ( l + 1 ) * O k ( l ) + b p ( l + 1 )
where * represents the convolution operation. ξ and ψ are, respectively, the real and imaginary parts of a complex-valued domain. V m ( l + 1 ) is the mth output feature map of the (l + 1)th layer. w p K ( l + 1 ) denotes the filter banks. O k ( l ) and b p ( l + 1 ) represent the previous layer’s input feature maps and the bias, respectively. f ( ) is the nonlinear function, which represents the sigmoid function.
The pooling layer usually follows the convolutional layer, which can not only simplify the spatial structure, but can also merge similar features of the input feature maps. Among the many pooling functions, the max pooling and average pooling layers are the two most commonly used functions. Hence, pooling layers can be regarded as downsampling layers.
In general, one or more fully connected layers are employed in the CV-CNN model, and can be regarded as special convolutional layers. In addition, each neuron of the fully connected layer is connected to all the neurons in the front layer. The output can be expressed as:
O n ( l + 1 ) = f ( ξ ( V n ( l + 1 ) ) + j f ( ψ ( V n ( l + 1 ) ) )
V p ( l + 1 ) = m = 1 M w p m ( l + 1 ) · O m ( l ) + b p ( l + 1 )
where M denotes the quantity of neurons at the lth fully connected layer.
The final output layer is a classifier represented in a 1 * C complex-valued vector, which indicates the probability of the pixel belonging to the Cth class. After this, all of the parameters in the network are learned in a supervised way by minimizing a loss function E, which can be written as:
E = 1 2 1 N n = 1 N k = 1 K [ ( ξ ( T k [ n ] ) ψ ( O k [ n ] ) ) 2 ( ψ ( T k [ n ] ) ψ ( O k [ n ] ) ) 2 ]
where T [ n ] is the nth input data. The covariance matrix T is a Hermitian matrix. Hence, we only take the upper triangular elements of the matrix as the input data.

3. The Proposed Approach

3.1. The Deep Learning Method

Deep Learning has shown great advantages in PolSAR image processing task. The deep learning method is a series of algorithms based on a complex structure using nonlinear transformation to abstract the multiple high-level characteristics, whose essence is acquiring high-level characteristics by a multi-layer network. This is based on a large amount of data, so as to eventually improve the accuracy of classification and provide a new idea for high-level characteristics. Auto encoders (AE), deep belief network (DBN), and RV-CNN are the main methods that have been widely used in PolSAR image application.
The CV-CNN model, as an advanced form of RV-CNN, has achieved great performances in PolSAR image supervised classification. However, its classification performance is heavily reliant on the number of labeled training samples, and limited training samples can lead to an overfitting phenomenon, resulting in the low generalization performance of the trained model. The CV-CNN model may have difficulties in discriminating objects with similar textures, but with different scattering patterns. On the contrary, the Wishart classifier can employ the scattering traits of PolSAR data, but the spatial information in the image has not been fully utilized. Generally, the single classifier is unable to accurately abstract high-level description of ground objects. Hence, discovering a method that can take the scattering traits and the multiple high-level characteristics deserves extensive research.
In this paper, with the aim of solving the problem of insufficient labeled samples, a novel supervised classification method is proposed. The main idea is to take advantage of each classifier, and expand the number of high-reliability samples, so as to improve the classification accuracy of the CV-CNN model.

3.2. Configuration of the Proposed Method

As illustrated in Figure 2, the flowchart of the proposed method consists of two main parts. Firstly, a small number of labeled samples are randomly selected to classify the PolSAR images by three base classification methods, which are SVM, Wishart, and CV-CNN. A majority voting strategy is then employed to fuse the base classifier results, which is done by expanding the labeled training samples by dividing the classification result into a strong dataset and a weak dataset. In this majority voting process, if the three classifiers are unanimous in the class of a certain pixel, this pixel is categorized into the strong dataset with the voted class label; otherwise, it is categorized into the weak dataset with the uncertain class label.
After acquiring the strong dataset and expanding the number of training samples, more samples could, in theory, be employed to classify the PolSAR data. In general, it is easier and faster to classify the weak dataset only once by the CV-CNN model. In order to make full use of the strong dataset and suppress the interference of misclassified pixels in the initial majority voting process, samples are selected three times from the strong dataset as pseudo-labels to train the CV-CNN classifier three times, followed by normal majority voting, which is based on the predicted class with the most votes. Finally, by combining the strong dataset and the reclassification result of the weak dataset, the final classification results are obtained.

3.3. Preprocessing of PolSAR Data for CV-CNN

In order to speed up the gradient descent and improve the classification accuracy, it is necessary to preprocess the PolSAR data. For each element of the coherency matrix of the input data, the average Tave and the standard deviation Tstd of the training samples are calculated. For the off-diagonal terms, for example T12, the average and standard deviation can be calculated as:
T 12 _ ave = 1 K k = 1 K T 12 ( k )
T 12 _ std = k = 1 K ( T 12 ( k ) T 12 _ ave ) ( T 12 ( k ) T 12 _ ave ¯ ) K
and the final normalized data are obtained as:
T 12 = T 12 T 12 _ ave T 12 _ std
The diagonal terms are similarly defined.

4. Experimental Section

In this section, we describe the experiments on two real PolSAR images to demonstrate the performance of the proposed algorithm (Table 1). The OA and Kappa were adopted to quantitatively evaluate the performance of the different methods. The refined Lee filter [52] was applied to reduce the speckle noise before conducting the experiments. For SVM classification on PolSAR data, selecting proper parameter settings and the kernel parameter was an important step. In this study, we selected all the terms of the coherency matrix as the input polarimetric indicators and chose the Gaussian radial basis function (RBF) as the kernel function, which was implemented in PolSARpro V6.0 software. All of the experiments were implemented in the MATLAB R2018a environment on a 3.2 GHz machine with GTX 1060 GPU and 16-GB memory. The CV-CNN model consists of an input layer, an output layer, two convolutional layers and one pooling layer. The size of the input layer was 12 × 12 × 6 , which means that the channel number was 6, and the local patch was 12 × 12 . Then, the input data was filtered by six convolution filters with the size of 3 × 3 × 6 , producing six feature maps of size 10 × 10 . The size of average pooling layer was 2 × 2 and stride was 1. The size of the second convolutional layer filter was 3 × 3 × 6 × 12 , resulting in 12 feature maps and 3 × 3 size. The fully connected layer consisted of 108 neurons. The output layer contained q neurons and q is the number of classification classes.

4.1. Experiments with the Flevoland Dataset

The first dataset used for the verification of the proposed method was acquired by the NASA/JPL AIRSAR platform in 1989, over an agricultural area in Flevoland in the Netherlands. The size of the image is 750 × 1024 pixels. Figure 3a displays the Pauli RGB image of the data, which was formed by the intensities of the Pauli de-composition. The ground truth and the corresponding legend are shown in Figure 3b. In total, there are 13 classes, including stem beans, peas, forest, lucerne, beet, potatoes, grass, rapeseed, barley, water, and three types of wheat.
In this experiment, we randomly selected only 509 labeled samples. That is to say, only 39 pixels were taken as the training samples for each class. The training sample ratio was only 0.34%, which is quite a low level when compared with other studies. The classification results are shown in Figure 3 and the quantitative assessment results are listed in Table 2. For the first CV-CNN experiment on Flevoland dataset, the hyperparameters were set as follows. The learning rate was 0.6, and the batch size was 10 with 200 epochs. The loss curves are shown in Figure 4. From the loss curves, we can easily find that with the increasing of epoch, the loss decreased quickly. When the epoch is 100, the loss was stable at 0.006.
From Table 2 we can see that the OAs of the Wishart, SVM, and CV-CNN base classifiers, the strong dataset, and the proposed method, were 87.04%, 80.78%, 81.77%, 97.34%, and 90.75%, respectively. The highest OA was achieved by the proposed method, and the classification accuracies for all categories with the proposed method are higher than those with CV-CNN. It can also be seen that the accuracy for every class with the strong dataset was higher than for the base classifiers. The OA for the strong dataset was higher than for the base classifiers, which validates the high reliability of the strong dataset. On one hand, majority voting can, on the whole, improve the accuracy. On the other hand, the OA for the strong dataset was very high, so this dataset can be taken as pseudo-labels to reclassify the weak dataset. As shown in Table 2, the number of training samples of stem beans was only 41, however, the number of samples of stem beans in the strong dataset was 10,934 with a high accuracy of 99.98%, which indicates that the labeled training samples were expanded by the proposed method. The low classification accuracy of CV-CNN can be predominantly attributed to the low accuracy obtained for the rapeseed class, while the low accuracy of SVM can be attributed to the low accuracy obtained for the potatoes class and rapeseed class. This directly degrades the performance of the majority voting approach, thus leading to the poor performance of the proposed method in classifying the rapeseed and potato classes. There are also many fragments in the classification results of Wishart, SVM, and the proposed method, as shown in Figure 5e–g, respectively. From Figure 5c, it is clear that the misclassification is severe in the right side and top of the Flevoland image classified by CV-CNN. In addition, the results of the proposed method in Figure 5h appear much smoother than the results of the other methods. This indicates that the proposed method can maintain a good classification result, even with a small number of labeled samples.
To further analyze the improvement of the proposed method, a series of experiments were conducted. For the Flevoland dataset, the trends of accuracy with the increase of the training samples are shown in Figure 6. It can be seen that, with the increase in the number of training samples, the OA of each method generally increased. The superiority of the proposed method was more significant when the number of samples was small, especially when compared with CV-CNN. The accuracy of SVM was relatively poor, and its accuracy fluctuates with the increase of the samples. With 90 training samples in particular, the classification accuracy was only 74.3%. In general, the Wishart classifier was stable with different numbers of samples, and the OA is around 88.0%.

4.2. Experiments with the Oberpfaffenhofen Dataset

The second experiment was carried out with the dataset obtained over the German Aerospace Center at Oberpfaffenhofen in Germany. This image was acquired by the ESAR system. The Oberpfaffenhofen image, as a benchmark dataset, is widely used for PolSAR data classification research. According to the ground truth shown in Figure 7, there are three identified classes in the image, i.e., built-up areas, wood land, and open areas. The size of the image is 1300 × 1200.
In this experiment, we selected only 446 labeled samples; in other words, only 149 pixels were taken as the training samples for each class. The training sample ratio was only 0.18%, which is quite a low ratio. For the first CV-CNN experiment, on Oberpfaffenhofen dataset, the hyperparameters were set as follows. The learning rate was 0.8, and the batch size was 10 with 200 epochs. The loss curves are shown in Figure 8. With the increasing of epoch, the loss fluctuated, but the tendency decreased.
The classification results are shown in Figure 9 and listed in Table 3. From Table 3, we can see that the OAs for the Wishart, SVM, and CV-CNN base classifiers, the strong dataset and the proposed method are 75.70%, 77.01%, 78.89%, 87.47%, and 82.76%, respectively. The highest OA was again achieved by the proposed method, and the classification accuracies obtained by the proposed method for most categories were higher than those ob-tained by CV-CNN, except for the open areas class. The classification accuracies obtained by the Wishart, SVM, and CV-CNN base classifiers for the built-up area class were all below 50%, thus leading to the poor performances of the strong dataset and the proposed method in classifying the built-up area class. The black rectangle in Figure 9 indicates the distinctly different results of the above classifiers. From Figure 9g, it can be seen that the misclassification in the built-up area class for CV-CNN is serious, especially in the right of the image, which is consistent with the data in Table 3. There are also too many fragments in the Wishart and SVM classification maps (Figure 9e,f). For the built-up area class, the accuracy of the proposed method shows a significant improvement over CV-CNN, thanks to the good classification result of the strong dataset. It is clearly demonstrated that, compared with base classifiers, the accuracy of the proposed method on every land-cover type shows a distinct improvement. In addition, the visual effect of the proposed method in the classification map is much smoother than the effect of the other methods.
A series of experiments were carried out to further demonstrate the advantages of the proposed method. As shown in Figure 10, we can observe that, generally speaking, with the increase of the training samples, the OAs of each method show a gradual increase. The lowest OA was realized by the Wishart classifier. The Wishart classifier’s performance fluctuates with the increase of the training samples. The OA of SVM was about 82%, which is a similar performance to CV-CNN. The OA of the proposed method is always higher than that of the base classifiers, especially when the training samples are limited. At the right side of the Oberpfaffenhofen image, the buildings are sparsely distributed, and there are some forests, which makes the scattering characteristics similar to wood land. As shown in Table 3, the majority of accuracy in every land cover class was lower than open areas, which leads to a low accuracy of build-up area and wood land, and this leads to a small improvement in the Oberpfaffenhofen dataset experiment, compared to the Flevoland dataset experiment.

5. Conclusions

Deep learning methods have become more and more widely used in the field of PolSAR image classification. However, when the labeled training dataset is insufficient, the classification results are usually unsatisfactory. A novel semi-supervised classification method, which combines the deep learning technique with the traditional classifiers, has been proposed in this paper. The proposed method can take advantage of each classifier to generate a strong dataset and a weak dataset. The strong training set are used as pseudo-labels to reclassify the weak dataset by CV-CNN, which can make full use of the high-accuracy strong dataset. The final classification results are obtained by combining the strong training set and the reclassification results. The innovation of this method is that a majority voting approach is used to increase the training samples of the CV-CNN model, to further improve the classification performance. The experimental results obtained on two real PolSAR images confirmed the superiority of the proposed method with regard to the traditional supervised classifiers, especially when the number of original training samples was small. One of the limitations is that the settings of hyperparameters is slightly sophisticated in the proposed method and this makes it slightly time consuming. On one hand, our future work will focus on developing an automatic parameter optimization for the proposed method, and further investigation of the interactive connection between deep learning and the traditional scattering trait-based classifiers; on the other hand, our future work will carry out experiments on mountain areas to further verify the validity of the proposed method, and finally, space-borne PolSAR data also deserve to be researched.

Author Contributions

All the authors designed the study and discussed the basic structure of the manuscript. L.Z. and X.M. carried out the experiments and finished the first version of this manuscript. P.W. and J.X. reviewed and edited this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 41701390 and No. 61671334, and by the Natural Science Foundation of Anhui Province under Grant No. 1808085QD108.

Institutional Review Board Statement

Not application.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are openly available in https://earth.esa.int/eogateway/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, J.S.; Pottier, E. Polarimetric Radar Imaging: From Basics to Applications; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
  2. Nie, X.; Ding, S.; Huang, X.; Qiao, H.; Zhang, B.; Jiang, Z.P. An Online Multiview Learning Algorithm for PolSAR Data Real-Time Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 302–320. [Google Scholar] [CrossRef]
  3. Ronny Hänsch, O.H. Classification of Polarimetric SAR Data by Complex Valued Neural Networks; ISPRS Hannover Workshop-high-resolution Earth Imaging for Geospatial Information; Berlin Institute of Technology: Berlin, Germany, 2014. [Google Scholar]
  4. Liu, C.; Gierull, C.H. A new application for Po1SAR imagery in the field of moving target indication/ship detection. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3426–3436. [Google Scholar] [CrossRef]
  5. Voormansik, K.; Jagdhuber, T.; Zalite, K.; Noorma, M.; Hajnsek, I. Observations of Cutting Practices in Agricultural Grasslands Using Polarimetric SAR. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 1382–1396. [Google Scholar] [CrossRef]
  6. Yamaguchi, Y. Disaster Monitoring by Fully Polarimetric SAR Data Acquired With ALOS-PALSAR. Proc. IEEE 2012, 100, 2851–2860. [Google Scholar] [CrossRef]
  7. Gao, H.; Wang, C.; Wang, G.; Zhu, J.; Tang, Y.; Shen, P.; Zhu, Z. A Crop Classification Method Integrating GF-3 PolSAR and Sentinel-2A Optical Data in the Dongting Lake Basin. Sensors 2018, 18, 3139. [Google Scholar] [CrossRef] [Green Version]
  8. Nemni, E.; Bullock, J.; Belabbes, S.; Bromley, L. Fully Convolutional Neural Network for Rapid Flood Segmentation in Synthetic Aperture Radar Imagery. Remote Sens. 2020, 12, 2532. [Google Scholar] [CrossRef]
  9. Kong, J.A.; Swartz, A.A.; Yueh, H.A.; Novak, L.M.; Shin, R.T. Identification of Terrain Cover Using the Optimum Polarimetric Classifier. J. Electromagn. Waves Appl. 1988, 2, 171–194. [Google Scholar]
  10. Lee, J.S.; Grunes, M.R. Classification of multi-look polarimetric SAR data based on complex Wishart distribution. In Proceedings of the NTC-92: National Telesystems Conference, Washington, DC, USA, 19–20 May 1992; pp. 2299–2311. [Google Scholar]
  11. Liu, B.; Hu, H.; Wang, H.; Wang, K.; Liu, X.; Yu, W. Superpixel-Based Classification With an Adaptive Number of Classes for Polarimetric SAR Images. IEEE Trans. Geosci. Remote Sens. 2013, 51, 907–924. [Google Scholar] [CrossRef]
  12. Wen, X.; Ziwei, X.; Feng, Z.; Bo, R. POLSAR image classification via Clustering-WAE classification model. IEEE Access 2018, 6, 40041–40049. [Google Scholar]
  13. Freeman, A.; Villasenor, J.; Klein, J.D.; Hoogeboom, P.; Groot, J. On the use of multi-frequency and polarimetric radar backscatter features for classification of agricultural crops. Int. J. Remote Sens. 1994, 15, 1799–1812. [Google Scholar] [CrossRef]
  14. Kouskoulas, Y.; Ulaby, F.T.; Pierce, L.E. The Bayesian hierarchical classifier (BHC) and its application to short vegetation using multifrequency polarimetric SAR. IEEE Trans. Geosci. Remote Sens. 2004, 42, 469–477. [Google Scholar] [CrossRef]
  15. Cheng, X.; Huang, W.; Gong, J. An Unsupervised Scattering Mechanism Classification Method for PolSAR Images. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1677–1681. [Google Scholar] [CrossRef]
  16. Zhao, L.W.; Zhou, X.G.; Jiang, Y.M.; Kuang, G.Y. Iterative classification of polarimetric SAR image based on the freeman decomposition and scattering entropy. In Proceedings of the 2007 1st Asian and Pacific Conference on Synthetic Aperture Radar, Huangshan, China, 5–9 November 2007. [Google Scholar]
  17. Lee, J.-S.; Grunes, M.; Ainsworth, T.; Du, L.-J.; Schuler, D.; Cloude, S. Unsupervised Classification Using Polarimetric Decomposition and the Complex Wishart Classifier. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2249–2258. [Google Scholar]
  18. Cloude, S.R.; Pottier, E. An entropy based classification scheme for land applications of polarimetric SAR. IEEE Trans. Geosci. Remote Sens. 1997, 35, 68–78. [Google Scholar] [CrossRef]
  19. Chang, G.; Oh, Y. Polarimetric SAR Image Classification Based on the Degree of Polarization and Co-polarized Phase-Difference Statistics. J. Korean Inst. Electromagn. Eng. Sci. 2007, 18, 1345–1351. [Google Scholar] [CrossRef]
  20. Hao, D.; Xin, X.; Lei, W.; Pu, F. Gaofen-3 PolSAR Image Classification via XGBoost and Polarimetric Spatial Information. Sensors 2018, 18, 611. [Google Scholar]
  21. Hou, B.; Kou, H.; Jiao, L. Classification of Polarimetric SAR Images Using Multilayer Autoencoders and Superpixels. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3072–3081. [Google Scholar] [CrossRef]
  22. Liu, F.; Jiao, L.; Tang, X. Task-Oriented GAN for PolSAR Image Classification and Clustering. IEEE Trans. Neural. Netw. Learn. Syst. 2019, 30, 2707–2719. [Google Scholar] [CrossRef]
  23. Li, J.; Tang, Y.; Zou, X.; Lin, G.; Wang, H. Detection of Fruit-Bearing Branches and Localization of Litchi Clusters for Vision-Based Harvesting Robots. IEEE Access 2020, 8, 117746–117758. [Google Scholar] [CrossRef]
  24. Wang, L.; Xu, X.; Hao, D.; Rong, G.; Pu, F. Multi-Pixel Simultaneous Classification of PolSAR Image Using Convolutional Neural Networks. Sensors 2018, 18, 769. [Google Scholar] [CrossRef] [Green Version]
  25. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep Learning in Remote Sensing. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  26. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; The MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  27. Hirose, A.; Asano, Y.; Hamano, T. Developmental Learning With Behavioral Mode Tuning by Carrier-Frequency Modulation in Coherent Neural Networks. IEEE Trans. Neural. Netw. 2006, 17, 1532–1543. [Google Scholar] [CrossRef] [PubMed]
  28. Zhang, Z.; Wang, H.; Xu, F.; Jin, Y.-Q. Complex-Valued Convolutional Neural Network and Its Application in Polarimetric SAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
  29. Shao, Z.; Tang, P.; Wang, Z.; Saleem, N.; Sommai, C. BRRNet: A Fully Convolutional Neural Network for Automatic Building Extraction From High-Resolution Remote Sensing Images. Remote Sens. 2020, 12, 1050. [Google Scholar] [CrossRef] [Green Version]
  30. Gao, J.; Deng, B.; Qin, Y.; Wang, H.; Li, X. Enhanced Radar Imaging Using a Complex-Valued Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2019, 16, 35–39. [Google Scholar] [CrossRef] [Green Version]
  31. Li, Y.; Zhang, Y.; Zhu, Z. Learning Deep Networks under Noisy Labels for Remote Sensing Image Scene Classification. In Proceedings of the 2019 IEEE International Geoscience And Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 3025–3028. [Google Scholar]
  32. Tang, Y.; Chen, M.; Wang, C.; Luo, L.; Zou, X. Recognition and Localization Methods for Vision-Based Fruit Picking Robots: A Review. Front. Plant Sci. 2020, 11, 510. [Google Scholar] [CrossRef] [PubMed]
  33. Li, F.; Xu, L.; Siva, P.; Wong, A.; Clausi, D.A. Hyperspectral Image Classification With Limited Labeled Training Samples Using Enhanced Ensemble Learning and Conditional Random Fields. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2427–2438. [Google Scholar] [CrossRef]
  34. Xie, W.; Ma, G.; Zhao, F.; Liu, H.; Zhang, L. PoISAR image classification via a novel semi-supervised recurrent complex-valued convolution neural network. Neurocomputing 2020, 388, 255–268. [Google Scholar] [CrossRef]
  35. Liu, W.; Qin, R.; Su, F. Weakly supervised classification of time-series of very high resolution remote sensing images by transfer learning. Remote Sens. Lett. 2019, 10, 689–698. [Google Scholar] [CrossRef]
  36. Xu, L.; Clausi, D.A.; Li, F.; Wong, A. Weakly Supervised Classification of Remotely Sensed Imagery Using Label Constraint and Edge Penalty. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1424–1436. [Google Scholar] [CrossRef]
  37. Persello, C.; Bruzzone, L. Active and Semisupervised Learning for the Classification of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2014, 52, 6937–6956. [Google Scholar] [CrossRef]
  38. Li, Y.; Xing, R.; Jiao, L.; Chen, Y.; Shang, R. Semi-Supervised PolSAR Image Classification Based on Self-Training and Superpixels. Remote Sens. 2019, 11, 1933. [Google Scholar] [CrossRef] [Green Version]
  39. Sun, F.; Fang, F.; Wang, R.; Wan, B.; Wu, X. An Impartial Semi-Supervised Learning Strategy for Imbalanced Classification on VHR Images. Sensors 2020, 20, 6699. [Google Scholar] [CrossRef] [PubMed]
  40. Liu, H.; Zhu, D.; Yang, S.; Hou, B.; Gou, S.; Xiong, T.; Jiao, L. Semisupervised Feature Extraction With Neighborhood Constraints for Polarimetric SAR Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 9, 1–15. [Google Scholar] [CrossRef]
  41. Hou, B.; Wu, Q.; Wen, Z.; Jiao, L. Robust Semisupervised Classification for PolSAR Image With Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6440–6455. [Google Scholar] [CrossRef]
  42. Huang, Z.; Datcu, M.; Pan, Z.; Lei, B. Deep SAR-Net: Learning objects from signals. ISPRS J. Photogramm. Remote Sens. 2020, 161, 179–193. [Google Scholar] [CrossRef]
  43. Steele, B.M. Combining Multiple Classifiers—An Application Using Spatial and Remotely Sensed Information for Land Cover Type Mapping. Remote Sens. Environ. 2000, 74, 545–556. [Google Scholar] [CrossRef]
  44. Ma, X.; Shen, H.; Yang, J.; Zhang, L.; Li, P. Polarimetric-Spatial Classification of SAR Images Based on the Fusion of Multiple Classifiers. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 961–971. [Google Scholar]
  45. Kim, H.; Kim, H.; Moon, H.; Ahn, H. A weight-adjusted voting algorithm for ensembles of classifiers. J. Korean Stat. Soc. 2011, 40, 437–449. [Google Scholar] [CrossRef]
  46. Cloude, S.R.; Pottier, E. A review of target decomposition theorems in radar polarimetry. IEEE Trans. Geosci. Remote Sens. 1996, 34, 498–518. [Google Scholar] [CrossRef]
  47. Jiao, L.; Liu, F. Wishart Deep Stacking Network for Fast POLSAR Image Classification. IEEE Trans. Image Process. 2016, 25, 3273–3286. [Google Scholar] [CrossRef]
  48. Vapnik, V.N. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 1995. [Google Scholar]
  49. Maghsoudi, Y.; Collins, M.; Leckie, D.G. Polarimetric classification of Boreal forest using nonparametric feature selection and multiple classifiers. Int. J. Appl. Earth Obs. Geoinf. 2012, 19, 139–150. [Google Scholar] [CrossRef]
  50. Negri, R.G.; Frery, A.C.; Silva, W.B.; Mendes, T.S.G.; Dutra, L.V. Region-based classification of PolSAR data using radial basis kernel functions with stochastic distances. Int. J. Digit. Earth 2019, 12, 699–719. [Google Scholar] [CrossRef] [Green Version]
  51. Lv, W.J.; Wang, X.F. Overview of Hyperspectral Image Classification. J. Sens. 2020, 2020, 4817234. [Google Scholar] [CrossRef]
  52. Lee, J.S.; Grunes, M.R.; de Grandi, G. Polarimetric SAR speckle filtering and its implication for classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2363–2373. [Google Scholar]
Figure 1. Framework of the CV-CNN model.
Figure 1. Framework of the CV-CNN model.
Sensors 21 03006 g001
Figure 2. Framework of the proposed model.
Figure 2. Framework of the proposed model.
Sensors 21 03006 g002
Figure 3. Original Flevoland dataset. (a) Pauli RGB image for the Flevoland dataset. (b) Ground truth and legend of (a).
Figure 3. Original Flevoland dataset. (a) Pauli RGB image for the Flevoland dataset. (b) Ground truth and legend of (a).
Sensors 21 03006 g003
Figure 4. Loss curves on the Flevoland dataset.
Figure 4. Loss curves on the Flevoland dataset.
Sensors 21 03006 g004
Figure 5. Classification results for the Flevoland dataset. (ad) Classification results of the different approaches: (a) Wishart, (b) SVM, (c) CV-CNN, (d) proposed method. (eh) Results overlaid with the ground-truth maps of (ad), respectively.
Figure 5. Classification results for the Flevoland dataset. (ad) Classification results of the different approaches: (a) Wishart, (b) SVM, (c) CV-CNN, (d) proposed method. (eh) Results overlaid with the ground-truth maps of (ad), respectively.
Sensors 21 03006 g005
Figure 6. Trends of the OA with different numbers of training samples for the Flevoland dataset.
Figure 6. Trends of the OA with different numbers of training samples for the Flevoland dataset.
Sensors 21 03006 g006
Figure 7. Original Oberpfaffenhofen dataset. (a) Pauli RGB image. (b) Ground truth and legend of (a).
Figure 7. Original Oberpfaffenhofen dataset. (a) Pauli RGB image. (b) Ground truth and legend of (a).
Sensors 21 03006 g007
Figure 8. Loss curves on Oberpfaffenhofen dataset.
Figure 8. Loss curves on Oberpfaffenhofen dataset.
Sensors 21 03006 g008
Figure 9. Classification results for the Oberpfaffenhofen dataset. (ad) Classification results of the different approaches: (a) Wishart, (b) SVM, (c) CV-CNN, (d) proposed method. (eh) Results overlaid with the ground-truth maps of (ad), respectively.
Figure 9. Classification results for the Oberpfaffenhofen dataset. (ad) Classification results of the different approaches: (a) Wishart, (b) SVM, (c) CV-CNN, (d) proposed method. (eh) Results overlaid with the ground-truth maps of (ad), respectively.
Sensors 21 03006 g009
Figure 10. Trends of the OA with different numbers of training samples for the Oberpfaffenhofen dataset.
Figure 10. Trends of the OA with different numbers of training samples for the Oberpfaffenhofen dataset.
Sensors 21 03006 g010
Table 1. Basic parameters of the dataset.
Table 1. Basic parameters of the dataset.
Data SetPlatformPolarizationSpatial ResolutionBandNumber of LooksSize
FlevolandAIRSARQuad-polarization10 m × 10 mL41024 × 750
OberpfaffenhofenESARQuad-polarization3 m × 3 mL11300 × 1200
Table 2. Classification precision for the Flevoland dataset (%).
Table 2. Classification precision for the Flevoland dataset (%).
ClassNumber of SamplesWishartSVMCV-CNNStrong DatasetNumber of Samples in Strong DatasetProposed Method
Stem beans4199.8797.2881.9199.9810,93497.72
Peas6189.3183.2397.0799.2212,28897.93
Forest4589.9486.1791.9699.0120,61496.83
Lucerne3897.7094.5490.7299.9815,46996.82
Wheat4892.8481.8194.9599.8918,44296.63
Beet3295.3597.1984.5399.6662,29595.64
Potatoes1783.2753.5758.1485.2425,78080.93
Grass2586.1280.3278.9994.1027,15584.59
Rapeseed3365.8250.3949.2478.4512,92784.41
Barley1778.7388.9089.9999.8419,93491.29
Wheat 24767.0276.9784.7292.42852082.64
Wheat 35793.3992.6391.5299.6624,82297.41
Water4890.7888.9587.1299.5819,65396.19
OA--87.0480.7881.7797.34--90.75
Kappa--85.9679.1880.2597.11--89.96
Table 3. Classification precision for the Oberpfaffenhofen dataset (%).
Table 3. Classification precision for the Oberpfaffenhofen dataset (%).
ClassWishartSVMCV-CNNStrong DatasetProposed Method
Built-up area43.6646.2436.8661.2855.46
Wood land90.0581.2977.3993.1978.67
Open areas85.1689.2898.0699.8996.29
OA75.7077.0178.8987.4782.76
Kappa61.4862.4663.4175.2171.47
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, L.; Ma, X.; Wu, P.; Xu, J. Multiple Classifiers Based Semi-Supervised Polarimetric SAR Image Classification Method. Sensors 2021, 21, 3006. https://doi.org/10.3390/s21093006

AMA Style

Zhu L, Ma X, Wu P, Xu J. Multiple Classifiers Based Semi-Supervised Polarimetric SAR Image Classification Method. Sensors. 2021; 21(9):3006. https://doi.org/10.3390/s21093006

Chicago/Turabian Style

Zhu, Lekun, Xiaoshuang Ma, Penghai Wu, and Jiangong Xu. 2021. "Multiple Classifiers Based Semi-Supervised Polarimetric SAR Image Classification Method" Sensors 21, no. 9: 3006. https://doi.org/10.3390/s21093006

APA Style

Zhu, L., Ma, X., Wu, P., & Xu, J. (2021). Multiple Classifiers Based Semi-Supervised Polarimetric SAR Image Classification Method. Sensors, 21(9), 3006. https://doi.org/10.3390/s21093006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop