Next Article in Journal
Determination of Long-Term Soil Apparent Thermal Diffusivity Using Near-Surface Soil Temperature on the Tibetan Plateau
Previous Article in Journal
Influence of the Solar Spectra Models on PACO Atmospheric Correction
Previous Article in Special Issue
Standoff Infrared Measurements of Chemical Plume Dynamics in Complex Terrain Using a Combination of Active Swept-ECQCL Laser Spectroscopy with Passive Hyperspectral Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ES2FL: Ensemble Self-Supervised Feature Learning for Small Sample Classification of Hyperspectral Images

PLA Strategic Support Force Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2022, 14(17), 4236; https://doi.org/10.3390/rs14174236
Submission received: 26 June 2022 / Revised: 21 August 2022 / Accepted: 23 August 2022 / Published: 27 August 2022
(This article belongs to the Special Issue Hyperspectral Remote Sensing: Current Situation and New Challenges)

Abstract

:
Classification with a few labeled samples has always been a longstanding problem in the field of hyperspectral image (HSI) processing and analysis. Aiming at the small sample characteristics of HSI classification, a novel ensemble self-supervised feature-learning (ES 2 FL) method is proposed in this paper. The proposed method can automatically learn deep features conducive to classification without any annotation information, significantly reducing the dependence of deep-learning models on massive labeled samples. Firstly, to utilize the spatial–spectral information in HSIs more fully and effectively, EfficientNet-B0 is introduced and used as the backbone to model input samples. Then, through constraining the cross-correlation matrix of different distortions of the same sample to the identity matrix, the designed model can extract the latent features of homogeneous samples gathering together and heterogeneous samples separating from each other in a self-supervised manner. In addition, two ensemble learning strategies, feature-level and view-level ensemble, are proposed to further improve the feature-learning ability and classification performance by jointly utilizing spatial contextual information at different scales and feature information at different bands. Finally, the concatenations of the learned features and the original spectral vectors are inputted into classifiers such as random forest or support vector machine to complete label prediction. Extensive experiments on three widely used HSI data sets show that the proposed ES 2 FL method can learn more discriminant deep features and achieve better classification performance than existing advanced methods in the case of small samples.

1. Introduction

Hyperspectral image (HSIs) classification is widely used in earth observation systems [1,2]. However, the sparsity of labeled samples has long restricted the development of HSI classification technologies and results in that only a few labeled samples can be used for training models [3,4]. In this case, HSI classification directly using classifiers such as nearest neighbor, random forest (RF) and support vector machine (SVM) often could not obtain high-precision results [5]. Therefore, it is often said that HSI classification faces the challenge of small samples. Feature extraction, which can extract more easily distinguishable features by transforming hyperspectral data, is one of the effective methods to improve classification effects in the case of small samples [6,7]. At present, the development of HSI feature extraction has roughly experienced three stages: spectral feature extraction, spatial–spectral feature extraction and feature extraction based on deep learning.
HSIs could provide nearly continuous spectral curves of ground objects which contain rich property information for distinguishing different ground objects [8]. Therefore, the early feature-extraction methods are mainly aimed at spectral features. For example, the principal component analysis (PCA) method can be applied along the spectral dimension and produce a few principal components for classification, which can reduce spectral redundancy while maintaining classification accuracy [9]. However, the PCA method is actually a linear transform function, which cannot fit the high-dimensional and nonlinear characteristics of HSIs well. Therefore, nonlinear feature-extraction methods represented by manifold learning have been widely studied [10,11], and have obtained better classification performance. At the same time, the kernel methods map spectral vectors to higher dimensional features through kernel functions so that the samples that are linearly inseparable in the original spectral space become linearly separable in the high-dimensional feature space [12,13].
HSIs not only contain rich spectral features, but also provide rich spatial information [14]. Therefore, only considering spectral information in classification processes undeniably has certain limitations. To this end, the neighborhood spatial information of samples is introduced in the process of feature extraction [15,16]. For example, the morphological methods are used to extract spatial shape features, and Gabor or local binary pattern are both used to extract texture features [17,18]. Moreover, the spatial feature-extraction methods can be extended along the spectral dimension to fully extract the so-called spatial–spectral features so as to further improve the classification accuracy. For example, the Gabor filter can be extended from two dimensions to three dimensions, to directly extract the three-dimensional spatial–spectral features contained in HSI cubes [19]. Before the extensive application of deep-learning models, the spatial–spectral feature-extraction methods attracted great attention. However, the generalization ability of the above manually designed methods is generally weak. Different hyperparameters need to be set for different HSIs, which often requires strong professional knowledge and tedious trial-and-error tests, limiting the further application of these methods [20].
Recently, deep-learning methods have achieved impressive results in many computer vision tasks [21,22]. Hyperspectral curve detection and feature extraction combined with deep-learning methods have been widely used in many fields and sectors, such as fine agriculture [23], mine recognition [24] and environmental monitoring [25]. To automatically extract deep features suitable for the target tasks from input data, supervised deep-learning methods have been introduced into hyperspectral remote sensing image classification [26]. Among many deep-learning models, one-dimensional convolutional neural network (1D CNN) was first applied to the spectral dimension, and then extended to the spatial dimension and the spatial–spectral dimension [27,28,29]. At the same time, researchers attempt to treat HSIs as sequence data along the spectral dimension, and use the RNN models to fit them [30,31]. Then, 2D CNN, 3D CNN, capsule network and other advanced deep-learning models were introduced to make full use of the unique advantage of the “spatial–spectral unity” of HSIs [32]. Benefiting from the powerful feature-extraction capability, 2D CNN and 3D CNN have become the most widely used model structures in existing deep-learning-based methods, constantly refreshing the classification accuracy with sufficient samples. To further improve the accuracy and robustness of classification results, many advanced structures and modules such as residual connection, attention mechanism and transformer have also been applied [33,34,35,36,37].
The above supervised deep-learning models, which can integrate feature extraction and classification into a complete workflow, often require sufficient labeled samples to ensure excellent classification performance. However, The sparsity of HSI-labeled samples inevitably limits the performance of these methods in the case of small samples [32]. Considering this issue, the unsupervised feature-learning methods are adopted, which not only avoid the complicated process of designing feature-extraction rules by hand, but also effectively learn deep features with zero labeled sample [38,39]. For example, the encoder–decoder paradigms based on stacked autoencoder (SAE) are widely used in HSI unsupervised feature extraction [40,41,42]. Mou et al. integrate fully convolutional units, residual connections and a new unpooling operations into an autoencoder network, achieving better classification results than classical deep models [43]. The generative deep models represented by the generative adversarial network (GAN) are also used to extract abstract features in an unsupervised form [44,45,46]. For example, Zhang et al. improve the GAN model by introducing the Wasserstein distance and the fully convolutional network, effectively extracting the deep features with better invariance [45].
Undeniably, the unsupervised models of the encoder–decoder paradigm cannot extract deep features with strong robustness and discrimination by training only on reconstruction errors, while the training process of the GAN-based models is a very tricky issue due to the difficulty of convergence. Therefore, none of the above unsupervised feature-learning methods can obtain satisfactory classification results in the case of small samples. Self-supervised learning, as an effective representation-learning method, has had rapid development and close attention in the field of natural language processing and computer vision in recent years. It has a powerful data-learning ability, and can construct supervised information from massive unlabeled data through pretext tasks so as to carry out efficient training similar to supervised learning methods [47,48]. In view of the success of self-supervised learning methods in feature learning on unlabeled data, we propose an ensemble self-supervised feature-learning method, ES 2 FL, to improve HSI classification accuracy in the case of small samples. Specifically, the proposed method first utilizes a large number of unlabeled samples for self-supervised training and improves the ability of model-learning discriminant features by making the cross-correlation matrix of samples equal to the identity matrix [49]. Then, the optimized model is used to extract deep features from HSIs. Finally, the concatenations of the extracted features and the original spectral vectors are inputted into the machine-learning classifier to predict labels in the case of small samples to verify the feature-extraction effect of the proposed method. With the advanced network structure EfficientNet-B0 [50] as the backbone, the proposed method can better fit the high-dimensional and nonlinear hyperspectral data. In addition, two ensemble learning strategies, feature-level and view-level ensemble, are proposed and integrated into the proposed method to further improve classification performance. Therefore, our method is actually a novel self-supervised feature-learning method integrated with ensemble learning strategies. In summary, the contributions of this paper can be summarized as follows:
  • A novel ensemble self-supervised feature-learning method is proposed for small sample classification of HSIs. Through constraining the cross-correlation matrix of different distortions of the same sample to the identity matrix, the proposed method can learn deep features of homogeneous samples gathering together and heterogeneous samples separating from each other in a self-supervised manner, effectively improving HSI classification performance in the case of small samples.
  • To utilize the spatial–spectral information in HSIs more fully and effectively, a deep feature-learning network is designed based on the EfficientNet-B0. Experimental results show that the designed model possesses a greater feature-learning ability than general CNN-based models and deep residual models.
  • Two simple but effective ensemble learning strategies, feature-level and view-level ensemble, are proposed to further improve the feature-learning effect and HSI small sample classification accuracy by jointly utilizing spatial contextual information at different scales and feature information at different bands.
  • Extensive experiments on three HSIs are conducted and the statistical results demonstrate that the proposed method not only achieves better classification performance than existing advanced methods in the case of small samples, but also obtains higher classification accuracy with sufficient training samples.
The remainder of this paper is organized as follows. In Section 2, the proposed method is described in detail. In Section 3, extensive experiments are conducted and results analyses are given to verify the effectiveness of the proposed method. The conclusion is presented in Section 4.

2. Methodology

2.1. Workflow of the Proposed Method

To improve HSI classification accuracy in the case of small samples, an ensemble self-supervised feature-learning method is proposed. The whole workflow can be divided into two parts: self-supervised training on unlabeled samples (Figure 1), and feature learning and classification based on ensemble learning strategies (Figure 2).
In the self-supervised training process, the target HSI is first transformed into many cubes of equal size, and each cube is actually all the data in the neighborhood of the center pixel. Then, each cube is evenly divided into two sub-cubes along the spectral dimension [51] and the PCA method is used to extract the first three principal components. Next, several data augmentation methods such as random crop, flip and rotation are applied to generate the distortions, which can effectively enhance the difference between the two sub-cubes. In this way, for a given sample X, two three-band images (distortions) Y A and Y B can be obtained. Finally, the distortions are inputted into an arbitrary model to generate the deep feature vectors Z A and Z B , and the model is optimized based on the self-supervised loss between different feature vectors and the back-propagation algorithm. In the process of feature learning and classification, the optimized model is directly used for extracting deep features of the target HSI. After the extracted features are concatenated with the original spectral vector, the obtained features can be inputted into arbitrary classifiers such as RF or SVM to complete classification. As shown in Figure 2, the feature- and view-level ensemble strategies are proposed to further improve the feature-learning effect and classification performance.
Next, we will introduce in detail the loss function for model self-supervised training, the network structure of the deep model for feature learning and the two ensemble strategies proposed in this paper.

2.2. Self-Supervised Loss Function for Model Training

To effectively improve the ability of the more discriminative deep model learning features, the following loss function is applied:
L = i ( 1 C i i ) 2 + λ i j i C i j 2 .
In Equation (1), C is the cross-correlation matrix of Z A and Z B along the batch dimension. The whole loss function contains two objectives. Minimizing i ( 1 C i i ) 2 would cause the deep features generated from different distortions belonging to the same sample to become closely clustered together. Minimizing i j i C i j 2 would cause the deep features generated from different samples to become separated far from each other. The first term i ( 1 C i i ) 2 is invariant, while the hyperparameter λ is used to adjust the weight of the second term and balance the two optimization objectives. The cross-correlation matrix C is calculated according to the following formula:
C i j = b z b , i A z b , j B b ( z b , i A ) 2 b ( z b , j B ) 2 .
In Equation (2), C i j represents the value of row i and column j in the cross-correlation matrix C, b is the index value of the sample batch and i and j index the dimensions of feature vectors Z A and Z B , respectively.
Actually, minimizing the whole loss function is equivalent to making the cross-correlation matrix of Z A and Z B close to the identity matrix, which enables the learned features from the same sample to cluster with each other and the features generated from different samples to separate from each other so that the features with more discrimination can be obtained.

2.3. Deep Network for Feature Learning

As described above, arbitrary deep models can be used for mapping distortions into deep feature vectors, but different network structures will greatly affect the performance of feature learning. Designed by neural architecture search, EfficientNet-B0 can achieve stronger and more efficient feature learning by balancing the depth, width and resolution of the network to utilize the spatial–spectral information in HSIs more fully. The preliminary experimental results (Section 3.2) also show that, compared with CNN-based models and deep residual models, the designed deep model based on EfficientNet-B0 can further improve classification performance in the case of small samples. Therefore, EfficientNet-B0 is finally chosen as the backbone of the deep model for feature learning.
As shown in Table 1, EfficientNet-B0 consists of nine stages, where the first stage is a 3 × 3 convolution layer with 32 channels, the second stage is an MBConv1 structure, the stages 3–8 are MBConv6 structures with different kernel sizes, respectively, and the last stage is a classification layer composed of a 1 × 1 convolution layer, a pooling layer and a fully connected (FC) layer. All the above network structure settings are consistent with reference [50]. As shown in Figure 3, the MBConv structure consists of two 1 × 1 convolution layers, a depthwise separable convolution layer (kernel size is 3 × 3 or 5 × 5 ), a squeeze-and-excitation (SE) module, a dropout layer and a shortcut connection, of which the first 1 × 1 convolution layer and the depthwise separable convolution layer are both followed by a batch normalization (BN) layer and a swish activation function, to speed up training speed and improve training effect. In the SE module, adaptive average pooling is first applied to the input feature maps, and then two FC layers are used for nonlinear transformation, which are followed by the swish and sigmoid activation functions, respectively. The output of the SE module is obtained by multiplying the input feature maps and the output vectors of the second FC layer along the channel dimension. After the SE module, the feature maps go through a convolution layer, a BN layer and a dropout layer, and is added pixel-by-pixel to the input feature maps to generate the final output features of the MBConv structure. Besides the widely used convolution and pooling operations, the mathematical formulas of the BN layer, sigmoid function and swish function are listed below.
(1)
The BN layer can accelerate the convergence speed of deep models and make the training process more effective. For the input of the qth layer, the BN result can be represented as:
X ^ q = X q E ( X q ) V a r ( X q ) .
In Equation (3), X q and X ^ q represent the input and output for the BN layer, and E ( · ) and V a r ( · ) denote the expectation and standard deviation, respectively.
(2)
The sigmoid activation function can map the output into the interval [0, 1], which is conducive to fast calculation and back propagation. For the input of the qth layer, the result of sigmoid activation can be represented as:
S i g ( x q ) = 1 1 + e x q .
In Equation (4), e is the natural logarithm.
(3)
The swish activation function is smooth and non-monotonic. Existing studies have shown that the swish activation function can further improve the performance of deep models in many computer vision tasks [52]. For the input of the qth layer, the result of swish activation can be represented as:
S w i ( x q ) = x × S i g ( β x q ) .
In Equation (5), β is either a constant or a trainable parameter.
Only the first eight stages of EfficientNet-B0 are used as the backbone of the designed deep model for deep feature learning on HSIs. Two FC layers with equal dimensions are connected to the backbone, and the output vectors of the last layer are regarded as the deep features learned by the deep model. Obviously, the number of neurons in the two FC layers directly determines the dimension of the learned features, which has an important influence on classification results. In Section 3.2, the influence of the dimensions of the learned features on classification accuracy will be explored in detail.

2.4. Feature Extraction and Classification Based on Two Ensemble Learning Strategies

Ensemble learning, a simple but effective method to improve classification and recognition performance, has been widely applied in HSI classification. In recent years, with the extensive application of advanced data-driven deep-learning models in HSI classification, new and effective ensemble learning methods have been actively explored. For example, Li et al. use a voting strategy to perform joint classification on the basis of extracting deep pixel-pair features, effectively improving the classification accuracy and robustness of the designed 1D CNN model [53]. Based on the core idea of ensemble learning, two ensemble strategies, feature-level and view-level ensemble, are proposed and integrated into the proposed method to further improve the feature-learning effect and classification performance of the designed model.
(1)
Feature-level ensemble: As shown in Figure 4a, the first ensemble strategy attempts to improve the feature-learning effect by fusing multiscale spatial information. Firstly, the cubes with different scales around the center pixel are used to train the designed model in the self-supervised form, so that different models are more focused on spatial contextual information at different scales. In other words, deep models training on cubes with different scales are multiple base learners. Then, in the classification process, feature vectors generated by different base learners are fused to jointly utilize multiscale spatial information. Obviously, there are many different fusion methods, such as mean and concatenation, that can be employed. Moreover, there are many options for the size of the spatial scale. In the next section, the setting of the two hyperparameters will be analyzed in detail.
(2)
View-level ensemble: Inspired by the work of Liu et al. [51], different bands of HSIs can be regarded as different views. The proposed view-level ensemble strategy aims to improve the accuracy and robustness of classification results by comprehensively utilizing feature information from different views. As shown in Figure 4b, the target HSI is firstly reduced to multiple three-band images by random band selection, skilfully ensuring the consistency of the data dimension between the self-supervised training and classification process. Then, the voting strategy is used to integrate the classification results of different three-band images into the final classification results.
Figure 2 illustrates the combined application of the two ensemble learning strategies in detail. The feature-level ensemble is contained in the view-level ensemble, which can effectively improve HSI small sample classification accuracy by jointly utilizing spatial contextual information at different scales and feature information at different bands.

3. Experimental Results and Analysis

In this section, the experimental data sets are first introduced. Then, the setting of several important hyperparameters is analyzed in detail, and the classification results are given in the case of small samples and sufficient samples. Finally, ablation studies are conducted. In the experiments, all the algorithms are developed with Python 3.9 and machine learning libraries such as Pytorch and TensorFlow. The hardware environment includes a computer equipped with an Intel(R) Xeon(R) Gold 6152 CPU and an Nvidia A100 PCIE GPU, and other related devices.

3.1. Data Sets

Three widely used public HSIs are selected as experimental data sets to evaluate the classification performance of different methods, including University of Pavia (UP), Indian Pines (IP) and Salinas (SA). As shown in Table 2, the three HSIs possess different ground objects, resolutions and spectral ranges, providing different data environments to comprehensively evaluate the classification performance of different models. In the self-supervised training process, all samples in target HSIs are treated as unlabeled samples. In the feature learning and small sample classification process, referring to existing studies, five labeled samples per class are selected randomly from target HSIs for training classifiers, and all samples are used to evaluate classification results.

3.2. Hyperparameter Settings

As analyzed above, the network structure of the backbone, the setting of the spatial scale in the feature-level ensemble strategy and the dimension of the learned features have a great influence on classification results. In this section, therefore, the above hyperparameters are studied in detail to optimize the classification performance of the proposed method. It should be noted that RF is used as the classifier in all the experiments in this section.
In the proposed method, arbitrary networks can be used as the backbone of the deep model, but different network structures have an important impact on the feature-learning effect and classification accuracy. Considering the characteristics of input data, four networks including a deep 2D CNN-based model DCNN [54] and three advanced deep models, VGG16 [55], Resnet18 [56] and EfficientNet-B0, are selected as the backbone, and their influence on classification accuracy is explored. According to the results in Table 3, adopting EfficientNet-B0 as the backbone can achieve excellent and better classification performance than other deep networks. Especially in the IP data set, its classification accuracy is 5.89% higher than the second place. This indicates that EfficientNet-B0 can utilize the highly discriminative spatial–spectral information in HSIs more fully, thus effectively improving classification accuracy in the case of small samples.
The proposed feature-level ensemble strategy attempts to improve feature learning and classification performance by jointly utilizing spatial contextual information at different scales. Therefore, the setting and combination of different spatial scales is one of the problems that should be first explored. Specifically, the influence of six different spatial scale combinations on classification results is analyzed under the condition that the feature fusion method is set as the mean. As can be seen from Figure 5, within a certain range (from 16 to 80), gradually increasing the size of the spatial scale can effectively improve classification accuracy. Moreover, appropriate spacing between different spatial scales can further improve classification performance. For example, the classification accuracy of combination (32-64-96) is significantly higher than that of combination (48-64-80). It is believed that the appropriate spacing can make the model better integrate features at different scales while reducing spatial information redundancy. Comparing all the combinations, the setting of (32-64-96) can enable the proposed method to obtain the best classification performance. On the one hand, a large-scale setting ensures that the model can make full use of the contextual information in large spatial regions; on the other hand, appropriate spacing enables the effective joint utilization of spatial information at different scales.
The proposed method, ES 2 FL, is actually a self-supervised feature-learning method integrated with ensemble learning strategies. Therefore, the quality of the learned deep features directly determines the level of classification results. Dimension is one of the important properties of feature vectors: a too large dimension will lead to information redundancy and increase the burden of the classifier, while a too small dimension will result in the incomplete inclusion of important discriminant information in HSIs. Figure 6 shows the relationship between the dimension of the learned features and the overall classification accuracy. As can be seen, when the dimension value is set to 1024, the proposed method can achieve the highest classification accuracy. Generally, with the increase in dimension value, the classification accuracy shows a trend of increasing first and then decreasing, and the classification accuracy will decrease to varying degrees if the dimension value is too large or too small.
In addition to the above hyperparameters, other experimental settings are directly given by referring to relevant studies. In the self-supervised training process, stochastic gradient descent is adopted as the optimization algorithm, the learning rate is set to 0.001 and the iteration epochs on the UP, IP and SA data sets are set to 50, 15 and 50, respectively, in accordance with the preliminary experiments. The hyperparameter λ in the self-supervised loss function is set to 0.005, and the number of views in the view-level ensemble strategy is 10. In addition, different distortions of the same sample are generated by random cropping, cutout, rotation and flipping.

3.3. Classification Results in the Case of Small Samples

A classical semi-supervised machine-learning classifier, two semi-supervised deep-learning models and two advanced meta learning-based methods are selected and compared with the proposed method to verify its classification performance in the case of small samples. The proposed method, actually a deep feature-learning method, can be combined with different classifiers. In the experiments, two classifiers, support vector machine and random forest, are tested, and are denoted as ES 2 FL + SVM and ES 2 FL + RF, respectively. In addition to the proposed method, the other six methods used for comparison are as follows.
(1)
TSVM [57]: TSVM, short for Transductive Support Vector Machine, is one of the most representative models in semi-supervised support vector machines. TSVM can utilize the information contained in unlabeled samples to improve classification performance in the case of small samples by treating each unlabeled sample as a positive example or a negative example.
(2)
EMP + TSVM: EMP can utilize the spatial texture information in HSIs and retain the main spectral features, and its combination with TVSM can effectively improve the classification effect in the case of small samples.
(3)
EMP + GCN [58]: This method firstly extracts EMP features from HSIs, then organizes them into a graph structure by the k-neighbors method, and finally performs classification based on graph convolution. EMP + GCN is actually a semi-supervised classification method based on the designed graph convolution network, which can make full use of the information contained in labeled and unlabeled samples simultaneously.
(4)
3D-HyperGAMO [59]: Aiming at the issue of imbalanced data in HSI classification, this method attempts to generate more samples for minority classes using a novel generative adversarial minority oversampling strategy. The generated and original samples are both used for training the deep model, effectively improving classification performance in the case of small samples.
(5)
DFSL + SVM [60]: Based on the prototype network, this method designs a novel feature-learning and few-shot classification framework, which can make full use of the large number of source-labeled samples to improve classification accuracy in the case of small samples. This method is the first one to have explored the performance of meta learning methods in HSI classification.
(6)
RN-FSC [32]: RN-FSC is an end-to-end few-shot classification framework based on the relation network. By comparing the similarity between different samples, this mothod can automatically extract the deep features of homogeneous samples gathering together and heterogeneous samples separating from each other to effectively improve classification accuracy in cases of small samples.
Existing studies have shown that, compared with the process of training and classification with sufficient samples, the case of small samples will undoubtedly make the classification task more difficult and the classification results more unstable. This is because only five label samples per class are selected, making it a daunting task to fully train the model. Furthermore, the quality of these randomly selected samples will directly lead to the level of classification accuracy. Therefore, in order to fairly compare different models, the same random seeds are set for each method in the experiments to ensure that the classification results of different methods will not be affected by differences in randomly selected samples. In addition, referring to the existing studies, all methods are run 10 times each, and the final results are presented in the form of mean value and standard deviation to further improve the credibility and persuasiveness of experimental results [29,36,45,61].
Consistent with other existing studies, the overall accuracy (OA), average accuracy (AA) and kappa coefficient are used to quantitatively evaluate the classification performance of different methods. The OA is used to measure the accuracy of all classification results, that is, the ratio of the number of correctly classified samples to the total number of samples. The AA is the mean value of classification accuracy per class, reflecting the adaptability of the model to different classes. The kappa coefficient is an index to measure the overall classification results, which can better reflect the classification performance on imbalanced data sets, and its value is usually between 0 and 1. In the statistical results of this paper, the kappa value is displayed by multiplying it by 100, which is consistent with the value range of OA and AA, and is convenient for comparison.
The classification results of different methods in the case of small samples are listed in Table 4, Table 5 and Table 6, from which several observations can be obtained.
(1)
The results of the TSVM-based methods are significantly worse than those of other deep models in terms of OA, AA and kappa. In addition, extracting EMP features can significantly improve the classification performance of TSVM.
(2)
Compared with the TSVM-based methods, EMP + GCN and 3D-HyperGAMO can both significantly improve the accuracy and robustness of classification results in the case of small samples. It can be observed that EMP + GCN can acquire higher accuracy on the UP and SA data sets, while 3D-HyperGAMO has better classification performance on the IP data set.
(3)
The two advanced meta learning-based methods, DFSL + SVM and RN-FSC, can further improve classification performance. The meta-training with a large number of source-labeled samples and the feature-extraction network based on metric learning can significantly improve the learning ability and classification accuracy of the deep models in the case of small samples.
(4)
The proposed method achieves the best classification performance in all three HSIs, and both the classification results of ES 2 FL + SVM and ES 2 FL + RF are significantly better than other methods. In the UP and IP data sets, ES 2 FL + RF can obtain higher classification accuracy, and its OA is improved by 6.83% and 8.78%, respectively, compared with the second place, which is undoubtedly significant. In the SA data set, the classification performance of ES 2 FL + SVM is slightly better than ES 2 FL + RF, which is 2.99% higher than the second place in terms of OA.
Experimental results have shown that the proposed method can effectively improve HSI classification accuracy in the case of small samples, which can be attributed to the following points: On the one hand, the introduced self-supervised loss function and the deep model with the EfficientNet-B0 as its backbone can enable the proposed method to effectively learn the highly discriminant features based on the spatial–spectral information in HSIs, effectively improving the separability between different classes. On the other hand, the combined application of the feature- and view-level ensemble strategies can further improve classification accuracy by fusing spatial contextual information at different scales and feature information at different bands.
Classification maps are the direct products of HSI classification, which can visually present classification results and reflect the spatial distribution characteristics of ground objects. In the case of small samples, the visual representation effect of classification maps is particularly important. Figure 7, Figure 8 and Figure 9 show the classification maps of different methods on the three HSI data sets, where different colors represent different classes. It can be seen that, compared with other methods, the proposed method can effectively reduce the noise and misclassification in its classification maps which are closest to the real ground truths. Benefiting from the feature-level ensemble strategy, the proposed method significantly improves the classification effect of planar ground objects (red boxes mark out the areas for visual comparison). For example, in Figure 9, the noise in the corresponding regions of classes Vinyard_untrained and Grapes_untrained is significantly reduced and the regional coherence is significantly improved. Through qualitative comparison on different classification maps, it is verified that the classification performance of the proposed method is better than those of the other six methods in the case of small samples from the perspective of visualization.

3.4. Classification Results with Sufficient Samples

In the case of small samples, the proposed method has shown better classification performance than several advanced models. When the number of labeled samples is further increased, how will the classification accuracy of the proposed method change? In order to further explore the classification performance of the proposed method with sufficient samples, this section conducts experiments and compares it with several current advanced deep-learning models. Specifically, an unsupervised feature-learning method 3DCAE [40], a novel TNT (transformer in transformer)-based deep model EMP+TNT [37], two deep convolution models combined with attention mechanism CACNN [62] and DBMA [63], a metric-based meta learning algorithm combined with dynamic routing mechanism DIN-SSC [15] and an image-level fully convolutional classification model FOctConvPA [64] are used for comparison. For each HSI, 100 labeled samples per class are randomly selected for training, and all methods are run 10 times each. As can be seen from the statistical results in Table 7, the proposed method can achieve the most excellent classification performance, with OA of 99.05%, 96.44% and 99.19% in the UP, IP and SA data set at best, respectively. This shows that the proposed method has good adaptability to the change of the number of training samples, and its classification accuracy is improved significantly with the increase in training samples.

3.5. Ablation Studies

To verify the effectiveness of the proposed ensemble strategies in improving classification performance, taking the SA data set as an example, ablation studies are conducted and detailed analyses are given in this section. The experimental designs and results are listed in Table 8. By comparing the statistics in the third column and the fourth column, it can be found that the classification performance based on only one spatial scale is significantly worse than that based on the fusion of the contextual information at three different spatial scales, no matter what kind of feature fusion mode is adopted. The results in the fourth column show that different feature-level ensemble modes have different influences on the classification results. The mode of calculating the mean value of the learned features can enable the deep model to obtain better classification results. In addition, the introduction of spectral feature vectors can further improve classification accuracy, because it can provide more spectral details. The statistics in the fifth column are the classification results of the combined application of two ensemble learning strategies, which are obviously better than the results in other modes. This indicates that the classification accuracy can be further improved by jointly utilizing spatial contextual information at different scales and feature information at different bands. In summary, the above statistical results directly demonstrate the effectiveness of the two ensemble learning strategies proposed in this paper in improving small sample classification performance.
To more intuitively observe the influence of different settings in ablation studies on the learned features, visual analysis is carried out. Specifically, the t-SNE (t-distributed stochastic neighbor embedding) algorithm [65] is used to reduce the dimensionality of the learned features and separability among the obtained data is observed. Figure 10a–c correspond to the third column in Table 8, and Figure 10d,e correspond to the statistics under the mode of mean and mean+spectral vector, respectively, in the fourth column in Table 8. It can be easily observed that the data after feature-level ensemble has better separability, with closer aggregation between homogeneous samples and more distant separation between heterogeneous samples. Moreover, the introduction of spectral vectors can further improve the degree of aggregation and separation.

4. Conclusions

To further improve HSI classification accuracy in the case of small samples, an ensemble self-supervised feature-learning method is proposed in this paper. The proposed method firstly utilizes a large number of unlabeled samples for self-supervised training, and then carries out feature learning and classification with two ensemble learning strategies; this has the following advantages: (1) Through constraining the cross-correlation matrix of different distortions of the same sample to the identity matrix, the proposed method can learn the more discriminant and informative features from unlabeled samples in a self-supervised manner, significantly reducing the dependence of deep-learning models on massive labeled samples. (2) The designed deep model based on the advanced network EfficientNet-B0 can make full use of the spatial–spectral information in HSIs, effectively improving the feature-learning effect and classification performance. (3) The combined application of the feature- and view-level ensemble learning strategies can enable the model to jointly utilize the spatial contextual information at different scales and feature information at different bands to further improve classification accuracy in the case of small samples. Experimental results show that the proposed method can achieve an OA of 83.99%, 69.62% and 89.36% at most on the UP, IP and SA data sets, respectively, in the case of small samples. Moreover, in the case of sufficient samples, the proposed method can achieve an OA of 99.05%, 96.44% and 99.19% on the UP, IP and SA data sets respectively. These statistics demonstrate that the proposed method can achieve better classification performance than existing advanced models in the case of small samples or sufficient samples.
It should be pointed out that, although the proposed method effectively improves the HSI small sample classification accuracy, it still fails to solve the cross-domain issue in HSI classification, that is, the HSIs for model training and classification are completely different. Based on this work, future studies will explore the issue of HSI cross-domain small sample classification by introducing meta learning methods and domain adaptation techniques.

Author Contributions

Methodology, B.L. and K.G.; investigation, A.Y., L.D. and C.Q.; resources, B.L., A.Y.; writing–original draft preparation, K.G.; writing–review and editing, B.L.; visualization, B.L., A.Y. and J.L.; supervision, A.Y., L.D. and C.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under Grant 41801388, 42101458 and 42130112.

Data Availability Statement

Publicly available datasets were analyzed in this study, which can be found here: https://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 16 April 2022).

Acknowledgments

The authors would like to thank all the professionals for kindly providing the codes associated with the experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhong, Y.; Wang, X.; Xu, Y.; Wang, S.; Jia, T.; Hu, X.; Zhao, J.; Wei, L.; Zhang, L. Mini-UAV-Borne Hyperspectral Remote Sensing: From Observation and Processing to Applications. IEEE Geosci. Remote Sens. Mag. 2018, 6, 46–62. [Google Scholar] [CrossRef]
  2. Xiao, J.; Li, J.; Yuan, Q.; Zhang, L. A Dual-UNet with Multistage Details Injection for Hyperspectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5515313. [Google Scholar] [CrossRef]
  3. Liu, B.; Yu, X.; Zhang, P.; Yu, A.; Fu, Q.; Wei, X. Supervised Deep Feature Extraction for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1909–1921. [Google Scholar] [CrossRef]
  4. Liu, B.; Yu, X. Patch-Free Bilateral Network for Hyperspectral Image Classification Using Limited Samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10794–10807. [Google Scholar] [CrossRef]
  5. Zheng, C.; Wang, N.; Cui, J. Hyperspectral Image Classification with Small Training Sample Size Using Superpixel-Guided Training Sample Enlargement. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7307–7316. [Google Scholar] [CrossRef]
  6. Li, W.; Chen, C.; Su, H.; Du, Q. Local Binary Patterns and Extreme Learning Machine for Hyperspectral Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  7. Zhong, Y.; Zhang, L. An Adaptive Artificial Immune Network for Supervised Classification of Multi-/Hyperspectral Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2012, 50, 894–909. [Google Scholar] [CrossRef]
  8. Khan, M.J.; Khan, H.S.; Yousaf, A.; Khurshid, K.; Abbas, A. Modern Trends in Hyperspectral Image Analysis: A Review. IEEE Access 2018, 6, 14118–14129. [Google Scholar] [CrossRef]
  9. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-Based Edge-Preserving Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  10. Ma, L.; Zhang, X.; Yu, X.; Luo, D. Spatial Regularized Local Manifold Learning for Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 609–624. [Google Scholar] [CrossRef]
  11. Duan, Y.; Huang, H.; Li, Z.; Tang, Y. Local Manifold-Based Sparse Discriminant Learning for Feature Extraction of Hyperspectral Image. IEEE T Cybernetics 2021, 51, 4021–4034. [Google Scholar] [CrossRef] [PubMed]
  12. Gu, Y.; Chanussot, J.; Jia, X.; Benediktsson, J.A. Multiple Kernel Learning for Hyperspectral Image Classification: A Review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
  13. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  14. Zhang, S.; Xu, M.; Zhou, J.; Jia, S. Unsupervised Spatial-Spectral CNN-Based Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5524617. [Google Scholar] [CrossRef]
  15. Gao, K.; Guo, W.; Yu, X.; Liu, B.; Yu, A.; Wei, X. Deep Induction Network for Small Samples Classification of Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3462–3477. [Google Scholar] [CrossRef]
  16. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  17. Li, W.; Du, Q. Gabor-Filtering-Based Nearest Regularized Subspace for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1012–1022. [Google Scholar] [CrossRef]
  18. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Local Binary Pattern-Based Hyperspectral Image Classification With Superpixel Guidance. IEEE Trans. Geosci. Remote Sens. 2018, 56, 749–759. [Google Scholar] [CrossRef]
  19. Jia, S.; Shen, L.; Zhu, J.; Li, Q. A 3-D Gabor Phase-Based Coding and Matching Framework for Hyperspectral Imagery Classification. IEEE Trans. Cybern. 2018, 48, 1176–1188. [Google Scholar] [CrossRef]
  20. Xue, Z.; Yu, X.; Tan, X.; Liu, B.; Yu, A.; Wei, X. Multiscale Deep Learning Network with Self-Calibrated Convolution for Hyperspectral and LiDAR Data Collaborative Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514116. [Google Scholar] [CrossRef]
  21. Wang, H.; Lin, Y.; Xu, X.; Chen, Z.; Wu, Z.; Tang, Y. A Study on Long-Close Distance Coordination Control Strategy for Litchi Picking. Agronomy 2022, 12, 1520. [Google Scholar] [CrossRef]
  22. Wu, F.; Duan, J.; Ai, P.; Chen, Z.; Yang, Z.; Zou, X. Rachis detection and three-dimensional localization of cut off point for vision-based banana robot. Comput. Electron. Agric. 2022, 198, 107079. [Google Scholar] [CrossRef]
  23. Cui, Q.; Yang, B.; Liu, B.; Li, Y.; Ning, J. Tea Category Identification Using Wavelet Signal Reconstruction of Hyperspectral Imagery and Machine Learning. Agriculture 2022, 12, 1085. [Google Scholar] [CrossRef]
  24. Booysen, R.; Lorenz, S.; Thiele, S.T.; Fuchsloch, W.C.; Marais, T.; Nex, P.A.; Gloaguen, R. Accurate hyperspectral imaging of mineralised outcrops: An example from lithium-bearing pegmatites at Uis, Namibia. Remote Sens. Environ. 2022, 269, 112790. [Google Scholar] [CrossRef]
  25. Obermeier, W.; Lehnert, L.; Pohl, M.; Makowski Gianonni, S.; Silva, B.; Seibert, R.; Laser, H.; Moser, G.; Müller, C.; Luterbacher, J.; et al. Grassland ecosystem services in a changing environment: The potential of hyperspectral monitoring. Remote Sens. Environ. 2019, 232, 111273. [Google Scholar] [CrossRef]
  26. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  27. Lee, H.; Kwon, H. Going Deeper with Contextual CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef]
  28. Praveen, B.; Menon, V. Study of Spatial-Spectral Feature Extraction Frameworks with 3-D Convolutional Neural Network for Robust Hyperspectral Imagery Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1717–1727. [Google Scholar] [CrossRef]
  29. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.I. A Simplified 2D-3D CNN Architecture for Hyperspectral Image Classification Based on Spatial–Spectral Fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2485–2501. [Google Scholar] [CrossRef]
  30. Mei, S.; Li, X.; Liu, X.; Cai, H.; Du, Q. Hyperspectral Image Classification Using Attention-Based Bidirectional Long Short-Term Memory Network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5509612. [Google Scholar] [CrossRef]
  31. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef]
  32. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep Relation Network for Hyperspectral Image Few-Shot Classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef]
  33. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.I. Feedback Attention-Based Dense CNN for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501916. [Google Scholar] [CrossRef]
  34. Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual Spectral-Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 449–462. [Google Scholar] [CrossRef]
  35. Yang, K.; Sun, H.; Zou, C.; Lu, X. Cross-Attention Spectral Spatial Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5518714. [Google Scholar] [CrossRef]
  36. Gao, K.; Liu, B.; Yu, X.; Yu, A. Unsupervised Meta Learning with Multiview Constraints for Hyperspectral Image Small Sample set Classification. IEEE Trans. Image Process. 2022, 31, 3449–3462. [Google Scholar] [CrossRef]
  37. Tan, X.; Gao, K.; Liu, B.; Fu, Y.; Kang, L. Deep global-local transformer network combined with extended morphological profiles for hyperspectral image classification. J. Appl. Remote Sens. 2021, 15, 38509. [Google Scholar] [CrossRef]
  38. Romero, A.; Gatta, C.; Camps-Valls, G. Unsupervised Deep Feature Extraction for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1349–1362. [Google Scholar] [CrossRef]
  39. Wei, W.; Xu, S.; Zhang, L.; Zhang, J.; Zhang, Y. Boosting Hyperspectral Image Classification with Unsupervised Feature Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502315. [Google Scholar] [CrossRef]
  40. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised Spatial-Spectral Feature Learning by 3D Convolutional Autoencoder for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  41. Shi, C.; Pun, C.M. Multiscale Superpixel-Based Hyperspectral Image Classification Using Recurrent Neural Networks With Stacked Autoencoders. IEEE Trans. Multimedia 2020, 22, 487–501. [Google Scholar] [CrossRef]
  42. Feng, J.; Liu, L.; Cao, X.; Jiao, L.; Sun, T.; Zhang, X. Marginal Stacked Autoencoder with Adaptively-Spatial Regularization for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3297–3311. [Google Scholar] [CrossRef]
  43. Mou, L.; Ghamisi, P.; Zhu, X.X. Unsupervised Spectral-Spatial Feature Learning via Deep Residual Conv-Deconv Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 391–406. [Google Scholar] [CrossRef]
  44. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  45. Zhang, M.; Gong, M.; Mao, Y.; Li, J.; Wu, Y. Unsupervised Feature Extraction in Hyperspectral Images Based on Wasserstein Generative Adversarial Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2669–2688. [Google Scholar] [CrossRef]
  46. Yu, W.; Zhang, M.; He, Z.; Shen, Y. Convolutional Two-Stream Generative Adversarial Network-Based Hyperspectral Feature Extraction. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5506010. [Google Scholar] [CrossRef]
  47. Jing, L.; Tian, Y. Self-Supervised Visual Feature Learning With Deep Neural Networks: A Survey. IEEE Trans. Pattern Anal. 2021, 43, 4037–4058. [Google Scholar] [CrossRef]
  48. Liu, X.; Zhang, F.; Hou, Z.; Wang, Z.; Mian, L.; Zhang, J.; Tang, J. Self-supervised Learning: Generative or Contrastive. IEEE Trans. Knowl. Data Eng. 2020. [Google Scholar] [CrossRef]
  49. Zbontar, J.; Jing, L.; Misra, I.; LeCun, Y.; Deny, S. Barlow Twins: Self-Supervised Learning via Redundancy Reduction. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  50. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the Machine Learning Research, Vancouver, BC, Canada, 13 December 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  51. Liu, B.; Yu, A.; Yu, X.; Wang, R.; Gao, K.; Guo, W. Deep Multiview Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7758–7772. [Google Scholar] [CrossRef]
  52. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  53. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  54. Zhi, L.; Yu, X.; Liu, B.; Wei, X. A dense convolutional neural network for hyperspectral image classification. Remote Sens. Lett. 2019, 10, 59–66. [Google Scholar] [CrossRef]
  55. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  57. Joachims, T. Transductive Inference for Text Classification using Support Vector Machines. In Proceedings of the Sixteenth International Conference on Machine Learning, Bled, Slovenia, 27–30 June 1999; pp. 200–209. [Google Scholar]
  58. Liu, B.; Gao, K.; Yu, A.; Guo, W.; Wang, R.; Zuo, X. Semisupervised graph convolutional network for hyperspectral image classification. J. Appl. Remote Sens. 2020, 14, 26516. [Google Scholar] [CrossRef]
  59. Roy, S.K.; Haut, J.M.; Paoletti, M.E.; Dubey, S.R.; Plaza, A. Generative Adversarial Minority Oversampling for Spectral–Spatial Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5500615. [Google Scholar] [CrossRef]
  60. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  61. Xu, Q.; Xiao, Y.; Wang, D.; Luo, B. CSA-MSO3DCNN: Multiscale Octave 3D CNN with Channel and Spatial Attention for Hyperspectral Image Classification. Remote Sens. 2020, 12, 188. [Google Scholar] [CrossRef]
  62. Guo, H.; Liu, J.; Yang, J.; Xiao, Z.; Wu, Z. Deep Collaborative Attention Network for Hyperspectral Image Classification by Combining 2-D CNN and 3-D CNN. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4789–4802. [Google Scholar] [CrossRef]
  63. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-Branch Multi-Attention Mechanism Network for Hyperspectral Image Classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef]
  64. Sun, Y.; Liu, B.; Yu, X.; Yu, A.; Xue, Z.; Gao, K. Resolution reconstruction classification: Fully octave convolution network with pyramid attention mechanism for hyperspectral image classification. Int. J. Remote Sens. 2022, 43, 2076–2105. [Google Scholar] [CrossRef]
  65. Van der Maaten, L.; Hinton, G. Viualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Illustration of the self-supervised training process of the proposed method.
Figure 1. Illustration of the self-supervised training process of the proposed method.
Remotesensing 14 04236 g001
Figure 2. Illustration of feature learning and classification based on ensemble learning strategies of the proposed method. The symbols s1, s2 and s3 represent different spatial scales.
Figure 2. Illustration of feature learning and classification based on ensemble learning strategies of the proposed method. The symbols s1, s2 and s3 represent different spatial scales.
Remotesensing 14 04236 g002
Figure 3. Detailed architecture of MBConv. Conv denotes a convolution layer, FC denotes a fully connected layer, BN denotes a batch normalization layer and SE denotes a squeeze-and-excitation module.
Figure 3. Detailed architecture of MBConv. Conv denotes a convolution layer, FC denotes a fully connected layer, BN denotes a batch normalization layer and SE denotes a squeeze-and-excitation module.
Remotesensing 14 04236 g003
Figure 4. Illustrations of the proposed ensemble strategies.
Figure 4. Illustrations of the proposed ensemble strategies.
Remotesensing 14 04236 g004
Figure 5. Influence of different spatial scale combinations on classification results.
Figure 5. Influence of different spatial scale combinations on classification results.
Remotesensing 14 04236 g005
Figure 6. Relationship between the dimensions of the learned features and the overall classification accuracy in three HSIs.
Figure 6. Relationship between the dimensions of the learned features and the overall classification accuracy in three HSIs.
Remotesensing 14 04236 g006
Figure 7. The classification maps resulting from different methods on the UP data set in the case of small samples.
Figure 7. The classification maps resulting from different methods on the UP data set in the case of small samples.
Remotesensing 14 04236 g007
Figure 8. The classification maps resulting from different methods on the IP data set in the case of small samples.
Figure 8. The classification maps resulting from different methods on the IP data set in the case of small samples.
Remotesensing 14 04236 g008
Figure 9. The classification maps resulting from different methods on the SA data set in the case of small samples.
Figure 9. The classification maps resulting from different methods on the SA data set in the case of small samples.
Remotesensing 14 04236 g009
Figure 10. Visual presentation of the learned features under different settings in ablation studies. Different colors represent different classes.
Figure 10. Visual presentation of the learned features under different settings in ablation studies. Different colors represent different classes.
Remotesensing 14 04236 g010
Table 1. Network architecture of EfficientNet B0.
Table 1. Network architecture of EfficientNet B0.
StagesOperationsChannel NumberLayer Number
1Conv 3 × 3 321
2MBConv1 3 × 3 161
3MBConv6 3 × 3 242
4MBConv6 5 × 5 402
5MBConv6 3 × 3 803
6MBConv6 5 × 5 1123
7MBConv6 5 × 5 1924
8MBConv6 3 × 3 3201
9Conv 1 × 1 &Pooling&FC12801
Table 2. Details of the HSI data sets. University of Pavia (UP), Indian Pines (IP), Salinas (SA), Ground Sample Distance (GSD) (m), Spatial size (pixel), Spectral range (nm), Airborne Visible Infrared Imaging Spectrometer (AVIRIS), Reflective Optics System Imaging Spectrometer (ROSIS).
Table 2. Details of the HSI data sets. University of Pavia (UP), Indian Pines (IP), Salinas (SA), Ground Sample Distance (GSD) (m), Spatial size (pixel), Spectral range (nm), Airborne Visible Infrared Imaging Spectrometer (AVIRIS), Reflective Optics System Imaging Spectrometer (ROSIS).
UPIPSA
Spatial size 610 × 340 145 × 145 512 × 217
Spectral range430–860400–2500400–2500
No. of bands103200204
GSD1.3203.7
Sensor typeROSISAVIRISAVIRIS
AreasPaviaIndianaCalifornia
No. of classes91616
Labeled samples42,77610,24954,129
Table 3. Overall accuracy (OA, %) of the proposed method for the target HSIs with different deep networks as the backbone.
Table 3. Overall accuracy (OA, %) of the proposed method for the target HSIs with different deep networks as the backbone.
No.1234
BackboneDCNNVGG16Resnet18EfficientNet-B0
UP79.35 ± 3.7480.63 ± 3.8582.76 ± 3.6983.99 ± 3.15
IP63.07 ± 3.8362.94 ± 4.1763.73 ± 3.9769.62 ± 3.73
SA85.09 ± 3.0786.74 ± 2.6988.31 ± 2.7388.81 ± 2.12
Table 4. The classification results of different methods on the UP data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
Table 4. The classification results of different methods on the UP data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
CriteriaTSVM
Mean ± SD
EMP + TSVM
Mean ± SD
EMP + GCN
Mean ± SD
3D-HyperGAMO
Mean ± SD
DFSL + SVM
Mean ± SD
RN-FSC
Mean ± SD
ES 2 FL + SVM
Mean ± SD
ES 2 FL + RF
Mean ± SD
Class 186.03 ± 10.3793.74 ± 5.1192.50 ± 5.0080.46 ± 7.2886.86 ± 6.3984.11 ± 9.6884.95 ± 12.8192.62 ± 9.67
Class 282.60 ± 6.1083.62 ± 5.7585.94 ± 2.5686.95 ± 3.6491.81 ± 3.2594.88 ± 3.6595.64 ± 4.1196.81 ± 3.05
Class 343.42 ± 8.9549.25 ± 12.1449.20 ± 8.7149.78 ± 8.4159.08 ± 12.0354.74 ± 9.3494.57 ± 4.4280.28 ± 15.16
Class 458.98 ± 12.2865.18 ± 18.1774.83 ± 17.8259.53 ± 13.4892.79 ± 4.1875.76 ± 9.5178.26 ± 10.1971.07 ± 11.56
Class 593.21 ± 3.4698.09 ± 1.3599.73 ± 0.0998.90 ± 1.0399.92 ± 0.2486.73 ± 8.2298.77 ± 2.4193.57 ± 9.01
Class 638.23 ± 14.8541.42 ± 13.8548.96 ± 10.9847.40 ± 8.1040.26 ± 7.4862.22 ± 18.9069.16 ± 14.2568.95 ± 13.90
Class 742.16 ± 4.5349.64 ± 13.8154.95 ± 17.3552.43 ± 17.1544.49 ± 9.2454.02 ± 11.1757.24 ± 8.3761.00 ± 12.96
Class 866.56 ± 8.1573.02 ± 7.9273.44 ± 6.2871.37 ± 9.1665.45 ± 8.6266.58 ± 11.3372.92 ± 12.9386.10 ± 6.85
Class 999.78 ± 0.1899.88 ± 0.0996.53 ± 6.8782.90 ± 12.3986.42 ± 23.1074.34 ± 12.9399.10 ± 0.5893.41 ± 7.92
OA64.72 ± 6.8969.58 ± 7.5575.23 ± 3.9670.42 ± 4.9171.64 ± 6.0777.16 ± 5.8281.95 ± 3.3083.99 ± 3.15
AA67.89 ± 2.6072.65 ± 3.7075.12 ± 3.2769.97 ± 3.6074.12 ± 4.6572.60 ± 4.3883.36 ± 2.0782.64 ± 2.85
kappa56.17 ± 6.6861.83 ± 7.6768.02 ± 4.1962.59 ± 5.4564.41 ± 6.5370.86 ± 6.6677.05 ± 3.7379.64 ± 3.75
Table 5. The classification results of different methods on the IP data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
Table 5. The classification results of different methods on the IP data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
CriteriaTSVM
Mean ± SD
EMP + TSVM
Mean ± SD
EMP + GCN
Mean ± SD
3D-HyperGAMO
Mean ± SD
DFSL + SVM
Mean ± SD
RN-FSC
Mean ± SD
ES 2 FL + SVM
Mean ± SD
ES 2 FL + RF
Mean ± SD
Class 128.66 ± 4.2535.89 ± 8.6037.72 ± 12.9838.59 ± 14.1737.00 ± 9.4232.93 ± 19.8760.49 ± 13.5361.71 ± 19.88
Class 236.00 ± 5.0339.47 ± 8.2147.32 ± 10.0151.92 ± 9.9946.78 ± 10.0758.53 ± 16.8156.92 ± 10.0367.90 ± 11.78
Class 327.44 ± 9.4443.43 ± 9.8040.54 ± 8.7140.86 ± 10.1339.90 ± 11.5552.43 ± 21.4934.08 ± 5.3946.66 ± 9.17
Class 425.78 ± 6.1037.20 ± 8.4126.48 ± 6.7739.36 ± 12.8228.08 ± 7.0848.14 ± 13.9531.82 ± 8.8347.90 ± 15.40
Class 552.84 ± 7.8046.29 ± 12.1452.02 ± 11.4264.54 ± 18.7277.88 ± 10.1872.62 ± 18.6982.82 ± 4.5880.99 ± 8.48
Class 683.38 ± 3.9579.37 ± 4.8481.03 ± 9.7583.82 ± 10.0294.13 ± 4.6283.11 ± 13.5989.35 ± 4.6686.26 ± 7.06
Class 728.17 ± 5.7430.87 ± 3.7331.48 ± 13.0524.67 ± 10.5030.82 ± 9.4829.48 ± 12.6530.81 ± 10.3723.05 ± 5.57
Class 894.87 ± 3.2996.93 ± 1.9895.89 ± 4.2595.70 ± 3.4398.04 ± 2.4594.81 ± 6.9897.38 ± 2.0287.84 ± 5.77
Class 915.57 ± 6.6721.48 ± 9.6116.00 ± 14.5610.37 ± 4.4519.36 ± 6.9513.55 ± 5.3420.65 ± 8.5127.00 ± 12.61
Class 1039.79 ± 7.9349.57 ± 11.2352.40 ± 7.7348.84 ± 12.6958.26 ± 12.9258.13 ± 17.9049.23 ± 10.0258.12 ± 12.20
Class 1162.98 ± 3.6766.46 ± 5.4770.44 ± 6.7471.18 ± 10.2874.33 ± 6.8372.51 ± 11.7476.45 ± 4.5380.87 ± 6.08
Class 1223.10 ± 3.2427.73 ± 5.5224.23 ± 7.4034.20 ± 9.1634.27 ± 5.5339.75 ± 9.1541.81 ± 10.8864.15 ± 11.38
Class 1371.75 ± 8.0674.89 ± 10.0991.92 ± 11.3772.02 ± 11.2175.01 ± 11.5249.34 ± 15.3087.65 ± 7.1093.35 ± 5.03
Class 1488.75 ± 4.1686.18 ± 6.9781.50 ± 5.1589.00 ± 7.2294.21 ± 3.0293.31 ± 4.3197.28 ± 2.3697.07 ± 1.80
Class 1533.87 ± 7.8850.30 ± 12.5643.06 ± 9.6651.54 ± 10.6655.58 ± 7.6862.08 ± 8.3866.75 ± 12.9973.98 ± 13.18
Class 1683.89 ± 24.8396.94 ± 5.1584.29 ± 20.8071.13 ± 11.6086.99 ± 9.8845.91 ± 24.5193.71 ± 15.3356.46 ± 20.95
OA49.45 ± 2.2255.09 ± 3.5155.98 ± 2.9357.02 ± 3.0060.18 ± 3.5360.84 ± 3.1563.16 ± 3.2369.62± 3.73
AA49.80 ± 1.7055.19 ± 1.9654.77 ± 1.8955.48 ± 2.4759.41 ± 1.7356.66 ± 4.6563.58 ± 2.6365.83 ± 3.41
kappa43.64 ± 2.3349.62 ± 3.8050.65 ± 3.0052.17 ± 3.3055.66 ± 3.7256.52 ± 3.4158.84 ± 3.3865.99 ± 4.01
Table 6. The classification results of different methods on the SA data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
Table 6. The classification results of different methods on the SA data set in the case of small samples. SD denotes the standard deviation of 10 experimental results.
CriteriaTSVM
Mean ± SD
EMP + TSVM
Mean ± SD
EMP + GCN
Mean ± SD
3D-HyperGAMO
Mean ± SD
DFSL + SVM
Mean ± SD
RN-FSC
Mean ± SD
ES 2 FL + SVM
Mean ± SD
ES 2 FL + RF
Mean ± SD
Class96.11 ± 5.4694.81 ± 5.9299.93 ± 0.2189.82 ± 11.3099.92 ± 0.2491.105 ± 7.3887.74 ± 15.2793.92 ± 10.86
Class 298.12 ± 1.1898.92 ± 0.4798.22 ± 0.3696.78 ± 3.7899.63 ± 1.0297.21 ± 4.3395.07 ± 7.2698.88 ± 2.30
Class 386.09 ± 6.8585.83 ± 5.3692.04 ± 1.4392.82 ± 4.4093.04 ± 3.4088.68 ± 7.0296.95 ± 3.2794.14 ± 5.04
Class 496.35 ± 1.7697.27 ± 0.7596.67 ± 2.7292.82 ± 4.7298.48 ± 1.6386.56 ± 9.4690.04 ± 4.4181.09 ± 8.35
Class 594.29 ± 4.8194.08 ± 4.4596.59 ± 3.3793.76 ± 3.8994.32 ± 4.1697.06 ± 2.9898.54 ± 2.4596.89 ± 1.58
Class 699.74 ± 0.2699.97 ± 0.0499.93 ± 0.0399.27 ± 1.1499.99 ± 0.0297.89 ± 2.6699.98 ± 0.0399.22 ± 0.77
Class 792.88 ± 5.1791.24 ± 3.1198.04 ± 1.7796.52 ± 4.8299.13 ± 2.3993.60 ± 4.0299.47 ± 0.5798.36 ± 1.51
Class 872.93 ± 4.0374.59 ± 5.7371.15 ± 1.6975.35 ± 11.7180.08 ± 5.7984.95 ± 7.2491.83 ± 9.8189.89 ± 5.27
Class 997.25 ± 3.1898.23 ± 1.3598.67 ± 1.0597.80 ± 1.9796.34 ± 5.8095.78 ± 3.0097.81 ± 2.1697.90 ± 1.13
Class 1078.75 ± 8.4486.68 ± 4.6384.89 ± 5.7488.90 ± 6.0191.74 ± 8.6588.66 ± 9.6298.82 ± 1.1595.20 ± 3.80
Class 1163.99 ± 9.9869.43 ± 6.9478.57 ± 8.2086.46 ± 11.8163.45 ± 3.7587.01 ± 8.1590.37 ± 5.1590.41 ± 9.54
Class 1291.68 ± 5.1290.51 ± 3.8692.56 ± 4.1892.59 ± 7.7798.67 ± 1.3896.83 ± 3.6296.13 ± 5.1692.74 ± 7.90
Class 1382.34 ± 10.9181.16 ± 11.2090.58 ± 7.0489.34 ± 13.4999.84 ± 0.3395.12 ± 4.5689.62 ± 5.2188.39 ± 7.01
Class 1480.62 ± 13.6292.90 ± 3.2580.06 ± 16.4879.95 ± 12.6599.63 ± 0.5381.26 ± 11.6690.18 ± 5.3369.64 ± 12.28
Class 1555.43 ± 5.6958.29 ± 6.8364.91 ± 5.2255.17 ± 11.4354.62 ± 9.9265.69 ± 11.1565.39 ± 12.1267.27 ± 6.46
Class 1693.61 ± 4.1186.66 ± 7.2195.11 ± 5.5692.83 ± 6.0197.82 ± 5.3288.60 ± 9.9199.44 ± 0.6096.34 ± 2.19
OA82.25 ± 1.3683.58 ± 1.6585.93 ± 0.9983.70 ± 4.4084.52 ± 3.3286.37 ± 3.3289.36± 4.4288.81 ± 2.12
AA86.26 ± 1.3087.54 ± 0.7989.87 ± 0.9888.76 ± 2.5991.67 ± 0.8489.75 ± 1.6192.96 ± 2.4090.63 ± 1.56
kappa80.31 ± 1.4881.78 ± 1.8184.33 ± 1.1081.97 ± 4.8482.90 ± 3.6084.91 ± 3.6588.23 ± 4.8787.61 ± 2.33
Table 7. The classification results of different methods with sufficient samples. SD denotes the standard deviation of 10 experimental results.
Table 7. The classification results of different methods with sufficient samples. SD denotes the standard deviation of 10 experimental results.
HSICriteria3DCAE
Mean ± SD
EMP + TNT
Mean ± SD
CACNN
Mean ± SD
DBMA
Mean ± SD
DIN-SSC
Mean ± SD
FOctConvPA
Mean ± SD
ES 2 FL + SVM
Mean ± SD
ES 2 FL + RF
Mean ± SD
UPOA88.65 ± 1.7893.65 ± 0.8197.49 ± 1.3998.03 ± 1.3097.54 ± 0.7483.57 ± 3.4899.05 ± 0.3197.89 ± 0.28
AA86.96 ± 1.2992.34 ± 0.7997.95 ± 0.8197.11 ± 1.4996.93 ± 0.9473.22 ± 7.2998.55 ± 0.5996.80 ± 0.57
kappa85.24 ± 2.2091.72 ± 1.0596.67 ± 1.8397.41 ± 1.6896.76 ± 0.9778.04 ± 4.8198.75 ± 0.4197.22 ± 0.37
IPOA82.00 ± 1.0294.60 ± 2.1493.79 ± 0.8995.70 ± 1.4494.02 ± 1.6195.87 ± 0.9096.17 ± 0.4996.44 ± 0.69
AA82.76 ± 0.8994.73 ± 1.7995.65 ± 0.6696.13 ± 1.1093.81 ± 1.5495.34 ± 1.4096.01 ± 0.5496.04 ± 0.77
kappa79.11 ± 1.1993.69 ± 2.4892.67 ± 1.0494.97 ± 1.6693.02 ± 1.8795.14 ± 1.0795.51 ± 0.5795.83 ± 0.81
SAOA90.73 ± 0.6197.19 ± 4.3995.95 ± 0.6096.18 ± 1.4997.76 ± 1.0698.21 ± 0.6499.19 ± 0.2798.20 ± 0.41
AA93.86 ± 0.4398.64 ± 1.2398.19 ± 0.2798.25 ± 0.5998.32 ± 0.7696.50 ± 1.5299.48 ± 0.1898.02 ± 0.48
kappa89.70 ± 0.6796.89 ± 4.8395.48 ± 0.6895.75 ± 1.6597.51 ± 1.1798.01 ± 0.7199.09 ± 0.3098.00 ± 0.45
Table 8. The experimental designs and results of ablation studies.
Table 8. The experimental designs and results of ablation studies.
ClassifierCriteriaNo EnsembleFeature-Level Ensemble ModeFeature- and
View-Level
Ensembles
Scale = 32Scale = 64Scale = 96MeanElement
Multiplication
ConcatenationMean+
Spectral Vector
SVMOA70.37 ± 2.0663.87 ± 3.1379.27 ± 2.4283.74 ± 4.3182.59 ± 5.3575.94 ± 5.6887.12 ± 4.7589.36± 4.42
AA67.09 ± 0.8773.23 ± 2.2279.55 ± 2.3988.22 ± 2.4187.75 ± 2.6986.80 ± 2.3891.80 ± 1.8392.96 ± 2.40
kappa67.38 ± 2.1960.57 ± 3.5177.02 ± 2.7581.99 ± 4.7580.72 ± 5.9173.26 ± 6.7685.77 ± 5.2088.23 ± 4.87
RFOA71.73 ± 2.9970.96 ± 3.7877.78 ± 1.9783.70 ± 4.1783.21 ± 3.8781.62 ± 5.3288.16 ± 2.3488.81 ± 2.12
AA68.68 ± 2.1975.28 ± 2.9975.74 ± 1.1486.62 ± 2.4985.56 ± 2.6284.90 ± 3.0490.81 ± 1.1590.63 ± 1.56
kappa68.86 ± 3.1968.14 ± 4.0775.53 ± 2.2281.98 ± 4.5981.43 ± 4.2679.70 ± 5.8486.92 ± 2.5787.61 ± 2.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, B.; Gao, K.; Yu, A.; Ding, L.; Qiu, C.; Li, J. ES2FL: Ensemble Self-Supervised Feature Learning for Small Sample Classification of Hyperspectral Images. Remote Sens. 2022, 14, 4236. https://doi.org/10.3390/rs14174236

AMA Style

Liu B, Gao K, Yu A, Ding L, Qiu C, Li J. ES2FL: Ensemble Self-Supervised Feature Learning for Small Sample Classification of Hyperspectral Images. Remote Sensing. 2022; 14(17):4236. https://doi.org/10.3390/rs14174236

Chicago/Turabian Style

Liu, Bing, Kuiliang Gao, Anzhu Yu, Lei Ding, Chunping Qiu, and Jia Li. 2022. "ES2FL: Ensemble Self-Supervised Feature Learning for Small Sample Classification of Hyperspectral Images" Remote Sensing 14, no. 17: 4236. https://doi.org/10.3390/rs14174236

APA Style

Liu, B., Gao, K., Yu, A., Ding, L., Qiu, C., & Li, J. (2022). ES2FL: Ensemble Self-Supervised Feature Learning for Small Sample Classification of Hyperspectral Images. Remote Sensing, 14(17), 4236. https://doi.org/10.3390/rs14174236

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop