Next Article in Journal
A Study of Two Impactful Heavy Rainfall Events in the Southern Appalachian Mountains during Early 2020, Part I; Societal Impacts, Synoptic Overview, and Historical Context
Previous Article in Journal
Accuracy Assessment in Convolutional Neural Network-Based Deep Learning Remote Sensing Studies—Part 1: Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A 3D Cascaded Spectral–Spatial Element Attention Network for Hyperspectral Image Classification

1
School of Information Science and Technology, Northwest University, Xi’an 710127, China
2
Anyang Institute of Technology, College of Computer Science and Information Engineering, Anyang 455000, China
3
Shaanxi Province Silk Road Digital Protection and Inheritance of Cultural Heritage Collaborative Innovation Center; Xi’an 710127, China
4
Xi’an Microelectronics Technology Institute, Xi’an 710000, China
5
School of Information Engineering, Northwest A&F University, Xi’an 712100, China
6
Academy of Space Electronic Information Technology, Xi’an 710127, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(13), 2451; https://doi.org/10.3390/rs13132451
Submission received: 7 May 2021 / Revised: 12 June 2021 / Accepted: 18 June 2021 / Published: 23 June 2021

Abstract

:
Most traditional hyperspectral image (HSI) classification methods relied on hand-crafted or shallow-based descriptors, which limits their applicability and performance. Recently, deep learning has gradually become the mainstream method of HSI classification, because it can automatically extract deep abstract features for classification. However, it remains a challenge to learn more meaningful features for HSI classification from a small training sample set. In this paper, a 3D cascaded spectral–spatial element attention network (3D-CSSEAN) is proposed to solve this issue. The 3D-CSSEAN integrates the spectral–spatial feature extraction and attention area extraction for HSI classification. Two element attention modules in the 3D-CSSEAN enable the deep network to focus on primary spectral features and meaningful spatial features. All attention modules are implemented though several simple activation operations and elementwise multiplication operations. In this way, the training parameters of the network are not added too much, which also makes the network structure suitable for small sample learning. The adopted module cascading pattern not only reduces the computational burden in the deep network but can also be easily operated via plug–expand–play. Experimental results on three public data sets show that the proposed 3D-CSSEAN achieved comparable performance with the state-of-the-art methods.

Graphical Abstract

1. Introduction

With the development of remote sensing technology, hyperspectral images (HSIs) have been of wide concern and gradually applied in many fields [1,2]. In the field of HSIs, as a fundamental task, HSI classification is a task of assigning category labels to each pixel in the HSI and has attracted more and more attention.
An HSI usually contains hundreds of spectral bands, so it has abundant spectral information in addition to the usual spatial information of the image. In the early stages of HSI classification, there were many works based on spectral or spatial characteristics [3]. Support vector machines (SVMs) were used to address the problem by using spectral information [4]. In the past ten years, many works were based on spectral–spatial feature learning for HSI classification [5,6]. The performance of sparse representation was improved by using the spatial neighborhood information of samples [7]. In [8], principal component analysis (PCA) was used for unsupervised extraction of spectral features and data dimensionality reduction, and edge-preserving features were obtained by edge-preserving filtering, and the resulting features were classified by an SVM classifier. A hierarchical subspace and ensemble learning algorithm was proposed to solve the problem of hyperspectral image classification, in which spectral–spatial features were also applied [9]. Although most of these methods based on spectral–spatial features have achieved better results than those based on spectral information alone, they usually rely on hand-crafted or shallow-based descriptors. Therefore, the robustness and classification accuracy of these traditional methods still need to be improved.
In recent years, deep learning has been widely adopted in HSI classification because of its advantage of automatically learning discrimination features from raw data [10]. Autoencoders (AEs) were applied to extract the deep features of the image in an unsupervised manner [11,12]. In [13], the spectral information of each pixel was regarded as a sequence, and sequence features were extracted by recurrent neural networks (RNNs) for HSI classification. In [14], AE and RNN were combined to construct a new network for HSI classification.
Convolutional neural networks (CNNs) have been widely used in the field of HSI classification because of the advantages of spatial extraction and weight sharing mechanisms [15,16]. In [17], 1D CNNs were employed to extract the spectral features for HSI classification. The spectral images in HSIs were treated as the channels of conventional images, and then 2D CNNs were designed to extract the spatial features for HSI classification [18]. A 3D CNN that combined spectral and spatial information was used for HSI classification [19]. A spectral–spatial residual network (SSRN) adopted a 3D CNN and residual connections to improve the classification accuracy [20]. Batch normalization (BN) was used to regularize the training process in SSRN, making the training processing of the deep learning model more efficient. A 3D CNN has advantages over a 1D CNN and a 2D CNN in simultaneously extracting spectral and spatial features, while it requires more computation. To reduce the computational burden of 3D CNNs, 3D and 2D CNNs were mixed in a hybrid network (HybridSN) for HSI classification [21]. Overall, deep spectral–spatial feature learning has become a new trend in the classification of HSIs. Among these deep learning methods, it is difficult to achieve satisfactory results with the existing unsupervised network methods. Although these deep learning methods trained in a supervised manner can obtain encouraging results, they usually require sufficient labelled samples for training. However, obtaining labelled samples of hyperspectral images often consumes a lot of human and material resources. Therefore, training a deep learning model for hyperspectral image classification with limited samples is still a challenge.
Recently, some deep learning methods have introduced the attention mechanism to alleviate these problems in HSI classification [22,23]. The attention mechanism is inspired by the human visual mechanism [24,25]. When people observe a scene, they always pay more attention to the area of interest to obtain more meaningful information. In [26], the global pooling operations in the spectral dimension and spatial dimension were used to assign the attention to the interesting features. In [27], the spatial correlation and spectral band correlation were used to compute the attention weights of feature learning. In [28], a cascaded dual-scale crossover network (CDSCN) was proposed for HSI classification, which can obtain the parts of interest in the images through the multiplication of dual branch features. These methods use different ways to obtain attention features, thereby improving the classification performance. In addition to these attention methods, there may be other ways to extract attention features.
In this paper, a 3D cascaded spectral–spatial element attention network (3D-CSSEAN) is proposed for HSI classification. In 3D-CSSEAN, an element attention mechanism is used to extract spectral and spatial attention features, as shown in Figure 1. This method is different from the attention method mentioned above. It uses several activation functions to assign weights to all elements in the 3D feature tensor and obtains attention features through elementwise multiplication. The overall framework of 3D-CSSEAN is shown in Figure 2. It first uses convolution operations for data dimensionality reduction and shallow feature extraction. Then two attention modules are used to extract attention features. The following pooling operation is used to reduce the dimensionality of features. Finally, a fully connected layer and softmax activation layer are used to generate classification results. The main contributions of this work can be summarized in the following three aspects.
First, a cascade element attention network is proposed to extract meaningful features, which can give different weight responses to each element in the 3D data. Two element attention modules are employed to enhance the important spectral features and strengthen the interesting spatial features, respectively.
Second, the proposed element attention modules are implemented through several simple activation operations and elementwise multiplication operations. Therefore, the implementation of the attention module does not add too many parameters, which makes the network model suitable for small sample learning.
Third, the proposed attention modules can be easily plug and play, and can be achievable based on a single branch, so it is more time-efficient.
The rest of this paper is organized as follows: In Section 2, the existing attention methods for HSI classification are discussed. The proposed 3D-CSSEAN model is described in detail in Section 3. Experimental results and analysis are presented in Section 4. In Section 5, the influence of attention block numbers and different training sample numbers on the model are discussed. Finally, conclusions are summarized in Section 6.

2. Related Work

In this section, the existing attention methods for HSI classification are reviewed briefly. According to the different ways of paying attention to spectral and spatial features, these methods can be roughly divided into three categories:
  • Global operation-based methods. These methods use a global operation on an HSI or its feature map, such as global pooling or global convolution, to obtain the spectral attention weight or spatial attention weight [26,29]. As shown in Figure 3a, a spectral weight vector of the HSI is obtained by global operation of spatial dimension, and then the weight vector is multiplied by the HSI to achieve the spectral attention. Similarly, in Figure 3b, a spatial weight plane of the HSI is obtained by global operation of the spectral dimension, and subsequently, spatial attention features are obtained by multiplying the spatial weight plane by the HSI.
  • Correlation-based methods. Spatial location correlation and inter-channel correlation are used to describe the degree of attention [27,30]. The channel attention module can be illustrated as Figure 4a. Firstly, the original HSI or 3D feature tensor is reshaped to a plane with C height and N width, where C is the spectral dimension and N is the number of pixels. Next, matrix multiplication is performed on the plane and its transpose to obtain the channel correlation matrix. Finally, the channel attention features are obtained by multiplying the channel correlation matrix with the transpose matrix. The spatial attention features can also be obtained in a similar way, and the spatial attention module is shown in Figure 4b.
  • Multifeature-based methods. These methods usually appear in the form of two branches; the rough network structure is shown in Figure 5. The attention module is composed of a trunk and mask [31], and the trunk branch is composed of some residual blocks, and the mask branch is composed of a symmetrical downsampler–upsampler structure. Different features can be extracted by different network structures in two branches. Finally, the attention features are obtained by multiplying different features between the trunk branch and mask branch. Similarly, attention modules are composed of two branches extracting different scale spectral–spatial features [28]. The parts of interest in the images are obtained by multiplying different scale features between two branches. By adopting different structures or utilizing different scales, these attention models can extract meaningful information and improve the performance of classification tasks.
The above three kinds of attention methods may help deep networks pay more attention to the region of interest in space and important spectral bands. Recently, a multiattention fusion network (MAFN) [32] was proposed to merge multiple attention features for classification. MAFN is a method that combines the global operation-based method and the correlation-based method. However, these methods still have room for improvement. For global operation-based methods, the global pooling is too simple and crude to capture certain local attention features. For correlation-based methods, they have too high a computational burden due to matrix multiplication. For multifeature-based methods, they suffer from the small sample learning issue and computational burden because two branch networks inevitable increase the parameters. In this paper, the element attention mechanism is used to extract the spectral–spatial attention features, which is more meaningful for HSI classification. At the same time, the design of a single branch network structure can produce a network with less computing burden and higher time efficiency.

3. Proposed Method

As illustrated in Figure 2, the proposed 3D-CSSEAN contains four main modules: data dimension reduction module, spectral element attention module, spatial element attention module, and prediction layers. The 3D-CSSEAN firstly uses several 3D convolution operations for data-dimension reduction and spectral–spatial feature extraction. Then, the element attention mechanism is used to make the model focus on the primary spectral features and strengthen meaningful spatial features as well as to suppress unnecessary features. Finally, prediction layers are used to obtain the classification results. To fully utilize the spectral–spatial information of the HSI, each labeled pixel is first expanded into a 3D image patch centered on it, and then the patch is used as the input of the 3D-CSSEAN for training and testing. The training objective of the network is to update the parameters of the 3D-CSSEAN by minimizing cross-entropy loss between the predictive output and the truth label of the patch center pixel.

3.1. Data Dimension Reduction Module

Commonly, the utilization of hundreds of bands in the HSI is not only not optimal for classification but also increases the computational burden, especially for deep learning with a limited training data set. Therefore, data dimension reduction is necessary to improve the classification effect and time efficiency. The input of our model is a 3D image patch. Let the patch size be ω × ω × B , where ω × ω represents the spatial neighborhood of the centered pixel, and B is the band number of the HSI. In the proposed framework, for shallow feature extraction and spectral dimension reduction, a data dimension reduction module is designed based on a 3D convolutional operation, as shown in Figure 2. The i -th output of ( k + 1 ) -th 3D convolutional layer can be formulated as
P i k + 1 = j = 1 n k P j k W i k + 1 + b i k + 1 ,   i = 1 , 2 , n k + 1
P k + 1 = G ( P k + 1 )
where P j k R ω × ω × c k , 1 is the j -th component of P k , P k R ω × ω × c k , n k represents the input feature tensor of the ( k + 1 ) -th convolutional layer, ω × ω × c k is the size of the feature tensor, ω × ω represents the spatial size and c k represents the spectral size, n k is the number of the convolutional kernel in the k -th convolutional layer, W i k + 1 and b i k + 1 indicate weights and the bias of the i -th convolutional operation in the ( k + 1 ) -th layer, respectively, and denotes the 3D convolutional operation. After each convolution operation, batch normalization (BN) is used to regularize the training process, as in prior work [20]. Moreover, G ( · ) represents the BN operation and rectified linear unit (ReLU) activation function.
If the output data dimension of the convolution operation is expected to be smaller than the input data, then the convolution stride needs to be set greater than 1 or the convolution kernel size needs to be greater than 1 without a boundary padding. In the proposed model, three 3D convolutional layers, C 1 , C 2 , and C 3 , are used for spectral-dimension reduction, as shown in Figure 2. These convolutional layers used a 3D convolution kernel with 1 × 1 × L i , L i > 1 and added the subsampling procedure with a stride of ( 1 , 1 , S i ) , S i 1 , where   i is 1 , 2 , or 3 corresponding to C 1 , C 2 , and C 3 . The kernel size 1 × 1 × L i specify the height, width, and spectral dimensionality of the 3D convolution window, respectively. In particular, the convolutional layer C 3 integrates all the spectral features into one dimension by not padding the boundary, which is convenient for subsequent spatial feature extraction.
To better understand this process, an example diagram is used to illustrate the data dimension reduction module on the Indian Pines data set. As shown in Figure 6, let the input of the model be a tensor with a size of 7 × 7 × 200 where 7 × 7 represents the spatial size of the tensor, 200 is the spectral dimensionality. The first convolutional layer C 1 uses a convolution operation with a stride size of 2 to reduce the spectral dimension. The spectral dimension has been reduced from 200 to 97. The second convolutional layer C 2 uses a convolution kernel with 1 × 1 × 7 without a boundary padding to reduce the spectral dimension. The spectral dimension has been reduced from 97 to 91. Finally, the convolutional layer C 3 uses a convolution kernel with 1 × 1 × 91 without a boundary padding to integrate all the spectral features into one dimension.

3.2. Spectral Element Attention Module

Following the data dimension reduction module, a spectral element attention module is designed to extract deep meaningful spectral features for each patch. The spectral element attention module is composed of several attention blocks, which are shown in Figure 7. The red dotted box in Figure 7 represents an attention block, which can be defined as follows:
t e m p = t a n h ( P k W k + 1 + b k + 1 )
w e i g h t e d _ P = s o f t m a x ( t e m p )  
P k + 1 = G ( w e i g h t e d _ P × P k )
where P k is the input tensor of the spectral element attention block, P k + 1 is the output of the spectral element attention block, W k + 1 and b k + 1 indicate weights and the bias of the convolutional operation in the ( k + 1 ) -th layer, respectively, represents the 3D convolutional operation, and × represents the elementwise multiplication operation. To extract spectral features, a 1 × 1 × L e , L e > 1 convolution kernel is used, where L e represents the kernel size of spectral dimension. Moreover, t a n h ( · ) and s o f t m a x ( · ) represent the t a n h and s o f t m a x activation function, respectively. The t a n h activation function can play a role in contrast stretching, which can increase the relative separability of data around zero. The s o f t m a x activation function can map the outputs to a probability distribution ranging from 0 to 1, which are considered to be the weight map (or mask) of the spectral features. The attention block can pay the different levels of attention to spectral features via elementwise multiplication operation between w e i g h t e d _ P and P k . Finally, the output of the element attention block is obtained through the BN layer and the activation layer. Since this method can give different attention weight for each element in the tensor, this attention block is called an element attention block. It should be noted that the output tensors of the convolution operation are the same size as the input tensors through the padding strategy, and thus the implementation of elementwise multiplication can be guaranteed.
To illustrate the method more clearly, an example diagram is used to illustrate the spectral element attention block. As shown in Figure 8, let the input of a spectral element attention block be a feature tensor with size of ( 7 × 7 × 91 , 24 ) , where 7 × 7 represents the spatial size of feature map, 91 is the spectral dimensionality, and 24 is the number of the 3D feature map. First, a convolution layer with kernel size 1 × 1 × 3 is used to extract spectral features from the input data. The t a n h activation and s o f t m a x activation are utilized to transform spectral features to attention weights. Finally, spectral attention features are obtained by elementwise multiplication between the original feature tensor and the attention weights.
From the above process, it can be seen that the spectral element attention block first extracts the features by 3D convolution. Then it converts the features into attention weights by two simple activation functions. Finally, the elementwise multiplication between the weights and the features of the previous layer is performed. The element attention method can give different weights to any element in the tensor, thereby achieving more attention to detail features. This method considers all the elements of the feature map, so local details will not be lost. Meanwhile, this single-branch implementation does not add many training parameters, so the model is easy to converge and implement for small data sets. However, there are still several limitations to this module. Because the values of   w e i g h t e d _ P are in the range [0, 1], its multiplication over P k features may degrade them in deeper layers. Drawing on the idea of a residual network [20], this problem can be mitigated by adding P k + i + 1 and P k + i . Equation (5) is reformulated as follows:
P k + i + 1 = G ( w e i g h t e d _ P × P k + i ) + P k + i   i = 1 , 2 , , M
where + denotes the elementwise addition, and P k + i and P k + i + 1 represent the input and output of i-th attention block, respectively.

3.3. Spatial Element Attention Module

The spatial element attention module has a similar structure to the spectral element attention module. Unlike the spectral element attention module, the convolutional kernel size is L a × L a × 1 , L a > 1 in the spatial element attention module for the spatial feature extraction. The structure of a spatial element attention block is shown in Figure 9. A convolution layer with kernel size 7 × 7 × 1 is used to extract spatial features from the input data firstly. Then spatial attention weights and attention features are obtained in the same way as the spectral element attention module. It should be noted that the input of the spatial module is ( 7 × 7 × 1 , 24 ) , because the C 3 convolutional layer reduces the spectral dimension to 1 , as shown in Figure 2. The spatial element attention module is also composed of several spatial element attention blocks, as shown in Figure 7.
As can be seen from the above introduction, regardless of the spectral feature or the spatial feature, different attention degrees can be obtained in this way of element attention, so this model does not need to design different global pooling methods based on the spectral feature and the spatial feature.
Finally, in the prediction layers, the average pooling layer is used to reduce the dimensions of the feature tensor, while a flatten layer, a fully connected layer, and a s o f t m a x activation function are adopted for classification.

3.4. Analysis of the Role of the t a n h Function

In this section, the influence of the t a n h function on the data is briefly analyzed. The function curve of the t a n h function in the interval of [ 5 ,   5 ]   is shown in Figure 10. For values outside the interval of [ 5 ,   5 ] , the value of the tanh function was infinitely close to 1 as the value of the horizontal axis became smaller and smaller. On the other hand, the larger the number of the horizontal axis, the closer the value of the function became to 1 . It can be seen that the t a n h function had a higher slope at the 0 point and its surroundings compared to the other positions. This also means that the image contrast stretch in this area was greater than in other areas. Moreover, the preprocessed data conformed to the Gaussian distribution with 0 mean unit variance, so there were many values distributed near 0 . Thus, the t a n h function could increase the relative separability of most data. At the same time, the t a n h function could also suppress the contrast at some too large or too small values. In order to show the effect of the t a n h function, the visualization result of the image after t a n h transformation is provided in Figure 11. Figure 11a–c show the images before transformation, and Figure 11d–f show the results transformed by the t a n h function. It can be clearly seen from the figure that most details of Figure 11d–f are clearer and easier to identify than in Figure 11a–c.

4. Experimental Results

4.1. Experimental Setup

This section evaluates the performance of our method on three public hyperspectral image data sets. The Indian Pines data set includes 16 vegetation classes and has 224 bands from 400 to 2500 nm. After removing water absorption bands, it had 145 × 145 pixels with 200 bands. The Kennedy Space Center data set includes 13 classes and has 224 bands from 400 to 2500 nm. After removing water absorption bands, it had 512 × 453 pixels with 176 bands. The Salinas Scene data set includes 16 classes and has 224 bands from 360 to 2500 nm. After removing water absorption bands, it had 512 × 217 pixels with 204 bands.
In the Indian Pines data set, the labeled samples were unbalanced. In the Kennedy Space Center data set, the number of labeled samples was small. Compared with the Indian Pines and Kennedy Space Center, the labeled samples in the Salinas Scene data set were larger and more balanced. Therefore, these three data sets represented three different situations. The performance of the proposed method was verified in three different cases, which could better demonstrate the generalization ability of the method. For the Indian Pines and Kennedy Space Center data sets, about 5%, 5%, and 90% of the labeled samples were randomly select as training, validation, and testing data sets, respectively. For the Salinas Scene data set, due to the large number of overall labeled samples, a smaller training ratio was set. The ratio was about 1%:1%:98% for the Salinas Scene data set. Moreover, all three data sets were normalized to a Gaussian distribution with zero mean and unit variance. The overall accuracy (OA%), average accuracy (AA%), and Kappa coefficient ( Kappa × 100 ) were used to evaluate the classification performance of the proposed methods. The higher these index values, the better the classification performance of the method. Each method was randomly run ten times, and the mean and standard deviation of the classification index were reported. All the experiments were implemented with a GTX 2080Ti GPU, 16 GB of RAM, Python 3.6, TensorFlow 1.10, and the Keras 2.1.0 framework.
To express more clearly, Table 1 shows the shape of input data and output data and the specific parameters of the convolutional operation in the 3D-CSSEAN for the Indian Pines data set. The settings of Kennedy Space Center and Salinas Scene data sets are same as Indian Pines except for the band number of the input data. C s p e and C s p a in Table 1 indicate the convolution operation in the spectral element attention module and spatial element attention module, respectively. For each convolutional layer, n k were set to be 24 for each convolutional layer, and experiments show that the change of n k in a small range had little impact on the result.

4.2. Comparison and Analysis of Experimental Results

To evaluate the superiority and effectiveness of the proposed 3D-CSSEAN model, some machine learning and deep learning classification methods were compared with it. These methods included a traditional machine learning method SVM, state-of-the-art 3D deep learning models such as SSRN [20] and HybridSN [21], and the latest attention networks, such as CDSCN [28] and MAFN [32]. SVM was implemented by scikit-learn tools of the machine learning. The Radial Basis Function (RBF) was selected as the kernel function on the three data sets. The grid search method was used to determine the best values of parameters C and g a m m a . Other comparison methods were implemented through code published in their papers [20,21,28,32]. For fairness of comparison, the input image patch size was set to 7 × 7 × B for all methods except HybridSN, where B was the band number of the HSI. For HybridSN, in order to make the network work without changing the network structure, the input image patch size was set to 11 × 11 × B , which was the closest parameter setting. For SVM and HybridSN, the number of PCA principal components was set to 30, which is the same as in the literature on HybridSN [21].
Classification results of the different methods on testing data of the three data set are reported in Table 2, Table 3 and Table 4. As shown, 3D-CSSEAN achieved the best results on most indicators compared with the other methods. In our cases, the classification performances of all deep learning methods were better than those of SVM, which indicates that these deep learning models are generally superior to the traditional machine learning method in HSI classification. On the Indian Pines data set, the 3D-CSSEAN, MAFN, and CDSCN achieved better results than other methods. These results show that in the case of imbalanced categories, these attention models pay more attention to meaningful features, so they achieved better results. Compared with the two other attention methods, the 3D-CSSEAN increased the score at least 0.89%, 1.52%, and 1.01% in the OA, AA, and Kappa, respectively. Moreover, the AA of the 3D-CSSEAN was 0.89% higher than the best result of the other compared methods. These results indicate that the proposed method has good stability and robustness under the condition of unbalanced samples.
On the Kennedy Space Center data set, the 3D-CSSEAN, SSRN, CDSCN, and MAFN achieved at least 22% improvement compared to HybridSN and SVM. The reasons for this may be that HybridSN and SVM use PCA for dimension reduction, while the 3D-CSSEAN, SSRN, CDSCN, and MAFN are end-to-end network structures. The data dimension reduction module in the end-to-end is implemented in a supervised way, so the effect is better than the unsupervised way of PCA. Compared with SSRN and CDSCN, the 3D-CSSEAN achieved 2% and 1.75% improvement on OA, respectively. As for the latest MAFN, the 3D-CSSEAN also achieved comparable results. MAFN was slightly better than the 3D-CSSEAN on AA. The possible reason is that the spatial distribution of some categories in the Kennedy Space Center data set was relatively scattered. MAFN uses the correlation-based attention method to extract spatial features. Correlation-based methods may better capture the connections between scattered samples of these categories, so as to obtain more ideal results. The increase in accuracy of these categories can improve AA. On the Salinas Scene data set, all methods achieved higher than the 94% overall accuracy, while the 3D-CSSEAN was 0.42, 0.47, and 0.55 higher than the best result of the other methods on OA, Kappa, and AA, respectively.
In general, the three attention methods, CDSCN, MAFN and 3D-CSSEAN, achieved good results, indicating that the attention features extracted by them are beneficial to classification. These results indicated that the proposed element attention method can also effectively improve the classification performance. According to the results of the three data sets, the 3D-CSSEAN has good generalization ability on different data sets.
The classification maps of the five methods and the corresponding ground truth maps of the three data sets are shown in Figure 12, Figure 13 and Figure 14. It can be clearly seen from these results that the higher the classification accuracy, the better the continuity of the classification map. For the Indian Pines data set, there were obvious noise and discontinuous regions, as shown in Figure 12b, while the classification effect of the 3D-CSSEAN was relatively good. As shown in Figure 13, although there are very few labeled samples in the Kennedy Space Center data set, the 3D-CSSEAN still achieved good results. On the contrary, many obvious misclassified pixels can be seen in Figure 13b,c. All methods achieved over 94% overall accuracy on Salinas Scene data sets; however, there were still significant differences, which can be observed in Figure 14. It can be seen from Figure 14g that the 3D-CSSEAN still performed well at the edge of the category and the easily confused area.
Training and testing times provide a direct measure of the computational efficiency of HSI classification methods. In Table 5, the training time and the test time on the test data of different methods are shown. As presented in Table 5, because their inputs were the data under dimension reduction through PCA, the training time of SVM and HybridSN was significantly lower than that of other methods. Additionally, the time efficiency of the 3D-CSSEAN was higher than that of SSRN, CDSCN, and MAFN. As for MAFN, this may be because it uses a mixture of global operation-based and correlation-based methods to extract attention features, so it is relatively time-consuming. In particular, the training and testing time of the 3D-CSSEAN was about half that of the CDSCN method. The possible reason for this is that CDSCN adopts the dual branches mode, while the 3D-CSSEAN adopts the single branch mode, and thus it can save about half of the running time.

4.3. Ablation Studies

Three ablation experiments were conducted to analyze the contribution of different attention modules to HSI classification. The results are shown in Table 6. NONE means the 3D-CSSEAN without spectral and spatial attention module. SPE-EAN indicates the 3D-CSSEAN only with the spectral attention module, and SPA-EAN indicates the 3D-CSSEAN only with the spatial attention module. The experimental results showed that any kind of attention module is helpful for classification. The role of the spatial attention module is more obvious than that of the spectral attention module. In terms of OA indicators, SPA-EAN increased 1.25%, 0.87%, and 1.06% more than SPE-EAN on Indian Pines, Kennedy Space Center, and Salinas Scene data sets, respectively. These results suggest that the spatial element attention module is more conducive to acquiring discriminative features for classification. The OA obtained by the 3D-CSSEAN had obvious improvement compared with the module without spectral–spatial attention. The OA of the 3D-CSSEAN was 3.17%, 3.66%, and 1.99% higher than without attention modules on Indian Pines, Kennedy Space Center, and Salinas Scene data sets, respectively. It can be seen from the results of ablation experiments that the proposed cascaded spectral–spatial element attention module can obtain more meaningful spectral and spatial features, thereby improving the final classification results.
To verify the contribution of t a n h activation function to the classification task, a series of experiments was conducted on the three data sets. Experiment results are shown in Table 7. As can be seen from Table 7, AA, Kappa, and OA were all improved on the three data sets by using the t a n h function. Compared with the model without t a n h , the OA score’s enhancements obtained by the 3D-CSSEAN with   t a n h were 0.56% (Indian Pines), 0.49% (Kennedy Space Center), and 0.12% (Salinas Scene). The AA score’s increases were 0.49% (Indian Pines), 0.82% (Kennedy Space Center), and 0.04% (Salinas Scene). The Kappa coefficient’s improvements were 0.64% (Indian Pines), 0.55% (Kennedy Space Center), and 0.14% (Salinas Scene). These results indicate that the t a n h function is beneficial to enhance the separability of features and improve the classification performance. In addition, the standard deviation of all the results also decreased through using the   t a n h function. This also shows that the stability of the model is improved by using the t a n h function.

5. Discussion

5.1. Influence of the Attention Block Number

On three public data sets, the influence of the attention block number on the classification performance was analyzed. The experimental results are shown in Figure 15. In the figure, i S P E _ j S P A of the horizontal axis represents i   attention blocks in the spectral element attention module and j attention blocks in the spatial element attention module. Figure 15a–c, respectively, show the influence of the attention block number on overall accuracy, average accuracy, and Kappa coefficient. As can be seen from the figure, on the Salinas Scene data set, the number of attention blocks had little effect on the results. Particularly, the model with 1 S P E _ 1 S P A achieved good performance of OA at over 98%, indicating that the network structure with only one spectral element attention block cascading to one spatial element attention block extracted enough features for the improvement of the classification performance.
On the Indian Pines and Kennedy Space Center data sets, when the number of the spectral attention block was 1, three indicators all fluctuated greatly with the increase of spatial attention modules. In the case of 1 S P E _ 3 S P A , all the indicators were significantly reduced. This result shows that when the spectral features are not sufficiently extracted, blindly adding spatial depth features will not bring good results. When the spectral feature block was greater than 2, the indicators on the Kennedy Space Center data set tended to be stable, and at the same time, the fluctuation range on the Indian Pines data set was also narrowing.
When the number of spectral attention modules was 2, and the number of spatial attention modules was from 1 to 2, both OA and Kappa increased slightly on the three data sets. In the case of 2 S P E _ 2 S P A , the best OA was achieved on Kennedy Space Center and Salinas Scene data sets. As for the Indian Pines data set, when the number of attention module increased, the improvement in classification performance was limited. Furthermore, as the number of attention block increased, the time efficiency was bound to decrease. Overall, the network with 2 S P E _ 2 S P A could achieve the best or very close to the best on three indicators. In addition, it had good performance on the three data sets, indicating that its generalization performance was better. Based on the above analysis, the network structure of our final model is 2 S P E _ 2 S P A .

5.2. Influence of Different Training Sample Numbers

To evaluate the performance of the proposed 3D-CSSEAN, in this paper, under different numbers of training samples, four groups of labeled samples with different percentages were randomly selected as training samples for experiments. Specifically, 1%, 3%, 5%, and 10% of each category were randomly selected from the labeled samples as training samples on the Indian Pines data set and Kennedy Space Center data set, and 0.1%, 0.5%, 1%, and 3% of each category were randomly selected from the labeled samples as training samples on the Salinas Scene data set. The experiment results are shown in Figure 16.
On the Indian Pines data set, the advantages were more obvious when 1% and 3% of the labeled samples were used for training. Meaningful features extracted by the 3D-CSSEAN were more conducive to improving the classification performance in the case of small samples. Moreover, there was a significant decrease in the OA of CDSCN when only 3% of the labeled samples were used for training, indicating that CDSCN is prone to overfitting small training data. However, the 3D-CSSEAN did not increase many training parameters in the implementation of the attention module, and thus this problem can be avoided to some extent. On the Kennedy Space Center data set, the three different attention models, the 3D-CSSEAN, MAFN, and CDSCN, achieved better results than other methods, especially at 1% and 3%. These results indicate that these three attention features are beneficial for classification on the Kennedy Space Center data set. On the Salinas Scene data set, all methods achieved relatively close results, but the results of the 3D-CSSEAN were always the highest. In most cases, all methods could achieve good results, but in 0.10% of cases, the 3D-CSSEAN and MAFN had more obvious advantages.
In general, on Indian Pines and Salinas Scene data sets, the 3D-CSSEAN consistently outperformed the other approaches on all the training samples. As for the Kennedy Space Center data set, the results of the 3D-CSSEAN and MAFN were very close, and these results were better than those from the other comparison methods. Through these experimental investigations, it can be concluded that the 3D-CSSEAN has better classification performance and robustness in different training sample sets, and especially in the case of small samples, this advantage is more obvious. In addition, the MAFN method based on multiple attention combinations also demonstrated its competitiveness, especially on the Kennedy Space Center data set, where the spatial distribution of categories was relatively scattered. This shows that the combination of multiple attention methods is a promising research direction. In the future, perhaps the combination of the proposed element attention method and other attention methods will also produce more competitive results.

6. Conclusions

In this paper, a 3D cascaded spectral–spatial element attention network (3D-CSSEAN) is proposed to extract the meaningful features for hyperspectral image classification. The spectral element attention module and the spatial element attention module can make the network focus on primary spectral features and meaningful spatial features. Two element attention modules were implemented through several simple activation functions and elementwise multiplication. Therefore, the proposed model not only can obtain features that facilitate classification, but also has high computational efficiency. Since the implementation of the attention module does not add too many training parameters, it also makes the network structure suitable for small sample learning.
To evaluate the effectiveness of the method, extensive experiments were implemented on three public data sets: Indian Pines, Kennedy Space Center and Salinas Scene. Compared with the machine learning method, the popular deep learning methods and the attention methods, the proposed method obtained better classification performance. In cases with small samples, the advantages of the proposed method are more obvious. These results verify that the attention features obtained by the 3D-CSSEAN are beneficial for classification, and the 3D-CSSEAN is suitable for small sample learning. To evaluate the effectiveness of attention modules, several ablation experiments were conducted. From the results of the ablation experiments, both the spectral element attention module and the spatial element attention module have improved classification performance.
Extensive experiments showed that in the case of limited training samples, how to extract more meaningful features for classification is a direction worth exploring. In addition, the fusion of multiple attention features may be a kind of potential method, but how to ensure time efficiency may be a direction to be studied in the future.

Author Contributions

Conceptualization, H.Y., E.Z. and L.T.; methodology, H.Y.; project administration and resources, J.W., J.P. and L.T.; software, K.Y. (Kai Yu) and K.Y. (Kun Yan); supervision, J.W. and J.P.; writing—original draft, H.Y.; writing—review and editing, H.Y., E.Z. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

The Work was supported by the Xi’an Key Laboratory of Intelligent Perception and Cultural Inheritance (No. 2019219614SYS011CG033), the Key Research and Development Program of Shaanxi (No. 2021ZDLGY15-06), the National Natural Science Foundation of China (Program No. 62006188), the Program for Changjiang Scholars and Innovative Research Team in University (No. IRT 17R87), and the Special scientific research project of Shaanxi Provincial Department of Education (NO. 20JK0940).

Data Availability Statement

Three publicly available data sets were analyzed in this work. These data sets can be found http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes, accessed on 7 May 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, X.G.; Yu, Y. Estimating Soil Salinity under Various Moisture Conditions: An Experimental Study. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2525–2533. [Google Scholar] [CrossRef]
  2. Peng, J.Y.; Yu, K.; Wang, J.; Zhang, Q.X.; Wang, L.; Fan, P. Mining painted cultural relic patterns based on principal component images selection and image fusion of hyperspectral images. J. Cult. Herit. 2019, 36, 32–39. [Google Scholar] [CrossRef]
  3. He, L.; Li, J.; Liu, C.Y.; Li, S.T. Recent Advances on Spectral-Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  4. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, E.L.; Zhang, X.R.; Liu, H.Y.; Jiao, L.C. Fast Multifeature Joint Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 12, 1397–1401. [Google Scholar] [CrossRef]
  6. Imani, M.; Ghassemian, H. An overview on spectral and spatial information fusion for hyperspectral image classification: Current trends and challenges. Inform. Fusion. 2020, 59, 59–83. [Google Scholar] [CrossRef]
  7. Peng, J.T.; Jiang, X.; Chen, N.; Fu, H.J. Local adaptive joint sparse representation for hyperspectral image classification. Neurocomputing 2019, 334, 239–248. [Google Scholar] [CrossRef]
  8. Kang, X.D.; Xiang, X.L.; Li, S.T.; Benediktsson, J.A. PCA-Based Edge-Preserving Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  9. Li, Y.M.; Xie, T.J.; Wang, P.; Wang, J.; Liu, S.J.; Zhou, X.C.; Zhang, X.Z. Joint spectral-spatial hyperspectral image classification based on hierarchical subspace switch ensemble learning algorithm. Appl. Intell. 2018, 48, 4128–4148. [Google Scholar] [CrossRef]
  10. Li, S.T.; Song, W.W.; Fang, L.Y.; Chen, Y.S.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  11. Chen, Y.S.; Lin, Z.H.; Zhao, X.; Wang, G.; Gu, Y.F. Deep Learning-Based Classification of Hyperspectral Data. IEEE J.-Stars 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  12. Zhang, X.R.; Liang, Y.J.; Li, C.; Ning, H.Y.; Jiao, L.C.; Zhou, H.Y. Recursive Autoencoders-Based Unsupervised Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 14, 1928–1932. [Google Scholar] [CrossRef] [Green Version]
  13. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Scalable recurrent neural network for hyperspectral image classification. J. Supercomput. 2020, 76, 8866–8882. [Google Scholar] [CrossRef]
  14. Shi, C.; Pun, C.M. Multiscale Superpixel-Based Hyperspectral Image Classification Using Recurrent Neural Networks With Stacked Autoencoders. IEEE Trans. Multimed. 2020, 22, 487–501. [Google Scholar] [CrossRef]
  15. Yang, X.F.; Zhang, X.F.; Ye, Y.M.; Lau, R.Y.K.; Lu, S.J.; Li, X.T.; Huang, X.H. Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sens. 2020, 12, 2033. [Google Scholar] [CrossRef]
  16. Xu, H.; Yao, W.; Cheng, L.; Li, B. Multiple Spectral Resolution 3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sens. 2021, 13, 1248. [Google Scholar] [CrossRef]
  17. Hu, W.; Huang, Y.Y.; Wei, L.; Zhang, F.; Li, H.C. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, X.F.; Ye, Y.M.; Li, X.T.; Lau, R.Y.K.; Zhang, X.F.; Huang, X.H. Hyperspectral Image Classification With Deep Learning Models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  19. Ben Hamida, A.; Benoit, A.; Lambert, P.; Ben Amar, C. 3-D Deep Learning Approach for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  20. Zhong, Z.L.; Li, J.; Luo, Z.M.; Chapman, M. Spectral-Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  21. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D-2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  22. Li, Z.W.; Cui, X.S.; Wang, L.Q.; Zhang, H.; Zhu, X.; Zhang, Y.J. Spectral and Spatial Global Context Attention for Hyperspectral Image Classification. Remote Sens. 2021, 13, 771. [Google Scholar] [CrossRef]
  23. Qing, Y.H.; Liu, W.Y. Hyperspectral Image Classification Based on Multi-Scale Residual Network with Attention Mechanism. Remote Sens. 2021, 13, 335. [Google Scholar] [CrossRef]
  24. Jie, H.; Li, S.; Gang, S. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; p. 27. [Google Scholar]
  26. Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual Spectral-Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 449–462. [Google Scholar] [CrossRef]
  27. Tang, X.; Meng, F.; Zhang, X.; Cheung, Y.-M.; Ma, J.; Liu, F.; Jiao, L. Hyperspectral Image Classification Based on 3-D Octave Convolution With Spatial-Spectral Attention Network. IEEE Trans. Geosci. Remote Sens. 2020, 1–18. [Google Scholar] [CrossRef]
  28. Cao, F.L.; Guo, W.H. Cascaded dual-scale crossover network for hyperspectral image classification. Knowl.-Based Syst. 2020, 189, 105122. [Google Scholar] [CrossRef]
  29. Mou, L.C.; Zhu, X.X. Learning to Pay Attention on Spectral Domain: A Spectral Attention Module-Based Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 110–122. [Google Scholar] [CrossRef]
  30. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–Spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3232–3245. [Google Scholar] [CrossRef]
  31. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Visual Attention-Driven Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  32. Li, Z.; Zhao, X.; Xu, Y.; Li, W.; Zhai, L.; Fang, Z.; Shi, X. Hyperspectral Image Classification with Multiattention Fusion Network. IEEE Geosci. Remote Sens. Lett. 2021, 1–5. [Google Scholar] [CrossRef]
Figure 1. The elemental attention mechanism used in the 3D-CSSEAN.
Figure 1. The elemental attention mechanism used in the 3D-CSSEAN.
Remotesensing 13 02451 g001
Figure 2. The proposed framework of the 3D-CSSEAN.
Figure 2. The proposed framework of the 3D-CSSEAN.
Remotesensing 13 02451 g002
Figure 3. Global operation-based attention mechanism approaches for HSI classification: (a) schematic diagram of obtaining spectral attention features; (b) schematic diagram of obtaining spatial attention features. The symbol represents the dot multiplication.
Figure 3. Global operation-based attention mechanism approaches for HSI classification: (a) schematic diagram of obtaining spectral attention features; (b) schematic diagram of obtaining spatial attention features. The symbol represents the dot multiplication.
Remotesensing 13 02451 g003
Figure 4. Correlation-based attention mechanism approach for HSI classification: (a) schematic diagram of obtaining spectral attention features; (b) schematic diagram of obtaining spatial attention features. The symbol   represents the matrix multiplication.
Figure 4. Correlation-based attention mechanism approach for HSI classification: (a) schematic diagram of obtaining spectral attention features; (b) schematic diagram of obtaining spatial attention features. The symbol   represents the matrix multiplication.
Remotesensing 13 02451 g004
Figure 5. Multifeature-based attention mechanism approach of the HSI. The symbol represents the element multiplication between features of two branches.
Figure 5. Multifeature-based attention mechanism approach of the HSI. The symbol represents the element multiplication between features of two branches.
Remotesensing 13 02451 g005
Figure 6. Diagram of the data dimension reduction process on the Indian Pines data set.
Figure 6. Diagram of the data dimension reduction process on the Indian Pines data set.
Remotesensing 13 02451 g006
Figure 7. The attention module in the proposed 3D-CSSEAN.
Figure 7. The attention module in the proposed 3D-CSSEAN.
Remotesensing 13 02451 g007
Figure 8. The spectral element attention block.
Figure 8. The spectral element attention block.
Remotesensing 13 02451 g008
Figure 9. The spatial element attention block.
Figure 9. The spatial element attention block.
Remotesensing 13 02451 g009
Figure 10. The curve of the tanh function.
Figure 10. The curve of the tanh function.
Remotesensing 13 02451 g010
Figure 11. Comparison of visualization results before and after tanh function transformation: (a) the 100th band in the Salinas data set; (b) the 50th band in the Kennedy Space Center data set; (c) the 100th band in the Indian Pines data set; (df) represent the image after tanh transformation of (ac), respectively.
Figure 11. Comparison of visualization results before and after tanh function transformation: (a) the 100th band in the Salinas data set; (b) the 50th band in the Kennedy Space Center data set; (c) the 100th band in the Indian Pines data set; (df) represent the image after tanh transformation of (ac), respectively.
Remotesensing 13 02451 g011
Figure 12. Classification map for the Indian Pines data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Figure 12. Classification map for the Indian Pines data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Remotesensing 13 02451 g012
Figure 13. Classification map for the Kennedy Space Center data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Figure 13. Classification map for the Kennedy Space Center data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Remotesensing 13 02451 g013
Figure 14. Classification map for the Salinas Scene data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Figure 14. Classification map for the Salinas Scene data set: (a) ground truth; (b) SVM; (c) HybridSN; (d) SSRN; (e) CDSCN; (f) MAFN; (g) 3D-CSSEAN.
Remotesensing 13 02451 g014
Figure 15. Classification performance of the 3D-CSSEAN with different numbers of attention blocks. IN, KSC, and SA represent the Indian Pines, Kennedy Space Center, and Salinas Scene data sets, respectively. (a) Overall accuracy; (b) average accuracy; (c) Kappa coefficient.
Figure 15. Classification performance of the 3D-CSSEAN with different numbers of attention blocks. IN, KSC, and SA represent the Indian Pines, Kennedy Space Center, and Salinas Scene data sets, respectively. (a) Overall accuracy; (b) average accuracy; (c) Kappa coefficient.
Remotesensing 13 02451 g015
Figure 16. Overall accuracy (%) of the 3D-DSSEAN with different training sample proportions on the three data sets: (a) Indian Pines; (b) Kennedy Space Center; (c) Salinas Scene.
Figure 16. Overall accuracy (%) of the 3D-DSSEAN with different training sample proportions on the three data sets: (a) Indian Pines; (b) Kennedy Space Center; (c) Salinas Scene.
Remotesensing 13 02451 g016
Table 1. The input, output, and parameters of convolutional operation for the Indian Pines data set.
Table 1. The input, output, and parameters of convolutional operation for the Indian Pines data set.
LayerKernel SizeStrideInput ShapeOutput Shape
C 1 1 × 1 × 7 (1,1,2) 7 × 7 × 200 , 1 7 × 7 × 97 , 24
C 2 1 × 1 × 7 (1,1,1) 7 × 7 × 97 , 24 7 × 7 × 91 , 24
C s p e 1 × 1 × 3 (1,1,1) 7 × 7 × 91 , 24 7 × 7 × 91 , 24
C 3 1 × 1 × 91 (1,1,1) 7 × 7 × 91 , 24 7 × 7 × 1 , 24
C s p a 3 × 3 × 1 (1,1,1) 7 × 7 × 1 , 24 7 × 7 × 1 , 24
Table 2. Classification results for Indian Pines data set. Bold represents the best results.
Table 2. Classification results for Indian Pines data set. Bold represents the best results.
ClassSVMHybridSNSSRNCDSCNMAFN3D-CSSEAN
110089.0190.0089.8296.53100
258.0887.4193.7794.2495.2296.24
373.7886.0488.6094.8993.1596.66
489.8189.5692.2891.4090.7494.79
596.7095.6997.4098.8397.5098.41
698.3196.7597.7398.3998.9797.92
790.0096.3150.0098.3883.6994.39
893.2492.4296.8597.7899.3699.75
980.0078.8130.0098.7597.9598.57
1070.7987.3290.8394.3394.4996.35
1171.3490.7294.8995.0398.2898.25
1261.9189.2894.1191.1593.8497.38
1310094.5599.7398.8196.8998.07
1493.5494.6297.1197.5199.2998.33
1589.7293.1694.5892.4995.0994.66
1699.1790.4998.3497.9092.5394.30
AA85.40 ± 3.1290.76 ± 2.4787.89 ± 6.4895.61 ± 0.7395.22 ± 1.1697.13 ± 0.83
Kappa73.65 ± 0.8889.25 ± 1.2393.40 ± 1.6094.59 ± 1.0695.99 ± 0.8897.00 ± 0.65
OA77.17 ± 0.7490.60 ± 1.0794.21 ± 1.4195.26 ± 0.9296.48 ± 0.7897.37 ± 0.57
Table 3. Classification results for the Kennedy Space Center data set. Bold represents the best results.
Table 3. Classification results for the Kennedy Space Center data set. Bold represents the best results.
ClassSVMHybridSNSSRNCDSCNMAFN3D-CSSEAN
179.5487.3198.3398.8399.0799.68
246.9055.7797.1896.3010098.43
352.2554.3889.8086.4697.7993.81
441.5242.9285.1086.8399.5193.42
553.9565.4586.7085.2298.5194.13
665.7253.6793.0695.5998.9997.73
782.0974.4193.9989.8295.5095.53
858.5569.2296.8896.8198.1998.99
985.1893.6699.6399.7693.1399.94
1034.2248.3299.9499.8110099.94
1110095.2599.0599.5310099.40
1253.5465.2799.5699.7510099.67
1394.2090.4210099.70100100
AA65.21 ± 1.3968.93 ± 2.7595.34 ± 2.0294.95 ± 1.1998.02 ± 0.5797.74 ± 0.99
Kappa66.96 ± 1.1971.48 ± 2.7296.26 ± 1.6396.54 ± 0.7098.42 ± 0.3898.48 ± 0.59
OA70.43 ± 1.0574.41 ± 2.4596.64 ± 1.4696.89 ± 0.6298.60 ± 0.6298.64 ± 0.53
Table 4. Classification results for the Salinas Scene data set. Bold represents the best results.
Table 4. Classification results for the Salinas Scene data set. Bold represents the best results.
ClassSVMHybridSNSSRNCDSCNMAFN3D-CSSEAN
110099.9510099.9799.9299.98
299.6699.7499.8799.9199.9499.79
310099.8799.5998.7399.598.49
499.8499.0599.2099.4697.0499.28
597.9996.1199.4299.0598.8799.42
610099.8499.9899.9199.5399.97
799.7899.7299.9999.9599.8599.84
883.9293.9092.8490.4297.2397.33
999.6299.5099.8699.7399.7199.79
1099.3098.3699.3997.8798.4999.31
1199.9898.7097.6097.6797.3697.27
1299.3999.0799.0499.3399.1699.74
1397.1897.599.4999.5695.8698.86
1499.7894.5197.2997.2898.6198.44
1588.5990.6893.8787.5793.2294.40
1699.6498.9610099.9898.7699.99
AA97.79 ± 0.2297.84 ± 0.4698.59 ± 0.1497.9 ± 0.3698.32 ± 0.2898.87 ± 0.30
Kappa93.94 ± 0.4496.38 ± 0.5497.08 ± 0.2895.39 ± 0.8397.7 ± 0.2998.17 ± 0.34
OA94.57 ± 0.3996.75 ± 0.4997.38 ± 0.2595.86 ± 0.7597.93 ± 0.2698.35 ± 0.31
Table 5. Training and testing times of different models for the three HSI data sets.
Table 5. Training and testing times of different models for the three HSI data sets.
Indian PinesKennedy Space CenterSalinas Scene
SVMTrain. (s)0.100.740.80
Test. (s)1.346.3037.55
HybridSNTrain. (s)13.8210.4815.28
Test. (s)0.620.333.60
SSRNTrain. (s)89.2345.1682.11
Test. (s)3.251.4116.94
CDSCNTrain. (s)114.4761.49121.19
Test. (s)4.362.0425.15
MAFNTrain. (s)374.62264.07389.35
Test. (s)10.385.4188.43
3D-CSSEANTrain. (s)60.2033.9164.00
Test. (s)2.321.0514.40
Table 6. OA (%) of the 3D-CSSEAN with different attention modules on the three data sets. Bold represents the best results.
Table 6. OA (%) of the 3D-CSSEAN with different attention modules on the three data sets. Bold represents the best results.
Attention ModuleIndian PinesKennedy Space CenterSalinas Scene
NONE94.20 ± 0.8394.98 ± 0.9596.36 ± 0.98
SPE-EAN96.01 ± 0.4896.85 ± 0.8097.16 ± 1.35
SPA-EAN97.26 ± 0.6997.72 ± 0.6998.22 ± 0.31
3D-CSSEAN97.37 ± 0.5798.64 ± 0.5398.35 ± 0.31
Table 7. Experiment results of the 3D-CSSEAN without or with   t a n h activate function on Indian Pines, Kennedy Space Center (KSC), and Salinas Scene data sets. Bold represents the best results.
Table 7. Experiment results of the 3D-CSSEAN without or with   t a n h activate function on Indian Pines, Kennedy Space Center (KSC), and Salinas Scene data sets. Bold represents the best results.
Data SetAA (without)AA (with)Kappa (without)Kappa (with)OA (without)OA (with)
Indian Pines96.64 ± 1.0697.13 ± 0.8396.36 ± 0.7697.00 ± 0.6596.81 ± 0.6797.37 ± 0.57
KSC96.92 ± 1.3297.74 ± 0.9997.93 ± 0.8398.48 ± 0.5998.15 ± 0.7598.64 ± 0.53
Salinas Scene98.83 ± 0.3398.87 ± 0.3098.03 ± 0.3798.17 ± 0.3498.23 ± 0.3498.35 ± 0.31
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yan, H.; Wang, J.; Tang, L.; Zhang, E.; Yan, K.; Yu, K.; Peng, J. A 3D Cascaded Spectral–Spatial Element Attention Network for Hyperspectral Image Classification. Remote Sens. 2021, 13, 2451. https://doi.org/10.3390/rs13132451

AMA Style

Yan H, Wang J, Tang L, Zhang E, Yan K, Yu K, Peng J. A 3D Cascaded Spectral–Spatial Element Attention Network for Hyperspectral Image Classification. Remote Sensing. 2021; 13(13):2451. https://doi.org/10.3390/rs13132451

Chicago/Turabian Style

Yan, Huaiping, Jun Wang, Lei Tang, Erlei Zhang, Kun Yan, Kai Yu, and Jinye Peng. 2021. "A 3D Cascaded Spectral–Spatial Element Attention Network for Hyperspectral Image Classification" Remote Sensing 13, no. 13: 2451. https://doi.org/10.3390/rs13132451

APA Style

Yan, H., Wang, J., Tang, L., Zhang, E., Yan, K., Yu, K., & Peng, J. (2021). A 3D Cascaded Spectral–Spatial Element Attention Network for Hyperspectral Image Classification. Remote Sensing, 13(13), 2451. https://doi.org/10.3390/rs13132451

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop