Next Article in Journal
Effective Training and Inference Strategies for Point Classification in LiDAR Scenes
Previous Article in Journal
Imagery Time Series Cloud Removal and Classification Using Long Short Term Memory Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral-Spatial Center-Aware Bottleneck Transformer for Hyperspectral Image Classification

1
State Key Laboratory for Strength and Vibration of Mechanical Structures, Xi’an Jiaotong University, Xi’an 710049, China
2
School of Aerospace Engineering, Xi’an Jiaotong University, Xi’an 710049, China
3
School of Automation Science and Engineering, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(12), 2152; https://doi.org/10.3390/rs16122152
Submission received: 24 March 2024 / Revised: 26 May 2024 / Accepted: 10 June 2024 / Published: 13 June 2024

Abstract

:
Hyperspectral image (HSI) contains abundant spectral-spatial information, which is widely used in many fields. HSI classification is a fundamental and important task, which aims to assign each pixel a specific class label. However, the high spectral variability and the limited labeled samples create challenges for HSI classification, which results in poor data separability and makes it difficult to learn highly discriminative semantic features. In order to address the above problems, a novel spectral-spatial center-aware bottleneck Transformer is proposed. First, the highly relevant spectral information and the complementary spatial information at different scales are integrated to reduce the impact caused by the high spectral variability and enhance the HSI’s separability. Then, the feature correction layer is designed to model the cross-channel interactions, thereby promoting the effective cooperation between different channels to enhance overall feature representation capability. Finally, the center-aware self-attention is constructed to model the spatial long-range interactions and focus more on the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel. Experimental results on the common datasets show that compared with the state-of-the-art classification methods, S2CABT has the better classification performance and robustness, which achieves a good compromise between the complexity and the performance.

Graphical Abstract

1. Introduction

Hyperspectral image (HSI) is obtained by continuous sampling on the spectral band using the spectral sensor [1,2]. It has been successfully applied in many fields such as geological examination, urban and rural planning, and agricultural production due to its abundant spectral-spatial information [3,4,5]. HSI classification is a fundamental and important task, which aims to assign each pixel a specific class label [6,7,8]. It is generally composed by the feature extraction and the label prediction. Among them, the feature extraction is used to integrate the spectral-spatial information, while the label prediction is conducted to find the optimal decision boundary according to the extracted features [9].
In the past decades, various HSI classification methods have been proposed [10]. They can be grouped into the non-deep learning (non-DL)-based classification model and the DL-based classification network. The non-DL model was widely applied to HSI classification in earlier works, including the support vector machine (SVM) [11], random forest (RF) [12], and multinomial logistic regression (MLR) [13]. They usually take the original HSI data as the input features. However, the higher spectral dimension makes it difficult for the non-DL models to estimate the optimal model parameters [14]. In recent years, DL has shown enormous development potential in the computer vision tasks, which improves the expressive ability for high-dimensional data. Therefore, the DL networks become the mainstream tool for HSI classification, such as the autoencoder (AE) [15], deep belief network (DBN) [16,17], recurrent neural network (RNN) [18,19,20], and convolutional neural network (CNN) [21,22,23], etc. The DL networks combine the feature extraction and the label prediction in a unified framework, thereby extracting high-level semantic features that are the most beneficial for classification performance during network optimization. Compared with other DL networks, the CNN can learn spectral-spatial features efficiently, so it has become the most popular DL network in HSI classification [24].
The CNN-based classification networks usually require a large number of labeled samples to estimate the optimal network parameters. However, it is difficult to obtain data labels in most cases [25]. The limited labeled samples make it difficult to learn highly discriminative semantic features from the complex spectral-spatial structure. Therefore, optimizing the feature representation learning framework is the key to solving this problem.
In the research on the CNN-based classification network framework, the channel attention plays a crucial role, which can optimize the feature learning and obtain the diverse semantic information from different channels [26]. The channel attentions widely used in HSI classification include the convolutional block attention module (CBAM) [27], selective kernel network (SK-Net) [28], and efficient channel attention network (ECA-Net) [29]. These methods can improve the network’s sensitivity to the important features, but they have certain commonalities and limitations in their implementation approaches. Specifically, the above channel attentions all use the global pooling layer to map the multi-dimensional channel features into a one-dimensional vector along the channel direction. Subsequently, the fully connected layer is used to assign an attention weight to each channel to reflect its relative importance. However, this process inevitably leads to compression and potential loss of some information that is crucial for classification decisions. In addition, the attention weight learned by the above channel attentions only represents the independent importance of a single channel, and fails to fully capture the interactions and synergies between channels. In order to promote the effective cooperation between channels to enhance the overall feature representation capability, this paper reveals the dependencies between channels by modeling the cross-channel interactions. When the given channel has significant interactions with multiple channels, it indicates that this channel contains more valuable information and should be given a higher attention weight to highlight its role. Compared with the traditional channel attention that only focuses on the independent importance of a single channel, the cross-channel interactions can enhance the discriminative ability of different channels more comprehensively, thereby providing powerful feature representation. Therefore, it is necessary to explore a novel channel attention to model the cross-channel interactions and optimize attention allocation on this basis, thereby improving the representation capabilities of the semantic features.
CNN is characterized by local connection, so it is difficult to model the spatial long-range interactions [30], especially for HSI classification task that requires cross-regional information integration. In order to address this problem, researchers explored various more advanced network frameworks [31]. More recently, the vision Transformer based on the multi-head self-attention (MHSA) has achieved great success in computer vision tasks [32]. MHSA can divide the hidden state vector into multiple heads to form multiple sub-semantic spaces, which can allow the network to attend to the information in different dimensional semantic spaces. The self-attention can model the spatial long-range interactions in the global receptive field, thus better capturing the global contextual information, so the Transformer is widely introduced into the CNN-based HSI classification networks to learn the global spectral-spatial properties [33]. As a key component in the Transformer, the self-attention is essentially a spatial attention and it can capture the dependencies between pixels. The HSI classification network takes the 3D patch cropped from the original HSI as the input and learns the high-level spectral-spatial features to predict the class label for the central pixel, so how to better characterize the center pixel is a key to improving classification performance. However, it is difficult for the existing work to capture the contextual information that is beneficial to the central pixel classification decision. Meanwhile, the complex spectral-spatial structure results in that the input 3D patch usually contains some neighboring pixels with different spectral-spatial properties from the central pixel, which will have an impact on the accurate prediction of the class label for the center pixel. Since the existing self-attention cannot measure the different contributions of neighboring pixels to the central pixel, it is difficult to learn the discriminative information for the central pixels from the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel. Therefore, it is necessary to introduce the spectral-spatial property similarities measure to allow the central pixel to learn discriminative features from the neighboring pixels to improve classification performance.
Moreover, in addition to the impact caused by the limited labeled samples on classification performance, the high spectral variability [34] inherent in HSI is also a problem worth considering, which will lead to a poor data separability, thereby increasing uncertainty in identifying the land covers and making it difficult to obtain high-quality training samples. Therefore, increasing the HSI’s separability to provide a highly descriptive feature space for subsequent feature learning is the key to solving this problem.
The existing works usually enhance the HSI’s separability by extracting the highly representative spectral-spatial features [35], which can project the original HSI data into a feature space that integrates spectral-spatial information. Recently, various spectral-spatial feature extraction methods have been proposed, such as the extended morphological profile (EMP) [36], local binary pattern (LBP) [37], and extended morphological attribute profile (EMAP) [38]. The above features deal with HSI in a completely independent manner and do not consider the spectral-spatial correlation [39]. In contrast, the covariance matrix can combine multiple related features to better characterize land covers, and it is widely used in the spectral-spatial feature extraction, in which the element expresses the correlation between two spectral bands in the neighborhood window. However, since the high spectral variability leads to the weak spectral correlations between pixels, using all pixels in the neighborhood window to calculate the covariance matrix will easily introduce some erroneous information. Meanwhile, the covariance matrix is extracted within the neighborhood window with a fixed size, so it is difficult to fully adapt to the complex spectral-spatial structure. For the regions containing the features with multiple scales (such as the homogeneous regions in a small range coexisting with the heterogeneous regions in a large range), it is difficult to simultaneously explore the detailed information and the effective representation for the larger structures. Based on the above analysis, it can be seen that integrating spectral-spatial information from the neighboring pixels with the high spectral correlation at different scales can further increase the HSI’s separability.
In order to enhance the data separability while optimizing the design for the existing channel attentions and the self-attention, a novel Spectral-Spatial Center-Aware Bottleneck Transformer (S2CABT) is proposed. S2CABT constructs a parallel dual-branch structure, which first extracts the multi-scale spectral-spatial features. Then, the heterogeneous convolution kernels are used to separate the spectral and spatial features, and the high-level spectral and spatial features are learned in parallel. Finally, the learned spectral and spatial features are concatenated and a fully connected layer is used to predict the class label. The main contributions of this paper can be summarized as follows:
(1)
The Spectral-Spatial Feature Extraction Module (S2FEM) is designed to extract the Multi-Scale Covariance Matrix (MSCM). It constructs the spectral correlation selection block to remove the neighboring pixels with weak spectral correlation, thereby reducing the impact caused by the high spectral variability and obtaining highly relevant spectral information. Meanwhile, S2FEM introduces the multi-scale analysis to explore the complementary spatial information at different scales and improve the representation power for the land covers.
(2)
A novel channel attention, i.e., the Feature Correction Layer (FCL), is designed to model the cross-channel interactions using the inner products and considers all the information in each channel. Compared with the existing channel attentions, FCL can promote the effective cooperation between channels to enhance overall feature representation capability. Furthermore, FCL is combined with the convolutional layer to construct the Residual Convolution Module (RCM), which can exploit the local spectral-spatial properties.
(3)
The Center-Aware Bottleneck Transformer (CABT) is designed to capture the global spectral-spatial properties, which develops the Center-Aware Multi-Head Self-Attention (CAMHSA) to model the spatial long-range interactions. Meanwhile, CAMHSA introduces the spectral-spatial feature embedding to measure the property similarities between pixels, thereby learning discriminative features that are beneficial to characterizing the central pixel from the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel.
The experimental results on the common datasets show that compared with the state-of-the-art classification methods, S2CABT has the better performance and robustness, which achieves a good compromise between the complexity and the performance.
The rest of this paper is organized as follows. Section 2 introduces the related works. Section 3 presents the S2CABT. Section 4 validates and analyzes the proposed network. Section 5 discusses the S2CABT’s advantages. Section 6 draws the conclusion.

2. Related Works

This section will elaborate on the existing HSI classification networks from two aspects: the CNN in HSI classification and the Transformer in HSI classification.

2.1. CNN in HSI Classification

With the in-depth research on deep learning and related technologies, CNN has gradually been widely used in HSI classification. For example, Hu et al. first proposed a five-layer 1D CNN to extract the spectral features [40]. Li et al. proposed a pixel-level 1D CNN that obtains the pixel-pair spectral features based on the correlation between pixels [41]. In order to integrate the spatial information, Zhao et al. proposed a 2D CNN [42], which takes the patch cropped from the original HSI as the input, thereby capturing the spatial correlation between pixels. Hamida et al. proposed a 3D CNN to extract the joint spectral-spatial features [43]. In order to learn more discriminative spectral and spatial features, the classification network with a dual-branch structure is proposed [44,45]. Guo et al. proposed a serial dual-branch structure [44]. Yu et al. proposed a parallel dual-branch structure [45]. In the serial dual-branch structure, the subsequent feature will be impacted by the previous feature. In contrast, the parallel dual-branch structure can learn spectral and spatial features independently without impacting each other, which can extract more robust features and is more suitable for HSI classification.
The network depth is crucial for computer vision tasks, especially for processing HSI with complex spectral-spatial information. However, over-increasing network depth will lead to degradation. The residual connection is designed to address this problem with a better performance [46], which has become a basic structure in the CNN-based HSI classification networks [47,48,49]. For example, Zhong et al. proposed the spectral and spatial residual blocks to sequentially learn the spectral and spatial features [47]. Paoletti et al. proposed a pyramid residual block, which gradually increases the feature dimension as the network depth increases to learn more semantic features [48]. Zhang et al. partitioned HSI into multiple sub-bands and used several residual blocks to construct the feature extraction stage to extract semantic features on each sub-band [49]. However, the above residual blocks only focus on the feature learning, and ignore the feature optimization.
Since the convolutional layer usually uses multiple convolutional kernels to learn multi-channel features, and different channels can provide different semantic information for HSI classification, researchers usually introduce the channel attention into the residual block to adaptively emphasize the channel that contains the valuable information [50]. For example, Zhu et al. combined the spectral-spatial attention design based on CBAM with the convolutional layers to construct a spectral-spatial residual block [51]. Roy et al. combined the efficient feature recalibration layer design based on ECA-Net with the convolutional layer to construct the different residual blocks in different network depths [52]. Shu et al. combined the spectral and spatial attention layer design based on SK-Net with the convolutional layer to construct the spectral and spatial residual blocks [53]. The existing channel attentions usually use the global pooling layer to compress the multi-channel features into a 1D vector along the channel, and learn an attention weight for each channel using the fully connected layer. Specifically, let the feature size be C × H × W . The existing channel attentions first use the global pooling layer to compress the feature into C × 1 × 1 . Then, the fully connected layer is used to learn the attention weight vector with size C × 1 × 1 . Finally, the attention weight vector is multiplied with the multi-channel feature to obtain the emphasized feature. Each element in the attention weight vector represents the importance of the corresponding channel. Therefore, the existing channel attention easily loses some significant information due to the use of the global pooling layer, and it is difficult for the learned importance to describe the cross-channel interactions.

2.2. Transformer in HSI Classification

CNN is characterized by local connection, so it restricts the receptive field to a local window, which makes it difficult to learn the global semantic feature. As a key component in the Transformer, the self-attention has the global receptive field to better capture global context information, which shows great potential in modeling the spatial long-range interactions. The framework of the self-attention is shown in Figure 1.
Let the input feature be F . The self-attention uses three linear projections to obtain query ( Q ), key ( K ), and value ( V ),
Q = F W Q ,       K = F W K ,       V = F W V
where W Q , W K and W V are the projection matrixes which are trainable. The self-attention can be formulated as
A t t e n t i o n Q , K , V = s o f t m a x Q K T d V
where d is the dimension of the input feature. MHSA is designed based on the self-attention. It first uses W i Q , W i K , and W i V to obtain Q i , K i , and V i in i - t h head. Then, Q i , K i , and V i is used to compute the self-attention according to the framework shown in Figure 1. Finally, the self-attention of all heads is concatenated and the linear projection is performed on the concatenated self-attention. MHSA with h heads can be formulated as
M H S A = C o n c a t A 1 , A 2 , , A h W o
where W o is a learnable projection matrix, C o n c a t · is the concatenate operation. A i is the output of i - t h head, which can be formulated as
A i = A t t e n t i o n Q i , K i , V i
It is currently a mainstream approach to overcome the shortcoming that CNN has difficulty in modeling the spatial long-range interactions by introducing the Transformer into the CNN-based HSI classification networks. For example, Sun et al. proposed the spectral-spatial feature tokenization Transformer (SSFTT) [54]. SSFTT uses the feature tokenizer with Gaussian weights to enhance the semantic information in the low-level spectral-spatial features, and uses the Transformer to model the spatial long-range interactions. Liao et el. proposed the spectral-spatial fusion Transformer network (S2FTNet) [55]. S2FTNet constructed the spectral and spatial Transformer modules and designed the multi-head double self-attention (MHDSA) to model the spatial long-range interactions of the spectral and spatial features, respectively. Ouyang et al. proposed the hybrid Transformer (HF) [56]. HF constructs the multi-granularity feature generator, and designs the spectral-spatial self-attention to focus on the channels and positions with more differences.
According to the above calculation process, it can be seen that the computational complexity of MHSA is affected by the feature channel dimension. In order to balance the computational complexity and performance, Srinivas et al. proposed the bottleneck Transformer [57], which combines MHSA’s modeling capabilities for the spatial long-range interactions with the lightweight design of bottleneck structures. The bottleneck structure was proposed by He et al. [46] in ResNet, as shown in Figure 2a. It contains three convolution layers. The first layer uses 1 × 1 convolution to reduce the channel dimension, and the second layer uses 3 × 3 convolution to extract semantic features, and the third layer uses 1 × 1 convolution to rescale the channel dimension. The bottleneck Transformer replaces the second layer with MHSA, which is shown in Figure 2b. The self-attention in the bottleneck Transformer can be formulated as
A t t e n t i o n Q , K , V = s o f t m a x Q P T + Q K T d V
where P is a relative position encoding, which consists of the relative height encoding R h and the relative width encoding R w .
In recent years, the bottleneck Transformers have also been introduced into the HSI classification. For example, Song et al. introduced the bottleneck Transformer into HSI classification to extract global spectral and spatial features, thereby proposing the spectral-spatial bottleneck Transformer (S2BT) [58]. Zhang et al. designed a multilayer residual convolution module to extract the low-level spectral-spatial features, and input the extracted features into the dual-dimension spectral-spatial bottleneck Transformer to simultaneously model the global channel-spatial attention [59].
Unlike other computer vision tasks, HSI classification requires predicting class labels for the central pixel of input 3D patches. However, the input 3D patch usually contains some neighboring pixels with different spectral-spatial properties from the central pixel, which will have an impact on accurately predicting the class label for the central pixel. It is difficult for MHSA to measure the different contributions of neighboring pixels to the central pixel, making it cannot explore discriminative information from the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel.

3. Spectral-Spatial Center-Aware Bottleneck Transformer

In this section, we first show the overall framework. Then, the main modules in S2CABT are described in detail. Finally, an implementation example is provided.

3.1. Overall Framework

As shown in Figure 3, a novel Spectral-Spatial Center-Aware Bottleneck Transformer (S2CABT) is proposed, which can be divided into four modules: Spectral-Spatial Feature Extraction Module (S2FEM), Residual Convolution Module (RCM), Center-Aware Bottleneck Transformer (CABT), and classification module.
Specifically, since the high spectral variability results in the poor HSI’s separability, S2FEM is constructed to extract the Multi-Scale Covariance Matrix (MSCM). S2FEM uses spectral correlation selection to remove the neighboring pixels with weak spectral correlation and introduces multi-scale analysis to obtain complementary spatial information at different scales. Compared with the original HSI data, MSCM integrates highly relevant spectral information and multi-scale spatial information, so it has stronger separability. In order to build a reasonable network framework to learn high discriminative semantic features from the limited labeled samples, this paper designs RCM and CABT with a parallel dual-branch structure. Specifically, RCM separates the spectral and spatial features from MSCM using two heterogeneous convolutional kernels, and stacks spectral and spatial residual blocks on the spectral and spatial branches to learn the local semantic features, respectively. Among them, the spectral and spatial residual blocks are composed by the convolutional layer and the Feature Correction Layer (FCL). FCL can model the cross-channel interactions to promote the effective cooperation between channels, thereby enhancing overall feature representation capability. Similarly, CABT stacks the center-aware bottleneck Transformer encoders on the spectral and spatial branches to learn the global semantic features, respectively. In the center-aware bottleneck Transformer encoders, the Center-Aware Multi-Head Self-Attention (CAMHSA) is designed to model the spatial long-range interactions. Moreover, CAMHSA introduces the spectral-spatial feature embedding to measure the property similarities between the neighboring pixels to the central pixel. The classification module concatenates the learned high-level spectral feature X S p e O u t and spatial feature X S p a O u t , and predicts the class labels using the fully connected layer, whose calculation process can be formulated as
y ^ i = S o f t m a x F C C o n c a t X S p e O u t , X S p a O u t
where y ^ i denotes the predicted class probabilities, F C · is the full connection layer. C o n c a t · is the concatenation operator. In the next sections, we will detail the principles associated with S2FEM (Section 3.2), RCM (Section 3.3), and CABT (Section 3.4).

3.2. Spectral-Spatial Feature Extraction Module

As shown in Figure 4, in order to reduce the impact cause by the high spectral variability and enhance the HSI’s separability, S2FEM is designed to extract MSCM.
First, in order to remove the redundant spectral bands and decrease the memory requirements, the maximum noise fraction (MNF) [60] is used to generate reduced HSI X R R H × W × R , where R is the reduced spectral dimension. Second, let Ω x s = x 1 , x 2 , , x D s 2 R D s × D s × R be the neighborhood window centered at pixel x 1 with size D s × D s , where D s is the window size at scale s . However, the high spectral variability results in weak spectral correlations between pixels, and calculating the covariance matrix for the central pixel using all neighboring pixels in Ω x s easily introduces some erroneous information. In order to address this problem, we designed the spectral correlation selection block. Specifically, the cosine distance is used to measure the spectral correlations between the central pixel and the neighboring pixel, which can be formulated as
s i m x 1 , x i = x 1 , x i x 1 2 · x i 2 , i = 2 , , D s × D s
where · and · are the inner product and the Frobenius norm, respectively. According to Equation (7), the spectral correlations of all neighboring pixels and the central pixel can be obtained, and the average correlation is calculated. Only those neighboring pixels whose correlation to the central pixel is greater than or equal to the average correlation are retained, and the retained pixels are recorded as x k k = 2,3 , , K . Therefore, a new neighborhood window is constructed, i.e., x k k = 1,2 , 3 , , K . The covariance matrix of the central pixel x 1 is computed using the new neighborhood window, which can be formulated as
C M x 1 = 1 K 1 i = 1 K x i μ x i μ T R R × R
where μ is the average spectral vector of the new neighborhood window. · T is the transpose operation. Then, in order to obtain the complementary spatial information in different scales, this paper extracts the covariance matrix of the central pixel x 1 in the neighborhood window with different sizes according to the above computational process, respectively. The MSCM of the central pixel x 1 can be obtained by concatenating the covariance matrixes at all scales, and its size is 1 × 1 × B c , where B c = N s × R × R , and N s denotes the number of scales. Finally, the HSI’s MSCM can be obtained by combining the MSCMs of all pixels according to the corresponding positions, and its size is H × W × B c , where H and W are the height and width of HSI. In summary, compared with the original HSI, MSCM integrates the highly relevant spectral information and the complementary spatial information at different scales, thereby enhancing the HSI’s separability.

3.3. Residual Convolution Module

As shown in the orange dashed box in Figure 3, in order to model the cross-channel interactions to better describe the local spectral-spatial properties, RCM is constructed based on the convolution layer and the channel attention. Specifically, RCM first crops the 3D patch from MSCM. Then, it uses the 1 × 1 convolution to adjust the feature channel. Finally, the Spectral Residual Convolution Branch (ERCB) and the Spatial Residual Convolution Branch (ARCB) are built to learn local spectral and spatial features, respectively.
As shown in Figure 5, let the size of the input 3D patch be S × S and the adjusted dimension using 1 × 1 convolutional layer be D 1 . First, inspired by [61], ERCB and ARCB use the convolutional kernels of sizes 1 × 1 × 7 and 3 × 3 × D 1 to separate the spectral and spatial features from the input 3D patch. Then, ERCB and ARCB learn local spectral and spatial features by stacking the spectral and spatial residual blocks, which are composed with the convolutional layer, batch normalization, rectified linear unit (ReLU) activation function, and FCL. Among them, the sizes of convolutional kernels in the spectral and spatial residual blocks are 1 × 1 × 7 and 3 × 3 × 1 , respectively. Finally, ERCB and ARCB use the convolutional kernels of sizes 1 × 1 × D 2 and 3 × 3 × 1 to output the local spectral and spatial features, respectively, where D 2 = D 1 6 / 2 .
In order to promote the effective cooperation between channels and enhance overall feature representation capability, FCL is designed to model the cross-channel interactions. Inspired by the self-attention, the inner product is used to compute the interactions between every two channels, which represents the projection of one channel on another channel. The larger the projection means the stronger the interaction between these two channels. As shown in Figure 6, the different colors in the input multiple channel features represent different channels. First, FCL calculates the inner product between every two channels and saves the calculated inner product in the attention matrix according to their channel indexes. Then, the average pooling layer and the S o f t m a x function are used to eliminate the outliers and normalize the attention matrix. Each row in the attention matrix represents the interactions between the corresponding channel and other channels. Finally, the matrix multiplication is performed on the input features and the attention matrix to emphasize the channels that contain valuable information, thereby outputting the enhanced multiple channel features. The FCL’s calculation process can be formulated as
X F C L O U T = F F C L X F C L I N
F F C L X F C L I N = X C A O U T + X F C L I N
X C A O U T = S o f t m a x A v g P o o l X F C L I N · X F C L I N T X F C L I N
where X F C L I N and X F C L O U T are the input and output features. FCL uses the inner product to calculate the cross-channel interactions and considers all information in each channel, which overcomes the drawbacks that traditional channel attention easily loses some significant information and only learns the independent importance of a single channel.
The residual connection is widely used in HSI classification networks to prevent network degradation. Compared with the existing residual blocks, RCM has significant advantages. On the one hand, RCM adopts a dual-branch structure and uses the heterogeneous convolution kernels with different receptive fields to learn the spectral and spatial features respectively, which helps S2CABT more comprehensively understand the HSI’s intrinsic structure. On the other hand, unlike the channel attention used in the existing residual blocks, this paper introduces FCL into RCM, which enables RCM to reveal the interactions and synergies between channels and optimize attention allocation on this basis, thereby improving the ability to characterize the complex spectral-spatial structure.

3.4. Center-Aware Bottleneck Transformer

As shown in the blue dashed box in Figure 3, in order to better describe the global spectral-spatial properties, CABT is designed based on the bottleneck Transformer. Similar to RCM, CABT contains two branches: the Spectral Center-Aware Bottleneck Transformer (ECABT) and the Spatial Center-Aware Bottleneck Transformer (ACABT). ECABT and ACABT first learn global hierarchical features by stacking the Center-Aware Bottleneck Transformer Encoders (CABTE). Then, the global average pooling layer is used to compress features for the subsequent prediction of class labels. As shown in Figure 7, CABTE uses the 1 × 1 convolutional kernel to ensure the consistency of channel. Moreover, CAMHSA can learn discriminative information from the neighboring pixels for the central pixel while modeling the spatial long-range interactions to improve the classification performance, whose framework is shown in Figure 8.
The existing self-attention is usually used to model the spatial long-range interactions, while neglecting to capture the contextual information that is beneficial to the central pixel classification decision. In order to address this problem, this paper introduces the spectral-spatial feature embedding ( E R D × S × S ) to measure the different contributions of neighboring pixels to the central pixel, thereby enabling CAMHSA to focus more on the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel, where D is the feature dimension in each head. Specifically, CAMHSA calculates the similarities between the central pixel and the neighboring pixels in the feature map using the Feature Similarity Measure (FSM) to measure the spectral-spatial properties consistency. Let e 1 R D × 1 × 1 be the central pixel and e i R D × 1 × 1 be the neighboring pixel; the similarity between them can be formulated as
F S M e i = e 1 ,   e i e 1 2 · e i 2 R 1 × S × S ,         i = 1,2 , , S 2
In the center-aware self-attention, we use four linear projection layers to generate Q , K , V and the spectral-spatial feature embedding ( E ) from the input feature. The calculation process of the center-aware self-attention can be formulated as
A t t e n t i o n Q , K , V , E = s o f t m a x Q R h + R w T + Q K T D V · F S C E
where A t t e n t i o n · is the output of the center-aware self-attention. R h and R w denote the relative height encoding and the relative width encoding, respectively. Let the input feature map be F . CAMHSA with h heads can be formulated as
C A M H S A = C o n c a t A 1 , A 2 , , A h W o
where W o is the learnable projection matrix. A i is the output of i - t h head, which can be formulated as
A i = A t t e n t i o n F W i Q , F W i K , F W i V , F W i E
where W i Q , W i K , W i V , and W i E are the learnable projection matrixes in i - t h head, respectively. CAMHSA concatenates the attention calculated by each head and uses a linear projection to obtain the desired attention.
Compared with the original self-attention, the center-aware self-attention designed in this paper can not only can model the spatial long-range interactions, but also can learn the discriminative feature that is beneficial to characterizing the central pixel from the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel, thereby enhancing the S2CABT’s ability to capture the contextual information and interpret the HSI data with a complex spectral-spatial structure.

3.5. Implementation Example

In order to show S2CABT more clearly, this section will take the Indian Pines dataset with size 145 × 145 × 200 as an example to describe the computation process of S2CABT in detail. Let the Indian Pines dataset be X R 145 × 145 × 200 . It is worth noting that all parameters appearing below are assumed to facilitate the presentation of implementation.
In S2FEM, MNF is firstly used to generate the reduced HSI X R R 145 × 145 × 25 . Second, we extract the covariance matrix of pixel x at different scales, and the window sizes at different scales are 3, 15, 27, respectively. When the window size is 3, the neighborhood window with size 3 × 3 × 25 centered at x is selected. The correlations between all pixels and the central pixel are calculated based on the cosine distance, and the pixels with correlation greater than or equal to the average correlation are retained. The x ’s covariance matrix is calculated using the retained pixels according to Equation (8) and its size is reshaped to 1 × 1 × 625 . Then, when the window sizes are 15 and 27, the covariance matrixes are calculated according to the above steps, and they are concatenated to obtain the x ’s MSCM with size 1 × 1 × 1875 . Finally, the MSCMs of all pixels are combined according to the corresponding positions to obtain the whole MSCM with size 145 × 145 × 1875 .
In RCM, we crop the input 3D patch x p R 9 × 9 × 1875 from the MSCM and reshape the input 3D patch to 9 × 9 × 200 using the 1 × 1 convolution. Assume that ERCB and ARCB have two spectral residual blocks and spatial residual blocks, respectively. ECABT and ACABT have two CABTEs, respectively. The implementation of ERCB and ARCB are shown in Table 1 and Table 2, respectively. The implementation of ECABT and ACABT are the same, so they are shown in Table 3. For ease of notation, the convolutional layer, batch normalization, and activation function are abbreviated as Conv-BN-ReLU.
The global spectral and spatial features are concatenated to obtain the fusion feature of size 64 × 1 × 1 and mapped into class probabilities using a fully connected layer. According to the above analysis, the implementation process of the proposed S2CABT is shown in Algorithm 1.
Algorithm 1: Implementation process of S2CABT.
Input: HSI dataset, ground truth, train sample indexes, validation sample indexes, test sample indexes, reduced dimension, multiscale window sizes, patch size, the number of convolutional kernels, batch size, epochs, learning rate.
Output: The predicted class labels on test samples.
1: 
Extract MSCM from the original HSI data.
2: 
Crop training patch, validation patch, and test patch from MSCM. Generate training label, validation label, and test label from the ground truth.
3: 
Create training loader, validation loader, and test loader.
4: 
For  e = 1 to epochs
5:   
Use 2D convolutional kernel to adjust feature dimension.
6:   
Use ERCB and ARCB to learn the local spectral and spatial features.
7:   
Use ECABT and ACABT to learn the global spectral and spatial features.
8:   
Concatenate the learned global spectral and spatial features.
9:   
Input the concatenated features into the classification module to predict class label.
10:    
Compute loss function and update network parameters on training loader.
11:    
Apply the trained network to the validation loader and compute the loss function. If the loss function is smaller than the loss function in the last epoch, the network parameter in this epoch will be saved. Conversely, it will not be saved.
12: 
End for
13: 
Predict class labels on test loader using the best network.

4. Experiments and Analysis

This section first describes the common HSI datasets. Second, the relevant experimental setup is introduced. Then, we discuss some important parameters. Next, S2CABT is compared with 3 non-DL-based models and 13 CNN-based networks. Moreover, the ablation experiments are implemented. Finally, we analyze the complexity of S2CABT.

4.1. Dataset’s Description

In order to validate the S2CABT’s classification performance, the experiments are carried on the Indian Pines, Kennedy Space Center, Houston, and Pavia university datasets.
Indian Pines dataset: This dataset was taken by the AVIRIS on Indian Pines in northwestern Indiana, USA. It contains 145 × 145 pixels and 200 spectral bands. Its ground truth provides 16 land-cover classes for 10,249 labeled pixels.
Kennedy Space Center (KSC) dataset: This dataset was taken by the AVIRIS on Kennedy Space Center in Florida, USA. It contains 512 × 614 pixels and 176 spectral bands. Its ground truth provides 13 land-cover classes for 5211 labeled pixels.
Houston dataset: This dataset was taken by the ITRES CASI-1500 on the University of Houston in Texas, USA. It contains 349 × 1905 pixels and 144 spectral bands. Its ground truth provides 15 land-cover classes for 15,029 labeled pixels.
Pavia University (PaviaU) dataset: This dataset was taken by the ROSIS on Pavia University in Pavia, Italy. It contains 610 × 340 pixels and 103 spectral bands. Its ground truth provides 9 land-cover classes for 42,776 labeled pixels.

4.2. Experimental Setup

Evaluation Metrics: Three common evaluation metrics are used to measure the classification performance: Overall Accuracy (OA), Average Accuracy (AA), and Kappa ( κ ).
Configuration Detail: The training and test sets are selected in the form of percentages. In order to eliminate the effect of randomness, we carry out five repeated experiments. The average classification performance and the standard deviation are reported.
Parameters Setting: For the comparison methods, their parameters are set to the optimal values suggested in the corresponding papers, and the optimal classification performance is achieved on the HSI dataset to the maximum extent possible. Meanwhile, S2CABT is trained using Adam optimizer, and the window sizes at different scales are 3, 15, 27, e p o c h s = 200 . Moreover, other parameters will be discussed in the next section.

4.3. Parameters Discussion

Batch Size: The batch size is the number of training samples input to the classification network in each iteration. A smaller batch size means that the updated gradient in each iteration is calculated based on fewer samples, which increases the randomness of the training process and may speed up the convergence. However, this may also cause the training process to be more volatile, thus affecting the classification performance. A larger batch size reduces this randomness, thereby making the gradient updates smoother and the network convergence process more stable. However, it may fall into a wide valley region due to the excessive smoothness, thereby causing the network to converge to a local minimum, and affecting the classification performance. Therefore, in order to determine an appropriate batch size, this paper validates the impact of the batch size on the classification performance. We select different batch sizes from the set 16,96 with step 16 to carry out the experiments with 2% training samples. The average OA, AA, κ and their standard deviations are shown in Figure 9a–c. As can be seen from Figure 9, the batch size has basically no impact on the classification performances of the Indian Pines, Houston, and PaviaU datasets. In contrast, the impact of the batch size on the classification performance of the KSC dataset is more obvious. When the batch size is 32 or 48, its classification performance is better than when the batch size is set to other values. Considering that the batch size is directly related to the memory occupied by the network parameters in each iteration, this paper sets the batch size of the PaviaU datasets to 32. In order to maintain the consistency, the batch sizes of other three datasets are also set to 32.
Learning rate: The learning rate is the extent to which the network parameters are updated along the gradient direction during the optimization process. A smaller learning rate will make the network parameters gradually approach the optimal solution, but the convergence speed will be very slow and it is easy to fall into a wide local minimum region, thereby making it difficult to obtain the optimal value and affecting the classification performance. A large learning rate will cause the network parameters to decrease too much in each iteration, thereby making it easy to ignore the optimal solution. Meanwhile, a larger learning rate will cause the network parameters to stop updating quickly, thereby making the network unable to mine valuable information and affecting the classification performance. Therefore, in order to determine an appropriate learning rate, this paper validates the impact of the learning rate on the classification performance. We select different learning rates from the set 0.00003 , 0.0001 , 0.0003 , 0.001 , 0.003 , 0.01 to carry out the experiments with 2% training samples. The average OA, AA, κ and their standard deviations are shown in Figure 10a–c. As the learning rate increases, the classification performances of the Indian Pines and KSC datasets gradually increase and then decrease, and reach the optimal classification performance when the learning rate is 0.0003. The classification performance of the KSC dataset fluctuates up and down as the learning rate increases, and it has an optimal value when the learning rate is 0.003. Meanwhile, it can be seen that when the learning rate is 0.0001 or 0.0003, the classification performance of the KSC dataset is very close to the optimal value. In addition, the learning rate has basically no impact on the classification performance of the PaviaU dataset. As the learning rate increases, its classification performance remains basically unchanged. In order to unify the learning rate in the proposed S2CABT corresponding to different datasets as much as possible, this paper sets the learning rate of four datasets to 0.0003.
Reduced spectral dimension: S2FEM uses the MNF to reduce the spectral dimension to remove the redundant spectral bands and decrease the memory requirements. Since the spectral bands contain abundant spectral information, a smaller reduced spectral dimension will not only remove the redundant spectral bands, but also easily remove some bands containing the discriminative spectral information, thereby resulting in the loss of the spectral features and affecting the identification for land covers. A larger reduced spectral dimension easily introduces some redundant spectral information, which makes it more difficult to extract the useful spectral features and increases the computational complexity. Therefore, in order to determine an appropriate reduced spectral dimension, this paper validates the impact of the reduced spectral dimension on the classification performance. We select different reduced spectral dimensions from the set 5 , 50 with step 5 to carry out the experiments with 2% training samples. The average OA, AA, κ and their standard deviation are shown in Figure 11a–c. It can be seen from Figure 11 that as the reduced spectral dimension increases, the classification performances of the four datasets gradually increase. Moreover, when the reduced spectral dimension is equal to 15 or greater, their classification performances remain basically unchanged. In other words, reducing the spectral dimension does lose the spectral features to a certain extent, thereby affecting the classification performance, but this is only limited to the smaller reduced spectral dimension. When the reduced spectral dimension is appropriately increased, the impact of the spectral reduction will become smaller.
Patch size: The patch size relates to the spatial features that can be learned by the classification network. A smaller patch size will lose some important spatial information and the pixels with high correlation to the current pixel, which will result in the unsmooth predicted class label map. A larger patch size will introduce some useless or erroneous spatial information, and increase the computational cost. Therefore, in order to determine an appropriate patch size, this paper validates the impact of the patch size on the classification performance. We select different patch sizes from the set 3 , 21 with step 2 to carry out the experiments with 2% training samples. The average OA, AA, κ and their standard deviations are shown in Figure 12a–c. It can be seen from Figure 12 that as the patch size increases, the classification performances of the Indian Pines and Houston datasets gradually increase and then decrease. The Indian Pines dataset has the optimal classification performance when the patch size is 5. The Houston dataset has essentially the same classification performance when the patch sizes are 7 and 9. Considering the computational complexity, its patch size is set to 7. As the patch size increases, the classification performances of the KSC and PaviaU datasets fluctuate up and down. They reach the optimal classification performances when their patch sizes are 13 and 7, respectively.
Convolutional kernels: the number of convolutional kernels is related to the high-level semantic features that can be learned by the classification network. The fewer convolutional kernels cannot learn enough high-level semantic features, which affects the representational ability of the classification network and leads to underfitting. The more convolutional kernels tend to learn redundant or erroneous high-level semantic features, which leads to overfitting and increases the computational complexity. Therefore, in order to determine an appropriate number of the convolutional kernels, this paper validates the impact of the number of convolutional kernels on the classification performance. We select different numbers of convolutional kernels from the set 4 , 40 with step 4 to carry out the experiments with 2% training samples. The average OA, AA, κ and their standard deviations are shown in Figure 13a–c. It can be seen from Figure 13 the classification performances of the Indian Pines, KSC, and PaviaU datasets gradually increase with the convolutional kernels’ increases, and then fluctuate up and down. The Indian Pines and KSC datasets have the optimal classification performances when the number of convolutional kernels is 20. The PaviaU dataset reaches the optimal classification performance when the number of convolutional kernels is 28. As the convolutional kernels increase, the classification performance of the Houston dataset gradually increases and then decreases. It has an optimal classification performance when the number of convolutional kernels is 28.

4.4. Classification Results

In order to demonstrate the classification performance, our network is compared with 3 non-DL models: SVM [11], RF [12], MLR [13] and 13 CNN networks: ResNet-34 [40], SSRN [47], DPRN [48], A2S2K [52], RSSAN [51], FADCNN [45], SPRN [49], FGSSCA [44], SSFTT [54], BS2T [58], S2FTNet [55], S3ARN [53], HF [56].
Table 4, Table 5, Table 6 and Table 7 show the statistical classification results of different methods on the Indian Pines, KSC, Houston, and PaviaU datasets with 2% training samples, respectively. The best results are in bold. We report the average ± standard deviation on the classification accuracy in each class (CA), OA, AA and κ . It can be seen that our network outperforms the other methods in most CA, all OA, all AA, and all κ on the four datasets. The non-DL-based models perform better than CNN-based networks on classes with fewer training samples. This is because fewer training samples cannot accurately estimate the network parameters, thereby resulting in underfitting. In the classes with more training samples, CNN-based networks perform better than the non-DL-based models. This is because CNN-based networks have strong feature learning capabilities and can predict land-cover classes more accurately using sufficient training samples.
Specifically, for the Indian Pines dataset, the proposed network has the highest CA in 10 classes (16 classes in total). Its evaluation metrics on the Indian Pines dataset are the best, which are 94.80%, 94.20%, and 0.9407, respectively. SPRN has the highest CA in 5 classes, among which it reaches 100% on class 8. Apart from S2CABT, it is the best performing network on CA. SPRN partitions the spectral bands into sub-bands and learns spectral-spatial features on each sub-band. Each sub-band can be regarded as a new dataset, so SPRN increases the available training samples by spectral partitioning. However, since it only uses the convolutional layer to learn the local features and ignores the cross-channel interactions as well as the global features, the features learned by SPRN make it difficult to accurately describe all classes. Therefore, its performance in other classes is not outstanding or even poor. Moreover, its OA, AA, and κ are 4.56%, 4.29%, and 0.0519 different from our network. The closest classification performance to our network is BS2T; its evaluation metrics are 94.01%, 91.46%, and 0.9316, respectively. The differences between S2CABT and BS2T on evaluation metrics are 0.79%, 2.74%, and 0.0091, respectively. BS2T learns the local and global features using CNN and Transformer, respectively, but it ignores the enhancement of semantic features by cross-channel interactions and spectral-spatial properties. Moreover, it does not consider the impact of limited labeled samples. Therefore, its CA reaches 100% only on class 8, and is worse than S2CABT on other classes.
For the KSC dataset, the proposed network has the highest CA in 10 classes (13 classes in total). Its evaluation metrics are the best, which are 99.11%, 98.42%, and 0.9901, respectively. It can be seen that the KSC dataset is more separable than the Indian Pines dataset, especially on classes 10 and 13, where 7 and 10 classification methods reach 100% CA, respectively. The main reason is that the number of training samples of each class in the KSC dataset is more balanced, which avoids overfitting to a certain extent. On this basis, the proposed network extracts more representative multi-scale spectral-spatial features to further enhance the KSC’s separability. The closest classification performance to our network is FGSSCA; its evaluation metrics are 98.36%, 97.43%, and 0.9818, respectively. The differences between S2CABT and FGSSCA on evaluation metrics are 0.75%, 0.99%, and 0.0083, respectively. Similar to SPRN, FGSSCA groups the shallow features and constructs the serial dual-branch structure to learn spectral and spatial features sequentially, thereby increasing the availability of training samples. However, in the serial dual-branch structure, the spatial features will be affected by the spectral features, which makes it difficult to learn highly discriminative semantic features.
For the Houston dataset, S2CABT has the highest CA in 2 classes (15 classes in total). Moreover, the difference in CA between our network and the method with optimal CA is small. For example, the difference between the CA of our network and the optimal CA is only 0.45%, 0.61%, 0.08%, 0.03%, and 0.17% on the classes 2, 3, 5, 7, and 11. The evaluation metrics of the proposed network are the best, which are 94.24%, 94.35%, and 0.9377, respectively. The closest classification performance to our network is FGSSCA; its evaluation metrics are 94.04%, 93.96%, and 0.9355, respectively. The differences between S2CABT and FGSSCA on the evaluation metrics are 0.2%, 0.39%, and 0.0022, respectively. For the PaviaU dataset, S2CABT has the highest CA in 2 classes (9 classes in total). The difference in CA between our network and the method with optimal CA is small. For example, the difference between the CA of our network and the optimal CA is only 0.26%, 0.03%, 1.14%, 0.4%, 0.09%, and 0.54% on classes 1, 2, 4, 5, 6, and 8. The evaluation metrics of the proposed network are the best, which are 99.35%, 98.75%, and 0.9914, respectively. The closest classification performance to our network is SPRN; its evaluation metrics are 99.27%, 98.61%, and 0.9903, respectively. The differences between S2CABT and SPRN on evaluation metrics are 0.08%, 0.14%, and 0.0011, respectively.
The visual classification performance of different methods on the Indian Pines dataset with 2% training samples is shown in Figure 14. Overall, the CNN-based networks perform better than the non-DL-based models. The main reason is that CNN-based networks usually take the 3D patch as the input feature, and comprehensively consider the spectral and spatial information. They combine the feature extraction and the label prediction in the unified framework, and can dynamically learn more discriminative high-level semantic features based on the loss function. The non-DL-based models take 1D spectral vectors as input features, which only consider the spectral information and ignore the spatial information, thereby resulting in many misclassified pixels in the class label map. They generally take the input feature directly as the classification basis, which makes it difficult to estimate the optimal model parameters. In contrast, S2CABT can obtain a smoother class label map and predict the labels for the most labeled samples more accurately.
Specifically, the land covers are closer to each other in the Indian Pines dataset, so the feature extraction and the label prediction are easily affected by the adjacent land covers. The Indian Pines dataset has a large spatial resolution (20 m), which leads to the higher spectral variability. Since the land covers in the Indian Pines dataset have size differences, the number of training samples of different classes are not balanced. Both the spectral variability and the unbalanced training samples will create difficulties in estimating the optimal network parameters. As shown in Figure 14, most methods are unable to accurately classify the land cover and recognize the image edges. The non-DL-based models and a few CNN-based networks (e.g., ResNet and DPRN) produce many discrete predicted labels due to the fact that the non-DL-based models ignore the spatial information, while ResNet and DPRN are underfitting because the training samples are too few. The remaining CNN-based networks are prone to over-smoothness, i.e., the land covers with different classes are recognized as the same class. The main reason is that these networks use original HSI data as the input feature, which makes it difficult for them to learn spectral-spatial features with high discriminative power from the original HSI data with low separability. In contrast, S2CABT enhances the data separability by integrating the highly relevant spectral information and the complementary spatial information at different scales. Meanwhile, it optimizes the designs for the channel attention and the self-attention to learn more discriminative semantic features, which can enhance the S2CABT’s ability to interpret the HSI data with the complex spectral-spatial structure and capture the contextual information, thereby better describing the essential properties of land covers.
In order to compare the classification performances of different methods on four datasets more comprehensively, we use different training sizes 2%, 4%, 6%, 8%, 10%, 15%, and 20% to complete the classification task, and the experimental results are shown in Figure 15. It can be seen that the classification performances of non-DL-based models outperform some CNN-based networks when the training size is small. For example, the non-DL-based models are better than DPRN, RSSAN, FADCNN in the Indian Pines dataset, better than RSSAN, SSFTT, S2FTNet in the KSC dataset, better than DPRN, RSSAN, S3ARN in the Houston dataset, and better than FADCNN in the PaviaU dataset. As the training size increases, the CNN-based networks gradually perform better than the non-DL-based models. This is mainly because more training samples allow the CNN-based networks to learn highly discriminative semantic features, thereby improving the classification performance. Different from the comparison methods, S2CABT extracts MSCM with stronger separability and builds the more efficient feature learning module, so it can still achieve exciting classification performance with smaller training size, and it is better than other methods. As the training size increases, the difference between S2CABT and other comparison methods gradually decreases, but our network is still optimal.
The classification’s robustness is easily affected by many factors such as the sample quality and the feature representation capacity. In order to improve the S2CABT’s robustness, this paper projects HSI into a feature space with higher descriptive power by integrating the highly relevant spectral information and the complementary spatial information at different scales, thereby increasing the data separability and providing high-quality training samples for subsequent feature learning. In addition, this paper designs FCL and CAMHSA to improve the feature representation learning capacity. Among them, FCL can model the cross-channel interactions and optimize attention allocation on this basis. CAMHSA introduces the spectral-spatial property similarity measure to allow the central pixel to learn discriminative information from the neighboring pixels.
It can be seen from Figure 15 that S2CABT still maintains better classification performance even with fewer training samples. Meanwhile, as the training size increases, the S2CAB’s classification performance gradually becomes better. Moreover, compared with other classification methods, S2CAB can converge faster. These phenomena all show that our network has better robustness and stability.

4.5. Ablation Study

In order to validate the impact of the important contributions in S2CABT on the classification performance, we designed an ablation study. In this experiment, the impact of the important contributions will be analyzed from three aspects: the ablation study on the input feature, the ablation study on the framework, and the ablation study on the attention.
Ablation study on the input feature: In order to reduce the impact caused by the high spectral variability and enhance the HSI’S separability, MSCM is extracted to reconstruct the input feature space. MSCM integrates the highly relevant spectral information and the complementary spatial information at different scales using the Spectral Correlation Selection (SCS) and the spatial multi-scale analysis, respectively. In this experiment, we compare the classification performance when the original HSI, single-scale covariance matrix, MSCM without SCS, and MSCM are used as input feature. Since we use three scales to extract MSCM, the window sizes at different scales are 3, 15, and 27, respectively. Therefore, different window sizes, i.e., D 1 = 3 , D 2 = 15 , and D 3 = 27 need to be considered when using the single-scale covariance matrix as the input feature. In order to analyze the differences between input features more intuitively, T-SNE is used to project the input features into a two-dimensional space and visualize them. As shown in Figure 16, in each subgraph, the different colors represent the samples with different classes. Moreover, the experimental results on four HSI datasets are shown in Table 8, and the best results are in bold. We report the average ± standard deviation on OA, AA, and κ .
As can be seen from Figure 16, affected by the high spectral variability, the samples with different class labels overlap with each other in the original HSI data. The distribution of samples with the same class label is relatively divergent, thereby making it difficult for the classification network to use the training samples obtained from the original HSI data to learn the optimal parameters. Meanwhile, it can be seen from Table 8 that when the original HSI data is used as the input features, its classification performance is poor. In contrast, there are clear decision boundaries between the samples with different class labels in the MSCM, and the distribution of samples with the same class label is also more concentrated. Moreover, the classification performance when using the MSCM as the input feature is significantly better than that when using the original HSI data as the input feature. According to the above analysis, it can be seen that the separability of input features will directly affect the classification performance. In order to validate whether the spectral feature is lost when S2FEM uses the spectral correlation selection to remove the neighboring pixels with low spectral correlation, this section visualizes the MSCM without SCS. It can be seen from the visualization results that using all pixels in the neighborhood window to calculate the covariance matrix for the central pixel will lead to larger intra-class distances and smaller inter-class distances. Therefore, the introduction of the spectral correlation selection block not only does not lose the spectral features, but also can obtain the highly discriminative spectral information, thereby improving classification performance. Meanwhile, the multi-scale analysis enriches the feature representation to a certain extent. The small-scale covariance matrix can capture the fine features, while the large-scale covariance matrix can focus more on the overall representation. It can be seen from the visualization results of three single-scale covariance matrixes that it is difficult to obtain valuable information from the Indian Pines dataset on the small-scale region. However, in the large-scale region, a feature space with high descriptive power can be constructed for the Indian Pines dataset. The KSC, Houston, and PaviaU datasets can obtain effective feature representations at different scales, especially in the small-scale region, where they can extract spectral-spatial features with stronger separability.
Ablation study on the framework: In order to comprehensively characterize the spectral-spatial properties, the residual convolution module and the center-aware bottleneck Transformer are designed. In addition, the parallel dual-branch structure is built to learn robust spectral and spatial features. The feature learning module can be divided into four sub-modules: ERCB, ARCB, ECABT, and ACABT. ERCB and ARCB are constructed based on CNN to learn the local spectral and spatial features, respectively. ECABT and ACABT are designed based on the bottleneck Transformer to learn the global spectral and spatial features, respectively. In this experiment, we remove one or more sub-modules to validate the impact of different sub-modules on classification performance. The experimental results are shown in Table 9 and the best results are in bold. It is worth noting that ✘ means the module is removed, while ✓ means the module is not removed. We report the average ± standard deviation on OA, AA, and κ . It can be seen that each sub-module plays a vital role in improving classification performance. After removing one or more modules, the classification performance is degraded to varying degrees.
Specifically, this experiment can be divided into three groups: removing one sub-module, removing two sub-modules, and removing all sub-modules. On the KSC, Houston, and PaviaU datasets, the classification performance after removing one sub-module is better than other experiments. The classification performance after removing all sub-modules is worse than the other experiments. On the Indian Pines dataset, overall, the classification performance after removing one sub-module is better than other experiments. However, AA after removing two sub-modules is worse than AA after removing all modules. This phenomenon is manifested in the removal of the spectral or spatial branches, i.e., removing both ERCB and ECABT or removing both ARCB and ACABT. The main reason is that the spectral or spatial features are lost after removing a branch, which makes it difficult to accurately describe spectral-spatial properties. Especially for the land covers that only provide one training sample, removing one branch makes these classes have smaller CA, which has less effect on OA and κ but more effect on AA.
Ablation study on the attention: In order to optimize feature extraction and measure the different contributions of neighboring pixels to the central pixel, FCL and CAMHSA are designed. Among them, FCL uses the inner product between channels to model the cross-channel interactions, which suppresses the drawbacks that the existing channel attention easily loses some important information and only learns the independent importance for a single channel. CAMHSA introduces the feature similarity measure based on MHSA to focus more on the neighboring pixels that have relatively consistent spectral-spatial properties with the central pixel. In this experiment, in order to validate the advantages of FCL and CAMHSA, we replace FCL and CAMHSA with the existing channel attention (CBAM, SK-Net, and ECA-Net, etc.) and MHSA, respectively. The experimental results are shown in Table 10; the best results are in bold. We report the average ± standard deviation on OA, AA, and κ . It can be seen from Table 10 that both FCL and CAMHSA improve the classification performance. Compared with the existing channel attention and MHSA, they demonstrate stronger feature optimization and learning capabilities.

4.6. Network Complexity Analysis

In order to describe the proposed network more comprehensively, the network complexity analysis experiments are designed in this section. Table 11 shows the complexity on four datasets of the CNN-based classification networks involved in this paper from four aspects: the trainable parameters that need to be updated during the backpropagation process, the computational cost, the training time, and the testing time, respectively. The trainable parameters are in M , and the computational cost is in G F L O P s , and the training and testing times are in s . Among them, 1 M = 10 6 , 1 G F L O P s = 10 9 F L O P s .
As can be seen from Table 11, although this paper constructs the parallel dual-branch structure and designs four sub-modules to learn highly discriminative spectral-spatial features, the network complexity of our network is not the highest. This is due to the smaller convolution kernel used in the residual convolution module. In addition, the learned semantic features compress the spectral dimension, which provides an ideal feature size for subsequent feature learning. In the center-aware bottleneck Transformer, the proposed network uses 1 × 1 convolution to adjust the channels, thereby reducing the network complexity. However, there are still some networks that have a smaller network complexity than our network. It is worth noting that the difference in complexity between our network and these networks is small, but the classification performance of our network is far better than these networks. This demonstrates that S2CABT does not rely on the excessive increases in the trainable parameters, the computational costs, the training and testing times to improve classification performance. In other words, S2CABT achieves a good compromise between network complexity and classification performance.

5. Discussion

This section mainly discusses the significant advantages of S2CABT compared to the existing state-of-the-art classification networks from the following aspects:
(1)
Network framework: The significant advantages in the network framework are reflected in the reconstruction for the input feature space and the optimization for the feature representation learning. Among them, S2FEM integrates highly relevant spectral information and multi-scale complementary spatial information. Compared with the state-of-the-art classification networks that use original HSI as the input features, S2CABT can increase the HSI’s separability to provide a highly descriptive feature space for subsequent feature learning stages.
Unlike the residual blocks used in existing classification networks [47,48,49,51,52,53], RCM processes the spectral and spatial information in parallel, thereby improving computational efficiency. Meanwhile, RCM introduces FCL to reveal the interactions and synergies between different channels and optimize the attention allocation, thereby enhancing the overall feature representation ability.
CABT is a novel bottleneck Transformer, which develops the CAMHSA to model the spatial long-range interactions. CAMHSA is different from the self-attentions used in the state-of-the-art classification networks [54,55,56,57,58,59]. It introduces spectral-spatial feature embedding to measure the property differences between pixels, thereby capturing the contextual information that is beneficial to the central pixel classification decision.
(2)
Classification performance: This paper carries out the comparison experiments between S2CABT and 13 existing state-of-the-art classification networks on four common HSI datasets, and reports the average evaluation metrics and the corresponding standard deviations. The experimental results show that S2CABT effectively addresses some key challenges in HSI classification through the innovative design. It shows better classification performance and robustness, and outperforms the existing state-of-the-art classification networks.
(3)
Network complexity: This paper compares S2CABT with 13 existing state-of-the-art classification networks on the network complexity and running time. The introduction of the novel modules inevitably increases the complexity and running time, but it is still better than some existing classification networks. In other words, S2CABT achieves a good compromise between the network complexity and the classification performance.

6. Conclusions

This paper proposes a novel HSI classification network framework to reduce the impact caused by the high spectral variability and the limited labeled samples. The high spectral variability leads to poor data separability, so this paper designs the spectral correlation selection block to remove the neighboring pixels with low spectral correlation. Meanwhile, the complementary spatial information at different scales is integrated. The limited labeled samples make it difficult for the classification network to effectively learn feature representations, so this paper designs FCL and CAMHSA to optimize feature learning. FCL is used to model the interactions and synergies between channels to enhance the overall feature representation capability. CAMHSA is used to model the spatial long-range interactions while capturing the contextual information that is beneficial to the central pixel classification decisions.
This paper validates the proposed S2CABT on the common HSI datasets. The experimental results show that compared with the state-of-the-art classification methods, S2CABT has the better classification performance and robustness, which achieves a good compromise between the complexity and the performance.
Although S2CABT has achieved satisfactory performance, in order to consolidate and improve the current research results, we will continue to optimize the HSI classification network from the following aspects in the future:
(1)
We will design more efficient feature extraction methods to improve the ability to represent the complex spectral-spatial structure.
(2)
We will continue to explore the novel channel attention to enable the classification network to dynamically adjust the attention weights according to the input content and the feature distribution to achieve more refined feature learning.
(3)
Considering the prominent position of the central pixel in the HSI classification, we will continue to research the interactions between pixels at different scales, thereby enhancing the feature representation learning for the central pixel.
In addition, this work mainly focuses on the HSI classification tasks. In order to expand its application scope, applying S2CABT to other remote sensing image processing tasks such as the object detection and scene classification is also an important direction for our future work. For the object detection tasks, we will integrate the cross-channel interaction and the spatial long-range attention into the object detection frameworks to enhance their detection capacities for the small or complex distributed objects. In the scene classification tasks, since scenes usually involve wider context information and complex structures, the attention mechanism in S2CABT can be used to emphasize semantic features that are crucial to the scene’s understanding. By adjusting the network structure, such as adding the self-attention that handles the larger scene contexts, the cross-regional semantic features can be better extracted, thereby providing more precise clues when distinguishing the different scene classes.

Author Contributions

Conceptualization, M.Z., Y.Y. and D.H.; methodology, M.Z. and Y.Y.; software, M.Z., S.Z. and P.M.; validation M.Z.; writing—original draft preparation, M.Z.; writing—review and editing, Y.Y. and D.H.; funding acquisition, Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number U22A2045.

Data Availability Statement

The Indian Pines and KSC datasets can be obtained at https://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes. The Houston dataset can be obtained at https://hyperspectral.ee.uh.edu/?page_id=459 (all accessed on 23 March 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sun, H.; Wang, L.; Liu, H.; Sun, Y. Hyperspectral image classification with the orthogonal self-attention ResNet and two-step support vector machine. Remote Sens. 2024, 16, 1010. [Google Scholar] [CrossRef]
  2. Yang, J.; Qin, J.; Qian, J.; Li, A.; Wang, L. AL-MRIS: An active learning-based multipath residual involution siamese network for few-shot hyperspectral image classification. Remote Sens. 2024, 16, 990. [Google Scholar] [CrossRef]
  3. Guo, H.; Liu, W. S3L: Spectrum Transformer for self-supervised learning in hyperspectral image classification. Remote Sens. 2024, 16, 970. [Google Scholar] [CrossRef]
  4. Cui, B.; Wen, J.; Song, X.; He, J. MADANet: A lightweight hyperspectral image classification network with multiscale feature aggregation and a dual attention mechanism. Remote Sens. 2023, 15, 5222. [Google Scholar] [CrossRef]
  5. Diao, Q.; Dai, Y.; Wang, J.; Feng, X.; Pan, F.; Zhang, C. Spatial-pooling-based graph attention U-Net for hyperspectral image classification. Remote Sens. 2024, 16, 937. [Google Scholar] [CrossRef]
  6. Zhao, F.; Zhang, J.; Meng, Z.; Liu, H.; Chang, Z.; Fan, J. Multiple vision architectures-based hybrid network for hyperspectral image classification. Expert Syst. Appl. 2023, 234, 121032. [Google Scholar] [CrossRef]
  7. Islam, T.; Islam, R.; Uddin, P.; Ulhaq, A. Spectrally segmented-enhanced neural network for precise land cover object classification in hyperspectral imagery. Remote Sens. 2024, 16, 807. [Google Scholar] [CrossRef]
  8. Shi, C.; Wu, H.; Wang, L. A positive feedback spatial-spectral correlation network based on spectral slice for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503417. [Google Scholar] [CrossRef]
  9. Shi, C.; Yue, S.; Wang, L. A dual branch multiscale Transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5504520. [Google Scholar] [CrossRef]
  10. Liu, S.; Li, H.; Jiang, C.; Feng, J. Spectral–spatial graph convolutional network with dynamic-synchronized multiscale features for few-shot hyperspectral image classification. Remote Sens. 2024, 16, 895. [Google Scholar] [CrossRef]
  11. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  12. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  13. Haut, J.; Paoletti, M.; Paz-Gallardo, A.; Plaza, J.; Plaza, A.; Vigo-Aguiar, J. Cloud implementation of logistic regression for hyperspectral image classification. In Proceedings of the 17th International Conference on Computational and Mathematical Methods in Science and Engineering (CMMSE), Costa Ballena (Rota), Cádiz, Spain, 4–8 July 2017. [Google Scholar]
  14. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef]
  15. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  16. Li, T.; Zhang, J.; Zhang, Y. Classification of hyperspectral image based on deep belief networks. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014. [Google Scholar]
  17. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  18. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
  19. Ma, A.; Filippi, A.; Wang, Z.; Yin, Z. Hyperspectral image classification using similarity measurements-based deep recurrent neural networks. Remote Sens. 2019, 11, 194. [Google Scholar] [CrossRef]
  20. Seydgar, M.; Alizadeh Naeini, A.; Zhang, M.; Li, W.; Satari, M. 3-D convolution-recurrent networks for spectral-spatial classification of hyperspectral images. Remote Sens. 2019, 11, 883. [Google Scholar] [CrossRef]
  21. Paoletti, M.E.; Moreno-Álvarez, S.; Xue, Y.; Haut, J.M.; Plaza, A. AAtt-CNN: Automatical attention-based convolutional neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5511118. [Google Scholar] [CrossRef]
  22. Hu, Y.; Tian, S.; Ge, J. Hybrid convolutional network combining multiscale 3D depthwise separable convolution and CBAM residual dilated convolution for hyperspectral image classification. Remote Sens. 2023, 15, 4796. [Google Scholar] [CrossRef]
  23. Li, S.; Liang, L.; Zhang, S.; Zhang, Y.; Plaza, A.; Wang, X. End-to-end convolutional network and spectral-spatial Transformer architecture for hyperspectral image classification. Remote Sens. 2024, 16, 325. [Google Scholar] [CrossRef]
  24. Fang, S.; Li, X.; Tian, S.; Chen, W.; Zhang, E. Multi-level feature extraction networks for hyperspectral image classification. Remote Sens. 2024, 16, 590. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Gao, D.; Liu, D.; Shi, G. Spectral-spatial domain attention network for hyperspectral image few-shot classification. Remote Sens. 2024, 16, 592. [Google Scholar] [CrossRef]
  26. Zhou, H.; Luo, F.; Zhuang, H.; Weng, Z.; Gong, X.; Lin, Z. Attention multi-hop graph and multi-scale convolutional fusion network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5508614. [Google Scholar]
  27. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  28. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  29. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  30. Ma, Y.; Lan, Y.; Xie, Y.; Yu, L.; Chen, C.; Wu, Y.; Dai, X. A Spatial–Spectral Transformer for Hyperspectral Image Classification Based on Global Dependencies of Multi-Scale Features. Remote Sens. 2024, 16, 404. [Google Scholar] [CrossRef]
  31. Arshad, T.; Zhang, J. Hierarchical attention transformer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5504605. [Google Scholar] [CrossRef]
  32. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  33. Zhang, Z.; Wang, S.; Zhang, W. Dilated spectral–spatial Gaussian Transformer net for hyperspectral image classification. Remote Sens. 2024, 16, 287. [Google Scholar] [CrossRef]
  34. Zhang, G.; Hu, X.; Wei, Y.; Cao, W.; Yao, H.; Zhang, X.; Song, K. Nonlocal correntropy matrix representation for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5502305. [Google Scholar] [CrossRef]
  35. Jia, S.; Zhu, Z.; Shen, L.; Li, Q. A two-stage feature selection framework for hyperspectral image classification using few labeled samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1023–1035. [Google Scholar] [CrossRef]
  36. Beirami, B.A.; Mokhtarzade, M. Band grouping SuperPCA for feature extraction and extended morphological profile production from hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1953–1957. [Google Scholar] [CrossRef]
  37. Fang, Y.; Ye, Q.; Sun, L.; Zheng, Y.; Wu, Z. Multi-attention joint convolution feature representation with lightweight Transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5513814. [Google Scholar] [CrossRef]
  38. Wei, Y.; Zheng, Y.; Wang, R.; Li, H. Quaternion convolutional neural network with EMAP representation for multisource remote sensing data classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5508805. [Google Scholar] [CrossRef]
  39. Torti, E.; Marenzi, E.; Danese, G.; Plaza, A.; Leporati, F. Spatial-spectral feature extraction with local covariance matrix from hyperspectral images through hybrid parallelization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 7412–7421. [Google Scholar] [CrossRef]
  40. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  41. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2016, 55, 844–853. [Google Scholar] [CrossRef]
  42. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  43. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  44. Guo, W.; Ye, H.; Cao, F. Feature-grouped network with spectral–spatial connected attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5500413. [Google Scholar] [CrossRef]
  45. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.-I. Feedback attention-based dense CNN for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501916. [Google Scholar] [CrossRef]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  47. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  48. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep pyramidal residual networks for spectral–spatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 740–754. [Google Scholar] [CrossRef]
  49. Zhang, X.; Shang, S.; Tang, X.; Feng, J.; Jiao, L. Spectral partitioning residual network with spatial attention mechanism for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5507714. [Google Scholar] [CrossRef]
  50. Gao, H.; Zhang, Y.; Chen, Z.; Li, C. A multiscale dual-branch feature fusion and attention network for hyperspectral images classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8180–8192. [Google Scholar] [CrossRef]
  51. Zhu, M.; Jiao, L.; Liu, F.; Yang, S.; Wang, J. Residual spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 449–462. [Google Scholar] [CrossRef]
  52. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-based adaptive spectral–spatial kernel ResNet for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7831–7843. [Google Scholar] [CrossRef]
  53. Shu, Z.; Liu, Z.; Zhou, J.; Tang, S.; Yu, Z.; Wu, X.-J. Spatial–spectral split attention residual network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 419–430. [Google Scholar] [CrossRef]
  54. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization Transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  55. Liao, D.; Shi, C.; Wang, L. A spectral–spatial fusion Transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5515216. [Google Scholar] [CrossRef]
  56. Ouyang, E.; Li, B.; Hu, W.; Zhang, G.; Zhao, L.; Wu, J. When multigranularity meets spatial–spectral attention: A hybrid Transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4401118. [Google Scholar]
  57. Srinivas, A.; Lin, T.-Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck Transformers for visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  58. Song, R.; Feng, Y.; Cheng, W.; Mu, Z.; Wang, X. BS2T: Bottleneck spatial–spectral Transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5532117. [Google Scholar] [CrossRef]
  59. Zhang, L.; Wang, Y.; Yang, L.; Chen, J.; Liu, Z.; Bian, L.; Yang, C. D 2 S 2 BoT: Dual-dimension spectral-spatial bottleneck Transformer for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 2655–2669. [Google Scholar] [CrossRef]
  60. Green, A.A.; Berman, M.; Switzer, P.; Craig, M.D. A transformation for ordering multispectral data in terms of image quality with implications for noise removal. IEEE Trans. Geosci. Remote Sens. 1988, 26, 65–74. [Google Scholar] [CrossRef]
  61. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
Figure 1. Framework of the self-attention.
Figure 1. Framework of the self-attention.
Remotesensing 16 02152 g001
Figure 2. Different bottleneck structures. (a) Bottleneck structure in ResNet; (b) bottleneck Transformer encoder.
Figure 2. Different bottleneck structures. (a) Bottleneck structure in ResNet; (b) bottleneck Transformer encoder.
Remotesensing 16 02152 g002
Figure 3. Framework of S2CABT.
Figure 3. Framework of S2CABT.
Remotesensing 16 02152 g003
Figure 4. Framework of S2FEM.
Figure 4. Framework of S2FEM.
Remotesensing 16 02152 g004
Figure 5. Framework of ERCB and ARCB. (a) ERCB; (b) ARCB.
Figure 5. Framework of ERCB and ARCB. (a) ERCB; (b) ARCB.
Remotesensing 16 02152 g005
Figure 6. Framework of FCL.
Figure 6. Framework of FCL.
Remotesensing 16 02152 g006
Figure 7. Framework of CABTE.
Figure 7. Framework of CABTE.
Remotesensing 16 02152 g007
Figure 8. Framework of CAMHSA.
Figure 8. Framework of CAMHSA.
Remotesensing 16 02152 g008
Figure 9. The impact of the batch size on the classification performance. (a) OA; (b) AA; (c) κ .
Figure 9. The impact of the batch size on the classification performance. (a) OA; (b) AA; (c) κ .
Remotesensing 16 02152 g009
Figure 10. The impact of the learning rate on the classification performance. (a) OA; (b) AA; (c) κ .
Figure 10. The impact of the learning rate on the classification performance. (a) OA; (b) AA; (c) κ .
Remotesensing 16 02152 g010
Figure 11. The impact of the reduced spectral dimension on the classification performance. (a) OA; (b) AA; (c) κ .
Figure 11. The impact of the reduced spectral dimension on the classification performance. (a) OA; (b) AA; (c) κ .
Remotesensing 16 02152 g011
Figure 12. The impact of the patch size on the classification performance. (a) OA; (b) AA; (c) κ .
Figure 12. The impact of the patch size on the classification performance. (a) OA; (b) AA; (c) κ .
Remotesensing 16 02152 g012
Figure 13. The impact of the number of convolutional kernels on the classification performance. (a) OA; (b) AA; (c) κ .
Figure 13. The impact of the number of convolutional kernels on the classification performance. (a) OA; (b) AA; (c) κ .
Remotesensing 16 02152 g013
Figure 14. Visual classification results for Indian Pines dataset. (a) Ground truth; (br) the class label maps predicted by the SVM, RF, MLR, ResNet, SSRN, DPRN, A2S2K, RSSAN, FADCNN, SPRN, FGSSCA, SSFTT, BS2T, S2FTNet, S3ARN, HF, and S2CABT.
Figure 14. Visual classification results for Indian Pines dataset. (a) Ground truth; (br) the class label maps predicted by the SVM, RF, MLR, ResNet, SSRN, DPRN, A2S2K, RSSAN, FADCNN, SPRN, FGSSCA, SSFTT, BS2T, S2FTNet, S3ARN, HF, and S2CABT.
Remotesensing 16 02152 g014
Figure 15. Statistical classification results of different methods on the Indian Pines, KSC, Houston, and PaviaU datasets under different training sizes: 2%, 4%, 6%, 8%, 10%, 15%, and 20%: (ac) the OA, AA, and κ of the Indian Pines dataset; (df) the OA, AA, and κ of the KSC dataset; (gi) the OA, AA, and κ of the Houston dataset; (jl) the OA, AA, and κ of the PaviaU dataset.
Figure 15. Statistical classification results of different methods on the Indian Pines, KSC, Houston, and PaviaU datasets under different training sizes: 2%, 4%, 6%, 8%, 10%, 15%, and 20%: (ac) the OA, AA, and κ of the Indian Pines dataset; (df) the OA, AA, and κ of the KSC dataset; (gi) the OA, AA, and κ of the Houston dataset; (jl) the OA, AA, and κ of the PaviaU dataset.
Remotesensing 16 02152 g015
Figure 16. T-SNE visualization of different input features. The first to sixth rows are the original HSI, three single-scale covariance matrixes, MSCM without SCS and MSCM, respectively. (a) The Indian Pines dataset, (b) the KSC dataset, (c) the Houston dataset, (d) the PaviaU dataset.
Figure 16. T-SNE visualization of different input features. The first to sixth rows are the original HSI, three single-scale covariance matrixes, MSCM without SCS and MSCM, respectively. (a) The Indian Pines dataset, (b) the KSC dataset, (c) the Houston dataset, (d) the PaviaU dataset.
Remotesensing 16 02152 g016aRemotesensing 16 02152 g016b
Table 1. The implementation of ERCB.
Table 1. The implementation of ERCB.
Layer NameKernel SizeOutput Size
Input- 9 × 9 × 200
3D Conv. 1 × 1 × 7 32 × 9 × 9 × 97
3D Conv-BN-ReLU 1 × 1 × 7 32 × 9 × 9 × 97
FCL- 32 × 9 × 9 × 97
Residual Connection- 32 × 9 × 9 × 97
3D Conv-BN-ReLU 1 × 1 × 7 32 × 9 × 9 × 97
FCL- 32 × 9 × 9 × 97
Residual Connection- 32 × 9 × 9 × 97
3D Conv. 1 × 1 × 97 32 × 9 × 9 × 1
Squeeze- 32 × 9 × 9
Table 2. The Implementation of ARCB.
Table 2. The Implementation of ARCB.
Layer NameKernel SizeOutput Size
Input- 9 × 9 × 200
3D Conv. 3 × 3 × 200 32 × 9 × 9 × 1
3D Conv-BN-ReLU 3 × 3 × 1 32 × 9 × 9 × 1
FCL- 32 × 9 × 9 × 1
Residual Connection- 32 × 9 × 9 × 1
3D Conv-BN-ReLU 3 × 3 × 1 32 × 9 × 9 × 1
FCL- 32 × 9 × 9 × 1
Residual Connection- 32 × 9 × 9 × 1
3D Conv. 3 × 3 × 1 32 × 9 × 9 × 1
Squeeze- 32 × 9 × 9
Table 3. The Implementation of ECABT and ACABT.
Table 3. The Implementation of ECABT and ACABT.
Layer NameKernel SizeOutput Size
Input- 32 × 9 × 9
2D Conv-BN-ReLU 1 × 1 12 × 9 × 9
CAMHSA- 12 × 9 × 9
2D Conv-BN-ReLU 1 × 1 32 × 9 × 9
Residual Connection- 32 × 9 × 9
2D Conv-BN-ReLU 1 × 1 12 × 9 × 9
CAMHSA- 12 × 9 × 9
2D Conv-BN-ReLU 1 × 1 32 × 9 × 9
Residual Connection- 32 × 9 × 9
Global Average Pooling- 32 × 1 × 1
Table 4. Statistical classification results of different methods on Indian Pines dataset.
Table 4. Statistical classification results of different methods on Indian Pines dataset.
ClassTrainTestSVMRFMLRResNetSSRNDPRNA2S2KRSSANFADCNNSPRNFGSSCASSFTTBS2TS2FTNetS3ARNHFS2CABT
114513.78 ± 7.603.11 ± 4.872.67 ± 3.658.89 ± 8.1717.78 ± 12.478.89 ± 17.4376.00 ± 15.5112.89 ± 6.9221.78 ± 22.5878.67 ± 30.5064.44 ± 12.8684.89 ± 12.8086.67 ± 4.1624.00 ± 22.3652.89 ± 8.8044.89 ± 11.9195.56 ± 6.29
229139961.77 ± 2.4648.25 ± 6.1758.13 ± 3.4869.12 ± 8.6287.91 ± 2.2748.91 ± 17.5187.99 ± 3.2859.70 ± 10.5367.91 ± 20.3187.13 ± 17.9892.75 ± 2.0989.75 ± 1.1893.03 ± 3.0889.55 ± 1.8788.84 ± 4.4791.87 ± 2.3893.35 ± 3.02
31781347.18 ± 5.9929.37 ± 11.4234.46 ± 6.5556.80 ± 12.7078.23 ± 10.1630.36 ± 19.5382.27 ± 8.3340.17 ± 9.3458.13 ± 9.1169.69 ± 31.6088.14 ± 6.7186.79 ± 8.5391.00 ± 3.2490.65 ± 5.6290.95 ± 4.7187.50 ± 8.9691.51 ± 8.03
4523232.41 ± 10.9119.40 ± 6.9216.47 ± 10.497.84 ± 5.5468.45 ± 10.2213.54 ± 15.3080.26 ± 16.1815.26 ± 4.5717.84 ± 8.1493.11 ± 8.3586.73 ± 11.3586.47 ± 13.3892.07 ± 11.1581.47 ± 11.3653.53 ± 25.0371.81 ± 12.6688.97 ± 13.21
51047374.97 ± 4.2463.38 ± 6.9563.47 ± 7.2155.35 ± 14.4389.05 ± 3.4449.85 ± 29.0290.28 ± 1.9160.68 ± 7.1341.18 ± 23.2193.40 ± 2.8291.08 ± 3.8389.51 ± 1.7993.23 ± 1.2390.82 ± 2.2689.18 ± 4.8189.77 ± 4.6693.49 ± 2.75
61571589.20 ± 1.6386.24 ± 6.6690.91 ± 2.2391.02 ± 10.0497.26 ± 1.7176.73 ± 21.8498.07 ± 1.3588.36 ± 8.9469.40 ± 28.3597.98 ± 1.3995.36 ± 3.1797.42 ± 1.3795.97 ± 1.9298.35 ± 0.5593.12 ± 2.2794.46 ± 2.8897.90 ± 1.60
712746.67 ± 11.891.48 ± 3.315.93 ± 5.623.70 ± 5.2414.07 ± 8.450.00 ± 0.0092.59 ± 10.8016.29 ± 10.0114.07 ± 19.3195.55 ± 4.0689.63 ± 8.4585.93 ± 11.2388.15 ± 17.0541.48 ± 35.8573.33 ± 16.2371.11 ± 18.78100.00 ± 0.00
81046898.42 ± 0.7098.08 ± 2.4094.53 ± 4.3174.74 ± 11.6499.96 ± 0.0943.50 ± 30.1499.87 ± 0.1997.10 ± 2.9068.04 ± 22.47100.00 ± 0.0099.83 ± 0.2899.70 ± 0.41100.00 ± 0.0099.96 ± 0.0994.49 ± 5.5999.87 ± 0.29100.00 ± 0.00
911933.68 ± 19.197.37 ± 8.815.26 ± 6.454.21 ± 4.404.21 ± 9.414.21 ± 6.8660.00 ± 18.0816.84 ± 6.8616.84 ± 14.6168.42 ± 30.0150.52 ± 18.8378.95 ± 15.3564.21 ± 29.6341.05 ± 28.9270.53 ± 25.4170.53 ± 20.9290.53 ± 9.41
102095257.79 ± 6.4642.67 ± 6.8047.42 ± 4.3356.37 ± 8.8481.05 ± 4.9961.07 ± 11.8482.37 ± 6.0253.42 ± 4.7766.79 ± 7.6184.81 ± 3.1883.17 ± 2.8582.80 ± 4.3885.31 ± 4.3483.82 ± 3.5078.02 ± 2.2284.45 ± 2.8586.41 ± 6.07
1150240568.70 ± 4.3178.30 ± 7.0070.37 ± 2.6886.29 ± 4.0489.35 ± 4.4873.71 ± 12.9892.16 ± 2.4374.90 ± 2.8485.55 ± 4.4489.32 ± 9.3695.63 ± 2.2991.24 ± 5.1295.30 ± 3.8393.91 ± 2.9790.54 ± 3.4996.06 ± 2.9797.58 ± 2.38
121258138.18 ± 4.0820.86 ± 9.1528.57 ± 5.8432.01 ± 17.1982.75 ± 7.4322.65 ± 14.1686.20 ± 9.6934.08 ± 14.0730.05 ± 11.7093.60 ± 4.2888.78 ± 4.3576.70 ± 5.4792.67 ± 3.5382.89 ± 6.1066.16 ± 7.3987.37 ± 4.8689.12 ± 7.25
13520090.80 ± 3.2188.30 ± 5.1391.70 ± 8.2871.80 ± 8.1797.00 ± 3.8650.20 ± 23.5699.20 ± 1.7998.40 ± 2.3356.10 ± 26.4799.60 ± 0.6597.50 ± 0.7999.50 ± 0.7198.40 ± 2.1698.60 ± 1.7896.10 ± 2.5395.30 ± 3.6599.30 ± 1.10
1426123989.43 ± 3.7590.69 ± 7.7990.17 ± 4.2289.56 ± 6.3996.63 ± 4.0583.07 ± 7.2594.71 ± 4.4495.00 ± 1.9582.87 ± 9.6699.28 ± 0.6498.13 ± 1.7397.51 ± 3.0498.34 ± 1.8698.27 ± 2.1696.01 ± 2.8397.50 ± 1.2198.92 ± 1.24
15837835.08 ± 6.7323.60 ± 4.1038.10 ± 4.4936.40 ± 31.5883.60 ± 10.9030.10 ± 35.7790.16 ± 9.8230.63 ± 4.3636.08 ± 16.3994.60 ± 6.7089.95 ± 11.2089.95 ± 2.9297.78 ± 1.4093.23 ± 8.4279.52 ± 14.8190.00 ± 6.7298.20 ± 2.87
1629183.96 ± 2.5348.13 ± 16.7576.92 ± 11.0710.55 ± 3.5376.49 ± 22.024.18 ± 3.3398.02 ± 3.3318.90 ± 4.6213.85 ± 11.9593.40 ± 6.2186.59 ± 8.1386.59 ± 4.2191.21 ± 4.0388.13 ± 6.3335.16 ± 19.8157.36 ± 23.7486.37 ± 12.30
OA (%)66.99 ± 1.5461.58 ± 1.2663.29 ± 0.9468.88 ± 1.2387.67 ± 2.4756.67 ± 11.7590.22 ± 1.9165.74 ± 2.8866.22 ± 5.9990.24 ± 1.8492.61 ± 0.7890.46 ± 1.7694.01 ± 0.7991.61 ± 1.3086.80 ± 0.9491.62 ± 0.9894.80 ± 0.64
AA (%)60.13 ± 2.1846.83 ± 1.7550.94 ± 1.0447.17 ± 2.1472.74 ± 3.4137.56 ± 11.1088.13 ± 2.9650.79 ± 2.0746.65 ± 7.1889.91 ± 5.5087.39 ± 1.0788.98 ± 1.2091.46 ± 2.1681.01 ± 4.8678.02 ± 2.7483.12 ± 1.1194.20 ± 1.07
κ 0.6223 ± 0.01710.5529 ± 0.01410.5763 ± 0.01170.6375 ± 0.01710.8590 ± 0.02820.4976 ± 0.13480.8885 ± 0.02200.6066 ± 0.03230.6096 ± 0.07230.8888 ± 0.02140.9155 ± 0.00880.8911 ± 0.01980.9316 ± 0.00890.9042 ± 0.01490.8491 ± 0.010990.42 ± 1.10900.9407 ± 0.0069
Table 5. Statistical classification results of different methods on the KSC dataset.
Table 5. Statistical classification results of different methods on the KSC dataset.
ClassTrainTestSVMRFMLRResNetSSRNDPRNA2S2KRSSANFADCNNSPRNFGSSCASSFTTBS2TS2FTNetS3ARNHFS2CABT
11674590.26 ± 5.6390.47 ± 5.1390.87 ± 5.0899.46 ± 0.5699.36 ± 0.9099.17 ± 0.5299.14 ± 0.7378.87 ± 44.0999.44 ± 0.81100.00 ± 0.00100.00 ± 0.0098.20 ± 2.8799.87 ± 0.3097.61 ± 1.1899.33 ± 0.51100.00 ± 0.00100.00 ± 0.00
2523876.22 ± 5.2170.59 ± 6.0484.79 ± 6.8585.55 ± 12.4492.69 ± 5.2684.45 ± 16.8890.00 ± 10.0619.33 ± 16.3162.86 ± 21.6193.36 ± 7.7995.55 ± 8.3530.59 ± 6.8698.07 ± 2.7264.96 ± 9.5778.91 ± 24.9493.87 ± 3.9598.22 ± 2.79
3625073.28 ± 6.1773.68 ± 6.3382.72 ± 4.8590.88 ± 6.5698.08 ± 1.6377.60 ± 12.6792.88 ± 7.6247.04 ± 32.1867.36 ± 16.9599.28 ± 1.0099.04 ± 1.7148.48 ± 37.9196.32 ± 8.0190.72 ± 6.9767.92 ± 32.8492.56 ± 12.1396.50 ± 7.00
4624643.82 ± 7.3644.72 ± 6.2537.72 ± 10.0041.06 ± 21.0462.76 ± 10.6754.39 ± 11.7588.94 ± 3.8916.26 ± 9.0043.98 ± 12.6889.02 ± 13.4087.48 ± 13.3429.19 ± 9.6182.44 ± 7.5330.49 ± 22.5143.74 ± 19.7373.01 ± 4.8597.05 ± 5.90
5415750.06 ± 10.8848.54 ± 3.9446.88 ± 14.8729.04 ± 17.9179.24 ± 15.1719.36 ± 20.7481.78 ± 16.9421.53 ± 15.8651.21 ± 18.7289.81 ± 8.2791.34 ± 9.3536.30 ± 12.1090.57 ± 8.8027.64 ± 15.9547.90 ± 15.0784.59 ± 10.6989.81 ± 9.69
6522436.34 ± 10.6725.80 ± 14.5547.05 ± 5.6239.91 ± 31.9791.79 ± 8.0136.07 ± 24.7998.04 ± 3.4335.09 ± 25.5573.66 ± 11.4399.02 ± 1.4299.29 ± 0.7520.62 ± 26.0899.46 ± 1.2025.80 ± 19.9972.14 ± 26.6796.70 ± 0.60100.00 ± 0.00
7310275.69 ± 15.6665.29 ± 17.5277.65 ± 16.8062.16 ± 31.3993.14 ± 4.1070.59 ± 9.4894.90 ± 8.8157.06 ± 32.9183.53 ± 13.3996.08 ± 8.2399.22 ± 1.7547.65 ± 29.7798.43 ± 3.512.74 ± 3.8282.94 ± 15.6392.55 ± 7.36100.00 ± 0.00
8942270.76 ± 13.5144.93 ± 7.0863.03 ± 7.0380.71 ± 12.7695.59 ± 4.7682.56 ± 6.8797.63 ± 4.2737.25 ± 21.7179.90 ± 22.9298.77 ± 2.3898.01 ± 3.8261.66 ± 17.9498.58 ± 3.0568.34 ± 15.6797.35 ± 3.1498.67 ± 1.0899.11 ± 0.78
91150985.19 ± 6.4282.04 ± 10.4288.17 ± 6.2293.36 ± 4.9799.68 ± 0.4994.34 ± 7.1098.94 ± 2.3791.71 ± 6.4692.85 ± 7.4999.65 ± 0.7999.37 ± 1.4084.95 ± 6.69100.00 ± 0.0089.70 ± 5.4096.42 ± 7.3598.23 ± 3.95100.00 ± 0.00
10939583.29 ± 14.5067.95 ± 8.1888.35 ± 6.1799.19 ± 1.67100.00 ± 0.0096.96 ± 4.34100.00 ± 0.0026.73 ± 19.3997.06 ± 2.78100.00 ± 0.00100.00 ± 0.0071.75 ± 4.08100.00 ± 0.0081.17 ± 3.6999.54 ± 0.63100.00 ± 0.00100.00 ± 0.00
11941089.51 ± 3.5692.49 ± 3.0091.12 ± 5.6296.10 ± 3.3398.15 ± 3.6292.44 ± 9.3398.29 ± 3.8274.93 ± 41.9992.98 ± 6.4099.12 ± 1.9698.63 ± 3.0584.88 ± 10.0898.54 ± 2.4897.27 ± 4.0392.05 ± 11.3398.54 ± 3.2798.97 ± 1.29
121149288.94 ± 2.8179.23 ± 6.4792.93 ± 3.1295.12 ± 2.7798.41 ± 1.8193.54 ± 5.0998.86 ± 0.9827.88 ± 16.9271.95 ± 27.7298.78 ± 1.7098.70 ± 1.5362.48 ± 9.1798.94 ± 1.6284.11 ± 14.6176.66 ± 20.8499.67 ± 0.7399.80 ± 0.41
131990899.98 ± 0.0599.71 ± 0.1799.89 ± 0.25100.00 ± 0.00100.00 ± 0.0099.93 ± 0.15100.00 ± 0.0076.74 ± 41.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0098.55 ± 1.71100.00 ± 0.0099.89 ± 0.19100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
OA (%)81.51 ± 2.0476.24 ± 1.5783.13 ± 1.0987.27 ± 1.8195.84 ± 0.8286.48 ± 1.5797.23 ± 0.7455.56 ± 25.0685.01 ± 5.3598.36 ± 0.6298.36 ± 0.6172.09 ± 4.3598.18 ± 0.4279.89 ± 2.5987.96 ± 8.9596.84 ± 0.6799.11 ± 0.12
AA (%)74.10 ± 1.8668.11 ± 2.1076.24 ± 1.3577.89 ± 2.9392.99 ± 1.5477.03 ± 2.4095.34 ± 1.7046.96 ± 22.1278.21 ± 4.9097.15 ± 0.8697.43 ± 0.5859.64 ± 6.4397.02 ± 0.4166.19 ± 2.9881.15 ± 12.1094.49 ± 1.1298.42 ± 0.28
κ 0.7940 ± 0.02250.7349 ± 0.01730.8121 ± 0.01210.8574 ± 0.02050.9536 ± 0.00920.8488 ± 0.01760.9691 ± 0.00830.5014 ± 0.28050.8330 ± 0.05930.9818 ± 0.00690.9818 ± 0.00680.69 ± 0.05010.9797 ± 0.00460.7750 ± 0.02910.8653 ± 0.10080.9648 ± 0.00740.9901 ± 0.0014
Table 6. Statistical classification results of different methods on the Houston dataset.
Table 6. Statistical classification results of different methods on the Houston dataset.
ClassTrainTestSVMRFMLRResNetSSRNDPRNA2S2KRSSANFADCNNSPRNFGSSCASSFTTBS2TS2FTNetS3ARNHFS2CABT
126122593.03 ± 7.1493.06 ± 2.6694.81 ± 2.8693.96 ± 2.3095.79 ± 2.3193.60 ± 3.6397.60 ± 1.7077.70 ± 3.5394.55 ± 1.5797.49 ± 1.8598.66 ± 0.8797.49 ± 0.5397.06 ± 1.8596.47 ± 2.1773.58 ± 42.2497.55 ± 1.6793.63 ± 3.08
226122888.36 ± 5.5988.50 ± 4.0192.04 ± 5.4587.98 ± 2.3595.05 ± 5.0285.65 ± 5.0693.34 ± 4.3582.25 ± 5.0385.80 ± 5.3692.72 ± 3.6992.47 ± 7.5496.29 ± 5.2093.55 ± 5.6793.19 ± 3.4161.52 ± 35.7493.32 ± 4.1895.84 ± 2.13
31468361.58 ± 41.4291.83 ± 1.7899.50 ± 0.2297.92 ± 2.4999.88 ± 0.2699.38 ± 0.8399.97 ± 0.0782.46 ± 4.8798.36 ± 1.9299.82 ± 0.1999.93 ± 0.1198.86 ± 1.4499.68 ± 0.2299.59 ± 0.8479.47 ± 44.4398.83 ± 0.8799.36 ± 0.84
425121993.58 ± 2.6893.57 ± 2.8293.75 ± 2.7990.62 ± 4.7497.49 ± 2.7089.68 ± 2.9696.54 ± 1.9578.93 ± 8.8185.76 ± 11.0598.10 ± 2.5599.72 ± 0.2997.80 ± 2.4697.43 ± 3.0297.87 ± 2.7775.78 ± 41.5896.13 ± 3.5991.42 ± 3.62
525121797.52 ± 1.6690.45 ± 5.5197.50 ± 0.7898.47 ± 1.56100.00 ± 0.0099.57 ± 0.3899.29 ± 1.1392.31 ± 4.4599.06 ± 1.0499.98 ± 0.04100.00 ± 0.0098.96 ± 2.32100.00 ± 0.00100.00 ± 0.0067.16 ± 42.97100.00 ± 0.0099.92 ± 0.14
6731881.70 ± 2.0580.57 ± 4.5883.08 ± 0.9882.96 ± 4.2593.84 ± 4.3873.84 ± 28.2494.21 ± 3.2751.07 ± 13.7274.78 ± 18.9089.94 ± 4.5988.68 ± 0.0092.08 ± 5.0490.50 ± 5.1691.70 ± 6.0772.83 ± 18.4088.74 ± 2.1691.40 ± 5.84
726124278.45 ± 8.6679.34 ± 3.6886.22 ± 4.4679.52 ± 6.2290.52 ± 2.4971.71 ± 11.0189.66 ± 3.4877.25 ± 4.4870.45 ± 5.3494.86 ± 4.5694.45 ± 3.5393.01 ± 3.0295.65 ± 3.2095.17 ± 2.8643.59 ± 41.8686.78 ± 4.3795.62 ± 2.62
825121951.11 ± 10.3565.23 ± 7.0054.39 ± 3.5964.40 ± 7.5176.13 ± 7.3955.96 ± 2.6375.95 ± 5.3548.97 ± 6.7168.86 ± 7.7979.70 ± 6.2676.29 ± 0.4776.49 ± 9.5176.16 ± 6.9676.28 ± 10.1740.95 ± 36.3379.08 ± 3.9889.17 ± 4.81
926122678.48 ± 2.7569.58 ± 5.0766.05 ± 5.1551.97 ± 19.9075.34 ± 5.5950.91 ± 8.5884.68 ± 8.8358.42 ± 9.2269.04 ± 10.0985.04 ± 3.8789.52 ± 5.1381.03 ± 3.3684.54 ± 6.2585.87 ± 5.7746.05 ± 41.9480.41 ± 4.1986.71 ± 5.75
1025120228.35 ± 18.3771.26 ± 2.8577.40 ± 4.7169.62 ± 8.3288.29 ± 11.0072.76 ± 14.6787.59 ± 9.2164.36 ± 10.8878.20 ± 13.3097.77 ± 1.3497.09 ± 4.1294.64 ± 4.8198.59 ± 1.9296.96 ± 4.6449.07 ± 47.0792.28 ± 4.1295.37 ± 1.13
1125121067.97 ± 4.74570.25 ± 8.2068.38 ± 4.0270.99 ± 7.7390.55 ± 6.2067.87 ± 10.7592.51 ± 7.0363.36 ± 9.2583.16 ± 4.4497.72 ± 0.94100.00 ± 0.0097.12 ± 3.8496.11 ± 4.1099.06 ± 1.3674.59 ± 41.9195.34 ± 2.8099.83 ± 0.29
1225120823.76 ± 7.4959.24 ± 4.3953.96 ± 5.9670.58 ± 17.0086.89 ± 3.6567.67 ± 12.5183.58 ± 9.8049.97 ± 11.3477.10 ± 7.8887.24 ± 6.3589.12 ± 9.7892.86 ± 4.0389.98 ± 6.5490.37 ± 3.0656.74 ± 40.6493.26 ± 4.3891.17 ± 3.57
13104595.23 ± 5.1812.77 ± 5.0420.65 ± 4.9052.11 ± 8.4086.62 ± 5.7647.32 ± 14.0785.45 ± 7.0043.92 ± 13.2372.42 ± 10.6481.74 ± 9.2783.44 ± 5.5492.46 ± 2.4887.23 ± 3.1987.93 ± 1.5546.80 ± 42.4080.70 ± 2.8887.29 ± 3.40
14941989.50 ± 8.8290.84 ± 2.6994.18 ± 4.4599.62 ± 0.3299.95 ± 0.1198.57 ± 1.27100.00 ± 0.0080.00 ± 8.4694.08 ± 10.0099.95 ± 0.11100.00 ± 0.0097.52 ± 5.1599.95 ± 0.11100.00 ± 0.0057.57 ± 51.7399.38 ± 0.90100.00 ± 0.00
151464699.04 ± 0.2584.55 ± 2.6098.67 ± 0.1499.97 ± 0.07100.00 ± 0.0099.97 ± 0.07100.00 ± 0.0078.98 ± 7.5995.20 ± 5.88100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0073.93 ± 42.27100.00 ± 0.0098.45 ± 1.21
OA (%)69.84 ± 2.8477.39 ± 0.8379.10 ± 1.1379.67 ± 2.3790.83 ± 0.6977.46 ± 2.6491.20 ± 1.4869.54 ± 1.8882.55 ± 2.4093.45 ± 0.6994.04 ± 1.0593.31 ± 1.3593.50 ± 0.9293.72 ± 0.7560.38 ± 34.2091.96 ± 0.8794.24 ± 0.71
AA (%)69.18 ± 3.0776.07 ± 0.7578.71 ± 0.9880.71 ± 2.3691.75 ± 0.6478.30 ± 3.5592.02 ± 1.3968.66 ± 2.3883.12 ± 1.5993.47 ± 0.4993.96 ± 0.6293.77 ± 1.1793.76 ± 1.0294.03 ± 0.8661.31 ± 34.1192.12 ± 0.7094.35 ± 0.55
κ 0.6732 ± 0.03090.7552 ± 0.00890.7738 ± 0.01220.7800 ± 0.02570.9009 ± 0.00740.7561 ± 0.02860.9049 ± 0.01600.6702 ± 0.02040.8113 ± 0.02590.9292 ± 0.00750.9355 ± 0.01130.9277 ± 0.01460.9297 ± 0.00990.9321 ± 0.00810.5698 ± 0.37240.9130 ± 0.00940.9377 ± 0.0078
Table 7. Statistical classification results of different methods on the PaviaU dataset.
Table 7. Statistical classification results of different methods on the PaviaU dataset.
ClassTrainTestSVMRFMLRResNetSSRNDPRNA2S2KRSSANFADCNNSPRNFGSSCASSFTTBS2TS2FTNetS3ARNHFS2CABT
1133649892.71 ± 2.0588.04 ± 1.8289.51 ± 2.0991.79 ± 2.9997.40 ± 0.9991.83 ± 2.1698.62 ± 0.3492.82 ± 2.4387.74 ± 4.7799.75 ± 0.3099.77 ± 0.1999.46 ± 0.3898.98 ± 0.4399.56 ± 0.2197.08 ± 1.2098.98 ± 0.4399.51 ± 0.55
23731827698.90 ± 0.3396.49 ± 0.4595.63 ± 0.2799.38 ± 0.3799.91 ± 0.1199.51 ± 0.3099.88 ± 0.0998.32 ± 0.8299.54 ± 0.3099.99 ± 0.02100.00 ± 0.0199.96 ± 0.0499.99 ± 0.0199.98 ± 0.0299.85 ± 0.1599.99 ± 0.0199.97 ± 0.03
342205725.65 ± 12.5148.66 ± 5.3168.49 ± 2.2476.84 ± 10.2588.63 ± 3.6776.71 ± 4.5587.18 ± 5.0765.12 ± 6.4675.19 ± 13.2294.16 ± 2.7591.94 ± 3.7096.84 ± 1.5294.58 ± 1.1595.09 ± 1.7890.97 ± 5.1894.58 ± 1.1598.27 ± 1.85
461300385.66 ± 1.5982.38 ± 3.3580.09 ± 3.3190.44 ± 3.6798.17 ± 0.7388.81 ± 3.3698.04 ± 1.4095.57 ± 1.6890.02 ± 6.5897.44 ± 0.5198.25 ± 1.0796.68 ± 1.2697.39 ± 1.1597.81 ± 1.3898.65 ± 0.6697.39 ± 1.1597.51 ± 0.87
527131899.17 ± 0.3597.75 ± 0.7798.74 ± 0.7199.98 ± 0.04100.00 ± 0.0099.98 ± 0.04100.00 ± 0.0099.12 ± 1.1899.19 ± 1.0599.77 ± 0.1499.92 ± 0.1399.77 ± 0.3999.71 ± 0.44100.00 ± 0.0099.89 ± 0.1599.71 ± 0.4499.60 ± 0.44
6101492845.17 ± 7.6449. 90 ± 4.0772.49 ± 1.3690.79 ± 2.2798.13 ± 1.2488.56 ± 5.6698.96 ± 1.2489.65 ± 8.4598.38 ± 1.94100.00 ± 0.00100.00 ± 0.0099.65 ± 0.49100.00 ± 0.0099.82 ± 0.2899.52 ± 0.46100.00 ± 0.0099.91 ± 0.14
72713030.05 ± 0.0766.31 ± 8.1352.52 ± 7.2688.20 ± 8.0393.14 ± 5.0474.75 ± 4.7597.37 ± 2.3767.80 ± 5.0173.10 ± 28.1399.94 ± 0.1099.65 ± 0.2399.25 ± 0.9699.23 ± 1.3699.98 ± 0.0492.39 ± 1.7599.23 ± 1.3699.98 ± 0.04
874360890.86 ± 1.4882.47 ± 1.7285.05 ± 0.9971.29 ± 12.9290.69 ± 2.7268.70 ± 6.3492.77 ± 1.5878.23 ± 9.8775.24 ± 3.3197.90 ± 0.6798.32 ± 0.9996.80 ± 0.9795.49 ± 2.3097.08 ± 1.7397.20 ± 1.0895.49 ± 2.3097.78 ± 1.02
91992899.76 ± 0.1299.27 ± 0.2699.72 ± 0.1082.24 ± 10.7299.53 ± 0.1577.80 ± 17.6099.63 ± 0.5297.61 ± 1.4377.05 ± 6.9898.58 ± 0.9198.36 ± 0.6296.72 ± 0.1696.96 ± 1.2997.16 ± 2.2699.20 ± 0.6796.96 ± 1.2996.21 ± 2.02
OA (%)83.35 ± 0.9784.30 ± 0.4987.45 ± 0.2692.32 ± 1.1697.62 ± 0.4591.26 ± 0.6898.13 ± 0.2491.95 ± 2.8692.27 ± 1.9299.27 ± 0.1799.25 ± 0.2099.09 ± 0.2198.90 ± 0.2599.19 ± 0.1098.39 ± 0.5198.90 ± 0.2599.35 ± 0.12
AA (%)70.88 ± 1.1279.03 ± 1.3182.47 ± 1.0887.88 ± 1.5896.18 ± 0.8885.18 ± 1.4796.94 ± 0.5787.14 ± 3.5486.16 ± 4.3098.61 ± 0.3098.47 ± 0.4798.35 ± 0.3598.04 ± 0.4398.50 ± 0.2097.20 ± 0.8598.04 ± 0.4398.75 ± 0.25
κ 0.7707 ± 0.01430.7865 ± 0.00750.8316 ± 0.00370.8976 ± 0.01530.9684 ± 0.00600.8829 ± 0.00940.9752 ± 0.00320.8928 ± 0.03830.8973 ± 0.02580.9903 ± 0.00220.9900 ± 0.00270.9879 ± 0.00280.9854 ± 0.00330.9893 ± 0.00130.9786 ± 0.00680.9854 ± 0.00330.9914 ± 0.0015
Table 8. Statistical classification results of ablation study on the input feature.
Table 8. Statistical classification results of ablation study on the input feature.
DatasetMetricsOriginal HSI Single   Scale   D 1 Single   Scale   D 2 Single   Scale   D 3 Without SCSMSCM
Indian PinesOA (%)91.90 ± 1.4181.94 ± 1.0994.26 ± 1.1593.93 ± 0.6991.36 ± 0.8294.80 ± 0.64
AA (%)77.47 ± 1.6371.16 ± 1.4892.73 ± 1.9992.88 ± 1.4888.97 ± 1.9194.20 ± 1.07
κ 0.9074 ± 0.01600.7924 ± 0.01210.9345 ± 0.01310.9307 ± 0.00780.9015 ± 0.00920.9407 ± 0.0069
KSCOA (%)97.30 ± 1.3298.08 ± 1.0298.21 ± 0.6598.83 ± 0.6298.13 ± 0.9399.11 ± 0.12
AA (%)96.07 ± 1.9696.98 ± 1.5897.07 ± 0.6898.10 ± 0.9197.15 ± 1.3998.42 ± 0.28
κ 0.9699 ± 0.01470.9786 ± 0.01140.9801 ± 0.00720.9870 ± 0.00690.9792 ± 0.01040.9901 ± 0.0014
HoustonOA (%)90.85 ± 0.8191.42 ± 0.6892.64 ± 0.7792.21 ± 1.4490.38 ± 0.8494.24 ± 0.71
AA (%)91.42 ± 0.6192.05 ± 0.6392.90 ± 0.6192.75 ± 1.3791.10 ± 0.7294.35 ± 0.55
κ 0.9011 ± 0.00860.9072 ± 0.00730.9205 ± 0.00830.9158 ± 0.01550.8960 ± 0.00910.9377 ± 0.0078
PaviaUOA (%)98.05 ± 0.2799.16 ± 0.2298.91 ± 0.1699.20 ± 0.2399.14 ± 0.2099.35 ± 0.12
AA (%)96.69 ± 0.8698.68 ± 0.3798.17 ± 0.4598.20 ± 0.7297.96 ± 0.8898.75 ± 0.25
κ 0.9742 ± 0.00360.9888 ± 0.00290.9856 ± 0.00210.9894 ± 0.00300.9886 ± 0.00270.9914 ± 0.0015
Table 9. Statistical classification results of ablation study on the framework.
Table 9. Statistical classification results of ablation study on the framework.
ERCB
ARCB
ECABT
ACABT
Indian PinesOA (%)94.41 ± 0.6994.48 ± 0.6794.44 ± 0.9094.73 ± 0.8793.34 ± 1.6793.17 ± 1.8093.25 ± 1.0593.67 ± 1.8792.92 ± 1.7694.80 ± 0.64
AA (%)92.42 ± 2.3093.21 ± 3.1693.43 ± 1.4393.31 ± 1.8591.62 ± 3.2591.68 ± 3.3086.01 ± 5.0681.80 ± 4.2589.82 ± 4.9294.20 ± 1.07
κ 0.9362 ± 0.00780.9370 ± 0.00760.9365 ± 0.01020.9398 ± 0.00990.9244 ± 0.0180.9220 ± 0.01910.9228 ± 0.0210.9278 ± 0.01190.9190 ± 0.01860.9407 ± 0.0069
KSCOA (%)98.63 ± 0.4398.63 ± 0.4998.66 ± 0.3598.67 ± 0.4798.28 ± 1.4397.57 ± 0.9198.21 ± 2.3198.10 ± 0.9596.94 ± 2.7099.11 ± 0.12
AA (%)97.67 ± 0.4597.86 ± 0.6697.87 ± 0.7597.89 ± 0.7497.47 ± 1.6496.74 ± 1.6197.54 ± 3.4697.36 ± 1.3195.26 ± 4.1098.42 ± 0.28
κ 0.9847 ± 0.00470.9848 ± 0.00550.9851 ± 0.00390.9851 ± 0.00530.9808 ± 0.01480.9729 ± 0.01260.9808 ± 0.02570.9788 ± 0.01060.9659 ± 0.02780.9901 ± 0.0014
HoustonOA (%)92.89 ± 1.1593.37 ± 1.1992.75 ± 1.1992.87 ± 1.2392.09 ± 1.7591.50 ± 1.8791.86 ± 1.8591.54 ± 1.5390.74 ± 2.9294.24 ± 0.71
AA (%)93.05 ± 0.9593.48 ± 1.0893.01 ± 0.8493.09 ± 0.9192.14 ± 1.7391.97 ± 1.5392.66 ± 1.5492.00 ± 1.1991.20 ± 2.7794.35 ± 0.55
κ 0.9231 ± 0.01250.9283 ± 0.01290.9216 ± 0.01290.9229 ± 0.01330.9144 ± 0.01810.9080 ± 0.02020.9119 ± 0.02000.9086 ± 0.01660.8999 ± 0.02300.9377 ± 0.0078
PaviaUOA (%)99.29 ± 0.1899.28 ± 0.1799.28 ± 0.1899.24 ± 0.2399.13 ± 0.2099.11 ± 0.2999.08 ± 0.3299.04 ± 0.3298.57 ± 1.2199.35 ± 0.12
AA (%)98.57 ± 0.4898.62 ± 0.3699.58 ± 0.4098.52 ± 0.3898.43 ± 0.2898.35 ± 0.4598.27 ± 0.4998.05 ± 0.6697.55 ± 1.5098.75 ± 0.25
κ 0.9907 ± 0.00240.9905 ± 0.00220.9905 ± 0.00240.9900 ± 0.00300.9898 ± 0.00270.9888 ± 0.00360.9885 ± 0.00370.9887 ± 0.00350.9810 ± 0.01280.9914 ± 0.0015
Table 10. Statistical classification results of ablation study on the attention.
Table 10. Statistical classification results of ablation study on the attention.
DatasetMetricsCBAM + CAMHSASK-Net + CAMHSAECA-Ne + CAMHSAFCL + MHSAFCL + CAMHSA
Indian PinesOA (%)94.42 ± 0.9394.40 ± 0.6994.47 ± 0.7293.55 ± 1.0294.80 ± 0.64
AA (%)91.84 ± 3.1192.80 ± 1.0293.28 ± 2.7891.90 ± 2.0494.20 ± 1.07
κ 0.9368 ± 0.01060.9361 ± 0.00790.9369 ± 0.00820.9266 ± 0.01160.9407 ± 0.0069
KSCOA (%)98.22 ± 0.3898.43 ± 0.4598.21 ± 0.7997.84 ± 0.6099.11 ± 0.12
AA (%)97.31 ± 0.7297.61 ± 0.4097.37 ± 1.0997.00 ± 0.6798.42 ± 0.28
κ 0.9802 ± 0.00420.9826 ± 0.00500.9801 ± 0.00880.9760 ± 0.00670.9901 ± 0.0014
HoustonOA (%)92.71 ± 1.1192.52 ± 1.4393.16 ± 1.3292.26 ± 1.0194.24 ± 0.71
AA (%)92.95 ± 1.0092.84 ± 1.1093.29 ± 1.0192.45 ± 0.7394.35 ± 0.55
κ 0.9212 ± 0.01200.9192 ± 0.01550.9260 ± 0.01430.9163 ± 0.01090.9377 ± 0.0078
PaviaUOA (%)99.16 ± 0.2199.08 ± 0.2799.10 ± 0.3099.08 ± 0.3099.35 ± 0.12
AA (%)98.50 ± 0.4098.51 ± 0.4398.42 ± 0.2798.35 ± 0.5198.75 ± 0.25
κ 0.9895 ± 0.00270.9884 ± 0.00350.9888 ± 0.00390.9875 ± 0.00470.9914 ± 0.0015
Table 11. Computational complexity analysis of different networks on four datasets.
Table 11. Computational complexity analysis of different networks on four datasets.
DatasetMetricsResNetSSRNDPRNA2S2KRSSANFADCNNSPRNFGSSCASSFTTBS2TS2FTNetS3ARNHFS2CABT
Indian PinesTrainable Params (M)21.91070.364222.38890.37070.148435.67770.20540.53070.14850.37810.91768.11043.39180.4545
Computation Cost (GFLOPs)0.08430.31680.08520.34320.01580.66650.03350.46790.02280.21640.29681.07520.62580.0825
Training Time (s)23.7744.1228.5450.8113.3558.2913.0553.988.6687.6536.84107.1149.1929.89
Testing Time (s)2.124.204.895.161.155.471.135.400.709.372.3611.984.674.27
KSCTrainable Params (M)21.83390.327222.30900.33320.131435.71450.19750.47690.14850.33390.91576.83293.10820.4543
Computation Cost (GFLOPs)0.08050.27820.08140.30130.01460.67840.03230.41050.03380.19010.29680.83460.50270.5580
Training Time (s)12.3421.0522.4322.0314.0536.726.2525.233.7044.4918.7450.3322.7124.67
Testing Time (s)1.131.942.512.241.282.010.562.510.374.361.215.262.372.45
HoustonTrainable Params (M)21.73450.278122.21180.28330.124835.70150.19330.40620.14850.27580.91395.26442.80420.5235
Computation Cost (GFLOPs)0.07550.22670.07640.24550.01310.67420.03150.33390.00730.15490.29680.56120.36280.2787
Training Time (s)33.8547.0549.1049.3529.5070.4314.7659.439.7399.6541.21111.1760.7845.07
Testing Time (s)3.134.883.465.251.635.581.596.201.0111.032.2412.826.415.52
PaviaUTrainable Params (M)21.60290.216522.07390.22070.094635.62590.18500.31890.14850.20210.91053.48042.47540.5178
Computation Cost (GFLOPs)0.06900.16240.07000.17580.01110.64980.03030.23820.02280.11100.29680.29120.22390.2120
Training Time (s)94.97103.68104.00104.4342.20174.0638.49127.7027.62234.0197.66230.22165.20102.42
Testing Time (s)8.5410.929.5311.314.5514.484.3613.722.9325.866.4428.1317.8410.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, M.; Yang, Y.; Zhang, S.; Mi, P.; Han, D. Spectral-Spatial Center-Aware Bottleneck Transformer for Hyperspectral Image Classification. Remote Sens. 2024, 16, 2152. https://doi.org/10.3390/rs16122152

AMA Style

Zhang M, Yang Y, Zhang S, Mi P, Han D. Spectral-Spatial Center-Aware Bottleneck Transformer for Hyperspectral Image Classification. Remote Sensing. 2024; 16(12):2152. https://doi.org/10.3390/rs16122152

Chicago/Turabian Style

Zhang, Meng, Yi Yang, Sixian Zhang, Pengbo Mi, and Deqiang Han. 2024. "Spectral-Spatial Center-Aware Bottleneck Transformer for Hyperspectral Image Classification" Remote Sensing 16, no. 12: 2152. https://doi.org/10.3390/rs16122152

APA Style

Zhang, M., Yang, Y., Zhang, S., Mi, P., & Han, D. (2024). Spectral-Spatial Center-Aware Bottleneck Transformer for Hyperspectral Image Classification. Remote Sensing, 16(12), 2152. https://doi.org/10.3390/rs16122152

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop