Next Article in Journal
Research on Three-Dimensional Cloud Structure Retrieval and Fusion Technology for the MODIS Instrument
Next Article in Special Issue
Domain Feature Decomposition for Efficient Object Detection in Aerial Images
Previous Article in Journal
A Multi-Level Robust Positioning Method for Three-Dimensional Ground Penetrating Radar (3D GPR) Road Underground Imaging in Dense Urban Areas
Previous Article in Special Issue
Dynamic Task Planning Method for Multi-Source Remote Sensing Satellite Cooperative Observation in Complex Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification

1
School of Electronic and Information Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing University of Information Science and Technology, Nanjing 210044, China
3
School of Computer Science, Nanjing University of Information Science and Technology, Nanjing 210044, China
4
College of Intelligence and Computing, Tianjin University, Tianjin 300072, China
5
Institute of Intelligent Network and Information System, Nanjing University of Information Science and Technology, Nanjing 210044, China
6
Guangxi Forest Resources and Environment Monitoring Center, Nanning 530028, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(9), 1560; https://doi.org/10.3390/rs16091560
Submission received: 19 March 2024 / Revised: 22 April 2024 / Accepted: 25 April 2024 / Published: 27 April 2024
(This article belongs to the Special Issue Deep Learning for the Analysis of Multi-/Hyperspectral Images II)

Abstract

:
Hyperspectral image classification (HSIC) has garnered increasing attention among researchers. While classical networks like convolution neural networks (CNNs) have achieved satisfactory results with the advent of deep learning, they are confined to processing local information. Vision transformers, despite being effective at establishing long-distance dependencies, face challenges in extracting high-representation features for high-dimensional images. In this paper, we present the multiscale efficient attention with enhanced feature transformer (MEA-EFFormer), which is designed for the efficient extraction of spectral–spatial features, leading to effective classification. MEA-EFFormer employs a multiscale efficient attention feature extraction module to initially extract 3D convolution features and applies effective channel attention to refine spectral information. Following this, 2D convolution features are extracted and integrated with local binary pattern (LBP) spatial information to augment their representation. Then, the processed features are fed into a spectral–spatial enhancement attention (SSEA) module that facilitates interactive enhancement of spectral–spatial information across the three dimensions. Finally, these features undergo classification through a transformer encoder. We evaluate MEA-EFFormer against several state-of-the-art methods on three datasets and demonstrate its outstanding HSIC performance.

1. Introduction

Hyperspectral images (HSIs) are widely used in remote sensing (RS) due to their abundance of spatial and spectral information [1]. Compared with natural images, HSIs consist of numerous dense and narrow spectral bands [2], allowing for precise identification of land categories [3,4,5,6]. Consequently, HSIs have distinct advantages in various fields based on these characteristics, including ground material identification [7], precision agriculture [8,9] and scene understanding [10]. Among these applications, HSI classification emerges as a critical task.
In recent years, various HSI classification methods have been suggested, including support vector machine (SVM) [11,12], k-nearest neighbors (KNN) [13,14] and random forest (RF) [15]. These algorithms have achieved remarkable results by utilizing spectral information effectively. While SVM performs well in high-dimensional problems, it needs to select some indispensable parameters [16]. In [17], linear discriminant analysis (LDA) was used for HSI classification. However, these classification methods still have much room for improvement due to their lack of spatial characteristics. Therefore, the extended morphological attribute profile (EMAP) took texture and morphological features into consideration [18]. Subsequently, there were also methods to capture texture and edge features in images based on Gabor filters [19,20]. Nevertheless, these methods fail to consider the correlation between spatial and spectral information. Thus, in [21,22,23], spectral and spatial information were jointly extracted for HSI classification. Nevertheless, these traditional methods extract shallow texture features and fail to reflect deep connections between spectral–spatial features. In this thesis, we develop a lower branch beyond the spectral–spatial features of the HSI itself to perform the processing of the local binary pattern (LBP) features under the first component band. LBP is an efficient approach for describing the local texture of an image and generates a binary encoding by comparing the differences in gray values between a pixel and its neighboring pixels. This operation can enhance the representation of spatial information by modeling spatial structures such as edges and corner points in the image. The upper and lower branches combine spectral and spatial information in order to more comprehensively and accurately describe the properties of HSI.
Compared with traditional machine learning classification technology, deep learning (DL) [24] has the characteristics of automatic learning and strong classification ability and is widely used in HSIC. It is important to note that the CNN model [25,26] stands as the most prevalent choice because of its proficiency in extracting features efficiently, its adaptability in handling high-dimensional data processing, its retention of spatial information, and its capability in managing large-scale data processing. Yang et al. [27] introduced four novel deep learning models, encompassing both two-dimensional CNN and three-dimensional CNN. Their research revealed that, while the 2D-CNN model excelled at exploiting spatial characteristics, it lacked consideration of spectral correlations. On the other hand, 3D-CNN models, despite having a higher number of network parameters compared to 2D models, effectively utilize spectral information alongside spatial features. To leverage both spatial and spectral information, Roy et al. [28] developed a hybrid spectral CNN (HybridSN) that integrates a spectral–spatial 3D-CNN with a spatial 2D-CNN. Unlike models solely reliant on 3D-CNNs, HybridSN incorporates elements of 2D-CNNs to extract a more abstract spatial representation, consequently streamlining the model’s complexity. Zhu et al. [29] devised the deformable HSI classification network (DHCNet): a CNN-based method tailored for hyperspectral image classification. DHCNet integrates deformable convolutional sampling locations that dynamically conform in size and shape to accommodate the intricate spatial characteristics found in HSIs. This adaptive feature enables enhanced extraction of spatial features, leveraging complex structural information more efficiently. Jia et al. [30] proposed a lightweight convolutional neural network (LWCNN) for HSIC and designed a two-scale convolutional (DSC) module to process joint spatial–spectral information features [31]. By merging operations, the parameter sizes are greatly reduced. It has the advantage of efficiency and robustness when solving small-sample-set problems. Gong et al. [32] developed a multiscale convolutional and diversified metric CNN (DPP-DML-MS-CNN). Diversifying depth measurements rooted in multiscale features and determinantal point processes (DPPs) [33] has enhanced the characterization and classification abilities of HSI. However, general CNN for HSIC tends to focus overly on local information, and it is difficult to comprehensively capture the trends of spectral band curves. We thus use a hybrid 3D- and 2D-CNN as a feature extractor to refine representative local features and combine it deeply with a transformer, which is excellent at global modeling, to adequately understand the global spectral trend features.
Although a deep neural network provides an invaluable contribution to HSI classification, HSIs have a high data dimension and a large number of spectral channels, which greatly increases the number of parameters of the CNN model, requires more computing resources and is prone to overfitting problems. Recurrent neural networks (RNNs) [34,35] can utilize each band in the spectrum by applying the cyclic operator layer by layer, thus obtaining fewer training parameters than convolutional neural networks (CNNs) and making the training and reasoning phases more efficient. By generating GANs [36,37,38], the discriminator training process continues to be effective through network confrontation and competition, which can alleviate the overfitting phenomenon in the training process. ResNet [39,40,41] mitigates the problem of disappearing gradients and avoids the loss of accuracy as the network deepens. In addition to the above methods, there are many classic deep learning methods such as autoencoders (AEs) [42,43], deep confidence networks (DBNs) [44,45], complete convolutional networks (FCNs) [46,47] and capsule networks (CapsNets) [48,49].
Transformers have significant advantages when processing sequential data and can establish global relationships, but they still encounter many challenges, such as limited spatial feature extraction capabilities or high computational costs [50]. A transformer is a neural network architecture based on a self-attention mechanism. Its emergence abandons the traditional RNN or CNN and allows the model to be trained in parallel and to have global information. The key feature of the transformer model is that it relies on multi-head self-attention mechanism (MHSAs) to capture dependencies between different elements in the input sequence, regardless of their position or distance in the sequence. With their powerful parallel computing capabilities, good scalability, ability to handle long-distance dependencies and advantages in processing long sequence data, transformers have broad application prospects in various fields and are still being continuously improved and expanded. Hong et al. [51] used transformers in HSI classification tasks for the first time. Later, Sun et al. [52] developed a spectral–spatial feature tokenization transformer (SSFTT) model to capture spectral–spatial features and high-level semantic features. Tu et al. [53] introduced an architecture named the local semantic feature aggregation transformer (LSFAT), which employs local semantic feature aggregation. This design enhances the capability of transformers to effectively capture long-term dependencies within multiscale features. Qiao et al. [54] recently developed a new type of hierarchical dual-frequency transformer network (DFTN) in which a frequency domain feature extraction (FDFE) block was proposed to capture high-frequency and low-frequency features separately, allowing the network to effectively utilize the input data. The multi-layer feature information in the system improves the modeling ability for complex relationships. Wang et al. [55] proposed a novel extended spectral spatial attention network (ESSAN) for HSI data classification when training samples are insufficient. For the whole network structure of a transformer, the information data for its global modeling all come from the token data generated by local patch transformation. In addition, some transformer networks based on the mask technique can effectively improve the classification performance in scenarios with insufficient samples [56,57]; such a technical improvement is also eye-catching. Our proposed SSEA module achieves further enhancement of the spectral–spatial features by computing the attention in three dimensions and also skillfully incorporates the LBP information. This operation achieves high-performance feature refinement for the subsequently generated token and also eliminates redundant information to a certain extent, providing accurate and representative global modeling information for the MHSA operation in the transformer.
In this paper, a novel multiscale efficient attention with enhanced feature transformer is presented for HSI classification. It mainly includes a multiscale efficient attention feature extraction module, a spectral–spatial enhancement attention module and a transformer encoder. The ingenious feature extraction method adequately exploits the abundant spatial and spectral information in HSI. The SSEA module enhances the interaction of spectral information with spatial features and LBP features from multiple perspectives. The transformer encoder fully integrates the key features through multi-head self-attention to optimize the feature representation. The main contributions of this paper are listed as follows:
(1)
MEA-EFFormer is a multiscale efficient attentional feature extraction module that incorporates an efficient channel attention mechanism with multiscale convolution. It facilitates the mining of details in spectral–spatial information and solves the problem of fine-grained feature loss during single-scale sampling.
(2)
MEA-EFFormer uses an SSEA module. Based on three directions, C-H, C-W and H-W, it captures the dependencies between spectral–spatial LBP information, refines the scale of the features and improves the perception of the attention mechanisms.
(3)
The classification performance of MEA-EFFormer outperforms several classical and SOTA methods. Experiments on all three well-known datasets show that the proposed method has excellent classification performance.

2. Materials and Methods

Figure 1 depicts the overarching structure of the HSIC task incorporating the innovative MEA-EFFormer. This architecture is built upon three core components: a spectral–spatial multiscale efficient attention feature extraction module, a spectral–spatial enhancement attention module, and a transformer encoder module.

2.1. Spectral–Spatial Multi-Feature Convolution Extraction

2.1.1. Multiscale Efficient Attention Feature Extraction Module

Original HSI data contain abundant spatial–spectral features that are manifested through specific spectrum reflectances and spatial morphologies. Preserving the relationship between spatial and spectral information is crucial for accurately distinguishing various land cover in classification tasks. Therefore, we employ a multiscale 3D-CNN that enables simultaneous extraction from multiple perspectives. The 3D-CNN utilizes three-dimensional kernels to perform convolution operations on the HSI along the spectral domain. However, sampling at a single scale may lead to the loss of fine-grained features. To address this issue, we utilize three distinct types of convolution kernels, each characterized by a unique scale size, to enable specialized extraction of relevant features. To prevent redundancy in high-dimensional data and mitigate the risk of overfitting, we conduct principal component analysis (PCA) dimensionality reduction on the raw HSI I R m × n × l to I R m × n × b prior to 3D convolution operations, where m × n represents the spatial size, l represents the number of bands in the original hyperspectral data, and b is the number of bands after PCA. This step aims to retain the most significant bands in order to optimize the model’s performance and ensure the extracted features are more representative. Subsequently, we partition the HSI data into numerous 3D cubes P R s × s × b to facilitate convolution operations. The multiscale propagation operation for the jth feature cube at the ( x , y , z ) position on the ith layer is expressed as:
α i j x y z = θ m h = 0 H L i 1 w = 0 W L i 1 r = 0 R L i 1 w i j m h w r α ( i 1 ) m ( x + h ) ( y + w ) ( z + r ) + b i j β i j x y z = θ m h = 0 H M i 1 w = 0 W M i 1 r = 0 R M i 1 w i j m h w r β ( i 1 ) m ( x + h ) ( y + w ) ( z + r ) + b i j δ i j x y z = θ m h = 0 H S i 1 w = 0 W S i 1 r = 0 R S i 1 w i j m h w r δ ( i 1 ) m ( x + h ) ( y + w ) ( z + r ) + b i j
Here, m denotes the feature cube associated with the jth feature cube. The three types of convolution kernel sizes are represented by H L × W L × R L , H M × W M × R M and  H S × W S × R S . The variables w and b correspond to the weight and bias parameters, respectively, while θ represents the activation function. The functions α i j x y z , β i j x y z and  δ i j x y z are the extracted features at different scales. To mitigate the sensitive volatility of the spectral information due to scale changes, we input them into efficient channel attention (ECA) separately to realize the alignment fusion of the feature maps at different scales. The process of ECA is illustrated in Figure 2.
Efficient channel attention is a mechanism for enhanced modeling of neural networks by focusing on the importance of different channels within the feature map. It utilizes lightweight and highly efficient computation to break through the cost limitations of complex band computation for HSI data. For the spectral–spatial data α i j x y z output from one scale branch, we use the global average pool (GAP) to compress it into a feature map with a spatial dimension of 1 × 1 . Subsequently, the one-dimensional convolution of the adaptive convolution kernel is utilized to compute the weights under each band, and finally, a  S i g m o i d function is introduced to map the weights between 0 and 1. These weights are then utilized to modulate the original feature map, highlighting information-rich bands while suppressing relatively irrelevant bands. The size of the adaptive convolution kernel is generically related to the number of spectral bands under the branch of the scale, as expressed by the formula:
k = l o g 2 ( C ) γ + b γ
where C is the number of spectral bands characterized, and b and r are set to 1 and 2, respectively. The advantage of this module is that global dependencies between spectral bands are captured using a small computational cost, and the whole process can be represented as follows:
o u t = S i g C o n v G A P F , C F × F , F = α i j x y z , β i j x y z , δ i j x y z
After performing ECA computation on the three scale branches, we cascade fuse them with 1 × 1 × 1 unit convolution and extract 2DCNN features uniformly. This is done as follows:
v i j x y z = cat ( α i j x y z , β i j x y z , δ i j x y z ) Filter 1 × 1 × 1
v i j x y = θ m h = 0 H i 1 w = 0 W i 1 w i j m h w v ( i 1 ) m ( x + h ) ( y + w ) + b i j
Here, m represents the feature map associated with the jth feature, while H × W denotes the dimensions of the 2D convolution kernel. The parameters w and b correspond to the weights and biases, respectively, and  θ represents the activation function. The HSI features at this time have extensive detailed information and high expressive capability. To further enhance the surface texture information and capture the spatial fine-grained features, we next extract the LBP features from the original HSI image.

2.1.2. LBP Convolution Feature Processing

The local binary pattern (LBP) is an operator used to characterize local features of an image and has significant advantages such as grayscale invariance and rotational invariance. The LBP feature provides a description of the texture characteristics of the HSI surface, which utilizes spatially localized pixel grayscale differences to highlight landscape detail information. Each pixel in the image is compared as a center pixel with its domain pixel’s gray value, and binary bits are set for the domain pixel depending on the result. Subsequently, this is converted to decimal to get the LBP code. The specific computational operations are as follows:
L B P P , R = p = 0 P 1 2 p · S lbp ( x p , y p ) S lbp ( x ) = 1 , if I ( x , y ) I c 0 , if I ( x , y ) < I c
where S lbp is the binary bit of the neighboring pixel, I ( x , y ) is the grayscale value of the neighboring pixel, and  I c is the grayscale value of the center pixel. P is the number of pixels in the neighborhood, R is the radius of the neighborhood, and  ( x p , y p ) are the coordinates of the pixels in the neighborhood. In the following, we perform a 2D convolution operation on the LBP spatial features to obtain a feature map with the same dimensions as the spectral–spatial convolution in the previous section. The operation here is the same as Equation (5).

2.2. Spectral–Spatial Enhancement Attention Module

After extracting the LBP convolution features, we superimpose them onto the spectral–spatial branching features from the spectral dimension, i.e.,  M R b + 1 × h × w , with h and w being the convolution feature map dimensions. Subsequently, the LBP convolution and spectral–spatial convolution were merged through a 2D convolution unit employing a 1 × 1 kernel size, resulting in a b-band convolution feature map. This fusion process effectively combines the complementary information from both convolutions, enriching the representational power of the feature map. However, the correlation between the original spectral features and the improved spatial features can significantly impact the final classification results. To address this, we propose a spectral–spatial enhancement attention (SSEA) module that facilitates the interactive enhancement of spatial and spectral information from two vertical directions and one horizontal direction. The specific process of SSEA is shown in Figure 3. SSEA establishes the dependency relationship between spectral dimension C and spatial dimensions H and W by rotational operation and calculates the attention weights in the three directions C H , C W and H W . Following this, the three types of correlation weights are combined to promote interaction and strengthen the representation of spectral features, thereby ensuring a more comprehensive and nuanced understanding of the data.
In this stage, the features undergo 90-degree counterclockwise rotation along the H axis, resulting in dimensions of W × H × C . Following this transformation, both average and maximum pooling operations are performed in the W dimension. The average pooling operation calculates the mean value of image regions to smooth the data and suppress noise, thereby enhancing the stability and coupling of subsequent spectral–spatial information. In contrast, maximum pooling discards other information by retaining the maximum pixel value of each region in both the spatial and spectral dimensions. This operation helps to emphasize the salient features in the image. Finally, an activation function is applied to generate attention weights between the H and C dimensions; these weights are then used to modulate the convolution features. The second branch emphasizes the interaction between the spectral dimension C and spatial dimension W. In this phase, the convolution features undergo a 90-degree counterclockwise rotation along the W axis, resulting in dimensions of H × C × W . Attention weights for this branch are computed similarly to those in the first branch. The third branch is focused on the interaction between spatial dimensions H and W. It directly applies average and maximum pooling to the convolution features, enhancing both the original spatial information and the spatial information derived from LBP features. This approach further sharpens the ability to capture intricate ground category features. Once the three branches have been computed, their outputs are averaged and aggregated, fostering interaction between spectral and spatial dimensions and resulting in a richer and more comprehensive feature representation.

2.3. Transformer Encoder Module

The transformer encoder block mainly consists of two residual structures, which are represented by the multi-head self-attention mechanism (MHSA) module and the multi-layer perceptron (MLP) module, as shown in Figure 4a. For the enhanced spectral–spatial features N R c × h × w , which were extracted in the previous section, we perform an embedding flattening operation on their spatial data to obtain a 2D structure N R c × h × w that is suitable as input for the transformer, where d = h × w . As shown in Figure 4b, we initialize three update matrices W Q , W K , W V to map the input features N to Q , K , and V matrices, respectively, which are composed of h head components:
Q = { Q 1 , Q 2 , , Q i , , Q h } K = { K 1 , K 2 , , K i , , K h } V = { V 1 , V 2 , , V i , , V h }
where h is the number of heads, and Q i , K i , V i R c × ( d / h ) . Next, we use them to compute the attention scores:
SA i = Attention ( Q i , K i , V i ) = Softmax Q i K i T d V i
After obtaining the attention results, we combine the multiple results and perform a linear transformation to obtain the final global interaction result:
MSA ( Q , K , V ) = Concat ( SA 1 , SA 2 , , SA h ) W
where W R d × d is a parameter matrix used to perform the linear transformation.
Next, the data proceed to the second residual structure MLP. it is composed of two linear layers. Following it, highly expressive and discriminative spectral–spatial features are fitted into multiple categories to realize the classification task for HSI.
Figure 4. Graphical representation of the transformer encoder: (a) The general structure of encoder blocks. (b) Multi-head self-attention mechanism.
Figure 4. Graphical representation of the transformer encoder: (a) The general structure of encoder blocks. (b) Multi-head self-attention mechanism.
Remotesensing 16 01560 g004

2.4. Algorithm Summarization for MEA-EFFormer

The overall process of the proposed MEA-EFFormer network is shown in Algorithm 1.
Algorithm 1: MEA-EFFormer network.
Require:
  • HSI data I R m × n × l ; ground truth Y R m × n ; PCA bands number b; input patch size s; training sample rate μ % ; attention head number h; epochs E; batch size = 128.
Ensure:
   Predicted classification labels for the test dataset.
1:
Extract the I p c a features from the HSI data and transform them into multiscale convolution features α i j x y z , β i j x y z , δ i j x y z by Equation (1).
2:
Perform efficient channel attention operation on α i j x y z , β i j x y z , δ i j x y z , and stack to perform unit convolution and 2D convolution to get V i j x y by Equations (3)–(5).
3:
Obtain the I l b p features by Equation (6) from HSI data and transform them into the convolution features V l b p x y by Equation (5).
4:
for  i = 1 to E do
5:
   Perform SSEA operation on V i j x y and V l b p x y to achieve enhancement of spectral–spatial features.
6:
   Perform multi-head self-attention in the transformer encoder.
7:
end for
8:
Employ the attention outputs and feed them into a linear layer to determine the corresponding labels.
9:
Employ the trained model on the test dataset to acquire predicted labels.

3. Experiment and Analysis

In this section, we employ three well-known HSI datasets—Indian Pines (IP), Salinas (SA) and Pavia University (PU)—to evaluate the effectiveness of the proposed method, and we use three metric indicators to give a quantitative assessment of the classification results.

3.1. Data Description

3.1.1. Indian Pines

This dataset was acquired by the AVIRIS sensor over a test site in northwestern Indiana, USA. It has a spatial resolution of 20 m per pixel and covers an area of 145 × 145 pixels. After removing the water absorption bands, the dataset contains 200 spectral bands for analysis. The ground truth for Indian Pines identifies 16 distinct classes, including various crops, forests and other natural vegetation types. This dataset is frequently used to benchmark HSIC algorithms. Figure 5 illustrates the false-color image and labeling map of Indian Pines, and the specific division of the training and testing sets of the samples is shown in Table 1.

3.1.2. Salinas

This dataset was captured by the AVIRIS sensor; the Salinas dataset focuses on the Salinas Valley in California. It offers a higher spatial resolution of 3.7 m per pixel with dimensions of 512 × 217 pixels. Similar to Indian Pines, the Salinas dataset typically uses 204 spectral bands after water absorption band removal. The ground truth comprises 16 classes representing agricultural fields, vineyards, and bare soil. Researchers often use this dataset to explore the challenges of classifying crops with finer spatial details. Figure 6 illustrates the false-color image and labeling map of Salinas, and the specific division of the training and testing sets of the samples is shown in Table 1.

3.1.3. Pavia University

The ROSIS sensor collected this dataset over an urban area in Pavia, Italy. It boasts a high spatial resolution of 1.3 m per pixel. The dataset contains 103 spectral bands and covers an image size of 610 × 340 pixels. Pavia University offers 9 land-cover classes focused on urban features. This dataset is commonly used to study the classification of urban environments and to address the challenge of working with noisy spectral bands. Figure 7 illustrates the false-color image and labeling map of Pavia University, and the specific division of the training and testing sets of the samples is shown in Table 1.

3.2. Experimental Setting

3.2.1. Evaluation Criteria

In order to quantitatively evaluate the experimental results, three quantitative evaluation metrics were employed: overall accuracy (OA), average accuracy (AA) and kappa coefficient. First, OA measures the ratio between the number of correctly classified samples in a dataset and the total number of samples. OA provides an overall assessment of classification performance by indicating the model’s ability to correctly classify samples. Second, AA calculates the average accuracy for each category in the dataset. It provides an evaluation of the model’s performance on different categories and helps to determine whether the model performs well uniformly across all categories. Finally, the kappa coefficient measures the agreement between the predictions and the true classification while taking into account the stochastic agreement. A kappa value close to 1 indicates that there is strong agreement between prediction and true categorization beyond random consistency.

3.2.2. Environment Configuration

The proposed method was implemented using the PyTorch 2.2.0, while the traditional classical methods used for comparison were executed in the MATLAB R2018b environment. The computational setup included an Intel Xeon Silver 4314 CPU (Intel Corporation, Santa Clara, CA, USA) with 256 GB of RAM, along with an NVIDIA GeForce RTX 4090 GPU server (ASUS, Taipei, Taiwan) equipped with 24 GB of memory. In the comparison involving deep learning and transformer-based methods, parameters were configured as follows: the number of epochs was set to 100, and a batch size of 128 was employed.

3.2.3. Parameter Setting Adjustment

In this subsection, we analyze the impact of several important parameters on the classification results of the proposed network. These parameters are patch size, reduced spectral dimension, learning rate of the network, and the number of attention heads.
Figure 8 illustrates the impact of patch size on the classification metrics OA, AA and kappa. On all three datasets, the accuracy generally increases as the patch size increases. This is likely because larger patches capture more spectral information, which can help the model better distinguish between different classes. However, there is a point of diminishing returns at which increasing the patch size further does not improve accuracy. This is because larger patches may also include irrelevant information that can confuse the model. The specific patch size that yields the best accuracy varies depending on the dataset. For Indian Pines and Pavia University, the highest accuracies are achieved with a patch size of 13 × 13 . For Salinas, the best accuracy is achieved with a patch size of 15 × 15 . Moreover, our proposed network achieves relatively good and stable performance over a wide range of patch sizes, e.g., [ 11 × 11 , 19 × 19 ] . This demonstrates that our proposed network has certain robustness to the parameter of patch size.
Figure 9 shows the classification results as a function of the reduced spectral dimensions. The reduced spectral dimensions indeed have a strong impact on the performance of the proposed network; however, the proposed network can achieve relatively stable results when the reduced dimensions lie in the range of [15, 35] for all three datasets. Specifically, it achieves the best performance when the reduced dimension is 20 for Indian Pines. For the Salinas and Pavia University datasets, the optimal values of the reduced dimensions are 25 and 30, respectively.
Figure 10 plots the impact of the learning rate on the classification accuracy. For three datasets, the learning rate of the proposed network indeed has a strong impact on the performance. It can be seen that on the three datasets, the OA, AA and kappa curves of the proposed network all show a trend of first increasing and then decreasing, indicating that these three indicators all have an optimal value. In addition, when the learning rate is in the range [ 5 × 10 5 , 1 × 10 3 ], the quantitative indicator values obtained by the proposed network are relatively stable, indicating that the network has certain robustness to the learning rate parameter. In the subsequent classification, we set the learning rate to 1 × 10 3 .
Figure 11 plots the number of attention heads as a function of the classification accuracy. The performance of the proposed network is quite stable when the number of attention heads lies in the range of [2, 32]. When the number is eight, the network achieves optimal performance.

3.3. Ablation Study

We performed an ablation experiment employing a 5% sample rate on the Indian Pines dataset. The framework was deconstructed into six discernible sections: principal component analysis (PCA), multiple scales (MS), efficient channel attention (ECA), LBP feature branch (LBP), spectral–spatial enhancement attention module (SSEA) and transformer encoder (TE). Subsequently, a comprehensive evaluation was conducted utilizing performance metrics including OA, AA and the kappa coefficient. The outcomes of these ablation experiments are meticulously tabulated in Table 2 for reference and analysis.
In Case 1, we eliminate the PCA component of the network and input all 200 bands of the Indian Pines dataset into the model. The amount of data computed is more than six times that of our proposed method. The large amount of redundant data being fed into the model computation also brings about a slight decrease in the accuracy metric.
In Case 2, we cancel the MS component of the network and employ a single-scale CNN to extract the HSI features. At this time, the accuracy metrics are all significantly decreased, especially the AA accuracy. This demonstrates the remarkable advantage of the MS component for exploiting the details of the imbalanced category samples.
In Case 3, we eliminate the ECA component of the network and do not compute the attention to the spectral dimension information. At this point, the degradation of AA accuracy is also obvious due to the direct and rough integration of the spectral information at multiple scales. This demonstrates that the ECA component can significantly mitigate the sensitivity of spectral information to scale transformations.
In Case 4, we directly cut the lower branch of the network and do not use the LBP features for spatial information enhancement. The AA accuracy decreases significantly in this case. This indicates that LBP features can provide effective spatial feature enhancement for samples with unbalanced distributions so as to provide the model’s capture of the data as a whole.
In Case 5, we remove the SSEA module so that spectral–spatial information and LBP features are merely stacked and fed into the network. Each accuracy metric at this time is slightly decreased. This suggests that the fusion and de-redundancy operations within the SSEA module on the features can extract finer representations, which is conducive to the downstream recognition of the ground surface categories.
In Case 6, the transformer encoder is replaced with a deep residual convolutional network, and all accuracy metrics drop severely. This shows that with purely local features, it is difficult to carry out effective data modeling and the model lacks the overall consideration of global information. And it also further proves that the transformer is favorable for capturing trends within the global spectral curve and for calculating long-distance spatial information.
In addition, we conducted additional ablation experiments for the three branches of the SSEA module to explore the impact of each branch on the final classification results. The results, as shown in Table 3, show that the computation of attention with the absence of any branch causes a decrease in the accuracy metric. This also proves that computation using each branch of our proposed SSEA module effectively strengthens the degree of coupling between spectral–spatial and LBP-HSI for refinement of feature representation.

3.4. Classification Results

In this subsection, we compare the proposed MEA-EFFormer network with state-of-the-art classifiers using quantitative and qualitative measures. These classifiers include random forest (RF) [58], support vector machine (SVM) [11], 1D-CNN [59], 2-DCNN [60], 3DCNN [61], HybridSN [28], GAHT [62], SpectralFormer [51], SSFTT [52] and GSC-ViT [63]. Table 4, Table 5 and Table 6, respectively, provide quantitative results of the compared algorithms on the Indian Pines, Salinas and Pavia University datasets. The parameter settings of the comparison methods were set according to the optimal settings of the reference source texts. To ensure the generality of the experimental results, we conducted ten separate rounds of each experiment and retained the means and variances. From the tables, it can be seen that traditional classification methods have a significant gap compared to deep learning methods. Among deep learning classifiers, methods based on transformers generally achieved better results; this is a benefit of the deep exploration of long-distance relationships between features. It is evident that our MEA-EFFormer method achieves the highest OA, AA and kappa values across all three datasets.
As well-known methods for HSIC in recent years, SSFTT and GSC-ViT effectively integrate spatial–spectral features from HSI and enhance feature discrimination through the integration of transformer networks. Compared with SSFTT on the Indian Pines dataset, MEA-EFFormer demonstrated a 0.47% increase in OA, a 1.26% increase in AA and a 0.53% increase in kappa while reducing bias by 0.11, 1.17, and 0.13, respectively. This comparison highlights the superior performance of MEA-EFFormer for enhancing classification accuracy and stability through the incorporation of multiscale information and an SSEA strategy for exploring discriminative features of land-cover objects.
For the Salinas dataset, GAHT obtained the second-best classification accuracies in terms of OA, AA and kappa. This is a benefit of the group-aware hierarchical transformer strategy in the network, which is good at classifying scenarios with objects that are relatively concentrated and uniform. However, MEA-EFFormer still achieves classification results that are similar to or even better than GAHT. Specifically, it improves OA by more than 0.16%, AA by more than 0.53% and kappa by more than 0.18%. This result further demonstrates the effectiveness of the proposed network for HSIC.
Finally, for the Pavia University dataset, which is known for its high spatial complexity, our method also achieves the best classification results, especially in terms of AA. This shows that MEA-EFFormer can effectively utilize the existing samples for global information even when dealing with scenarios with uneven sample distributions.
In a word, the proposed network has a significant advantage over the state-of-the-art transformer classifiers on these three well-known datasets and achieves the best results in terms of OA, AA and kappa.

3.5. Visual Evaluation

To qualitatively compare the performance of different algorithms, we illustrate the classification maps of different methods on the Indian Pines, Salinas and Pavia University datasets in Figure 12, Figure 13 and Figure 14, respectively.
It can be clearly seen from the figures that the traditional methods exhibit numerous noisy points in the classification maps across the three datasets, indicating that their classification accuracies are relatively low. This is primarily attributed to the inherent limitations of traditional methods in terms of feature representation and exploration in high-dimensional data, which results in an inability to capture deep-level feature representations effectively. In contrast, the classification maps generated by deep neural network classifiers are generally smoother compared to those produced by traditional methods, aligning with the quantitative results. Particularly, transformer methods stand out for their superior performance: yielding classification maps that deliver satisfactory results both within categories and at their boundaries.
Notably, on the Indian Pines and Salinas datasets, the classification maps generated by the proposed MEA-EFFormer demonstrate enhanced classification performance, and the classification results of the boundary pixels are all relatively precise. At the same time, the smoothness and consistency of the classification maps are also appealing inside each category. For example, the “orange area” in Figure 13j is more accurate than for the other classification maps.
For the Pavia dataset, due to its high spatial resolution, the distribution of objects is more dispersed and the boundaries are more complex. The classification accuracy of most classifiers is lower, and there are many noise points in the classification map. However, the proposed network MEA-EFFormer can still obtain a relatively satisfactory result. For example, the “gray area” in Figure 14l is significantly more accurate than that of other classification maps.

3.6. Model Complexity and Efficiency Analysis

We analyzed the computational efficiency of several common deep-learning-based methods on the Pavia University dataset with a sampling rate of 1%. The results are shown in Table 7; our proposed method achieves a moderate advantage in terms of training time and parameter size while achieving the leading classification accuracy.
For training time, we ranked second among the several classes of transformer-based methods that we compared. As for SSFTT, the simple fact is that it only performs convolutional extraction of a hybrid on raw HSI data, whereas MEA-EFFormer additionally uses LBP data for spatial information augmentation, which results in a slight increase in time. With the 3D-CNN and HybridSN methods, the runtime is faster since they only use convolutional networks.
For the model parameter sizes, our method is also preferred to most of the methods. The smaller parameter sizes of SpectralFormer and SSFTT are due to the fact that they are too simple, as they only have a single scale in the feature extraction stage, while our method adopts a multiscale strategy to fully exploit the spectral–spatial information in the HSI data. This is an important reason why MEA-EFFormer achieves higher accuracy. For GSC-ViT, the method itself is known for its light weight, and there is an obvious gap with the proposed method in terms of accuracy.
In summary, our proposed MEA-EFFormer can keep the computational efficiency as low as possible with small parameter sizes under the premise of leading accuracy, which again proves the superiority of our method.

4. Conclusions

In this paper, we propose a method called multiscale efficient attention and enhanced feature transformer (MEA-EFFormer) for hyperspectral image classification. We obtain further-refined spectral–spatial information through multiscale efficient attention feature extraction module. Then, we combine two-dimensional convolution features with local binary mode (LBP) spatial information, which effectively improves the representation of features. Then, we use the spectral–spatial enhancement attention module to make the feature enhanced. Finally, we classify these features through transformer encoders. Our experimental results on three HSI datasets show that the proposed method has excellent performance compared with other classification methods. Although our method achieves remarkable results in experiments, we also recognize that there is room for improvement. In the future, we will continue to work to improve our approach. We plan to introduce more efficient feature extraction and attention mechanisms as well as to optimize model structure and parameter settings to achieve better classification performance. At the same time, we also encourage other researchers to explore and innovate in this field and to jointly promote the development of HSIC technology.

Author Contributions

Conceptualization, Q.S. and L.S.; methodology, G.Z. and C.F.; software, Y.F.; validation, G.Z., Y.F. and X.L.; formal analysis, L.S.; investigation, Y.F.; data curation, G.Z. and C.F.; writing—original draft preparation, Q.S.; writing—review and editing, L.S. and X.L.; visualization, G.Z. and Y.F.; supervision, L.S. and X.L.; project administration, X.L.; funding acquisition, Q.S. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62076137, in part by the Startup Foundation for Introducing Talent of NUIST, grant number 2022r075, also in part by Guangxi Forestry Technology Promotion and Demonstration Application Project, grant number 2024GXLK19.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors thank the anonymous reviewers and the editors for their insightful comments and helpful suggestions that helped improve the quality of our manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liu, S.; Liu, S.; Zhang, S.; Li, B.; Hu, W.; Zhang, Y.D. SSAU-Net: A Spectral–Spatial Attention-Based U-Net for Hyperspectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542116. [Google Scholar] [CrossRef]
  2. Sun, G.; Pan, Z.; Zhang, A.; Jia, X.; Ren, J.; Fu, H.; Yan, K. Large Kernel Spectral and Spatial Attention Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5519915. [Google Scholar] [CrossRef]
  3. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in Hyperspectral Image and Signal Processing: A Comprehensive Overview of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef]
  4. Sun, L.; Wang, Q.; Chen, Y.; Zheng, Y.; Wu, Z.; Fu, L.; Jeon, B. CRNet: Channel-enhanced Remodeling-based Network for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5618314. [Google Scholar] [CrossRef]
  5. Sun, L.; Wang, X.; Zheng, Y.; Wu, Z.; Fu, L. Multiscale 3-D–2-D Mixed CNN and Lightweight Attention-Free Transformer for Hyperspectral and LiDAR Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 2100116. [Google Scholar] [CrossRef]
  6. Sun, L.; Zhang, H.; Zheng, Y.; Wu, Z.; Ye, Z.; Zhao, H. MASSFormer: Memory-Augmented Spectral-Spatial Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  7. Yao, J.; Hong, D.; Xu, L.; Meng, D.; Chanussot, J.; Xu, Z. Sparsity-Enhanced Convolutional Decomposition: A Novel Tensor-Based Paradigm for Blind Hyperspectral Unmixing. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5505014. [Google Scholar] [CrossRef]
  8. Dalponte, M.; Ørka, H.O.; Gobakken, T.; Gianelle, D.; Næsset, E. Tree Species Classification in Boreal Forests with Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2632–2645. [Google Scholar] [CrossRef]
  9. Zhang, J.; Tao, D. Empowering Things With Intelligence: A Survey of the Progress, Challenges, and Opportunities in Artificial Intelligence of Things. IEEE Internet Things J. 2021, 8, 7789–7817. [Google Scholar] [CrossRef]
  10. Hang, R.; Liu, Q.; Li, Z. Spectral Super-Resolution Network Guided by Intrinsic Properties of Hyperspectral Imagery. IEEE Trans. Image Process. 2021, 30, 7256–7265. [Google Scholar] [CrossRef]
  11. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  12. Fauvel, M.; Chanussot, J.; Benediktsson, J. Evaluation of Kernels for Multiclass Classification of Hyperspectral Remote Sensing Data. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 2, p. II. [Google Scholar] [CrossRef]
  13. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-Based Representation of Superpixels for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  14. Samaniego, L.; Bardossy, A.; Schulz, K. Supervised Classification of Remotely Sensed Imagery Using a Modified k-NN Technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Cao, G.; Li, X.; Wang, B. Cascaded Random Forest for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1082–1094. [Google Scholar] [CrossRef]
  16. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of Hyperspectral Images With Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  17. Lobo, A. Image segmentation and discriminant analysis for the identification of land cover units in ecology. IEEE Trans. Geosci. Remote Sens. 1997, 35, 1136–1145. [Google Scholar] [CrossRef]
  18. Huang, J.; Liu, K.; Xu, M.; Perc, M.; Li, X. Background Purification Framework With Extended Morphological Attribute Profile for Hyperspectral Anomaly Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8113–8124. [Google Scholar] [CrossRef]
  19. Liu, C.; Li, J.; He, L.; Plaza, A.; Li, S.; Li, B. Naive Gabor Networks for Hyperspectral Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 376–390. [Google Scholar] [CrossRef]
  20. Jiang, C.; Su, J. Gabor Binary Layer in Convolutional Neural Networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3408–3412. [Google Scholar] [CrossRef]
  21. Li, H.; Ye, Z.; Xiao, G. Hyperspectral Image Classification Using Spectral–Spatial Composite Kernels Discriminant Analysis. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2341–2350. [Google Scholar] [CrossRef]
  22. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized Composite Kernel Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  23. Tang, Y.Y.; Lu, Y.; Yuan, H. Hyperspectral Image Classification Based on Three-Dimensional Scattering Wavelet Transform. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2467–2480. [Google Scholar] [CrossRef]
  24. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  25. Romero, A.; Gatta, C.; Camps-Valls, G. Unsupervised Deep Feature Extraction for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1349–1362. [Google Scholar] [CrossRef]
  26. Cao, X.; Yao, J.; Xu, Z.; Meng, D. Hyperspectral Image Classification With Convolutional Neural Network and Active Learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4604–4616. [Google Scholar] [CrossRef]
  27. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.K.; Zhang, X.; Huang, X. Hyperspectral Image Classification With Deep Learning Models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  28. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  29. Zhu, J.; Fang, L.; Ghamisi, P. Deformable Convolutional Neural Networks for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1254–1258. [Google Scholar] [CrossRef]
  30. Jia, S.; Lin, Z.; Xu, M.; Huang, Q.; Zhou, J.; Jia, X.; Li, Q. A Lightweight Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4150–4163. [Google Scholar] [CrossRef]
  31. Czaja, W.; Ehler, M. Schroedinger Eigenmaps for the Analysis of Biomedical Data. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1274–1280. [Google Scholar] [CrossRef]
  32. Gong, Z.; Zhong, P.; Yu, Y.; Hu, W.; Li, S. A CNN With Multiscale Convolution and Diversified Metric for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3599–3618. [Google Scholar] [CrossRef]
  33. Xie, P.; Salakhutdinov, R.; Mou, L.; Xing, E.P. Deep Determinantal Point Process for Large-Scale Multi-label Classification. 2017; 473–482. [Google Scholar] [CrossRef]
  34. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef]
  35. Zhang, X.; Sun, Y.; Jiang, K.; Li, C.; Jiao, L.; Zhou, H. Spatial Sequential Recurrent Neural Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4141–4155. [Google Scholar] [CrossRef]
  36. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  37. Jia, S.; Wang, Z.; Li, Q.; Jia, X.; Xu, M. Multiattention Generative Adversarial Network for Remote Sensing Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5624715. [Google Scholar] [CrossRef]
  38. Neagoe, V.E.; Diaconescu, P. CNN Hyperspectral Image Classification Using Training Sample Augmentation with Generative Adversarial Networks. In Proceedings of the 2020 13th International Conference on Communications (COMM), Bucharest, Romania, 8–20 June 2020; pp. 515–519. [Google Scholar] [CrossRef]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  40. Zhong, Z.; Li, J.; Ma, L.; Jiang, H.; Zhao, H. Deep residual networks for hyperspectral image classification. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 1824–1827. [Google Scholar] [CrossRef]
  41. Li, T.; Zhang, X.; Zhang, S.; Wang, L. Self-Supervised Learning With a Dual-Branch ResNet for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5512905. [Google Scholar] [CrossRef]
  42. Zhang, X.; Liang, Y.; Li, C.; Huyan, N.; Jiao, L.; Zhou, H. Recursive Autoencoders-Based Unsupervised Feature Learning for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1928–1932. [Google Scholar] [CrossRef]
  43. Koda, S.; Melgani, F.; Nishii, R. Unsupervised Spectral–Spatial Feature Extraction With Generalized Autoencoder for Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2020, 17, 469–473. [Google Scholar] [CrossRef]
  44. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  45. Chen, C.; Ma, Y.; Ren, G. Hyperspectral Classification Using Deep Belief Networks Based on Conjugate Gradient Update and Pixel-Centric Spectral Block Features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4060–4069. [Google Scholar] [CrossRef]
  46. Sun, H.; Zheng, X.; Lu, X. A Supervised Segmentation Network for Hyperspectral Image Classification. IEEE Trans. Image Process. 2021, 30, 2810–2825. [Google Scholar] [CrossRef]
  47. Zheng, Z.; Zhong, Y.; Ma, A.; Zhang, L. FPGA: Fast Patch-Free Global Learning Framework for Fully End-to-End Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5612–5626. [Google Scholar] [CrossRef]
  48. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.; Li, J.; Pla, F. Capsule Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2145–2160. [Google Scholar] [CrossRef]
  49. Jiang, X.; Liu, W.; Zhang, Y.; Liu, J.; Li, S.; Lin, J. Spectral–Spatial Hyperspectral Image Classification Using Dual-Channel Capsule Networks. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1094–1098. [Google Scholar] [CrossRef]
  50. Liang, L.; Zhang, Y.; Zhang, S.; Li, J.; Plaza, A.; Kang, X. Fast Hyperspectral Image Classification Combining Transformers and SimAM-Based CNNs. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5522219. [Google Scholar] [CrossRef]
  51. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5518615. [Google Scholar] [CrossRef]
  52. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  53. Tu, B.; Liao, X.; Li, Q.; Peng, Y.; Plaza, A. Local Semantic Feature Aggregation-Based Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5536115. [Google Scholar] [CrossRef]
  54. Qiao, X.; Huang, W. A Dual Frequency Transformer Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 10344–10358. [Google Scholar] [CrossRef]
  55. Wang, S.; Liu, Z.; Chen, Y.; Hou, C.; Liu, A.; Zhang, Z. Expansion Spectral–Spatial Attention Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6411–6427. [Google Scholar] [CrossRef]
  56. Zhou, F.; Xu, C.; Yang, G.; Hang, R.; Liu, Q. Masked Spectral–Spatial Feature Prediction for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4400913. [Google Scholar] [CrossRef]
  57. Huang, L.; Chen, Y.; He, X. Spectral–Spatial Masked Transformer with Supervised and Contrastive Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5508718. [Google Scholar] [CrossRef]
  58. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  59. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar] [CrossRef]
  60. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  61. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar]
  62. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral Image Classification Using Group-Aware Hierarchical Transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5539014. [Google Scholar] [CrossRef]
  63. Zhao, Z.; Xu, X.; Li, S.; Plaza, A. Hyperspectral Image Classification Using Groupwise Separable Convolutional Vision Transformer Network. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5511817. [Google Scholar] [CrossRef]
Figure 1. The architecture of the proposed MEA-EFFormer network. The network can be divided into three stages: data preprocessing, feature extraction and processing, and the transformer encoder. The data preprocessing stage includes principal component analysis (PCA) to extract the main bands from the raw HSI and local binary pattern (LBP) extraction. The feature extraction and processing stage is mainly a multiscale efficient attention feature extraction module and a spectral–spatial enhancement attention module. Finally, the obtained refined features are fed into the transformer encoder for classification operations.
Figure 1. The architecture of the proposed MEA-EFFormer network. The network can be divided into three stages: data preprocessing, feature extraction and processing, and the transformer encoder. The data preprocessing stage includes principal component analysis (PCA) to extract the main bands from the raw HSI and local binary pattern (LBP) extraction. The feature extraction and processing stage is mainly a multiscale efficient attention feature extraction module and a spectral–spatial enhancement attention module. Finally, the obtained refined features are fed into the transformer encoder for classification operations.
Remotesensing 16 01560 g001
Figure 2. Illustration of ECA. It uses global average pooling and a one-dimensional convolution operation with an adaptive convolution kernel to compute the weights under each band, followed by an activation function to implement the mapping of the attention weights.
Figure 2. Illustration of ECA. It uses global average pooling and a one-dimensional convolution operation with an adaptive convolution kernel to compute the weights under each band, followed by an activation function to implement the mapping of the attention weights.
Remotesensing 16 01560 g002
Figure 3. Illustration of SSEA. It consists of three branches that establish the dependencies between the spectral dimension C and the spatial dimensions H and W by means of rotational operations, and it computes the attention weights in each of the three directions.
Figure 3. Illustration of SSEA. It consists of three branches that establish the dependencies between the spectral dimension C and the spatial dimensions H and W by means of rotational operations, and it computes the attention weights in each of the three directions.
Remotesensing 16 01560 g003
Figure 5. Indian Pines dataset. (a) False-color map. (b) Ground-truth map.
Figure 5. Indian Pines dataset. (a) False-color map. (b) Ground-truth map.
Remotesensing 16 01560 g005
Figure 6. Salinas dataset. (a) False-color map. (b) Ground-truth map.
Figure 6. Salinas dataset. (a) False-color map. (b) Ground-truth map.
Remotesensing 16 01560 g006
Figure 7. Pavia University dataset. (a) False-color map. (b) Ground-truth map.
Figure 7. Pavia University dataset. (a) False-color map. (b) Ground-truth map.
Remotesensing 16 01560 g007
Figure 8. Patch size as a function of OA, AA and kappa. (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Figure 8. Patch size as a function of OA, AA and kappa. (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Remotesensing 16 01560 g008
Figure 9. Effect of reducing spectral dimensionality on OA, AA and kappa coefficient: (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Figure 9. Effect of reducing spectral dimensionality on OA, AA and kappa coefficient: (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Remotesensing 16 01560 g009
Figure 10. Effect of learning rate on OA, AA and kappa coefficient: (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Figure 10. Effect of learning rate on OA, AA and kappa coefficient: (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Remotesensing 16 01560 g010
Figure 11. Effect of the number of attention heads on the OA, AA and kappa coefficient. (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Figure 11. Effect of the number of attention heads on the OA, AA and kappa coefficient. (a) Indian Pines (IP). (b) Salinas (SA). (c) Pavia University (PU).
Remotesensing 16 01560 g011
Figure 12. Maps depicting the classifications on Indian Pines dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Figure 12. Maps depicting the classifications on Indian Pines dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Remotesensing 16 01560 g012
Figure 13. Maps depicting the classification of Salinas dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Figure 13. Maps depicting the classification of Salinas dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Remotesensing 16 01560 g013
Figure 14. Maps depicting the classification of Pavia University dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Figure 14. Maps depicting the classification of Pavia University dataset using various methods. (a) Ground truth. (b) RF. (c) SVM. (d) 1D-CNN. (e) 2D-CNN. (f) 3D-CNN. (g) HybridSN. (h) GAHT. (i) SpectralFormer. (j) SSFTT. (k) GSC-ViT. (l) MEA-EFFormer.
Remotesensing 16 01560 g014
Table 1. Training and test samples in Indian Pines, Salinas and Pavia University datasets.
Table 1. Training and test samples in Indian Pines, Salinas and Pavia University datasets.
IDIndian PinesSalinasPavia University
Land Cover ClassTrainingTestLand Cover ClassTrainingTestLand Cover ClassTrainingTest
C01Alfalfa343Brocoli_green_weeds_1111998Asphalt676564
C02Corn-notill721356Brocoli_green_weeds_22193707Meadows18718,462
C03Corn-mintill42788Fallow101966Gravel212078
C04Corn12225Fallow_rough_plow71387Trees313033
C05Grass-pasture25458Fallow_smooth142664Painted metal sheets141331
C06Grass-tree37693Stubble203939Bare Soil514978
C07Grass-pasture-mowed226Celery183561Bitumen141316
C08Hay-windrowed24454Grapes_untrained5711,214Self-Blocking Bricks373645
C09Oats119Soil_vinyard_develop326171Shadows10937
C10Soybean-notill49923Corn_senesced_green_weeds173261  
C11Soybean-mintill1232332Lettuce_romaine _4wk61062  
C12Soybean-clean30563Lettuce_romaine_5wk101917  
C13Wheat11194Lettuce_romaine_6wk5911  
C14Woods641201Lettuce_romaine_7wk61064  
C15Buildings-Grass-Trees20366Vinyard_untrained377231  
C16Stone-Steel-Towers588Vinyard_vertical_trellis101797  
Total5139736Total27153,858Total42842,348
Table 2. Ablation experiment results (The optimal results are bolded).
Table 2. Ablation experiment results (The optimal results are bolded).
CasesComponentIndicators
PCAMSECALBPSSEATEOA (%)AA (%) k 100
1×97.2794.1996.88
2×96.9293.0796.48
3×97.0392.6696.61
4×97.1892.0496.78
5×97.1493.6196.74
6×95.3190.9792.24
797.4494.6997.07
Table 3. Ablation experimental results on spectral–spatial enhancement attention module (The optimal results are bolded).
Table 3. Ablation experimental results on spectral–spatial enhancement attention module (The optimal results are bolded).
CasesCombination of BranchesIndicators
Spe-C Spa-WSpe-C Spa-HSpa-W Spa-HOA (%)AA (%) k 100
1×97.1194.4396.76
2×97.0392.6696.61
3×97.1892.0496.78
497.4494.6997.07
Table 4. Classification accuracy of Indian Pines dataset using various methods (optimal results highlighted in bold).
Table 4. Classification accuracy of Indian Pines dataset using various methods (optimal results highlighted in bold).
No.Traditional ClassifiersDeep-Learning-Based Classifiers
RFSVM1D-CNN2D-CNN3D-CNNHybridSNGAHTSpectralFormerSSFTTGSC-ViTMEA-EFFormer
11.95 ± 1.833.41 ± 2.4922.09 ± 6.630.0 ± 7.3943.72 ± 11.7632.79 ± 4.762.56 ± 11.3227.67 ± 6.9372.56 ± 17.1840.7 ± 4.9187.67 ± 7.79
257.73 ± 18.248.14 ± 4.7590.16 ± 1.1690.16 ± 2.0692.19 ± 1.4889.93 ± 2.5294.26 ± 1.3880.06 ± 1.8394.67 ± 0.6995.76 ± 1.1295.19 ± 0.54
334.24 ± 10.3837.99 ± 3.294.94 ± 2.1994.65 ± 2.0596.12 ± 1.5191.04 ± 2.4595.73 ± 2.8891.65 ± 2.6196.5 ± 1.1498.64 ± 0.9898.81 ± 0.69
44.98 ± 1.511.36 ± 1.9382.41 ± 7.2782.63 ± 5.278.66 ± 4.0377.5 ± 6.690.94 ± 2.8268.3 ± 4.4592.9 ± 2.794.91 ± 3.0494.73 ± 2.15
564.46 ± 5.9961.2 ± 2.1798.87 ± 1.4598.67 ± 1.5199.54 ± 0.698.85 ± 0.797.73 ± 3.2697.45 ± 0.899.93 ± 0.198.82 ± 0.8499.98 ± 0.07
655.76 ± 17.5261.62 ± 12.9198.67 ± 0.798.99 ± 0.4299.74 ± 0.1799.06 ± 0.6897.75 ± 1.2297.12 ± 0.5698.87 ± 0.5398.29 ± 0.5899.22 ± 0.48
74.0 ± 5.061.6 ± 1.9654.44 ± 23.4959.26 ± 22.2898.52 ± 1.8155.19 ± 15.9361.48 ± 31.9112.59 ± 7.2694.44 ± 9.8388.89 ± 10.6197.04 ± 6.79
854.98 ± 21.3157.81 ± 12.6899.96 ± 0.0999.87 ± 0.18100.0 ± 0.099.85 ± 0.499.4 ± 0.4998.15 ± 2.399.6 ± 0.6499.98 ± 0.0799.45 ± 0.51
92.22 ± 4.441.11 ± 2.2263.16 ± 12.2361.58 ± 11.0572.63 ± 16.9464.74 ± 16.4886.32 ± 9.7626.32 ± 7.0679.47 ± 10.3880.0 ± 10.9980.53 ± 4.74
1050.79 ± 19.7438.4 ± 5.394.45 ± 1.793.7 ± 1.7791.84 ± 2.5894.21 ± 1.7792.18 ± 9.1386.04 ± 1.2596.46 ± 1.0497.38 ± 1.1297.8 ± 0.52
1175.19 ± 14.4968.72 ± 3.1195.94 ± 1.8895.72 ± 0.8895.02 ± 1.6595.46 ± 0.8897.39 ± 0.5991.89 ± 1.2498.76 ± 0.2998.06 ± 0.4298.73 ± 0.32
1223.11 ± 8.6919.74 ± 3.683.57 ± 6.982.2 ± 4.4881.76 ± 3.2979.86 ± 4.1590.53 ± 2.6766.71 ± 5.5591.47 ± 1.3192.86 ± 1.2390.39 ± 1.77
1318.37 ± 7.5845.0 ± 15.0799.38 ± 0.599.38 ± 0.7998.67 ± 0.798.56 ± 0.9988.41 ± 8.3498.21 ± 1.88100.0 ± 0.097.18 ± 1.38100.0 ± 0.0
1484.84 ± 10.075.72 ± 6.7399.56 ± 0.3199.7 ± 0.2898.5 ± 1.5396.57 ± 1.798.8 ± 0.4998.95 ± 0.6599.17 ± 0.3199.85 ± 0.1099.26 ± 0.32
1513.93 ± 4.6327.75 ± 4.6485.48 ± 5.8283.95 ± 5.9286.57 ± 2.9681.61 ± 3.2491.66 ± 3.5290.0 ± 3.1594.69 ± 3.890.84 ± 3.3195.86 ± 1.65
161.69 ± 1.810.12 ± 3.0192.07 ± 6.3192.76 ± 5.9588.28 ± 6.6885.98 ± 11.7563.91 ± 8.9756.21 ± 5.3385.29 ± 4.2387.13 ± 3.5580.46 ± 6.93
OA (%)56.08 ± 8.2452.67 ± 1.9693.98 ± 0.8493.78 ± 0.5493.91 ± 0.6892.66 ± 0.7395.12 ± 1.2888.57 ± 0.6296.97 ± 0.2897.01 ± 0.3097.44 ± 0.17
AA (%)34.26 ± 4.6435.61 ± 2.2484.70 ± 2.4585.20 ± 2.1988.86 ± 1.7983.82 ± 1.7788.07 ± 2.4274.21 ± 0.8993.43 ± 1.7491.21 ± 1.5094.69 ± 0.57
k × 10049.48 ± 9.0946.02 ± 2.2693.13 ± 0.9692.90 ± 0.6293.05 ± 0.7791.62 ± 0.8494.42 ± 1.4886.92 ± 0.7096.54 ± 0.3296.59 ± 0.3497.07 ± 0.19
Table 5. Classification accuracy of Salinas dataset using various methods (optimal results highlighted in bold).
Table 5. Classification accuracy of Salinas dataset using various methods (optimal results highlighted in bold).
No.Traditional ClassifiersDeep-Learning-Based Classifiers
RFSVM1D-CNN2D-CNN3D-CNNHybridSNGAHTSpectralFormerSSFTTGSC-ViTMEA-EFFormer
123.82 ± 18.2756.82 ± 28.6799.71 ± 0.3799.88 ± 0.1799.99 ± 0.0299.78 ± 0.399.93 ± 0.1699.38 ± 0.399.98 ± 0.0399.32 ± 0.599.98 ± 0.03
250.95 ± 42.4275.92 ± 19.78100.0 ± 0.0100.0 ± 0.0100.0 ± 0.099.85 ± 0.42100.0 ± 0.0100.0 ± 0.0100.0 ± 0.0100.0 ± 0.0100.0 ± 0.0
314.88 ± 22.8957.54 ± 5.599.99 ± 0.02100.0 ± 0.099.99 ± 0.0299.47 ± 0.45100.0 ± 0.099.98 ± 0.0599.98 ± 0.0499.83 ± 0.21100.0 ± 0.0
44.18 ± 6.3234.52 ± 13.4896.28 ± 1.4494.26 ± 2.6893.82 ± 2.7494.86 ± 2.8298.48 ± 1.5394.58 ± 1.4998.33 ± 0.9992.54 ± 4.1797.62 ± 0.99
56.73 ± 8.3259.23 ± 7.0399.95 ± 0.1199.86 ± 0.2199.21 ± 0.6499.21 ± 0.4799.65 ± 0.2498.86 ± 0.4799.33 ± 1.2597.36 ± 0.7699.92 ± 0.14
668.04 ± 34.374.01 ± 5.14100.0 ± 0.099.76 ± 0.799.9 ± 0.2499.94 ± 0.1197.67 ± 0.7199.95 ± 0.0999.85 ± 0.1499.87 ± 0.1699.96 ± 0.07
791.08 ± 2.4579.04 ± 13.6799.97 ± 0.0699.95 ± 0.0699.92 ± 0.1499.58 ± 0.3799.87 ± 0.1199.91 ± 0.0999.16 ± 1.0199.89 ± 0.1299.1 ± 0.62
879.07 ± 16.9967.52 ± 8.1688.55 ± 1.0488.02 ± 1.4790.42 ± 1.7986.82 ± 3.6394.65 ± 0.884.85 ± 0.6993.46 ± 0.9489.99 ± 1.4795.09 ± 0.94
977.78 ± 38.8664.66 ± 15.31100.0 ± 0.0100.0 ± 0.0100.0 ± 0.099.83 ± 0.14100.0 ± 0.0100.0 ± 0.099.99 ± 0.0299.92 ± 0.05100.0 ± 0.0
1033.34 ± 36.6546.85 ± 2.7798.25 ± 0.7497.74 ± 0.9298.79 ± 0.3997.88 ± 1.7899.94 ± 0.0798.23 ± 0.4898.73 ± 0.4998.03 ± 0.4698.36 ± 0.86
1118.51 ± 33.6510.01 ± 11.8399.38 ± 0.4799.13 ± 1.2199.33 ± 0.6498.43 ± 1.6797.61 ± 1.1798.51 ± 0.8199.74 ± 0.2697.06 ± 1.9199.75 ± 0.19
125.41 ± 8.7939.61 ± 13.5499.81 ± 0.2599.84 ± 0.2199.61 ± 0.3798.73 ± 1.0799.66 ± 0.399.81 ± 0.3299.48 ± 0.3795.61 ± 4.2599.39 ± 0.37
1317.27 ± 34.4918.33 ± 12.494.34 ± 10.695.63 ± 5.7894.96 ± 4.192.13 ± 6.5495.15 ± 2.8597.14 ± 1.9495.42 ± 3.0774.4 ± 14.4594.25 ± 2.64
1415.62 ± 31.0931.65 ± 13.2697.82 ± 2.1897.81 ± 2.2897.06 ± 2.1396.57 ± 4.1186.67 ± 1.1496.07 ± 1.6595.86 ± 1.993.03 ± 3.5595.39 ± 2.28
1527.18 ± 19.3856.11 ± 6.9387.09 ± 1.0187.33 ± 1.7285.67 ± 4.2286.27 ± 4.7992.68 ± 1.1685.17 ± 1.4689.49 ± 1.1882.6 ± 2.8291.79 ± 1.15
1621.14 ± 27.8949.0 ± 13.9699.28 ± 0.1699.16 ± 0.1199.09 ± 0.1498.6 ± 1.299.99 ± 0.0299.94 ± 0.0599.49 ± 0.1796.89 ± 1.0399.72 ± 0.19
OA (%)49.29 ± 9.6559.91 ± 3.8995.48 ± 0.3395.32 ± 0.4295.60 ± 0.4294.70 ± 0.6197.26 ± 0.2194.35 ± 0.1996.81 ± 0.2294.19 ± 0.5597.42 ± 0.27
AA (%)34.69 ± 14.3251.30 ± 4.1097.53 ± 0.6797.40 ± 0.4497.36 ± 0.4196.75 ± 0.5397.62 ± 0.2597.02 ± 0.1798.02 ± 0.2394.77 ± 1.1298.15 ± 0.23
k × 10041.70 ± 10.8955.34 ± 4.3394.97 ± 0.3794.79 ± 0.4695.10 ± 0.4894.10 ± 0.6796.95 ± 0.2493.72 ± 0.2196.44 ± 0.2493.53 ± 0.6297.13 ± 0.30
Table 6. Classification accuracy of Pavia University using various methods (optimal results highlighted in bold).
Table 6. Classification accuracy of Pavia University using various methods (optimal results highlighted in bold).
No.Traditional ClassifiersDeep-Learning-Based Classifiers
RFSVM1D-CNN2D-CNN3D-CNNHybridSNGAHTSpectralFormerSSFTTGSC-ViTMEA-EFFormer
163.15 ± 27.6340.51 ± 12.6995.42 ± 2.5296.09 ± 1.1795.43 ± 1.2494.54 ± 2.0796.68 ± 2.3693.78 ± 0.8497.05 ± 1.2189.17 ± 1.6297.37 ± 0.69
293.89 ± 9.4478.54 ± 4.5199.41 ± 0.4799.74 ± 0.1599.52 ± 0.7499.36 ± 0.4799.82 ± 0.1199.77 ± 0.1699.89 ± 0.0599.78 ± 0.1899.94 ± 0.03
34.08 ± 2.8321.1 ± 6.2677.62 ± 6.9682.52 ± 5.7873.12 ± 4.4176.41 ± 4.9387.29 ± 4.3872.25 ± 3.1985.35 ± 6.6576.03 ± 4.4288.67 ± 2.07
419.31 ± 10.739.51 ± 4.9491.32 ± 2.7490.35 ± 1.9887.72 ± 3.0486.44 ± 3.6482.98 ± 3.7284.94 ± 2.3693.65 ± 1.0488.48 ± 1.5193.33 ± 1.22
528.12 ± 26.3548.56 ± 17.3599.76 ± 0.3399.68 ± 0.3799.85 ± 0.1599.81 ± 0.2699.5 ± 0.3598.64 ± 0.4499.55 ± 0.4299.89 ± 0.1599.13 ± 0.48
628.18 ± 25.3731.82 ± 4.7393.84 ± 3.2294.54 ± 1.1797.19 ± 1.7196.18 ± 1.8698.84 ± 1.1991.58 ± 1.5499.5 ± 0.3991.13 ± 2.6899.98 ± 0.04
725.13 ± 19.4415.78 ± 9.693.15 ± 5.2593.12 ± 3.7496.26 ± 3.0894.31 ± 4.0399.02 ± 1.1593.55 ± 2.9999.86 ± 0.2688.56 ± 4.7799.96 ± 0.09
819.26 ± 11.625.26 ± 1.9785.58 ± 4.5782.65 ± 8.2989.18 ± 3.0386.27 ± 4.7895.35 ± 2.6381.06 ± 1.0492.3 ± 2.7778.65 ± 2.5892.85 ± 1.1
910.75 ± 12.8727.65 ± 26.789.36 ± 6.7188.77 ± 4.7992.34 ± 6.6689.1 ± 4.3273.62 ± 4.1871.51 ± 4.9194.4 ± 2.5292.78 ± 3.3392.4 ± 3.08
OA (%)59.18 ± 8.4452.93 ± 1.5994.89 ± 0.7295.12 ± 0.5595.33 ± 0.5094.69 ± 0.5796.40 ± 0.6393.00 ± 0.3397.46 ± 0.3592.82 ± 0.3497.72 ± 0.24
AA (%)32.43 ± 9.5736.53 ± 2.8191.72 ± 1.6291.94 ± 1.0592.29 ± 1.3491.38 ± 1.1692.57 ± 1.0387.45 ± 0.6695.73 ± 0.8689.38 ± 0.8895.96 ± 0.38
k × 10042.45 ± 12.5837.36 ± 2.0093.19 ± 0.9693.50 ± 0.7393.79 ± 0.6792.93 ± 0.7695.21 ± 0.8490.64 ± 0.4596.62 ± 0.4790.42 ± 0.4696.97 ± 0.32
Table 7. Comparison of trainable parameters, testing times and accuracy of approaches based on deep learning on Pavia University dataset (the optimal results are bolded).
Table 7. Comparison of trainable parameters, testing times and accuracy of approaches based on deep learning on Pavia University dataset (the optimal results are bolded).
Deep-Learning-Based Approaches
3D-CNNHybridSNGAHTSpectralFormerSSFTTGSC-ViTMEA-EFFormer
Testing Time (s)6.397.2913.6918.527.2410.858.54
Params. (K)462.486 K797.57946.83128.8148.377.90436.625
OA (%)95.33 ± 0.5094.69 ± 0.5796.40 ± 0.6393.00 ± 0.3397.46 ± 0.3592.82 ± 0.3497.72 ± 0.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, Q.; Zhao, G.; Fang, Y.; Fang, C.; Sun, L.; Li, X. MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification. Remote Sens. 2024, 16, 1560. https://doi.org/10.3390/rs16091560

AMA Style

Sun Q, Zhao G, Fang Y, Fang C, Sun L, Li X. MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification. Remote Sensing. 2024; 16(9):1560. https://doi.org/10.3390/rs16091560

Chicago/Turabian Style

Sun, Qian, Guangrui Zhao, Yu Fang, Chenrong Fang, Le Sun, and Xingying Li. 2024. "MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification" Remote Sensing 16, no. 9: 1560. https://doi.org/10.3390/rs16091560

APA Style

Sun, Q., Zhao, G., Fang, Y., Fang, C., Sun, L., & Li, X. (2024). MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification. Remote Sensing, 16(9), 1560. https://doi.org/10.3390/rs16091560

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop