Next Article in Journal
The Spatiotemporal Changes in Ecological–Environmental Quality Caused by Farmland Consolidation Using Google Earth Engine: A Case Study from Liaoning Province in China
Next Article in Special Issue
Attention-Based Multi-Level Feature Fusion for Object Detection in Remote Sensing Images
Previous Article in Journal
Deep Contrastive Self-Supervised Hashing for Remote Sensing Image Retrieval
Previous Article in Special Issue
A Dual-Generator Translation Network Fusing Texture and Structure Features for SAR and Optical Image Matching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Branch-AttentionNet: A Novel Deep-Learning-Based Spatial-Spectral Attention Methodology for Hyperspectral Data Analysis

by
Bishwas Praveen
1,*,† and
Vineetha Menon
2,†
1
Computer and Information Sciences, University of Alabama in Huntsville, Huntsville, AL 35899, USA
2
Computer Science and the Big Data Analytics Lab, University of Alabama in Huntsville, Huntsville, AL 35899, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2022, 14(15), 3644; https://doi.org/10.3390/rs14153644
Submission received: 8 June 2022 / Revised: 24 July 2022 / Accepted: 26 July 2022 / Published: 29 July 2022

Abstract

:
Recently, deep learning-based classification approaches have made great progress and now dominate a wide range of applications, thanks to their Herculean discriminative feature learning ability. Despite their success, for hyperspectral data analysis, these deep learning based techniques tend to suffer computationally as the magnitude of the data soars. This is mainly because the hyperspectral imagery (HSI) data are multidimensional, as well as giving equal importance to the large amount of temporal and spatial information in the HSI data, despite the redundancy of information in the temporal and spatial domains. Consequently, in literature, this equal information emphasis has proven to affect the classification efficacy negatively in addition to increasing the computational time. As a result, this paper proposes a novel dual branch spatial-spectral attention based classification methodology that is computationally cheap and capable of selectively accentuating cardinal spatial and spectral features while suppressing less useful ones. The theory of feature extraction with 3D-convolutions alongside a gated mechanism for feature weighting using bi-directional long short-term memory is used as a spectral attention mechanism in this architecture. In addition, a union of 3D convolutional neural network (3D-CNN) and a residual network oriented spatial window-based attention mechanism is proposed in this work. To validate the efficacy of our proposed technique, the features collected from these spatial and spectral attention pipelines are transferred to a feed-forward neural network (FNN) for supervised pixel-wise classification of HSI data. The suggested spatial-spectral attention based hyperspectral data analysis and image classification methodology outperform other spatial-only, spectral-only, and spatial-spectral feature extraction based hyperspectral image classification methodologies when compared, according to experimental results.

1. Introduction

Hyperspectral imaging, widely referred to as image spectroscopy, aims at capturing electromagnetic energy emitted off the land cover under observation, over hundreds of contiguous, narrow spectral bands between infrared and visible wavelength ranges. Hyperspectral remote sensing data thus collected with the aid of multiple unmanned aerial devices such as drones, satellites, etc. have strongly favored the emergence of numerous hyperspectral remote sensing applications built around land cover classification for farming [1] for earth observation and monitoring, planning the cities [2], aerial devices based supervision [3], weather forecasting and climate monitoring [4], observations on changes in climatic conditions [5], etc. Over the last decade, breakthroughs in artificial intelligence (especially in machine learning and deep learning) have prompted the hyperspectral remote sensing community to focus on developing efficient hyperspectral image classification approaches for feature extraction [6,7], HSI data classification [6,8] and object detection for HSI based applications [9].
With most of the aforementioned remote sensing application frameworks, classification of HSI data acts as the center piece. While traditional machine learning-based HSI data classification and object detection frameworks rely primarily on spectral information as features [10,11], spectral feature-based classification approaches are unable to capture and use the spatial variability present in high-dimensional HSI data [6,8]. Distance measure [12], KNN (K-Nearest Neighbors) [13], maximum likelihood criteria [14], AdaBoost [15] , random forest classification [16], and logistic regression [17] are some of the frameworks that operate on spectral-only features that have been described in the literature and have proven to be useful in categorizing HSI data. Furthermore, literature has shown that combining a large number of spectral and spatial data in a complementary form can considerably improve the efficacy of HSI classification frameworks [18,19,20].
Over the past few years, deep learning based HSI classification frameworks with stratified feature learning ability have dominated the domain of computer vision, and have produced superlative results in various remote sensing applications [21,22,23]. Not only do these deep learning-based models have the capability to learn more convoluted and abstract features in the data [24,25], they also hold the intrinsic capability of learning higher level features present in the shallow part of the network. Such models are typically not affected by the changes to the input, and have produced unprecedented results in spectral-spatial feature extraction-based HSI classification frameworks [6,8].
Unsupervised and supervised feature extraction and data analysis methodologies applying convolution operation based neural networks, recurrent neural networks, residual networks, and additional models are used in deep learning-based frameworks for HSI classification, which fosters process automation through progressive data-learning. These approaches generally operate on HSI data that are represented in a lower dimensional space in comparison with the raw HSI data which is high dimensional in nature through incorporation of an effective dimensionality reduction technique producing promising results. A few such dimensionality reduction techniques that we come across in literature are principal component analysis (PCA) [26], linear discriminant analysis (LDA) [27], local fisher discriminant analysis (LFDA) [28], random projections (RPs) [6,8,29,30], etc. However, it is evidently noticeable in the hyperspectral remote sensing community that the following questions have not been addressed effectively in literature to date [31]:
  • Do all the spatial and spectral information extracted contribute equally towards effectively classifying HSI data with a deep learning framework?
  • If no, is there a technique to discriminate between less informative and more informative spectral bands that contribute towards effective HSI classification?
  • Is there a strategy that can improve HSI data classification in a comprehensive classification methodology by emphasizing the contribution of data points which happen to be more informative in a local spatial windowed area around the data point of interest?
  • Is it possible to improve the efficacy of an HSI classification methodology based on deep learning by highlighting more informative features and suppressing less valuable ones?
These questions fuel our motivation to devise a novel functional methodology to efficiently emphasize more informative pixels in a spatial neighborhood and spectral bands in the temporal domain which aid in boosting the efficacy of a HSI data classification framework. This phenomenon has widely been recognised as “ATTENTION” in the machine learning and deep learning community [7,32,33]. Attention methodologies have recently been used extensively in language modeling and computer vision applications. Its accomplishment is based mostly on the fair presumption that human vision only pivots on specific sections of the entire visual expanse when and where required. Researchers over the past few years have extensively worked on such attention-based methodologies in the remote sensing community and produced exceptional results. For instance, Zhang et al. in [34] proposed a saliency detection technique to extract salient areas in the input data which can be used for data representation and sampling. In a similar work, Diao et al. [35] proposed an alternative technique where norm gradient maps could be effectively used as saliency maps for data representation. However, Wang et al. in [36] claim that salient areas are not necessarily cardinal in achieving exceptional results in applications related to remote sensing data classification. Following this claim, they proposed to use LSTM networks to recurrently extract attention maps from deep convolutional features of a CNN based architecture. However, combining LSTM and CNN to extract attention features made the proposed methodology too complex and introduces scaling issues [36]. Later, Haut et al. in [37] proposed a two-branch attention module, where one branch is utilized to generate an attention mask and the other branch is utilized to extract convolutional features. Then, attention features are obtained by multiplying the attention mask with the convolutional features which proved to be an effective technique for hyperspectral data analysis and classification. With this phenomenon as an inspiration, we propose dual branch spectral–spatial attention based methodology for hyperspectral image classification. We aim to use attention mechanisms to improve the classification framework’s representation capacity, allowing us to focus on more selective spectral bands and spatial positions while repressing those that are not.
Hence, in this work, we propose a dual branch attention and classification framework for HSI data in order to address the aforementioned problems. To begin, principal component analysis was used as a computationally efficient dimensionality reduction component in this study to effectively extract spectral information, to reduce noise, and to bring down the redundancy in spectral information. Then, using a 3D-convolution based automated feature engineering technique and a bi-directional long short term memory (LSTM) based gated attention mechanism, a spectral attention and feature encoding mechanism that delivers enhanced hyperspectral data learning that emphasizes spectral information that is important for hyperspectral data review. Another significant contribution of this work is the proposal of a 3D-convolutional neural network and softmax activation based spatial attention technique to adaptively diversify spatial information around the pixel of interest into features of higher importance and into features that contribute less towards making classification decisions on input data. This is followed by a ResNet based feature encoding mechanism to encode this spatial information into a feature vector for classification. In addition, to evaluate the efficacy of this automated hyperspectral image classification methodology, a FNN-based supervised classification on the feature vectors derived from the spatial and spectral attention blocks is included.
Therefore, the novel contributions of the proposed dual-branched spatial-spectral attention and classification framework are summarized as follows:
  • A novel framework of spectral attention mechanism and feature encoding scheme using 3D-convolutions for volumetric feature analysis and feature extraction jointly with bi-directional LSTMs for spectral features based attention is proposed for hyperspectral image classification. Bi-directional LSTMs, in our work, offer an effective means to discriminate between features that are of less importance and emphasize more principle spectral information through a gating mechanism with softmax activation function. This technique also presents an effective way to encode spectral information, thus aiding in restoration and enhancement of spectral relationship between spectral bands in hyperspectral imagery data.
  • A unique combination of 3D-CNN based spatial attention mechanism and a ResNet based information encoding scheme is proposed, which not only constructively preserves the spatial relationship between pixels in a windowed region around the input data point, but also accents the weightage for features which have a dominant contribution in the ground truth based decision-making process of a HSI classification framework.
  • This work also proposes a capable technique for combining encoded characteristics from the spectral and spatial attention frameworks before being produced as an input to a feed-forward neural network (FNN) based supervised classification network to demonstrate the efficacy of our approach. In this work, the proposed dual-branched spatial-spectral attention and classification architecture is aimed at improving automated hyperspectral data analysis.
  • This paper also introduces a number of unique spatial-only and spectral-only attention based approaches and frameworks to conduct a brief evaluation on the effects of incorporation of attention methodology in hyperspectral image classification.
The following is a summary of the remainder of this paper: The theoretical description of each component in the proposed design is presented in Section 2. This is continued by a detailed discussion on the proposed architecture in Section 3, followed by a brief discussion about all the other methodologies used in our work as a means to compare the efficacy of our proposed methodology in Section 4, which are then evaluated in Section 5. Finally, in the conclusion Section 6, the analysis on the efficacy of our proposed work is summarized.

2. Approach Overview

This section examines the different components of the proposed dual-branched spatial-spectral attention and classification framework from a theoretical standpoint.

2.1. Principal Component Analysis

Principal Component Analysis (PCA) is a dimensionality technique which is linear in nature used to transfer input data to a new coordinate system such that the first coordinate (which is generally referred to as the first principal component) achieved through scalar projection has the highest variance, the second projected coordinate has the second highest variance, and so on. With this data transformation methodology, the goal is to project every data point to a new coordinate system and retain the first few principal components and preserve the maximum amount of variance possible and discard the rest, which indirectly transforms the input to a lower dimension when compared to that of the high dimensional input. In addition, it is to be noted that the ith principal component (PC) retains the maximum variance of the data projected data in a direction that is orthogonal to the first i 1 principal components [26,38,39].
Let P R n × d be the data with a column-wise empirical mean of 0, n rows representing the number of data vectors, and d columns representing the data’s features or dimensions. The PCA transformation is defined mathematically as a linear combination of d-dimensional basis vectors u ( k ) = ( u 1 , , u d ) ( k ) which maps every row vector p ( i ) of P to a new vector of principal component scores t ( i ) = ( t 1 , , t l ) ( i ) computed as:
t k ( i ) = p ( i ) . u ( k )
where i = { 1 , , n } , k = { 1 , , l } and l d . For the variance to be maximized, the initial weight vector u ( 1 ) which is in direct correlation to the first principal component has to satisfy:
u ( 1 ) = arg m a x u = 1 { Σ ( t 1 ) ( i ) 2 } = arg m a x u = 1 { Σ ( p ( i ) . u ) 2 }
As a result, Equation (2) can be enunciated in the form of a matrix as:
u ( 1 ) = arg m a x u = 1 { P u 2 } = arg m a x u = 1 { u T P T P u }
As u ( 1 ) has been specified as a unit vector, it likewise satisfies (4), as shown below:
u ( 1 ) = arg m a x { u T C m u u T u }
where the highest eigen values (decreasing order) corresponding to the largest eigen vectors u of the data are represented by C m = P T P , which is the covariance matrix of the data. Finally, the initial d r principal components of all the computed D principal components are retained where d r d [6].

2.2. 3D-Convolutional Neural Network

Convolutional neural network (CNN) is a kind of an artificial neural network with a convolution layer as its building block. The integral part of this convolution layer is a mathematical convolution operator, which comprises of a sequence of kernels/filters which has a receptive field that is usually narrow in nature but traverses across the depth of the volume of the input data. CNN-based techniques, like traditional deep learning-based classification frameworks, are hierarchical designs with several convolution layers stacked one on top of the other. In traditional CNN-based frameworks, 2D convolutional kernels are used for any operations that take place on 2D image inputs. 2D convolution operations, on the other hand, are incapable of capturing differentiating features along the spectral/temporal dimension and can only extract spatial information. To overcome this issue, Ji et al. demonstrated the application of a 3D convolution operation on video-based classification which acted along both space (which is two-dimensional in nature) and also along the third dimension, which is time in this case [40]. Since HSI data are usually extracted and processed as a data cube which is three-dimensional in nature, extraction of low level features for HSI data cube can be effectively constructed using 3D-convolutions. As a result, 3D convolutions are considered for learning and capturing 3D local patterns, as well as preserving local spatial dependency around a neighborhood and spectral dependencies across the entire hyperspectral data cube for comprehensive volumetric data analysis, hence enhancing the effectiveness of the underlying framework. The value at position ( p , q , r ) on the jth feature map in the ith layer is produced using 3D convolution, as demonstrated in Equation (5):
h p , q , r i , j = f ( ( W i , j V i 1 ) p , q , r + b i , j )
where the weights and biases for the corresponding jth feature map are denoted by W i , j and b i , j , respectively; V i 1 represents the input feature maps from the ( i 1 ) th layer connected to the current layer; the nonlinear function is denoted using f; and ∗ is the obvious convolution operation. It is also noteworthy that there is a direct relationship between how complex the learned patterns/features are of a data and the total number of 3D convolutional kernels used for the purpose of feature extraction. However, the network usually tends to be adversely affected by over-fitting with a drastic increase in the number of kernels used to extract meaningful patterns from the input data. However, a general notion is that the underlying network should consist of a big enough number of convolutional layers to extract finer and deeper data patterns and lesser number of feature maps in every layer so that it does not adversely take a toll on the overall computational cost (training and testing) [8].

2.3. Bi-Directional Long Short Term Memory

Bi-Directional Long Short Term Memory (LSTM) is an extension of traditional LSTMs and was introduced by Paliwal et al. in 1997 [41]. They are constructed by just putting two independent LSTMs together. This structure inherently facilitates the network to encode and extract information in both the directions, i.e., in the forward direction from time step t 0 , , t n and also in the backward direction from time step t n , , t 0 , where n is the total number of time steps in input data. This encoded information along both the directions is put together and produced as an input to the subsequent layer in the underlying neural network framework. As a result, bi-directional LSTMs have the potential in enhancing the efficacy of information extraction and encoding (spectral information) while applied on high dimensional hyperspectral data, which results in the underlying network to perform robustly and produce better classification results.
Mathematically, let t be the time step, given a minibatch input X t R n × d (where n is the number of samples and d is the number of inputs per example), and let ϕ be the hidden layer activation function, and assuming that H t R n × h is the forward hidden state and H t R n × h is the backward hidden unit for this particular time step, where h is the number of neurons in the hidden layer in the network, the forward and backward hidden state updates are as follows:
H t = ϕ ( X t W x h ( f ) + H t 1 W h h ( f ) + b h ( f ) )
H t = ϕ ( X t W x h ( b ) + H t + 1 W h h ( b ) + b h ( b ) )
where the weights W x h ( f ) R d × h , W h h ( f ) R h × h , W x h ( b ) R d × h , W h h ( b ) R h × h , and biases b h ( f ) R 1 × h and b h ( b ) R 1 × h are all the model parameters.
Consequently, the forward and backward hidden states, H t and H t are concatenated to obtain the resultant hidden state output, H t R n × 2 h . In the case of deep bi-directional LSTMs with multiple hidden layers, H t is produced on as an input to the sequential bi-directional LSTM layer. Lastly, the output layer computes the output O t R n × q (q: number of outputs). It is denoted mathematically in Equation (8) where the output O t is obtained as a result of matrix multiplication between H t R n × 2 h and W h q R 2 h × q , which is finally added to a bias term b q R 1 × q :
O t = H t W h q + b q
In fact, the two directional units (forward and backward) can contain a varied number of hidden units within them in the underlying classification network.

2.4. Residual Neural Networks

A residual neural network (ResNet) is building blocks in a type where skip connections, or shortcuts, are used in ResNets to leap over some layers in the underlying network, simulating this capability of pyramidal cells in the cerebral cortex of a human brain. Most ResNet-based neural network topologies use double- or triple-layer skip connections with a nonlinear activation layer, such as the ReLU and batch normalization between these skip connections [42].
The two principal reasons to introduce skip connections in ResNets are as follows:
  • To effectively avoid and overcome the issue of vanishing/diminishing gradients;
  • To avoid the deterioration (accuracy saturation) problem, which occurs when a deep neural network has too many hidden layers, resulting in increased training errors.
As there are just a few layers to transport gradients through in the backward direction, introducing skip connections effectively simplifies the neural network’s training process, lessening the influence of vanishing gradients.
Mathematically, for a single skip connection, the layers of a neural network can either be indexed as l 2 to l or as l to l + 2 . These indexing paradigms work conveniently while defining skip connections in the backward or forward direction. It is convenient to describe the skip as l + k from the current layer of the neural network as information is propagated forward through the network, but as a learning rule during back propagation, it is more convenient to describe the activation layer that is being reused as l k , where k 1 turns out to be the skip number. Given a weight matrix W l 1 , l , for weights between layer l 1 and l, and a weight matrix W l 2 , l , for weights between layer l 2 and l, the resultant output of the residual connection is defined as shown in Equation (9):
a l = g ( W l 1 , l · a l 1 + b l + W l 2 , l · a l 2 ) = g ( Z l + W l 2 , l · a l 2 )
where
  • a l is the activation output of neurons in layer l,
  • g is the activation function of layer l,
  • W l 1 , l is the weight matrix between layer l 1 and layer l in the neural network, and
  • Z l = W l 1 , l · a l 1 + b l

2.5. Feed-Forward Neural Network

A feed-forward neural network (FNN) is the simplest form of neural networks. It consists of a number of simple neuron-like processing units, usually organised in layers, and there is a connection between any two neurons in adjacent layers. While the preceding statement is true, each of those connections may have different strengths, which are vastly recognised as weights. The weights on these connections encode the knowledge of the defined neural network. In a classification framework, which is the area of focus here, all the FNN weights are modified in a way that, when an input pattern is presented to the network, the output units rightly categorize the input into one of the ground-truth classes present in the data set [43].
The FNN uses a supervised learning approach to know what category the input pattern belongs to. FNNs majorly rely on backpropagation to propagate the gradients from the classification decisions through a mathematically formulated loss function at the end of the forward pass. These propagation of gradients results in alteration of weights in the hidden layers pushing the gross value of the loss function to a global minimum. This technique is repeated a number of times, generally denoted as iterations or epochs, until either the loss function hits a global minimum or the changes in weights are no more significant [43].

3. Proposed Classification Methodology (SPAT-SPEC-HYP-ATTN)

In our proposed methodology, the input high-dimensional hyperspectral cube is dimensionally reduced utilizing principal component analysis (PCA). PCA on raw input data minimizes noise, redundancy in data along the spectral dimension, and makes the process of mapping decision boundaries for classification easier. As a result, the input hyperspectral data cube, with a spatial resolution of ( M × N ) and a temporal resolution/dimension of D, has been reduced in dimension to ( M × N × P ) , where P is the new reduced spectral dimension. Furthermore, the result from PCA is now sent into two parallel attention modules as denoted in Figure 1, one for spatial attention mechanism and the other for spectral attention mechanism where more informative spatial and spectral features are selectively emphasized and the less useful ones are suppressed respectively for a superior classification efficacy. The architectural details of the spatial and spectral attention modules are as follows.

3.1. Spectral Attention Module

In this module, we propose a 3D-convolution based feature extraction and a bi-directional LSTM based spectral attention methodology to emphasize data along spectral bands that are of greater importance in achieving higher classification accuracy and suppressing information from spectral bands that are of lesser importance. The input to the spectral attention module is of dimension ( M × N × P ) from which 3D windows of data around the image pixel of interest ( 11 × 11 in our work) are extracted to preserve closely related dependency between pixels in the windowed spatial area in tandem with the spectrally/temporally constructed features for the purpose of feature engineering with 3D convolutions. As a result, the shape of the modified input for the spectral attention model turns out to be ( 11 × 11 × P ) .
This windowed information is now passed on to a 3D convolution layer as an input with 16 kernels of shape ( 3 × 3 × 32 ) , which is followed immediately by a 3D average pooling layer of size ( 2 × 2 × 2 ) . The resultant features from this pooling layer are forwarded into a second 3D convolution based feature extraction layer with 32 filters of the shape ( 3 × 3 × 8 ) , which efficiently extracts local spatial neighborhood information from the input data whilst conserving spectral band correlation. The output from this phase of the feature extraction and attention module is flattened to produce an output of shape ( L × 1 ) as denoted in Figure 2.
As explained in Equations (6)–(11), this vector of size ( L × 1 ) was successively passed as an input to a bi-directional LSTM-based spectral attention gating mechanism. This attention gating technique highlights relevant informative pixels while suppressing data from less informative spectral regions. Using forward and backward hidden state implementation of a bi-directional LSTM, where H t R n × h represents the forward hidden state implementation and H t R n × h represents the backward hidden state implementation, and their outputs are defined in Equations (6) and (7), where the number of hidden units is represented by h and the number of inputs is denoted by n; the attention gating mechanism is mathematically represented as follows:
The output of the final hidden state output, defined as H t , is obtained as a result of element-wise multiplication between H t and H t , which is depicted in Equation (10). The mathematical calculations outlined in Equations (6), (7), and (10) are performed twice before the attention gating mechanism’s final output O t 3 is achieved, as illustrated in Equations (11) and (12). The notation ⊗ in Equations (10) and (12) denotes element-wise multiplication, and ⊕ in Equation (12) denotes element-wise addition.
For a 50-dimensional feature embedding vector, these produced input data representations are now utilized as an input for a two dense layer based feed forward neural network (FNN) with 100 and 50 nodes in the first and second layer, respectively, with an inclusion of a dropout layer between these two dense layers and has a value of 0.2. The complete proposed spectral feature extraction and bi-directional LSTM-based attention module is denoted in Figure 2:
H t = H t H t
O t 1 = s o f t m a x ( H t )
O t 2 = O t 1 X t and O t 3 = O t 2 X t

3.2. Spatial Attention Module

In this module, we propose a 3D-convolution based spatial attention paradigm which selectively amplifies weightage to specific areas of feature maps through a softmax activation-based gating mechanism which effectively captures the spatial relationship between data pixels in a window, boosting the feature representation capability. In addition, a ResNet-based feature extraction and encoding methodology has been discussed in this work, which effectively extracts spatial information and encodes it into a 50-dimensional vector to achieve superior classification performance.
Firstly, a spatial window of ( 11 × 11 ) around every data point is extracted which transforms the input shape of the spatial attention module to ( 11 × 11 × P ) , from ( M × N × P ) , which was the output shape from the PCA implementation block. Furthermore, this input is forwarded to a 3D-convolution-based feature extraction layer with 1 filter of shape ( 1 × 1 × 25 ) , which is serially followed by a second 3d-convolution layer with 1 filter of shape ( 1 × 1 × 26 ) . The choice of the number of convolutional kernels and their shape are astutely made to ensure that the output spectral dimension of the convolution operations precisely matches the spectral dimension of the input data block. These convolution operations are followed by a softmax activation layer which produces an attention map which produces the probabilistic weightage of pixels in the feature map ranging between 0 and 1. Consequently, the spatial attention gating mechanism proposed in our work is clearly described in Figure 3.
If X ∈∈ R 11 × 11 × P is the input to the spatial attention module, then the output of the first and the second 3D-convolution operations are obtained as shown in Equations (13) and (14):
O 1 = f ( ( W 1 X ) + b 1 )
O 2 = f ( ( W 2 O 1 ) + b 2 )
where O 1 and O 2 are the outputs of the first and the second 3D-convolution operations, respectively, W 1 and W 2 are the weights of the first and second convolution layer in the defined network, the nonlinear activation function is denoted by f, while the convolution operator is denoted by ∗. Furthermore, Equations (15) and (16) denote the gated weighting spatial attention mechanism implemented in our work where O f i n a l is the final output. The notation ⊗ denotes element-wise multiplication and ⊕ denotes element-wise addition as seen in Equation (16):
O 3 = s o f t m a x ( O 2 )
O 4 = O 3 X and O f i n a l = O 4 X
The resultant output from the 3D-CNN based attention mechanism is produced as an input to a set of three identical and sequential residual blocks for spatial feature extraction. An advantage that residual blocks brings to the table is their ability to solve the vanishing gradient problem when the depth of neural network increases by allowing an alternate path through skip connections for the gradient to flow through. In addition, these skip connections aid in facilitating the underlying neural network framework to learn the identity functions which ensure that the deeper layers in the network perform equally as well as the shallow layers in the network. This proposed 3D-CNN-based attention methodology and ResNet-based feature extraction mechanism is shown in Figure 3. The three 2D-convolution layers in every residual block are equipped with 32 feature extraction filters of shape ( 3 × 3 ) , 64 kernels of shape ( 3 × 3 ) and 64 filters with a dimension of ( 1 × 1 ) . The features extracted from the residual blocks are now input to FNN for 50-dimensional feature encoding. The number of layers in the FNN used is two, which consists of 100 neurons in the first hidden layer of the FNN and 50 neurons in the second layer of the FNN network.
Finally, the encoded features from both the spatial attention module and spectral attention module and added together element-wise and input to another FNN for classification are depicted in Figure 1. This FNN based network has two fully connected layers where the first layer has 25 neurons and the second fully connected layer, which is the final layer of the network, consists of C neurons with an added dropout functionality between the two layers with a dropout value of 0.2 . C here denotes the total number of classes present in the datasets used for assessing the efficacy of our proposed methodology.

4. Methodologies for Comparison

4.1. SPEC-HYP-ATTN

There is no explicit spatial attention mechanism in this technique. The goal of this research work is to see how well a hyperspectral image (HSI) classification framework based solely on a light-weighted, computationally efficient spectral attention mechanism performs. As a result, this model uses principal component analysis directly on input data for spectral feature extraction, noise reduction, and spectral information redundancy minimization. Section 5 briefly discusses the experimentally measured size of the lower dimensional space P. Now, data, which are in a lower dimensional feature space, are spatially windowed to extract patches for bi-directional LSTM-based spectral attention and FNN-based classification, where window size is fixed to ( 11 × 11 × P ) . The generalized framework for this methodology SPEC-HYP-ATTN is exactly as depicted in Figure 2 with a minute alteration to the FNN architecture at the tail-end of the architecture. The FNN network in this approach has three dense layers with 100, 50 and C neurons in the first, second, and third layer of FNN, respectively. There also is an inclusion of a dropout of 0.2 in the first two fully connected layers. C here denotes the total number of classes present in the datasets used for assessing the efficacy which is discussed in Section 5 of this work.

4.2. SPAT-HYP-ATTN

In this methodology, we study the contributions of a spatial attention-based classification framework on hyperspectral data analysis using this methodology. In SPAT-HYP-ATTN, a PCA based dimensionality reduction module is integrated to linearly project input HSI data in its original dimensional space to an orthogonally projected lower dimensional space to extract temporal information off the input data and also as a noise reduction technique. Therefore, this PCA based hyperspectral data analysis technique used for noise reduction characterizes the robustness of the proposed methodology especially in conditions where the input data are noisy. The output from PCA is windowed to a shape of ( 11 × 11 × P ) and passed to a 3D-CNN based spatial attention module exactly as depicted in Figure 3 which is forwarded to a ResNet based spatial feature extraction and feature embedding framework. Finally, these encoded features are classified with a FNN-based classification framework. The framework for this methodology is a carbon copy of the architecture presented in Figure 3 with a tiny modification to the FNN at the end of the network. Here, the FNN has three dense layers for classification, with 100 neurons in the first layer, 50 neurons in the second layer, and C number of neurons in the final layer of the FNN-based architecture, with a 0.2 dropout between the first two dense layers. In addition, as in Section 4.1, C here denotes the total number of classes present in the datasets used for assessing the efficacy in this comprehensive HSI data analysis.

4.3. GAB-RP-S-3DCNN

We compare the proposed attention models introduced in this work with a prior work in [8], which developed a unique spatial-spectral feature extraction and HSI classification based analysis model. The methodology in [8] consisted of a spatial feature extraction technique using Gabor filtering on raw hyperspectral data cube in their research. This technique achieved spatial feature extraction by convolving individual temporal bands of HSI data with distinct Gabor filters which differed from one another in terms of frequency responses and orientation information embedded within the individual Gabor filters [44].
In prior work [8], the Gabor filtering greatly varies based on the values chosen for frequency f and the orientation of the Gabor filters denoted by θ , and that the choice of these values majorly impact the results achieved in this HSI data classification and analysis methodology. As a result, the empirical values of f = 0.8 and θ = 0 were picked since they generated the best outcomes in their research. In conjunction, sparse random projections (sparse RPs) are also used for the purpose of dimensionality reduction, and to offer a computationally light-weighted spectral information extraction. In this study, the size of the subspace to which high dimensional HSI data are projected to using sparse RPs is set to 30. Finally, the data are windowed to ( 21 × 21 × 30 ) and fed into a 3D-CNN based classification network.
The initial layer of this 3D-CNN based classification method is a convolutional layer with 8 convolution filters of size ( 3 × 3 × 7 ) which is immediately followed by a max pooling layer with ( 2 × 2 × 2 ) as a dimension. This is followed by two 3D convolution operations with 16 and 32 filters with filter dimensions of ( 3 × 3 × 5 ) and ( 3 × 3 × 3 ) respectively. This part of the classification methodology is now followed by a 2D convolution operation for spatial feature construction and engineering with 64 filters of size ( 3 × 3 ) . The final part of the network consists of three dense layers with 256 nodes in the first layer, 128 nodes in the second layer and C number of nodes in the final layer with a dropout layer with a value of 0.4 between every two of three layers. C here denotes the total number of classes present in the datasets used for assessing the efficacy in this comprehensive HSI data analysis [8].

4.4. SSAN (Spectral-Spatial Attention Network)

Sun et al. in [45] introduce a comprehensive spatial-spectral attention-based methodology for pixel-wise HSI data classification. Their proposed framework has a separate spatial and spectral modules for feature extraction in which 3D-convolutions play a prominent role. The spectral-attention module introduced in this work is comprised of two back-to-back convolution operations with 32 filters each which are of ( 1 × 1 × 7 ) in dimension, with a combination of two layers of batch normalisation and rectified linear units (ReLU) based activation layers. However, the spatial module is composed of two convolutional layers with 64 kernels of shape ( 3 × 3 × q ) where q is the spectral dimension of the input, two batch normalization layers, two ReLU activation layers, and two attention modules after each 3D-convolution layer. Additionally, a 3D-convolution, matrix reshaping and multiplication, matrix addition-based attention module is introduced in this work for a comprehensive analysis on improvements in classification efficacy through the attention mechanism. The window size of HSI data around the pixel of interest is set to ( 7 × 7 ) in this work. The detailed discussion on this generalized classification framework and its results is clearly documented in [45].

4.5. SVM-CK

In this work, we include the state-of-the-art composite kernel oriented support vector machine (SVM) proposed in [20] to test our proposed dual branch spatial-spectral attention-based HSI pixel-wise data classification algorithm. For spatial-spectral information retrieval and classification, this technique combined a spatial mean kernel which is of size ( 3 × 3 ) for spatial feature learning in combination with an RBF spectral kernel for spectral feature analysis and learning. The input for the RBF based spectral kernel is the raw input hyperspectral pixel vector along the temporal domain in SVM-CK, whereas the spatial mean kernel is applied on the input data which is windowed in the shape of ( 3 × 3 ) to effectively include spatial information present around the pixel of interest. LIBSVM is the platform that was used to analyze and conduct all the experimentation related to SVM-CK, and no dimensionality reduction technique was applied on the input data here and instead data in its original format and dimensions were used for experimentation.

5. Experimental Results

The efficacy of our proposed methodology, dual-branch spatial-spectral attention-based HSI classification framework SPAT-SPEC-HYP-ATTN, as defined in Section 3, is validated and compared against five other classification methodologies, namely SPAT-HYP-ATTN, SPEC-HYP-ATTN, GAB-RP-S-3DCNN, SSAN, and SVM-CK in this part of the paper.

5.1. Datasets

All our studies were conducted on three datasets, two of which were collected using an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor and are named Salinas Valley and Indian Pines. The third dataset was collected using a Reflective Optics System Imaging Spectrometer (ROSIS) sensor by flying the sensor over the University of Pavia in Northern Italy [46]. The Salinas dataset consists of 224 spectral bands; however, after deletion of 20 water absorption bands present in the data, the number of spectral dimensions are brought down to 204. Similarly, Indian Pines has 220 spectral bands and, with the removal of 20 water absorption bands, the final spectral dimension turns out to be 200. The Salinas datasets have the spatial resolution of ( 512 × 217 ) with 16 ground-truth classes which includes fruits, vegetables, different phases in growing lettuce, vineyard fields, etc. The Indian Pines dataset has a spatial resolution of ( 145 × 145 ) with 16 classes consisting of alfalfa, different variations of corn vegetation, grass pastures, etc. The ROSIS sensor captures data with a spectral dimension of 103, and each of those spectral bands are made up of a spatial dimension of ( 610 × 340 ) with a spatial resolution of 1.3 m, spanning across nine land-cover categories. As a result, in our proposed work, the value of C is set to 16, 16, and 9 for the Salinas dataset, Indian Pines dataset, and the University of Pavia dataset respectively. The training set for each dataset was picked at random, ranging from 1% to 10% of the total.

5.2. Parameter Tuning and Experimental Setup

In our work, all the parameters that were used in the experimentation phase were carefully chosen and optimized empirically. All the experiments were run thrice and averaged over these three trials to effectively avoid introduction of bias that may be caused by sampling the training and testing data points randomly. Alongside this, the average accuracies produced by experimentation in our work with a detailed documentation of the execution times are made in this study. Categorical cross-entropy is used as the objective function in all of our experiments, with a learning rate of 0.0001 and a decay of 10 6 . All the experiments related to SPAT-SPEC-HYP-ATTN, SPAT-HYP-ATTN, SPEC-HYP-ATTN and GAB-RP-S-3DCNN use the Adams optimizer. The suggested SPAT-SPEC-HYP-ATTN architecture has, in total, 223,820 number of trainable parameters, and a batch size of 32 was determined empirically. On Salinas and Pavia University datasets, the proposed classification frameworks were trained for 60 epochs, whereas the networks were trained for 100 epochs in the case of the Indian Pines dataset. Apart from experiments related to SSAN, no GPUs were made us of while conducting experiments on any of the aforementioned techniques in our research work. All our experiments were conducted on a workstation with an Intel(R) Core(TM) i7-7700HQ processor and 16 GB RAM, whereas SSAN was executed on a workstation with an Intel Core i7-5930K processor with 64 GB RAM and a GeForce GTX Titan X graphics card [45].
Additionally, it is to be noted that the value of P, which is representation of the data in the reduced feature space and the output of the PCA module on the input raw HSI data cube, as denoted in Figure 1, Figure 2 and Figure 3, is empirically set to 50 in all the experiments related to SPAT-SPEC-HYP-ATTN, SPAT-HYP-ATTN and SPEC-HYP-ATTN for all the three datasets as it generated optimal results. In addition, the value of L in Figure 2 is set to be 256 in number for all the three datasets, namely, Salinas, Pavia University and Indian Pines dataset.

5.3. Discussion

The classification maps for 5 % of randomly chosen training samples spanning the proposed dual-branch spatial-spectral attention and classification framework and other data analysis approaches for the Salinas dataset are presented in Figure 4 in this paper. The classification maps for all of the approaches covered in our study for the Pavia University scene are shown in Figure 5. Similarly, the classification maps for the Indian Pines dataset along with their respective classification accuracies are clearly documented in Figure 6. Table 1, Table 2 and Table 3, respectively, show the result of randomly sampling data points for training and testing purposes from every class for a train–test split of 5 % and and 95 % , respectively, used in all the experimentation detailed in Section 3 and Section 4.
When compared to other methodologies, our proposed bi-directional LSTM-based and 3D-CNN based spectral and spatial attention techniques respectively coupled with FNN-based HSI classification framework produce more coherent classification regions in the classification maps and very few misclassifications, as shown in Figure 4, Figure 5 and Figure 6. Additionally, Table 4, Table 5 and Table 6 present the effectiveness of the proposed methodology SPAT-SPEC-HYP-ATTN through their class-wise classification performance evaluation for all the hyperspectral data analysis models in this study for the Salinas, Pavia University, and Indian Pines datasets, respectively. From Table 4, Table 5 and Table 6, it is evident that SPAT-SPEC-HYP-ATTN indeed outperformed other comparison hyperspectral data analysis frameworks in terms of its classification performance.
The overall classification accuracy performance depicted in Figure 7, Figure 8 and Figure 9 for the Salinas dataset and Pavia University dataset, respectively, further affirms the superior classification performance of our proposed methodology, SPAT-SPEC-HYP-ATTN. In particular, SPAT-SPEC-HYP-ATTN outperformed all other HSI data analysis methodologies of comparison used in this work, particularly when compared with the traditional 3D-CNN based feature construction and hyperspectral image classification model, GAB-RP-S-3DCNN, a bi-directional LSTM-based spectral attention-only classification framework, SPEC-HYP-ATTN, and a conventionally used SVM-based spatial-spectral information inclusion model, SVM-CK as demonstrated in the aforementioned figures.
The major reason for the proposed approach SPAT-SPEC-HYP-ATTN exhibits superior classification performance when compared to all the other methodologies used for comparison is that the SPAT-SPEC-HYP-ATTN framework has two separate intrinsic modules for spatial attention and spectral attention which bring forward the enhanced ability of the model to selectively emphasize information, spatially through a 3DCNN-based spatial attention and a ResNet-based feature encoding scheme and spectrally through a bi-directional LSTM-based spectral attention mechanism. These softmax activation function based gating attention-mechanism schemes aid in engineering and weighting features that effectively boost the overall performance of the framework. Even when only 5 % of the samples were used to train the model, our technique SPAT-SPEC-HYP-ATTN achieved the best classification results.
The overall execution time of the approaches presented in our study for the Salinas, Pavia University, and Indian Pines datasets is tabulated in Table 7. Furthermore, Table 8 presents the overall classification results of the proposed approach SPAT-SPEC-HYP-ATTN on varying spatial window sizes along with their execution times for 5 % training data samples for all the datasets used in our study. In addition, to empirically justify that our proposed methodology does not suffer over-fitting due to training with limited number of samples (i.e., 5% in our case), a graphical plot of loss value versus number of epochs during training and validation phases of our experimentation is documented in Figure 10 for all the three datasets. Thus, Figure 10 shows that both the training and validation losses gracefully converge to their optimal values without any over-fitting issues for all the three datasets during training and validation in a limited samples scenario.
Furthermore, Table 8 shows the impact of different window sizes on the proposed technique SPAT-SPEC-HYP-ATTN’s performance and supports our argument for using a window size of (11 × 11), since it produces the best trade-off between computational efficiency and the classification performance that is desired when contrasted with the other state-of-the-art methodologies discussed in our work. Table 7 shows that, when compared to alternative HSI data analysis models discussed in this work, SPAT-SPEC-HYP-ATTN gave improved classification performance to the desired extent.

6. Conclusions

In this work, several novel deep-learning-based spatial-spectral, spatial-only, and spectral-only attention and hyperspectral classification frameworks, namely, SPAT-SPEC-HYP-ATTN, SPAT-HYP-ATTN, and SPEC-HYP-ATTN, respectively, are introduced. Compared to the traditional HSI data analysis and classification methodologies in literature, a gating mechanism-based spatial and spectral attention technique proposed in our work aids in selectively emphasizing important features present in hyperspectral data; at the same time, it suppresses information that is of lesser importance, thus helping the classification framework to produce superior results. When compared to other spatial attention- and spectral-attention-only based classification approaches, the proposed deep-learning-based spatial-spectral attention and classification methodology SPAT-SPEC-HYP-ATTN yields outstanding classification performance while being robust under limited training sample scenarios, paving the way for new research avenues in hyperspectral remote sensing.

Author Contributions

Conceptualization, B.P. and V.M.; methodology, B.P.; software, B.P.; validation, B.P. and V.M.; formal analysis, B.P. and V.M.; investigation, B.P.; resources, B.P.; data curation, B.P.; writing—original draft preparation, B.P. and V.M.; writing—review and editing, V.M.; visualization, B.P.; supervision, V.M.; project administration, V.M.; funding acquisition, V.M. All authors have read and agreed to the published version of the manuscript.

Funding

Publication costs were supported by the Army Research Laboratory and were accomplished under Cooperative Agreement No. W911NF-21-2-0266.

Data Availability Statement

Statistical and computational models used are fully detailed in the main text. All datasets used are publicly available at the url: http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 1 June 2022).

Acknowledgments

This research was sponsored by the Army Research Laboratory and was accomplished under Cooperative Agreement Number W911NF-21-2-0266. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the Army Research Office or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright notation herein.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, K.; Cheng, T.; Deng, X.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W. Assessment of spectral variation between rice canopy components using spectral feature analysis of near-ground hyperspectral imaging data. In Proceedings of the 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016. INSPEC Accession Number 17261771. [Google Scholar]
  2. Abbate, G.; Fiumi, L.; De Lorenzo, C.; Vintila, R. Evaluation of remote sensing data for urban planning. Applicative examples by means of multispectral and hyperspectral data. In Proceedings of the GRSS/ISPRS Joint Workshop on Remote Sensing and Data Fusion over Urban Areas, Berlin, Germany, 22–23 May 2003; pp. 201–205. [Google Scholar]
  3. Vakil, M.I.; Megherbi, D.B.; Malas, J.A. An efficient multi-stage hyper-spectral aerial image registration technique in the presence of differential spatial and temporal sensor uncertainty with application to large critical infrastructures and key resources (CIKR) surveillance. In Proceedings of the IEEE International Symposium on Technologies for Homeland Security (HST), Waltham, MA, USA, 14–16 April 2015; pp. 1–6. [Google Scholar]
  4. Wickert, L.M.; Percival, J.B.; Morris, W.A.; Harris, J.R. XRD and infrared spectroscopic validation of weathering surfaces from ultramafic and mafic lithologies examined using hyperspectral imagery, Cross Lake Area, Cape Smith Belt, Northern Quebec, Canada. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Boston, MA, USA, 7–11 July 2008; Volume 3, p. III-362. [Google Scholar]
  5. Heldens, W.; Esch, T.; Heiden, U. Supporting urban micro climate modelling with airborne hyperspectral data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 1598–1601. [Google Scholar]
  6. Praveen, B.; Menon, V. Novel deep-learning-based spatial-spectral feature extraction for hyperspectral remote sensing applications. In Proceedings of the IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 5444–5452. [Google Scholar]
  7. Praveen, B.; Menon, V. A Bidirectional Deep-Learning-Based Spectral Attention Mechanism for Hyperspectral Data Classification. Remote Sens. 2022, 14, 217. [Google Scholar] [CrossRef]
  8. Praveen, B.; Menon, V. Study of spatial–spectral feature extraction frameworks with 3D convolutional neural network for robust hyperspectral imagery classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1717–1727. [Google Scholar] [CrossRef]
  9. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Loupos, K. Deep learning-based man-made object detection from hyperspectral data. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 14–16 December 2015; pp. 717–727. [Google Scholar]
  10. Mojaradi, B.; Abrishami-Moghaddam, H.; Zoej, M.J.V.; Duin, R.P. Dimensionality reduction of hyperspectral data via spectral feature extraction. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2091–2105. [Google Scholar] [CrossRef]
  11. Chang, C.I. Hyperspectral Imaging: Techniques for Spectral Detection and Classification; Springer Science and Business Media: Mexico City, Mexico, 2003; p. 1. [Google Scholar]
  12. Du, Q.; Chang, C.I. A linear constrained distance-based discriminant analysis for hyperspectral image classification. Pattern Recognit. 2001, 34, 361–373. [Google Scholar] [CrossRef]
  13. Samaniego, L.; Bárdossy, A.; Schulz, K. Supervised classification of remotely sensed imagery using a modified k-NN technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  14. Ediriwickrema, J.; Khorram, S. Hierarchical maximum-likelihood classification for improved accuracies. IEEE Trans. Geosci. Remote Sens. 1997, 35, 810–816. [Google Scholar] [CrossRef]
  15. Chan, J.C.W.; Paelinckx, D. Evaluation of Random Forest and Adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens. Environ. 2008, 112, 2999–3011. [Google Scholar] [CrossRef]
  16. Xia, J.; Bombrun, L.; Berthoumieu, Y.; Germain, C.; Du, P. Spectral–spatial rotation forest for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4605–4613. [Google Scholar] [CrossRef] [Green Version]
  17. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised hyperspectral image segmentation using multinomial logistic regression with active learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar] [CrossRef] [Green Version]
  18. Qian, Y.; Ye, M.; Zhou, J. Hyperspectral image classification based on structured sparse logistic regression and three-dimensional wavelet texture features. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2276–2291. [Google Scholar] [CrossRef] [Green Version]
  19. Cao, X.; Xu, L.; Meng, D.; Zhao, Q.; Xu, Z. Integration of three-dimensional discrete wavelet transform and Markov random field for hyperspectral image classification. Neurocomputing 2017, 226, 90–100. [Google Scholar] [CrossRef]
  20. Menon, V.; Prasad, S.; Fowler, J.E. Hyperspectral classification using a composite kernel driven by nearest-neighbor spatial features. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 2100–2104. [Google Scholar]
  21. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  22. Liu, H.; Li, W.; Xia, X.G.; Zhang, M.; Gao, C.Z.; Tao, R. Central attention network for hyperspectral imagery classification. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef]
  23. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  24. Acharya, D.B.; Zhang, H. Data Points Clustering via Gumbel Softmax. SN Comput. Sci. 2021, 2, 1–13. [Google Scholar] [CrossRef]
  25. Acharya, D.B.; Zhang, H. Feature selection and extraction for graph neural networks. In Proceedings of the 2020 ACM Southeast Conference, Tampa, FL, USA, 2–4 April 2020; pp. 252–255. [Google Scholar]
  26. Jolliffe, I.T. Principal Components in Regression Analysis; Springer: New York, NY, USA, 1986; pp. 129–155. [Google Scholar]
  27. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of hyperspectral images with regularized linear discriminant analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  28. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  29. Menon, V.; Du, Q.; Christopher, S. Improved Random Projection with K-Means Clustering for Hyperspectral Image Classification. IEEE Int. Geosci. Remote Sens. Symp. 2017, 14, 4768–4771. [Google Scholar]
  30. Xu, J.L.; Esquerre, C.; Sun, D.W. Methods for performing dimensionality reduction in hyperspectral image classification. J. Infrared Spectrosc. 2018, 26, 61–75. [Google Scholar] [CrossRef]
  31. Mou, L.; Zhu, X.X. Learning to pay attention on spectral domain: A spectral attention module-based convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 110–122. [Google Scholar] [CrossRef]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  33. Praveen, B.; Menon, V. HYPER-VIT : A novel light-weighted visual transformer-based supervised classification framework for hyperspectral remote sensing applications. In Proceedings of the 12th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Rome, Italy, 13–16 September 2022. [Google Scholar]
  34. Zhang, F.; Du, B.; Zhang, L. Saliency-guided unsupervised feature learning for scene classification. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2175–2184. [Google Scholar] [CrossRef]
  35. Diao, W.; Sun, X.; Zheng, X.; Dou, F.; Wang, H.; Fu, K. Efficient saliency-based object detection in remote sensing images using deep belief networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 137–141. [Google Scholar] [CrossRef]
  36. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene classification with recurrent attention of VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1155–1167. [Google Scholar] [CrossRef]
  37. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Visual attention-driven hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  38. Wang, F.; Zhang, R.; Wu, Q. Hyperspectral image classification based on PCA network. In Proceedings of the Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016. INSPEC Accession Number 17261748. [Google Scholar]
  39. Deepa, P.; Thilagavathi, K. Data reduction techniques of hyperspectral images: A comparative study. In Proceedings of the 3rd International Conference on Signal Processing, Communication and Networking (ICSCN), Chennai, India, 26–28 March 2015; pp. 1–6. [Google Scholar]
  40. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Dreyfus, G. Neural Networks: Methodology and Applications; Springer Science and Business Media: New York, NY, USA, 2005. [Google Scholar]
  44. Zhu, Z.; Jia, S.; He, S.; Sun, Y.; Ji, Z.; Shen, L. Three-dimensional Gabor feature extraction for hyperspectral imagery classification using a memetic framework. Inf. Sci. 2015, 298, 274–287. [Google Scholar] [CrossRef]
  45. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3232–3245. [Google Scholar] [CrossRef]
  46. Gamba, P. A collection of data for urban area characterization. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Anchorage, AK, USA, 20–24 September 2004; p. 1. [Google Scholar]
Figure 1. The proposed spatial-spectral attention-based classification framework (SPAT-SPEC-HYP-ATTN).
Figure 1. The proposed spatial-spectral attention-based classification framework (SPAT-SPEC-HYP-ATTN).
Remotesensing 14 03644 g001
Figure 2. The architecture of the proposed bi-directional LSTM-based spectral attention module from Figure 1.
Figure 2. The architecture of the proposed bi-directional LSTM-based spectral attention module from Figure 1.
Remotesensing 14 03644 g002
Figure 3. The architecture of the proposed 3D-CNN-based spatial attention and ResNet-based feature extraction module from Figure 1.
Figure 3. The architecture of the proposed 3D-CNN-based spatial attention and ResNet-based feature extraction module from Figure 1.
Remotesensing 14 03644 g003
Figure 4. Classification maps of Salinas dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 99.01 % ); (c) SPAT-HYP-ATTN ( 98.00 % ); (d) SPEC-HYP-ATTN ( 96.10 % ); (e) GAB-RP-S-3DCNN ( 97.24 % ); (f) SSAN ( 98.24 % ); (g) SVM-CK ( 93.36 % ).
Figure 4. Classification maps of Salinas dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 99.01 % ); (c) SPAT-HYP-ATTN ( 98.00 % ); (d) SPEC-HYP-ATTN ( 96.10 % ); (e) GAB-RP-S-3DCNN ( 97.24 % ); (f) SSAN ( 98.24 % ); (g) SVM-CK ( 93.36 % ).
Remotesensing 14 03644 g004
Figure 5. Classification maps of Pavia University dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 99.52 % ); (c) SPAT-HYP-ATTN ( 98.33 % ); (d) SPEC-HYP-ATTN ( 96.77 % ); (e) GAB-RP-S-3DCNN ( 97.13 % ); (f) SSAN ( 99.25 % ); (g) SVM-CK ( 91.62 % ).
Figure 5. Classification maps of Pavia University dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 99.52 % ); (c) SPAT-HYP-ATTN ( 98.33 % ); (d) SPEC-HYP-ATTN ( 96.77 % ); (e) GAB-RP-S-3DCNN ( 97.13 % ); (f) SSAN ( 99.25 % ); (g) SVM-CK ( 91.62 % ).
Remotesensing 14 03644 g005
Figure 6. Classification maps of Indian Pines dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 96.21 % ); (c) SPAT-HYP-ATTN ( 94.06 % ); (d) SPEC-HYP-ATTN ( 83.86 % ); (e) GAB-RP-S-3DCNN ( 91.57 % ); (f) SSAN ( 88.71 % ); (g) SVM-CK ( 87.93 % ).
Figure 6. Classification maps of Indian Pines dataset for the proposed dual-branch spatial-spectral attention-based classification framework and methodologies used for comparison using 5% training data. (a) Ground Truth; (b) SPAT-SPEC-HYP-ATTN ( 96.21 % ); (c) SPAT-HYP-ATTN ( 94.06 % ); (d) SPEC-HYP-ATTN ( 83.86 % ); (e) GAB-RP-S-3DCNN ( 91.57 % ); (f) SSAN ( 88.71 % ); (g) SVM-CK ( 87.93 % ).
Remotesensing 14 03644 g006
Figure 7. Overall classification accuracies of the Salinas dataset for varying size of training samples.
Figure 7. Overall classification accuracies of the Salinas dataset for varying size of training samples.
Remotesensing 14 03644 g007
Figure 8. Overall classification accuracies of Pavia University dataset for varying size of training samples.
Figure 8. Overall classification accuracies of Pavia University dataset for varying size of training samples.
Remotesensing 14 03644 g008
Figure 9. Overall classification accuracies of Indian Pines dataset for varying size of training samples.
Figure 9. Overall classification accuracies of Indian Pines dataset for varying size of training samples.
Remotesensing 14 03644 g009
Figure 10. Plots of training loss versus validation loss for all the three datasets used during experimentation. (a) Salinas (Epochs: 60); (b) Pavia University (Epochs: 60); (c) Indian Pines (Epochs: 100).
Figure 10. Plots of training loss versus validation loss for all the three datasets used during experimentation. (a) Salinas (Epochs: 60); (b) Pavia University (Epochs: 60); (c) Indian Pines (Epochs: 100).
Remotesensing 14 03644 g010
Table 1. Total number of class-specific training and testing samples used for a Salinas dataset for 5% training data.
Table 1. Total number of class-specific training and testing samples used for a Salinas dataset for 5% training data.
#Class Name# of Training Samples# of Testing Samples
1Brocoli-green-weeds-11001909
2Brocoli-green-weeds-21863540
3Fallow991877
4Fallow-rough-plow701324
5Fallow-smooth1342544
6Stubble1983761
7Celery1793400
8Grapes-untrained56410,707
9Soil-vinyard-develop3105893
10Corn-senesced-green-weeds1643114
11Lettuce-romaine-4wk531015
12Lettuce-romaine-5wk961831
13Lettuce-romaine-6wk46870
14Lettuce-romaine-7wk541016
15Vinyard-untrained3636905
16Vinyard-vertical-trellis901717
Total2706 51,423
Table 2. Total number of class-specific training and testing samples used for Pavia University dataset for 5% training data.
Table 2. Total number of class-specific training and testing samples used for Pavia University dataset for 5% training data.
#Class Name# of Training Samples# of Testing Samples
1Asphalt3326299
2Meadows93217,717
3Gravel1051994
4Trees1532911
5Painted Metal Sheets671278
6Bare Soil2514778
7Bitumen671263
8Self-Blocking Bricks1843498
9Shadows47900
Total213840,638
Table 3. Total number of class-specific training and testing samples used for Indian Pines dataset for 5% training data.
Table 3. Total number of class-specific training and testing samples used for Indian Pines dataset for 5% training data.
#Class Name# of Training Samples# of Testing Samples
1Alfalfa244
2Corn-notill711357
3Corn-mintill41789
4Corn12225
5Grass-pasture24459
6Grass-trees37693
7Grass-pasture-mowed127
8Hay-Windrowed24454
9Oats119
10Soybean-notill49923
11Soybean-mintill1232332
12Soybean-clean30563
13Wheat10195
14Woods631202
15Buildings-Grass-Trees-Drives19367
16Stone-Steel-Towers588
Total512 9737
Table 4. Class-specific accuracies of the Salinas dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
Table 4. Class-specific accuracies of the Salinas dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
#Class NameSPAT-SPEC-HYP-ATTNSPAT-HYP-ATTNSPEC-HYP-ATTNGAB-RP-S-3DCNNSSANSVM-CK
1Brocoli-green-weeds-1100 99.910099.910099.6
2Brocoli-green-weeds-210010010010010099.8
3Fallow10010099.910010099.9
4Fallow-rough-plow98.999.199.599.599.698.3
5Fallow-smooth99.799.499.399.599.797.6
6Stubble10010010010010099.8
7Celery99.910010010010099.1
8Grapes-untrained99.495.592.394.096.290.3
9Soil-vinyard-develop10010099.899.810098.6
10Corn-senesced-green-weeds99.699.599.099.099.998.7
11Lettuce-romaine-4wk99.910099.199.099.899.5
12Lettuce-romaine-5wk99.910099.899.999.999.5
13Lettuce-romaine-6wk99.999.792.899.699.498.3
14Lettuce-romaine-7wk98.199.592.299.399.196.5
15Vinyard-untrained10093.887.290.294.083.2
16Vinyard-vertical-trellis10099.899.489.799.899.3
OA (%)99.0198.0096.1097.2498.2493.36
κ (%)98.9497.9095.8296.9398.1692.37
Table 5. Class-specific accuracies of Pavia University dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
Table 5. Class-specific accuracies of Pavia University dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
#Class NameSPAT-SPEC-HYP-ATTNSPAT-HYP-ATTNSPEC-HYP-ATTNGAB-RP-S-3DCNNSSANSVM-CK
1Asphalt99.0 99.097.796.897.891.1
2Meadows10099.898.998.910094.7
3Gravel98.895.386.391.694.686.5
4Trees99.198.998.597.297.985.7
5Painted Metal Sheets10099.410099.199.499.4
6Bare Soil99.910096.096.599.990.9
7Bitumen99.899.895.699.898.993.8
8Self-Blocking Bricks98.795.987.992.195.882.7
9Shadows97.799.498.897.198.097.5
OA (%)99.5298.3396.7797.1399.2591.62
κ (%)99.3898.1296.0296.5398.8791.03
Table 6. Class-specific accuracies of Indian Pines dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
Table 6. Class-specific accuracies of Indian Pines dataset for 5% of training data for the proposed classification methodologies and models used for comparison.
#Class NameSPAT-SPEC-HYP-ATTNSPAT-HYP-ATTNSPEC-HYP-ATTNGAB-RP-S-3DCNNSSANSVM-CK
1Alfalfa61.186.292.882.894.389.6
2Corn-notill96.0 94.486.982.792.189.4
3Corn-mintill94.783.584.178.283.489.5
4Corn95.492.277.979.482.692.3
5Grass-pasture95.694.199.394.894.790.3
6Grass-trees97.797.865.098.793.680.1
7Grass-pasture-mowed89.457.880.687.992.192.4
8Hay-Windrowed96.899.470.798.575.287.0
9Oats75.268.387.552.688.886.1
10Soybean-notill93.295.789.483.491.489.4
11Soybean-mintill87.392.094.083.891.591.7
12Soybean-clean99.495.495.696.790.784.5
13Wheat98.998.398.594.692.593.0
14Woods96.597.686.687.193.889.2
15Buildings-Grass-Trees-Drives94.398.183.781.490.090.7
16Stone-Steel-Towers95.893.397.896.686.491.9
OA (%)96.2194.0683.8691.5788.7187.93
κ (%)95.7293.6683.0190.4588.1386.34
Table 7. Overall execution time (in minutes) of all the methodologies in comparison for 5% training data.
Table 7. Overall execution time (in minutes) of all the methodologies in comparison for 5% training data.
DatasetSPAT-SPEC-SPAT-SPEC-GAB-RP-SSANSVM-
(5% Training)HYP-ATTNHYP-ATTNHYP-ATTNS-3DCNN CK
# of Trainable Parameters224,214150,84615,957313,472180,930-
Salinas103.3135.7728.1840.3686.2814.42
Pavia University103.7933.9722.9566.2053.2514.29
Indian Pines60.5118.6713.7121.7516.9513.46
Table 8. Classification results of the proposed SPAT-SPEC-HYP-ATTN methodology on varying spatial window sizes for 5% training data.
Table 8. Classification results of the proposed SPAT-SPEC-HYP-ATTN methodology on varying spatial window sizes for 5% training data.
Salinas DatasetPavia University DatasetIndian Pines Dataset
Window SizeAccuracy (%)Time (min)Accuracy (%)Time (min)Accuracy (%)Time (min)
( 3 × 3 ) 96.3820.5396.6214.7781.294.01
( 5 × 7 ) 96.8724.2297.5418.9682.394.53
( 7 × 7 ) 97.9347.3898.6030.6091.1414.33
( 9 × 9 ) 98.7145.0099.3359.0593.8020.28
( 11 × 11 ) 99.01103.3199.52103.7996.2160.51
( 13 × 13 ) 99.83246.1999.67290.6097.3299.08
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Praveen, B.; Menon, V. Dual-Branch-AttentionNet: A Novel Deep-Learning-Based Spatial-Spectral Attention Methodology for Hyperspectral Data Analysis. Remote Sens. 2022, 14, 3644. https://doi.org/10.3390/rs14153644

AMA Style

Praveen B, Menon V. Dual-Branch-AttentionNet: A Novel Deep-Learning-Based Spatial-Spectral Attention Methodology for Hyperspectral Data Analysis. Remote Sensing. 2022; 14(15):3644. https://doi.org/10.3390/rs14153644

Chicago/Turabian Style

Praveen, Bishwas, and Vineetha Menon. 2022. "Dual-Branch-AttentionNet: A Novel Deep-Learning-Based Spatial-Spectral Attention Methodology for Hyperspectral Data Analysis" Remote Sensing 14, no. 15: 3644. https://doi.org/10.3390/rs14153644

APA Style

Praveen, B., & Menon, V. (2022). Dual-Branch-AttentionNet: A Novel Deep-Learning-Based Spatial-Spectral Attention Methodology for Hyperspectral Data Analysis. Remote Sensing, 14(15), 3644. https://doi.org/10.3390/rs14153644

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop