Next Article in Journal
Social–Ecological Factors and Ecosystem Service Trade-Offs/Synergies in Vegetation Change Zones of Qilian Mountain National Park During 2000–2020
Previous Article in Journal
Case Study on the Use of an Unmanned Aerial System and Terrestrial Laser Scanner Combination Analysis Based on Slope Anchor Damage Factors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Reconstruction Based on Blur–Kernel–Prior and Spatial–Spectral Attention

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of the Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(8), 1401; https://doi.org/10.3390/rs17081401
Submission received: 15 January 2025 / Revised: 21 March 2025 / Accepted: 24 March 2025 / Published: 15 April 2025

Abstract

:
Given the problem of spatial detail loss and spectral feature degradation in hyperspectral images (HSIs) characterized as blur, often caused by noise during image acquisition, and methods of removing blur noise designed on HSIs being insufficient, we propose an HSI reconstruction network based on a Blur–Kernel–Prior (BKP) method and Spectral–Spatial Attention (SSA) strategy for noise removal and reconstruction of HSIs. Specifically, a grouping strategy is designed to segment the HSIs into spectral dimension sub-images, and the BKP module, based on U-Net, learns the spatially adaptive blur kernel to extract and remove blurred features from each sub-image while preserving spatial features with spatial resolution. Subsequently, the SSA block is employed to extract shallow features, details, and edge information using a hybrid 2D–3D convolution from the sub-images, followed by deep feature extraction using a deep ResNet and multi-head attention (MSA) on the merged image to maximize the preservation of spectral dimension information. The L 1 loss function, combined with spectral dimension loss and peak signal-to-noise ratio loss, is utilized to constrain and ensure reconstruction accuracy. Experiments on both synthetic and real datasets demonstrate that our method exhibits excellent performance in reconstructing HSIs affected by blurred noise, outperforming existing methods in terms of quantitative quality and recovery of spectral dimension information.

1. Introduction

Unlike RGB and multispectral images (MSIs), hyperspectral images (HSIs) provide more spectral bands and rich spectral information, revealing detailed intrinsic properties of real-world objects and subtle differences that are often difficult to detect in RGB and MSIs. As a result, HSIs are extensively utilized in various domains, including computer vision and remote sensing, with applications in classification [1,2,3], medical diagnosis [4], detection [5,6], and geological exploration [7]. It is essential that the degradation characterized as blur caused by external and internal noise simultaneously affects the spatial and spectral dimensions of HSIs. In particular, during the acquisition process, factors such as platform vibration [8], scanning errors [9], atmospheric turbulence [10], and sensor electronic noise [11] can lead to the loss of spatial details and the smoothing of spectral features, further impacting related applications. Therefore, the removal of blur noise and the reconstruction that enhances spatial resolution are critical preprocessing steps necessary to improve the quality of HSIs and augment their practical utility [12,13,14].
Unlike machine-learning using a low-rank matrix, tensor representations [2,15] and Singular Value Decomposition (SVD) [16] which result in prolonged processing times and often yields unstable outcomes that are highly sensitive to the configuration of regularization parameters [17,18], deep-learning breakthroughs have revolutionized HIS reconstruction on deblurring through data-driven feature learning [19]. One of the early contributions in this domain is the SSF–CNN method [20], which employs convolution neural networks (CNNs) combined with techniques such as residual learning, dilated convolution, and multi-channel filtering to effectively denoise HSI data. Similarly, a 3D-CNN method [21] was proposed, which integrates 2D CNNs for spectral feature extraction and 3D CNNs with larger convolution kernels to expand the receptive field. The objective of this network is to remove noise by learning cross-band spectral features of HSIs, thereby addressing image degradation. However, while this design allows the trained network to generalize to new images with different spectral bands, it does not fully utilize all spectral bands during the denoising process. Building on these foundations, various reconstructions on deblur algorithms based on 2D CNNs or 3D CNNs have been proposed [22,23,24,25]. For instance, Nonlocal-CNN [24] groups nonlocal similar patches by leveraging the high spectral correlation of HSIs data, using subspace representations where the corresponding representation coefficients are termed eigenimages. Hou et al. [25] introduced a cooperative self-supervised CNN combined with non-convex regularization, which enhances the local smoothness of extracted features while preserving sharp edges. Despite their effectiveness, CNNs often overemphasize local contextual information while neglecting long-range dependencies within the data. Additionally, CNN-based methods that limit receptive field (<15 × 15 pixels) compromised long-range dependency modeling typically incur >15% mutual information loss across spectral bands due to localized feature extraction mechanisms [25,26].
The attention mechanism, initially introduced for sequential data analysis in natural language processing (NLP), has shown great potential in addressing the limitations of CNNs [27,28]. Unlike CNNs, the attention mechanism excels in capturing global information by focusing on relevant features while processing data sequentially and has been applied successfully for RGB image denoising and reconstruction [29]. However, when applied to high-dimensional data such as HSIs, the attention mechanism may encounter challenges due to the large number of parameters in the network. This can result in slow training processes and reduced effectiveness in tasks like HSI denoising and reconstruction. To overcome these challenges, researchers have explored various strategies. One approach involves applying downsampling methods to HSIs with blurred noise, which helps reduce the number of parameters through CNN layers while still effectively capturing global features [26]. Another strategy focuses on the correlation between spectral bands, employing a band-attention mechanism to lower training costs while preserving essential spectral information [30,31,32]. These advancements demonstrate how optimized attention mechanisms can strike a balance between computational efficiency and performance, making them a promising solution for HSI denoising and reconstruction tasks.
One of the significant challenges in HSI reconstruction on deblur is the presence of various types of noise introduced by imaging sensors and acquisition conditions. Most deep-learning algorithms typically limit their degradation models to address only resolution degradation of input images due to inherent defects in the sensor and optical system, so that relative research often uses bicubic interpolation for downsampling [19,20,33]. However, real-world HSI degradation involves not only resolution loss but also blurring introduced during acquisition. These issues arise from factors such as scanning errors, atmospheric turbulence, and electronic noise, manifesting as a mixture of multiple blurring noises, including motion blur and Gaussian blur. Currently, there is no deep-learning-based reconstruction on the deblur method specifically targeting the challenges posed by low spatial resolution HSIs contaminated by such noise. Recent studies have shown that as image data passes through deep neural networks, the original noise may be amplified [17,34]. This can lead the network to extract features from noisy textures, resulting in inaccurate outcomes. Therefore, our research focuses on mitigating noise in HSIs, which is expected to improve the accuracy of subsequent feature recognition and ground object classification.
To tackle the issues previously mentioned, this paper introduces an HSI reconstruction network grounded in spatial prior denoising, which is designated as BKPSSAnet. The primary contributions of this paper are outlined as follows:
  • We present a novel HSI reconstruction method that leverages a Blur–Kernel–Prior neural network for the denoising and reconstruction of HSIs affected by mixed blur noise, marking the first application of this technique in the field.
  • The architecture incorporates two distinct modules: a Blur–Kernel–Prior denoising module-based U-Net backbone and a spatial attention feature-reconstruction module. The module implements an end-to-end encoding-decoding process utilizing a dual U-Net architecture, which effectively isolates the blur kernel from the original image. Meanwhile, the spatial attention feature-reconstruction module employs hybrid 2D–3D convolution provided shallow feature extraction locally and attention mechanisms with multi-head attention to facilitate comprehensive deep feature extraction across both spatial and spectral dimensions globally.
  • Experimental evaluations conducted on the Cave, Chikusei, and Pavia University datasets, which include mixed simulated noise, demonstrate that our research surpasses the state-of-the-art of deep-learning-based approaches across different image-quality metrics.

2. Related Work

In this section, a review of the algorithms is provided for deblurring noise and image reconstruction in HSIs.

2.1. Traditional Image Blind Deconvolution Model

The classical approach to addressing the blind deconvolution problem in blurry images adopts a joint optimization framework in which both the blur kernel and the latent image are updated with iteration to minimize a loss function that integrates prior knowledge of the kernel and the image. For cases involving high noise levels, a combination of L 1 regularization and TV priors is employed to effectively suppress noise and preserve structural details [35,36]. This joint optimization framework proves advantageous in mitigating degenerate solutions, such as the trivial case where ( x * , h * ) = ( y , I ) , with I representing the identity operator. Moreover, some approaches explicitly model the blur kernel by considering the camera’s motion trajectory, leveraging optimization techniques or supervised-learning methods to recover both the trajectory and the underlying clean image.
In the HSI acquisition process, external environmental factors tend to be more complex and challenging to manage. In the practically significant non-blind deconvolution tasks, where the blur kernel is presumed to be known, the Poisson deconvolution problem has been a focus of study for several decades, beginning with the foundational Richardson-Lucy algorithm. Contemporary methods have evolved to include advanced techniques such as Plug-and-Play (PnP) frameworks and optimization strategies grounded in Maximum A Posteriori (MAP) estimation [37]. These modern approaches offer enhanced performance and robustness in dealing with the intricacies of deconvolution in HSI data.
There is a method for estimating spatially varying kernels from a single image, which aims to address challenges posed by non-uniform motion blur. Early attempts at estimating these kernels were primarily focused on scenarios involving camera self-motion, under the assumption that the scene was planar and could be mapped to the image plane via a homography. To tackle spatially varying blur resulting from depth differences and moving objects, a strategy was developed to segment the image into a limited number of layers at fixed depths. However, this piece-wise constant model has the drawback of being highly sensitive to how the blurred image is segmented. Another approach, more closely aligned with recent methods, addresses both depth variation within the scene and the presence of moving objects by incorporating local predictions of motion blur. This approach utilizes a set of predefined linear motion kernels parameterized by their length and direction. CNNs are employed to predict the probability distribution of these kernel parameters at the patch level. Subsequently, a Markov Random Field (MRF) is used to enforce motion smoothness, transforming patch-level distributions into a dense motion field [18]. For a given blurred image, a Fully Convolution Network (FCN) can directly estimate the dense linear motion field, parameterized by its horizontal and vertical components. To train the FCN, simulated motion fields are used to create synthetic pairs of blurred images and corresponding motion fields. Both of these methods advocate for using a non-blind deblurring approach with the estimated kernels to restore the image, combining an L 2 data fitting term with an Expected Patch Log Likelihood (EPLL) image prior [38]. These advanced techniques enhance the capability to effectively manage complex blurring scenarios in single images.

2.2. Effective Estimation of Blur Kernels Based on Deep Learning

In recent years, kernel prediction networks (KPNs) have demonstrated significant potential in various low-level visual tasks, such as burst denoising, optical flow estimation, frame interpolation, stereo matching, and video prediction [39]. KPNs have been particularly effective in the context of burst denoising, where multiple noisy images from the same scene are used to achieve a cleaner output. Cai et al. [40] utilized KPNs to generate denoised estimates for each pixel by computing a weighted average of locally noisy pixels across all input frames in a burst. The weights, or kernels, for this averaging process, are predicted directly from the noisy input bursts using CNNs. This approach enables adaptive and content-aware processing, ensuring the denoising operation is tailored to the local image structure. To further enhance computational efficiency, Xia et al. [41] proposed a basis prediction network. Instead of predicting unique kernels for each pixel, this method predicts a set of global basis kernels shared across the image. Alongside these global kernels, the network predicts pixel-specific mixing coefficients that determine how the global kernels are combined for each pixel in the input burst. This strategy significantly reduces computational complexity while maintaining high-quality denoising performance, making it well-suited for resource-constrained applications. These advancements in kernel prediction networks have paved the way for more efficient and effective solutions to complex denoising and other low-level vision tasks, particularly in scenarios where burst imaging or multi-frame data are available.
In this paper, a dense non-uniform motion-blurred estimation method based on a low-rank representation of the pixel-direction blurred kernel function is used. As in, the per-pixel kernel is modeled as a linear combination of basic elements, but instead of having a single pre-computed basis of elements, Blur–Kernel–Prior (BKP) based on KPN is used to infer a non-parametric basis specific to each input image. In addition, rather than learning kernels to solve the image recovery problem, we research them to fit forward-degenerate models, thus enabling us to utilize existing techniques to solve the inverse problem.

2.3. Attention Mechanisms for HSIs Reconstruction

Originally introduced in the field of natural language processing, the attention mechanism has demonstrated a remarkable capability to manage nonlocal similarities and has since achieved significant success in various computer vision tasks [3,6]. Considering the unique characteristics of HSIs, which possess richer spectral resolution and intricate spatial–spectral relationships compared to RGB images, most existing transformer-based methods incorporate dual constraints on both spatial and spectral information. In the context of HSI reconstruction, attention mechanisms have been employed to enhance the extraction and utilization of spatial–spectral features. Researchers have investigated the integration of transformers with 3D convolutions to facilitate the learning of these features. For instance, Liu et al. proposed a network named Interformer, which incorporates Transformer modules alongside 3D convolutions to capture complex spatial–spectral interactions [42]. Similarly, Hu et al. developed a multi-stage progressive network (MPNet) that utilizes learning and nonlocal channel attention to refine feature details [43].
Despite the innovative nature of these methods, they often encounter challenges due to the quadratic computational complexity intrinsic to attention mechanisms. This perplexity limits the efficiency and scalability of the models, as they predominantly emphasize long-range dependencies within the spectral dimension, potentially underutilizing both global and local spatial information. To address these challenges, a spatial–spectral attention (SSA) model has been proposed. The SSA model is designed to effectively extract and integrate both spatial and spectral information by leveraging local-global spatial–spectral relationships. This approach aims to enhance the performance of HSI reconstruction tasks by balancing computational efficiency with the comprehensive utilization of spatial–spectral features, thereby overcoming the limitations of existing methods.

3. Methodology

In this section, the problem formulation of HSIs denoising the research is first illustrated. Then, the proposed HISR method BKPSSA is described in detail, including four parts: general framework-based group strategy, BKP, SSA, and loss function.

3.1. Overall Architecture

The overall structure of the proposed BKPSSA frame for HIS denoising is illustrated in Figure 1. The BKPSSA HSIs denoising network proposed in this study is based on a group strategy, which will result in a model with fewer parameters, thereby facilitating a more efficient convergence during the training process than traditional CNNs and Attentions [31,44]. The feature extractor and reconstructor use innovative components that simultaneously capture both local and global spatial–spectral interactions. The first part, the BKP module, is designed to learn the blurred kernel and the spatial dimensional relationships of the HSIs. By leveraging U-Net’s symmetric architecture, the BKP module efficiently captures the blur kernel’s prior knowledge and helps restore the spatial consistency of the image. This structure enables the network to focus on the essential spatial–spectral features by preserving key spatial information while mitigating the effects of noise and blur. The second part, the SSA module, is responsible for learning the locally and globally diverse spatial–spectral interactions. The SSA architecture, which combines a hybrid 2D-3D convolutional network with an attention mechanism, can capture both short-range spatial correlations and long-range spectral dependencies. The integration of these two components within the BKPSSA framework not only improves the efficiency of the model but also contributes to its robustness in handling complex degradation issues, making it a powerful approach for HSI reconstruction on denoise tasks. It is essential to emphasize that, unlike conventional denoising models that focus on the removal of Gaussian noise and strip noise, as well as super-resolution tasks that aim to enhance sensor resolution limits, our task concentrates on addressing imaging blur noise caused by external factors, such as Gaussian blur and motion blur, during the acquisition process of hyperspectral camera.
Rather than being limited to conventional Gaussian noise models that consider only additive noise, the degradation of HSIs, whether in the form of blurring or resolution reduction, is often caused by convolution with a blur kernel.
It is agreed that the real HSIs are I H R R C × H × W , and the blurred HSIs are I B R C × H × W , the spatial and spectral resolution of the blurred HSIs are unchanged compared to the real HSIs, but the signal-to-noise ratio is lower. The relationship between the blurred HSIs I B and the real HSIs I H R is showing as Equation (1),
I B = K I H R + N
where K is the blurred kernel, the essence of image degradation is the convolution operation of the image with the blurred kernel, shown by the symbol ⊙. N indicates that the blurred image is affected by other linear noises, such as the pretzel noise, Gaussian noise, and so on. Therefore, the recovery reconstruction of blurred HSIs can be regarded as the inverse problem of the degradation process. To establish the model M R e b u i l d , the reconstructed HSIs I R minimizes the information difference in spatial and spectral dimensions between the reconstructed image I H R and the original image I H R . The acquisition of I R belongs to the following Equation (2),
I R = M R e b u i l d ( I B )
To fully utilize the strong correlation between adjacent bands and improve efficiency, many methods based on grouping strategies have been proposed and shown to be superior to other methods. On this basis, the input I B HSIs are first grouped into G groups according to the spectral channel mode, I B = [ I B 1 , , I B g , , I B G ] .
In this model, the blurred image I B first undergoes the BKP process. For the local information I B g , a U-net-based blurred kernel estimation is designed to extract the blurred kernel features K g R C G × H × W and the equivalent degraded hyperspectral sub-images I L R g R C G × s × s and the equivalent degenerate hyperspectral sub-image I L R g R C G × h × w to the following Equation (3),
K g , I L R g = M B K P ( I B g )
The dimension size of K g depends on the initially given hyperparameters. The spatial-temporal dimensional degradation blurred kernel estimate K for the whole HSIs can be finally obtained after inter-spectral cascading, with the following Equation (4):
K = c a t ( K 1 , K 2 , , K G ) )
As for the equivalent degraded hyperspectral sub-image I L R g , it already contains prior degradation information, so it needs to be reconstructed in the spatial–spectral dimension in the SSA structure. In order to finely control the reconstruction, as well as to ensure that the reconstruction accuracy is improved, the sub-images are first reconstructed within the corresponding channel of each sub-image to obtain the SSA reconstructed sub-images I R g R C G × H × w , a process known as the Shallow Feature Extraction Block (SFEB), including hybrid 2D-3D convolution which combines 2DCNN utilized at spatial dimension and 3DCNN utilized at spectral dimension. Then cascade each sub-map to obtain the cascade reconstruction image in spectral dimension I R R C × H × w to the following Equation (5),
I R = c a t ( I R 1 , I R 2 , , I R G )
Obviously, I R is only a simple cascade of channels, and thus, the error with the original image is still not small, and the spatial dimension is lower than the original image, so it is necessary to carry out reconstruction of the cascade reconstructed image I R , to establish a deep feature extraction module Deep Feature Extraction Block (DFEB) that unlike SFEB can extract spatial–spectral features from the full spectrum, restoring texture detail information in the spatial dimension and spectral characteristics in the spectral dimension, thereby correcting the global spectral loss.
I d e c o n v = d e c o n v I B , K I R = U p ( I d e c o n v + U p ( I R ) )
Finally, in order to reduce the difficulty of training the network and to focus on the high-frequency information, the blurred kernel is deconvoluted with the input blurred image by a filter function I d e c o n v is added directly to the tail of the network. The 1 × 1 convolution after the double-third time is used to adjust the spatial size of the addition operation. Moreover, 1 × 1 convolution after summation is performed to make its spatial size consistent with the original high-resolution HSIs. The above upsampling processes can be represented as Equation (6).

3.2. Blur Kernel Denoise Prior

HSIs denoising constitutes an ill-posed challenge that necessitates the incorporation of additional priors, or regularization, to appropriately constrain the reconstruction process, which traditional approaches strive to address by manually designing intricate regularization terms such as total variation (TV), sparsity, and low-rank constraints. Consequently, the effectiveness of these algorithms is largely contingent upon the capability of the designed priors to accurately represent the observed data. In the context of the HSIs super-resolution problem, it becomes imperative to efficiently leverage the intrinsic characteristics of HSIs, which include nonlocal self-similarity within the spatial domain and a high degree of correlation across spectral bands [45].
To address the limited sample size in hyperspectral imaging, the U-Net architecture employs a symmetric U-shaped structure comprising contracting and expanding paths (Figure 2). At the downsample stage, the contracting path progressively doubles the number of convolutional kernels at each level to expand feature channels, formulated as follows, where learnable kernels and b represent bias terms.
Global average pooling (GAP) follows each convolution to reduce feature map dimensions:
G A P ( I ) , w l = 1 H W i = 1 j = 1 I i , j ,
the output of the downsampling process is the weight w l of the convolutional layer, which represents the learning of the hyperspectral image-blur kernel.
The upsampling process is achieved through deconvolution. The expanding path utilizes transposed convolution to progressively restore spatial resolution while halving channel numbers. Skip connections fuse multi-level features through concatenation
| I c o m b i n e d = [ I d o w n ; I u p ] ,
integrating positional details from the contracting path with semantic abstractions from the expanding path.
The channel dimension follows exponential scaling: C ι = 2 l · C 0 (contracting) and C l = 2 L l · C 0 (expanding). Information compensation is quantified as
Δ I = l 1 α l · I s k i p l ,
with α l weighting skip-connection contributions. At the same time, after the intermediate layer ConvResnet (mixed residual convolution) further image-blur kernel extraction, through the layer-by-layer unsqueeze to achieve the size of the blur convolution kernel recovery, and the final output for the blur kernel after the BKPUnet network.
Figure 2. The Blur–Kernel–Prior Block-based U-Net.
Figure 2. The Blur–Kernel–Prior Block-based U-Net.
Remotesensing 17 01401 g002
Based on the above description, the BKP module is designed wherein a U-Net architecture that serves as the foundational network of the proposed methodology, as depicted in Figure 2, and utilizes 3 × 3 convolutions with average pooling strategy as downsampling blocks with a stride of 2, while employing transposed convolutions as upsampling blocks alongside multiple layers functioning as feature extraction blocks. Meanwhile, ResNet with convolution-2D is chosen as the middle layer, learning abstract and sophisticated features. These feature extraction blocks are predicated on pointwise and 3 × 3 depthwise convolutions with a stride of 1, complemented by simple ReLU functions that are utilized as activation mechanisms. The final output is subsequently compared to the downsampled patches of the input blurry hyperspectral sub-images in conjunction with the learned blurry feature kernels. If the input sub-image I i n g R C G × h × w and output sub-image I o u t g R C G × h × w are defined, and the downsampling process is represented as Conv AvgPool , the upsampling process as TransConv , and the intermediate layer as ConvResNet , the overall process can be formalized as Equation (10),
I d o w n s a m p l e g = Conv AvgPool ( I i n g ) P I m i d d l e g = TransConv ( I d o w n s a m p l e g ) Q I o u t g = ConvResNet ( I m i d d l e g ) R
while P, Q, R are layers of dowmsampler, middle, and upsampler.

3.3. Spatial–Spectral Attention Feature Rebuild

To remove noise from HSIs, it is important to explore the similarity information in the spatial domain, which implies that similar pixels can be aggregated together for denoising. Existing deep-learning-based HSI denoising methods mainly utilize the convolutional layer to extract the local information with spatially invariant kernels, limiting the flexibility to model the nonlocal similarity. The Spatial–Spectral Attention (SSA) is specifically designed as an adaptive channel feature extraction mechanism within the network illustrated by SSPSR [44], with its structure shown in Figure 3. This module begins with layer normalization (LN) and enters the spatial–spectral Attention Block (SSAB) to generate preliminary spectral feature maps.
HSIs is a three-dimensional cubic image whose rich and detailed spatial and spectral features can be effectively extracted using 2DCNN. When a 2D CNN is used directly, preprocessing of the original HSIs is required, which leads to a reduction in spectral dimensionality. On the other hand, 3D convolutions can directly take HSIs as the input to the network without the need for complex preprocessing, thereby allowing simultaneous extraction of both spatial and spectral features. However, 3D CNNs involve a larger number of parameters to be learned, which presents a limitation in terms of computational complexity. Given that there exists a significant amount of redundant information and noise in the HSIs spectral bands, simple 3D convolution alone does not yield optimal HSIC results. To address these issues, a hybrid 2D-3D convolution is proposed, which combines the 2D convolution, 3D convolution, and residual mechanism of the harvester, effectively mitigating the deficiencies in feature extraction and enhancing computational efficiency. The Shallow spatial–spectral Attention Block (SSSAB) applies attention mechanisms separately to the spatial and spectral dimensions, aggregating contextual information from each patch in the spatial dimension while adaptively learning key information from each channel in the spectral dimension, selectively emphasizing relevant spectral features while suppressing irrelevant ones.
Many studies have proposed spectral attention-based methods for HSIs restoration and fidelity enhancement by leveraging spectral correlations [30,31]. However, standard transformers typically aggregate similar features based on global query-key pairs, which can result in the interaction of unrelated information, potentially interfering with the restoration outcomes. To overcome this disadvantage, stimulated by NLP [46,47], Spectral Similarity Attention (SpeSA) and Spatial Similarity Attention (SpaSA) mechanism-based multi-head attention are a replacement for the traditional spectral attention method.
In detail, SpeSA and SpaSA address these challenges through co-optimized spectral reliability weighting and spatial geometry-aware regularization, enabling physics-informed disentanglement of degradation factors, which adaptively score the similarity of features, retaining the relevant components while discarding the irrelevant ones, thereby preventing the interference of redundant information during the information interaction process. Considering X R C × W · H is HSIs contractive feature at spectral dimension and X R C · W × C · H is HSIs contractive feature at spatial dimension, MSA processes can be represented as Equation (11),
SpeSA ( Q , K , V ) = SoftMax m = 1 n T K m Q K d V
where d is the feature dimension, and T m is the token studied by the feature. Then spectral features are learned by
F spe = R e L U ( W 2 δ ( W 1 SpeSA ) ) ,
where W 1 R C × C / r and W 2 R C × C / r form a bottleneck structure with reduction ratio r, δ denotes ReLU activation, and R e L U ( · ) represents the ReLU nonlinear function.
Meanwhile, the position-aware attention is computed through affinity matrices,
e i j = exp ϕ ( y ) T ψ ( y j ) j = 1 N exp ϕ ( y i ) T ψ ( y j ) ,
where ϕ ( · ) and ψ ( · ) are linear projections of the input features and denote spatial positions. The refined features are obtained through,
F spa = j = 1 N e i j θ ( y j ) ,
Finally, the spectral and spatial streams are combined through element-wise multiplication,
F output = F spe F spa + F input
where F output represents the extractor of deep spatial and spectral features of HSIs, adding input features F input as a residual structure for avoiding the vanishing gradient problem and enable the network to train more layers efficiently.
The Deep Spatial–Spectral Attention Block (DSSAB), shown in Figure 4, compared to SSSAB, chooses a strategy including additional residual layers (ResBlock) and Multi-Linear Layers (MLP) to extract detailed information more comprehensively, such as edges, textures, and other features. Meanwhile, SpaSA and SpeSA are provided for global information restoration. After extracting features from all channels, this module extensively learns the most informative spectral features. SSA can simultaneously focus on detailed features in different spatial positions and different spectral channels, thus achieving the purpose of SSA and enhancing the robustness of HSI restoration.

3.4. Overall Loss Function

A large body of work has demonstrated the positive impact of L 1 and L 2 losses in image restoration tasks. Since the L 2 loss usually leads to over-smoothed results, while the L 1 loss provides a more balanced error distribution and better convergence, the L 1 loss is employed to measure the accuracy of reconstructed images. L 1 loss calculates the pixel difference between the reconstructed I R HSIs and the original I H R HSIs, which can be formulated as Equation (16),
L 1 ( Θ ) = 1 N n = 1 N | | H h r n H s r n | | 1
where N is the number of images in the training batch, Θ denotes the parameter set of our proposed network, and H h r n and H s r n denote the nth reconstructed HSIs and the original HR HSIs, respectively. Although the L 1 loss can achieve good performance for natural images of SR, it ignores the spectral properties of HSIs and may lead to spectral distortion. SAM loss is introduced to maintain spectral consistency and spatial details. SAM loss can be formulated as Equation (17),
L s p e ( Θ ) = 1 N n = 1 N 1 π a r c c o s ( H h r n · H s r n p a r a l l e l H h r n 2 · H s r n 2 )
Furthermore, to measure the loss of informativeness of the blurred degraded HSIs, a peak signal-to-noise ratio loss PSNR [48] is introduced as Equation (18),
L P S N R ( Θ ) = 1 M N i = 1 M j = 1 N l o g 10 R M S E I i , j R , I i , j H R m a x ( I i , j R
where I i , j R , I i , j H R denote the spectral dimension information of the t ( i , j ) coordinate point in space, and RMSE(·) denotes the root-mean-square error of the spectral information of a point on the original HSIs and the reconstructed image.
The total loss for training the proposed network can be expressed as Equation (19),
L total ( Θ ) = L 1 + λ 1 L s p e + λ 2 L P S N R
where λ 1 and λ 2 denote trade-off parameters with different losses. Specifically, λ 1 controls the weighting of the spectral term to avoid excessively large values, which would lead to an excessive focus on spectral similarity at the expense of spatial detail. Similarly, λ 2 controls the weight of the PSNR term to avoid oversharpening in the spatial and spectral structure. The combination of the above losses can achieve superior reconstruction performance in both spatial and spectral dimensions. As a rule of thumb, λ 1 is set to 0.5 and λ 2 to 0.1 [31].

4. Results

To conduct a rigorous evaluation of the proposed method, several related approaches are selected for comparison, including the Singular Value Decomposition (SVD) [16], which epitomizes classic machine-learning techniques for deblur, alongside FPNSR [49], CEGATSR [50], which represent a classic CNN network, EUNet [45], MSDformer [31], and DSST [32], which are representative of novel deep-learning methodologies for denoise. In our model, as the application of group strategy, the number of bands processed by the spatial–spectral attention module was configured to be 32. Additionally, the batch size was designated as 32, and the patch size was set to 4. All methods are trained to utilize the PyTorch 2.1.0 framework, which was implemented on a graphics processing unit (GPU) employing an NVIDIA RTX 4090. Furthermore, the aforementioned methods underwent training for a total of 300 epochs. All deep-learning-based methods are trained and tested under the same conditions to ensure fairness.

4.1. Data Description

As illustrated in Figure 5, three publicly available HIS datasets, namely the CAVE, Chikusei, and Pavia University standard datasets, and an actual hyperspectral remotely sensed image dataset named the XiongAn dataset are used in the experiment as illustrated in Figure 6.
To systematically evaluate algorithm robustness, we established a hybrid degradation model combining parameterized blur types. The synthetic input generation process is formulated as
K e r n e l = μ 1 · G σ + μ 2 · M θ , Δ x ( μ 1 + μ 2 = 1 ) ,
In our simulation experiments, a publicly available dataset is utilized, from which sub-images with a spatial resolution of 256 × 256 pixels are extracted. To generate the input for the low spatial resolution blurred HSIs, images are applied of a motion blur with random image shifts ranging from 0 to 20 pixels and rotations from 0 to 0.5 radians, along with a Gaussian blur characterized by a mean of 0 and a standard deviation of 3. This resulted in blurred images with average signal-to-noise ratio (SNR) gradients of 20 dB, 30 dB, and 40 dB. For the training set, 1000 sub-images are selected, while 200 sub-images are chosen for the validation and test sets from each of the CAVE, Chikusei, and Pavia University datasets. In all experiments conducted in this study, several evaluation metrics are employed: spectral angular similarity (SAM), peak signal-to-noise ratio (PSNR), global normalized loss (ERGAS), root-mean-square error (RMSE), structural similarity index (ASSIM), and cross-correlation (CC). Additionally, the model size, including reference model size and model parameters, was considered to assess the ease of deployment and migration of the model. Furthermore, ablation experiments are provided on the XiongAn dataset to demonstrate the importance of each module in enhancing model accuracy.

4.1.1. Cave Dataset

This contains 32 HSIs with a size of 512 × 512 and 31 spectral bands. In addition, each HSI also has a corresponding RGB image of size 512 × 512 and three spectral bands. This simulated dataset somehow characterizes the blurred features of the feature HSIs as well as the HSIs with a medium number of bands [51].

4.1.2. Pavia University Dataset

The Pavia University dataset [52] was acquired by the Reflectance Optical System Imaging Spectrometer (ROSIS) over the city center of Pavia in northern Italy. The HSIs in this dataset cover the wavelength range from 430 nm to 860 nm and are divided into 102 bands after removing the noise bands. The spatial resolution of the dataset is 1.3 m per pixel, and the image size is 1096 × 1096 pixels.

4.1.3. Chikusei Dataset

The Chikusei dataset [53] captures a variety of urban and agricultural landscapes in the Chikusei region of Ibaraki Prefecture, Japan. The dataset covers the wavelength range from 363 nm to 1018 nm with 128 spectral bands. Each image has a high spatial resolution of 2048 × 2048 pixels. These images cover different scenes, including urban areas, rice fields, forests, and roads, making them suitable for various remote sensing applications.

4.1.4. XiongAn Dataset

Including HSIs data of Horseshoe Bay Village in Xiong’an New Area (https://www.ygxb.ac.cn/zh/article/doi/10.11834/jrs.20209065/, accessed on 23 March 2025), collected by the full-spectrum multi-modal imaging spectrometer of the aerial system developed by the Shanghai Institute of Physics and Technology of the Chinese Academy of Sciences, with a spectral range of 400–1000 nm, 250 bands, and an image size of 3750 × 1580 pixels, with a spatial resolution of 0.5 m. The simultaneous field research on the distribution of land types is also conducted on 19 species.

4.2. Evaluation Metrics

To quantitatively evaluate the performance of the proposed method and the existing methods, six commonly used HSI reconstruction metrics are chosen. Let I s o l and I r e s denote the denoised and super-resolution reconstructed HSIs and the reference HSIs, respectively.

4.2.1. Spectral Angle Mapping (SAM)

Spectral Angle Mapper (SAM) [54] is a technique used for HSI similarity on spectral dimension. In SAM, the spectrum of each pixel is considered to be a multidimensional vector (each band corresponds to one dimension), and then the angle between these vectors is calculated. Obviously, the closer the spectral angle between two spectra is to 0° indicates better similarity. The calculation formula is as Equation (21),
SAM I s o l , I r e f = a r c c o s I s o l · I r e f I s o l 2 I r e f 2

4.2.2. Root-Mean-Square Error (RMSE)

The Root-Mean-Square Error (RMSE) between spectral signals is a useful metric of the difference between two spectral signals. It is the square root of the mean of the sum of the squares of the differences between the observed and true values. When processing HSIs or analyzing spectral data, RMSE can be used to assess the quality of reconstructed spectra, compare the similarity between different spectra, or verify the accuracy of model predictions. The formula is as Equation (22),
RMSE I s o l , x r e f = I s o l , I r e f F n × l
where | | · | | F denotes the Frobenius paradigm. The optimal RMSE is 0.0.
In practice, the RMSE provides a quantitative metric for evaluating the differences between two spectral signals, but it does not provide information about the source of those differences. For example, a larger RMSE may be due to a difference in overall signal intensity or a small but significant difference at a specific wavelength. Therefore, RMSE is often used in conjunction with other analytical methods to gain a more complete understanding.

4.2.3. Peak Signal-to-Noise Ratio (PSNR)

The Peak Signal-to-Noise Ratio (PSNR) [48] of a spectral signal is a significant metric of the quality of HSI reconstruction, for example, to assess the degree of distortion between a compressed image and the original image. The PSNR is calculated by comparing the maximum possible power of the signal to the noise power of the signal and is usually expressed in decibels (dB) units. The calculation formula is as Equation (23),
PSNR I s o l , I r e f = 1 l i = 1 l 10 l o g 10 m a x I s o l i R M S E I s o l i , I r e f i
PSNR is a useful metric for evaluating the quality of image reconstruction, reflecting sensor-perceived differences. the higher the PSNR value, the better the quality.

4.2.4. Structural Similarity Index (SSIM)

Structural Similarity Index (SSIM) [48] for spectral signals is a metric of similarity between two images that takes into account three aspects of similarity: luminance, contrast, and structure. SSIM is widely used in the image processing field to assess the quality of an image. Although SSIM was originally designed for images, it can also be applied to spectral signals to assess the similarity between two HSIs on the spatial dimension. The calculation formula is as Equation (24),
SSIM I s o l , I r e f = 2 μ I s o l μ I r e f + C 1 2 σ I s o l I r e f + C 2 μ I s o l 2 + μ I r e f 2 + C 1 σ I s o l 2 + σ I r e f 2 + C 2
where σ I s o l I r e f denotes the covariance between I s o l and I r e f . The optimal value of SSIM is 1.

4.2.5. Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS)

ERGAS (Erreur Relative Globale Adimensionnelle de Synthèse [55], Integrated Relative Global Error) is a metric commonly used to assess the quality of hyperspectral data or image fusion. ERGAS is commonly used in remote sensing, especially in HSI processing, to evaluate the performance of image fusion, compression, or other processing techniques. The calculation formula is as Equation (25),
ERGAS I s o l , I r e f = 100 1 γ 2 1 l i = 1 l RMSE I s o l i , I r e f i μ i
where RMSE i denotes the RMSE between I s o l and I r e s in the i-th spectral band, and γ is the scaling factor. The optimal ERGAS is 0.0.

4.2.6. Cross-Correlation (CC)

In hyperspectral data analysis, cross-correlation [56] is a metric of the similarity of two signals at different time displacements. For HSIs, in order to distinguish between spatial dimension and spectral dimension loss, this usually means comparing the image similarity in the same band and comparing the similarity between two bands; this is also known as Spatial Cross-correlation and Spectral Cross-correlation. The formula is as per Equation (26),
CC s p a ( I s o l , I r e f ) = 1 l i = 1 l j = 1 n I s o l j μ I s o l I r e f j μ I r e f j = 1 n I s o l j μ I s o l 2 I r e f j μ I r e f 2 CC s p e l s o l , l r e f = 1 m n i = 1 m j = 1 n j = 1 n l s o l i j μ l s o l l r e f i j μ l r e f j = 1 n l s o l i j μ l s o l 2 l r e f i j μ l r e f 2

4.2.7. Model Scale

Model size typically refers to the storage space requirements of the model, usually expressed in units such as bytes (Bytes), kilobytes (KB), or megabytes (MB). It primarily depends on the number of parameters in the model and the storage format of these parameters (e.g., floating-point numbers (FLOPs), integers, etc.). The model size directly affects storage requirements and computational resources, especially when deploying to edge devices or mobile devices.
The number of parameters, also known as computational complexity, characterizes the number of layers in the model and the number of neurons in each layer, which directly influence the parameter count. Generally, more parameters indicate stronger expressive capability of the model but may lead to increased demands for computational resources and energy consumption, which can, in turn, result in slower training and inference speeds as well as higher storage requirements. These challenges make it difficult to deploy such models on resource-constrained devices. Therefore, when designing a model, it is essential to strike a balance between the model’s accuracy and its reasonable scale, ensuring that the model remains efficient and practical for deployment without compromising its performance.

4.3. Comparison with State-of-the-Art HSIs Reconstruction Method

4.3.1. Experiments on the Cave Dataset

Table 1 presents the quantitative results of our proposed method alongside other comparative methods on the Chikusei dataset at different blurred noise signal-to-noise ratios (SNRs), with the best results highlighted in bold. The CAVE dataset, characterized by its rich spatial information, falls under the category of close-range hyperspectral imaging. Our model consistently outperforms other methods at blurred noise SNRs of 40 dB, 30 dB, and 20 dB across all six evaluation metrics. This indicates the robustness and effectiveness of the proposed method in handling various noise levels in HSI reconstruction. Figure 7 displays the qualitative results of HSIs reconstruction on the CAVE dataset at a 30 dB blurred noise SNR. Specifically, channels 31, 14, and 4, selected as the R-G-B channels, are shown to enhance visualization. While the Singular Value Decomposition (SVD) method can perform the reconstruction task, its performance is limited in terms of the recovery of fine spatial and spectral details. Although methods like EuNet and MSDformer demonstrate significant improvements in reconstruction quality, as observed in the red-boxed image, the BKPSSA model achieves superior pixel-level recovery results. The visual comparison highlights the effectiveness of BKPSSA in preserving both spatial and spectral information, resulting in more accurate and sharper reconstructed images. This further emphasizes the advantage of the proposed BKPSSA approach, which not only provides superior quantitative performance but also ensures better visual fidelity in the reconstructed HSIs, even under different levels of noise interference. The combination of advanced spatial–spectral feature extraction and noise suppression mechanisms allows BKPSSA to outperform existing methods in both objective metrics and subjective visual evaluation.

4.3.2. Experiments on the Pavia University Dataset

Table 2 presents the quantitative results of all methods on the Pavia University dataset across six evaluation metrics, expressed in terms of blurred noise signal-to-noise ratios (SNRs) of 40 dB, 30 dB, and 20 dB. Due to the limited number of training samples available in the Pavia University dataset, each method tends to perform poorly on this dataset in comparison to others, as smaller training sets often lead to overfitting or insufficient generalization. Nevertheless, the proposed BKPSSA method consistently achieves better results across all six evaluation metrics and at every level of blurred noise SNR compared to other methods. This demonstrates the robustness and efficiency of BKPSSA in handling noise and preserving image quality under different noise conditions. The experimental results further highlight that the BKPSSA method is particularly well-suited for scenarios involving small sample training datasets. Despite the challenges posed by the limited amount of training data, BKPSSA can effectively learn from the available samples by leveraging both local and global spatial–spectral information, along with prior noise knowledge. This combination allows the model to recover fine details in the HSIs while suppressing noise, even when the training data are sparse. By incorporating these advanced techniques, the proposed BKPSSA method provides strong robustness in both spatial and spectral dimensions, enabling high-quality HSI denoising and reconstruction. This capability is crucial for real-world applications where data may be limited, yet the need for accurate and reliable image reconstruction remains paramount. The ability of BKPSSA to effectively exploit available information, even with small datasets, underscores its potential for a wide range of HSI processing tasks. And Figure 8 shows the error of reconstructed images from different methods compared with the Ground Truth.

4.3.3. Experiments on the Chikusei Dataset

Table 3 presents the quantitative results of our method compared to other approaches on the Chikusei dataset under different levels of blurry SNR. The best results are highlighted in bold, while the second-best results are underlined. It is noteworthy that the SVD, a non-deep-learning method, demonstrates average denoising and reconstruction performance and struggles with nonlinear blurry noise. In contrast, deep-learning methods show significant improvements.
The FPNSR and CEGATSR methods utilize grouping strategies to extract spectral information, achieving better super-resolution (SR) results than SVD. The MSDformer builds upon the grouping strategy by incorporating a Transformer to capture long-range dependencies in spectral information; however, it lacks attention to local spatial details. Notably, our proposed BKPSSA captures global spatial–spectral dependencies, demonstrating superior spatial–spectral information extraction and fusion capabilities, outperforming other methods across all metrics, regardless of the scaling factor at 40 dB, 30 dB, or 20 dB.
Figure 9 provides visualizations of the HSIs reconstructions from the Chikusei test set at a 30 dB blurry noise SNR. Channels 20, 40, and 60 are selected for RGB visualization to enhance visual interpretation. It is evident that while the SVD method performs SR on the HSIs, it results in the blurriest edges and poorest smooth region reconstructions. Eunet and MSDformer show improved reconstructions, especially regarding line details, but still have deficiencies in detail recovery. In contrast, our BKPSSA achieves superior edge and detail restoration.

4.3.4. Comprehensive Evaluation in the XiongAn Database

The XiongAn dataset was selected as the real data to test the network’s robustness. Table 4 shows the model sizes and evaluation metrics within the XiongAn dataset. Specifically, our proposed BKPSSA achieved the highest scores across all metrics, convincingly demonstrating its excellent ability to restore spatial details and preserve spectral features. MSDFormer and DSST also exhibited good spectral reconstruction capabilities but still lagged behind DL-based methods in the spatial domain. The matrix-based method, SVD, suffered from severe distortion, leading to a loss of spatial details and spectral features.
On the other hand, in terms of model size, our proposed BKPSSA has a moderate number of FLOPs and achieves better reconstruction performance with a lower number of parameters. Figure 10 shows the visual results from the XiongAn dataset. We can see that methods such as CEGATSR, EuNet, and MSDFormer exhibit noticeable distortions, highlighting their limited capability in the reconstruction domain. BKPSSA, on the other hand, produces more satisfactory results with fewer artifacts.
In summary, after comparing different deblurring algorithms applied to hyperspectral images, most of the CNN and Attention-based methods of reconstruction on deblur rely on the assumption that the spectral and spatial degradation lie on the resolution. As a result, shown in Figure 11, the spectral details are smoothed out based on traditional methods, and even if the signal-to-noise ratio is improved, it cannot be directly used to remove the blurred noise.

5. Discussion

In this section, the effect of different modules is reported on the robustness performance of different models on more subdivided scale blur noise and the effectiveness of the different modules of the proposed BKPSSA.

5.1. Comparison of Robustness Performance of Different Models at a More Subdivided Noise Scale

Based on the above model comparison, the blur noise is set from 20 to 40 dB to a more detailed scale, and a single dataset is trained and tested at 5 dB intervals. The results are shown in Figure 12, and the evaluation indicators are PSNR and SAM. It can be seen that on the Chikusei and CAVE datasets, the proposed BKPSSA has better robustness, and when the noise level increases, the response is stable, and there is no mutation compared with other methods.

5.2. Effectiveness of the Grouping Strategy

At the beginning of the network, the spectral bands are divided into subgroups and efficiently extract the local spatial–spectral information. Since the grouping strategy was studied in depth and relevant effects of the different settings are analyzed in the experimental results of the research, the grouping processes have not been studied here but demonstrate the effectiveness of the grouping strategy. Models without grouping strategies are donated as w/o Groups and train the network in a band-by-band function. As a result, the complexity of the model increases as the number of spectral bands in the image increases, which reduces the efficiency of the network.

5.3. Effectiveness of the BKP Block

In the proposed a priori denoising module, BKP, a U-net-based neural network is employed to extract global noise features and decouple the input hyperspectral sub-image from the blurred kernel, thereby enhancing the model’s representation capability. To evaluate the effectiveness of the BKP module, a network composed entirely of CNN modules is designed. Specifically, the U-net network is replaced with the AE (Autoencode-decode) module [57], which is commonly used in deep learning. The values of all metrics deteriorate significantly, demonstrating the importance of the designed BKP. The Autoencode-decode network is denoted as “w/o AE” in Table 5. While the AE module slightly preserves spectral similarity, the spatial similarity is severely degraded, which further justifies the significance of the proposed BKP.

5.4. Effectiveness of Spatial and Spectral Attention

In BKPSSA, Spatial and Spectral Attention (SSA) is designed to learn features from the spatial and spectral dimensions, respectively. To validate the effectiveness of the proposed SSA structure, various variants were designed using either spatial attention or spectral attention alone. As shown in Table 4, the performance of the model is significantly degraded when using “w/o Spatial Attention Only” (w/o SA1) or “Spectral Attention Only” (w/o SA2). The likely reason for this is that each module specializes in either spatial or spectral information, and their parallel operation enables the integration of information, facilitating the comprehensive learning of spatial and spectral features in the HSIs. When operating independently, significant shortcomings in the efficient extraction of spatial or spectral information arise, which fail to address the long-range dependencies of spatial information or the redundancy present in spectral information.

6. Conclusions

In this study, we propose a novel HSI deblurring reconstruction network, referred to as BKPSSA, to address the limitations of existing transformation-based methods in mitigating image-blur interference during HSI reconstruction. Specifically, BKPSSA is founded on a group strategy learning framework, wherein the BKP module is employed to establish prior knowledge of the blur kernel within the learned subspace, thereby restoring spatial consistency through the estimation of the blur kernel. The SSA module incorporates shallow spatial–spectral feature extraction based on a hybrid 2D-3D convolution mechanism alongside a spatial–spectral residual attention network based on multi-head attention, which effectively extracts global information. Extensive experiments on both natural and remote sensing images demonstrate that the proposed method outperforms other approaches in terms of visual perception and quantitative metrics. Additionally, ablation studies validate the effectiveness of each proposed module.
In future research, the proposed method will be further improved in aspect, which is to obtain prior noise information from the acquisition perspective and compare it with the network. Additionally, efforts will be made to reduce the parameter burden while ensuring high-quality reconstruction.

Author Contributions

Conceptualization, H.X. and M.Z.; methodology, H.X.; software, H.X. and H.H.; validation, Q.J. and W.Z.; investigation, H.X. and M.Y.; writing—original draft preparation, H.X.; writing—review and editing, M.Y. and X.T.; visualization, H.X.; supervision L.X.; funding acquisition, X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Changchun Science and Technology Development Plan Project (22S H03); Capital construction funds in Jilin Province in 2022 (2022C026); Jilin Province Science and Technology Development Plan Project (20240402029GH); Jilin Province Science and Technology Development Plan Projects (20240601051RC; 20220204079YY; 20230204095YY).

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tejasree, G.; Agilandeeswari, L. An extensive review of hyperspectral image classification and prediction: Techniques and challenges. Multimed. Tools Appl. 2024, 83, 80941–81038. [Google Scholar] [CrossRef]
  2. Nie, X.; Xue, Z.; Lin, C.; Zhang, L.; Su, H. Structure-prior-constrained low-rank and sparse representation with discriminative incremental dictionary for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5506319. [Google Scholar] [CrossRef]
  3. Shu, Z.; Wang, Y.; Yu, Z. Dual attention transformer network for hyperspectral image classification. Eng. Appl. Artif. Intell. 2024, 127, 107351. [Google Scholar] [CrossRef]
  4. Namburu, H.; Munipalli, V.N.; Vanga, M.; Pasam, M.; Sikhakolli, S.; Chinnadurai, S. Cholangiocarcinoma Classification using MedisawHSI: A Breakthrough in Medical Imaging. In Proceedings of the 2024 Second International Conference on Emerging Trends in Information Technology and Engineering (ICETITE), Vellore, India, 22–23 February 2024; pp. 1–6. [Google Scholar] [CrossRef]
  5. Li, S.; Song, Q.; Liu, Y.; Zeng, T.; Liu, S.; Jie, D.; Wei, X. Hyperspectral imaging-based detection of soluble solids content of loquat from a small sample. Postharvest Biol. Technol. 2023, 204, 112454. [Google Scholar] [CrossRef]
  6. Gong, M.; Jiang, F.; Qin, A.K.; Liu, T.; Zhan, T.; Lu, D.; Zheng, H.; Zhang, M. A spectral and spatial attention network for change detection in hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5521614. [Google Scholar] [CrossRef]
  7. Petropoulos, G.P.; Kalivas, D.P.; Georgopoulou, I.A.; Srivastava, P.K. Urban vegetation cover extraction from hyperspectral imagery and geographic information system spatial analysis techniques: Case of Athens, Greece. J. Appl. Remote Sens. 2015, 9, 096088. [Google Scholar] [CrossRef]
  8. Zia, A.; Zhou, J.; Gao, Y. Exploring chromatic aberration and defocus blur for relative depth estimation from monocular hyperspectral image. IEEE Trans. Image Process. 2021, 30, 4357–4370. [Google Scholar] [CrossRef]
  9. Geladi, P.; Burger, J.; Lestander, T. Hyperspectral imaging: Calibration problems and solutions. Chemom. Intell. Lab. Syst. 2004, 72, 209–217. [Google Scholar] [CrossRef]
  10. Gao, B.C.; Montes, M.J.; Davis, C.O.; Goetz, A.F. Atmospheric correction algorithms for hyperspectral remote sensing data of land and ocean. Remote Sens. Environ. 2009, 113, S17–S24. [Google Scholar] [CrossRef]
  11. Jia, J.; Zheng, X.; Guo, S.; Wang, Y.; Chen, J. Removing stripe noise based on improved statistics for hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 5501405. [Google Scholar] [CrossRef]
  12. Rasti, B.; Scheunders, P.; Ghamisi, P.; Licciardi, G.; Chanussot, J. Noise reduction in hyperspectral imagery: Overview and application. Remote Sens. 2018, 10, 482. [Google Scholar] [CrossRef]
  13. He, H.; Cao, M.; Gao, Y. Noise learning of instruments for high-contrast, high-resolution and fast hyperspectral microscopy and nanoscopy. Nat. Commun. 2024, 15, 754. [Google Scholar] [CrossRef]
  14. Jiang, T.X.; Zhuang, L.; Huang, T.Z. Adaptive hyperspectral mixed noise removal. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
  15. Peng, J.; Sun, W.; Li, H.C.; Li, W.; Meng, X.; Ge, C.; Du, Q. Low-rank and sparse representation for hyperspectral image processing: A review. IEEE Geosci. Remote Sens. Mag. 2021, 10, 10–43. [Google Scholar] [CrossRef]
  16. Guo, Q.; Zhang, C.; Zhang, Y.; Liu, H. An efficient SVD-based method for image denoising. IEEE Trans. Circuits Syst. Video Technol. 2015, 26, 868–880. [Google Scholar] [CrossRef]
  17. Rasti, B.; Ghamisi, P.; Benediktsson, J.A. Hyperspectral mixed Gaussian and sparse noise reduction. IEEE Geosci. Remote Sens. Lett. 2019, 17, 474–478. [Google Scholar] [CrossRef]
  18. Li, W.; Prasad, S.; Fowler, J.E. Hyperspectral image classification using Gaussian mixture models and Markov random fields. IEEE Geosci. Remote Sens. Lett. 2013, 11, 153–157. [Google Scholar] [CrossRef]
  19. Wang, X.; Hu, Q.; Cheng, Y.; Ma, J. Hyperspectral image super-resolution meets deep learning: A survey and perspective. IEEE/CAA J. Autom. Sin. 2023, 10, 1668–1691. [Google Scholar] [CrossRef]
  20. Han, X.H.; Shi, B.; Zheng, Y. SSF-CNN: Spatial and spectral fusion with CNN for hyperspectral image super-resolution. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2506–2510. [Google Scholar] [CrossRef]
  21. Wang, L.; Bi, T.; Shi, Y. A frequency-separated 3D-CNN for hyperspectral image super-resolution. IEEE Access 2020, 8, 86367–86379. [Google Scholar] [CrossRef]
  22. Arun, P.V.; Buddhiraju, K.M.; Porwal, A.; Chanussot, J. CNN-based super-resolution of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6106–6121. [Google Scholar] [CrossRef]
  23. Dixit, A.; Gupta, A.K.; Gupta, P.; Srivastava, S.; Garg, A. UNFOLD: 3D U-Net, 3D CNN and 3D transformer based hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5529710. [Google Scholar] [CrossRef]
  24. Wang, Z.; Ng, M.K.; Zhuang, L.; Gao, L.; Zhang, B. Nonlocal self-similarity-based hyperspectral remote sensing image denoising with 3-D convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531617. [Google Scholar] [CrossRef]
  25. Hou, R.; Li, F. Hyperspectral image denoising via cooperated self-supervised CNN transform and nonconvex regularization. Neurocomputing 2024, 616, 128912. [Google Scholar] [CrossRef]
  26. Shi, H.; Cao, G.; Zhang, Y.; Ge, Z.; Liu, Y.; Fu, P. H2A2 Net: A hybrid convolution and hybrid resolution network with double attention for hyperspectral image classification. Remote Sens. 2022, 14, 4235. [Google Scholar] [CrossRef]
  27. Galassi, A.; Lippi, M.; Torroni, P. Attention in natural language processing. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4291–4308. [Google Scholar] [CrossRef]
  28. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; Wang, Y. Transformer in transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  29. Zhang, D.; Zhou, F. Self-supervised image denoising for real-world images with context-aware transformer. IEEE Access 2023, 11, 14340–14349. [Google Scholar] [CrossRef]
  30. Zhao, Y.; Zhai, D.; Jiang, J.; Liu, X. ADRN: Attention-based deep residual network for hyperspectral image denoising. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 2668–2672. [Google Scholar] [CrossRef]
  31. Chen, S.; Zhang, L.; Zhang, L. Msdformer: Multi-scale deformable transformer for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5525614. [Google Scholar] [CrossRef]
  32. Yang, J.; Lin, T.; Liu, F.; Xiao, L. Learning degradation-aware deep prior for hyperspectral image reconstruction. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5531515. [Google Scholar] [CrossRef]
  33. Zhuang, L.; Ng, M.K.; Gao, L.; Michalski, J.; Wang, Z. Eigenimage2Eigenimage (E2E): A self-supervised deep learning network for hyperspectral image denoising. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 16262–16276. [Google Scholar] [CrossRef]
  34. Geng, P.; Zhang, M.; Li, X. Hyperspectral image deblurring based on joint utilization of spatial-spectral information. In Proceedings of the 2023 16th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 16–17 December 2023; pp. 156–160. [Google Scholar] [CrossRef]
  35. Fan, H.; Li, C.; Guo, Y.; Kuang, G.; Ma, J. Spatial–Spectral Total Variation Regularized Low-Rank Tensor Decomposition for Hyperspectral Image Denoising. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6196–6213. [Google Scholar] [CrossRef]
  36. He, W.; Zhang, H.; Zhang, L.; Shen, H. Total-variation-regularized low-rank matrix factorization for hyperspectral image restoration. IEEE Trans. Geosci. Remote Sens. 2015, 54, 178–188. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Zheng, Y.; Yuan, Q.; Song, M.; Yu, H.; Xiao, Y. Hyperspectral image denoising: From model-driven, data-driven, to model-data-driven. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 13143–13163. [Google Scholar] [CrossRef] [PubMed]
  38. Zhou, X.; Zhang, X.; Guo, P. Expected patch log likelihood with a prior of mixture of matrix normal distributions for image denoising. In Proceedings of the 2018 Ninth International Conference on Intelligent Control and Information Processing (ICICIP), Wanzhou, China, 9–11 November 2018; pp. 344–348. [Google Scholar] [CrossRef]
  39. Mildenhall, B.; Barron, J.T.; Chen, J.; Sharlet, D.; Ng, R.; Carroll, R. Burst denoising with kernel prediction networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2502–2510. [Google Scholar] [CrossRef]
  40. Cai, J.; Zuo, W.; Zhang, L. Dark and bright channel prior embedded network for dynamic scene deblurring. IEEE Trans. Image Process. 2020, 29, 6885–6897. [Google Scholar] [CrossRef]
  41. Xia, Z.; Perazzi, F.; Gharbi, M.; Sunkavalli, K.; Chakrabarti, A. Basis Prediction Networks for Effective Burst Denoising with Large Kernels. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11841–11850. [Google Scholar] [CrossRef]
  42. Liu, Y.; Hu, J.; Kang, X.; Luo, J.; Fan, S. Interactformer: Interactive transformer and CNN for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531715. [Google Scholar] [CrossRef]
  43. Hu, J.; Liu, Y.; Kang, X.; Fan, S. Multilevel progressive network with nonlocal channel attention for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5543714. [Google Scholar] [CrossRef]
  44. Jiang, J.; Sun, H.; Liu, X.; Ma, J. Learning spatial-spectral prior for super-resolution of hyperspectral imagery. IEEE Trans. Comput. Imaging 2020, 6, 1082–1096. [Google Scholar] [CrossRef]
  45. Liu, D.; Li, J.; Yuan, Q.; Zheng, L.; He, J.; Zhao, S.; Xiao, Y. An efficient unfolding network with disentangled spatial-spectral representation for hyperspectral image super-resolution. Inf. Fusion 2023, 94, 92–111. [Google Scholar] [CrossRef]
  46. Cordonnier, J.B.; Loukas, A.; Jaggi, M. Multi-head attention: Collaborate instead of concatenate. arXiv 2020, arXiv:2006.16362. [Google Scholar] [CrossRef]
  47. Hu, M.; Wu, C.; Zhang, L. GlobalMind: Global multi-head interactive self-attention network for hyperspectral change detection. ISPRS J. Photogramm. Remote Sens. 2024, 211, 465–483. [Google Scholar] [CrossRef]
  48. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  49. Sun, H.; Zhong, Z.; Zhai, D.; Liu, X.; Jiang, J. Hyperspectral image super-resolution using multi-scale feature pyramid network. In Proceedings of the Digital TV and Wireless Multimedia Communication: 16th International Forum, IFTC 2019, Shanghai, China, 19–20 September 2019; Revised Selected Papers 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 49–61. [Google Scholar] [CrossRef]
  50. Liu, C.; Dong, Y. CNN-Enhanced graph attention network for hyperspectral image super-resolution using non-local self-similarity. Int. J. Remote Sens. 2022, 43, 4810–4835. [Google Scholar] [CrossRef]
  51. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef] [PubMed]
  52. Huang, X.; Zhang, L. A comparative study of spatial approaches for urban mapping using hyperspectral ROSIS images over Pavia City, northern Italy. Int. J. Remote Sens. 2009, 30, 3205–3221. [Google Scholar] [CrossRef]
  53. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; SAL-2016-5-27; The University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  54. Sohn, Y.; Rebello, N.S. Supervised and unsupervised spectral angle classifiers. Photogramm. Eng. Remote Sens. 2002, 68, 1271–1282. [Google Scholar] [CrossRef]
  55. Renza, D.; Martinez, E.; Arquero, A. A new approach to change detection in multispectral images by means of ERGAS index. IEEE Geosci. Remote Sens. Lett. 2012, 10, 76–80. [Google Scholar] [CrossRef]
  56. Lei, J.; Liu, P.; Xie, W.; Gao, L.; Li, Y.; Du, Q. Spatial–spectral cross-correlation embedded dual-transfer network for object tracking using hyperspectral videos. Remote Sens. 2022, 14, 3512. [Google Scholar] [CrossRef]
  57. Sun, H.; Wang, L.; Zhang, L.; Gao, L. Hyperbolic space-based autoencoder for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5522115. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of Blur–Kernel–Prior and Spatial–Spectral Attention.
Figure 1. The overall architecture of Blur–Kernel–Prior and Spatial–Spectral Attention.
Remotesensing 17 01401 g001
Figure 3. The Shallow Feature Extract: Hybrid 2D-3D Convolution.
Figure 3. The Shallow Feature Extract: Hybrid 2D-3D Convolution.
Remotesensing 17 01401 g003
Figure 4. The Deep Feature Extraction: Spatial–Spectral Attention Block.
Figure 4. The Deep Feature Extraction: Spatial–Spectral Attention Block.
Remotesensing 17 01401 g004
Figure 5. Train Dataset: (a) cave dataset. (b) Pavia University dataset (Local). (c) Chikusei dataset (Local).
Figure 5. Train Dataset: (a) cave dataset. (b) Pavia University dataset (Local). (c) Chikusei dataset (Local).
Remotesensing 17 01401 g005
Figure 6. Test Dataset: XiongAn dataset.
Figure 6. Test Dataset: XiongAn dataset.
Remotesensing 17 01401 g006
Figure 7. Reconstructed test HSIs in the CAVE dataset with spectral bands 31-14-4 as R-G-B at blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Figure 7. Reconstructed test HSIs in the CAVE dataset with spectral bands 31-14-4 as R-G-B at blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Remotesensing 17 01401 g007
Figure 8. Reconstructed test HSIs in the Pavia University dataset with spectral bands 59-41-12 as R-G-B at a blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Figure 8. Reconstructed test HSIs in the Pavia University dataset with spectral bands 59-41-12 as R-G-B at a blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Remotesensing 17 01401 g008
Figure 9. Reconstructed test HSIs in the Chikusei dataset with spectral bands 60-40-20 as R-G-B at a blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Figure 9. Reconstructed test HSIs in the Chikusei dataset with spectral bands 60-40-20 as R-G-B at a blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Remotesensing 17 01401 g009
Figure 10. Reconstructed test HSIs in the XiongAn dataset with spectral bands 120-72-36 as R-G-B at blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Figure 10. Reconstructed test HSIs in the XiongAn dataset with spectral bands 120-72-36 as R-G-B at blur noise scale of 30 dB. From left to right, ground truth, then results of ground truth, SVD, CEGATSR, EuNet, MSDformer, and the proposed BKPSSA method.
Remotesensing 17 01401 g010
Figure 11. Reconstructed test HSIs in the XiongAn dataset normalization spectral difference with a blur noise scale of 30 dB.
Figure 11. Reconstructed test HSIs in the XiongAn dataset normalization spectral difference with a blur noise scale of 30 dB.
Remotesensing 17 01401 g011
Figure 12. Comparison of robustness performance of different models at a more subdivided noise scale.
Figure 12. Comparison of robustness performance of different models at a more subdivided noise scale.
Remotesensing 17 01401 g012
Table 1. Quantitative evaluation of different blur HSI reconstruction methods on the Cave dataset. The best and second-best results are bolded and underlined, respectively.
Table 1. Quantitative evaluation of different blur HSI reconstruction methods on the Cave dataset. The best and second-best results are bolded and underlined, respectively.
MethodNoise SNRPSNR↑SSIM↑SAM↓CC↑RMSE↓ERGAS↓
SVD 41.03350.91263.89440.90250.03085.3216
FPNSR 42.65210.92073.65670.92120.02383.1894
CEGATSR 45.60180.92653.15170.93040.02012.5681
EUNet40 dB47.02160.94802.89120.94110.01561.9260
MSDformer 48.05000.96102.23650.96980.01021.5724
DSST 48.35700.96602.09510.96610.00981.5654
Ours 49.53790.97012.00150.97640.00711.3256
SVD 32.56100.90273.86910.87120.05754.2561
FPNSR 35.41070.91003.35680.89980.04523.9430
CEGATSR 35.70750.92713.01760.90080.03993.6508
EUNet30 dB36.54170.93652.93150.91060.03683.2654
MSDformer 40.21250.94592.77760.91700.03042.9969
DSST 39.89310.94413.00140.92460.03472.8469
Ours 41.00150.95032.89580.93210.02112.3673
SVD 31.26590.81645.09180.85650.08997.7517
FPNSR 33.00230.88014.20190.87970.07656.6162
CEGATSR 33.36540.88344.00250.88650.07155.5681
EUNet20 dB34.26110.90683.98970.89440.05605.0007
MSDformer 35.56100.92273.66130.90170.04544.2526
DSST 36.00210.92963.32180.90650.04884.0524
Ours 37.01470.94033.19940.91690.04503.9367
Table 2. Quantitative evaluation of different blur HSI reconstruction methods on the Pavia University dataset. The best and second-best results are bolded and underlined, respectively.
Table 2. Quantitative evaluation of different blur HSI reconstruction methods on the Pavia University dataset. The best and second-best results are bolded and underlined, respectively.
MethodNoise SNRPSNR↑SSIM↑SAM↓CC↑RMSE↓ERGAS↓
SVD 32.69010.92104.54640.92010.05093.9651
FPNSR 33.61580.93774.01070.94470.04523.2654
CEGATSR 34.45040.94053.97900.95640.04013.1145
EUNet40 dB35.11230.94993.79840.96500.03472.9897
MSDformer 35.81850.95113.61390.97440.02662.7710
DSST 35.60890.95973.56120.97810.02642.7798
Ours 36.00610.96023.57210.97900.02542.6066
SVD 24.65110.76775.23150.82100.05527.0056
FPNSR 25.41010.77584.98900.85260.05086.8709
CEGATSR 26.51400.79124.92080.86700.04326.6511
EUNet30 dB26.59640.78294.90850.87940.04016.1625
MSDformer 28.78940.80204.77970.89070.03795.9773
DSST 28.89750.80894.51200.89910.03975.7171
Ours 29.00170.82074.17110.90240.03355.4107
SVD 20.68410.66157.79780.72000.07799.9841
FPNSR 21.05640.68797.50090.76020.07059.6759
CEGATSR 21.65480.69777.21540.78770.06749.1555
EUNet20 dB21.89040.69547.02560.78890.06469.1606
MSDformer 22.64240.70696.67060.80010.06088.3564
DSST 23.12120.71026.51320.81150.05988.2356
Ours 23.56170.72566.07990.82100.05018.1979
Table 3. Quantitative evaluation of different blur HSI reconstruction methods on the Chikusei dataset. The best and second-best results are bolded and underlined, respectively.
Table 3. Quantitative evaluation of different blur HSI reconstruction methods on the Chikusei dataset. The best and second-best results are bolded and underlined, respectively.
MethodNoise SNRPSNR↑SSIM↑SAM↓CC↑RMSE↓ERGAS↓
SVD 40.21250.90211.88800.93110.03996.5613
FPNSR 41.96900.91251.65100.95450.02135.6556
CEGATSR 41.39150.93601.26650.96010.01174.9001
EUNet40 dB42.807000.92901.61050.97260.02084.6324
MSDformer 47.28240.97221.14980.98000.01292.4315
DSST 47.56580.98011.20560.98110.01022.3001
Ours 48.56010.98211.12570.98520.00972.1871
SVD 30.25010.88712.92510.82100.045410.0235
FPNSR 32.00240.90152.88640.85990.04219.5001
CEGATSR 33.38710.92952.26650.84790.03978.9802
EUNet30 dB32.92070.90342.40980.83250.03159.1060
MSDformer 35.09110.93102.29140.85340.02568.9203
DSST 36.00910.95412.60500.87360.02128.7215
Ours 38.60610.95442.72810.88340.02017.9957
SVD 22.56400.80125.46890.79210.093212.0309
FPNSR 24.05450.82114.65070.81720.080411.8907
CEGATSR 25.68490.82634.41900.84100.082611.5119
EUNet20 dB24.91020.83624.50230.83490.075111.0029
MSDformer 25.59140.84024.25310.85630.069910.5203
DSST 25.90140.84154.19890.86230.066510.4911
Ours 28.06510.85214.72150.85900.060110.0718
Table 4. Comparisons of the Parameters, FLOPs on the XiongAn Dataset.
Table 4. Comparisons of the Parameters, FLOPs on the XiongAn Dataset.
FPNSRCEGATSREUNetMSDformerDDSTOurs
Parameters4.42 M13.55 M12.83 M14.90 M20.65 M12.77 M
FLOPs5.762 G29.925 G16.613 G53.915 G59.34 G38.89 G
PSNR33.161434.031634.987735.023535.672436.9011
SSIM0.90010.91020.91250.92710.92350.9460
SAM3.56523.45403.16543.10273.05622.8965
CC↑0.89850.90229.90750.91080.91560.9279
RMSE0.02040.02990.02850.02560.02440.0217
ERGAS5.45105.30215.36545.23475.21945.2024
Table 5. Ablation experiments of some variants of the proposed method over the Chikusei dataset at SNR 30 dB. Bold represents the best.
Table 5. Ablation experiments of some variants of the proposed method over the Chikusei dataset at SNR 30 dB. Bold represents the best.
VariantParams/FLOPsPSNR↑SSIM↑SAM↓RMSE↓
w/o Groups8.48 M/26.77 G30.02910.86713.56890.0588
w/o AE5.56 M/15.02 G30.24010.88234.09050.0627
w/o SA17.81 M/18.65 G31.56890.85643.65800.0460
w/o SA25.56 M/16.52 G30.65220.90882.65120.0362
Ours12.77 M/38.89 G35.09110.93102.29140.0256
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, H.; Yang, M.; Huang, H.; Zhang, M.; Zhang, W.; Jiao, Q.; Xu, L.; Tan, X. Hyperspectral Image Reconstruction Based on Blur–Kernel–Prior and Spatial–Spectral Attention. Remote Sens. 2025, 17, 1401. https://doi.org/10.3390/rs17081401

AMA Style

Xie H, Yang M, Huang H, Zhang M, Zhang W, Jiao Q, Xu L, Tan X. Hyperspectral Image Reconstruction Based on Blur–Kernel–Prior and Spatial–Spectral Attention. Remote Sensing. 2025; 17(8):1401. https://doi.org/10.3390/rs17081401

Chicago/Turabian Style

Xie, Hongyu, Mingyu Yang, Huansong Huang, Mingle Zhang, Wei Zhang, Qingbin Jiao, Liang Xu, and Xin Tan. 2025. "Hyperspectral Image Reconstruction Based on Blur–Kernel–Prior and Spatial–Spectral Attention" Remote Sensing 17, no. 8: 1401. https://doi.org/10.3390/rs17081401

APA Style

Xie, H., Yang, M., Huang, H., Zhang, M., Zhang, W., Jiao, Q., Xu, L., & Tan, X. (2025). Hyperspectral Image Reconstruction Based on Blur–Kernel–Prior and Spatial–Spectral Attention. Remote Sensing, 17(8), 1401. https://doi.org/10.3390/rs17081401

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop