Multiscale Spatial–Spectral Dense Residual Attention Fusion Network for Spectral Reconstruction from Multispectral Images
Abstract
:1. Introduction
- First, the 3-D MHR module is proposed to extract multiscale spatial and spectral features from MSIs. Specifically, the proposed 3-D MHR module not only expands the receptive field through the hierarchical residual structure but also enables fine-grained multiscale feature extraction by dividing the input feature into subsets, where each subset is processed sequentially to capture specific spatial and spectral details. In this way, the feature diversity and representation are enhanced.
- Second, a novel two-branch network architecture is designed, comprising a spectral and a spatial feature extraction branch, each equipped with an SSDR and an RCSA module. The SSDR module in the spectral branch employs a 1 × 1 × 3 convolution to capture spectral dependencies, whereas, in the spatial branch, it uses a 3 × 3 × 1 convolution to extract local spatial patterns. Moreover, instead of using dense connections, the SSDR module uses single-shot connections to facilitate feature reuse, which reduces complexity while ensuring effective feature propagation. The RCSA module is connected behind the SSDR module, which can adaptively recalibrate feature responses by emphasizing important spectral and spatial features while enhancing the global representation capability.
- Third, the AWFF module is introduced to integrate spectral and spatial features extracted from different branches to improve the reconstruction quality. Extensive experiments on four MSI datasets, including two simulated and two real-world datasets, using eight state-of-the-art methods, consistently demonstrates that the proposed MS2Net achieves superior reconstruction performance.
2. Related Works
2.1. Sparse Representation-Based Methods for SR
2.2. CNN-Based Methods for SR
2.3. Attention-Based Methods for SR
3. Proposed Method
3.1. Overall Pipeline
3.2. 3-D MHR Module
3.3. SSDR Module
3.4. RCSA Module
3.5. AWFF Module
4. Experiments
4.1. Dataset Description
- Pavia University (PU) Dataset: The first dataset was acquired by the reflective optics system imaging spectrometer airborne sensor over the University of Pavia, Italy. The spatial size of this dataset is 610 × 340 and includes 207,400 pixels, and it contains 9 labeled land-cover types. The ground sampling distance (GSD) is 1.3 m. Every pixel has 115 bands with a spectral wavelength ranging from 430 to 860 nm. After dropping the noise-contaminated bands, the number of spectral bands used for the experiment was 103. To match the requirements of the SR experimental setup, the corresponding MSI was simulated by the black-SWIR1 (i.e., 450–520 nm, 520–600 nm, 630–690 nm, and 760–900 nm) spectral response functions (SRFs) of Quick Bird. The size of the simulated MSI dataset is 610 × 340 × 4. As shown in Table 2, the training set size is 305 × 340 × 4, and the remaining regions were used as the test set.
- Indian Pines (IP) Dataset: The second dataset was photographed by the airborne visible/infrared imaging dpectrometer sensor over the Indian Pines test site in Northwestern Indiana, USA. The spatial size of this dataset is 145 × 145 and includes a total of 21,025 pixels, and it contains 16 labeled land-cover types. The GSD is 20 m. Every pixel has 220 bands with a spectral wavelength ranging from 400 to 2500 nm. To match the requirements of the SR experimental setup, the corresponding MSI was simulated by the SRFs of Sentinel-2. The size of the simulated MSI dataset is 145 × 145 × 13. As shown in Table 2, the training set size is 75 × 145 × 13, and the remaining regions were used as the test set.
- Chongqing (CQ) Dataset: The third dataset includes the real data collected by the ZY-1 02D satellite in Chongqing, China, which contains paired HSI and MSI for the same region and time. The advanced hyperspectral imager has a GSD of 30 m, while the multispectral imager provides MSIs with a GSD of 10 m. To make the GSD of the MSIs uniform with those of the HSIs, we performed a spatial downsampling operation on the MSI. We selected MSI and HSI with a size of 400 × 1000, containing 400,000 pixels. Every pixel of the HSI has 94 bands with a spectral wavelength ranging from 395 to 1341 nm. The MSI includes eight bands with a spectral wavelength ranging from 452 to 1047 nm. As shown in Table 2, the training set and test set sizes are 400 × 400 × 8 and 400 × 600 × 8, respectively.
- Jiaxing (JX) Dataset: The last dataset was collected by the ZY-1 02D satellite in Jiaxing, China. The GSDs for HSI and MSI are the same as those for the Chongqing dataset. We also performed a spatial downsampling operation on the MSI to ensure that the GSDs of the MSI and HSI were consistent. In this paper, we selected MSI and HSI with a size of 500 × 500, containing a total of 250,000 pixels. Each pixel has 76 bands for HSI and 8 bands for MSI. As shown in Table 2, the training set and test set sizes are 250 × 500 × 8 and 250 × 500 × 8, respectively.
4.2. Experimental Setup
- J-SLoL [9]: This method performs SR by learning the low-rank HS and MS dictionaries and their corresponding sparse representations.
- AWAN [27]: The backbone of this network is composed of several dual residual attention modules, where the long and short skip connections ensure adequate feature flow. Adaptive weighted channel attention is proposed to make the network focus on representative channels. In addition, the PReLU replaces the ReLU activation function to increase the nonlinear expressiveness.
- HSCNN [15]: This network consists of three stages to perform the SR task. It first applies 1 × 1 convolution for feature extraction; then, it stacks multiple densely connected blocks with a convolution kernel size of 3 × 3 for feature mapping. Finally, the 1 × 1 convolution is used for spectral reconstruction.
- SSJSR [45]: This method contains two sub-networks, namely, spatial super-resolution sub-network and SR sub-network. In this experiment, only the SR sub-network is adopted for comparison. This spectral sub-network consists of five 3-D convolutional layers with a kernel size of , and the network ends with a sub-pixel convolutional layer with a kernel size of to reconstruct the HSI.
- SSRAN [31]: This network utilizes three identical dual-branch residual blocks to extract spatial and spectral features of MSI, simultaneously. The bottom branch employs two 2-D convolutional layers with a kernel size of to extract the spectral information. Two 2-D convolutional layers with a kernel size of are used for the top branch to extract the spatial information. Similarly, the neighboring spectral attention is introduced at the end of each block.
- MST++ [39]: This network mainly consists of several single-stage spectral-wise Transformer (SST) modules. Each SST adopts a U-shaped structure consisting of an encoder and a decoder to extract multi-resolution contextual information. The fundamental unit of both the encoder and decoder is the spectral-wise attention (SA) block. Unlike conventional Transformer architectures, the SA block introduces spectral-wise multi-head self-attention to replace standard self-attention, reducing computational complexity while preserving performance.
- RepCPSI [23]: This network mainly consists of several polymorphic residual context restructuring (PRCR) modules. The PRCR module is similar to a basic residual structure, with the difference that it uses polymorphic convolution to extract features and that it equips a lightweight coordinate-preserving proximity spectral-aware attention block to enhance representational capabilities.
- MSFN [40]: Similar to MST++, the main structure of this network mainly consists of multiple single-stage spatial–spectral fusion networks (SSFNs). SSFN also has the structure of the U-Net to extract multiscale features. In addition, it uses the codec of Swin Transformer as the basic unit for feature extraction.
4.3. Evaluation Metrics
4.4. Experimental Results and Analysis
5. Discussion
5.1. Analysis of Ablation Experiments
5.2. Analysis of Model Complexity
5.3. Analysis of Classification Validation
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Zhang, B. Current status and future prospects of remote sensing. Bull. Chin. Acad. Sci. (Chin. Ver.) 2017, 32, 774–784. [Google Scholar]
- Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
- Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2014, 31, 34–44. [Google Scholar] [CrossRef]
- Zhang, B.; Wu, D.; Zhang, L.; Li, S.; Chen, X.; Zhao, Y.; Wang, H. Application of hyperspectral remote sensing for environment monitoring in mining areas. Environ. Earth Sci. 2012, 65, 649–658. [Google Scholar] [CrossRef]
- He, L.; Li, J.; Liu, C.; Li, S. Recent advances on spectral–spatial hyperspectral image classification: An overview and new guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
- Sun, X.; Zhang, L.; Yang, H.; Wu, T.; Cen, Y.; Guo, Y. Enhancement of spectral resolution for remotely sensed multispectral image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2198–2211. [Google Scholar] [CrossRef]
- He, J.; Yuan, Q.; Li, J.; Xiao, Y.; Liu, X.; Zou, Y. DsTer: A dense spectral transformer for remote sensing spectral super-resolution. Int. J. Appl. Earth Obs. Geoinf. 2022, 109, 102773. [Google Scholar] [CrossRef]
- Yi, C.; Zhao, Y.-Q.; Chan, J.C.-W. Spectral super-resolution for multispectral image based on spectral improvement strategy and spatial preservation strategy. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9010–9024. [Google Scholar] [CrossRef]
- Gao, L.; Hong, D.; Yao, J.; Zhang, B.; Gamba, P.; Chanussot, J. Spectral superresolution of multispectral imagery with joint sparse and low-rank learning. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2269–2280. [Google Scholar] [CrossRef]
- Fotiadou, K.; Tsagkatakis, G.; Tsakalides, P. Spectral super resolution of hyperspectral images via coupled dictionary learning. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2777–2797. [Google Scholar] [CrossRef]
- Han, X.; Yu, J.; Luo, J.; Sun, W. Reconstruction from multispectral to hyperspectral image using spectral library-based dictionary learning. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1325–1335. [Google Scholar] [CrossRef]
- Arad, B.; Ben-Shahar, O. Sparse recovery of hyperspectral signal from natural RGB images. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016; Volume 9911, pp. 19–34. [Google Scholar]
- Aeschbacher, J.; Wu, J.; Timofte, R. In defense of shallow learned spectral reconstruction from RGB images. In Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017. [Google Scholar]
- Xiong, Z.; Shi, Z.; Li, H.; Wang, L.; Liu, D.; Wu, F. HSCNN: CNN-based hyperspectral image recovery from spectrally undersampled projections. In Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 518–525. [Google Scholar]
- Shi, Z.; Chen, C.W.; Xiong, Z.; Liu, D.; Wu, F. HSCNN+: Advanced CNN-based hyperspectral recovery from RGB images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1052–10528. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 630–645. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Can, Y.B.; Timofte, R. An efficient CNN for spectral reconstruction from RGB images. arXiv 2018, arXiv:1804.04647. [Google Scholar]
- Stiebel, T.; Koppers, S.; Seltsam, P.; Merhof, D. Reconstructing spectral images from RGB-images using a convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Zhao, Y.; Po, L.M.; Yan, Q.; Liu, W.; Lin, T. Hierarchical regression network for spectral reconstruction from RGB images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 422–423. [Google Scholar]
- Koundinya, S.; Sharma, H.; Sharma, M.; Upadhyay, A.; Manekar, R.; Mukhopadhyay, R.; Chaudhury, S. 2D-3D CNN-based architectures for spectral reconstruction from RGB images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 844–851. [Google Scholar]
- Fu, Y.; Zhang, T.; Zheng, Y.; Zhang, D.; Huang, H. Joint camera spectral response selection and hyperspectral image recovery. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 256–272. [Google Scholar] [CrossRef] [PubMed]
- Wu, C.; Li, J.; Song, R.; Li, Y.; Du, Q. RepCPSI: Coordinate-preserving proximity spectral interaction network with reparameterization for lightweight spectral super-resolution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5508313. [Google Scholar] [CrossRef]
- Hang, R.; Liu, Q.; Li, Z. Spectral super-resolution network guided by intrinsic properties of hyperspectral imagery. IEEE Trans. Image Process. 2021, 30, 7256–7265. [Google Scholar] [CrossRef]
- Dian, R.; Shan, T.; He, W.; Liu, H. Spectral super-resolution via model-guided cross-fusion network. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 10059–10070. [Google Scholar] [CrossRef]
- Ghaffarian, S.; Valente, J.; van der Voort, M.; Tekinerdogan, B. Effect of attention mechanism in deep learning-based remote sensing image processing: A systematic literature review. Remote Sens. 2021, 13, 2965. [Google Scholar] [CrossRef]
- Li, J.; Wu, C.; Song, R.; Li, Y.; Liu, F. Adaptive weighted attention network with camera spectral sensitivity prior for spectral reconstruction from RGB images. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1894–1903. [Google Scholar]
- Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
- Li, J.; Wu, C.; Song, R.; Xie, W.; Ge, C.; Li, B.; Li, Y. Hybrid 2-D–3-D deep residual attentional network with structure tensor constraints for spectral super-resolution of RGB images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2321–2335. [Google Scholar] [CrossRef]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
- Zheng, X.; Chen, W.; Lu, X. Spectral super-resolution of multispectral images using spatial–spectral residual attention network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
- Dai, T.; Cai, J.; Zhang, Y.; Xia, S.-T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 11057–11066. [Google Scholar]
- Wang, L.; Sole, A.; Hardeberg, J.Y. Densely residual network with dual attention for hyperspectral reconstruction from RGB images. Remote Sens. 2022, 14, 3128. [Google Scholar] [CrossRef]
- Li, J.; Du, S.; Song, R.; Wu, C.; Li, Y.; Du, Q. HASIC-Net: Hybrid attentional convolutional neural network with structure information consistency for spectral super-resolution of RGB images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
- Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
- Sun, W.; Wang, Y.; Liu, W.; Shao, S.; Yang, S.; Yang, G.; Ren, K.; Chen, B. STANet: A hybrid spectral and texture attention pyramid network for spectral super-resolution of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- Song, Q.; Li, J.; Li, C.; Guo, H.; Huang, R. Fully attentional network for semantic segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 22 February–1 March 2022; pp. 2280–2288. [Google Scholar]
- Cai, Y.; Lin, J.; Lin, Z.; Wang, H.; Zhang, Y.; Pfister, H.; Van Gool, L. MST++: Multi-stage spectral-wise transformer for efficient spectral reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 745–755. [Google Scholar]
- Wu, Y.; Dian, R.; Li, S. Multistage spatial–spectral fusion network for spectral super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2024; early access. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
- Li, S. Real HSI-MSI-PAN image dataset for the hyperspectral/multi-spectral/panchromatic image fusion and super-resolution fields. arXiv 2024, arXiv:2407.02387. [Google Scholar]
- Sha, L.; Zhang, W.; Zhang, B.; Liu, Z.; Li, Z. Spectral mixing theory-based double-branch network for spectral super-resolution. Remote Sens. 2023, 15, 1308. [Google Scholar] [CrossRef]
- Liu, M.; Pan, H.; Ge, H.; Wang, L. MS3Net: Multiscale stratified-split symmetric network with quadra-view attention for hyperspectral image classification. Signal Process. 2023, 212, 109153. [Google Scholar] [CrossRef]
- Mei, S.; Jiang, R.; Li, X.; Du, Q. Spatial and spectral joint super-resolution using convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4590–4603. [Google Scholar] [CrossRef]
- Ranchin, T.; Wald, L. Fusion of high spatial and spectral resolution images: The ARSIS concept and its implementation. Photogramm. Eng. Remote Sens. 2000, 66, 49–61. [Google Scholar]
- Dian, R.; Li, S. Hyperspectral image super-resolution via subspace-based low tensor multi-rank regularization. IEEE Trans. Image Process. 2019, 28, 5135–5146. [Google Scholar] [CrossRef]
- Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Summaries of the Third Annual JPL Airborne Geoscience Workshop, Volume 1: AVIRIS Workshop; JPL: Pasadena, CA, USA, 1992. [Google Scholar]
- Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
Module | Input Shape | Layer Operations | Kernel Size | Padding | Stride | Filters | Output Shape |
---|---|---|---|---|---|---|---|
Input | (5, 5, 4, 1) | Conv-3D and PReLU | (1, 1, 1) | (0, 0, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) |
3-D MHR | (5, 5, 4, 100) | Split | / | / | / | / | #4 (5, 5, 4, 25) |
(5, 5, 4, 25) | / | / | / | / | / | (5, 5, 4, 25) | |
(5, 5, 4, 25) | Conv-3D and PReLU | (3, 3, 3) | (1, 1, 1) | (1, 1, 1) | 25 | (5, 5, 4, 25) | |
#2 (5, 5, 4, 25) | Add and Conv-3D and PReLU | (3, 3, 3) | (1, 1, 1) | (1, 1, 1) | 25 | (5, 5, 4, 25) | |
#2 (5, 5, 4, 25) | Add and Conv-3D and PReLU | (3, 3, 3) | (1, 1, 1) | (1, 1, 1) | 25 | (5, 5, 4, 25) | |
#4 (5, 5, 4, 25) | Concatenate | / | / | / | / | (5, 5, 4, 100) | |
Spectral-SSDR | (5, 5, 4, 100) | Conv-3D and PReLU | (1, 1, 3) | (0, 0, 1) | (1, 1, 1) | 100 | (5, 5, 4, 100) |
(5, 5, 4, 100) | Conv-3D and PReLU | (1, 1, 3) | (0, 0, 1) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
(5, 5, 4, 100) | Conv-3D and PReLU | (1, 1, 3) | (0, 0, 1) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
#4 (5, 5, 4, 100) | Concatenate | / | / | / | / | (5, 5, 4, 400) | |
(5, 5, 4, 400) | Conv-3D and PReLU | (1, 1, 1) | (0, 0, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
#2 (5, 5, 4, 100) | Add | / | / | / | / | (5, 5, 4, 100) | |
(5, 5, 4, 100) | Conv-3D and PReLU | (1, 1, 4) | (0, 0, 0) | (1, 1, 1) | 100 | (5, 5, 100) | |
RCSA | (5, 5, 100) | Attention | / | / | / | / | (5, 5, 100) |
#2 (5, 5, 100) | Multiplication | / | / | / | / | (5, 5, 100) | |
Spatial-SSDR | (5, 5, 4, 100) | Conv-3D and PReLU | (3, 3, 1) | (1, 1, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) |
(5, 5, 4, 100) | Conv-3D and PReLU | (3, 3, 1) | (1, 1, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
(5, 5, 4, 100) | Conv-3D and PReLU | (3, 3, 1) | (1, 1, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
#4 (5, 5, 4, 100) | Concatenate | / | / | / | / | (5, 5, 4, 400) | |
(5, 5, 4, 400) | Conv-3D and PReLU | (1, 1, 1) | (0, 0, 0) | (1, 1, 1) | 100 | (5, 5, 4, 100) | |
#2 (5, 5, 4, 100) | Add | / | / | / | / | (5, 5, 4, 100) | |
(5, 5, 4, 100) | Conv-3D and PReLU | (1, 1, 4) | (0, 0, 0) | (1, 1, 1) | 100 | (5, 5, 100) | |
RCSA | (5, 5, 100) | Attention | / | / | / | / | (5, 5, 100) |
#2 (5, 5, 100) | Multiplication | / | / | / | / | (5, 5, 100) | |
AWFF | #2 (5, 5, 100) | Feature Fusion | / | / | / | / | (5, 5, 100) |
Output | (5, 5, 100) | Conv-2D and PReLU | (1, 1) | (0, 0) | (1, 1) | 103 | (5, 5, 103) |
Datasets | Training Set | Test Set | ||
---|---|---|---|---|
MSI | HSI | MSI | HSI | |
Pavia University | ||||
Indian Pines | ||||
Chongqing | ||||
Jiaxing |
Methods | Indian Pines Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
J-SLoL [9] | 2.927 | 38.374 | 2.071 | 0.961 | 3.404 |
AWAN [27] | 2.864 | 44.936 | 1.929 | 0.955 | 3.452 |
HSCNN [15] | 2.374 | 47.246 | 1.659 | 0.969 | 2.866 |
SSJSR [45] | 2.261 | 48.785 | 1.573 | 0.971 | 2.673 |
SSRAN [31] | 2.701 | 45.629 | 1.915 | 0.962 | 3.336 |
MST++ [39] | 2.088 | 49.148 | 1.453 | 0.975 | 2.474 |
RepCPSI [23] | 2.616 | 44.710 | 1.849 | 0.959 | 3.511 |
MSFN [40] | 2.025 | 49.635 | 1.410 | 0.977 | 2.385 |
MS2Net | 1.542 | 52.642 | 1.045 | 0.986 | 1.823 |
Methods | Pavia University Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
J-SLoL [9] | 3.247 | 42.357 | 3.647 | 0.975 | 10.456 |
AWAN [27] | 2.381 | 41.819 | 2.690 | 0.979 | 7.392 |
HSCNN [15] | 2.049 | 43.648 | 2.418 | 0.982 | 6.480 |
SSJSR [45] | 2.269 | 42.472 | 2.593 | 0.980 | 7.149 |
SSRAN [31] | 2.557 | 41.045 | 2.864 | 0.977 | 7.732 |
MST++ [39] | 2.121 | 43.058 | 2.448 | 0.981 | 6.581 |
RepCPSI [23] | 2.159 | 42.947 | 2.540 | 0.981 | 6.645 |
MSFN [40] | 2.357 | 41.773 | 2.665 | 0.979 | 7.224 |
MS2Net | 1.827 | 45.275 | 2.178 | 0.984 | 5.819 |
Methods | Chongqing Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
J-SLoL [9] | 11.217 | 27.973 | 6.862 | 0.582 | 29.763 |
AWAN [27] | 2.673 | 40.734 | 2.452 | 0.974 | 7.361 |
HSCNN [15] | 2.614 | 40.895 | 2.337 | 0.973 | 7.056 |
SSJSR [45] | 2.748 | 40.619 | 2.495 | 0.972 | 7.449 |
SSRAN [31] | 3.016 | 39.390 | 2.865 | 0.966 | 9.036 |
MST++ [39] | 3.083 | 39.530 | 2.485 | 0.969 | 7.709 |
RepCPSI [23] | 2.572 | 41.341 | 2.324 | 0.976 | 6.544 |
MSFN [40] | 3.133 | 39.182 | 2.655 | 0.967 | 8.346 |
MS2Net | 2.211 | 42.535 | 1.922 | 0.983 | 5.670 |
Methods | Jiaxing Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
J-SLoL [9] | 6.411 | 33.879 | 3.574 | 0.896 | 18.337 |
AWAN [27] | 3.403 | 39.962 | 1.310 | 0.963 | 7.123 |
HSCNN [15] | 3.033 | 41.645 | 1.136 | 0.972 | 5.413 |
SSJSR [45] | 3.104 | 41.249 | 1.215 | 0.969 | 5.919 |
SSRAN [31] | 3.154 | 40.875 | 1.283 | 0.969 | 6.360 |
MST++ [39] | 3.090 | 42.196 | 1.028 | 0.968 | 6.221 |
RepCPSI [23] | 2.921 | 41.275 | 1.008 | 0.971 | 6.585 |
MSFN [40] | 3.172 | 40.223 | 1.151 | 0.964 | 7.922 |
MS2Net | 2.209 | 43.932 | 0.737 | 0.984 | 4.443 |
Models | Indian Pines Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
W/O 3-D MHR | 1.668 | 52.080 | 1.139 | 0.984 | 1.964 |
W/O RCSA | 1.723 | 50.625 | 1.192 | 0.982 | 2.078 |
W/O Spe-RCSA | 1.612 | 52.392 | 1.102 | 0.985 | 1.896 |
W/O Spa-RCSA | 1.604 | 52.561 | 1.093 | 0.985 | 1.912 |
W/O Spe-Net | 1.656 | 52.015 | 1.143 | 0.984 | 1.946 |
W/O Spa-Net | 1.558 | 52.371 | 1.060 | 0.985 | 1.850 |
W/O AWFF | 1.551 | 52.355 | 1.055 | 0.986 | 1.849 |
MS2Net | 1.542 | 52.642 | 1.045 | 0.986 | 1.823 |
Models | Pavia University Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
W/O 3-D MHR | 1.841 | 45.216 | 2.180 | 0.984 | 5.849 |
W/O RCSA | 1.866 | 44.976 | 2.236 | 0.984 | 5.898 |
W/O Spe-RCSA | 1.835 | 45.231 | 2.192 | 0.984 | 5.866 |
W/O Spa-RCSA | 1.833 | 45.238 | 2.193 | 0.984 | 5.832 |
W/O Spe-Net | 1.868 | 44.915 | 2.214 | 0.984 | 5.941 |
W/O Spa-Net | 1.864 | 45.017 | 2.227 | 0.984 | 5.913 |
W/O AWFF | 1.859 | 45.040 | 2.211 | 0.984 | 5.914 |
MS2Net | 1.827 | 45.275 | 2.178 | 0.984 | 5.819 |
Models | Chongqing Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
W/O 3-D MHR | 2.729 | 40.624 | 2.534 | 0.971 | 7.049 |
W/O RCSA | 2.383 | 41.786 | 2.189 | 0.978 | 6.293 |
W/O Spe-RCSA | 2.725 | 40.806 | 2.519 | 0.972 | 6.966 |
W/O Spa-RCSA | 2.665 | 40.679 | 2.259 | 0.976 | 6.956 |
W/O Spe-Net | 2.844 | 40.454 | 2.611 | 0.969 | 7.151 |
W/O Spa-Net | 3.024 | 39.802 | 2.816 | 0.964 | 7.950 |
W/O AWFF | 2.611 | 41.082 | 2.338 | 0.975 | 6.616 |
MS2Net | 2.211 | 42.535 | 1.922 | 0.983 | 5.670 |
Models | Jiaxing Dataset | ||||
---|---|---|---|---|---|
RMSE (↓ 0) | PSNR (↑ ∞) | SAM (↓ 0) | SSIM (↑ 1) | EGRAS (↓ 0) | |
W/O 3-D MHR | 3.037 | 42.081 | 0.907 | 0.971 | 5.248 |
W/O RCSA | 2.386 | 43.143 | 0.980 | 0.979 | 4.932 |
W/O Spe-RCSA | 3.063 | 41.372 | 1.155 | 0.970 | 5.739 |
W/O Spa-RCSA | 3.047 | 41.416 | 1.136 | 0.971 | 5.748 |
W/O Spe-Net | 3.096 | 41.230 | 1.191 | 0.969 | 5.894 |
W/O Spa-Net | 3.973 | 38.427 | 1.952 | 0.945 | 9.524 |
W/O AWFF | 2.352 | 43.544 | 0.959 | 0.979 | 4.541 |
MS2Net | 2.209 | 43.932 | 0.737 | 0.984 | 4.443 |
Datasets | Metrics | AWAN [27] | HSCNN [15] | SSJSR [45] | SSRAN [31] | MST++ [39] | RepCPSI [23] | MSFN [40] | MS2Net |
---|---|---|---|---|---|---|---|---|---|
IP | Parameters (M) | 1.971 | 1.986 | 1.116 | 0.286 | 52.985 | 2.221 | 121.553 | 0.846 |
FLOPs (GMac) | 0.049 | 0.050 | 0.174 | 0.007 | 0.741 | 0.055 | 1.540 | 0.160 | |
Runtime (ms) | 7.280 | 10.946 | 0.983 | 3.326 | 16.592 | 11.347 | 55.421 | 4.786 | |
PU | Parameters (M) | 1.844 | 1.910 | 0.507 | 0.278 | 11.675 | 2.130 | 26.761 | 0.655 |
FLOPs (GMac) | 0.046 | 0.048 | 0.047 | 0.007 | 0.164 | 0.053 | 0.340 | 0.053 | |
Runtime (ms) | 6.953 | 10.513 | 0.955 | 3.290 | 16.024 | 10.559 | 53.082 | 4.634 | |
CQ | Parameters (M) | 1.839 | 1.907 | 0.533 | 0.278 | 9.737 | 2.127 | 22.310 | 0.734 |
FLOPs (GMac) | 0.046 | 0.048 | 0.094 | 0.007 | 0.137 | 0.053 | 0.284 | 0.100 | |
Runtime (ms) | 6.867 | 10.107 | 0.912 | 3.266 | 16.133 | 10.057 | 51.390 | 4.594 | |
JX | Parameters (M) | 1.822 | 1.896 | 0.505 | 0.277 | 6.395 | 2.114 | 14.617 | 0.732 |
FLOPs (GMac) | 0.045 | 0.047 | 0.093 | 0.007 | 0.091 | 0.053 | 0.186 | 0.100 | |
Runtime (ms) | 6.695 | 10.002 | 0.883 | 3.228 | 15.960 | 9.967 | 50.643 | 4.473 |
Class | Color | J-SLoL [9] | AWAN [27] | HSCNN [15] | SSJSR [45] | SSRAN [31] | MST++ [39] | RepCPSI [23] | MSFN [40] | MS2Net | R-HSI |
---|---|---|---|---|---|---|---|---|---|---|---|
C1 | 70.87 | 82.72 | 88.44 | 94.22 | 87.11 | 93.64 | 83.61 | 91.83 | 90.66 | 87.55 | |
C2 | 0.83 | 28.35 | 54.09 | 57.78 | 43.44 | 45.26 | 48.37 | 73.60 | 56.96 | 61.06 | |
C3 | 86.69 | 95.36 | 92.63 | 87.25 | 90.08 | 93.31 | 82.58 | 81.50 | 92.63 | 92.63 | |
C4 | 99.01 | 91.47 | 90.11 | 95.52 | 90.58 | 95.15 | 96.22 | 96.31 | 98.35 | 97.41 | |
C5 | 0 | 92.86 | 100 | 66.66 | 86.66 | 66.66 | 42.85 | 40.00 | 93.33 | 86.66 | |
C6 | 0 | 62.11 | 47.91 | 52.08 | 42.70 | 76.84 | 59.57 | 67.67 | 55.20 | 56.25 | |
C7 | 96.77 | 92.04 | 94.82 | 93.93 | 95.22 | 93.25 | 95.02 | 92.15 | 95.46 | 96.11 | |
C8 | 94.74 | 96.77 | 92.10 | 97.89 | 90.52 | 98.41 | 98.38 | 98.97 | 98.42 | 98.42 | |
C9 | 96.23 | 88.00 | 90.69 | 91.19 | 92.67 | 95.30 | 94.30 | 94.00 | 95.16 | 98.63 | |
OA (%) | 84.73 | 85.94 | 88.57 | 89.38 | 87.88 | 90.44 | 88.98 | 90.26 | 91.61 | 92.59 | |
AA (%) | 60.57 | 81.07 | 83.42 | 81.83 | 79.89 | 84.20 | 77.88 | 81.78 | 86.24 | 86.08 | |
k × 100 | 80.03 | 82.16 | 85.52 | 86.55 | 84.60 | 87.90 | 85.91 | 87.71 | 89.36 | 90.59 |
Class | Color | J-SLoL [9] | AWAN [27] | HSCNN [15] | SSJSR [45] | SSRAN [31] | MST++ [39] | RepCPSI [23] | MSFN [40] | MS2Net | R-HSI |
---|---|---|---|---|---|---|---|---|---|---|---|
C1 | 83.31 | 85.68 | 81.92 | 88.28 | 83.35 | 85.02 | 86.11 | 84.97 | 88.12 | 92.00 | |
C2 | 97.18 | 97.38 | 98.13 | 97.38 | 97.38 | 97.54 | 97.24 | 97.77 | 97.51 | 98.59 | |
C3 | 75.24 | 80.03 | 74.20 | 83.03 | 76.22 | 81.05 | 81.22 | 78.27 | 81.24 | 89.29 | |
C4 | 74.93 | 80.49 | 88.05 | 86.30 | 83.21 | 89.68 | 85.93 | 86.63 | 88.72 | 93.92 | |
C5 | 63.31 | 60.50 | 66.83 | 59.08 | 65.09 | 75.05 | 70.67 | 71.00 | 74.22 | 90.66 | |
C6 | 84.88 | 86.57 | 85.60 | 83.80 | 83.68 | 68.41 | 71.84 | 67.25 | 88.11 | 92.53 | |
C7 | 77.92 | 86.39 | 85.71 | 81.12 | 83.29 | 83.17 | 83.54 | 69.29 | 81.06 | 87.59 | |
C8 | 99.23 | 99.69 | 100 | 99.84 | 99.86 | 99.83 | 99.47 | 99.84 | 99.85 | 99.69 | |
OA (%) | 86.47 | 87.84 | 88.48 | 88.01 | 87.66 | 89.25 | 88.67 | 87.39 | 90.22 | 94.82 | |
AA (%) | 82.01 | 84.59 | 85.05 | 84.86 | 84.00 | 84.97 | 84.50 | 81.88 | 87.35 | 93.03 | |
k × 100 | 80.87 | 82.81 | 83.78 | 83.11 | 82.61 | 84.97 | 84.09 | 82.33 | 86.30 | 92.82 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, M.; Zhang, W.; Pan, H. Multiscale Spatial–Spectral Dense Residual Attention Fusion Network for Spectral Reconstruction from Multispectral Images. Remote Sens. 2025, 17, 456. https://doi.org/10.3390/rs17030456
Liu M, Zhang W, Pan H. Multiscale Spatial–Spectral Dense Residual Attention Fusion Network for Spectral Reconstruction from Multispectral Images. Remote Sensing. 2025; 17(3):456. https://doi.org/10.3390/rs17030456
Chicago/Turabian StyleLiu, Moqi, Wenjuan Zhang, and Haizhu Pan. 2025. "Multiscale Spatial–Spectral Dense Residual Attention Fusion Network for Spectral Reconstruction from Multispectral Images" Remote Sensing 17, no. 3: 456. https://doi.org/10.3390/rs17030456
APA StyleLiu, M., Zhang, W., & Pan, H. (2025). Multiscale Spatial–Spectral Dense Residual Attention Fusion Network for Spectral Reconstruction from Multispectral Images. Remote Sensing, 17(3), 456. https://doi.org/10.3390/rs17030456