Next Article in Journal
A Feasibility Study on Monitoring Earthquake-Caused Furniture Vibrations Using Radiofrequency Identification Sensor Tags
Next Article in Special Issue
Spatially Explicit Modeling of Anthropogenic Heat Intensity in Beijing Center Area: An Investigation of Driving Factors with Urban Spatial Forms
Previous Article in Journal
The Potential of Blockchain Technology in Dental Healthcare: A Literature Review
Previous Article in Special Issue
PCNN Model Guided by Saliency Mechanism for Image Fusion in Transform Domain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pansharpening Model of Transferable Remote Sensing Images Based on Feature Fusion and Attention Modules

1
School of Information Science and Engineering, Xinjiang University, Urumqi 830014, China
2
Key Laboratory of Software Engineering, Xinjiang University, Urumqi 830008, China
3
Key Laboratory of Signal Detection and Processing, Xinjiang Uygur Autonomous Region, Urumqi 830046, China
4
School of Software, Xinjiang University, Urumqi 830008, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3275; https://doi.org/10.3390/s23063275
Submission received: 6 February 2023 / Revised: 16 March 2023 / Accepted: 17 March 2023 / Published: 20 March 2023

Abstract

:
The purpose of the panchromatic sharpening of remote sensing images is to generate high-resolution multispectral images through software technology without increasing economic expenditure. The specific method is to fuse the spatial information of a high-resolution panchromatic image and the spectral information of a low-resolution multispectral image. This work proposes a novel model for generating high-quality multispectral images. This model uses the feature domain of the convolution neural network to fuse multispectral and panchromatic images so that the fused images can generate new features so that the final fused features can restore clear images. Because of the unique feature extraction ability of convolution neural networks, we use the core idea of convolution neural networks to extract global features. To extract the complementary features of the input image at a deeper level, we first designed two subnetworks with the same structure but different weights, and then used single-channel attention to optimize the fused features to improve the final fusion performance. We select the public data set widely used in this field to verify the validity of the model. The experimental results on the GaoFen-2 and SPOT6 data sets show that this method has a better effect in fusing multi-spectral and panchromatic images. Compared with the classical and the latest methods in this field, our model fusion obtained panchromatic sharpened images from both quantitative and qualitative analysis has achieved better results. In addition, to verify the transferability and generalization of our proposed model, we directly apply it to multispectral image sharpening, such as hyperspectral image sharpening. Experiments and tests have been carried out on Pavia Center and Botswana public hyperspectral data sets, and the results show that the model has also achieved good performance in hyperspectral data sets.

1. Introduction

In recent years, with the emergence of many high-resolution Earth observation satellites, such as “GeoEye1”, “SPOT6”, and “GaoFen-2”, remote sensing applications have been widely used in research fields such as geography and land surveying. In these research fields, remote sensing images are often required to have high spectral resolution and high spatial resolution. However, as far as the design of the current remote sensing system is concerned, the spectral and spatial resolution often cannot be maintained at a high level at the same time. The images acquired by different sensors are different in terms of geometric features, spectral resolution, and spatial resolution. Some sensors acquire rich scene spectral information but lack sufficient spatial information, such as multi-spectral images (MS). On the other hand, some sensors are good at capturing spatial information, but cannot capture reliable spectral information, such as panchromatic images (PAN). Images with high spatial resolution provide subtle geometric features, while images with high spectral resolution provide rich spectral information, which can be used to identify and analyze targets. To make full use of the information provided by multi-spectral images and panchromatic images, the usual method is to fuse low-resolution multi-spectral images with high-resolution panchromatic images of the same scene to generate images with more detailed spatial and spectral structures, that is, pansharpening.
Remote sensing images need to be preprocessed. One of the most basic preprocessing methods is the panchromatic sharpening of remote sensing images, which plays a vital role in subsequent tasks such as target detection, classification, and the semantic segmentation of remote sensing images [1,2,3,4]. In early research and development, panchromatic sharpening can be broadly classified into the following four categories: (1) multiresolution analysis (MRA) [5,6,7,8,9,10,11]; (2) component substitution (CS) [12,13,14,15,16,17,18,19]; (3 ) mixed methods (combining CS and MRA) [20,21,22]; and (4) model-based methods [23,24,25,26,27]. Among the above four algorithms, the component substitution method has the characteristics of high space and high fidelity, and this method is straightforward to implement. However, the component replacement method also has some disadvantages, such as ignoring the local differences between MS and PAN images, resulting in obvious spectral distortions in the final image fusion stage. Secondly, in terms of preserving spectral information, although the MRA method has achieved good results, the effect of image fusion is significantly affected by the number of image decomposition and filter types. There are also strict requirements for image registration. Based on the hybrid method, although the high spatial fidelity of CS and the ability of MRA to preserve spectral information are combined, this is not enough, and the fused image still has the problems of spectral distortion and spatial structure distortion. The model-based method does an excellent job of solving the spectral distortion problem, but the solution process of the inversion model is time-consuming and complicated.
In recent years, in the field of computer vision, convolutional neural networks (CNNs) have been widely used and have achieved excellent results. Therefore, for the problem of panchromatic sharpening, many remote sensing researchers have attempted to use deep learning methods to solve it. For example, Masi et al. [28] applied a convolution neural network to panchromatic sharpening and made a major research breakthrough. Based on the single-image super-resolution reconstruction algorithm [29], a three-layer convolutional neural network (PNN) was proposed, which regards panchromatic sharpening as image super-resolution reconstruction. A panchromatic sharpening network based on domain knowledge was designed by Yang et al. [30]. In order to deal with the panchromatic sharpening problem, Liu et al. [31] designed a two-branch fusion network.
In order to obtain a better fusion effect, methods based on deep learning can take advantage of the slight spectral distortion and the feature extraction ability of the robust convolutional neural network. In existing methods, either neural networks are used to extract spatial details, or the panchromatic sharpening problem is treated as a super-resolution reconstruction problem. However, artifacts or spectral distortion still exist in the fusion results produced by these methods. This is because previous methods often assume that panchromatic and multispectral images contain different information. There needs to be a better solution for defining spatial and spectral information and how to extract these two types of information separately. In addition, the defining characteristic of a PAN image is that it contains less spectral information but more spatial information. The defining characteristic of an MS image is that it contains more spectral information but has a lower spatial resolution. Therefore, how to design a model that can combine the advantages of the two types of images to generate an applicable hyperspectral panchromatic image is an urgent problem to be solved. However, the existing deep learning methods usually only consist of the linear stitching of simple feature maps in feature fusion, and not all feature maps play a positive role in the final fusion process.
To overcome these problems, this article proposes a dual-branch fusion net-work based on the attention mechanism to optimize feature fusion and solve the problem of pansharpening. In addition, we applied the proposed model to hyperspectral image sharpening (HSI sharpening), and the experimental results show that the model also achieved better results on hyperspectral datasets. This study makes the following contributions:
  • We provide an end-to-end deep learning model for remote sensing image panchromatic sharpening and realize the reconstruction in the CNN feature domain to generate higher quality panchromatic sharpened images;
  • In the feature fusion module, a channel attention mechanism is introduced to optimize feature fusion, allowing the network to focus on crucial information;
  • Our network is a general model that can be directly applied to the hyperspectral image (HSI) sharpening and panchromatic image sharpening. For example, our experiments show that our method can achieve state-of-the-art performance both qualitatively and quantitatively.
The organization of the rest of this paper is as follows: In the second part, related works and the materials needed for this article are introduced. The third part elaborates on the pansharpening model of remote sensing images based on the attention optimization feature fusion proposed in this paper. The fourth part provides the experimental details, including the experimental settings, comparison experiments, and ablation experiments. In the fifth part, the proposed model is applied to HSI sharpening to prove the universality of the proposed model. Finally, this article is summarized.

2. Related Work

2.1. Pansharpening Based on Deep Learning

In recent years, the application of convolution neural networks to remote sensing images has attracted increasing attention from researchers. For example, the design of a high-performance panchromatic sharpening algorithm represents one such study, using the characteristics of remote sensing images to perform spatiotemporal fusion and spatial spectrum fusion without increasing the costs relating to equipment and manpower. Yuan [32] and others improved PNN by adding a multi-scale feature extraction module, which could make full use of the spatial features in high-resolution images and achieve an improved fusion effect. Ma et al. [33] proposed an unsupervised generation countermeasure network (GAN), which could better preserve the spatial features in PAN images. Liu et al. [34] proposed a GAN network for the panchromatic sharpening of remote sensing images and the effective fusion of MS and PAN images. On the basis of the literature [30], Fu et al. [35] proposed a grouping multi-scale gap network structure to expand the perception domain of each network layer, effectively obtain fine-grained multi-scale context features, and improve the quality of fused images. Zhou et al. [36] proposed an unsupervised pansharpening network based on perceptual loss and an automatic encoder. Liu et al. [37] based on the fusion results of different adaptive ground averaging methods, combined with the complementary properties of CS and MRA methods, proposed a generalized sharpening weighted network. Li et al. [38] proposed a multi-scale perceptual dense coding convolutional neural network to generate high-quality pansharpened images. The authors in [39] proposed a model-based deep sharpening method called the gradient projection-based pansharpening neural network (GPPNN). This model regards panchromatic images and low-resolution multispectral images as two optimization problems related to depth prior to regularization and uses a gradient projection algorithm to solve these two problems. The edge information part-guided convolutional sparse coding network, SCSC-PNN, proposed by Xu et al. [40] is applied to panchromatic sharpening in the field of remote sensing. This model mainly uses edge information regularization to segment low-resolution multispectral images and obtain a panchromatic image correlation feature map and a panchromatic image non-correlation feature map. HARNN [41] is a residual neural network based on the mixed attention mechanism. The encoder attention module is developed in the feature extraction part to solve the accuracy calculation of ground object recognition caused by spectral distortion and a lack of spatial detail in the pansharpening method. Yan et al. [42] proposed a model-driven and data-driven network, which combines model-driven and data-driven methods, and introduces a depth prior as its implicit regularization, thus improving its data adaptability and representative ability to be applied to multimodal fusion tasks. Guan et al. [43] proposed a hyperspectral pansharpening method of a multi-level double-attention guided fusion network (MDA-Net). This method uses a three-stream structure to enable the network to solve the hyperspectral pansharpening problem by combining the inherent characteristics of each input and their correlation at the same time.
Although the above methods have made some achievements, their spectral and spatial information have not been fully used for MS and PAN image fusion, respectively; that is to say, the lack of valuable information fusion leads to a certain spectral distortion and blur in the final fusion image. In addition, when performing feature fusion, only simple linear stitching is performed on the feature map, which usually does not meet the requirements of the pansharpening task. Therefore, in order to overcome the existing problems, this paper proposes an attention-based two-branch fusion network to optimize feature fusion.

2.2. Attention Mechanism

Attention mechanisms have been widely used in several research areas, such as natural language processing and computer vision. They enable the network to focus on relevant information and filter unnecessary information like a human. The attention mechanism is divided into the channel, spatial, and hybrid domains. The purpose of the attention mechanism is to better adjust the feature learning process by assigning different weights to different positions.
SENet [44] is an early visual model used to explore the channel attention mechanism. Through the “squeeze-incentive” process, different weights are obtained for each feature channel, different weights are assigned to different channels, and the channels are added to the attention mechanism. Zhang et al. [45] introduced SENet into the field of super-resolution reconstruction and constructed very deep residual channel attention networks (RCAN), which achieved better super resolution (SR) performance. The convolutional block attention module (CBAM) [46] is used to focus the two latitudes of the network space and the channel at the same time through the series space and channel attention modules, in order to help the network to understand the “what” and the “where” aspects of the attention. The attention mechanism improves the feature representation ability of neural networks by focusing more on critical features while reducing the attention paid to other features. Li et al. [47] proposed a new way of using the attention mechanism, using channel attention to merge the characteristics of the two branches. Furthermore, Dai et al. [48] proposed a multi-scale attention feature fusion module to replace the traditional summation and splicing methods for completing feature fusion.

3. Methods

The structure of the converged network proposed in this paper is shown in Figure 1. It consists of three modules: feature extraction, optimized feature fusion, and image reconstruction. This section first introduces the specific content of these three modules, and then introduces the loss function used.

3.1. Feature Extraction Module

The feature extraction module in this paper uses two sub-networks with the same structure but different weights. As shown, the upper subnetwork takes a single-band panchromatic (PAN) image as input, and the lower subnetwork takes a four-band multispectral (MS) image as input. Both feature extraction subnetworks consist of three consecutive convolution kernels as a 3 × 3 convolutional layer, followed by a parametric rectified linear unit (PReLU). Most CNN architectures use the average or max pooling to obtain rotation- and scale-invariant features, but the pooling operation is not used in the fusion network described in this article because in pansharpening, detailed information is more critical.

3.2. Optimized Feature Fusion Module

The complementary information of the panchromatic image and the multispectral image is extracted using two feature maps obtained by the feature extraction module. For panchromatic sharpening, we fuse the extracted feature maps to obtain a multispectral image with high spatial and spectral resolution.
The importance of the feature map to be fused with the fusion result varies, as does the amount of information carried. Therefore, valuable information needs to be enhanced, and useless information needs to be suppressed, meaning each channel must be weighted. Therefore, we first linearly concatenate the characteristic graphs of the two sets of complementary information, and then, through a channel attention mechanism SENet [44], we obtain a one-dimensional vector containing the number of channels of the characteristic graph, which represents the importance weight of each channel, and then apply the weight to the corresponding channels, so as to enhance the useful information and suppress the role of useless information. The network structure diagram of the optimized feature fusion module (OFFM) is shown in Figure 2.

3.3. Image Reconstruction Module

After feature fusion, we must recover high-resolution multispectral images from the fused features. Here, we use the convolution of the three-layer 3 × 3 convolution kernels to reconstruct the fused feature images and recover the HRMS images of the four bands. At the same time, we also use a long jump connection to directly transfer the input MS image, which has been upsampled to the same size as the PAN image by spectral mapping, to the output, which complements the spectral information of the reconstructed image. This point will also be proved in the ablation experiment, demonstrating the influence of this long jump connection on the fusion result.

3.4. Loss Function

The loss function is another key factor affecting the image quality of super-resolution reconstruction, besides the network structure. The  2  loss function is often used in image reconstruction tasks [28,29,30].  2  loss can smooth the image and punish larger outliers while being less sensitive to smaller outliers. For further improvement, because the  1  loss function can handle other small outliers and preserve the edge information of the image, we use the  1  loss function to train the network. The  1  loss function can be expressed as Equation (1). In the ablation experiment, we also prove through experiments that the  1  paradigm can obtain better fusion results than the  2  paradigm.
1 = 1 N i = 1 N h X p ( i ) , X M ( i ) , θ Y ( i ) 1 ,
where N is the number of small batch training samples,  X p ( i )  and  X M ( i )  are PAN images and low-resolution MS images,  Y ( i )  is the corresponding high-resolution MS image, and  θ  is the parameter of the fusion network.

3.5. Evaluation Indicators

In order to compare the proposed method with some previous methods, we used six widely used metrics to evaluate them quantitatively.
(1)
The peak signal-to-noise ratio (PSNR) [49] is defined as
P S N R = 10 · log 10 M A X I 2 M S E = 20 · log 10 M A X I M S E
where MAXI is the maximum value that represents the color of the image point. The higher the PSNR value between two images, the less distorted the reconstructed image relative to the high-resolution image. MSE is defined as
M S E = 1 m n i = 0 m 1 j = 0 n 1 I ( i , j ) K ( i , j ) 2
where I and K are two images of size m × n, one of which is the noise approximation of the other.
(2)
The structural similarity (SSIM) index
SSIM [50] measures the overall fusion quality by calculating the mean, variance, and covariance of the fused image and the reference image. The SSIM measurement consists of three contrast modules, namely, brightness, contrast, and structure. Given two images, X and Y, of size M × N, the means and variances of X and Y and the covariance are represented by  u x u y δ x 2 δ y 2 , and  δ x y , respectively. The comparison functions that define the brightness, contrast, and structure are
ι ( X , Y ) = 2 μ x μ y + c 1 μ x 2 + μ y 2 + c 1
c ( X , Y ) = 2 δ x δ y + c 2 δ x 2 + δ y 2 + c 2
s ( X , Y ) = δ x y + c 3 μ x μ y + c 3
The combination of these three component factors is the SSIM indicator, which is defined as
S S I M ( X , Y ) = l ( X , Y ) α c ( X , Y ) β s ( X , Y ) γ
The closer the SSIM value is to 1, the higher the similarity between the two images.
(3)
Spectral angle mapper (SAM)
SAM [51] calculates the angle between the HRMS image and the fusion result to evaluate the spectral quality of the fusion result. The smaller the value, the better the spectral quality, with an ideal value of 0.
S A M = arccos I α J α I α J α
where  I a  and  J a  are the pixel vectors of the fused image and the reference image, respectively, at the distance point  α .
(4)
Relative dimensionless global error in synthesis (ERGAS)
ERGAS [52] can more comprehensively reflect the quality of the fusion result. The smaller the value of ERGAS, the better the result, with an ideal value of 0.
E R G A S = 100 h l i = 1 N ( R M S E 2 ( B i ) / M i 2 ) N
where h is the resolution of the high-resolution image, l is the resolution of the low-resolution image, N is the number of bands,  B i  is the  M S  image, and  M i  is the average of the emissivity value of the  M S  image.
(5)
Spatial correlation coefficient (SCC)
SCC [53] is used to evaluate the similarity of the spatial details of the fused image and the reference image, using a high-pass filter to extract the high-frequency information of the reference image, and to calculate the correlation coefficient (CC) [54]. This article uses a high Laplacian filter to obtain a high frequency, as follows:
F = 1 1 1 1 8 1 1 1 1
A higher SCC means that most of the spatial information of the PAN image is injected during the fusion process. The SCC is calculated between the fused image and the reference image. The final SCC is averaged over all bands of the MS image. The CC is calculated as
C C = i = 1 w j = 1 h ( X i , j μ X ) ( Y i , j μ Y ) i = 1 w j = 1 h ( X i , j μ X ) 2 ( Y i , j μ Y ) 2
where X is the fused image, Y is the reference image, w and h are the width and height of the image, respectively, and  μ  represents the average value of the image.
(6)
Quality index (Q)
Q [55] combines three factors to calculate image distortion: correlation loss, brightness distortion, and contrast distortion. It is defined as
Q = σ Z 1 , Z 2 σ Z 1 · σ Z 2 · 2 σ Z 1 · σ Z 2 σ Z 1 2 + σ Z 2 2 · 2 Z 1 ¯ · Z 2 ¯ Z 1 ¯ 2 · Z 2 ¯ 2
where  Z 1  and  Z 2  represent the b-th band of the fused image and the reference image, respectively. When Q is 1, this represents the best fidelity for reference.

4. Experiment

4.1. Dataset Introduction

We trained and tested our network on two datasets collected by GaoFen-2 and SPOT6, and compared it with a variety of advanced methods. Gaofen-2 is the first civilian optical remote sensing satellite independently developed by China, with a spatial resolution better than 1 m. It was launched on 19 August 2014. It has two cameras with high resolution (1 m panchromatic and 4 m multispectral). Launched on 9 September 2012, the SPOT6 satellite collects multispectral images with a spatial resolution of 6 m and full-color images with a spatial resolution of 1.5 m, including red, green, blue, and near-infrared. The relevant information on the GaoFen-2 and SPOT6 satellites is shown in Table 1.

4.2. Experimental Setup

We trained and tested our network on the GaoFen-2 and SPOT6 datasets, respectively. We cropped the sub-regions of 32 × 32 and 128 × 128 in the center from the MS and PAN image pairs as test images and used the remaining regions for training. Specifically, in each iteration of training, we randomly cut out 32 × 32 and 128 × 128 image pairs with the same spatial resolution from the training area for use as training images. Our training area and test area did not overlap, which was achieved by filling the test area with zeros in the training phase. Figure 3 shows the MS image of GaoFen-2 as an example to introduce the division of the training area and test area.
Our goal was to generate a multispectral image with the same size and spatial resolution as the PAN image. We evaluated the proposed model by comparing the obtained results with nonexistent reference images. According to the Wald, protocol [55], we first preprocessed the image with a 5 × 5 Gaussian filter with a standard deviation of 2. Using the raw MS image (HRMS) as a reference, four downsampled PAN and blurred low-resolution MS images were used as input. In the network, in order to make the input MS image match the resolution of the PAN image, the MS image was upsampled using the bicubic interpolation method. Additionally, our network was implemented using PyTorch, using the Adam optimizer to minimize the loss. The training was performed on an Nvidia 3090 GPU, and the learning rate was set to 0.0004. A total of 30,000 epochs were trained.

4.3. Ablation Experiment

To further verify the effect of the attention mechanism, spectral mapping, and the selection of the loss function on the impact of our proposed model, taking the Gaofen-2 dataset as an example, we first conducted ablation experiments on the following four models (as shown in Figure 4) with  2  as the loss function, and then replaced the loss function with our proposed  1  loss function, verifying that this loss function can improve the obtained results.
  • The original feature extraction and image reconstruction network (Original);
  • Using our attention-based optimization feature fusion module on the original network (Attention-original);
  • The addition of spectral mapping to the original network (Skip-original);
  • Using our optimized feature to fuse fuzzy and spectral mapping on the original network (Our- 2 ).
The quantitative indicators of the experimental results of these four models on the GaoFen-2 dataset are shown in Table 2. From Table 2, we can see that when the long-skip connection of spectral mapping is added separately, both SAM and ERGAS are improved. This shows that such spectral mapping is beneficial to the maintenance of the fusion result. At the same time, we can also find that the PSNR, SSIM, SCC, and Q indicators improved to a certain extent. This is because as the network deepens, the features obtained by convolution are often more advanced, and these more high-level features map the semantic and abstract information of the image. Therefore, recovering the detailed texture of an image is difficult for high-level features. Such a long-hop connection directly transfers the low-level features of the input to the output, which also solves this problem to a certain extent. From Table 2, we can also see that when using our optimized feature fusion module alone, PSNR, SSIM, SCC, and Q are all improved. This shows that optimizing feature fusion is effective in space preservation and that assigning different weights to fusion features is more conducive to pansharpening tasks.
In order to combine the advantages of the two, we used spectral mapping and optimized feature fusion, simultaneously, as shown in Figure 4d. It can be seen from Table 2 that the results of Our- 2 , both in terms of the spectral index and the spatial index, are significantly higher than those of the original, demonstrating that our- 2  model can save spectrum and space.
In order to verify that  1  is conducive to improving the performance of the model, we use the model with  2  and  1  as the loss function. The change in the loss function during the training process is shown in Figure 5. We can find that the  1  loss function can reduce the training error and improve the network convergence. Therefore, we use the  1  loss function to train our network. From the last row of Table 2, we can also see that when using  1  as the loss function, all indicators achieved the best results, which indicates that the  1  loss function is more conducive to pansharpening.

4.4. Comparison with Other Algorithms

There are currently several widely used techniques, and in this section, we compare the proposed method with these techniques, including principal component analysis (PCA) [56], the intensity-based saturation (IHS) transform fusion method [57], MTF-GLP with high-pass modulation (MGH) [58], deep-learning-based PNN [28], PanNet [30], TFNet [31], ResTFNet [31], and the multi-scale deep convolutional neural network (MSDCNN) [32].
We conducted a qualitative analysis of the simulated dataset. Figure 6 and Figure 7 show the pansharpening results of different algorithms on the two datasets. From the fusion results in Figure 6, we can see that IHS, MGH, and PCA have spectral distortions and some obvious spatial information loss, with PCA having the most obvious spatial and spectral distortions. Several other methods based on deep learning can produce visually satisfactory pansharpened images, but PanNet and TFNet have a certain loss and blurring of spatial details. Images obtained by our method can preserve both spectral information and relatively more detailed spatial information. From the fusion results on the GaoFen-2 dataset, it can be seen that PCA, MGH, and HIS also experience a severe loss of spectral and spatial information, especially PCA. Deep learning methods can achieve better fusion results, but RestFNet and PanNet have some details missing.
To more accurately evaluate the spatial and spectral distortion, we also highlight the difference between the fused image and the ground truth, that is, the residual map. The residual map in the second row shows that our model has relatively fewer details and textures, indicating that it is the best in terms of space preservation. At the same time, the overall smooth area of our model is shown in dark blue, indicating that all differences are close to 0, while other residual maps more or less contain obvious areas. Figure 6d–i and Figure 7d–f, in particular, show that the error is large. This shows that our method achieves better spectral retention.
At the same time, we also compare the algorithms from a quantitative perspective. Table 3 and Table 4 show the quantitative indicators on the two satellite datasets. The results in Table 3 and Table 4 show that although our model is not dominant in terms of model size, it has the best results for both spatial and spectral indicators and is also suboptimal in time. The results show that our model has the best performance in terms of spectrum preservation and spatial reconstruction. At the same time, MSDCNN, based on the deep learning method, also achieved better fusion performance, surpassing traditional methods in both spectral and spatial indicators. On the GaoFen-2 dataset, TFNet achieved great performance, and on the SPOT6 dataset, PNN and ResTFNet also achieved great results. The results show that deep learning is still a very effective method at present, showing great potential in solving problems related to pansharpening. Then, we added the quantitative analysis of time and model parameter quantity. From the data in the two data sets in Table 3 and Table 4, we can see that the model we proposed is only inferior to the PNN model and superior to other traditional models and models based on deep learning. We analyze that the reason why our model does not reach the optimal value in time and model parameters is that the double-branch structure increases the computational cost.

5. Hyperspectral Image Sharpening

Our model is directly applicable to other types of multispectral image sharpening models, and the proof is given below. In this section, the proposed model is applied to hyperspectral image (HSI) sharpening. In remote sensing image processing, HIS sharpening has attracted more and more attention. It was designed to fuse high-resolution MS images and low-resolution hyperspectral images in order to obtain images with high spatial and spectral resolution.
We compare the proposed model with the seven latest models of TF-Net [31], Res-TFNet [31], MSD-CNN [32], SSF-CNN [59], Con-SSFCNN [59], and SSR-NET [60] on the Pavia Center (Pavia) and Botswana hyperspectral datasets. Table 5 and Table 6 show the experimental results on the Pavia dataset and the Botswana dataset. It can be seen that our model has the best performance in all indicators. From the results, it can be seen that our network has the best performance in terms of preserving spatial and spectral information under the effect of HIS sharpening. This also proves that our network can be applied to different tasks and that it is a general model.
At the same time, in Figure 8 and Figure 9, we also give the sharpening results of our model and other models on the Pavia and Botswana datasets for visual comparison. We can see that the images generated by TFNet, ResTFNet, MSDCNN, SSFCNN, ConSSFCNN, and SSRNet are clearly blurred. On the contrary, our model can maintain space and spectrum at the same time. We use pseudo colors in the corresponding residual maps to reflect the sharp differences between the results and the ground truth. We can see that the residual map of our model is displayed in dark blue as a whole, indicating that all of the differences are close to 0. Other residual maps contain more or less obvious areas, indicating that the error is larger, which also indicates that our method achieves better spectral preservation.

6. Conclusions

In this article, we propose a dual-branch fusion network based on attention-optimized feature fusion for the panchromatic sharpening of remote sensing images in the feature domain. This network is an end-to-end model that only requires the input of panchromatic and multispectral images to generate high-resolution multispectral images. Complementary information is extracted from the input image through two sub-networks with the same structure and different weights. The feature fusion is optimized through a channel attention mechanism to consider the relationship between other channels of the fusion features so that the network focuses more on critical information to improve the retention performance of the fused image in terms of spatial and spectral information. Compared with existing algorithms, this method shows better performance.

Author Contributions

Conceptualization, H.L. and Y.Q.; methodology, H.L.; software, H.L., L.D. and Y.D.; validation, L.D., Y.D. and X.Z.; formal analysis, H.L.; investigation, H.L. and L.D.; resources, Y.Q.; data curation, H.L.; writing—original draft preparation, H.L.; writing—review and editing, L.D.; visualization, H.L.; supervision, Y.Q.; project administration, H.L.; funding acquisition, Y.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61966035 and 62266043), the National Science Foundation of China (U1803261), the Natural Science Foundation of XinJiang Uygur Autonomous Region (2021D01C077), and the Autonomous Region Graduate Innovation Project (XJ2019G069, XJ2021G062, and XJ2020G074).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

The authors would like to thank all of the reviewers for their valuable contributions to this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSMulti-spectral
PANPanchromatic
CSComponent substitution
MRAMultiresolution analysis
CNNConvolutional neural network
HSIHyperspectral image
GANGenerative adversation network
SRSuper resolution
HRMSHigh-resolution MS image
OFFMOptimized feature fusion module
PReLUParametric rectified linear unit
PSNRPeak signal-to-noise ratio
SSIMStructural similarity
SAMSpectral angle mapper
ERGASRelative dimensionless global error in synthesis
SCCSpatial correlation coefficient
QQuality
PCAPrincipal component analysis
IHSIntensity-based saturation
RCANVery deep residual channel attention networks
CBAMConvolutional block attention module
MGHMTF-GLP with high-pass modulation

References

  1. Yilmaz, C.S.; Yilmaz, V.; Gungor, O.; Shan, J. Metaheuristic pansharpening based on symbiotic organisms search optimization. ISPRS J. Photogramm. Remote Sens. 2019, 158, 167–187. [Google Scholar] [CrossRef]
  2. Qu, Y.; Qi, H.; Ayhan, B.; Kwan, C.; Kidd, R. DOES multispectral / hyperspectral pansharpening improve the performance of anomaly detection? In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2017, Fort Worth, TX, USA, 23–28 July 2017; pp. 6130–6133. [Google Scholar]
  3. Lanir, J.; Maltz, M.; Yatskaer, I.; Rotman, S.R. Comparing Multispectral Image Fusion Methods for a Target Detection Task. In Proceedings of the 9th International Conference on Information Fusion, FUSION 2006, Florence, Italy, 10–13 July 2006; pp. 1–5. [Google Scholar]
  4. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  5. Shensa, M.J. The discrete wavelet transform: Wedding the a trous and Mallat algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef] [Green Version]
  6. Burt, P.J.; Adelson, E.H. The Laplacian Pyramid as a Compact Image Code. IEEE Trans. Commun. 1983, 31, 532–540. [Google Scholar] [CrossRef]
  7. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [Green Version]
  8. Shah, V.P.; Younan, N.H.; King, R.L. An Efficient Pan-Sharpening Method via a Combined Adaptive PCA Approach and Contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  9. Choi, M.; Kim, R.Y.; Nam, M.; Kim, H.O. Fusion of multispectral and panchromatic Satellite images using the curvelet transform. IEEE Geosci. Remote Sens. Lett. 2005, 2, 136–140. [Google Scholar] [CrossRef]
  10. Ghahremani, M.; Ghassemian, H. Remote-sensing image fusion based on curvelets and ICA. Int. J. Remote Sens. 2015, 36, 4131–4143. [Google Scholar] [CrossRef]
  11. Ji, X.; Zhang, G. Image fusion method of SAR and infrared image based on Curvelet transform with adaptive weighting. Multimed. Tools Appl. 2017, 76, 17633–17649. [Google Scholar] [CrossRef]
  12. Carper, W.; Lillesand, T.; Kiefer, R. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  13. Tu, T.; Huang, P.S.; Hung, C.; Chang, C. A fast intensity-hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  14. González-Audícana, M.; Saleta, J.L.; Catalán, R.G.; García, R. Fusion of multispectral and panchromatic images using improved IHS and PCA mergers based on wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1291–1299. [Google Scholar] [CrossRef]
  15. Pohl, C.; Genderen, J.L.V. Review article multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sensing 1998, 19, 823–854. [Google Scholar] [CrossRef] [Green Version]
  16. Shahdoosti, H.R.; Ghassemian, H. Combining the spectral PCA and spatial PCA fusion methods by an optimal filter. Inf. Fusion 2016, 27, 150–160. [Google Scholar] [CrossRef]
  17. Gillespie, A.R.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images. II. Channel ratio and “chromaticity” transformation techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar] [CrossRef]
  18. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  19. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  20. Valizadeh, S.A.; Ghassemian, H. Remote sensing image fusion using combining IHS and Curvelet transform. In Proceedings of the 6th International Symposium on Telecommunications, IST 2012, Tehran, Iran, 6–8 November 2012; pp. 1184–1189. [Google Scholar]
  21. Luo, Y.; Liu, R.; Zhu, Y.F. Fusion of remote sensing image base on the PCA+ ATROUS wavelet transform. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1155–1158. [Google Scholar]
  22. Cheng, J.; Liu, H.; Liu, T.; Wang, F.; Li, H. Remote sensing image fusion via wavelet transform and sparse representation. ISPRS J. Photogramm. Remote Sens. 2015, 104, 158–173. [Google Scholar] [CrossRef]
  23. Aly, H.A.; Sharma, G. A Regularized Model-Based Optimization Framework for Pan-Sharpening. IEEE Trans. Image Process. 2014, 23, 2596–2608. [Google Scholar] [CrossRef] [Green Version]
  24. Wei, Q.; Dobigeon, N.; Tourneret, J. Bayesian Fusion of Multi-Band Images. IEEE J. Sel. Top. Signal Process. 2015, 9, 1117–1127. [Google Scholar] [CrossRef] [Green Version]
  25. Li, S.; Yang, B. A New Pan-Sharpening Method Using a Compressed Sensing Technique. IEEE Trans. Geosci. Remote Sens. 2011, 49, 738–746. [Google Scholar] [CrossRef]
  26. Zhu, X.; Bamler, R. A Sparse Image Fusion Algorithm With Application to Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
  27. He, X.; Condat, L.; Bioucas-Dias, J.M.; Chanussot, J.; Xia, J. A New Pansharpening Method Based on Spatial and Spectral Sparsity Priors. IEEE Trans. Image Process. 2014, 23, 4160–4174. [Google Scholar] [CrossRef]
  28. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. PanshaRpening Convolutional Neural Networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
  29. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J.W. PanNet: A Deep Network Architecture for Pan-Sharpening. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 1753–1761. [Google Scholar]
  31. Liu, X.; Liu, Q.; Wang, Y. Remote sensing image fusion based on two-stream fusion network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef] [Green Version]
  32. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef] [Green Version]
  33. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  34. Liu, Q.; Zhou, H.; Xu, Q.; Liu, X.; Wang, Y. PSGAN: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10227–10242. [Google Scholar] [CrossRef]
  35. Fu, X.; Wang, W.; Huang, Y.; Ding, X.; Paisley, J.W. Deep Multiscale Detail Networks for Multiband Spectral Image Sharpening. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2090–2104. [Google Scholar] [CrossRef]
  36. Zhou, C.; Zhang, J.; Liu, J.; Zhang, C.; Fei, R.; Xu, S. PercepPan: Towards Unsupervised Pan-Sharpening Based on Perceptual Loss. Remote Sens. 2020, 12, 2318. [Google Scholar] [CrossRef]
  37. Liu, J.; Feng, Y.; Zhou, C.; Zhang, C. PWNet: An Adaptive Weight Network for the Fusion of Panchromatic and Multispectral Images. Remote Sens. 2020, 12, 2804. [Google Scholar] [CrossRef]
  38. Li, W.; Liang, X.; Dong, M. MDECNN: A Multiscale Perception Dense Encoding Convolutional Neural Network for Multispectral Pan-Sharpening. Remote Sens. 2021, 13, 535. [Google Scholar] [CrossRef]
  39. Xu, S.; Zhang, J.; Zhao, Z.; Sun, K.; Liu, J.; Zhang, C. Deep gradient projection networks for pan-sharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1366–1375. [Google Scholar]
  40. Xu, S.; Zhang, J.; Sun, K.; Zhao, Z.; Huang, L.; Liu, J.; Zhang, C. Deep convolutional sparse coding network for pansharpening with guidance of side information. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  41. Liu, Q.; Han, L.; Tan, R.; Fan, H.; Li, W.; Zhu, H.; Du, B.; Liu, S. Hybrid attention based residual network for pansharpening. Remote Sens. 2021, 13, 1962. [Google Scholar] [CrossRef]
  42. Yan, Y.; Liu, J.; Xu, S.; Wang, Y.; Cao, X. MD3Net: Integrating Model-Driven and Data-Driven Approaches for Pansharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  43. Guan, P.; Lam, E.Y. Multistage dual-attention guided fusion network for hyperspectral pansharpening. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar] [CrossRef]
  44. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  45. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the Computer Vision, ECCV 2018, 15th European Conference, Munich, Germany, 8–14 September 2018; Volume 11211, pp. 294–310. [Google Scholar]
  46. Woo, S.; Park, J.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision, ECCV 2018, 15th European Conference, Munich, Germany, 8–14 September 2018; Volume 11211, pp. 3–19. [Google Scholar]
  47. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 510–519. [Google Scholar]
  48. Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional Feature Fusion. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, WACV 2021, Waikoloa, HI, USA, 3–8 January 2021; pp. 3559–3568. [Google Scholar]
  49. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
  50. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  51. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Summaries of the Third Annual JPL Airborne Geoscience Workshop; Jet Propulsion Laboratory: Pasadena, CA, USA, 1992; Volume 1, pp. 147–149. [Google Scholar]
  52. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the Third International Conference Fusion of Earth Data, Nice, France, 26–28 January 2000; pp. 99–103. [Google Scholar]
  53. Zhou, J.; Civco, D.; Silander, J. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  54. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L.M. Comparison of pansharpening algorithms: Outcome of the 2006 GRS-S data-fusion contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef] [Green Version]
  55. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  56. Kwarteng, P.; Chavez, A. Extracting spectral contrast in Landsat Thematic Mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  57. Rahmani, S.; Strait, M.; Merkurjev, D.; Möller, M.; Wittman, T. An Adaptive IHS Pan-Sharpening Method. IEEE Geosci. Remote Sens. Lett. 2010, 7, 746–750. [Google Scholar] [CrossRef] [Green Version]
  58. Vivone, G.; Restaino, R.; Mura, M.D.; Licciardi, G.; Chanussot, J. Contrast and Error-Based Fusion Schemes for Multispectral Image Pansharpening. IEEE Geosci. Remote Sens. Lett. 2014, 11, 930–934. [Google Scholar] [CrossRef] [Green Version]
  59. Han, X.; Shi, B.; Zheng, Y. SSF-CNN: Spatial and Spectral Fusion with CNN for Hyperspectral Image Super-Resolution. In Proceedings of the 2018 IEEE International Conference on Image Processing, ICIP 2018, Athens, Greece, 7–10 October 2018; pp. 2506–2510. [Google Scholar]
  60. Zhang, X.; Huang, W.; Wang, Q.; Li, X. SSR-NET: Spatial-Spectral Reconstruction Network for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5953–5965. [Google Scholar] [CrossRef]
Figure 1. The overall framework diagram of the dual-branch remote sensing image fusion network based on the attention mechanism to optimize feature fusion. The inputs are MS and PAN images, and the output is HRMS. The network architecture is mainly composed of the following three sub-modules: (1) feature extraction module; (2) optimized feature fusion module (OFFM); and (3) image reconstruction module.
Figure 1. The overall framework diagram of the dual-branch remote sensing image fusion network based on the attention mechanism to optimize feature fusion. The inputs are MS and PAN images, and the output is HRMS. The network architecture is mainly composed of the following three sub-modules: (1) feature extraction module; (2) optimized feature fusion module (OFFM); and (3) image reconstruction module.
Sensors 23 03275 g001
Figure 2. Network structure of the optimized feature fusion module.
Figure 2. Network structure of the optimized feature fusion module.
Sensors 23 03275 g002
Figure 3. Examples of MS image training and testing regions for the GaoFen-2 dataset.
Figure 3. Examples of MS image training and testing regions for the GaoFen-2 dataset.
Sensors 23 03275 g003
Figure 4. Ablation experiment. (a) Original. (b) Attention-original. (c) Skip-original. (d) Our- 2 .
Figure 4. Ablation experiment. (a) Original. (b) Attention-original. (c) Skip-original. (d) Our- 2 .
Sensors 23 03275 g004
Figure 5. Changes in the loss function when using  1  and  2  in the training process.
Figure 5. Changes in the loss function when using  1  and  2  in the training process.
Sensors 23 03275 g005
Figure 6. Pansharpening results on the SPOT6 satellite dataset. The first line is the fusion result of the SPOT6 image, and the second line is the image obtained by pseudo-coloring the differences between the fusion result of the first line and the ground truth RGB image (that is, (a)). All images are displayed in true color (red, green, and blue being the three bands).
Figure 6. Pansharpening results on the SPOT6 satellite dataset. The first line is the fusion result of the SPOT6 image, and the second line is the image obtained by pseudo-coloring the differences between the fusion result of the first line and the ground truth RGB image (that is, (a)). All images are displayed in true color (red, green, and blue being the three bands).
Sensors 23 03275 g006
Figure 7. Pansharpening results on the GaoFen-2 satellite dataset. The first line is the fusion result of the GaoFen-2 image, and the second line is the image obtained by pseudo-coloring the differences between the fusion result of the first line and the ground truth RGB image (that is, (a)). All images are displayed in true color (red, green, and blue being the three bands).
Figure 7. Pansharpening results on the GaoFen-2 satellite dataset. The first line is the fusion result of the GaoFen-2 image, and the second line is the image obtained by pseudo-coloring the differences between the fusion result of the first line and the ground truth RGB image (that is, (a)). All images are displayed in true color (red, green, and blue being the three bands).
Sensors 23 03275 g007
Figure 8. Sharpening results on the Botswana dataset. The first row is the fusion result, and the second row is an image obtained by pseudo-coloring the differences between the fusion result of the first row and the ground truth RGB image. All images are displayed in true color (red, green, and blue being the three bands).
Figure 8. Sharpening results on the Botswana dataset. The first row is the fusion result, and the second row is an image obtained by pseudo-coloring the differences between the fusion result of the first row and the ground truth RGB image. All images are displayed in true color (red, green, and blue being the three bands).
Sensors 23 03275 g008
Figure 9. Sharpening results on the Pavia dataset. The first row is the fusion result, and the second row is an image obtained by pseudo-coloring the differences between the fusion result of the first row and the ground truth RGB image. All images are displayed in true color (red, green, and blue being the three bands).
Figure 9. Sharpening results on the Pavia dataset. The first row is the fusion result, and the second row is an image obtained by pseudo-coloring the differences between the fusion result of the first row and the ground truth RGB image. All images are displayed in true color (red, green, and blue being the three bands).
Sensors 23 03275 g009
Table 1. Spectral and spatial features of multispectral and panchromatic images from the SPOT6 and GF-2 satellites.
Table 1. Spectral and spatial features of multispectral and panchromatic images from the SPOT6 and GF-2 satellites.
SatelliteSpectral Wavelength (nm)Spatial Resolution (m)
PANBlueGreenRedNirPANMS
GaoFen-2450–900450–520520–590630–690770–8900.83.2
SPOT6455–745455–525530–590625–695760–8901.56
Table 2. The results of ablation experiments on the GaoFen-2 dataset (bold indicates the best results).
Table 2. The results of ablation experiments on the GaoFen-2 dataset (bold indicates the best results).
ModuleOFFMLong Skip   2   1 PSNRSSIMSAMERGASSCCQ
Original 29.70710.84140.07153.51730.95610.7686
Attention-original 29.80530.84900.06953.46570.95710.7799
Skip-original 29.83390.85300.06703.48260.95720.7879
Our- 2 29.99080.85660.06593.40990.95870.7957
Our 30.12020.86590.06263.36090.96010.8092
Table 3. Quantitative evaluation on the SPOT6 dataset (bold indicates the best results).
Table 3. Quantitative evaluation on the SPOT6 dataset (bold indicates the best results).
MethodsPSNRSSIMSAMERGASSCCQTIME(MS)MODLESIZE(M)
PCA19.29910.53590.361710.91390.76190.4612--
IHS23.37560.58450.10637.5650.76690.4933--
MGH22.80810.61230.09888.1540.7980.5431--
PNN28.76440.8480.07623.08570.91590.74921.240.31
PanNet28.29920.82890.08023.3610.90220.711534.170.08
TFNet28.74040.84610.07853.11230.91460.747379.199.03
ResTFNet28.6270.84970.07753.07640.91420.7554103.538.56
MSDCNN28.87150.85090.07543.05560.91870.754598.491.01
Our29.45930.87210.07022.79030.93210.789222.88.93
Table 4. Quantitative evaluation on the GaoFen-2 dataset (bold indicates the best results).
Table 4. Quantitative evaluation on the GaoFen-2 dataset (bold indicates the best results).
MethodsPSNRSSIMSAMERGASSCCQTIME(MS)MODLESIZE(M)
PCA20.44680.37360.237513.3050.72120.3063--
IHS21.10520.37860.152112.31920.72420.3188--
MGH21.13580.4360.172312.26930.78920.3929--
PNN29.43230.83210.07213.64380.95240.760958.240.31
PanNet29.07340.81920.07523.79750.94770.744423.430.08
TFNet29.46380.83820.07233.60670.95270.769489.979.03
ResTFNet28.68920.81620.07613.94230.94370.741184.928.56
MSDCNN29.76410.85340.06853.50180.95620.791548.391.01
Our30.12020.86590.06263.36090.96010.809233.108.93
Table 5. Quantitative evaluation on the Pavia dataset (bold indicates the best results).
Table 5. Quantitative evaluation on the Pavia dataset (bold indicates the best results).
MethodsRMSEPSNRERGASSAMTIME(MS)
SSFCNN7.990430.07944.988410.479147.18
ConSSFCNN4.263435.53564.73844.541343.78
SSRNET3.397337.50823.87213.795638.47
MSDCNN4.240235.5834.77115.3778113.93
TFNET3.457937.35453.87064.4395168.31
ResTFNet3.240537.91853.63484.0871141.48
Our2.581339.89392.94133.552334.01
Table 6. Quantitative evaluation on the Botswana dataset (bold indicates the best results).
Table 6. Quantitative evaluation on the Botswana dataset (bold indicates the best results).
MethodsRMSEPSNRERGASSAMTIME(MS)
SSFCNN1.208329.582712.16635.7059130.24
ConSSFCNN1.731626.457618.06278.027991.67
SSRNET0.706734.242214.49733.428594.98
MSDCNN0.541136.56122.63022.5187197.18
TFNET0.417038.82472.13221.9714166.36
ResTFNet0.377339.69211.90421.7592178.96
Our0.340240.59291.36361.554532.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, H.; Deng, L.; Dou, Y.; Zhong, X.; Qian, Y. Pansharpening Model of Transferable Remote Sensing Images Based on Feature Fusion and Attention Modules. Sensors 2023, 23, 3275. https://doi.org/10.3390/s23063275

AMA Style

Liu H, Deng L, Dou Y, Zhong X, Qian Y. Pansharpening Model of Transferable Remote Sensing Images Based on Feature Fusion and Attention Modules. Sensors. 2023; 23(6):3275. https://doi.org/10.3390/s23063275

Chicago/Turabian Style

Liu, Hui, Liangfeng Deng, Yibo Dou, Xiwu Zhong, and Yurong Qian. 2023. "Pansharpening Model of Transferable Remote Sensing Images Based on Feature Fusion and Attention Modules" Sensors 23, no. 6: 3275. https://doi.org/10.3390/s23063275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop