Next Article in Journal
A Fast and Low-Cost Human Body 3D Scanner Using 100 Cameras
Previous Article in Journal
Automatic Recognition of Dendritic Solidification Structures: DenMap
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectrum Correction Using Modeled Panchromatic Image for Pansharpening

Graduate School of Engineering, Tohoku University, Sendai, Miyagi 980-8579, Japan
*
Author to whom correspondence should be addressed.
J. Imaging 2020, 6(4), 20; https://doi.org/10.3390/jimaging6040020
Submission received: 26 December 2019 / Revised: 18 March 2020 / Accepted: 4 April 2020 / Published: 6 April 2020

Abstract

:
Pansharpening is a method applied for the generation of high-spatial-resolution multi-spectral (MS) images using panchromatic (PAN) and multi-spectral images. A common challenge in pansharpening is to reduce the spectral distortion caused by increasing the resolution. In this paper, we propose a method for reducing the spectral distortion based on the intensity–hue–saturation (IHS) method targeting satellite images. The IHS method improves the resolution of an RGB image by replacing the intensity of the low-resolution RGB image with that of the high-resolution PAN image. The spectral characteristics of the PAN and MS images are different, and this difference may cause spectral distortion in the pansharpened image. Although many solutions for reducing spectral distortion using a modeled spectrum have been proposed, the quality of the outcomes obtained by these approaches depends on the image dataset. In the proposed technique, we model a low-spatial-resolution PAN image according to a relative spectral response graph, and then the corrected intensity is calculated using the model and the observed dataset. Experiments were conducted on three IKONOS datasets, and the results were evaluated using some major quality metrics. This quantitative evaluation demonstrated the stability of the pansharpened images and the effectiveness of the proposed method.

1. Introduction

The optical sensor of an Earth observation satellite receives radiances in the visible to infrared regions of the electromagnetic spectrum. The sensor simultaneously receives two kinds of data; the multi-spectral (MS) image with high spectral resolution and low spatial resolution, and a panchromatic (PAN) image with high spatial resolution and low spectral resolution. Satellites with such optical sensors include the IKONOS, QuickBird, GeoEye, and WorldView. Satellite data are widely used for various purposes such as change detection, object detection, target recognition, background for map application, and visual image analysis. Pansharpening is an image processing technique that generates high-spatial and high-spectral-resolution MS images using the spatial information from the PAN image and the spectral information from the MS images. It can be used for preprocessing in the data analysis and applications of satellite-image processing described above. The pansharpening methods can be divided into four categories: component substitution (CS), multi-resolution analysis (MRA), machine learning, and the hybrid methods. The CS methods substitute the spatial information of the MS images by the spatial information from the PAN image and then generate the MS images with high spatial resolution. This category of methods includes several techniques such as the intensity–hue–saturation (IHS) transform [1], principal component analysis [2], Gram–Schmidt (GS) transform [3,4], and the Brovey transform [2]. For these techniques, it is known that the difference in the spectral characteristics between the replaced spatial component and the original spatial component gives rise to spectral distortion. The MRA class extracts the spatial information from the PAN image and then adds it to the MS image. For these methods, several techniques such as the decimated wavelet transform [5], undecimated wavelet transform [6], a “trous” wavelet transform [7], Laplacian pyramid [8], curvelet transform [9], contourlet transform [10], nonsubsampled contourlet transform [11], and shearlet transform [12,13] are used to extract the detailed spatial information. Although these methods increase the quality of the spectral information, ringing artifacts called spatial distortion may occur. Machine learning methods use techniques such as dictionary learning, sparse modeling [14], deep learning [15], and the Bayesian paradigm [16]. The performance of these methods depends on the amount of the available data, and their computational complexity is greater than that of the other methods. For example, sparse modeling incurs a significant computational cost for creating a dictionary. The hybrid methods combine several techniques to exploit their advantageous features and therefore can achieve higher performance than the methods in other classes. Recently, machine learning techniques have been widely used. Wang et al. [17] proposed a method using sparse representation and a convolution neural network. Fei et al. [18] and Yin [19] proposed an improved Sparse Representation-based details injection (SR-D) [14]. For the CS-based category, Imani [20] proposed a method that removes the noise and redundant spatial features to improve the band-dependent spatial detail (BDSD) [21].
Many methods have been proposed over the past several decades in order to reduce the spectral distortion and to enhance the spatial resolution [22,23,24] and in particular, pansharpening includes a process for correcting the image intensity. Two types of techniques are used for this process: the first uses the relative spectral response graph, and the second uses an intensity model based on the observed images. However, the numerical image quality of these methods depends on the image dataset, and it is difficult to obtain consistent results.
In this study, we propose a technique for correcting the intensity based on the IHS method. The IHS method is a known pansharpening method and substitutes the intensity of the red–green–blue (RGB) images with the intensity of the PAN images. Since the intensity of the pansharpened (PS) image is replaced by that of the PAN image, it contains a high level of spatial information. However, the PS image exhibits spectral distortion (i.e., color distortion) because of the differences in the spectral characteristics between the intensities of the RGB and PAN images [25]. To address this drawback, several methods have been proposed. Tu et al. [26] proposed a generalized IHS (GIHS) transform that transforms the IHS into a simple linear transform. Since GIHS can accelerate the calculations, it has been frequently used as a framework for the development of other methods focusing on correcting the intensity. Tu et al. [26] presented a fast IHS that corrects the intensity using the mean value of the intensity of the MS images and also presented a simple spectral-adjustment IHS method (SAIHS) [27] that corrects the intensity of the green and blue bands by exploring the best value from 92 IKONOS images. Tradeoff IHS is a method that controls the tradeoff of the spectral characteristics between the intensities of the RGB and PAN images [28]. Choi et al. [29] presented an improved SAIHS (ISAIHS) that calculates the intensity correction coefficients of the red and near-infrared (NIR) bands using 29 IKONOS images in addition to SAIHS. These approaches have the advantage that strong outliers are not generated because the correction coefficient is obtained from the images as a constant. However, the values of the constant may not be optimal for the processed image. Audicana et al. [30] proposed an expanded fast IHS with the spectral response function (eFIHS-SRF) method to correct the intensity of the PS image using the mean value of the MS images and the fraction of the photons detected by the sensors (i.e., the MS and PAN sensors). In this method, the correction is performed using the correction coefficient obtained from the relative spectral response graph and the observation data. However, the obtained results differ depending on the processed data. Garzelli et al. [21] presented the BDSD method that applies the correction coefficients calculated by the minimum-variance-unbiased estimator [31] to the MS images. The practical replacement adaptive CS (PRACS) [32] calculates the correction coefficients using high-frequency information and the characteristics of the MS images for each image dataset. The fusion approach (IHS-BT-SFIM) proposed by Tu et al. [33] calculates the intensity correction coefficients using the modeled low-spatial-resolution intensity of the MS images.
These methods use either unique or non-unique correction coefficients. In pansharpening, spectral distortion may be caused by three different effects: the relative spectral response of the sensors [27], aging of the optical and electronic payload [34], and the observation conditions [35]. Since the observation conditions differ for each dataset, the correction coefficients must be calculated separately for each image dataset. Even in the conventional methods, the correction coefficient is obtained from the intensity modeled based on the image. However, the use of a model formula obtained by combining the intensity with and without color information has not been considered. We estimate the PAN image without color information using the intensity of the RGB image and the intensity of the NIR image, and perform a detailed correction including the relative spectral response and observation conditions at each intensity of the RGB image with color information. In this study, we propose a novel model for low-spatial-resolution PAN images using MS images. Compared to other related methods for intensity correction, our method showed consistently good performance in terms of the numerical image quality. Therefore, it was concluded that unlike for the methods based on IHS, the proposed method can reduce spectral distortion and obtain results that are independent of the processed image.
Note that there are multiple satellites that have sensors whose characteristics are similar to those of IKONOS, such as Quick Bird and GeoEye. Therefore, experiments have been conducted on IKONOS images in many studies in the literature. Our proposed method can also be applied to those satellite images.

2. Materials and Methods

2.1. Image Datasets

The three image datasets from the IKONOS used for the experiments are listed in Table 1. The first was collected in May 2008 and covers the city of Nihonmatsu, Japan. The second was collected in May 2006 and covers the city of Yokohama, Japan. Both the Nihonmatsu and Yokohama datasets were provided by the Japan Space Imaging Corporation, Japan. The third dataset covering Mount Wellington in Hobart, Tasmania in Australia was collected in February 2003 and was provided by Space Imaging, LLC. These datasets have PAN and MS images with the spatial resolutions of 1 m and 4 m, respectively. The original dataset contains: a PAN image with 1024 × 1024 pixels and MS images with 256 × 256 pixels for the Nihonmatsu region, a PAN image with 1792 × 1792 pixels and MS images with 448 × 448 pixels for the Yokohama region, and a PAN image with 12,112 × 13,136 pixels and MS images with 3028 × 3284 pixels for the Hobart region in Tasmania.
To evaluate the quality of the PS image, we experimented with the test images and original images according to the Wald protocol [36]. The test images were used to evaluate the numerical image quality, and the original images were used as reference images for numerical and visual evaluation. We regard the original images as ground truth images. The spatial resolution of the test PAN image was reduced from 1 to 4 m and that of the test MS image was reduced from 4 to 16 m. Hence, the test image datasets have a PAN image with 256 × 256 pixels and MS images with 64 × 64 pixels for the Nihonmatsu region, a PAN image with 448 × 448 pixels and MS images with 112 × 112 pixels for the Yokohama region, and a PAN image with 3028 × 3284 pixels and MS images with 757 × 821 pixels for the Hobart region in Tasmania.

2.2. Proposed Method

It is considered that a high-resolution PAN image can be estimated from the corresponding high-resolution MS images. However, it is not obvious how to combine the MS images, and it may differ depending on the spectral response and observation conditions. Based on this idea, we propose a novel technique for correcting the intensity that can reduce the spectral distortion in each image by modeling the PAN image. The technique does not require detailed knowledge of sensor characteristics; in other words, it only requires the image dataset. The procedure of the proposed method includes a technique for estimating the intensity correction coefficients and also a technique for image fusion. The former method first models the PAN image with a low spatial resolution, using the MS images, and the coefficients are then calculated using the method of least squares in the comparison between the PAN image and the modeled PAN image. The latter method calculates the high-spatial-resolution intensity of the RGB image and then generates a PS image using GIHS.

2.2.1. Notation

I h i g h ( i ) and I l o w ( i ) denote the i-th pixel intensities of the high- and low-spatial resolution RGB images, respectively. P A N h i g h ( i ) and P A N l o w ( i ) denote the i-th pixel high- and low-spatial resolution, respectively. M S b B l o w ( i ) denotes the i-th pixel low-spatial resolution MS image in the b band and B = { r e d , g r n , b l u , n i r } ; red, grn, blu, and nir are the red, green, blue, and NIR, respectively. | B | denotes the total number of MS bands. · denotes the scalar multiplication of matrices, × denotes multiplication, and denotes a matrix product. N denotes the total number of the pixels in an image.

2.2.2. High-Spatial-Resolution Intensity of the RGB Image

The intensity of the RGB image is calculated by the IHS transform and is represented by Equation (1).
I l o w ( i ) = M S r e d l o w ( i ) + M S g r n l o w ( i ) + M S b l u l o w ( i ) 3
The intensity ratio between the PAN and RGB images is the same for the high- and low-spatial resolutions. Thus, if we have both PAN and RGB images with high- and low-spatial resolutions with the same number of pixels, this intensity relationship can be expressed as:
I h i g h ( i ) : P A N h i g h ( i ) = I l o w ( i ) : P A N l o w ( i )
The dimensions of the PAN and MS images used for the above process are the same as those of the MS images before processing, P × P . The PAN image is generated by down-sampling with bicubic interpolation.
To obtain an expression for the high-spatial-resolution intensity of the RGB image, Equation (2) can be rewritten as
I h i g h ( i ) = P A N h i g h ( i ) P A N l o w ( i ) × I l o w ( i )

2.2.3. Low-Spatial-Resolution PAN Image Model

Since the observation wavelength range of the PAN sensor includes the wavelength ranges of all of the MS sensors, we modeled the low-spatial-resolution PAN image using low-spatial-resolution MS images. Based on the relative spectral response graph of IKONOS [37] depicted in Figure 1, we noticed the following features:
  • The relative spectral response of the PAN band shows low sensitivity from the blue to green bands [27].
  • The relative spectral responses of the PAN band show low sensitivity in some regions of the red and NIR bands.
  • The relative spectral responses of the red, green, blue, and NIR bands partly overlap with those of their neighboring bands [4,27].
  • The PAN band includes the NIR band.
We design the low-spatial-resolution PAN image model including the above features as follows:
P A N l o w ( i ) = I l o w ( i ) + α · M S n i r l o w ( i ) β · M S b l u l o w ( i ) γ · M S g r n l o w ( i ) ξ · M S r e d l o w ( i )
where α , β , γ , and ξ denote the correction coefficients of the NIR, blue, green, and red bands, respectively. For the right-hand side of Equation (4), the third and fourth terms are affected by feature 1, the second and fifth terms are affected by feature 2, the second to fifth terms are affected by feature 3, and the second term is affected by feature 4. The first and second terms on the right-hand side of Equation (3) estimate the PAN image using the intensities of the RGB and NIR images. The third, fourth, and fifth terms are considered to be the overflowing and overlapping parts of the MS bands in the relative spectral response graph. The overflowing part of the NIR band is included in the coefficient α .

2.2.4. Estimation of the Intensity Correction Coefficient

The high- and low-spatial images observed under the same conditions have similar intensity characteristics, and the relationship between the high- and low-spatial resolution PAN images can be expressed as P A N h i g h ( i ) P A N l o w ( i ) . Therefore, the intensity correction coefficients are obtained when the sum of the differences between the high- and low-spatial resolution PAN images reaches the minimum value, as expressed by the root mean square error that is computed as:
argmin α , β , γ , ξ 1 N i = 1 N ( P A N h i g h ( i ) P A N l o w ( i ) ) 2   s . t .   α , β , γ , ξ 0
Equation (5) is used with the observed data—that is, the PAN and MS images, because the observed data include the relative spectral response of the sensors and the observation conditions. In practice, we use Equation (6) instead of Equation (5). The term ( P A N h i g h ( i ) P A N l o w ( i ) ) in Equation (5) can be rearranged as ( P A N h i g h ( i ) I l o w ( i ) ) + ( α · M S n i r l o w ( i ) + β · M S b l u l o w ( i ) + γ · M S g r n l o w ( i ) + ξ · M S r e d l o w ( i ) ) .
To express the above formula in matrix representation, we define
A = [ M S n i r l o w ( 1 ) M S b l u l o w ( 1 ) M S g r n l o w ( 1 ) M S r e d l o w ( 1 ) M S n i r l o w ( N ) M S b l u l o w ( N ) M S g r n l o w ( N ) M S r e d l o w ( N ) ]
c = [ α β γ ξ ] , d = [ I l o w ( 1 ) P A N h i g h ( 1 ) I l o w ( N ) P A N h i g h ( N ) ] ,   A N × | B | ,   c | B | ,   d N  
We compute the intensity correction coefficients using the least square method, as described in Equation (6):
argmin c A c d 2 2     s . t .     c 0 .

2.2.5. Fusion Process

The fusion process generates a PS image with GIHS, using the intensity correction coefficients and the observed images. Since GIHS is a simple fusion technique, it is the optimal technique for the evaluation of the image quality performance. The validity of the GIHS formula was rigorously proved for RGB images in Tu et al. [26], and the extension to NIR images has been derived from the form of the equation. Therefore, this time we applied it to RGB images. Figure 2 shows the procedures of the fusion process, which are also listed below:
  • Change the MS images into the same size as that of the PAN image using bicubic interpolation, and produce enlarged MS images M S b B l o w .
  • Calculate the enlarged low-spatial-resolution intensity I l o w from M S b { r e d , g r n , b l u } l o w , as expressed by Equation (1).
  • Calculate the high-spatial-resolution intensity I h i g h of M S b { r e d , g r n , b l u } l o w using the estimated correction coefficients, M S b { r e d , g r n , b l u } l o w , I l o w , and P A N h i g h using Equations (3) and (4).
  • Synthesize the PS image from I l o w , M S b B l o w , and I h i g h with GIHS, as expressed by Equation (7).
    [ M S r e d P S ( i ) M S g r n P S ( i ) M S b l u P S ( i ) ] = [ M S r e d l o w   ( i ) + I h i g h ( i ) I l o w ( i ) M S g r n l o w   ( i ) + I h i g h ( i ) I l o w ( i ) M S b l u l o w   ( i ) + I h i g h ( i ) I l o w ( i ) ]

3. Results

3.1. Experimental Setup

The procedure of the proposed method used the image sizes listed in Table 2. Let the sizes of the original PAN and MS images be R P r o w × R P c o l and P r o w × P c o l , respectively, where R is the ratio of the number of vertical and horizontal pixels of the PAN image relative to that for the MS image. Then, the size of the down-sampled test PAN image and test MS image used for the estimation intensity correction coefficients in Table 2 is P r o w R × P c o l R , and the size of the Test PAN image and the up-sampled test MS image used for image fusion is P r o w × P c o l . The test image was generated by bicubic spline interpolation. Due to the possible loss of data, the procedure for the estimation of the intensity correction coefficients does not use up-sampling. The correction coefficients were determined by solving Equation (5) using the non-negative least-squares method. The correction coefficient is the value for which the closest agreement is obtained between the spectral characteristics of the high-resolution PAN image and those of the low-resolution PAN image. The bicubic interpolation was used for down-sampling and up-sampling of the images used for the experiment of the proposed method shown in Table 2. For a fair comparison, the up-sampling of the MS images of the related works was also carried out using bicubic interpolation.

3.2. Image Quality Metric

3.2.1. Notation

O b ( i ) and O b ¯ denote the i-th pixel value of the b-band reference image and its mean value, respectively, and P S b ( i ) and P S b ¯ denote the i-th pixel value of the b-band PS image and its mean value, respectively. N and | B | are the total number of pixels in the entire image for each band and the number of bands in the PS image, respectively. σ O b and σ P S b are the variances of the reference and PS images in the b-band, respectively, and σ O b , P S b denotes the covariance of the reference and PS images in the b-band. h and l denote the spatial resolution of the PAN and MS images, respectively.

3.2.2. Numerical Quality Metrics

To evaluate the numerical image quality, we employed four metrics: the correlation coefficient (CC), university image quality index (UIQI) [38], erreur relative globale adimensionnelle de synthese (ERGAS) [39], and the spectral angle mapper (SAM) [40]. All of the metrics measure the spectral distortion, and UIQI, ERGAS, and SAM are global metrics.
The CC measures the correlation between the images and ranges from –1.0 to 1.0. A CC value closer to 1.0 implies a stronger correlation between the spectral information of the PS image and the original image. The CC is given by
CC = 1 | B | × b B C C b CC b = i = 1 N ( O b ( i ) O b ¯ ) × ( P S b ( i ) P S b ¯ ) i = 1 N ( O b ( i ) O b ¯ ) 2 × i = 1 N ( P S b ( i ) P S b ¯ ) 2
UIQI [38] comprehensively measures the value of the loss of correlation, intensity distortion, and contrast distortion. The loss of correlation measures the degree of the linear correlation between the images. The intensity distortion measures the closeness of the mean intensity values of the images. Contrast distortion measures the similarity of the contrasts of the images. These values range from –1.0 to 1.0. A UIQI value closer to 1.0 implies smaller values of the loss of correlation, intensity distortion, and contrast distortion, so that a higher UIQI value corresponds to a higher quality of the PS image. UIQI is given by
UIQI = 1 | B | × b B U I Q I b UIQI b = σ O b , P S b σ O b · σ P S b × 2 · O b ¯ · P S b ¯ ( O b ¯ ) 2 + ( P S b ¯ ) 2 × 2 · σ O b · σ P S b σ O b 2 + σ P S b 2
ERGAS [39] measures the global image quality with a lower ERGAS value corresponding to a higher spectral quality of the PS image, and it is given by
ERGAS = 100 × h l × 1 | B | × b B ( ( R M S E b ) 2 ( P S b ¯ ) 2 ) RMSE b = 1 N × i = 1 N ( O b ( i ) P S b ( i ) ) 2
RMSE b is the root-mean-square error between the reference image and the PS image in the b-band.
SAM [40] measures the global spectral distortion with the value closer to 0.0 corresponding to weaker spectral distortion, and is given by
SAM = 1 N i = 1 N S A M ( i ) SAM ( i ) = cos 1 ( b B O b ( i ) × P S b ( i ) b B ( O b ( i ) ) 2 × b B ( P S b ( i ) ) 2 )

3.3. Experimental Results

The intensity correction coefficients were estimated using Equation (5). Table 3 lists the estimated intensity correction coefficients for the three datasets, where α represents the fraction of the NIR included in the PAN image, and β ,     γ ,     ξ are the fractions of the image where the intensity of the RGB image that does not match the PAN image.
The PS image of the proposed method was compared to those obtained by the related methods for intensity correction, namely fast IHS [26], SAIHS [27], ISAIHS [29], Tradeoff IHS [28], eFIHS-SRF [30], PRACS [32], and Brovey Transform-Smoothing-Filter-based Intensity Modulation (BT-SFIM) [33]. PRACS [32] was performed using the code developed by Vivone et al. [23]. The detailed parameters of the existing methods were as follows: the weight parameter for SAIHS was (Green,Blue) = (0.75,0.25), the weight parameters for ISAIHS were (Red,Green,Blue,NIR) = (0.3,0.75,0.25,1.7), the tradeoff parameter for the Tradeoff IHS was 4.0, the fraction of the number of photons in the MS band detected by the PAN sensor of eFIHS-SRF was 0.8, and the mean filter kernel size and the weight parameter for BT-SFIM were 7 × 7 and (Red,Green,Blue,NIR) = (0.26,0.26,0.122,0.375), respectively. The numerical image quality was evaluated using the CC, UIQI [38], ERGAS [39], and SAM [40] metrics. The sliding window size of UIQI was 8 × 8 . Table 4 and Table 5 summarize the image quality results for CC and UIQI, and for ERGAS and SAM, respectively. It is observed that with the exception of the image of Tasmania, ISAIHS gave good results. The proposed method gave the best UIQI and ERGAS values for all of the images. The SAM values of eFIHS-SRF and the proposed method are not correlated with the size of the processed image, while the values obtained by the other methods decrease with the increasing number of pixels in the processed image. The Tradeoff IHS and the proposed method gave consistently good results for all of the images. Figure 3 shows the ranking of the quality metric results for the seven methods. Here, for each test image, the best result is worth three points, the second-best result is worth two points, and the third-best result is worth one point. The maximum possible number of points is 36, which is obtained when a method has the best values for all of the metrics. BT-SFIM uses the coefficients estimated from a relative spectral response graph and does not give good results. eFIHS-SRF using the coefficients estimated by the relative spectral response graph is the second-best method after Tradeoff IHS. In contrast, ISAIHS and Tradeoff IHS use the coefficients estimated from large image datasets and gave good results. These results show that techniques using image datasets tend to perform better than those using the relative spectral response graph. This demonstrates the need to consider other observation conditions in addition to the spectral response graphs, and it can be concluded that the observation dataset contains this information.
Visual analyses are shown in Figure 4 Figure 5 and Figure 6. In this evaluation, the PS images generated from the test images were compared to the ground truth RGB images. In Figure 4 and Figure 5, (d)–(l) are the expanded images corresponding to the area surrounded by the yellow box in (c). As shown in Figure 4, eFIHS-SRF (Figure 4i), BT-IHS (Figure 4k), and the proposed method (Figure 4l) reproduced the color tone of the forest (indicated by red arrows), while the color of the rice field in Figure 4i for eFIHS-SRF was darker (indicated by green arrows). The green component of PRACS (Figure 4j) was brighter than expected (indicated by white arrows). As shown in Figure 5, SAIHS (Figure 5f) and ISAIHS (Figure 5g) images were generally brighter, and the eFIHS-SRF image (Figure 5i) was darker. In Figure 6, eFIHS-SRF (Figure 6i) and the proposed method (Figure 6l) reproduced the overall color tone, while the PRACS (Figure 6j) image had a brighter appearance than the images obtained using the other methods, which were generally whitish. In summary, for the proposed method (Figure 6l), the color tone of the whole image was consistent with the ground truth image for all images, and the resolution was also good. The green component of the PRACS image (Figure 6j) is brighter. The results obtained by the other methods differed depending on the image.

4. Discussion

The experimental results show that for SAM, only the proposed method and eFIHS-SRF for which the coefficients are estimated by the relative spectral response graph do not depend on the size of the processed image. This indicates that the overall color was not destroyed. In other words, there is little spectral distortion. Tradeoff IHS performs well on average for all images; however, it does not exhibit best fitting for the processed images. Since each method has advantages and disadvantages, we calculated the ranking of the quality metric. According to the results of this ranking, the proposed method is the best, followed by Tradeoff IHS and eFIHS-SRF. Tradeoff IHS is a method of uniquely calculating the correction coefficients from the image datasets included in the modeling method, and eFIHS-SRF is a method of uniquely calculating the correction coefficients using the relative spectral response graph. The techniques for intensity correction use two methods: the first uses the relative spectral response graph and the other uses modeling of the obtained image datasets. The comparative results show that the latter technique provides better performance. Since this method calculates the intensity correction coefficients from the observed data that include the effect of all factors, it is able to obtain a good result. For modeling techniques such as SAIHS [27], ISAIHS [25], and PRACS [31], the modeled intensity is mostly expressed as I l o w = b B ω b M S b l o w or P A N l o w = b B ω b M S b l o w , where ω b denotes the b-band correction coefficients; subsequently, some correction coefficients are acquired by its formula. While the previously used modeling techniques generally obtain correction coefficients by optimization using image datasets of satellites or observations, the proposed technique is optimized using an image dataset. The proposed method obtains the best result in the quality metrics score for all of the image datasets; in particular, the UIQI and ERGAS metrics gave consistently good results. The results show that the proposed technique reduces the spectral distortion compared to other related conventional techniques. This suggests that the model can be considered to be adequate. For the estimation of the correction coefficients, the proposed technique calculates the intensity correction coefficients for each of the observed data, thus obtaining the better numerical quality metrics than some of the other modeling techniques that calculate the intensity correction coefficients separately for each satellite. This suggests that the estimation of the correction coefficients for each image dataset is adequate because of the different observation conditions and spectral characteristics of the PAN and MS sensors.

5. Conclusions

This study proposed a novel model of low-spatial-resolution PAN images for pansharpening, and the intensity correction coefficients were computed using this model and the obtained image dataset. The PS image is generated using its coefficients and GIHS. The proposed model is formulated according to the characteristics of the relative spectral response graph. Due to its inclusion of subtraction, the design of this model is different from that of a conventional model. Therefore, the correction coefficients calculated for each image correspond to the observation conditions and sensor characteristics. Compared to other related methods, the proposed method demonstrated consistently good performance. These results show that the proposed model is adequate and effective for estimating the intensity of pansharpening. However, the experiments were performed only on the images obtained from IKONOS; therefore, further verification experiments on images obtained from other optical sensors are required. We would like to consider finding new application as an important future work.

Author Contributions

Methodology, N.T. and Y.S.; Software, N.T.; Validation, Y.S.; Writing—Original Draft Preparation, N.T.; Writing—Review and Editing, Y.S. and S.O.; Supervision, S.O.; Project Administration, S.O.; Funding Acquisition, S.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the JSPS KAKENHI Grant Number 18K19772 and the Yotta Informatics Project by MEXT, Japan.

Acknowledgments

The authors thank the Japan Space Imaging Corporation and Space Imaging, LLC for providing the images.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pohl, C.; Van Genderen, J.L. Review article Multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar] [CrossRef] [Green Version]
  2. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  3. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  4. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS+Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  5. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 17, 674–693. [Google Scholar] [CrossRef] [Green Version]
  6. Nason, G.P.; Silverman, B.W. The Stationary Wavelet Transform and some Statistical Applications; Springer: New York, NY, USA, 1995; ISBN 9780387945644. [Google Scholar]
  7. Shensa, M.J. The Discrete Wavelet Transform: Wedding the À Trous and Mallat Algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef] [Green Version]
  8. Burt, P.J.; Adelson, E.H. The Laplacian Pyramid as a Compact Image Code. IEEE Trans. Commun. 1983, 31, 532–540. [Google Scholar] [CrossRef]
  9. Starck, J.L.; Candès, E.J.; Donoho, D.L. The curvelet transform for image denoising. IEEE Trans. Image Process. 2002, 11, 670–684. [Google Scholar] [CrossRef] [Green Version]
  10. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [Green Version]
  11. Zhang, Q.; Guo, B. Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process. 2009, 89, 1334–1346. [Google Scholar] [CrossRef]
  12. Guo, K.; Labate, D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007, 39, 298–318. [Google Scholar] [CrossRef] [Green Version]
  13. Luo, X.; Zhang, Z.; Wu, X. A novel algorithm of remote sensing image fusion based on shift-invariant Shearlet transform and regional selection. AEU Int. J. Electron. Commun. 2016, 70, 186–197. [Google Scholar] [CrossRef]
  14. Vicinanza, M.R.; Restaino, R.; Vivone, G.; Dalla Mura, M.; Chanussot, J. A pansharpening method based on the sparse representation of injected details. IEEE Geosci. Remote Sens. Lett. 2015, 12, 180–184. [Google Scholar] [CrossRef]
  15. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super—Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Fasbender, D.; Radoux, J.; Bogaert, P. Bayesian data fusion for adaptable image pansharpening. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1847–1857. [Google Scholar] [CrossRef]
  17. Wang, X.; Bai, S.; Li, Z.; Song, R.; Tao, J. The PAN and ms image pansharpening algorithm based on adaptive neural network and sparse representation in the NSST domain. IEEE Access 2019, 7, 52508–52521. [Google Scholar] [CrossRef]
  18. Fei, R.; Zhang, J.; Liu, J.; Du, F.; Chang, P.; Hu, J. Convolutional Sparse Representation of Injected Details for Pansharpening. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1595–1599. [Google Scholar] [CrossRef]
  19. Yin, H. PAN-Guided Cross-Resolution Projection for Local Adaptive Sparse Representation- Based Pansharpening. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4938–4950. [Google Scholar] [CrossRef]
  20. Imani, M. Band dependent spatial details injection based on collaborative representation for pansharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4994–5004. [Google Scholar] [CrossRef]
  21. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 228–236. [Google Scholar] [CrossRef]
  22. Thomas, C.; Ranchin, T.; Wald, L.; Chanussot, J. Synthesis of multispectral images to high spatial resolution: A critical review of fusion methods based on remote sensing physics. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1301–1312. [Google Scholar] [CrossRef] [Green Version]
  23. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2565–2586. [Google Scholar] [CrossRef]
  24. Ghamisi, P.; Rasti, B.; Yokoya, N.; Wang, Q.; Hofle, B.; Bruzzone, L.; Bovolo, F.; Chi, M.; Anders, K.; Gloaguen, R.; et al. Multisource and multitemporal data fusion in remote sensing: A comprehensive review of the state of the art. IEEE Geosci. Remote Sens. Mag. 2019, 7, 6–39. [Google Scholar] [CrossRef] [Green Version]
  25. Qiu, R.; Zhang, X.; Wei, B. An Improved Spatial/Spectral-Resolution Image Fusion Method Based on IHS-MC for IKONOS Imagery. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009; pp. 1–4. [Google Scholar]
  26. Tu, T.M.; Su, S.C.; Shyu, H.C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion 2001, 2, 177–186. [Google Scholar] [CrossRef]
  27. Tu, T.M.; Huang, P.S.; Hung, C.L.; Chang, C.P. A fast intensity-hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  28. Choi, M. A new intensity-hue-saturation fusion approach to image fusion with a tradeoff parameter. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1672–1682. [Google Scholar] [CrossRef] [Green Version]
  29. Choi, M.; Kim, H.H.; Cho, N. An improved intensity-hue-saturation method for IKONOS image fusion. Int. J. Remote Sens. 2006, 13. [Google Scholar]
  30. González-Audícana, M.; Otazu, X.; Fors, O.; Alvarez-Mozos, J. A low computational-cost method to fuse IKONOS images using the spectral response function of its sensors. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1683–1690. [Google Scholar] [CrossRef]
  31. Kay, S.M. Fundamentals of Statistical Signal Processing: Estimation Theory; PTR Prentice Hall: Upper Saddle River, NJ, USA, 1993; ISBN 9780133457117. [Google Scholar]
  32. Choi, J.; Yu, K.; Kim, Y. A new adaptive component-substitution-based satellite image fusion by using partial replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  33. Tu, T.M.; Hsu, C.L.; Tu, P.Y.; Lee, C.H. An adjustable pan-sharpening approach for IKONOS/QuickBird/GeoEye-1/ WorldView-2 imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 125–134. [Google Scholar] [CrossRef]
  34. Vivone, G.; Simoes, M.; Dalla Mura, M.; Restaino, R.; Bioucas-Dias, J.M.; Licciardi, G.A.; Chanussot, J. Pansharpening based on semiblind deconvolution. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1997–2010. [Google Scholar] [CrossRef]
  35. Otazu, X.; González-Audícana, M.; Fors, O.; Núñez, J. Introduction of Sensor Spectral Response into Image Fusion Methods. Application to Wavelet-Based Methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  36. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  37. IKONOS Relative Spectral Response. Available online: http://www.geoeye.com/CorpSite/assets/docs/technical-papers/2008/IKONOS_Relative_Spectral_Response.xls (accessed on 4 April 2010).
  38. Wang, Z.; Bovik, A.C. A Universal Image Quality Index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  39. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the Third Conference “Fusion of Earth Data: Merging Point Measurements, Raster Maps and Remotely Sensed Images”, Sophia Antipolis, France, 26–28 January 2000; Ranchin, T., Wald, L., Eds.; SEE/URISCA: Nice, France, 2000; pp. 99–103. [Google Scholar]
  40. Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W.; Heidebrecht, K.B.; Shapiro, A.T.; Barloon, P.J.; Goetz, A.F.H. The spectral image processing system (SIPS)-interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
Figure 1. IKONOS relative spectral response graph.
Figure 1. IKONOS relative spectral response graph.
Jimaging 06 00020 g001
Figure 2. Fusion process diagram of the proposed method.
Figure 2. Fusion process diagram of the proposed method.
Jimaging 06 00020 g002
Figure 3. Quality metric result rank in three tested datasets: the score of the quality metrics is expressed as a cumulative bar chart.
Figure 3. Quality metric result rank in three tested datasets: the score of the quality metrics is expressed as a cumulative bar chart.
Jimaging 06 00020 g003
Figure 4. Visual comparison of test Nihonmatsu images: (a) Test red–green–blue (RGB) image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d); The red arrow indicates where the forest colors are well reproduced. The green arrow indicates where the color of the rice field is dark. White arrows indicate where the green tones are light.
Figure 4. Visual comparison of test Nihonmatsu images: (a) Test red–green–blue (RGB) image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d); The red arrow indicates where the forest colors are well reproduced. The green arrow indicates where the color of the rice field is dark. White arrows indicate where the green tones are light.
Jimaging 06 00020 g004aJimaging 06 00020 g004b
Figure 5. Visual comparison of test Yokohama images: (a) Test RGB image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d).
Figure 5. Visual comparison of test Yokohama images: (a) Test RGB image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d).
Jimaging 06 00020 g005aJimaging 06 00020 g005b
Figure 6. Visual comparison of test Tasmania images: (a) Test RGB image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d).
Figure 6. Visual comparison of test Tasmania images: (a) Test RGB image; (b) Test PAN image; (c) Ground truth RGB image (i.e., Original RGB image); (d) Ground truth RGB image of the yellow box area in (c); (e–l) PS images generated by various methods corresponding to the area of (d).
Jimaging 06 00020 g006aJimaging 06 00020 g006b
Table 1. Characteristics of the original and test images of the image datasets. MS: multi-spectral, PAN: panchromatic.
Table 1. Characteristics of the original and test images of the image datasets. MS: multi-spectral, PAN: panchromatic.
ImageNihonmatsuYokohamaTasmania
Original imagePAN image1024 × 10241792 × 179212,112 × 13,136
MS image256 × 256448 × 4483028 × 3284
Test imagePAN image256 × 256448 × 4483028 × 3284
MS image64 × 64112 × 112757 × 821
Table 2. Image size for the estimation of the intensity correction coefficients and image fusion.
Table 2. Image size for the estimation of the intensity correction coefficients and image fusion.
ImageEstimation of Intensity Correction CoefficientsImage Fusion
PAN image
P A N h i g h
Down-sampled
Test PAN image ( P r o w R × P c o l R )
Test PAN image
( P r o w × P c o l )
MS images
M S b l o w
Test MS images
( P r o w R × P c o l R )
Up-sampled test MS images ( P r o w × P c o l )
Table 3. Estimated intensity correction coefficients.
Table 3. Estimated intensity correction coefficients.
Correction CoefficientNihonmatsuYokohamaTasmania
α   (NIR)0.38570.37890.5734
β   (Blue)0.21990.25490.2310
γ   (Green)0.19800.11230.0000
ξ   (Red)0.04860.10990.1039
Table 4. Correlation coefficient (CC) and university image quality index (UIQI) quality metrics results. IHS: intensity–hue–saturation, SAIHS: spectral-adjustment IHS method, ISAIHS: improved SAIHS, eFIHS-SRF: expanded fast IHS with the spectral response function, PRACS: practical replacement adaptive component substitution, BT-SFIM: brovey transform-smoothing filter based intensity modulation. The first row indicates the ideal values. For each metric, the best values are given in bold, the second-best values are underlined, and the third-best values are double underlined.
Table 4. Correlation coefficient (CC) and university image quality index (UIQI) quality metrics results. IHS: intensity–hue–saturation, SAIHS: spectral-adjustment IHS method, ISAIHS: improved SAIHS, eFIHS-SRF: expanded fast IHS with the spectral response function, PRACS: practical replacement adaptive component substitution, BT-SFIM: brovey transform-smoothing filter based intensity modulation. The first row indicates the ideal values. For each metric, the best values are given in bold, the second-best values are underlined, and the third-best values are double underlined.
MethodCCUIQI
NihonmatsuYokohamaTasmaniaNihonmatsuYokohamaTasmania
Ideal value1.01.0
fast IHS0.7830.9140.9340.7170.9010.686
SAIHS0.830 0.928 ¯ ¯ 0.9360.7430.9050.638
ISAIHS0.8870.9390.949 0.830 ¯ ¯ 0.908 ¯ ¯ 0.708
Tradeoff IHS0.843 0.928 ¯ ¯ 0.9540.8040.909 0.767 ¯ ¯
eFIHS-SRF0.8180.9140.9490.7330.8250.821
PRACS0.8670.8650.8620.8080.7910.508
BT-SFIM 0.879 ¯ ¯ 0.930 0.952 ¯ ¯ 0.851 0.908 ¯ ¯ 0.760
Proposed method0.883 0.928 ¯ ¯ 0.9660.8640.9100.935
Table 5. Erreur relative globale adimensionnelle de synthese (ERGAS) and spectral angle mapper (SAM) quality metrics results.
Table 5. Erreur relative globale adimensionnelle de synthese (ERGAS) and spectral angle mapper (SAM) quality metrics results.
MethodERGASSAM
NihonmatsuYokohamaTasmaniaNihonmatsuYokohamaTasmania
Ideal value0.00.0
fast IHS3.471 3.716 ¯ ¯ 11.4401.8982.3672.912
SAIHS6.6534.77217.7752.1982.3393.932
ISAIHS4.2584.47015.6121.8612.2973.643
Tradeoff IHS 2.910 ¯ ¯ 3.175 8.617 ¯ ¯ 1.783 ¯ ¯ 2.246 ¯ ¯ 2.650 ¯ ¯
eFIHS-SRF6.4077.6474.5781.638 2.251 ¯ ¯ 2.102 ¯ ¯
PRACS2.8704.7588.9321.9933.1925.246
BT-SFIM4.1934.1389.2222.0042.4082.712
Proposed method2.6732.6973.6251.7532.1761.984

Share and Cite

MDPI and ACS Style

Tsukamoto, N.; Sugaya, Y.; Omachi, S. Spectrum Correction Using Modeled Panchromatic Image for Pansharpening. J. Imaging 2020, 6, 20. https://doi.org/10.3390/jimaging6040020

AMA Style

Tsukamoto N, Sugaya Y, Omachi S. Spectrum Correction Using Modeled Panchromatic Image for Pansharpening. Journal of Imaging. 2020; 6(4):20. https://doi.org/10.3390/jimaging6040020

Chicago/Turabian Style

Tsukamoto, Naoko, Yoshihiro Sugaya, and Shinichiro Omachi. 2020. "Spectrum Correction Using Modeled Panchromatic Image for Pansharpening" Journal of Imaging 6, no. 4: 20. https://doi.org/10.3390/jimaging6040020

APA Style

Tsukamoto, N., Sugaya, Y., & Omachi, S. (2020). Spectrum Correction Using Modeled Panchromatic Image for Pansharpening. Journal of Imaging, 6(4), 20. https://doi.org/10.3390/jimaging6040020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop