Next Article in Journal
Research on the Construction of an Efficient and Lightweight Online Detection Method for Tiny Surface Defects through Model Compression and Knowledge Distillation
Previous Article in Journal
Edge-Computing-Enabled Abnormal Activity Recognition for Visual Surveillance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Optimized Inversion Method for Hyperspectral Image Fusion Based on a Hue–Intensity–Saturation, Wavelet, and Trust-Region Conjugate Gradient Method

1
School of Information Science and Technology, Fudan University, Shanghai 200433, China
2
Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(2), 252; https://doi.org/10.3390/electronics13020252
Submission received: 13 November 2023 / Revised: 31 December 2023 / Accepted: 3 January 2024 / Published: 5 January 2024
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)

Abstract

:
In hyperspectral remote sensing, achieving high spatial resolution holds paramount importance for an array of applications, such as environmental monitoring, geographic mapping, and precision agriculture. Nevertheless, conventional hyperspectral images frequently grapple with the issue of restricted spatial resolution. We apply optimized inversion methods to hyperspectral image fusion and present an innovative approach for hyperspectral image fusion which combines the Hue–Intensity–Saturation (HIS) transform, the wavelet transform, and the Trust-Region Conjugate Gradient technique. This amalgamation not only refines spatial precision but also augments spectral faithfulness, which is a pivotal aspect for applications like precise object detection and classification. In the context of our investigation, we conducted a thorough validation of our proposed HIS, Wavelet, and Trust-Region Conjugate Gradient (TRCG-HW) method for image fusion using a comprehensive suite of evaluation metrics. These metrics encompassed the Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Correlation Coefficient (CC), Spectral Angle Mapper (SAM), and Error Relative Global Accuracy Score (ERGAS). The findings incontrovertibly establish TRCG-HW as the preeminent method among those considered. Our study effectively tackles the pressing predicament of low spatial resolution encountered in hyperspectral imaging. This innovative paradigm harbors the potential to revolutionize high-resolution hyperspectral data acquisition, propelling the field of hyperspectral remote sensing forward and efficiently catering to crucial application.

1. Introduction

Hyperspectral imaging (HSI) stands as a versatile technology amalgamating imaging and spectroscopy to concurrently capture both spatial and spectral facets of targets. The resulting data are organized into a three-dimensional cube, comprising two spatial dimensions and a single spectral dimension, collectively forming a hypercube [1]. In the realm of hyperspectral remote sensing, this capability spans multi-band imaging across the visible and infrared spectra, enabling analyses at the molecular and even atomic scales. This surpasses the confines of traditional optical remote sensing, which is reliant solely on spectral data. Recently, hyperspectral remote sensing has showcased its prowess across varied domains, such as environmental monitoring [2], fire detection [3], geographical mapping [4], precision agriculture [5,6], and atmospheric and oceanic observation [7,8]. However, despite its high spectral resolution, hyperspectral imagery grapples with a limited ability to discern fine object details. This limitation stems from capturing information only when objects reflect light at specific wavelengths, resulting in subdued discriminatory power, thereby impacting precise boundary and shape depiction. To alleviate the challenges posed by low spatial resolutions and limited information detection, fusion techniques combining hyperspectral images with panchromatic imagery have been explored. Addressing the issue of hyperspectral image fusion has long been challenging for researchers. Traditional methods often grapple with balancing the richness of hyperspectral data with enhanced spatial resolution. Many existing approaches either compromise spatial resolution to preserve spectral data or forfeit crucial spectral information in favor of enhanced spatial resolution. This enduring trade-off issue poses a significant quandary in hyperspectral image fusion—how to enhance the spatial resolution while retaining the depth of hyperspectral data.
In recent times, there has been a notable surge in the exploration of panchromatic–hyperspectral image fusion techniques. Image fusion methodologies are categorized into three distinct levels based on their processing stages: pixel level [9], feature level [10,11], and decision level [12,13]. Pixel-level fusion focuses on the individual pixel points within two images, offering heightened accuracy and detailed information by directly manipulating the original data. However, pixel-level fusion demands extensive data processing, surpassing the complexity of the feature and decision levels, and requires meticulous alignment prior to fusion.
At the pixel level, color space-based methods, principal component analysis (PCA), and multi-resolution transformation techniques constitute common strategies [14,15]. Color space-based fusion involves transitioning images from the RGB color model to a sequential color system, employing methodologies like the HIS transform [16,17] and Brovey transform [18,19]. The HIS transform adeptly segregates spatial and spectral data. However, the principal component substitution technique’s limitation lies in its operation being confined solely to pixel-level functionality, making it prone to spectral aliasing. This issue causes the loss of intricate details in the fused image due to straightforward pixel-wise substitution. To address this shortcoming, a combined methodology intertwining the differential search algorithm, adaptive regional segmentation, IHS conversion, and RGB band processing was proposed [20]. Principal Component Analysis (PCA) is an image fusion technique that amalgamates multiple images by reducing data dimensions and extracting essential features [21]. However, its application may lead to information loss and the imposition of linear assumptions on intricate relationships, thereby constraining its efficacy.
Numerous researchers have delved into multi-resolution image fusion techniques. Toet introduced contrast pyramids in Gaussian pyramid-based fusion [22], while Burt and Kolczynski derived gradient pyramids from Gaussian pyramids [23,24]. However, the pyramid transformation lacks translational invariance, potentially leading to spurious Gibbs artifacts in the fused images [25]. Chipman proposed fusion using orthogonal wavelets [26], Li presented a digital filter for consistency verification [27], and Liu utilized a controlled pyramid algorithm [28]. Other methods involve Li Zhenhua’s pyramid frame transform [29] and Matsopoulos’ application of morphological pyramids in medical image fusion [30]. While these advanced methods demonstrate progress, they seem to overlook comprehensive spatial consistency, possibly resulting in color and brightness distortions in the fused outcomes. To address this, proposed solutions include guided filters [31] and bilateral filters [32], which effectively tackle spatial consistency concerns and reduce edge artifacts [33]. However, conventional bilateral filters exhibit limitations in effective image smoothing. To overcome this, Chen, B.H. introduced an innovative two-pass bilateral filtering approach for edge-preserving image smoothing, demonstrating exceptional performance [34,35]. Additionally, in the realm of multi-modal image fusion research, Goyal et al. focused on structure awareness and metric analysis, while Dogra and Kumar emphasized the use of guided filtering and multidirectional shearlet transform in medical image fusion [36,37].
The progress in Gaussian pyramid-based methods has not resolved the challenge of losing high-frequency detail during operations. To address this, a technique based on a Laplacian Pyramid direction filter bank was proposed to enhance fusion outcomes [38,39]. These advancements significantly impacted medical image processing, particularly in Laplacian Pyramid-based techniques. Methods employing Laplacian Pyramids and adaptive sparse representation were explored [40,41], notably improving lung cancer diagnosis through CT image fusion and the integration of multimodal medical images. Moreover, in reference [42], fusion methodologies underwent a revolution by combining a Laplacian Pyramid with deep learning, surpassing conventional techniques in image fusion capabilities.
Traditional multi-scale pyramid image fusion methods have undergone significant advancements and applications in the domain of deep learning. The work conducted by Ji, Peng, and Xu exemplifies the practical implementation of deep learning models in conjunction with multi-scale pyramid image enhancement techniques for real-time underwater river crab detection [43]. Notably, there has been a burgeoning interest in leveraging intelligent algorithms like Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs) [44,45]. These methodologies have exhibited adeptness in effectively fusing image information, thereby elevating image quality and augmenting features [46]. Specifically, the adoption of deep learning algorithms rooted in neural network theory has gained substantial prominence in the field of image fusion. Multifocal Image Fusion (MFIF) is capable of generating omnifocal images tailored to visual requirements, and ongoing research endeavors aim to mitigate the defocus spreading effects (DSEs) typically observed around focus/defocus boundaries. In response to DSE challenges, an innovative Generative Adversarial Network, termed MFIF-GAN, was introduced for the specific task of MFIF [47].
The application of the wavelet transform holds a paramount position among multi-resolution image fusion techniques [48]. Revered for its exceptional time–frequency constraints, the wavelet transform facilitates thorough multi-scale image analysis [49]. This influential method adeptly captures both spatial and frequency domain characteristics, offering a comprehensive and intuitive depiction of images. Its capability encompasses the analysis of elements varying in sizes and resolutions within a single image, ensuring a detailed representation. Advanced versions of the wavelet transform, such as the variational binary wavelet transform [50], multi-binary wavelet transform [51], and boosted structure wavelet transform [52], have expanded its potential, opening new avenues for further advancements in image fusion. However, there exists an exigent need for ongoing research to address computational redundancy within the wavelet transform and to devise innovative strategies to enhance computational efficiency. This endeavor is critical for ensuring spectral consistency in hyperspectral images, mitigating spectral aliasing, and preserving intricate high-frequency intensity components, which are often lost in the original HIS transform algorithm.
In this context, our research presents an innovative methodology for hyperspectral image fusion, integrating the HIS transform, wavelet transform, and Trust-Region Conjugate Gradient techniques. This pioneering approach is designed to enhance spatial resolution while preserving the abundance of spectral information. Our study is dedicated to resolving key questions regarding the preservation of spectral nuances in hyperspectral image fusion. We firmly anticipate that our contributions will propel advancements in the field, providing more detailed and accurate images tailored for practical applications.
The main contributions of this work are summarized as follows:
(1)
We present the Trust-Region Conjugate Gradient (TRCG) method, an optimization technique for enhancing image fusion accuracy and efficiency, exploring its principles, mathematics, and practical applications.
(2)
To approach the true value more accurately, our approach employed a two-tiered strategy, with the inner layer guided by the truncated conjugate gradient (TCG) for local optimization and the outer layer using the trust region algorithm (TRA) for global convergence.
(3)
We conducted extensive experiments on widely used datasets, consistently achieving satisfactory performance compared with the latest hyperspectral image fusion methods.

2. Materials and Methods

In this section, we aim to elaborate on the experimental setup of our hyperspectral imaging system and elucidate the methodologies adopted for hyperspectral image fusion. We provide an extensive explanation of the core principles and practical applications of crucial techniques, notably the Hue–Intensity–Saturation (HIS) transform, wavelet transform, and the Trust-Region Conjugate Gradient method. Additionally, we delve into the empirical data, the nuances of our experimental design, and the relevant evaluation metrics employed in our scholarly exploration.

2.1. Experimental Design of Hyperspectral Imaging System

We provide an overview of the hyperspectral imaging system employed in our study and detail our designed experimental procedure. Our experimental design is tailored to acquire high-quality hyperspectral image data, which serve as the foundation for subsequent processing and analysis.

2.1.1. Hyperspectral Imaging System

In the domain of contemporary hyperspectral imaging (HSI), the acquisition and processing of high-quality data hold paramount significance. This study delves into an innovative experimental design tailored to augment both data quality and subsequent fusion processes.
Figure 1 presents a comprehensive portrayal of our HSI system, structured into three distinct modules: a narrow-band light generator, an imaging section, and a control system. This schematic delineates the design and operational principles governing these modules.
The narrow-band light generator incorporates a Xenon source (55 W, 6000 K) to produce a broad spectrum of light. This source interfaces with several optical elements, including lenses and a reflective ruled diffraction grating (1800 lines/mm, angular dispersion rate 1.8 mrad/nm), inducing chromatic dispersion. An imaging lens (Φ = 38 mm, f = 200 mm) converges parallel rays. The optical components, including lenses and a small-aperture light stop (Φ = 1 mm), shape the light rays, ensuring parallel incidence onto the grating. The grating diffracts the light, focusing it at the imaging lens’s focal plane, generating a chromatic band known as the first-order diffraction spectrum. A small aperture at the focal point transmits narrow-band light at a specific wavelength. The grating’s rotation, managed by a rotating platform, adjusts emitted light wavelengths by varying the incident light angle. The lighting section and the narrow-band light generator connect solely via an optical fiber (Φ = 4 mm).
The selected narrow-band light illuminates tissue samples using a complementary metal oxide semiconductor (CMOS) camera featuring a 1280 × 1024 array and 5.2 μm square pixels and operating at 15 frames per second. The control system, comprising computer hardware and software, commands two modules: the MCS-51 microcontroller and an electromotor, controlling the rotation platform to adjust the narrow-band light’s center wavelength. The CMOS camera captures images at various wavelengths. The control system synchronizes wavelength switching and image acquisition, storing raw data as a hypercube.
This modular design offers potential integration into modern consumer imaging products. For this study, the HSI system was installed on a stereomicroscope XTZ-E, boasting magnification ranging from 7× to 45× (Shanghai Optical Instrument Factory, Shanghai, China).

2.1.2. Experimental Procedure

Our approach involved meticulous data collection using cutting-edge hyperspectral imaging systems, followed by an extensive preprocessing stage. This preprocessing included crucial tasks such as noise reduction, radiometric correction, and meticulous image registration.
The spectroscopic measurement of monochromatic light generated by an active monochromatic hyperspectral imaging system holds paramount importance in ensuring the precision and efficacy of the imaging device. This meticulous process, especially concerning the RGB spectral bands, is vividly depicted in Figure 2.
The top row exhibits the spectral power density curve, while the second and third rows showcase the CIE 1931 and CIE 1964 chromaticity diagrams, respectively. Detailed measurements, facilitated by the UPRtek MK350S spectrometer(UPRtek, New Taipei City, China), were conducted to ensure precise wavelength control and spectral fidelity.
Analysis of the spectral power density curve, CIE 1931 chromaticity diagram, and CIE 1964 chromaticity diagram allows us to evaluate the monochromatic performance of the spectrometer. The spectral power density curve illustrates the relative intensity of light across various wavelengths, displaying narrow and sharp peaks that denote the spectrometer’s exceptional monochromaticity. It effectively segregates light of different wavelengths. The CIE 1931 and CIE 1964 chromaticity diagrams indicate the positions of light at different wavelengths within the color space. The accurate representation of chromaticity coordinates in these diagrams confirms the spectrometer’s commendable monochromatic performance.
The spectrometer’s monochromatic performance is pivotal for its functionality, as it directly influences its capacity to precisely resolve and measure light of diverse wavelengths. Enhanced monochromaticity significantly improves the spectrometer’s accuracy in color measurement, spectral analysis, and other applications by allowing it to precisely differentiate and measure light wavelengths. This feature not only delivers high-resolution monochromatic imaging but also demonstrates exceptional wavelength stability and the capability to precisely select spectral bands. Our meticulous spectroscopic measurement and spectral analysis of monochromatic light, generated by the active monochromatic hyperspectral imaging system, not only yield top-tier data but also unleash the instrument’s full potential. This comprehensive approach not only enriches our understanding of spectral characteristics but also furnishes reliable spectral support across diverse domains, ultimately propelling advancements in research and the seamless integration of this technology into practical applications.
To exhibit the effectiveness of our preprocessing procedure, Figure 3 showcases preprocessed images captured using an RGB camera. These images encompass various lighting conditions, including (a) red lighting, (b) green lighting, (c) blue lighting, and (d) synthesized color hyperspectral images. Additionally, (e) grayscale images captured under full-spectrum illumination are included for comprehensive evaluation. These images unequivocally demonstrate the success of our preprocessing method in elevating the overall quality and consistency of hyperspectral data. The application of these methods significantly enhances our hyperspectral data, ensuring a robust and accurate foundation for the subsequent fusion process.

2.2. HIS Transformation

Section 2.2 delves into the HIS color model and its role in hyperspectral image fusion. The exploration commences with an introduction to the HIS color model (Section 2.2.1), followed by an elucidation of its application in hyperspectral image fusion (Section 2.2.2). These sections aim to offer comprehensive insight into the utilization of the HIS color model to augment image quality and information fusion.

2.2.1. HIS Color Model

HSI means intensity, saturation, and hue. Based on the RGB color system, the RGB color image can be decomposed into R, G, and B channels. The R, G, and B can be transformed into the H, I, and S by mathematical transformation, which completes the HIS transformation of RGB color images. The majority of images we encounter in our daily lives are colored images, although images are fundamentally two-dimensional data with pixels typically represented as m × n. Such images are referred to as grayscale images, commonly recognized as black and white images. However, the representation of colored images necessitates an understanding of colorimetry. The CIE 1931 RGB color space is the most prevalent standard, wherein the combination of the three primary colors R (red), G (green), and B (blue) is determined by their respective tristimulus values to create a colored image. Consequently, RGB color images can be decomposed into three separate images, corresponding to the R, G, and B channels.
While the RGB color space is employed for color mixing and computation, the perception of an object’s color in daily life requires a color perception system, known as a color order system. The Munsell color system, an example of a color order system, defines three parameters—brightness, hue, and chroma (saturation)—to characterize color. When observing objects, the “Munsell Color Chart” can be used for comparison, enabling the confirmation of an object’s color. Analogously to the Munsell color system, Munsell introduced the HIS color model, which comprises I (intensity/luminance), S (saturation), and H (hue). Based on the RGB color space, RGB color images can be decomposed into their R, G, and B channels, and mathematical transformations can be applied to convert the R, G, and B channel components into H, I, and S channel components. This process constitutes the HIS transformation of an RGB color image. In the context of the HIS transformation, intensity I conveys spatial information, while H and S represent spectral information, thereby achieving the separation of spectral and spatial information.

2.2.2. Hyperspectral Image Fusion Based on HIS Transform

In the process of fusing panchromatic and hyperspectral images based on the HIS transform, we began by performing the HIS transform on RGB color hyperspectral images. In Appendix A, we present the equations characterizing the linear RGB to HIS transformation. Following this transformation, the full-color grayscale image was preprocessed and introduced as a new component referred to as ‘I’ (intensity) within the color sequence system. This ‘I’ component was then fused with the ‘H’ (hue) and ‘S’ (saturation) components of the hyperspectral image. Subsequently, the HIS transformation was reversed, returning the data to the RGB color space and yielding the final RGB color fusion image.

2.3. Wavelet Transform (WT)

The wavelet transform represents a method capable of concurrently considering both the spatial and frequency domain attributes of an image. By decomposing the image into various frequency components across multiple scales, it enables a multi-scale analysis of the image. In terms of image fusion, the wavelet transform contributes to enhancing the spatial precision of the image. It achieves this by analyzing and integrating image details at different scales, thereby facilitating multi-scale processing of image details.
In Section 2.3, we delve into wavelet transform and its crucial application in image fusion. Firstly, we introduce wavelet transform and its mathematical principles (Section 2.3.1), elucidating its underlying concepts and mathematical foundations. Subsequently, we provide a detailed discussion on the application of wavelet transform in image processing (Section 2.3.2), encompassing specific transformation methods and processes. Finally, we explore the role of wavelet transform in image fusion (Section 2.3.3), offering an in-depth analysis of its pivotal contribution to the fusion process.

2.3.1. Wavelets’ Mathematical Principles

The mathematical representation of the Continuous Wavelet Transform (CWT) involves convolving a function, often referred to as the mother wavelet ψ ( t ) , with the signal f ( t ) across varying scales and translations. The CWT of a signal f ( t ) with respect to a mother wavelet ψ ( t ) at a scale a and translation τ   is given by
W T ( a , τ ) = 1 a f ( t ) ψ ( t τ a ) d t ,
where W T ( a ,   τ ) represents the Continuous Wavelet Transform of f ( t ) at scale a and translation τ , ψ ( t ) denotes the mother wavelet, a is the scale variable that can control the scaling of the wavelet basis, τ is the translation quantity that controls the translation of the wavelet basis, and a and τ correspond to the frequency inverse and time, respectively.
The considerable computational intricacy of Continuous Wavelet Transform (CWT) and its constrained implementation in discrete systems have restricted its broad utility in practical engineering and data processing. To surmount these constraints and furnish more effective analytical tools, the Discrete Wavelet Transform (DWT) was introduced and extensively embraced. DWT, a discrete counterpart of CWT, represents a technique for disassembling a signal into various components of diverse scales and frequencies. However, it employs an alternative approach to achieve this disassembly. While the conventional CWT entails convolving the signal with continuous wavelets, DWT utilizes sampling and filter bank methodologies, rendering it more suitable, particularly in the domain of image processing.
The operation of DWT on a discrete signal f ( x ) involves signal decomposition using a low-pass filter φ j 0 , k ( x ) and a high-pass filter ψ j , k ( x ) , followed by downsampling. In a single-level DWT, the signal f ( x ) can be decomposed into approximation coefficients W φ ( j 0 , k ) (representing low-frequency components) and detail coefficients W ψ ( j , k ) (representing high-frequency components):
W φ ( j 0 , k ) = 1 M x f ( x ) φ j 0 , k ( x ) W ψ ( j , k ) = 1 M x f ( x ) ψ j , k ( x ) ,
where 1 M is the normalization factor, usually j 0 = 0 ; M = 2 J ; x = 0 , 1 , 2 , , M 1 ; j = 0 , 1 , 2 , , J 1 ; k = 0 , 1 , 2 , , 2 j 1 ; W φ ( j 0 , k ) denotes the approximation coefficients; W ψ ( j , k ) represents the detail coefficients (representing high-frequency components); and φ j 0 , k ( x ) and ψ j , k ( x ) are, respectively, the low-pass and high-pass filters.

2.3.2. Wavelet Transform of Image

When dealing with two-dimensional data such as images, the extension of one-dimensional DWT to a two-dimensional domain becomes imperative. This extension is known as the Two-Dimensional Discrete Wavelet Transform (2D DWT). 2D DWT of size M × N image function f(x,y) is as follows:
W φ ( j 0 , m , n ) = 1 M N x = 0 M 1 y = 0 N 1 f ( x , y ) φ j 0 , m , n ( x , y ) W ψ i ( j , m , n ) = 1 M N x = 0 M 1 y = 0 N 1 f ( x , y ) ψ j , m , n i ( x , y ) ,
where the superscript i denotes H (horizontal direction), V (vertical direction), D (diagonal direction). The index j 0 signifies any starting scale. The approximation coefficient W φ ( j 0 , m , n ) defines the approximation value of f ( x , y ) at scale j 0 . The detail coefficient W ψ i ( j , m , n ) adds horizontal, vertical, and diagonal details for j j 0 . Approximation coefficients W φ ( j 0 , m , n ) represent the low-frequency information of the image, capturing its overall characteristics. Horizontal detail coefficients W ψ H ( j , m , n ) encompass high-frequency information in the horizontal direction of the image. Vertical detail coefficients W ψ V ( j , m , n ) encompass high-frequency information in the vertical direction of the image. Diagonal detail coefficients W ψ D ( j , m , n ) contain high-frequency information in the diagonal direction of the image. It is often set that j 0 = 0 ; N = M = 2 J ; j = 0 , 1 , 2 , , J 1 ; and m , n = 0 , 1 , 2 , , 2 j 1 . φ j 0 , m , n ( x , y ) and ψ j , m , n ( x , y ) are, respectively, the low-pass and high-pass filters.
In our preceding discussion regarding the 2D DWT, we presented significant variables, including the original image, denoted as f ( x ,   y ) , accompanied by its corresponding approximation coefficients W φ ( j 0 , m , n ) and detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) . Now, transitioning to the process of reconstructing the original image from these coefficients, we delve into the utilization of the Two-Dimensional Inverse Discrete Wavelet Transform (2D IDWT). This inverse transformation method harmonizes and combines W φ ( j 0 , m , n ) , W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) coefficients acquired from the 2D DWT to effectively regenerate the original two-dimensional image f ( x ,   y ) :
f x , y = 1 M N m n W φ j 0 , m , n φ j 0 , m , n x , y + 1 M N i = H , V , D j = j 0 + m n W ψ i j , m , n ψ j , m , n i x , y .
The Two-Dimensional Discrete Wavelet Transform (2D DWT) can be implemented using filtering and subsampling techniques. Initially, the 1D DWT is applied to each row of the image. Subsequently, the obtained results undergo another one-dimensional DWT in the column direction. In practical implementation, especially in computer programming, there might be a preference for filtering columns first and then rows, as this aligns better with the computational handling of image data. This approach can enhance efficiency or fulfill hardware requirements.
Figure 4 depicts this process. The image f ( x ,   y ) serves as the input to W φ ( j + 1 , m , n ) , undergoing convolution with h ψ ( n ) and h φ ( n ) along the columns separately, followed by subsampling. The high-pass components depict the image’s vertical directional details, while the low-pass approximate components portray low-frequency vertical information. This process yields two sub-images, each exhibiting a halving of their resolution by a factor of 2. Subsequently, the resulting two sub-images are then subjected to filtering and subsampling along the rows, generating four quarter-sized images denoted as W φ ( j , m , n ) , W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) . Approximation coefficients W φ ( j , m , n ) encapsulate the overall image characteristics, while detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) represent detailed information in the horizontal, vertical, and diagonal directions, respectively.
Figure 5 illustrates the image f ( x ,   y ) serving as the input for W φ ( j + 1 , m , n ) at scale ‘ j + 1 . Through a 2D DWT process, it generates the approximation coefficients W φ ( j , m , n ) and detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) at scale ‘ j ’. Subsequently, W φ ( j , m , n ) is utilized as input for another 2D DWT, yielding the approximation coefficients W φ ( j 1 , m , n ) and detail coefficients W ψ H ( j 1 , m , n ) , W ψ V ( j 1 , m , n ) , and W ψ D ( j 1 , m , n ) at scale ‘ j 1 .

2.3.3. Application of Wavelets in Image Fusion

The prior discourse delineated the fundamental principles of wavelet transforms in image processing, elucidating their role in feature extraction and analysis. Concerning image fusion, wavelet transforms amalgamate multiple images or diverse image features to generate a composite image enriched with comprehensive information. Leveraging the multi-scale nature of wavelet transforms aids in capturing intricate details at varying scales, better preserving vital features that might be challenging to depict entirely within individual images.
In practical application, initiating the process involves a 2D DWT performed on each original image. Post decomposition of each image based on designated wavelet types and decomposition levels, fusion processing is carried out on the different decomposition layers. Distinct fusion operators can be applied to the various frequency components in each decomposition layer, culminating in a fused wavelet pyramid. Ultimately, the fused wavelet pyramid undergoes reconstruction via the 2D IDWT to yield the fused image.
The wavelet transform decomposes images into frequency components at different scales, encompassing both low-frequency information (related to the overall structure and general features of the image) and high-frequency information (related to the finer details and texture of the image). By integrating information from various frequencies, particularly the high-frequency details, it is possible to retain the subtle features of the image. This fusion process can employ methods such as weighting, thresholding, or other suitable approaches to amalgamate details from different scales and orientations, thereby maintaining or enhancing spatial precision during image fusion.
Moreover, wavelet transforms aid in identifying essential image features, such as edges, textures, and more. Prioritizing the preservation of these crucial features during fusion notably enhances spatial accuracy within the image. Reasonable utilization of high-frequency information during image merging effectively amplifies image details. This can be achieved through the selection of specific segments of high-frequency components or employing a fusion strategy focused on detail enhancement, thus contributing to heightened spatial precision during the image fusion process.

2.4. HIS, Wavelet, and Trust-Region Conjugate Gradient (TRCG-HW)

The TRCG-HW image fusion methodology represents a comprehensive approach aimed at achieving superior performance in multispectral image fusion. This method seamlessly integrates the HIS (Hue, Intensity, Saturation) transformation, wavelet transformation, and trust region algorithms to optimize the image fusion process. The HIS transformation plays a pivotal role in preserving spectral information, while the wavelet transformation significantly enhances spatial accuracy. The incorporation of trust region algorithms orchestrates and optimizes the entire process cohesively. The primary objective is to procure high-quality fused images while excelling across various performance metrics.
The trust region methodology functions as an optimization strategy that intricately balances local and global models by confining a specific space around an iteration point to simulate the objective function. In contrast, the conjugate gradient approach is a dedicated optimization methodology focused on minimizing the objective function by reducing residuals during step size and direction adjustments. In the realm of image fusion, these methodologies wield significant influence, refining the fusion algorithm profoundly. The trust-region technique adeptly oversees the optimization process at each stage, ensuring a gradual refinement of fusion outcomes within a localized model. On the other hand, the conjugate gradient technique operates as an iterative process, pinpointing the most optimal direction at each step, enabling rapid convergence towards a globally optimal solution. These methodologies collaborate seamlessly to orchestrate and refine the entire image fusion process, striving to preserve image characteristics while attaining exceptional quality in the resultant fused image.

2.4.1. Mathematical Principles of TRCG

In this subsection, we delve into the mathematical foundations that underpin the TRCG method. We delve into key notions such as conjugate gradients, trust regions, and the rational fusion of these concepts to forge a pathway to efficient optimization. Consider an unconstrained nonlinear optimization problem:
m i n f ( x ) .
Using the trust region method to solve (5), we first give the current trust region trial step size Δ c (conventionally called the trust region radius) and then solve the quadratic subproblem of an approximation of problem (5):
m i n ψ ( x c + ξ ) = f ( x c ) + ( g ( x c ) , ξ ) + 1 2 ( H c ξ , ξ ) ,
where x c represents the reference point; ξ denotes a small increment or offset. f ( x c ) represents a function of x c , g ( x c ) might denote the gradient, and H c represents the Hessian matrix at x c . ( g ( x c ) , ξ ) represents the inner product between the vector g ( x c ) and ξ . ( H c ξ , ξ ) denotes the quadratic form of the matrix H c applied to the vector ξ .
s . t . ξ Δ c ,
where Δ c is trust region trial step. It describes the extent to which we can trust the quadratic approximation model.
Next, we consider using the trust region method to solve discrete operator equations:
κ f = h ,
where κ R m 2 × n 2 is a PSP matrix, f R n 2 is the input to be sought, and h R m 2 is the measured output.
First, we form the following unconstrained least squares problem:
m i n Μ [ f ] = 1 2 κ f h 2 .
The gradient and Hessian matrix of the functional M[f] can be explicitly calculated as g r a d ( M l f ) ) = κ * κ f κ * h , H e s s ( M [ f ] ) = κ * κ .
To solve with the trust region algorithm (TRA) (9), one needs to solve the following trust region subproblems (TRSs):
m i n   ϕ ( s ) = ( M [ f ] ,   s ) + 1 2 ( H e s s ( M [ f ] ) s ,   s ) ,
s . t .   s < Δ ,
In each step of the trust region iteration, the solutions of subproblems (TRSs) (10) and (11) do not have to be too precise, which can be achieved by using the truncated conjugate gradient (TCG) method. The point list generated by solving (10) is as follows:
s k + 1 = s k + α k d k ,
d k + 1 = g k + β k d k ,
g k = ϕ ( s k ) = H e s s ( M [ f ] ) s k + M [ f ] ,   α k = g k T d k / d k T H e s s ( M [ f ] ) d k ,   β k = g k + 1 2 / g k 2 ,
where g k   refers to the gradient vector at iteration step k, g k + 1   refers to the gradient vector at iteration step k + 1, g k denotes the Euclidean norm of vector g k , and g k + 1 denotes the Euclidean norm of vector g k + 1 . The initial value is
s 0 = 0 , d 0 = g 0 = M [ f ] .
If the current iteration s k + α k d k is in the trust domain, we accept it and transfer it to the next trust domain iterative process; if d k T H e s s ( M [ f ] ) d k 0 or s k + α k d k runs outside the trust domain, we take the longest value in the trust domain step d k and terminate the iterative process.

2.4.2. Methodology

The method described in Figure 6 begins by applying the HIS transformation to the RGB color hyperspectral image (HSI). Simultaneously, the panchromatic (PAN) image undergoes preprocessing and is incorporated into the color sequence system as an “I” (intensity) component. Next, the wavelet transformation is applied to the “I” components from both the panchromatic image and the hyperspectral image for improved fusion. Through the optimization of wavelet coefficients using TRCG, a new “I” component is obtained. This new “I” component is merged with the “H” (hue) and “S” (saturation) components of the hyperspectral image. Eventually, by reversing the HIS transformation, the data are restored to the RGB color space, producing the final RGB color fusion image. Several meticulously designed steps are employed in merging multispectral images to achieve superior performance. These steps involve defining evaluation metrics and an objective function, performing the optimization process using TRCG, and conducting evaluation and adjustment stages.
Prior to optimization, evaluation metrics and an objective function are defined. The evaluation metrics encompass various aspects of the image, including structural similarity, signal-to-noise ratio, spectral information, rate of change, and correlation. These evaluation metrics are integrated into an objective function that comprehensively evaluates the quality of the fused image. For normalization, Min–Max standardization was utilized to scale each metric within a range of 0 to 1. Metrics such as SSIM, PSNR, and SAM, with higher values indicating better performance, were used directly after normalization. However, ERGAS and CC values were computed by subtracting their normalized scores from 1, aiming for lower values to signify superior performance.
The objective function is represented as follows:
Objective Function = w1 × SSIM + w2 × P SNR + w3 × SAM + w4 × ERGAS + w5 × CC,
where w1, w2, w3, w4, and w5 symbolize the respective weights attributed to individual evaluation metrics. These weights were assigned to determine the relative significance of each metric within the objective function. Assuming equal impact from all metrics, assigning uniform weights of 0.2 to the normalized SSIM, SAM, ERGAS, CC, and PSNR values facilitated the formation of a unified objective function. This method ensures an equivalent contribution from each metric to the overall objective function. It is crucial to mention that altering these weights might be more suitable if particular metrics significantly influence the objective function in varying degrees of enhancement or degradation. In such instances, adjusting the weights based on the specific influence of each metric on the objective function could be more appropriate. The holistic assessment offers a comprehensive evaluation of image quality, minimizing potential biases inherent in individual metrics. Integrating objective functions not only saves time and energy but also reduces the potential misguidance of a single indicator, enhancing the reliability of decision making.
The TRCG method employs a two-tier strategy, comprising an inner layer and an outer layer, to ensure efficient and accurate image fusion. The inner layer utilizes truncated conjugate gradient (TCG) techniques, which concentrate on localized optimization within specific regions. By computing gradient information and preserving the most significant components, TCG refines the fusion process on a local scale. It achieves this through iterative updates of wavelet coefficients, thereby enhancing fusion quality within these localized areas. On the other hand, the outer layer operates using the trust region algorithm (TRA) to oversee global convergence. TRA calculates the gradient of the objective function and supervises the entire optimization process, ensuring effective convergence across the image. It collaborates with the inner layer and dynamically adjusts the trust region radius and step size to strike a balance between efficiency and accuracy throughout the fusion process. By integrating these two layers—local optimization via TCG and global convergence management through TRA—the TRCG method aims to achieve a synergy that balances efficiency and accuracy, ultimately enhancing the overall quality of the fused image.

3. Experiments and Results

In the experimental section, we conducted a series of experiments to evaluate the effectiveness of our TRCG-HW-based image fusion method. Additionally, we compared our TRCG-HW method against three state-of-the-art image fusion methods, including Principal Component Analysis (PCA), Hue–Intensity–Saturation (HIS), and wavelet transform (WT). The settings for each method were fine-tuned to produce optimal results based on reference recommendations. The order 4 Daubechies wavelet was selected for our hyperspectral image fusion study for its adeptness in preserving spectral details while efficiently capturing spatial intricacies across diverse scales. Its harmonious balance between the frequency and spatial domains proved highly beneficial in handling hyperspectral data, ensuring the conservation of vital spectral information during fusion.

3.1. Data Collection and Preparation

Experimental Dataset: We utilized hyperspectral images captured by the hyperspectral imaging system described in Section 2.1 for our fusion experiments. The collected high-spectral image data underwent cropping, registration, and normalization processes. Following these procedures, both the panchromatic and hyperspectral images were standardized to dimensions of 512 × 512 pixels. These images encompassed 31 spectral bands, with each band covering a 10 nm wavelength interval spanning the visible spectrum from 400 to 700 nm.
Publicly Available IKONOS-2 Remote Sensing Image Dataset: IKONOS-2, a high-resolution commercial satellite designed for Earth remote sensing and operated by the U.S. aerospace company DigitalGlobe, served as a vital data source. This dataset stands out due to its exceptional spatial resolutions, with a remarkable 1 m resolution for panchromatic imagery and a 4 m resolution for multispectral imagery. The IKONOS-2 satellite incorporates multiple spectral bands, typically encompassing blue, green, red, and near-infrared bands for multispectral imagery, in addition to a panchromatic band. IKONOS’s multispectral images (MSs) consist of four bands, with band settings including a panchromatic band spanning from 450 to 900 nm, a blue band ranging from 450 to 530 nm, a green band from 520 to 610 nm, and a red band from 640 to 720 nm. These bands cater to a wide array of Earth observation tasks. For our research, we meticulously selected 200 patches of IKONOS images from this dataset. During the preprocessing phase, we took great care to eliminate spectral bands that were impacted by water vapor absorption, ensuring the quality and accuracy of the data. These datasets provide a wealth of valuable information and diversity, rendering them suitable for a wide range of Earth observation tasks. The IKONOS-2 dataset is publicly accessible and can be downloaded from http://earthexplorer.usgs.gov (accessed on 1 November 2023).

3.2. Evaluation Metrics

We utilized five key evaluation indices to quantitatively assess the quality of the fusion results. These metrics are as follows:
(1) Peak Signal-to-Noise Ratio (PSNR) [53]: The PSNR measures the quality of image reconstruction by comparing the fused image to the original data. A higher PSNR value indicates superior image quality, with 30 or above typically considered excellent.
P S N R ( Z , Z ^ ) = 1 S i = 1 S P S N R ( Z i , Z ^ i ) ,
where Z ∈ RW×H×S and Z ^ ∈ RW×H×S are the reference and fused images. Z i ∈ RW×H and Z ^ i ∈ RW×H are the ith spectral bands of Z and Z ^ .
(2) Structural Similarity Index (SSIM) [54]: The SSIM evaluates the preservation of structural information in the fused image. A higher SSIM score signifies a better ability to retain fine details and structures, with a value of one indicating the best similarity.
S S I M ( Z , Z ^ ) = 1 S i = 1 S S S I M ( Z i , Z ^ i ) ,
(3) Correlation Coefficient (CC) [55]: The CC calculates the correlation between the pixel values of the fused image and the reference image. A high CC score indicates a strong correlation, suggesting that the fused image closely matches the original. A CC value closer to one is generally desired, indicating a strong linear relationship and higher image fusion quality.
C C ( Z , Z ^ ) = i = 1 N ( A i A ¯ ) ( B i B ¯ ) i = 1 N ( A i A ¯ ) 2 i = 1 N ( B i B ¯ ) 2 ,
where A i and B i   represent the pixel values of the fused image and the reference image, while A ¯   and B ¯ denote their respective mean values. N represents the total number of pixels.
(4) Spectral Angle Mapper (SAM) [56]: An SAM quantifies the spectral similarity between the fused image and the reference data. Lower SAM values indicate a higher degree of spectral similarity, which is essential in remote sensing applications.
S A M ( Z , Z ^ ) = 1 W H i = 1 W H a r c c o s z ^ i T z i z ^ i 2 z i 2 ,
where z ^ i and z i are pixels in Z ^ and Z .
(5) Error Relative Global Accuracy Score (ERGAS) [57]: The ERGAS evaluates the level of spectral distortion in the fused image. Lower ERGAS scores indicate reduced spectral distortions, emphasizing the quality of the spectral information in the fused result; the best value is zero.
E R C A S ( Z , Z ^ ) = 100 q 1 S i = 1 S M S E ( Z i Z ^ i ) μ Z ^ i 2 ,
where   q   denotes the spatial downsampling factor, μ Z ^ i 2   denotes the mean value of Z ^ i , and M S E ( Z i Z ^ i ) represents the mean square error between Z i and Z ^ i .

3.3. Results

In this section, we present the experimental results to demonstrate the performance of the proposed method. The results are evaluated using metrics and error maps. Our method consistently outperformed the comparison methods on the two datasets, achieving significantly better results. This underscores the effectiveness of our method in hyperspectral image fusion.

3.3.1. Results of Experimental Dataset

The average quantitative results are summarized in Table 1. In Table 1, TRCG-HW exhibits outstanding performance, with excellent scores in all evaluation metrics. It achieves a high PSNR of up to 29.59, indicating its remarkable capabilities in image quality reconstruction. Furthermore, it excels in SSIM and CC, with scores of 0.981 and 0.976, respectively, providing strong evidence of its ability to preserve structural information and maintain a high correlation with the original data. Additionally, TRCG-HW demonstrates remarkable spectral similarity, with an SAM score of 0 and an ERGAS score of 0.338, signifying effective preservation of spectral fidelity. This implies that TRCG-HW not only effectively retains image details but also significantly reduces noise levels.
We also provide qualitative results in the form of reconstruction error maps, as shown in Figure 7. These error maps illustrate that our method achieved minimal reconstruction errors, further validating its ability to better preserve spatial and spectral information. Based on the error maps of the reconstructed images, where colors transition from blue to red, indicating an increasing error, we have three rows representing three different sets of observations (abcd columns) using four different fusion methods. First, let us assess error uniformity. In the first row, all four methods exhibit some red areas, with the HIS method having predominantly green regions, indicating higher errors. The WT and PCA methods exhibit some red noise in their green areas. In contrast, the TRCG-HW method performed exceptionally well, with most of its area in deep blue. It demonstrated a uniform error distribution with minimal and concentrated red regions, making it convenient for subsequent noise reduction. Moving on to the second row, the WT method stands out as having the most red regions, followed by the HIS and PCA methods. The TRCG-HW method excelled once again, with most of its area in deep blue, a smaller green portion, and an evenly distributed error pattern. Now, looking at the third row, the PCA method shows the most red regions, followed by the HIS and WT methods. The TRCG-HW method performed remarkably well, with almost the entire area in deep blue and a uniform distribution of errors.
Let us also consider the maximum error values. In the first row, the HIS fusion method had the highest error with a maximum value of 0.2. The WT method performed relatively poorly, while the PCA and TRCG-HW methods exhibited better results, with TRCG-HW having smaller errors. In the second row, the WT fusion method had the highest error, reaching a maximum value of 0.25. The HIS and PCA methods did not perform as effectively, whereas the TRCG-HW method had the lowest error with a value of 0.1. In the third row, the PCA fusion method had the highest error, with a maximum value of 0.16, while the WT and HIS methods exhibited poorer performances. The TRCG-HW method once again achieved the lowest error, with a maximum value of 0.12.
To summarize, the PCA method had the highest errors in the third set, while the HIS method consistently produced high errors in all three sets. The WT method resulted in high errors in the first and second sets but lower errors in the third set. On the other hand, the TRCG-HW method consistently yielded low errors in all three sets. Overall, among these three sets of images from different observations, the TRCG-HW method demonstrated the best performance in terms of error reduction.

3.3.2. Results of IKONOS-2 Dataset

Results for the quality metrics for the IKONOS-2 dataset are presented in Table 2. In Table 2, for the fusion image quality metrics, we observe the following results: The TRCG-HW method achieved a significant PSNR score of 32.92, indicating a substantial improvement in image quality. It excelled in SSIM and CC, with respective scores of 0.974 and 0.984, underlining its ability to preserve structural information and maintain a strong correlation with the original data. Furthermore, TRCG-HW demonstrated exceptional spectral similarity, with an SAM score of 0.001 and an ERGAS score of 0.270. These low values highlight its effectiveness in minimizing spectral distortions and enhancing the overall spectral accuracy, which is a critical aspect in remote sensing applications.
Qualitative results: reconstruction error maps are also provided in Figure 8, where colors transitioned from blue to red with increasing errors. The error maps reveal that our method achieved minimal reconstruction errors, further confirming its ability to better preserve spatial and spectral information. These experiments encompassed five different locations (represented by rows) and involved four distinct fusion methods (designated by columns: a, b, c, d). First, let us analyze the uniformity of errors. In the first row, the PCA method exhibited the highest proportion of red regions, followed by the HIS method. However, the TRCG-HW method demonstrated superior performance, with the majority of its regions in deep blue and a smaller portion in light green. The error distribution was relatively uniform. Moving on to the second row, the HIS method had the most red areas, followed by the WT and PCA methods. Once again, the TRCG-HW method excelled, with most areas in deep blue, a smaller portion in green, and a uniform distribution. In the third row, the PCA method had the most red regions, followed by the WT method. The HIS and TRCG-HW methods displayed exceptional performances, with almost all areas in deep blue and a uniform distribution. The fourth row showed that the PCA method had the highest proportion of red regions, followed by the HIS and WT methods, with a significant green area. However, the TRCG-HW method performed exceptionally well, with the entire area in deep blue and an even distribution of errors. Finally, in the fifth row, the PCA and HIS methods had the most red regions, while the WT method had a considerable green area. Once again, the TRCG-HW method stood out, with almost the entire area in deep blue and a uniform distribution.
Now, let us assess the maximum error values. In the first row, the PCA fusion method had the highest maximum error value, reaching 0.4. The HIS method performed relatively poorly, while the WT and TRCG-HW methods demonstrated superior results with a maximum error value of 0.2. In the second row, the HIS fusion method had the highest maximum error with a value of 0.25. The WT method exhibited subpar performance, whereas the PCA and TRCG-HW methods had lower maximum errors at 0.05. In the third row, the PCA fusion method once again recorded the highest maximum error of 0.9. The WT method showed suboptimal performance, while the HIS and TRCG-HW methods excelled with a maximum error of 0.3. In the fourth row, the PCA fusion method had the highest maximum error at 0.1, while the HIS and WT methods demonstrated better results with a maximum error of 0.06. The TRCG-HW method showcased the best performance with a maximum error of 0.02. Finally, in the fifth row, both the PCA and HIS fusion methods exhibited the highest maximum errors, at 0.9, while the WT method had a substantial maximum error of 0.5. In contrast, the TRCG-HW method outperformed the others, with a maximum error value of 0.2.
In conclusion, the PCA method resulted in the highest errors in sets 1, 3, and 4, but lower errors in sets 2 and 5. The HIS method consistently yielded higher errors in sets 1, 2, and 5, while achieving lower errors in set 3. The WT method led to higher errors in sets 2, 3, and 4, with lower errors in sets 1 and 5. On the other hand, the TRCG-HW method consistently exhibited lower errors across all five sets. Overall, across these five different locations, the TRCG-HW method demonstrated the most stable and lowest error performance, surpassing the other fusion methods.

4. Discussion

The TRCG-HW technique has undergone thorough assessments, employing both visual inspections and quantitative analyses on simulated and real datasets. Its consistent excellence in producing high-quality images, maintaining both structural and spectral information, and minimizing reconstruction errors, has been convincingly showcased when compared to other methods for hyperspectral image fusion:
(1)
High Image-Quality Fidelity: The TRCG-HW method achieved outstanding scores in PSNR evaluations, indicating its ability to reconstruct images with high quality. It outperformed other methods in terms of image fidelity.
(2)
Preservation of Structural Information: The TRCG-HW method obtained significant scores in SSIM and CC evaluations, demonstrating its excellent performance in retaining structural information and maintaining a high level of correlation with the original data.
(3)
High Spectral Fidelity: SAM and ERGAS scores for the TRCG-HW method indicate its effectiveness in preserving spectral fidelity, allowing images to better reflect the spectral characteristics of objects.
(4)
Minimal Reconstruction Errors: Qualitative results in the form of error maps illustrate that the TRCG-HW method achieved the fewest reconstruction errors, further substantiating its outstanding performance in preserving both spatial and spectral information.
In comparison to previous research, the novelty of integrating HIS, wavelet, and TRCG techniques into a unified framework is a significant contribution. HIS transformation aids in spectral information preservation, while wavelet transformation enhances spatial accuracy. The unique feature of the TRCG technique, optimizing images at both the local and global levels, allows the method to excel in various aspects. Furthermore, the TRCG-HW method does not focus solely on one aspect but integrates multiple performance metrics, signifying its well-rounded excellence in different aspects, resulting in overall superior performance.
The TRCG-HW method significantly differs by employing a hierarchical optimization approach to handle high-dimensional hyperspectral data. It utilizes an inner-layer local optimization and an outer-layer global optimization strategy. Local optimization helps reduce computational complexity, while global optimization approximates the true values. This provides it with a computational efficiency advantage, especially in large-scale hyperspectral datasets. The local optimization employs the truncated conjugate gradient (TCG) algorithm at each wavelet decomposition level to ensure the effective extraction and preservation of image details during the image reconstruction process. This enhances spatial accuracy and structural preservation. The outer-layer global optimization introduces the trust region algorithm (TRA) to coordinate features between different local optimization levels to ensure their consistent coordination throughout the entire image. It ensures that the TRCG-HW method can achieve a global optimum solution in both the spectral and spatial dimensions, resulting in higher image quality. Another key feature of the TRCG-HW method is its utilization of multiscale information extraction via wavelet transformation from panchromatic images, which improves spatial accuracy and preserves structural information.
While our study has produced promising outcomes, it is imperative to acknowledge its limitations. One notable constraint pertains to the computational intricacy inherent in utilizing wavelet transformation. The time-consuming nature of this process might pose challenges, especially when handling extensive hyperspectral datasets. Furthermore, our method primarily concentrates on optimizing intensity components within the HIS transformation. Although this reduction in computational complexity is beneficial, it might constrain the comprehensive reconstruction of spatial structures and intricate spatial details.
To mitigate the computational complexity concern, future investigations could explore parallel processing methodologies and hardware acceleration. These approaches could substantially enhance our method’s efficiency, rendering it more viable for real-time applications. Moreover, to overcome the limitations related to spatial details, further exploration could focus on integrating non-local self-similarity attributes of panchromatic and hyperspectral images. Leveraging these characteristics could augment the local optimization process within the TRCG-HW framework, elevating the overall performance in hyperspectral fusion, particularly in spatial reconstruction.
Future research endeavors could be focused on various aspects. Primarily, there should be a concerted effort to adapt the TRCG-HW method for real-time applications, particularly in domains like automated monitoring and decision support systems. This adaptation necessitates heightened performance and computational efficiency, enabling the method to swiftly process data in real-time scenarios. Further refinements targeting the enhancement of the TRCG-HW method’s performance, particularly in computational efficiency and data processing speed, would render it more compatible with real-time applications and large-scale data processing.
Secondly, while the TRCG-HW method exhibits prowess in hyperspectral image fusion, future investigations could extend its utility to diverse fields such as medical imaging, remote sensing, and surveillance. Researchers could explore methods to tailor the TRCG-HW approach to handle various data types, broadening its application spectrum. By applying this method to different data types, like multimodal images or stereoscopic images, it can cater to a wider array of field-specific requirements.
Lastly, an avenue for exploration could involve the fusion of hyperspectral stereo images or other multimodal data to acquire more comprehensive information. Such an approach could bolster a broader range of application domains, including, but not limited to, environmental monitoring and geological exploration. This could lead to more nuanced and enriched data interpretations, amplifying the method’s utility across diverse fields.

5. Conclusions

In this research endeavor, we present the TRCG-HW methodology for the fusion of hyperspectral images, a technique harmonizing HIS, wavelet, and TRCG approaches. This fusion method amalgamates both HIS and wavelet transformations, not solely preserving the inherent traits of the HIS transform but also leveraging panchromatic images for wavelet conversion. This innovative approach augments spatial precision and markedly enhances spectral authenticity. To refine the fusion process and achieve veritable values, our methodology adopts a dual-tiered approach. The inner tier employs the truncated conjugate gradient (TCG) for localized optimization, while the outer layer employs the trust region algorithm (TRA) for global convergence. A meticulous evaluation through visual and quantitative analyses using simulated and empirical datasets substantiates the prowess of the TRCG-HW method. Impressively, this approach demonstrates outstanding performance metrics on both the IKONOS-2 and experimental datasets, registering PSNR scores of 32.92 and 29.59, SSIM scores of 0.974 and 0.981, CC scores of 0.984 and 0.976, SAM scores of 0.001 and 0, and ERGAS scores of 0.270 and 0.338, respectively. These results consistently affirm the superior capabilities of the TRCG-HW method when compared to recent advancements in hyperspectral image fusion methodologies.

Author Contributions

Conceptualization, J.W. and A.G.; methodology, J.W.; software, J.W.; validation, J.W., Q.W. and S.L.; formal analysis, J.W.; investigation, J.W.; resources, A.G.; data curation, J.W.; writing—original draft preparation, J.W.; writing—review and editing, J.W.; visualization, D.Z. and X.C.; supervision, A.G.; project administration, A.G.; funding acquisition, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Yiwu Research Institute of Fudan University [grant numbers 350 20-1-20].

Institutional Review Board Statement

The study did not require ethical approval.

Informed Consent Statement

Not applicable. The study did not involve humans.

Data Availability Statement

The data presented in this study are available from the corresponding author upon reasonable request.

Acknowledgments

Thanks to all institutions or units that published the datasets. Relevant links and citations are indicated in the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The linear RGB to HIS transformation is characterized by Equations (A1)–(A3), while the HIS to RGB transformation is depicted in Equation (A4).
I v 1 v 2 = 1 3 1 3 1 3 2 6 2 6 2 3 1 2 1 2 0 R G B ,
where v 1 and v 2 are intermediate variables, from which H and S can be computed.
S = v 1 2 + v 2 2 ,
H = a r c t a n v 2 v 1 ,
R G B = 1 1 2 1 2 1 1 2 1 2 1 2 0 I v 1 v 2 .
The non-linear transformation from RGB to HIS is expressed as follows:
H = θ                       ,       B G 360 θ ,       B > G ,
where θ = a r c c o s 1 2 [ ( R G ) + ( R B ) ] [ ( R G ) 2 + ( R B ) ( G B ) ] 1 2
I = R + G + B 3 ,
S = 1 3 m i n ( R , G , B ) R + G + B ,
The transformation from HIS to RGB is defined by the following equation:
if 0 ° H 120 ° ,
R = I 1 + S c o s H c o s ( 60 ° H ) ,
B = I ( 1 S ) ,
G = 3 I ( R + B ) .
if   120 ° H 240 ° ,   H = H 120 ° ,   R = I ( 1 S ) ,
G = I 1 + S c o s H c o s ( 60 ° H ) ,
B = 3 I ( R + B ) .
if   240 ° H < 360 ° ,   H = H 240 ° , B = I 1 + S c o s H c o s ( 60 ° H ) ,
G = I ( 1 S ) ,
R = 3 I ( R + B ) .

References

  1. Burger, J.; Gowen, A. Data handling in hyperspectral image analysis. Chemom. Intell. Lab. Syst. 2011, 108, 13–22. [Google Scholar] [CrossRef]
  2. Stuart, M.B.; McGonigle, A.J.S.; Willmott, J.R. Hyperspectral Imaging in Environmental Monitoring: A Review of Recent Developments and Technological Advances in Compact Field Deployable Systems. Sensors 2019, 19, 3071. [Google Scholar] [CrossRef] [PubMed]
  3. Schroeder, T.A.; Wulder, M.A.; Healey, S.P.; Moisen, G.G. Mapping wildfire and clearcut harvest disturbances in boreal forests with landsat time series data. Remote Sens. Environ. 2011, 115, 1421–1433. [Google Scholar] [CrossRef]
  4. Jin, S.; Yang, L.; Zhu, Z.; Homer, C. A land cover change detection and classification protocol for updating alaska nlcd 2001 to 2011. Remote Sens. Environ. 2017, 195, 44–55. [Google Scholar] [CrossRef]
  5. Avola, G.; Matese, A.; Riggi, E. An Overview of the Special Issue on “Precision Agriculture Using Hyperspectral Images”. Remote Sens. 2023, 15, 1917. [Google Scholar] [CrossRef]
  6. Xu, Y.; Tan, Y.; Abd-Elrahman, A.; Fan, T.; Wang, Q. Incorporation of Fused Remote Sensing Imagery to Enhance Soil Organic Carbon Spatial Prediction in an Agricultural Area in Yellow River Basin, China. Remote Sens. 2023, 15, 2017. [Google Scholar] [CrossRef]
  7. Govender, M.; Chetty, K.; Bulcock, H. A review of hyperspectral remote sensing and its application in vegetation and water resource studies. Water SA 2007, 33, 145–151. [Google Scholar] [CrossRef]
  8. Wang, J.Y.; Li, C. Development and prospects of hyperspectral remote sensing imaging technology. J. Space Sci. 2021, 41, 22–33. [Google Scholar] [CrossRef]
  9. Li, S.; Kang, X.; Fang, L.; Hu, J.; Yin, H. Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 2017, 33, 100–112. [Google Scholar] [CrossRef]
  10. Xiao, G.; Bavirisetti, D.P.; Liu, G.; Zhang, X. Feature-Level Image Fusion. In Image Fusion; Springer: Singapore, 2020; pp. 103–147. [Google Scholar] [CrossRef]
  11. Jiang, D.; Kim, J. Image retrieval method based on image feature fusion and discrete cosine transform. Appl. Sci. 2021, 11, 5701. [Google Scholar] [CrossRef]
  12. Xiao, G.; Bavirisetti, D.P.; Liu, G.; Zhang, X. Decision-Level Image Fusion. In Image Fusion; Springer: Singapore, 2020; pp. 149–170. [Google Scholar] [CrossRef]
  13. Zhang, H.; Xu, H.; Tian, X. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  14. Mhangara, P.; Mapurisa, W.; Mudau, N. Comparison of Image Fusion Techniques Using Satellite Pour l’Observation de la Terre (SPOT) 6 Satellite Imagery. Appl. Sci. 2020, 10, 1881. [Google Scholar] [CrossRef]
  15. Al-Doski, J.; Hassan, F.M.; Norman, M.; Najim, A.A. Interaction of image fusion techniques and atmospheric correction for improve SVM accuracy. Earth Sci. Inform. 2022, 15, 2673–2687. [Google Scholar] [CrossRef]
  16. Chen, C.M.; Hepner, G.F.; Forster, R.R. Fusion of hyperspectral and radar data using the ihs transformation to enhance urban surface features. ISPRS J. Photogramm. Remote Sens. 2003, 58, 19–30. [Google Scholar] [CrossRef]
  17. El-Mezouar, M.C.; Taleb, N.; Kpalma, K.; Ronsin, J. An IHS-Based Fusion for Color Distortion Reduction and Vegetation Enhancement in IKONOS Imagery. IEEE Trans. Geosci. Remote Sens. 2011, 49, 1590–1602. [Google Scholar] [CrossRef]
  18. Lin, H.; Zhang, A. Fusion of hyperspectral and panchromatic images using improved HySure method. In Proceedings of the International Conference on Image Vision and Computing, Chengdu, China, 2–4 June 2017. [Google Scholar] [CrossRef]
  19. Wang, B.; Choi, J.; Choi, S.; Lee, S.; Wu, P.; Gao, Y. Image Fusion-Based Land Cover Change Detection Using Multi-Temporal High-Resolution Satellite Images. Remote Sens. 2017, 9, 804. [Google Scholar] [CrossRef]
  20. Kurban, T. Region based multi-spectral fusion method for remote sensing images using differential search algorithm and IHS transform. Expert Syst. Appl. 2022, 189, 116135. [Google Scholar] [CrossRef]
  21. Chen, C.; He, X.; Guo, B. A pixel-level fusion method for multi-source optical remote sensing image combining the principal component analysis and curvelet transform. Earth Sci. Inform. 2020, 13, 1005–1013. [Google Scholar] [CrossRef]
  22. Toet, A.; van Ruyven, J.J.; Mathee Valeton, J. Merging thermal and visual images by a contrast pyramid. Opt. Eng. 1989, 28, 789–792. [Google Scholar] [CrossRef]
  23. Burt, P.; Adelson, E. The Laplacian pyramid as a compact image code. IEEE Trans. Commun. 1983, 31, 532–540. [Google Scholar] [CrossRef]
  24. Burt, P.J.; Kolczyski, R.J. Enhanced Image Capture Through Fusion. In Proceedings of the Fourth IEEE International Conference on Computer Vision, Berlin, Germany, 11–14 May 1993. [Google Scholar] [CrossRef]
  25. Dippel, S.; Stahl, M.; Wiemker, R.; Blaffert, T. Multiscale contrast enhancement for radiographies: Laplacian Pyramid versus fast wavelet transform. IEEE Trans. Med. Imaging 2002, 21, 343–353. [Google Scholar] [CrossRef] [PubMed]
  26. Chipman, L.J.; Orr, T.M.; Graham, L.N. Wavelets and image fusion. Int. Conf. Image Process. 1995, 3, 248–251. [Google Scholar] [CrossRef]
  27. Li, H.; Manjunath, B.S.; Mitra, S.K. Multisensor image fusion using the wavelet transform. Graph. Models Image Process. 1995, 57, 235–245. [Google Scholar] [CrossRef]
  28. Liu, Z.; Tsukada, K.; Hanasaki, K.; Ho, Y.K.; Dai, Y.P. Image Fusion by Using Steerable Pyramid. Pattern Recognit. Lett. 2001, 22, 929–939. [Google Scholar] [CrossRef]
  29. Li, Z.; Jing, Z.; Sun, S.; Liu, G. Remote Sensing Image Fusion Based on Steerable Pyramid Frame Transform. Acta Opt. Sin. 2005, 25, 598–602. [Google Scholar]
  30. Matsopoulos, G.K.; Marshall, S. Application of morphological pyramids: Fusion of MR and CT phantoms. J. Vis. Commun. Image Represent. 1995, 6, 196–207. [Google Scholar] [CrossRef]
  31. Li, S.T.; Kang, X.D.; Hu, J.W. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [CrossRef]
  32. Zhou, Z.Q.; Wang, B.; Li, S.; Dong, M.J. Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion 2016, 30, 15–26. [Google Scholar] [CrossRef]
  33. He, K.M.; Sun, J.; Tang, X.O. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef]
  34. Chen, B.; Cheng, H.; Tseng, Y.; Yin, J. Two-Pass Bilateral Smooth Filtering for Remote Sensing Imagery. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8006405. [Google Scholar] [CrossRef]
  35. Feng, Y.; Wu, J.; Hu, X.; Zhang, W. Medical image fusion using bilateral texture filtering. Biomed. Signal Process. Control 2023, 85, 1746–8094. [Google Scholar] [CrossRef]
  36. Goyal, B.; Lepcha, D.C.; Dogra, A.; Bhateja, V.; Lay-Ekuakille, A. Measurement and analysis of multi-modal image fusion metrics based on structure awareness using domain transform filtering. Measurement 2021, 182, 109663. [Google Scholar] [CrossRef]
  37. Dogra, A.; Kumar, S. Multi-modality medical image fusion based on guided filter and image statistics in multidirectional shearlet transform domain. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 12191–12205. [Google Scholar] [CrossRef]
  38. Chen, H.; Wang, Y. Research on image fusion algorithm based on Laplace pyramid transform. Laser Infrared 2009, 3, 1–10. [Google Scholar] [CrossRef]
  39. Xu, Y.; Sun, B.; Yan, X.; Hu, J.; Chen, M. Multi-focus image fusion using learning based matting with sum of the Gaussian-based modified Laplacian. Digit. Signal Process. 2020, 106, 102821. [Google Scholar] [CrossRef]
  40. Nazir, I.; Haq, I.U.; Khan, M.M.; Qureshi, M.B.; Ullah, H.; Butt, S. Efficient Pre-Processing and Segmentation for Lung Cancer Detection Using Fused CT Images. Electronics 2022, 11, 34. [Google Scholar] [CrossRef]
  41. Wang, Z.; Cui, Z.; Zhu, Y. Multi-modal medical image fusion by Laplacian pyramid and adaptive sparse representation. Comput. Biol. Med. 2020, 123, 103823. [Google Scholar] [CrossRef]
  42. Yao, J.; Zhao, Y.; Bu, Y.; Kong, S.G.; Chan, J.C.-W. Laplacian Pyramid Fusion Network with Hierarchical Guidance for Infrared and Visible Image Fusion. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4630–4644. [Google Scholar] [CrossRef]
  43. Ji, W.; Peng, J.; Xu, B. Real-time detection of underwater river crab based on multi-scale pyramid fusion image enhancement and MobileCenterNet model. Comput. Electron. Agric. 2023, 204, 107522. [Google Scholar] [CrossRef]
  44. Yin, H.; Xiao, J.; Chen, H. CSPA-GAN: A Cross-Scale Pyramid Attention GAN for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2023, 72, 5027011. [Google Scholar] [CrossRef]
  45. Han, X.; Shi, B.; Zheng, T. SSF-CNN: Spatial and Spectral Fusion with CNN for Hyperspectral Image Super-Resolution. In Proceedings of the 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018. [Google Scholar] [CrossRef]
  46. Wang, W.; Zeng, W.; Huang, Y.; Ding, P. Deep Multiscale Feedback Network for Hyperspectral Image Fusion. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5509205. [Google Scholar] [CrossRef]
  47. Wang, Y.; Xu, S.; Liu, J. MFIF-GAN: A new generative adversarial network for multi-focus image fusion. Signal Process. Image Commun. 2021, 96, 116295. [Google Scholar] [CrossRef]
  48. Fei, Z. Applied Mathematics in Digital Image Processing, 2nd ed.; Tsinghua University Press: Beijing, China, 2020; pp. 271–273. [Google Scholar]
  49. Karim, S.; Tong, G.; Li, J.; Qadir, A.; Farooq, U.; Yu, Y. Current advances and future perspectives of image fusion: A comprehensive review. Inf. Fusion 2023, 90, 185–217. [Google Scholar] [CrossRef]
  50. An, G.; Tong, Q.; Zhang, Y.; Liu, R.; Li, W.; Cao, J.; Lin, Y. An Improved Variational Mode Decomposition and Its Application on Fault Feature Extraction of Rolling Element Bearing. Energies 2021, 14, 1079. [Google Scholar] [CrossRef]
  51. Muqeet, M.A.; Holambe, R.S. Local binary patterns based on directional wavelet transform for expression and pose-invariant face recognition. Appl. Comput. Inform. 2019, 15, 163–171. [Google Scholar] [CrossRef]
  52. Wang, Y.; Guo, P.; Ma, N.; Liu, G. Robust Wavelet Transform Neural-Network-Based Short-Term Load Forecasting for Power Distribution Networks. Sustainability 2023, 15, 296. [Google Scholar] [CrossRef]
  53. Xu, T.; Huang, T.Z.; Deng, L.J.; Yokoya, N. An iterative regularization method based on tensor subspace representation for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5529316. [Google Scholar] [CrossRef]
  54. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  55. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  56. Yuhas, R.H.; Goetz, A.; Boardman, J. Discrimination among Semi-Arid Landscape Endmembers Using the Spectral Angle Mapper (SAM) Algorithm. In Proceedings of the Summaries of the Third Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; Volume 1, pp. 147–149. [Google Scholar]
  57. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the Third Conference “Fusion of Earth Data: Merging Point Measurements, Raster Maps and Remotely Sensed Images”, Sophia Antipolis, France, 26–28 January 2000; pp. 99–103. [Google Scholar]
Figure 1. Schematic diagram of our hyperspectral imaging system. A xenon source with a paraboloidal reflector provides stable illumination. Light is shaped by optical elements and cast in parallel onto optical grating. Grating diffraction causes dispersion and a chromatic band of light is shown at the focal plane of the lens. Optical fiber transmits selected light into the imaging section for illumination. Images of tissue are collected by a CMOS camera. A computer synchronizes the CMOS camera and grating rotation, and stores data as a form of hypercube.
Figure 1. Schematic diagram of our hyperspectral imaging system. A xenon source with a paraboloidal reflector provides stable illumination. Light is shaped by optical elements and cast in parallel onto optical grating. Grating diffraction causes dispersion and a chromatic band of light is shown at the focal plane of the lens. Optical fiber transmits selected light into the imaging section for illumination. Images of tissue are collected by a CMOS camera. A computer synchronizes the CMOS camera and grating rotation, and stores data as a form of hypercube.
Electronics 13 00252 g001
Figure 2. RGB spectral power distribution and color coordinates measured by spectrometer. First row: SPD curve. Second row: CIE1931 chromaticity. Third row: CIE1964 chromaticity. (a) Red lighting. (b) Green lighting. (c) Blue lighting.
Figure 2. RGB spectral power distribution and color coordinates measured by spectrometer. First row: SPD curve. Second row: CIE1931 chromaticity. Third row: CIE1964 chromaticity. (a) Red lighting. (b) Green lighting. (c) Blue lighting.
Electronics 13 00252 g002
Figure 3. Preprocessed photos taken with a camera. (a) Red lighting. (b) Green lighting. (c) Blue lighting. (d) Hyperspectral images. (e) Panchromatic images.
Figure 3. Preprocessed photos taken with a camera. (a) Red lighting. (b) Green lighting. (c) Blue lighting. (d) Hyperspectral images. (e) Panchromatic images.
Electronics 13 00252 g003
Figure 4. The 2D DWT filter establishes approximation coefficients W φ ( j , m , n ) and detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) at scale ‘j’ with the approximation coefficients W φ ( j + 1 , m , n ) at scale ‘ j + 1 . Each arrow in the diagram indicates a halving of the image’s resolution by a factor of 2.
Figure 4. The 2D DWT filter establishes approximation coefficients W φ ( j , m , n ) and detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) at scale ‘j’ with the approximation coefficients W φ ( j + 1 , m , n ) at scale ‘ j + 1 . Each arrow in the diagram indicates a halving of the image’s resolution by a factor of 2.
Electronics 13 00252 g004
Figure 5. Two-dimensional image decomposition results based on two successive 2D DWTs. Approximation coefficients W φ ( j , m , n ) encapsulate the overall image characteristics, while detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) represent detailed information in the horizontal, vertical, and diagonal directions, respectively.
Figure 5. Two-dimensional image decomposition results based on two successive 2D DWTs. Approximation coefficients W φ ( j , m , n ) encapsulate the overall image characteristics, while detail coefficients W ψ H ( j , m , n ) , W ψ V ( j , m , n ) , and W ψ D ( j , m , n ) represent detailed information in the horizontal, vertical, and diagonal directions, respectively.
Electronics 13 00252 g005
Figure 6. Schematic diagram of image fusion by combining HIS, Wavelet, and Trust-Region Conjugate Gradient.
Figure 6. Schematic diagram of image fusion by combining HIS, Wavelet, and Trust-Region Conjugate Gradient.
Electronics 13 00252 g006
Figure 7. Qualitative results: reconstruction error maps of the experimental dataset. Blue typically represents smaller errors, while red indicates larger errors. (a) PCA. (b) HIS. (c) WT. (d) TRCG-HW.
Figure 7. Qualitative results: reconstruction error maps of the experimental dataset. Blue typically represents smaller errors, while red indicates larger errors. (a) PCA. (b) HIS. (c) WT. (d) TRCG-HW.
Electronics 13 00252 g007
Figure 8. Qualitative results: reconstruction error maps of the IKONOS-2 dataset. Blue typically represents smaller errors, while red indicates larger errors. (a) PCA. (b) HIS. (c) WT. (d) TRCG-HW.
Figure 8. Qualitative results: reconstruction error maps of the IKONOS-2 dataset. Blue typically represents smaller errors, while red indicates larger errors. (a) PCA. (b) HIS. (c) WT. (d) TRCG-HW.
Electronics 13 00252 g008
Table 1. Average quantitative results of the test methods on the experimental dataset. (Note: ↑ indicates higher values are favorable, while ↓ indicates lower values are favorable.).
Table 1. Average quantitative results of the test methods on the experimental dataset. (Note: ↑ indicates higher values are favorable, while ↓ indicates lower values are favorable.).
MethodPSNR ↑SSIM ↑CC ↑SAM ↓ERGAS ↓
PCA29.500.8050.8764.8960.343
HIS26.980.6790.7711.1820.418
WT27.250.7890.8570.0040.411
TRCG-HW29.590.9810.97600.338
Table 2. Average quantitative results of the test methods on the IKONOS-2 dataset. (Note: ↑ indicates higher values are favorable, while ↓ indicates lower values are favorable.)
Table 2. Average quantitative results of the test methods on the IKONOS-2 dataset. (Note: ↑ indicates higher values are favorable, while ↓ indicates lower values are favorable.)
MethodPSNR SSIM CC SAM ERGAS
PCA30.880.7810.8353.9400.347
HIS29.390.7820.8971.8400.415
WT35.730.8170.8856.5200.285
TRCG-HW32.920.9740.9840.0010.270
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, J.; Ge, A.; Liu, S.; Wang, Q.; Zhu, D.; Chen, X. An Optimized Inversion Method for Hyperspectral Image Fusion Based on a Hue–Intensity–Saturation, Wavelet, and Trust-Region Conjugate Gradient Method. Electronics 2024, 13, 252. https://doi.org/10.3390/electronics13020252

AMA Style

Wu J, Ge A, Liu S, Wang Q, Zhu D, Chen X. An Optimized Inversion Method for Hyperspectral Image Fusion Based on a Hue–Intensity–Saturation, Wavelet, and Trust-Region Conjugate Gradient Method. Electronics. 2024; 13(2):252. https://doi.org/10.3390/electronics13020252

Chicago/Turabian Style

Wu, Jiangbo, Aiming Ge, Shuo Liu, Qiuyang Wang, Dongsheng Zhu, and Xindi Chen. 2024. "An Optimized Inversion Method for Hyperspectral Image Fusion Based on a Hue–Intensity–Saturation, Wavelet, and Trust-Region Conjugate Gradient Method" Electronics 13, no. 2: 252. https://doi.org/10.3390/electronics13020252

APA Style

Wu, J., Ge, A., Liu, S., Wang, Q., Zhu, D., & Chen, X. (2024). An Optimized Inversion Method for Hyperspectral Image Fusion Based on a Hue–Intensity–Saturation, Wavelet, and Trust-Region Conjugate Gradient Method. Electronics, 13(2), 252. https://doi.org/10.3390/electronics13020252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop