Next Article in Journal
FFLKCDNet: First Fusion Large-Kernel Change Detection Network for High-Resolution Remote Sensing Images
Previous Article in Journal
Study on GPR Image Restoration for Urban Complex Road Surfaces Using an Improved CycleGAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ATIS-Driven 3DCNet: A Novel Three-Stream Hyperspectral Fusion Framework with Knowledge from Downstream Classification Performance

1
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
2
College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
These authors are co-first authors of the article.
Remote Sens. 2025, 17(5), 825; https://doi.org/10.3390/rs17050825
Submission received: 17 January 2025 / Revised: 21 February 2025 / Accepted: 24 February 2025 / Published: 26 February 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Reconstructing high-resolution hyperspectral images (HR-HSIs) by fusing low-resolution hyperspectral images (LR-HSIs) and high-resolution multispectral images (HR-MSIs) is a significant challenge in image processing. Traditional fusion methods focus on visual and statistical metrics, often neglecting the requirements of downstream tasks. To address this gap, we propose a novel three-stream fusion network, 3DCNet, designed to integrate spatial and spectral information from LR-HSIs and HR-MSIs. The framework includes two dedicated branches for extracting spatial and spectral features, alongside a hybrid spatial–spectral branch (HSSI). The spatial block (SpatB) and the spectral block (SpecB) are designed to extract spatial and spectral details. The training process employs the global loss, spatial edge loss, and spectral angle loss for fusion tasks, with an alternating training iteration strategy (ATIS) to enhance downstream classification by iteratively refining the fusion and classification networks. Fusion experiments on seven datasets demonstrate that 3DCNet outperforms existing methods in generating high-quality HR-HSIs. Superior performance in downstream classification tasks on four datasets proves the importance of the ATIS. Ablation studies validate the importance of each module and the ATIS process. The 3DCNet framework not only advances the fusion process by leveraging downstream knowledge but also sets a new benchmark for classification-oriented hyperspectral fusion.

1. Introduction

Hyperspectral images (HSIs), boasting hundreds of spectral bands across various wavelengths, offer a treasure trove of spectral information. This richness enables their widespread application in key areas such as agricultural assessment [1], environmental monitoring [2], medical diagnostics [3], and other fields. However, the inherent limitation of a low spatial resolution in HSIs impedes their full application potential. Multispectral images (MSIs), while spectrally less informative, compensate with a higher spatial resolution that captures detailed textural information. The fusion of low-resolution hyperspectral images (LR-HSIs) and high-resolution multispectral images (HR-MSIs) to create high-resolution hyperspectral images (HR-HSIs) is thus a critical research endeavor, optimizing the strengths of both image types.
Recently, the HSI and MSI fusion field has achieved notable advancements, branching into traditional and deep learning methods. Traditional approaches include matrix factorization [4,5,6,7], tensor decomposition [8,9,10], and Bayesian estimation [11,12,13], each requiring substantial domain expertise and complicated, custom-tailored designs. Concurrently, deep learning methods have revolutionized numerous fields and have significantly propelled the frontiers of the HSI fusion field, encompassing models like single-stream [14] and dual-stream fusion architectures [15,16].
Despite these strides, existing studies predominantly concentrate on the visual and statistical metrics of fusion results, often sidelining their subsequent application in enhancing advanced downstream visual tasks. Nevertheless, it is crucial to acknowledge that the independent training of fusion and classification models may lead to unpredictable consequences, as it can introduce additional errors, redundancies, and inefficiencies during task transitions. To the best of our knowledge, only one study [17] considers the downstream object detection application while fusing HSIs and MSIs. However, it is based on traditional methods and does not leverage the simple and powerful fitting capabilities of neural networks. Meanwhile, deep learning-based fusion works have not taken into account downstream high-level visual tasks.
Our work addresses this gap with the introduction of a novel neural network-driven fusion framework, called the three-stream fusion network for downstream classification application (3DCNet). A concept map of 3DCNet and ATIS is shown in Figure 1. The main contributions of our research are summarized as follows:
  • A novel, efficient three-stream fusion network, 3DCNet, is proposed for high-precision LR-HSI and HR-MSI fusion. Three stream means that three branches distill the features in the LR-HSI, HR-MSI, and hybrid spatial–spectral image (HSSI), respectively.
  • For spatial and spectral feature extraction, low-cost linear transformation operations are utilized in spatial blocks (SpatBs) to effectively extract spatial information, while the channel attention mechanism in spectral blocks (SpecBs) is leveraged to capture global spectral information.
  • In order to direct our 3DCNet to acquire precise feature representations, the loss function is segmented into three distinct parts: L g l o b a l , L c a n n y , and L a n g l e . Beyond the global control of HR-HSI generation by L g l o b a l , the Canny operator is employed to constrain the neural network, focusing it on the reconstruction of spatial texture details, while L a n g l e is utilized to control the fidelity of the generated HR-HSI by minimizing the variation in the spectral dimension.
  • For downstream classification application, the alternating training iteration strategy (ATIS) is designed to iteratively train our fusion network alongside the classification network DBDA [18], with the aim of leveraging knowledge from each other, generating an HR-HSI that not only excels in visual and statistical performance but also minimizes the error in accuracy for the downstream high-level classification task.
The remainder of this article is organized as follows: Section 2 provides an overview of the related works on LR-HSI and HR-MSI fusion methods. In Section 3, 3DCNet and ATIS are proposed in detail. Section 4 designs and analyzes fusion experiments on seven datasets and classification experiments on four datasets, along with thorough ablation experiments. Conclusions are provided in Section 5.

2. Related Work

In this section, we first review the traditional methods and deep learning methods used in HSI and MSI fusion tasks. Second, we introduce some fusion methods leveraging knowledge from downstream high-level applications.

2.1. Traditional Fusion Methods

Prior to the emergence of deep learning, the traditional fusion methods were fundamentally structured around three principal categories.
(1) Matrix Factorization-based Methods: Matrix factorization-based fusion methods reinterpret the spatial dimensions of 3D HSIs alongside the spectral dimension, effectively flattening them into a two-dimensional matrix for processing. They decompose them into a spectral basis matrix and a coefficient matrix, and they consider the estimation of the spectral basis and coefficients as an optimization problem.
Kawakami et al. [4] employed sparse encoding algorithms to extract sparse coefficients from LR-HSIs, complemented by sparse dictionary learning methods to derive spectral dictionaries from HR-MSIs. Building upon this, Akhtar et al. [5] introduced a greedy algorithm for calculating sparse coefficients, incorporating the local similarity of the fused HSI into their calculations. Simoes et al. [6] expanded on this by considering both the spatial local similarity and the spectral characteristics of HSIs, employing a non-local sparse representation method to guide the optimization process. Dong et al. [7] further refined the approach by utilizing a sparse representation method with self-similarity constraints to guide the optimization.
Despite these advancements, matrix factorization-based methods have a critical limitation: they tend to overlook the intrinsic three-dimensional nature of HSI data. This oversight can result in a significant loss of information during the fusion process, highlighting the need for more sophisticated approaches that can preserve the rich spectral and spatial information inherent in HSIs.
(2) Tensor Decomposition-based Methods: Tensor decomposition-based methods take a direct approach to the fusion of LR-HSIs and HR-MSIs by treating HSI data as three-dimensional tensors, conceptualizing the fusion task as a tensor operation. Li et al. [8] introduced an innovative paradigm that leverages the Tucker decomposition model, coupled with Coupled Sparse Tensor Factorization (CSTF), for the fusion process. This approach decomposes the HR-HSI into a core tensor and three factor matrices, which are represented by dictionaries. The final estimation of the HR-HSI is derived from the estimation of these core tensors and dictionaries.
Kanatsoulis et al. [9] reformulated the factor matrix estimation into a least squares problem within the framework of canonical polynomial decomposition. In contrast, Dian et al. [10] employed singular value decomposition to learn spectral subspaces from LR-HSIs, estimated coefficients with the aid of low-tensor multi-rank priors, and utilized the clustering structure from HR-MSIs to group patches in the coefficients. The HR-HSI was then approximated using these spectral subspaces and coefficients.
Despite these methodological advancements, tensor decomposition-based methods are computationally intensive, demanding substantial computational resources, which can be a limiting factor in terms of algorithmic efficiency.
(3) Bayesian Estimation-based Methods: Bayesian estimation-based methods address the fusion of LR-HSIs and HR-MSIs from a statistical perspective, focusing on the prior distribution of the images. Hardie et al. [11] introduced a Maximum A Posteriori (MAP) estimator that employs a spatial variation statistical model derived from vector quantization to effectively harness local correlation information.
Zhang et al. [12] defined an innovative operator to articulate the spatial degradation inherent in HSI images. They employed a Bayesian framework with wavelet transform to describe and address the fusion problem by calculating this operator.
Akhtar et al. [13] took a probabilistic approach to infer the distribution and proportion of spectral information for different materials within HR-HSIs. Utilizing these distributions, they engaged in extensive research on the calculation of sparse encoding within HR-HSIs and proposed a Bayesian sparse encoding strategy. They further advanced the field by learning a Bayesian dictionary through the Beta process.
However, the probabilistic nature of these estimation methods inherently involves a degree of randomness, which may not always ensure optimal fusion outcomes.

2.2. Deep Learning-Based Fusion Methods

The development of deep learning has revolutionized the domain of image processing, with significant achievements made in recent years [19,20,21]. Its simplicity and effectiveness have also been demonstrated in the realm of HSI processing [22,23,24,25,26]. Deep learning approaches conceptualize the fusion of LR-HSIs and HR-MSIs as the establishment and refinement of fusion functions, which is tackled in a three-step process. Initially, a neural network is architected to serve as the fusion function, designed to exhaustively capture spatial and spectral image information. Following this, a loss function is carefully defined to guide the optimization process of the fusion function. The sequence of steps is ultimately concluded by employing a gradient descent algorithm to steer the loss function towards an optimal solution.
Masi et al. [27] employed a Convolutional Neural Network (CNN) with a stacked single-branch architecture to distill features from HSIs, marking one of the pioneering efforts directly applicable to the fusion challenge between HSIs and MSIs [28]. Palsson et al. [29] employed 3D-CNNs for image information extraction, incorporating PCA (Principal Component Analysis) priors to enhance the extraction process. Zhang et al. [14] introduced an innovative approach by integrating some spectral bands from HR-MSIs into LR-HSIs to create a composite image, subsequently leveraging a single-branch CNN to extract its features.
To fully leverage the complementary information in HSIs and MSIs, Yang et al. [15] crafted a dual-branch CNN that independently extracts spectral information from LR-HSIs and spatial information from HR-MSIs, merging these streams for a comprehensive fusion. Xu et al. [28] similarly designed a dual-branch network, progressively integrating extracted information from LR-HSIs and HR-MSIs across various scales. Yao et al. [30] introduced a cross-attention module to extract additional spatial information from HR-MSIs, which facilitates a more effective transmission of spatial–spectral information within the network architecture. Zhu et al. [31] proposed a new QIS strategy that combines the hierarchical structure of quadtrees with Implicit Neural Representations (INRs) and further enhances the fidelity of fused images by leveraging a Generative Adversarial Network (GAN) framework.
Neural networks have the advantages of modularity and efficiency, and fusion methods based on deep learning can lead to excellent fusion effects. However, the errors and redundancies introduced by them can reduce the effectiveness of downstream high-level visual applications.

2.3. Fusion Methods for Downstream Applications

The domain of image fusion, particularly in the context of visible and infrared images, has witnessed a surge of effective research on downstream high-level applications in recent years. Tang et al. [32] introduced a framework that concurrently trains models for visible and infrared image fusion and semantic segmentation of the fused output. This approach was later expanded by the same author [33], who transitioned from feature-level to image-level fusion, demonstrating that the latter could achieve superior performance in advanced visual tasks. However, although these works feed the performance of downstream applications back into the fusion process, they risk concentrating the focus towards those applications, potentially compromising the quality of fusion itself.
In the field of hyperspectral fusion, existing methods for fusing LR-HSIs and HR-MSIs have predominantly focused on visual quality and statistical metrics, with limited attention paid to their other roles in supporting downstream applications such as image classification [34] and object detection [35]. He et al. [17] introduced a novel strategy that aims to jointly optimize the super-resolution of HSIs and the downstream visual task of object detection in an effort to enhance both concurrently. However, this traditional method-based approach did not exploit the strengths of end-to-end trained neural networks.
To the best of our knowledge, there remains a lack of research in the hyperspectral and multispectral fusion domain that integrates low-level fusion knowledge and high-level task knowledge reciprocally within each other’s network training process. This highlights an opportunity for further exploration and innovation.

3. Methodology

In this section, the data preprocessing method is first introduced. Then, our three-stream fusion network for downstream classification application, 3DCNet, is proposed comprehensively, which is shown in Figure 2. Afterwards, the feature extraction blocks, SpatB and SpecB, are introduced in detail. Subsequently, we present the global fusion loss L g l o b a l , spatial canny loss L c a n n y , and spectral angle loss L a n g l e . Ultimately, the alternating training iteration strategy (ATIS) is designed to enhance the efficacy of the predicted output for downstream classification tasks.

3.1. Hybrid Spatial–Spectral Image

In the proposed 3DCNet, the reference HR-HSI and the predicted HR-HSI are written as R R H × W × C , Z R H × W × C , respectively, where H, W, and C represent the height, width, and number of channels or bands, respectively. The input LR-HSI and HR-MSI are denoted as X R h × w × C h H , w W and Y R H × W × c c C . First, the LR-HSI is upsampled spatially to match the spatial resolution of the HR-MSI.
X u p = Bilinear X , r
where X u p represents the upsampled LR-HSI, and r is the ratio. Bilinear means the bilinear interpolation operation.
To fully extract the information shared between the HR-MSI and LR-HSI, Y is inserted at equal intervals along the spectral dimension into X u p , resulting in hybrid spatial–spectral image (HSSI), written as I . The insertion position can be formulated as
p i = C · i c + 1 + 1 , i = 1 , 2 , , c
where p i denotes the insertion position of the i-th multispectral band. Then, the HSSI is computed as
I ( k ) = X u p ( k ) , if k { p 1 , p 2 , , p c } Y ( k ) , if k = p i , i = 1 , 2 , , c
where k is the k-th spectral band. By this approach, the resulting I integrates the spectral information from the LR-HSI with the spatial details from the HR-MSI.

3.2. Three-Stream HSI and MSI Fusion Network

In Figure 2, the overall architecture of our proposed 3DCNet is presented. It mainly consists of three branch networks and a global network. From top to bottom, the three branch networks extract the spatial features of Y , the spatial and spectral features of I , and the spectral features of X , respectively. Then, the features are combined into fused feature maps through the concatenation operation. The features in the fused feature maps are comprehensively extracted by the spatial and spectral blocks of the global network, with the final output being generated as Z .
(1) The Branch Network: Specifically, in order to extract the shallow feature maps of Y , a simple 3 × 3 convolutional layer is used.
Y c o n v = Conv ( Y , 3 )
where Conv represents the convolutional operation, and 3 represents the kernel size. Then, the shallow feature map is passed to the spatial block, with a residual connection. We have
Y o u t = Spat r e s ( Y c o n v )
where Spat refers to the spatial block, and r e s refers to the residual connection. The residual connection is used to alleviate the vanishing gradient problem, which allows the network to be built deeper, enabling the extraction of deep-level features. Similarly, the output of the HSI branch X o u t can be formulated as
X c o n v = Conv ( X , 3 ) X u p = Bilinear ( X c o n v , r ) X o u t = Spec r e s ( X u p )
where Spec represents the data flow passing through the spectal block.
The intermediate branch of the network converts I , which integrates spatial information and spectral information, into feature maps and extracts their deep-level features. The formula can be written as
I c o n v = Conv ( I , 3 ) I o u t = Spat-spec r e s ( I c o n v )
where Spat−spec denotes the module that connects the spatial block and the spectral block in series in order to extract spatial and spectral messages.
Before feeding the deep features into the following global network, we use the concatenation operation to gather feature maps,
Z i n = Conc ( Y o u t , I o u t , X o u t )
where Z i n represents the concatenated feature maps, and Conc denotes the concatenation operation. It should be noted that all convolutional layers and the spatial and spectral blocks used in the branch network are designed to maintain the same H, W, and C of the data flow for convenience.
(2) The Global Network: After obtaining the concatenated feature maps Z i n , the global network is utilized to further extract features. Here, the convolution layer using a 1 × 1 kernel size instead of a 3 × 3 kernel size is utilized to extract features. One reason for this is that Z i n has too many channels, and a 1 × 1 kernel size, in comparison, can reduce the number of parameters. The other reason is that the 1 × 1 convolutional operation is used to change the channels of feature maps into the same as R . This can be written as
Z c o n v = Conv ( Z i n , 1 )
Then, the feature maps are sent into the spatial block and the spectral block sequentially. A point to be aware of is that data flow in the global network after the first Conv 1 × 1 maintains the same number of channels as X , Y , Z , and R . This approach is taken for the convenience of using the loss function, which is introduced later. The data flow can be fomulated as
Z o u t = Spat-spec r e s ( Z c o n v )
Eventually, the feature maps are fed into the last 1 × 1 convolutional layer to generate the final output,
Z = Conv ( Z o u t , 1 )
where Z is the predicted HR-HSI.

3.3. SpatB and SpecB

Previous fusion research has made strides with dual-branch networks capable of extracting information from both LR-HSIs and HR-MSIs [28,36]. However, these works did not introduce specialized modules tailored to the distinct characteristics of spatial and spectral information. Drawing inspiration from Ghostnet [37], we design our spatial block (SpatB) to exhaustively capture spatial details. Similarly, we leverage the attention mechanism from CBAM [38] to conceive the spectral block (SpecB), ensuring the comprehensive extraction of spectral information across the global scale.
(1) SpatB: The upper left corner of Figure 2 shows the entire architecture of the spatial block (SpatB). In the SpatB, we denote the input as M i n and the output as M o u t . Firstly, M i n is fed into the Conv 1 × 1 layer, the batch normalization (BN) layer, and the rectified linear unit (ReLU) activation function sequentially, which can be formulated as
M c o n v = Conv ( M i n , 1 ) M b n = BN ( M c o n v ) M r e l u = ReLU ( M b n )
where BN denotes the batch normalization operation, which normalizes each mini-batch of data. This can reduce the risk of overfitting and help accelerate the training speed of the neural network. ReLU refers to the rectified linear unit activation function, which introduces nonlinearity into the neural network and enables it to model complex functional mappings.
Subsequently, the feature maps derived from M r e l u are separated individually along the channel dimension. From these separated feature maps, ghost feature maps are generated through a series of low-cost linear transformation operations denoted as Φ k . It should be noted that these operations are applied individually to each channel through convolution, thereby significantly reducing computational expenses. These ghost feature maps are then concatenated with the feature maps of M r e l u , and the concatenated output is collectively delivered from the SpatB, being the final output M o u t . These can be detailed as
M ϕ k = Φ k ( M r e l u k ) , k { 1 , . . . , K } M o u t = Conc ( M r e l u , M ϕ )
where k represents the kth channel within the feature maps, while K signifies the total channels of the feature maps. The notation Φ k refers to different low-cost linear transformation operations, which can be implemented by different Conv 1 × 1 operations mathematically.
(2) SpecB: In contrast to the SpatB, the SpecB concentrates on the spectral dimension. The upper right corner of Figure 2 shows the SpecB, which employs Conv 1 × 1 to extract shallow-level features from N i n , generating N c o n v . Subsequently, it utilizes an attention mechanism along the channel dimension, producing N a t t e .
The adjacent spectral bands in the HSI exhibit strong correlations, which is applied to HSI unmixing and HSI super-resolution tasks [39,40]. This rich spectral information is a distinctive attribute that sets HSIs apart from conventional RGB images. To fully leverage the abundant spectral information of the HSI, each channel in the ensuing feature maps can be regarded as a linear superposition of all spectral channels in the preceding feature maps [41]. The coefficients of this linear superposition correspond to the weights derived from the the Conv 1 × 1 operation mathematically, which is utilized in the SpecB. The formulation is as follows:
N c o n v = Conv ( N i n , 1 )
The channel-wise attention mechanism is inspired by CBAM [38], which adaptively modulates the feature scale in each channel by considering their cross-spectral interdependencies. The detailed channel-wise attention mechanism is shown in Figure 3. Specifically, max pooling and average pooling operations are applied to downsample the feature maps across spatial dimensions, thereby preserving the intrinsic channel characteristics. The max pooling operation is conducive to identifying prominent features, while average pooling ensures the inclusion of global features. These pooled features are then processed through a shared module, which is designed to refine feature representation. Subsequently, the outputs from the shared module are added together through an element-wise summation, and the resultant feature maps are then passed through a Sigmoid activation function to derive the final channel attention weights. The formulation process can be written as
N m a x p = Pool ( N c o n v , m a x ) N a v e p = Pool ( N c o n v , a v e ) N m l p = Shared ( N m a x p ) + Shared ( N a v e p ) N a t t e = Sigmoid ( N m l p )
where Pool denotes the pooling operation, and m a x and a v e represent the max pooling and average pooling operations. Shared refers to the shared module, which sequentially consists of a Conv 1 × 1 layer, a ReLU activation function, and another Conv 1 × 1 layer. Sigmoid refers to the Sigmoid activation function.
Back to the SpecB, N a t t e then serves as the coefficient to automatically adjust N c o n v . Following this, N a t t e and N c o n v go through element-wise multiplication at their corresponding positions, resulting in feature maps that are scaled through the attention mechanism. Finally, the output of the SpecB, N o u t , is obtained via a residual connection. The details are written as
N m u l t = N a t t e N c o n v N o u t = N i n + N m u l t
where ∗ represents the element-wise multiplication. N o u t denotes the result of applying the residual connection operation.

3.4. Loss Function

The inherent black-box nature of neural networks results in an uncontrollable feature map learning process. Therefore, a carefully crafted loss function for HSI data is essential to better guide the parameter updating process. Empirical evidence from various studies indicates that the L 1 norm loss is effective in penalizing minor errors, thereby reducing the blurring effects associated with image smoothing. However, the L 1 norm loss is sensitive to parameter initialization and necessitates careful hyperparameter tuning. In contrast, the L 2 norm loss exhibits robustness against the initialization of parameters, yielding more stable solutions. Its differentiability across the entire domain further contributes to the stability of the optimization process. By carefully designing the loss function for both spatial and spectral features, it becomes feasible to compensate for the image blurring problems that may arise from the L 2 norm loss. Specifically, the global fusion loss is formulated as
L g l o b a l = 1 H W C h = 1 H w = 1 W c = 1 C R ( h , w , c ) Z ( h , w , c ) 2
where R ( h , w , c ) and Z ( h , w , c ) represent the element in the reference HR-HSI and the predicted HR-HSI, respectively.
However, the aforementioned loss functions target the overall quality of the generated HR-HSI. Despite the low spatial resolution of HSIs, their spatial textural features play an important role in tasks such as classification and target detection. Therefore, our 3DCNet incorporates a spatial loss function.
Image edges, referring to locally discontinuous image features, are critical for capturing rich texture information. To effectively extract these edge features, we employ the Canny operator [42], a widely recognized and utilized algorithm for edge detection in the field of image processing. The Canny operator offers the dual advantages of noise suppression and the ability to detect fine-grained edges. The implementation specifics are as follows:
L c a n n y = MSE ( Canny ( Z s p a t ) , Canny ( R ) )
where MSE denotes the mean square error, and Canny denotes the Canny operator. It is crucial to note that our spatial loss function, L c a n n y , compares the output from the SpatB, Z s p a t , against the reference R . This comparison is instrumental in guiding the network’s data flow process.
Compared with RGB images, a notable feature of HSIs is that hundreds of spectral bands contain rich spectral information. This distinctive feature demands that our loss function precisely quantifies the spectral differences between the predicted HR-HSI and the reference HR-HSI. Given that HSIs are treated as 3D tensors in computation, the angular deviation between these tensors serves as a robust metric for assessing the similarity between the predicted and reference images. Inspired by the methodology presented in [28], we utilize L a n g l e to ensure the spectral fidelity, which is designed as follows:
Angle = 1 H W h = 1 H w = 1 W arcos R c · Z c R c 2 Z c 2 L a n g l e = MSE ( Angle ( Z o u t ) , Angle ( R ) )
where arcos represents the arccosine operation, and ⊤ refers to the transpose of the tensor. R c and Z c represent the vectors of the same pixel along the cth channel in R and Z , respectively.
In summary, the total loss function of our proposed 3DCNet consists of the above three parts and is written as
L t o t a l = λ 1 L g l o b a l + λ 2 L c a n n y + λ 3 L a n g l e
At last, it is worth pointing out that plenty of experiments show that the coefficients of these three loss functions should be set to λ 1 = 10 , λ 2 = λ 3 = 1 .

3.5. Alternating Training Iteration Strategy

In the domain of HSI processing, the HR-HSI fusion models and the downstream HSI classification models are typically trained in isolation. This method, although effective for individual model optimization, can result in discrepancies between the reference data and the fused HR-HSI, potentially impairing the efficacy of the downstream classification task.
Tang et al. [32] introduced a joint adaptive training strategy for the fusion and segmentation tasks of infrared and visible light images. However, this strategy primarily integrates the segmentation network’s performance into the fusion network, potentially overemphasizing the downstream network’s training. This could disrupt the balance between low-level and high-level visual tasks.
We refine this approach in the training fusion and downstream classification networks. Specifically, the fusion model and the classification model are trained repeatedly over multiple cycles. During each training cycle of the fusion network, the classification network’s inference loss L c e from the previous round is reintegrated into the fusion loss function. Reciprocally, during each iteration of the classification network’s training, the fusion network’s inference loss L t o t a l from the last training session is reintegrated into the classification loss function. It is worth noting that our classification network is the DBDA network from [18], which has demonstrated superior performance when trained on reference images. The details of our iterative strategy is shown in Algorithm 1.
In Algorithm 1, L c l a s s represents the classification loss function, and L c e refers to the cross-entropy loss, which is defined as
L c e = m = 1 L y m ( log ( n = 1 L e y n ^ ) y m ^ )
where L represents the total count of labeled pixels, and y m and y n denote the categories of the mth and nth pixels, respectively. The coefficients of these loss functions are set to λ 4 = 10 and λ 5 = 0.1 .
Algorithm 1 Alternating Training Iteration Strategy
  1: Input: LR-HSI Y and HR-MSI X
  2: Output: Fused HR-HSI Z
  3: Initial: Loop variable n = 0
  4: while  n N  do
  5:    for p iterations do
  6:       Randomly select the LR-HSI region Y i and
              the corresponding HR-MSI region X i ;
  7:       Calculate the fusion loss:
               L f u s i o n i = L t o t a l i + λ 4 · n · L c e i ;
  8:       Update parameters of the fusion network N F
              by Adam optimizer: N F ( L f u s i o n i ) ;
  9:    end for
10:    Generate fused HR-HSI Z from Y and X using
            the training set;
11:    Select t % labeled data from the fused HR-HSI
            as training data T ;
12:    for q epochs do
13:       Select batched training samples T i ;
14:       Calculate the classification loss:
               L c l a s s i = L c e i + λ 5 · n · L t o t a l i ;
15:       Update parameters of the classification network
               N C by another Adam optimizer: N C ( L c l a s s i ) ;
16:    end for
17:     n = n + 1 ;
18: end while

4. Experiments

This section presents a series of fusion experiments conducted on seven datasets, aimed at proving the generalization ability of our proposed 3DCNet. Downstream classification experiments are performed on the first four of these datasets to prove the effectiveness of our ATIS. The structure of this section is as follows: First, the experimental datasets and evaluation metrics are introduced. Following this, detailed experimental settings and a comparative analysis with nine cutting-edge methods are provided. Then, a series of ablation studies prove the indispensability of each model component and the validity of the ATIS. Finally, the section ends with efficiency experiments and a hyperparameter sensitivity analysis.

4.1. Dataset

In this article, seven datasets are used to validate the generalization performance of our 3DCNet, namely, the Pavia University, Pavia Center, Indian Pines, Botswana, Washington DC Mall, Urban, and CAVE datasets. The first four datasets are utilized to assess the ATIS’s efficacy, as their labeled samples are relatively dispersed, which helps to crop the testing data.
(1) Pavia University: The Pavia University dataset was captured by ROSIS sensors over Pavia University in Italy. After removing samples without any information, the dataset consists of 610 × 340 pixels and 103 spectral bands, with wavelengths ranging from 0.43 to 0.86 μm at a spatial resolution of 1.3 m. This dataset contains nine categories, such as asphalt, trees, and bare soil.
(2) Pavia Center: The Pavia Center dataset was captured by ROSIS sensors over Pavia Center in Italy. After removing samples that do not contain any information, the dataset consists of 1096 × 1096 pixels and 102 spectral bands, with wavelengths ranging from 0.43 to 0.86 μm at a spatial resolution of 1.3 m. This dataset contains nine categories, such as water, trees, and bare soil.
(3) Indian Pines: The Indian Pines dataset was collected by AVIRIS sensors over the Indian Pines test site in northwestern Indiana. After removing bands covering the water absorption area, the dataset consists of 145 × 145 pixels and 200 spectral bands, with wavelengths ranging from 0.4 to 2.5 μm. The available ground truth is divided into 16 categories, such as alfalfa, soybean mint, and woods.
(4) Botswana: The Botswana dataset contains a sequence of data captured by the Hyperion sensor on the EO-1 satellite over Okavango Delta, Botswana. After removing bands that cover the water absorption characteristics, the dataset consists of 1476 × 256 pixels and 145 bands, with spectral wavelengths ranging from 0.4 to 2.5 μm at a spatial resolution of 30 m. This dataset contains 14 determined categories, representing water, reeds, island interval, etc.
(5) Washington DC Mall: The Washington DC Mall dataset contains aerial hyperspectral images obtained by a Hydice sensor over the Washington shopping center. This dataset consists of 1280 × 307 pixels and 191 bands, with spectral bands ranging from 0.4 to 2.5 μm.
(6) Urban: The Urban dataset was collected in Copperas Cove, TX, USA. After removing the dense water vapor and atmospheric bands, the dataset consists of 307 × 307 pixels and 162 bands at a spatial resolution of 2 m, with wavelengths ranging from 0.4 to 2.5 μm.
(7) CAVE: The CAVE dataset consists of 32 hyperspectral images acquired indoors, each measuring 512 by 512 pixels, with 31 spectral bands spanning wavelengths from 0.4 μm to 0.7 μm at 10 nm intervals.

4.2. Evaluation Metrics

Five widely used fusion metrics are used to evaluate the performance of our 3DCNet and other comparison methods, and three commonly used classification metrics are used to evaluate the performance of downstream classification performance. The following five metrics represent fusion effects from different aspects.
(1) Root Mean Square Error: The RMSE measures the distance between the reference image and the predicted image but is sensitive to outliers in the data.
RMSE = c = 1 C h = 1 H w = 1 W ( R Z ) 2 H W C
where R represents the reference HR-HSI, and Z refers to the predicted HR-HSI. C, H, and W correspond to the number of channels or bands, height, and width of the HSI, respectively. The smaller the RMSE, the better the performance.
(2) Peak Signal to Noise Ratio: The PSNR is one of the most commonly used indicators to measure image quality:
PSNR = 10 · log 10 max R c 2 1 H W R c Z c 2 2
where R c and Z c represent the cth channel of the reference image and the predicted image, respectively. · 2 means the second norm. After adding up the PSNRs of all channels, we obtain the PSNR metric of the whole image. The larger the PSNR, the better the performance.
(3) Relative Dimensionless Global Error in Synthesis: ERGAS is an indicator used to evaluate the quality of remote sensing images [43]; it considers the mean square error and brightness information of the image to provide a comprehensive evaluation of the performance of the image. The definition formula is as follows:
ERGAS = 100 r 1 C c = 1 C R c Z c 2 2 μ 2 R c
where r refers to the spatial downsampling ratio when we obtain the LR-HSI. μ R c represents the mean value of the cth channel of the reference image. The lower the ERGAS value, the higher the image quality.
(4) Spectral Angle Mapper: The SAM is a spectral angle mapping that treats the spectrum of each pixel in an image as a high-dimensional vector [44]. The similarity between spectra is measured by calculating the angle between the two vectors. The SAM is defined as
SAM = cos 1 c = 1 C r c z c c = 1 C r c 2 c = 1 C z c 2
where r c and z c refer to the spectral vectors of the reference and predicted HR-HSIs. The smaller the SAM, the more similar the two spectra.
(5) Structural Similarity Index Measure: The SSIM is an indicator used to measure the degree of similarity between two digital images [45]. Structural similarity aligns more closely with the human visual perception of image quality, providing a metric that better reflects the discernment of human eyes. The SSIM is formulated as
SSIM = [ l ( R , Z ) ] α · [ c ( R , Z ) ] β · [ s ( R , Z ) ] γ
l ( R , Z ) = 2 μ R μ Z + C 1 μ R 2 + μ Z 2 + C 1
c ( R , Z ) = 2 σ R σ Z + C 2 σ R 2 + σ Z 2 + C 2
s ( R , Z ) = σ R Z + C 3 σ R σ Z + C 3
where l, c, and s represent a comparison of the brightness, contrast, and structure between the reference image and the predicted image, respectively. α , β , γ are weighting coefficients, typically set to 1. μ R and μ Z denote the average value of R and Z , while σ R and σ Z denote the standard deviation. σ R Z denotes the covariance of R and Z . The higher the SSIM, the better the performance.
In addition, the overall accuracy (OA) and average accuracy (AA) are used to evaluate the quality of classification performance. The Kappa coefficient (KAPPA) reflects the consistency between the ground truth and the classification results. The closer all three metrics to 1, the better.

4.3. Implementation Details

(1) Data Preprocessing: Due to the data scarcity of the remote sensing dataset, we select 128 × 128 pixels from the original dataset as the testing data, and we randomly crop 128 × 128 pixels from other regions in each iteration as the training data. During the training phase, we fill the testing area with 0 to ensure that the model does not use any testing data. In order to verify the effectiveness of our ATIS, we especially select regions containing more samples with different categories as the testing data. The specific regions to be cropped in each dataset are shown in Table 1. In particular, the Indian Pines dataset only has 145 × 145 pixels in total, so we replace the size of the training and testing sets with 64 × 64. In addition, the LR-HSI is obtained by the 5 × 5 Gaussian kernel blurring operation with a standard deviation of 2, and then it is downsampled with a ratio of 4, while the HR-MSI is simulated by selecting five bands located in the HR-HSI at equal intervals. All the above operations are proven effective in SSRNet [14]. For the CAVE dataset, we select 22 out of 32 pictures for training, while the remaining 10 are employed for testing purposes. We perform random translations, rotations, and shears on the training set of the CAVE dataset, expanding the number of images by a factor of 14.
(2) Training Details: When obtaining image edges, the Gaussian kernel size used in the Canny operator is 3 × 3 with a standard deviation of 1, while the low threshold is set to 0.15 , and the high threshold is set to 0.30 . Four traditional methods, namely, CNMF [39], LTTR [46], NSSR [7], and CTDF [47], as well as eight deep learning methods, namely, TFNet [36], ResTFNet [36], SSF [48], ConSSF [48], MSDCNN [49], SSRNet [14], MSST [16], and DCFormer [50], are used as a comparison.
For traditional methods, everything else remains consistent with the original papers, except for data preprocessing. To avoid the influence of different dimensions and numerical ranges, as well as to accelerate the convergence speed of the algorithms, we standardize all data to a range of 0–255 by using the Min–Max Scaler. For deep learning methods, the same experimental setup as the original papers [14] is adopted, specifically using Adam as the optimizer, with a learning rate of 1 × 10−4. We mask the test region of the dataset, designating the remaining area as the training region. From the training region, we randomly sample 10,000 images to form the training set, with a batch size of 32. It should be noted that we are unable to implement MSDCNN using the Caffe framework, but Pytorch is used as an alternative. Additionally, the original SGD optimizer in the MSDCNN model is replaced with the more effective Adam optimizer.
Our model uses Adam as the optimizer with an initial learning rate of 10 3 . The learning rate is adjusted using the exponential decay strategy, and the decay coefficient is set to 0.9995 , preventing network parameter oscillations in the later training stages.
As for the alternating training iteration strategy (ATIS), we set the best iteration count N to 4. When training the classification model, we fix the random seed to select the same training samples in each iteration, and the data preprocessing method is the same as that used by the fusion model. The batch size for the training samples of the classification model is set to 16, and the training epoch is set to 120, with all other parameters being consistent with the original settings in DBDA [18]. It should be noted that, we experiment with various sampling ratios for the training set of the DBDA network. An excessive number of samples can lead to model overfitting, while an insufficient number results in underfitting. Finally, for our task, the optimal training set sampling ratios are adopted as follows: Pavia University 1%, Pavia Center 0.5%, Indian Pines 5%, and Botswana 3%.
All deep learning methods are trained and tested in PyTorch 2.2.0 and Python 3.11 on an Intel Xeon Gold 6326 CPU (Intel Corporation, Santa Clara, CA, USA) and an NVIDIA GeForce RTX 4090 (NVIDIA Corporation, Santa Clara, CA, USA).

4.4. Fusion Performance Comparison

In this section, in order to verify the superiority of our 3DCNet, comparative experiments are conducted on the fusion performance of nine state-of-the-art methods on the Pavia University, Pavia Center, Indian Pines, Botswana, Washington DC Mall, and Urban datasets, both qualitatively and quantitatively. CNMF utilizes the matrix factorization-based method to synthesize HR-HSIs using an endmember matrix and a spatial abundance matrix. NSSR utilizes sparsity theory and leverages the spatial–spectral sparsity of HSIs by jointly learning a non-negative dictionary and estimating structured sparse codes to enhance image resolution. LTTR is a tensor decomposition-based method, and it learns the correlations among spatial and spectral modes considering nonlocal similarity, thereby achieving a super-resolution of the HSI. CTDF enhances feature extraction by representing a third-order HR-HSI with a higher-order spatial factor and a third-order spectral factor. The SSF method concatenates the HR-MSI and LR-HSI directly at the beginning of the network. ConSSF concatenates in the same way in each convolutional layer. TFNet utilizes the auto-encoder architecture to extract the spatial and spectral features separately. Skip connection is used in ResTFNet to learn deeper features. MSDCNN extracts information from different scales and different depths using convolutional layers and the residual learning trick. SSRNet designs a differential loss function and can achieve excellent performance using only three simple convolutional layers. MSST is a Transformer structure that incorporates masked band auto-encoders and masked patch auto-encoders, and it employs a self-supervised strategy to pre-train the network. DCFormer uses directional pairwise multi-head cross-attention to promote inter-modality information exchange, and the window attention mechanism is used to boost self-attention.
(1) Results on Pavia University: The quantitative comparison results are shown in Table 2, and the qualitative comparison RGB fusion results are shown in Figure 4. Based on Table 2, we can summarize that our 3DCNet outperforms all models in all metrics and that SSRNet ranks second only due to the detailed design of their loss functions. The better performance of ResTFNet than TFNet shows that residual connections can better help networks learn deep features. The better performance of our 3DCNet and SSRNet indicates that spatial and spectral loss functions play an important role in guiding networks during the training process. The MSST and DCFormer networks use a Transformer-based architecture, which requires more training data to support its large number of parameters. Therefore, their performance is not optimal on remote sensing images with limited data.
In Figure 4 and other comparison figures, it is worth noting that the first row shows the predicted RGB HR-HSIs of the algorithms or models and that the second row shows the difference RGB images between the predicted HR-HSIs and the reference HR-HSIs. The last column presents the ground truth. We can draw the conclusion that CNMF visually loses texture details. NSSR loses spatial resolution from the difference map. LTTR performs better, while it can still be seen in its difference map that there is a significant variation between it and the ground truth. We can observe from the difference image that CTDF performs relatively evenly across different areas of the image; however, overall, it introduces some noise. SSF achieves better performance than ConSSF, which indicates that the original image data introduced multiple times at different levels of depth may lead to the overfitting of the model.
(2) Results on Pavia Center: Table 3 shows the quantitative results of all models on the Pavia Center dataset. It can be seen that the proposed 3DCNet performs best among all models. The same conclusion as that regarding the Pavia University dataset can be drawn due to the similarity of these two scenes, except for in terms of MSST. This is likely because it effectively extracts information from both hyperspectral and multispectral images during the pretraining phase. ConSSF still achieves the lowest performance metrics. CTDF outperforms other traditional methods, likely due to its connection through the fourth-order spatial factor and third-order spectral factor, which enables it to achieve enhanced feature extraction capabilities. In Figure 5, it can be seen that ConSSF achieves relatively better difference images. We speculate that the reason for this is that it performs well in the bands selected for display but shows poor performance in other bands. TFNet, ResTFNet, and MSDCNN exhibit slight pixel distortion. However, our 3DCNet results in the minimal loss.
(3) Results on Indian Pines: As demonstrated in Table 4, 3DCNet and CTDF perform the best and second best, respectively, surpassing all other models in all metrics. However, the former has a worse ERGAS than the latter, indicating that the former’s detail features and texture information still need further improvement. Other traditional methods achieve comparable or even better results than deep learning methods, suggesting that CNMF captures both the local and global information of the dataset effectively. NSSR finds a better sparse representation for this dataset. LTTR also effectively captures and utilizes spatial and spectral similarity information. However, SSF and ConSSF overall perform noticeably worse than the other models, indicating that simple models, if not equipped with well-designed loss functions, lead to underfitting.
The RGB and difference maps in Figure 6 reveal that, although SSF achieves worse metrics, its RGB image is clearer than LTTR’s, and its difference image shows smaller errors compared to LTTR. This indicates that it is not enough to evaluate a model solely based on metrics. From the RGB images, it is clear that CNMF, CTDF, and our 3DCNet capture more texture information.
(4) Results on Botswana: The quantitative results of our 3DCNet and other comparative methods on the Botswana dataset are presented in Table 5. It is evident that, overall, our 3DCNet outperforms other models on this dataset. For the Botswana dataset, which has a relative substantial number of training samples, the residual connections enable ResTFNet to better capture deep features compared to TFNet, thus producing superior results. The outcomes of MSDCNN demonstrate that multiscale and multidepth CNNs are not sufficient. Moreover, the inferior results of ConSSF compared to those of SSF suggest that injecting shallow-layer information into subsequent layers via residual connections can lead to insufficient feature learning if the integration point is not properly selected. As observed in Figure 7, CNMF clearly fails to learn the characteristics of this dataset. The RGB images of LTTR and ConSSF appear brighter, indicating a significant deviation from the reference images. The difference maps of MSST and DCFormer show that window attention can capture more details when using a Transformer-based architecture.
(5) Results on Washington DC Mall: The experimental results of all models are shown quantitatively in Table 6 and qualitatively in Figure 8. Our 3DCNet ranks first in terms of all metrics. However, CNMF shows that alternating nonnegative matrix factorization algorithms can fully extract spatial and spectral information on the Washington DC Mall dataset. LTTR shows that nonlocal similarity has the potential to preserve the visual message in terms of the SSIM metric. We can see that ConSSF and SSF exhibit relatively large fluctuations in performance across different datasets. Moreover, 3DCNet uses a loss function similar to SSRNet, but its more complex and refined network architecture allows it to outperform the latter by a large margin. CTDF still performs stably, but the method of mapping data to higher dimensions may lead to convergence to a local optimum. In Figure 8, the difference results reveal that ConSSF and MSDCNN fail to show robust generalization. Moreover, the iterative integration of HSI and MSI data and the application of multi-scale convolutional layers within their fusion strategies do not have the desired effects. From the RGB images, it can be observed that NSSR, TFNet, ResTFNet, and SSRNet produce brighter images, indicating that these methods do not capture the luminance and color information effectively. DCFormer exhibits more detailed texture information than MSST, suggesting that the window attention mechanism is more capable of capturing local details.
(6) Results on Urban: Table 7 lists the experimental results of our 3DCNet and other comparative methods. The proposed 3DCNet obtains the best performance among all models in all metrics, thus demonstrating overall superiority. SSRNet ranks second for all metrics, showing the validity of the elaborate design of its loss function. The better performance of our 3DCNet than SSRNet demonstrates the advantage of using different branches to seperately extract features. CTDF achieves the second-best results in the RMSE, PSNR, and SSIM, indicating that the CTDF method handles pixel-level details better but overlooks the structural and global consistency of the image. In contrast, SSRNet performs better in terms of image structure. Furthermore, MSDCNN and ResTFNet outperform MSST and DCFormer overall, suggesting that convolutional operations perform better than Transformers on datasets with limited data. Figure 9 shows that CNMF, NSSR, and LTTR lose more spatial information due to their blurry output HR-HSIs and their difference images. When we zoom in on the difference images, we can see that, although DCFormer performs well on average in terms of RGB images, the variance in fluctuations at each pixel compared to the reference image is relatively large.
(7) Results on CAVE: In Table 8, we can see the weakness of our 3DCNet when it is applied with fewer bands. Its performance does not reach the best level but still achieves second place. This could be because fewer bands imply a lack of training data, and the shallower network depth of 3DCNet means that it cannot extract deeper high-dimensional features. In contrast, ResTFNet, although simpler in network structure and training approach, benefits from its deeper network layers and residual connections, which allow it to capture deeper information more effectively. ConSSF performs well on the SAM and SSIM metrics, indicating its strong ability to capture global information. In Figure 10, it can be seen that the RGB images of NSSR and MSDCNN have a reddish tint, while SSRNet exhibits a blackish tint. This indicates that these models fail to maintain the original color balance during the image reconstruction process, leading to the reddish or blackish color bias. Although 3DCNet does not achieve better metrics, its difference image suggests that it performs better than ResTFNet in terms of detail recovery.
To sum up, the proposed 3DCNet shows the best fusion performance on seven datasets. Deep learning methods do not always show superiority to traditional methods, with this depending on the careful design of the information extraction method. The advantages of feeding data to the model all in one stage, using residual connection structures, and carefully designing loss functions are apparent. Although all other methods may perform excellently on specific datasets, 3DCNet excels due to its strong generalization ability and stable performance across various metrics.

4.5. Downstream Classification Performance Comparison

In this subsection, we compare the performance of downstream classification applications on fused images generated by various deep learning-based fusion methods on four datasets: the Pavia University, Pavia Center, Indian Pines, and Botswana datasets. All the classes in the testing area are ordered in Table 9. DBDA [18] is utilized as the classification network. Initially, we apply the ATIS solely to our 3DCNet, iteratively refining both the fusion and classification networks. The quantitative classification results on the Pavia University dataset are presented in Table 10. It is important to note that our statistics on classification accuracy for each category, as well as the metrics OA, AA, and KAPPA, are based on the testing area. It is evident that our 3DCNet, enhanced by the ATIS, achieves the highest OA, AA, and KAPPA within the test area, and the prediction accuracy for each category reaches above 90%. Other deep learning models fail to predict any samples of trees, bitumen, and shadows. This may be attributed to the fact that, for these categories, the small number of training samples shows significant variation along the spectral dimension, which hinders the classification model from acquiring any meaningful knowledge. This observation indicates that, despite Figure 4 showing minimal visual differences in the predictions of all models, tiny variations can lead to significant prediction biases for downstream classification neural networks. The significant role of the ATIS in iteratively converging the two networks to generate images is thus highlighted. Consequently, in the subsequent comparative experiments for the downstream classification application, we employ the ATIS to iterate over all fusion networks.
(1) Results on Pavia University: Even though all deep learning fusion models employ the ATIS, our 3DCNet, due to its superior performance in various aspects, is capable of generating images that closely resemble the reference while ensuring that the spectral curves of samples in the same category are more similar. The visual results of classification using the ATIS on the Pavia University dataset are shown in Figure 11, with the quantitative results presented in Table 11. The color of the classification results can be known in DBDA [18]. The visual results clearly demonstrate that the downstream classification outcomes of our 3DCNet are the closest to the ground truth, followed by SSRNet, whereas other deep learning models misclassify a significant number of samples, particularly in the bare soil category. The quantitative results indicate that our 3DCNet achieves the highest OA, AA, and KAPPA, and the classification accuracy for all categories is beyond 90%, even near 100%. Upon examining the classification outcomes of SSF, ConSSF, TFNet, ResTFNet, and MSDCNN after the application of the ATIS, it is evident that the prediction accuracies for the two most abundant categories, bare soil and bitumen, are not high, which also explains their relatively lower OAs. Although the fusion results of MSST and DCFormer are slightly inferior, through multiple rounds of iterative training on the fusion and classification networks, they obtain sufficient training data, thereby achieving relatively optimal classification performance. The highest KAPPA achieved by 3DCNet indicates the highest level of consistency in the downstream classification outcomes. SSRNet, which similarly utilizes the image edge extraction loss function, achieves results close to those of 3DCNet, thereby indirectly validating the effectiveness of the Canny operator employed by 3DCNet.
(2) Results on Pavia Center: The classification results after the application of the ATIS on various models on the Pavia Center dataset are presented in Table 12. The quantitative results show that, for the most abundant category within the test area, bitumen, all models achieve an accuracy rate of over 99%; thus, all attained OAs of more than 90%. Regarding the asphalt category, although our 3DCNet has a lower prediction accuracy, it still surpasses MSST and DCFormer. Overall, our 3DCNet achieves the highest OA, AA, and KAPPA, demonstrating that, under the ATIS strategy, 3DCNet can perform more effectively with better convergence of both the fusion and classification models, and MSST follows closely. The lower AAs of the other models suggests that their prediction accuracy for different category samples is not stable.
(3) Results on Indian Pines: The classification results of 3DCNet and the other deep learning models on the Indian Pines dataset after employing the ATIS are presented in Table 13. Overall, our 3DCNet ranks first place in AA and second in terms of both OA and KAPPA, with a narrow margin behind MSDCNN. Despite its relatively poor fusion performance, the superior classification performance of MSDCNN on this dataset may be attributed to the ATIS making the fusion model and the classification model further converge. Moreover, the generally lower AAs achieved by all models on the Indian Pines dataset may be due to the fact that, within the prediction area, the grass-pasture category, although only consisting of 18 sample points, is not correctly predicted by any model. This is probably because the downstream classification DBDA network is sensitive to the spectral curve variations in the grass-pasture category. Compared to SSRNet and 3DCNet, SSF and ConSSF achieve poorer results, which suggests that simple fusion models, without the use of well-designed loss functions, cannot achieve satisfactory convergence effects given the condition of a small number of samples, thus leading to a lower accuracy in downstream classification applications. In contrast, TFNet and ResTFNet utilize deeper and more complex network architectures, enhancing the models’ representational capacity to some degree. MSST and DCFormer have the same condition as TFNet and ResTFNet.
(4) Results on Botswana: The results of the downstream classification application on the Botswana dataset are presented in Table 14. It should be noted that the distribution of the categories in Botswana is relatively dispersed. We specifically select the test area mentioned in Table 1 to cover as many categories as possible, thereby sacrificing the count of samples per category. Compared to other datasets, all models exhibit better classification performance on the Botswana dataset in terms of OA, AA, and KAPPA. However, SSF, ConSSF, and ResTFNet still achieve a prediction accuracy of 0% for the reeds 1 category, leading to their relatively lower AAs. SSF has a large improvement in both OA and KAPPA, which may imply that, although the fused images generated by SSF are not of high quality, they retain deep spectral characteristics that can be used in the classification process. This could also be attributed to the small sample size of the test area in the Botswana dataset, resulting in large variability in the results. Our 3DCNet achieves nearly 100% prediction accuracy in almost all categories, except for the exposed soils category, which also verifies that our 3DCNet, in conjunction with the ATIS strategy, can be applied to datasets with varying conditions.
In summary, 3DCNet achieves the best classification results across the four datasets. This is mainly attributed to the quality of the fused HR-HSI generated by the fusion networks. The higher the fusion assessment metrics of the generated images, the more the details that can be captured and the more faithfully they reproduce the spectral curves, which further confirms the conclusions of our fusion experiments.

4.6. Ablation Experiments

In this section, we first study the significance of various components in our 3DCNet on the Urban dataset, followed by the impact of different elements of the alternating training iteration strategy (ATIS) on the Pavia University dataset. Finally, another classification model, 3DCFormer [51], is used to validate the effectiveness of the ATIS.
(1) 3DCNet Ablation Experiments: As depicted in Table 15, the term w/o-spat, short for without spatial, signifies the removal of every SpatB as well as L c a n n y from the network. Similarly, w/o-spec denotes the absence of all SpecB and L a n g l e . W/o-2stream indicates the elimination of the X and Y branches from the 3DCNet model. The terms w/o-canny or w/o-angle represent the exclusion of only L c a n n y or L a n g l e from the loss function, respectively.
The findings reveal that all these components utilized are instrumental in the superior performance of 3DCNet. Notably, the absence of the SpatB significantly affects all metrics, emphasizing its role in capturing fine-grained spatial details. A comparative analysis between the scenarios of w/o-canny versus w/o-angle, and w/o-spat versus w/o-spec suggests that reconstructing the spatial information of the HR-HSI is more challenging than the spectral information within the Urban dataset.
It is also noteworthy that the w/o-2stream ablation experiment utilizes the same single-stream feature extraction network as SSRNet, thereby highlighting the effectiveness of our three-stream information extraction network in generating HR-HSIs. A thorough comparison of the outcomes from all these ablation architectures to the loss functions, along with input data, proves that 3DCNet’s excellence is a result of its multiple advantageous components.
(2) ATIS Ablation Experiments on Classification results: As shown in Table 16, the baseline for our study is established by classification on the reference strategy, where the classification network is trained directly on the reference HR-HSI. Notably, our implementation of the DBDA [18] network achieves results surpassing those of the original paper due to an increased training sample set. The transferred classification model strategy denotes the use of a model pretrained on the reference HR-HSI without further adaptation to the fused images produced by 3DCNet, while the one-stage training strategy reflects a non-iterative approach. The without loss interaction row illustrates the impact of disabling loss function communication during the alternating training of the fusion and classification networks, setting n = 0 throughout Algorithm 1.
The results clearly show that the OA, AA, and KAPPA metrics of our ATIS closely match those of classification on the reference strategy, indicating that the ATIS effectively refines the fused images to better serve the downstream classification task. We can observe that, after employing the ATIS strategy, our classification performance surprisingly surpasses the strategy of classification on the reference image. However, this does not guarantee that our method outperforms direct classification on the reference image for every individual category. This result falls within the acceptable margin of error. It is attributed to the overall performance improvement caused by the bare soil category, which has the largest number of samples, being predicted entirely correctly. The outcome of the transferred classification model strategy highlights the significant deviation between the fused HR-HSI generated by the fusion network and the reference HR-HSI, so the retraining of the classification network on the fused images is necessary. Furthermore, a comparison between the one-stage training strategy and our ATIS reveals that the initial fused images must undergo multiple iterative refinements to enhance their utility for classification. The without loss interaction scenario, where there is no information exchange between the training processes of the fusion and classification networks, leads to slightly inferior results.
Thus, it can be concluded that our ATIS integrates all the aforementioned elements: it trains the classification network on the generated fused HR-HSI, employs multiple iterations, and maintains information exchange between the two networks during the iteration process. Consequently, the downstream classification results achieved with the ATIS are the best.
(3) ATIS Ablation Experiments on Fusion results: We conduct ablation studies on the Pavia University dataset to observe the impact of the ATIS on the fusion results, as shown in Table 17. It can be seen that, when only the fusion task is considered, all fusion metrics reach their optimal values. However, using a one-stage ATIS framework, where the classification loss is incorporated into the fusion task only once, leads to a noticeable decline in fusion performance. However, with the full ATIS framework, which iteratively incorporates the classification loss into the fusion network and the fusion loss into the classification network, the results show some improvement but still fall short of the performance achieved by using the fusion network alone. It is worth noting that, although the ATIS has a slight impact on the fusion results, it still outperforms the other comparison methods shown in Table 2.
(4) Classification Model Ablation Experiments: Another classification model, 3DCFormer [51], is utilized to validate the effectiveness of the ATIS. We fix the upstream fusion network as our 3DCNet while replacing the downstream classification network from DBDA to 3DCFormer. All other settings remain the same. The generalization ability of the ATIS is verified sequentially on four datasets: the Pavia University, Pavia Center, Indian Pines, and Botswana datasets. The results are presented in Table 18. The results show that the ATIS is not only applicable to the classification model DBDA but also works well with other classification models.

4.7. Efficiency Experiments

While deep learning methods can build highly complex models that achieve excellent performance, the huge number of parameters often hampers their deployment, and the long inference time renders them impractical for real-time applications. This section evaluates the computational complexity of deep-learning models on the Urban dataset through three key metrics: the number of parameters, floating-point operations, and testing time. The detailed data are presented in Table 19. It should be noted that all models are tested on the testing area specified in Table 1.
According to the quantitative data, our 3DCNet ranks first on parameters and FLOPs and second for testing time. While SSRNet has a parameter count close to that of our 3DCNet, its FLOPs are 1.46 times higher, demanding greater computational resources. Although ResTFNet has the second-lowest FLOPs, its larger number of parameters and longer testing time increase its deployment barrier and decrease its suitability for real-time applications. MSST and DCFormer have a large number of parameters, resulting in the highest FLOPs and testing time. This makes it difficult to deploy them on edge devices and challenging to meet real-time requirements. It is important to note that, despite our 3DCNet having the fewest parameters and FLOPs, the low-cost linear transformation operations within the SpatB result in frequent memory access, which significantly increases the testing time. Consequently, the testing time of 3DCNet is 1.3 times longer than that of SSRNet, but it still ranks as the second fastest.

4.8. Hyperparameter Sensitivity Analysis

The training process involves many hyperparameters, and the values of these hyperparameters are crucial for the convergence of the network. We select the four most important hyperparameters for this experiment to conduct a hyperparameter analysis. These include the coefficients of the different loss functions, as well as the initial learning rate. As shown in Figure 12, the PSNR values corresponding to the coefficients of the loss functions are on the left side of the plot, while the PSNR values corresponding to the learning rates are on the right side. With the other parameters kept constant, we conduct 20 experiments for λ 1 , and the results are shown by the blue line. It can be seen that, as λ 1 increases, the PSNR value of the fused image generally rises first and then decreases. From the black and orange lines, it is clear that, as λ 2 or λ 3 increases, the PSNR value of the fused image gradually decreases. The red line represents the choice of the initial learning rates, specifically 10 1 , 10 2 , 10 3 , 10 4 , and 10 5 . Since a learning rate that is too high may prevent the model from converging, we decrease the learning rate gradually in the training process to maintain training stability. Even so, we observe that the choice of the initial learning rate has a significant impact on the convergence of the model. Only when the learning rate is 10 3 is the model able to find the global optimal region early in training, gradually approaching the optimal value in the later stages. Higher learning rates cause parameter oscillations, which may miss the global optimum, while lower learning rates result in the parameters staying close to the initial values, preventing the model from reaching the optimal region. In conclusion, the selection of these hyperparameters in this study is both reasonable and meaningful.

5. Conclusions

This paper presents a novel HSI and MSI fusion framework, the three-stream fusion network for downstream classification application (3DCNet), designed to generate HR-HSIs from LR-HSIs and HR-MSIs. The SpatB and SpecB are designed to effectively extract spatial and spectral features, respectively. The loss functions L c a n n y and L a n g l e are utilized to guide the learning process of spatial and spectral features. The proposed 3DCNet framework has been extensively evaluated and compared with state-of-the-art methods across seven datasets, demonstrating superior performance visually and quantitatively. Our 3DCNet also shows superiority in terms of parameter count and floating-point operations. In particular, the ATIS enhances the classification performance by transferring knowledge between the fusion and downstream classification networks during the training phrase, thereby overcoming the ineffective issue of classification on direct fusion results.
Although our 3DCNet achieves optimal fusion performance and ensures little inference time, the frequent memory access associated with low-cost linear transformation operations increases the deployment of the network. While the ATIS can be utilized for the downstream classification task, the choice of datasets limits its application to a broader range of downstream applications such as object detection and image captioning. In the future, we aim to further optimize the model architecture to reduce the frequency of memory access and extend the application of our ATIS to other hyperspectral datasets to complete more downstream tasks.

Author Contributions

Conceptualization, Q.Z. and J.L. (Jian Long); Methodology, Q.Z. and J.L. (Jian Long); Software, J.S.; Validation, Q.Z., J.L. (Jian Long) and C.L.; Formal analysis, J.L. (Jun Li) and J.S.; Investigation, C.L. and J.S.; Resources, J.L. (Jun Li) and Y.P.; Data curation, J.S.; Writing—original draft, Q.Z.; Writing—review & editing, J.L. (Jian Long); Visualization, C.L.; Supervision, J.L. (Jun Li) and Y.P.; Project administration, Y.P.; Funding acquisition, J.L. (Jun Li) and Y.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Opening Foundation of State Key Laboratory of High-Performance Computing, National University of Defense Technology, under Grant No. 202201-05. This work was partially supported by the Young Scientists Fund of the National Natural Science Foundation of China, under Grant No. 62401578.

Data Availability Statement

The data presented in this study are openly available in https://github.com/ZhangQuan-hub/3DCNet.git (accessed on 7 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Zhou, S.; Sun, L.; Xing, W.; Feng, G.; Ji, Y.; Yang, J.; Liu, S. Hyperspectral imaging of beet seed germination prediction. Infrared Phys. Technol. 2020, 108, 103363. [Google Scholar] [CrossRef]
  2. Ghanbari, H.; Antoniades, D. Convolutional neural networks for mapping of lake sediment core particle size using hyperspectral imaging. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102906. [Google Scholar] [CrossRef]
  3. Akbari, H.; Kosugi, Y.; Kojima, K.; Tanaka, N. Detection and analysis of the intestinal ischemia using visible and invisible hyperspectral imaging. IEEE Trans. Biomed. Eng. 2010, 57, 2011–2017. [Google Scholar] [CrossRef] [PubMed]
  4. Kawakami, R.; Matsushita, Y.; Wright, J.; Ben-Ezra, M.; Tai, Y.W.; Ikeuchi, K. High-resolution hyperspectral imaging via matrix factorization. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE: New York, NY, USA, 2011; pp. 2329–2336. [Google Scholar]
  5. Akhtar, N.; Shafait, F.; Mian, A. Sparse spatio-spectral representation for hyperspectral image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part VII 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 63–78. [Google Scholar]
  6. Simoes, M.; Bioucas-Dias, J.; Almeida, L.B.; Chanussot, J. A convex formulation for hyperspectral image superresolution via subspace-based regularization. IEEE Trans. Geosci. Remote. Sens. 2014, 53, 3373–3388. [Google Scholar] [CrossRef]
  7. Dong, W.; Fu, F.; Shi, G.; Cao, X.; Wu, J.; Li, G.; Li, X. Hyperspectral image super-resolution via non-negative structured sparse representation. IEEE Trans. Image Process. 2016, 25, 2337–2352. [Google Scholar] [CrossRef] [PubMed]
  8. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J.M. Fusing hyperspectral and multispectral images via coupled sparse tensor factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef]
  9. Kanatsoulis, C.I.; Fu, X.; Sidiropoulos, N.D.; Ma, W.K. Hyperspectral super-resolution: A coupled tensor factorization approach. IEEE Trans. Signal Process. 2018, 66, 6503–6517. [Google Scholar] [CrossRef]
  10. Dian, R.; Li, S. Hyperspectral image super-resolution via subspace-based low tensor multi-rank regularization. IEEE Trans. Image Process. 2019, 28, 5135–5146. [Google Scholar] [CrossRef] [PubMed]
  11. Hardie, R.C.; Eismann, M.T.; Wilson, G.L. MAP estimation for hyperspectral image resolution enhancement using an auxiliary sensor. IEEE Trans. Image Process. 2004, 13, 1174–1184. [Google Scholar] [CrossRef]
  12. Zhang, Y.; De Backer, S.; Scheunders, P. Noise-resistant wavelet-based Bayesian fusion of multispectral and hyperspectral images. IEEE Trans. Geosci. Remote. Sens. 2009, 47, 3834–3843. [Google Scholar] [CrossRef]
  13. Akhtar, N.; Shafait, F.; Mian, A. Bayesian sparse representation for hyperspectral image super resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3631–3640. [Google Scholar]
  14. Zhang, X.; Huang, W.; Wang, Q.; Li, X. SSR-NET: Spatial–spectral reconstruction network for hyperspectral and multispectral image fusion. IEEE Trans. Geosci. Remote. Sens. 2020, 59, 5953–5965. [Google Scholar] [CrossRef]
  15. Yang, J.; Zhao, Y.Q.; Chan, J.C.W. Hyperspectral and multispectral image fusion via deep two-branches convolutional neural network. Remote. Sens. 2018, 10, 800. [Google Scholar] [CrossRef]
  16. Jia, S.; Min, Z.; Fu, X. Multiscale spatial–spectral transformer network for hyperspectral and multispectral image fusion. Inf. Fusion 2023, 96, 117–129. [Google Scholar] [CrossRef]
  17. He, C.; Xu, Y.; Wu, Z.; Wei, Z. Connecting Low-Level and High-Level Visions: A Joint Optimization for Hyperspectral Image Super-Resolution and Target Detection. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5514116. [Google Scholar] [CrossRef]
  18. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote. Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  19. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face recognition: A convolutional neural-network approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. Acm 2017, 60, 84–90. [Google Scholar] [CrossRef]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Tang, X.; Li, C.; Peng, Y. Unsupervised joint adversarial domain adaptation for cross-scene hyperspectral image classification. IEEE Trans. Geosci. Remote. Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  23. Li, C.; Tang, X.; Shi, L.; Peng, Y.; Zhou, T. An efficient joint framework assisted by embedded feature smoother and sparse skip connection for hyperspectral image classification. Infrared Phys. Technol. 2023, 135, 104985. [Google Scholar] [CrossRef]
  24. Bhatti, U.A.; Huang, M.; Neira-Molina, H.; Marjan, S.; Baryalai, M.; Tang, H.; Wu, G.; Bazai, S.U. MFFCG–Multi feature fusion for hyperspectral image classification using graph attention network. Expert Syst. Appl. 2023, 229, 120496. [Google Scholar] [CrossRef]
  25. Li, C.; Rasti, B.; Tang, X.; Duan, P.; Li, J.; Peng, Y. Channel-Layer-Oriented Lightweight Spectral-Spatial Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5504214. [Google Scholar] [CrossRef]
  26. Zhu, C.; Dai, R.; Gong, L.; Gao, L.; Ta, N.; Wu, Q. An adaptive multi-perceptual implicit sampling for hyperspectral and multispectral remote sensing image fusion. Int. J. Appl. Earth Obs. Geoinf. 2023, 125, 103560. [Google Scholar] [CrossRef]
  27. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote. Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  28. Xu, S.; Amira, O.; Liu, J.; Zhang, C.X.; Zhang, J.; Li, G. HAM-MFN: Hyperspectral and multispectral image multiscale fusion network with RAP loss. IEEE Trans. Geosci. Remote. Sens. 2020, 58, 4618–4628. [Google Scholar] [CrossRef]
  29. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. Multispectral and hyperspectral image fusion using a 3-D-convolutional neural network. IEEE Geosci. Remote. Sens. Lett. 2017, 14, 639–643. [Google Scholar] [CrossRef]
  30. Yao, J.; Hong, D.; Chanussot, J.; Meng, D.; Zhu, X.; Xu, Z. Cross-attention in coupled unmixing nets for unsupervised hyperspectral super-resolution. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXIX 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 208–224. [Google Scholar]
  31. Zhu, C.; Deng, S.; Zhou, Y.; Deng, L.J.; Wu, Q. QIS-GAN: A lightweight adversarial network with quadtree implicit sampling for multispectral and hyperspectral image fusion. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  32. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  33. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Rethinking the necessity of image fusion in high-level vision tasks: A practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inf. Fusion 2023, 99, 101870. [Google Scholar] [CrossRef]
  34. Sun, H.; Zheng, X.; Lu, X. A supervised segmentation network for hyperspectral image classification. IEEE Trans. Image Process. 2021, 30, 2810–2825. [Google Scholar] [CrossRef] [PubMed]
  35. Zhang, G.; Zhao, S.; Li, W.; Du, Q.; Ran, Q.; Tao, R. HTD-Net: A deep convolutional neural network for target detection in hyperspectral imagery. Remote. Sens. 2020, 12, 1489. [Google Scholar] [CrossRef]
  36. Liu, X.; Liu, Q.; Wang, Y. Remote sensing image fusion based on two-stream fusion network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef]
  37. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
  38. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  39. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled nonnegative matrix factorization unmixing for hyperspectral and multispectral data fusion. IEEE Trans. Geosci. Remote. Sens. 2011, 50, 528–537. [Google Scholar] [CrossRef]
  40. Wycoff, E.; Chan, T.H.; Jia, K.; Ma, W.K.; Ma, Y. A non-negative sparse promoting algorithm for high resolution hyperspectral imaging. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; IEEE: New York, NY, USA, 2013; pp. 1409–1413. [Google Scholar]
  41. Jiang, J.; Sun, H.; Liu, X.; Ma, J. Learning spatial-spectral prior for super-resolution of hyperspectral imagery. IEEE Trans. Comput. Imaging 2020, 6, 1082–1096. [Google Scholar] [CrossRef]
  42. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 679–698. [Google Scholar] [CrossRef]
  43. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the Third Conference “Fusion of Earth Data: Merging Point Measurements, Raster Maps and Remotely Sensed Images”, Sophia Antipolis, France, 26–28 January 2000; SEE/URISCA: Nice, France, 2000; pp. 99–103. [Google Scholar]
  44. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop. Volume 1: AVIRIS Workshop, Pasadena, CA, USA, 1–5 June 1992. [Google Scholar]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  46. Dian, R.; Li, S.; Fang, L. Learning a low tensor-train rank representation for hyperspectral image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2672–2683. [Google Scholar] [CrossRef]
  47. Xu, T.; Huang, T.Z.; Deng, L.J.; Xiao, J.L.; Broni-Bediako, C.; Xia, J.; Yokoya, N. A Coupled Tensor Double-Factor Method for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5515417. [Google Scholar] [CrossRef]
  48. Han, X.H.; Shi, B.; Zheng, Y. SSF-CNN: Spatial and spectral fusion with CNN for hyperspectral image super-resolution. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), IEEE, Athens, Greece, 7–10 October 2018; 2018; pp. 2506–2510. [Google Scholar]
  49. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A multiscale and multidepth convolutional neural network for remote sensing imagery pan-sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef]
  50. Ma, Q.; Jiang, J.; Liu, X.; Ma, J. Reciprocal transformer for hyperspectral and multispectral image fusion. Inf. Fusion 2024, 104, 102148. [Google Scholar] [CrossRef]
  51. Wang, Y.; Yu, X.; Wen, X.; Li, X.; Dong, H.; Zang, S. Learning a 3D-CNN and Convolution Transformers for Hyperspectral Image Classification. IEEE Geosci. Remote. Sens. Lett. 2024, 21, 5504505. [Google Scholar]
Figure 1. The fusion of an LR-HSI and an HR-MSI to generate an HR-HSI. Initially, the interactions are captured by the hybrid information capturing operation. Then, three-stream data are passed to train 3DCNet. The alternating training iteration strategy is utilized to complete the downstream high-level classification task using the predicted HR-HSI.
Figure 1. The fusion of an LR-HSI and an HR-MSI to generate an HR-HSI. Initially, the interactions are captured by the hybrid information capturing operation. Then, three-stream data are passed to train 3DCNet. The alternating training iteration strategy is utilized to complete the downstream high-level classification task using the predicted HR-HSI.
Remotesensing 17 00825 g001
Figure 2. Three-stream LR-HSI and HR-MSI fusion network. SpatB refers to the spatial block, and SpecB refers to the spectral block.
Figure 2. Three-stream LR-HSI and HR-MSI fusion network. SpatB refers to the spatial block, and SpecB refers to the spectral block.
Remotesensing 17 00825 g002
Figure 3. Channel attention mechanism utilized in SpecB. Two data streams share parameters in the shared module.
Figure 3. Channel attention mechanism utilized in SpecB. Two data streams share parameters in the shared module.
Remotesensing 17 00825 g003
Figure 4. Fusion results of Pavia University based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (67-29-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 4. Fusion results of Pavia University based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (67-29-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g004
Figure 5. Fusion results of Pavia Center based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (67-29-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 5. Fusion results of Pavia Center based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (67-29-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g005
Figure 6. Fusion results of Indian Pines based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (29-15-4 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 6. Fusion results of Indian Pines based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (29-15-4 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g006
Figure 7. Fusion results of Botswana based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (48-15-4 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 7. Fusion results of Botswana based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (48-15-4 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g007
Figure 8. Fusion results of Washington DC Mall based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (55-35-11 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 8. Fusion results of Washington DC Mall based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (55-35-11 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g008
Figure 9. Fusion results of Urban based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (26-11-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 9. Fusion results of Urban based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (26-11-1 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g009
Figure 10. Fusion results of CAVE based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (31-21-11 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Figure 10. Fusion results of CAVE based on different models, where “GT” refers to the ground truth image. The (first row) shows the RGB images (31-21-11 bands) of the estimated HR-HSIs, and the (second row) shows the difference images between the estimated and reference RGB images, which are processed by a pseudo-color technique.
Remotesensing 17 00825 g010
Figure 11. Downstream classification maps for the Pavia University dataset using 1% training samples. “GT” refers to the classification ground truth. Overall accuracies are as follows: SSF, 82.30%; ConSSF, 90.18%; TFNet, 72.42%; ResTFNet, 62.01%; MSDCNN, 79.47%; SSRNet, 94.04%; MSST, 91.01%; DCFormer, 91.57%; 3DCNet, 99.00%. The classification colors are explained in DBDA [18].
Figure 11. Downstream classification maps for the Pavia University dataset using 1% training samples. “GT” refers to the classification ground truth. Overall accuracies are as follows: SSF, 82.30%; ConSSF, 90.18%; TFNet, 72.42%; ResTFNet, 62.01%; MSDCNN, 79.47%; SSRNet, 94.04%; MSST, 91.01%; DCFormer, 91.57%; 3DCNet, 99.00%. The classification colors are explained in DBDA [18].
Remotesensing 17 00825 g011
Figure 12. Hyperparameter analysis plot on the Urban dataset. Lambda 1, Lambda 2, and Lambda 3 represent the coefficients of the loss functions L g l o b a l , L c a n n y , and L a n g l e , respectively. The learning rate represents the initial learning rate.
Figure 12. Hyperparameter analysis plot on the Urban dataset. Lambda 1, Lambda 2, and Lambda 3 represent the coefficients of the loss functions L g l o b a l , L c a n n y , and L a n g l e , respectively. The learning rate represents the initial learning rate.
Remotesensing 17 00825 g012
Table 1. Testing data region.
Table 1. Testing data region.
DatasetPixel SizeLeft Top Position
Pavia University 128 × 128 ( 211 , 106 )
Pavia Center 128 × 128 ( 241 , 417 )
Indian Pines 64 × 64 ( 1 , 1 )
Botswana 128 × 128 ( 562 , 76 )
Washington DC Mall 128 × 128 ( 576 , 89 )
Urban 128 × 128 ( 89 , 89 )
Table 2. The fusion results of the experimentally compared methods on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 2. The fusion results of the experimentally compared methods on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodPavia University
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF11.04726.9515.43110.7580.845
NSSR6.46731.6023.2416.2680.939
LTTR4.99433.8462.5054.2790.948
CTDF2.13641.2241.4672.5510.977
SSF2.03341.6541.3412.2540.984
ConSSF2.68139.2501.6972.5630.977
TFNet2.31340.5331.5182.4870.979
ResTFNet2.14441.1931.4392.3750.981
MSDCNN2.43240.0971.5812.6710.980
SSRNet1.75642.9281.1751.9960.987
MSST2.83338.7721.7882.7670.976
DCFormer2.66639.2981.6142.0790.986
3DCNet1.60143.7291.1011.8850.988
Table 3. The fusion results of the experimentally compared methods on the Pavia Center dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 3. The fusion results of the experimentally compared methods on the Pavia Center dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodPavia Center
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF4.53035.0102.8205.4650.938
NSSR3.64336.9012.3964.4380.964
LTTR3.71436.7332.2474.0920.956
CTDF2.15841.4491.7893.3980.970
SSF1.81842.9411.5452.7250.981
ConSSF5.01534.1253.3416.0610.964
TFNet2.18541.3401.8183.1110.975
ResTFNet1.98842.1631.7082.9270.977
MSDCNN2.12641.5801.7763.1410.976
SSRNet1.64743.7961.4402.5090.984
MSST1.76943.1751.5662.6950.981
DCFormer2.17241.3951.6552.6530.983
3DCNet1.53144.4321.3822.4300.985
Table 4. The fusion results of the experimentally compared methods on the Indian Pines dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 4. The fusion results of the experimentally compared methods on the Indian Pines dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodIndian Pines
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF5.03834.0852.4754.4800.770
NSSR5.16833.8652.5364.2380.780
LTTR4.69534.6992.3084.6070.777
CTDF3.23737.0621.3872.5080.946
SSF9.27228.7879.7847.5820.838
ConSSF7.62330.48915.5875.9490.612
TFNet5.51533.3012.6733.7990.917
ResTFNet5.20333.8062.6153.6110.923
MSDCNN5.35533.5562.8323.7630.917
SSRNet5.02834.1037.9773.5980.874
MSST5.09933.1152.4503.4760.922
DCFormer4.19134.8171.6192.8060.945
3DCNet2.67938.7061.5991.9970.970
Table 5. The fusion results of the experimentally compared methods on the Botswana dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 5. The fusion results of the experimentally compared methods on the Botswana dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodBotswana
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF2.16626.4447.7878.9030.832
NSSR1.02532.9453.6405.7640.864
LTTR0.81734.9192.8086.9430.866
CTDF0.39639.2752.6821.9230.998
SSF0.84334.64511.5094.3220.982
ConSSF1.09532.36815.2774.8650.960
TFNet0.52138.8213.1982.5790.997
ResTFNet0.46939.7432.9812.3680.997
MSDCNN0.59137.7303.6022.9280.996
SSRNet0.52238.8096.7192.6570.992
MSST0.68334.5333.0002.6200.995
DCFormer0.35240.2922.5381.6170.998
3DCNet0.31942.5212.4741.5230.998
Table 6. The fusion results of the experimentally compared methods on the Washington DC Mall dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 6. The fusion results of the experimentally compared methods on the Washington DC Mall dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodWashington DC Mall
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF0.95445.8790.1750.3080.995
NSSR2.70836.8140.4970.8480.958
LTTR1.35642.8250.2490.3580.985
CTDF1.26343.4420.2300.5140.983
SSF19.54519.6463.6398.2200.941
ConSSF13.61122.7892.6025.7010.933
TFNet1.92339.7860.3370.6460.969
ResTFNet1.82440.2490.3190.6130.973
MSDCNN3.05635.7650.5351.0090.928
SSRNet2.29138.2660.4010.8010.957
MSST2.37237.9640.4130.7380.953
DCFormer1.55041.6590.2780.5410.992
3DCNet0.83047.0830.1480.3010.995
Table 7. The fusion results of the experimentally compared methods on the Urban dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 7. The fusion results of the experimentally compared methods on the Urban dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodUrban
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF7.13328.8183.6218.0830.923
NSSR6.12030.1483.1056.4590.950
LTTR7.02428.9523.5756.8390.911
CTDF2.27638.7411.3912.6660.988
SSF9.03226.7674.6968.7960.963
ConSSF3.95133.9481.9713.2360.972
TFNet3.12735.9791.7632.9570.983
ResTFNet2.94736.4961.6232.7380.984
MSDCNN2.96636.4381.6922.9920.983
SSRNet2.48337.9851.2802.4550.988
MSST3.46035.1021.9043.1330.977
DCFormer3.74734.4102.1212.0630.987
3DCNet2.06239.5981.0792.0620.991
Table 8. The fusion results of the experimentally compared methods on the Cave dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 8. The fusion results of the experimentally compared methods on the Cave dataset. The best scores are marked in red, and the second best scores are marked in blue.
MethodCAVE
RMSEPSNRERGASSAMSSIM
Best Value0 + 001
CNMF5.49333.4363.48413.8480.927
NSSR5.96234.5294.87110.3320.965
LTTR1.89844.5971.3214.1630.992
CTDF1.55646.0671.1273.6240.992
SSF2.06242.3461.3063.1980.992
ConSSF1.77743.9771.1052.8010.994
TFNet1.53145.1690.9633.0170.995
ResTFNet1.33845.9800.8502.9170.995
MSDCNN1.62944.4021.0312.9920.995
SSRNet1.86043.6461.2463.2680.995
MSST1.62345.8451.1613.7860.990
DCFormer2.32641.3571.4803.4600.991
3DCNet1.44646.0920.9162.9980.994
Table 9. Class labels for different datasets. The classes for each dataset are listed in the order shown in tables below.
Table 9. Class labels for different datasets. The classes for each dataset are listed in the order shown in tables below.
DatasetClass
Pavia Universityasphalt, meadows, trees, painted metal sheets, bare soil, bitumen, self-blocking bricks, shadows
Paviawater, trees, asphalt, self-blocking bricks, bitumen, tiles, bare soil
Indian Pinescorn-notill, corn-mintill, corn, grass-pasture, grass-trees, oats, soybean-notill, soybean-mintill, soybean-clean, buildings-grass-trees-drives, stone-steel-towers
Botswanawater, hippo grass, reeds 1, firescar 2, Acacia woodlands, exposed soils
Table 10. The results of the downstream classification experiments on the Pavia University dataset using 1% of available labeled data. ATIS is only applied to our 3DCNet, and other models directly train the classification models. The best scores are marked in red.
Table 10. The results of the downstream classification experiments on the Pavia University dataset using 1% of available labeled data. ATIS is only applied to our 3DCNet, and other models directly train the classification models. The best scores are marked in red.
ClassCountSSFConSSTFNetResTFMSDSSRMSSTDCForOurs
1365100.089.59100.094.8098.0896.7198.3696.99100.0
2570100.0100.099.83100.0100.0100.00100.0100.0100.0
311400000000100.0
458095.86100.098.62100.0100.098.83100.0100.0100.0
5353740.4350.7537.0436.5336.8754.8256.6343.99100.0
67530000000094.42
752100.0100.0100.0100.0100.0100.0100.0100.094.23
82330000000092.70
OA47.9253.5846.2345.7846.1656.3057.4550.1699.00
AA54.5455.0454.4453.9254.3756.3056.8755.1297.67
KAPPA35.9541.0134.6834.3834.5443.3544.9038.0998.43
Table 11. The results of the downstream classification experiments on the Pavia University dataset using 1% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
Table 11. The results of the downstream classification experiments on the Pavia University dataset using 1% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
ClassCountSSFConSSTFNetResTFMSDSSRMSSTDCForOurs
1365100.0100.0100.0100.0100.0100.0100.0100.0100.0
257097.19100.084.74100.0100.0100.096.6797.72100.0
3114100.0100.0100.0100.0100.0100.0100.0100.0100.0
4580100.0100.0100.0100.0100.098.45100.0100.0100.0
5353772.7585.5861.8637.3866.6496.1686.3288.01100.0
675384.4687.1263.8881.5487.6573.7194.1690.4494.42
752100.0100.0100.0100.0100.0100.0100.0100.094.23
823399.5799.1498.7198.7199.5788.4195.2893.9992.70
OA82.3090.1872.4262.0179.4794.0491.0191.5799.00
AA94.2596.4888.6589.7094.2394.5996.5596.2797.67
KAPPA75.1585.5362.6653.0671.8990.8386.6687.4398.43
Table 12. The results of the downstream classification experiments on the Pavia Center dataset using 0.5% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
Table 12. The results of the downstream classification experiments on the Pavia Center dataset using 0.5% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
ClassCountSSFConSSTFNetResTFMSDSSRMSSTDCForOurs
1563100.0100.0100.0100.0100.0100.0100.0100.0100.0
24119.5158.5446.3419.5119.5160.98100.0100.097.56
332799.6999.6974.62100.085.9373.0952.6045.8766.36
452443.7050.5760.1258.7875.3893.8999.24100.097.71
5343799.97100.099.8099.51100.099.94100.0100.0100.0
6635100.0100.098.27100.0100.099.84100.0100.0100.0
724100.0100.0100.0100.0100.0100.0100.0100.0100.0
OA94.0695.0194.0295.2196.2597.4697.1496.8197.78
AA80.4186.9782.7382.5482.9889.6893.1292.2794.52
KAPPA89.2491.0389.1991.4793.3595.5494.9594.3496.10
Table 13. The results of the downstream classification experiments on the Indian Pines dataset using 5% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
Table 13. The results of the downstream classification experiments on the Indian Pines dataset using 5% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
ClassCountSSFConSSTFNetResTFMSDSSRMSSTDCForOurs
173296.4589.2193.5894.9597.8198.0998.6393.4493.44
243485.9599.0899.0899.7797.4798.8594.7091.4898.16
323767.0973.8463.2988.1979.7573.0085.2380.1783.12
418000000000
5190100.098.95100.0100.0100.099.4799.4799.4794.21
6666.6766.67100.0100.083.33100.0100.0100.0100.0
76021.6740.0040.0068.3380.0046.6761.6761.6766.67
825282.9484.1398.4188.8978.9759.1353.1894.0593.25
951091.3781.3780.3954.9084.3187.0685.6982.5582.75
1089100.098.8895.5189.8995.5197.7592.1495.5195.51
119395.67100.095.70100.094.6297.85100.0100.098.93
OA87.6787.0788.4085.8890.5488.2988.2589.2490.27
AA73.4475.6578.7280.4581.0777.9979.1681.6782.37
KAPPA85.0684.4886.0283.0788.6285.8585.7587.0988.34
Table 14. The results of the downstream classification experiments on the Botswana dataset using 3% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
Table 14. The results of the downstream classification experiments on the Botswana dataset using 3% of the available labeled data. The best scores are marked in red, and the second best scores are marked in blue.
ClassCountSSFConSSTFNetResTFMSDSSRMSSTDCForOurs
134100.0100.0100.0100.097.06100.0100.0100.0100.0
258100.0100.089.6686.21100.098.28100.086.21100.0
3300100.00100.0100.0100.0100.0100.0
443100.0100.0100.0100.0100.0100.0100.0100.0100.0
533100.0100.0100.0100.0100.0100.0100.0100.0100.0
61662.5043.750.0050.0062.5037.5062.5087.5087.50
OA95.1993.5888.2489.8491.4494.1296.7994.6598.93
AA77.0873.9681.6172.7083.8989.3093.7595.6297.92
KAPPA93.8091.7785.2187.2389.0992.5195.9093.2698.63
Table 15. Ablation experiments of the proposed 3DCNet on the Urban dataset. The best data are marked in red.
Table 15. Ablation experiments of the proposed 3DCNet on the Urban dataset. The best data are marked in red.
StrategyUrban Ablation Experiment
RMSEPSNRERGASSAMSSIM
w/o-spat2.54737.7621.2492.5800.987
w/o-spec2.30238.6411.1522.2910.989
w/o-2stream2.26438.7871.1742.2520.990
w/o-canny2.35838.4331.1792.1350.990
w/o-angle2.14739.2451.1242.1260.990
3DCNet2.06239.5981.0792.0620.991
Table 16. Ablation experiments of our ATIS on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 16. Ablation experiments of our ATIS on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
StrategyClassMetric
12345678OAAAKAPPA
classification on the reference100.0100.0100.0100.098.9095.88100.099.5798.8699.2998.22
transferred classification model100.0100.0100.0100.080.2482.74100.099.5786.6295.3280.70
one stage training88.49100.095.6192.2473.400100.0067.4668.7254.89
without loss interaction100.093.33100.0100.093.3697.7492.3189.7094.8795.8192.15
our ATIS100.0100.0100.0100.0100.094.4294.2392.7099.0097.6798.43
Table 17. The impact of the ATIS on the fusion results on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
Table 17. The impact of the ATIS on the fusion results on the Pavia University dataset. The best scores are marked in red, and the second best scores are marked in blue.
StrategyPavia University
RMSEPSNRERGASSAMSSIM
one-stage1.77743.1391.3751.9930.986
ATIS1.64543.4911.1311.8990.987
only fusion1.60143.7291.1011.8850.988
Table 18. Classification results of our fusion model 3DCNet and downstream classification model DBDA or 3DCFormer using the ATIS. The best scores are shown in red.
Table 18. Classification results of our fusion model 3DCNet and downstream classification model DBDA or 3DCFormer using the ATIS. The best scores are shown in red.
DatasetOAAAKAPPA
3DCFormerDBDA3DCFormerDBDA3DCFormerDBDA
Pavia University97.1099.0099.1697.6795.5498.43
Pavia Center98.5897.7896.7894.5298.4696.10
Indian Pines89.6290.2780.0782.3791.1788.34
Botswana99.4798.9399.7197.9297.3398.63
Table 19. Model complexity of the proposed 3DCNet and other deep learning models. The best data are marked in red, and the second best data are marked in blue.
Table 19. Model complexity of the proposed 3DCNet and other deep learning models. The best data are marked in red, and the second best data are marked in blue.
ModelUrban Complexity Experiment
Params (M)FLOPs (G)Testing (ms)
SSF1.13637.217174
ConSSF1.16338.119115
TFNet2.50119.891159
ResTFNet2.37618.618111
MSDCNN1.82359.741177
SSRNet0.70923.21941
MSST41.01527.58464
DCFormer5.410318.51294
3DCNet0.70815.92394
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Long, J.; Li, J.; Li, C.; Si, J.; Peng, Y. ATIS-Driven 3DCNet: A Novel Three-Stream Hyperspectral Fusion Framework with Knowledge from Downstream Classification Performance. Remote Sens. 2025, 17, 825. https://doi.org/10.3390/rs17050825

AMA Style

Zhang Q, Long J, Li J, Li C, Si J, Peng Y. ATIS-Driven 3DCNet: A Novel Three-Stream Hyperspectral Fusion Framework with Knowledge from Downstream Classification Performance. Remote Sensing. 2025; 17(5):825. https://doi.org/10.3390/rs17050825

Chicago/Turabian Style

Zhang, Quan, Jian Long, Jun Li, Chunchao Li, Jianxin Si, and Yuanxi Peng. 2025. "ATIS-Driven 3DCNet: A Novel Three-Stream Hyperspectral Fusion Framework with Knowledge from Downstream Classification Performance" Remote Sensing 17, no. 5: 825. https://doi.org/10.3390/rs17050825

APA Style

Zhang, Q., Long, J., Li, J., Li, C., Si, J., & Peng, Y. (2025). ATIS-Driven 3DCNet: A Novel Three-Stream Hyperspectral Fusion Framework with Knowledge from Downstream Classification Performance. Remote Sensing, 17(5), 825. https://doi.org/10.3390/rs17050825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop