Next Article in Journal
Can One Series of Self-Organized Nanoripples Guide Another Series of Self-Organized Nanoripples during Ion Bombardment: From the Perspective of Power Spectral Density Entropy?
Previous Article in Journal
Super-Exponential Growth in Models of a Binary String World
Previous Article in Special Issue
BPDGAN: A GAN-Based Unsupervised Back Project Dense Network for Multi-Modal Medical Image Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GRPAFusion: A Gradient Residual and Pyramid Attention-Based Multiscale Network for Multimodal Image Fusion

1
Optoelectronic System Laboratory, Institute of Semiconductors, Chinese Academy of Sciences, Beijing 100083, China
2
College of Materials Science and Opto-Electronic Technology, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 169; https://doi.org/10.3390/e25010169
Submission received: 30 November 2022 / Revised: 7 January 2023 / Accepted: 12 January 2023 / Published: 14 January 2023
(This article belongs to the Special Issue Advances in Image Fusion)

Abstract

:
Multimodal image fusion aims to retain valid information from different modalities, remove redundant information to highlight critical targets, and maintain rich texture details in the fused image. However, current image fusion networks only use simple convolutional layers to extract features, ignoring global dependencies and channel contexts. This paper proposes GRPAFusion, a multimodal image fusion framework based on gradient residual and pyramid attention. The framework uses multiscale gradient residual blocks to extract multiscale structural features and multigranularity detail features from the source image. The depth features from different modalities were adaptively corrected for inter-channel responses using a pyramid split attention module to generate high-quality fused images. Experimental results on public datasets indicated that GRPAFusion outperforms the current fusion methods in subjective and objective evaluations.

1. Introduction

With the development of image sensor technology, it has become easy to obtain images at infrared wavelengths, which are invisible to the human eye, and at visible wavelengths. Infrared image sensors are noisy and have poor imaging quality; however, they can be used for imaging during the day as well as night. Visible image sensors have high image quality and clear textures but are susceptible to weather and environmental influences that prevent them from working around the clock. Multimodal image fusion is an effective means of image enhancement that can fuse useful information from different modal images into one image while removing redundant information. Fused images are easy for the human eye to see, and help to perform other downstream tasks such as object detection [1,2], object tracking [3], and object segmentation [4,5].
In recent years, researchers have proposed many traditional image fusion methods, including multiscale transform-based [6,7,8,9,10], sparse representation-based [11,12,13], and saliency map-based [14,15,16] methods. The multiscale transformation-based method transforms the original image according to the multiscale analysis method. The obtained transform coefficients are fused, and the corresponding inverse transform is performed to obtain the final fused image. The sparse representation-based method must construct a complete dictionary and then reconstruct the fused image according to the set fusion rules. The saliency map-based approach extracts a saliency map using saliency detection methods, and fuses the salient regions with other regions according to the fusion strategy. The fusion rules of these traditional methods are manually designed, and are computationally complex and time-consuming. In addition, these traditional methods do not consider the differences between different modalities, and use the same feature extraction method to extract the image features of different modalities, resulting in a large amount of redundant information in the fusion results, and poor fusion performance.
With the development of deep learning, researchers have proposed the use of convolutional neural networks (CNNs) for image fusion. The general steps of the deep learning-based method include feature extraction, fusion, and reconstruction. These methods include CNN-based [17,18], generative adversarial network (GAN)-based [19], and autoencoder (AE)-based [20] methods. Earlier, CNN-based methods used convolutional neural networks to only extract features, and they used traditional methods for image fusion, which is inefficient for fusion. Li et al. [21] proposed an end-to-end model for infrared and visible image fusion. The unsupervised network architecture avoided the use of fusion layers. However, this method used the same feature extraction layer to extract image features of different modes, resulting in an unbalanced feature representation of the reconstructed image to the original image. For example, the fused image is rich in texture information but lacks thermal-radiation information. Most autoencoder-based methods use manually designed fusion layers for feature fusion after feature extraction, and these manually designed fusion strategies limit fusion performance. The GAN-based approach makes it difficult to strike a balance between the generator and the discriminator, owing to the lack of ideal fusion images. To avoid this problem, Li et al. [22] proposed a dual discriminator network based on an attention mechanism. Instead of using the ideal fused image as the discriminator judgment criterion, this method used the saliency regions of different modal images as the discriminator criterion. This training method that only cared about local information made the fused images with artifacts and noise. Most current image fusion networks use only successive simple convolution operations for feature extraction and reconstruction, ignoring global dependencies and interchannel contextual information. SwinFusion [23], proposed by Ma et al. introduced the Swin transformer into image fusion to make full use of local and global information, and it achieved excellent fusion performance. However, the large number of parameters of transformer-based models brings a large computational overhead. Furthermore, multiscale feature extraction is a common means of obtaining global information. Frequently used methods include convolution with large convolution kernels or pooling operations with large strides. The convolution operation with large convolution kernels leads to an increase in computational effort and unavoidable block effects. The pooling operation causes a decrease in the image resolution, resulting in the loss of many vital details for image fusion tasks. Multi-task fusion frameworks have also made some progress in recent years. It is worth mentioning that SuperFusion [24], proposed by Tang et al., can implement image registration, image fusion, and segmentation tasks using only one framework.
This article proposes a multiscale image fusion framework called GRPAFusion, based on gradient residuals and pyramidal attention, to overcome the above problems. The framework is an end-to-end fusion network capable of extracting the features of different modalities based on the input image, thus performing adaptive image fusion. The fusion network is based on the encoder and decoder structures. The decoder uses a multiscale gradient residual (MGR) block to obtain more fine-grained multiscale information, with little computational effort, by increasing the receptive field. The pyramid split attention (PSA) module can acquire contextual information between the channels. The PSA module adaptively recalibrates the feature response between the channels, based on the dependencies between channels, to produce fused features that contain both infrared radiation and visible details. Adequate ablation and comparison experiments show that the proposed framework achieves the best fusion performance in subjective and objective evaluations.
To visually demonstrate the excellent performance of the fusion framework proposed in this study, Figure 1 shows the comparison effect of the proposed method with the AE-based method DenseFuse and the GAN-based method FusionGAN fused images. The two images on the left are the visible and infrared images, and the middle image and the two images on the right are the fusion results of DenseFuse, FusionGAN, and GRPAFusion, respectively. The thermal-radiation information of the infrared image is not obvious in the fusion results of DenseFuse. The fused images of FusionGAN have considerable noise in the image due to an excessive focus on the infrared thermal information. In contrast, the fusion result of GRPAFusion proposed in this study includes thermal-radiation information in the infrared image, and preserves texture information in the visible image. The main contributions of this study are as follows:
  • We propose an end-to-end multimodal image fusion framework that can adaptively fuse information from different modalities without human intervention. The training process of the fusion network is constrained using adaptive pixel intensity loss and maximum gradient loss, such that the fused image can highlight the thermal-radiation information of the infrared image while retaining the texture details of the visible image effectively.
  • The fusion framework proposed in this study extracts the multiscale features of the source images while keeping the computational effort small and avoiding the use of downsampling operations. The detailed residuals of multiple granularities are used together in the encoder to represent the joint residual information, effectively preventing the problem of detailed feature loss and gradient disappearance. The pyramid split attention module is introduced in the feature fusion to make the fusion network pay more attention to the thermal-radiation information and texture details in the original image, whereas the fused image has higher contrast.
  • Adequate ablation and comparative experimental studies were conducted using the TNO dataset. The results of the ablation experiments show that the fusion framework and trained loss function proposed in this study are effective. The subjective and objective experimental results of the comparative experiments show that the multimodal image fusion framework proposed in this paper has superior fusion performance compared to current state-of-the-art image fusion methods.
The remainder of this paper is organized as follows. Section 2 describes the fusion framework in detail. Section 3 explains the experiments and discussions. Finally, Section 4 concludes the study.

2. Proposed Method

This section describes the multiscale image fusion framework based on gradient residual and pyramid attention in detail. The general structure of the fusion network is first described, then the loss functions we use and their functions are explained, and the evaluation method of the fused image quality is presented.

2.1. Network Architecture

The overall structure of the proposed fusion framework is shown in Figure 2, and it includes three parts: encoder, feature fusion, and decoder. In the image fusion process, the encoder first extracts the input image with multiscale depth features. It is fused with multiscale features using the PSA [25] module of the feature fusion layer, and finally, the final fused image is obtained after feature reconstruction via successive 3 × 3 convolutions.
Encoder part: We used 1 × 1 convolution and two MGR blocks to extract the image features. The structure of the MGR block is illustrated in Figure 3. The MGR block is divided into two branches. The upper part is called the detailed branch, and the lower is called the structure branch. The detailed branch is designed to extract multigranularity detail information, whereas the structure branch is designed to extract multiscale structure information. We used the original feature maps in the detailed branch as coarse-grained features. The fine-grained detail features in the image were extracted using the Sobel gradient operator. Then, after adjusting the feature channels using the 1 × 1 convolution method, the feature maps were added to the multiscale feature maps element-wise. Assuming that the initial feature map is F, the output of the detailed branch F d e t i a l is expressed as follows:
F d e t i a l = C o n v 1 × 1 F C o n v 1 × 1 F ,
where C o n v 1 × 1 ( · ) represents the convolutional layer with a convolutional kernel size of 1, ⊕ stands for element-wise summation, and is the Sobel gradient operator. Inspired by Res2Net [26], in the structure branch, we acquired multiscale features by expanding the field of perception. The original feature map is first sliced into s feature groups according to scale s, denoted by x i , where i 1 , 2 , . . . , s . In the method described in this article, s is 4. The feature map of the first group is directly output. The feature map of the second group is output after 3 × 3 convolution. The feature map of the third group is superimposed on the feature map of the second group, which is output after 3 × 3 convolution. After sequentially completing the operation of s groups of feature channels, a set of feature maps with multiple scales is obtained by cascading the s groups of feature maps between channels. Thus, the multiscale feature maps F m i can be denoted as
F m i = x i i = 1 ; C o n v 3 × 3 x i i = 2 ; C o n v 3 × 3 x i F m i 1 2 < i s ,
where C o n v 3 × 3 ( · ) represents the convolutional layer with a convolutional kernel size of 3. The output of the structure branch F s t r u c t u r e is expressed as follows:
F s t r u c t u r e = C o n v 1 × 1 C o n c a t F m i , i 1 , 2 , . . . , s ,
where C o n c a t ( · ) denotes the concatenate operation on the channel dimension. Therefore, the output of the MGR block is expressed as follows:
O u t M G R = ReLU F d e t i a l F s t r u c t u r e ,
where ReLU ( · ) represents the linear rectification function. Two parallel channels are used in the encoder to extract the infrared and visible image features, respectively. The features of the two parallel channels are inter-channel cascaded to obtain the final output of the encoder.
Feature fusion part: We use the PSA module in the feature fusion layer for multiscale feature fusion. The structure of the PSA module is shown in Figure 4. PSA uses four convolutional layers to divide the input fused feature map F into four feature subsets with a convolutional kernel size of K = { 3 , 5 , 7 , 9 } . The number of channels is consistent for each feature subset. PSA uses group convolution to reduce the computational effort. The number of groups are G = { 1 , 4 , 8 , 16 } . The pyramid feature map generation function is given by
F p y r a m i d i = C o n v K i G i F , i 1 , 2 , 3 , 4 ,
where C o n v K i G i ( · ) represents the convolution operation with a convolution kernel size of K i and the number of groups G i . The results are then passed through SEWeight [27], and the channel is cascaded to obtain a multiscale channel attention map W a t t , which can be denoted as
W a t t = C o n c a t S E W e i g h t F p y r a m i d i ,
where S E W e i g h t ( · ) is used to obtain the attention weight from the input feature map. The interaction between global and local attention is implemented by SoftMax to obtain the final attention weights. Therefore, the output of the PSA module is expressed as follows:
O u t P S A = F Softmax ( W a t t ) ,
where ⊗ represents the channel-wise multiplication, and Softmax ( · ) is used to recalibrate the attention weight. The PSA module achieves an adaptive fusion of infrared and visible image features by adjusting the response between the feature fusion channels.
Decoder part: Decoder reconstructs the fused features to obtain the final fused image. Our decoder network consists of four convolutional layers. The convolutional kernels are all 3 × 3 in size, with a step size of 1. To avoid information loss, we obtain the final fused image via the successive convolutional adjustment of the feature channels without any up-sampling or down-sampling operations. This design allows our fusion network to accommodate image inputs of any size resolution, and feature maps and fused images can be output with the same resolution.

2.2. Loss Function

Infrared and visible image fusion is an enhancement method that aims to obtain a visible image that contains infrared thermal-radiation information and rich texture details. Therefore, this study proposes the use of content loss and detail loss as joint losses in the training phase to guide the optimization of network parameters. The total loss function is expressed as follows:
L t o t a l = L c o n t e n t + L d e t i a l .
Content loss is used to calculate the pixel intensity error between the fused image and the input image, and detail loss is used to calculate the edge texture difference between the fused image and the input image, which are, respectively, expressed as follows:
L c o n t e n t = 1 H W I f α I i r + β I v i s 1 ,
L d e t i a l = 1 H W I f max I i r , I v i s 1 ,
where I f indicates the fused image, I i r indicates the infrared image, H denotes the height of the image, W denotes the width of the image, α and β are hyperparameters, ∇ represents the Sobel gradient operation, and m a x ( · ) represents the maximum selection operation. · 1 represents the L1-norm, and the calculation is expressed as follows:
I 1 = i H j W I i , j .
The maximum selection strategy is used for detail loss because the texture-detail information in the fused image must be the maximum set of texture details in the infrared and visible images. We intend that the proposed network adaptively selects the pixel intensities of the infrared image, and the visible image be displayed in the fusion results, so α and β are used to adjust the ratio of the infrared image to the visible image content. Therefore, appropriately selecting α and β can enable the fusion network to have a better fusion effect, and this part will be described in detail in the ablation experiment.

2.3. Evaluation of the Fusion Results

Subjective and objective methods are used for evaluating the fusion image quality. The subjective evaluation method is based on the visual perception of the human eye. The evaluation criterion is whether the fused image contains valid information from the original image and removes redundant information. The subjective evaluation method is highly random, and the evaluator cannot accurately distinguish minor differences between the fused and source images. Thus, in this study, we used a combination of subjective and objective evaluations to assess the quality of fused images. Eight objective evaluation indices were selected from different perspectives to evaluate the fused images objectively. The evaluation metrics based on fused images include information entropy (EN) [28], spatial frequency (SF) [28], and average gradient (AG) [29]. The evaluation metrics based on the fused and original images include fusion quality ( Q a b f ) [30], pixel feature mutual information ( FMI p i x e l ), discrete cosine transform feature mutual information ( FMI d c t ), wavelet feature mutual information ( FMI w ) [31], and multiscale structural similarity (MS-SSIM) [32]. These evaluation metrics can effectively reflect the ability of the fusion network to fuse visual information, structural information, and detailed texture information. Larger values of evaluation metrics indicate a better performance of the fusion method.

3. Experiments and Discussion

This section provides the experimental validation of the performance of the proposed fusion framework. First, the detailed parameters for training and testing are presented. Then, an ablation study is conducted, focusing on the effects of the choice of hyperparameters, and the network structure on the fusion performance. Finally, we evaluate the proposed fusion method qualitatively and quantitatively against current advanced fusion methods.

3.1. Experiments Setting

3.1.1. Dataset

TNO [33] is an authoritative image fusion dataset, and we selected 21 representative pairs of scenes from the TNO dataset for ablation studies and comparison experiments. The data volume of the TNO dataset is small, to support the network’s training phase. Therefore, we selected LLVIP [34], which contains 14,588 pairs of infrared and visible images, to train the network.

3.1.2. Train Details

In the training phase, we adjusted the resolution of the input image to 256 × 256 , to speed up the training efficiency. The fusion network was trained on a PC containing an AMD R7-5800X 3.4 GHz CPU, 32 GB of RAM, and an NVIDIA GTX 3080Ti GPU. We used the deep learning framework Pytorch 1.8 and the Adam [35] optimizer to converge the loss function values to the minimum. The initial learning rate was 1e-4, the epoch was 4, and the batch size was 4.

3.1.3. Test Details

We selected two traditional and four deep-learning-based fusion methods for comparison with our proposed method: ADF [6], MSVD [7], DLF [18], FusionGAN [19], DenseFuse [20], U2Fusion [36], SuperFusion [24], and SwinFusion [23]. Traditional methods were implemented using the MATLAB toolbox, and deep learning-based image fusion methods were implemented using publicly available source code. To ensure the fairness of the experiments, we tuned these methods to achieve the best performance according to the parameters recommended in the references.

3.2. Ablation Study

In the ablation study, we first performed a sensitivity-analysis experiment on the setting of hyperparameters, and determined the optimal parameter setting. Then, the ablation experiments of the network structure were performed to analyze the effects of the gradient residual and the PSA module on the network performance.

3.2.1. Experimental Validation of Parameter Sensitivity Analysis

We use α and β as hyperparameters in the content loss function to adjust the ratio of infrared and visible image content. There are two ways to calculate α and β , based on global average pooling and global level map. The calculation based on the global level map method is as follows:
α x , y = I i r x , y I i r x , y + I v i s x , y ,
β x , y = I v i s x , y I i r x , y + I v i s x , y .
The calculation based on the global average pooling method is as follows:
S u = 1 H W x H y W I u x , y ,
α = S i r S i r + S v i s , β = S v i s S i r + S v i s ,
where I denotes the original image, and I ( x , y ) represents any pixel point in the image. With the network structure and other parameters being kept constant, we trained the network separately using the hyperparameters calculated via each of the above two methods. Figure 5 shows the representative results of the subjective experiments. The subjective experimental results show that the fusion results obtained using the network trained with the global level map-based approach to calculate hyperparameters are more prominent in the infrared thermal-radiation information and they avoid introducing additional artifacts. For example, in Kaptain_1123, the edges of the trees on the roof are sharper and free of artifacts, and the people on the bench are more prominent.
The average values of the objective evaluation metrics for 21 representative scenarios on the TNO dataset are listed in Table 1, with the best results in bold. The fusion results of the global-level map-based method outperformed those obtained using the global average pooling-based method for five of the eight objective evaluation metrics. In terms of the objective evaluation metrics, the fusion performances of the models obtained from the training of the global-level map-based approach were better. The hyperparameters chosen in the subsequent experiments in this study were calculated based on a global-level map.

3.2.2. Experimental Validation of Network Architecture

The fusion framework proposed in this study uses a gradient residual to provide more fine-grained, detailed information for the feature extraction part. It uses the PSA module as a feature fusion layer instead of the concatenate operation. The ablation study was conducted to analyze the impact of these two structures on the network performance. To verify the validity of the proposed network structure, we separately compared GRPAFusion with the model that excluded each of the two structures. The representative results of the subjective experiments are shown in Figure 6. In soldier_behind_smoke, fine-grained detailed features, such as the soldier’s body contour, are noticeably missing from the network fusion results without gradient residuals. In heather, the edges of the fence are clear and have less noise and artifacts in the fusion result of GRPAFusion. The fusion results using the PSA module as the fusion layer demonstrated higher contrast, and more realistic and natural images, because the PSA module can adaptively adjust the weight of infrared thermal-radiation and texture details during feature fusion.
The objective results of the network-structure validation experiments are presented in Table 2. Our proposed fusion network is optimal in all three structures in terms of metrics, except for EN and FMI w . The results of the subjective and objective experiments indicate that the structure of the fusion network proposed in this paper is effective. The results of the ablation study indicate that the gradient residual provides fine-grained detail features, and that the PSA module enables the fusion network to achieve adaptive image fusion.

3.3. Comparative Experiment

We conducted extensive comparative experiments on the TNO dataset to verify the excellent performance of the GRPAFusion. Figure 7 and Figure 8 show the representative results of the subjective experiments, from which we selected five representative scenes to demonstrate the excellent fusion effect of GRPAFusion. Rectangular boxes were used to mark key locations in the results. The red boxes mark the infrared thermal information that require attention, and the green boxes mark the texture details that require attention. In Figure 7, ABF, MSVD, DLF, DenseFuse, U2Fusion, and SuperFusion cannot clearly highlight the salient targets in infrared images, such as pedestrians on the road, with low luminance. FusionGAN can retain infrared thermal information; however, the fusion results contain considerable noise, resulting in blurred images. U2Fusion can preserve texture-detail information well; however, its ability to highlight infrared thermal information is weak. In the billboard of the Street scene, all comparison methods except U2Fusion and SwinFusion showed blurred text, whereas our method showed clear and sharp text with high contrast and a better fusion effect. In the Meeting scenario in Figure 8, artifacts appear to vary in degrees in the compared fusion methods, whereas our method avoids artifacts and provides a more realistic and natural fusion result. In the Ship scenario shown in Figure 8, the outline of the ship’s windows is significantly clearer in the fused image of GRPAFusion. The subjective experimental results indicate that GRPAFusion has better fusion performance, and the fused images can highlight the infrared saliency targets while retaining rich texture details. In the Street scene, GRPAFusion achieves the best subjective experimental result. However, the subjective experimental results in the Nato_camp and Sandpath scenes demonstrate that SwinFusion and GRPAFusion have similar fusion performances. To further compare the differences between the fusion performances of the comparison methods and GRPAFusion, we calculated the objective evaluation metrics for these methods.
Figure 9 shows the objective evaluation results of the above fusion methods on the TNO dataset, where the solid red pentagrams represent the GRPAFusion calculations. As can be observed in the graph, the fusion results of the proposed method are significantly better than those of other methods on EN, SF, AG, and Q a b f . To evaluate its performance more intuitively, we calculated the average values of eight objective evaluation metrics for different fusion methods on the TNO dataset. The results are shown in Table 3, with the best results shown in bold and the second-best results underlined. The proposed method achieves the optimum in three objective metrics, EN, SF, and AG, indicating that GRPAFusion performs better in fusing effective information and preserving source image details. Suboptimal results were achieved for Q a b f and FMI p i x e l , indicating that GRPAFusion can better preserve significant information in the source images. Although FMI d c t , FMI w , and MS-SSIM did not achieve the optimum, GRPAFusion achieved optimality or suboptimality in most objective evaluation metrics. Therefore, combining the results of subjective and objective experiments, the fusion performance of GRPAFusion is better than those of other comparative methods.

4. Conclusions

In this study, we proposed an efficient end-to-end multimodal image fusion framework in which the trained network can adaptively fuse multiple modal images without human intervention. The fusion framework consists of three parts: encoder, feature fusion layer, and decoder. The source image is passed through the encoder section for multiscale and multi-grained feature extraction. The features of different modalities are fused and fed into the decoder for feature reconstruction to obtain the final fused image. In the training phase, we propose the use of content loss and detail loss to guide the convergence direction of the fusion network to make the final fused image have rich texture details and high contrast. The encoder uses an MGR block to extract multi-grained detail features and multiscale structural features, which can preserve the texture details of the source image. In addition, this study also introduced the PSA module instead of the simple channel cascade as the fusion layer, which adaptively fuses the features of different modes by readjusting the responses of different feature channels. Finally, the number of channels was adjusted using successive convolutions to obtain the fused image. The results of the ablation and comparison experiments on the TNO dataset indicated that GRPAFusion performs better than the state-of-the-art infrared and visible-image fusion methods. In the future, we will investigate better-performing modules to improve objective evaluation metrics, and we will further focus on how to use image fusion to drive advanced vision tasks.

Author Contributions

The work presented here was carried out in collaboration between all authors. J.W. and X.X. performed the experiments and wrote the draft. G.Z. gave professional guidance and provided editing. D.L. and F.L. reviewed and edited the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, J.; Bhanu, B. Fusion of Color and Infrared Video for Moving Human Detection. Pattern Recognit. 2007, 40, 1771–1784. [Google Scholar] [CrossRef]
  2. Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image Fusion Meets Deep Learning: A Survey and Perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  3. Tu, Z.; Pan, W.; Duan, Y.; Tang, J.; Li, C. RGBT Tracking via Reliable Feature Configuration. Sci. China Inf. Sci. 2022, 65, 142101. [Google Scholar] [CrossRef]
  4. Tang, L.; Yuan, J.; Ma, J. Image Fusion in the Loop of High-Level Vision Tasks: A Semantic-Aware Real-Time Infrared and Visible Image Fusion Network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  5. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards Real-Time Semantic Segmentation for Autonomous Vehicles with Multi-Spectral Scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar] [CrossRef]
  6. Bavirisetti, D.P.; Dhuli, R. Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sensors J. 2015, 16, 203–209. [Google Scholar] [CrossRef]
  7. Naidu, V. Image fusion technique using multi-resolution singular value decomposition. Def. Sci. J. 2011, 61, 479. [Google Scholar] [CrossRef] [Green Version]
  8. Shreyamsha Kumar, B. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015, 9, 1193–1204. [Google Scholar] [CrossRef]
  9. Zhou, Z.; Dong, M.; Xie, X.; Gao, Z. Fusion of infrared and visible images for night-vision context enhancement. Appl. Opt. 2016, 55, 6480–6490. [Google Scholar] [CrossRef]
  10. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  11. Z, L.Y.L.S.W. A General Framework for Image Fusion Based on Multi-Scale Transform and Sparse Representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  12. Liu, Y.; Chen, X.; Ward, R.K.; Wang, Z.J. Image Fusion With Convolutional Sparse Representation. IEEE Signal Process. Lett. 2016, 23, 1882–1886. [Google Scholar] [CrossRef]
  13. Liu, C.H.; Qi, Y.; Ding, W.R. Infrared and Visible Image Fusion Method Based on Saliency Detection in Sparse Domain. Infrared Phys. Technol. 2017, 83, 94–102. [Google Scholar] [CrossRef]
  14. Bavirisetti, D.P.; Dhuli, R. Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 2016, 76, 52–64. [Google Scholar] [CrossRef]
  15. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  16. Han, J.; Pauwels, E.J.; De Zeeuw, P. Fast saliency-aware multi-modality image fusion. Neurocomputing 2013, 111, 70–80. [Google Scholar] [CrossRef]
  17. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and Visible Image Fusion with Convolutional Neural Networks. Int. J. Wavelets Multiresolution Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  18. Li, H.; Wu, X.J.; Kittler, J. Infrared and Visible Image Fusion Using a Deep Learning Framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2705–2710. [Google Scholar] [CrossRef] [Green Version]
  19. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A Generative Adversarial Network for Infrared and Visible Image Fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  20. Li, H.; Wu, X.J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [Green Version]
  21. Li, J.; Liu, J.; Zhou, S.; Zhang, Q.; Kasabov, N.K. Infrared and Visible Image Fusion Based on Residual Dense Network and Gradient Loss. Infrared Phys. Technol. 2023, 128, 104486. [Google Scholar] [CrossRef]
  22. Li, J.; Huo, H.; Li, C.; Wang, R.; Feng, Q. AttentionFGAN: Infrared and Visible Image Fusion Using Attention-Based Generative Adversarial Networks. IEEE Trans. Multimed. 2021, 23, 1383–1396. [Google Scholar] [CrossRef]
  23. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-Domain Long-Range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  24. Tang, L.; Deng, Y.; Ma, Y.; Huang, J.; Ma, J. SuperFusion: A Versatile Image Registration and Fusion Network with Semantic Awareness. IEEE/CAA J. Autom. Sin. 2022, 9, 2121–2137. [Google Scholar] [CrossRef]
  25. Zhang, H.; Zu, K.; Lu, J.; Zou, Y.; Meng, D. EPSANet: An efficient pyramid squeeze attention block on convolutional neural network. arXiv 2021, arXiv:2105.14447. [Google Scholar]
  26. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2Net: A New Multi-Scale Backbone Architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef] [Green Version]
  27. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  28. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar]
  29. Cui, G.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 2015, 341, 199–209. [Google Scholar] [CrossRef]
  30. Piella, G.; Heijmans, H. A New Quality Metric for Image Fusion. In Proceedings of the 2003 International Conference on Image Processing (Cat. No.03CH37429), Barcelona, Spain, 14–17 September 2003; Volume 3, p. III-173. [Google Scholar] [CrossRef]
  31. Haghighat, M.; Razian, M.A. Fast-FMI: Non-reference image fusion metric. In Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan, 15–17 October 2014; pp. 1–3. [Google Scholar]
  32. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  33. Toet, A. The TNO Multiband Image Data Collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef]
  34. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 19–25 June 2021; pp. 3496–3504. [Google Scholar]
  35. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  36. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef]
Figure 1. Example fusion results for FusionGAN, DenseFuse, and our GRPAFusion. GRPAFusion can retain both texture-detail information in the visible image and thermal-radiation information in the infrared image.
Figure 1. Example fusion results for FusionGAN, DenseFuse, and our GRPAFusion. GRPAFusion can retain both texture-detail information in the visible image and thermal-radiation information in the infrared image.
Entropy 25 00169 g001
Figure 2. Overall structure of the fusion framework proposed in this article. The framework includes an encoder, feature fusion layer, and decoder.
Figure 2. Overall structure of the fusion framework proposed in this article. The framework includes an encoder, feature fusion layer, and decoder.
Entropy 25 00169 g002
Figure 3. Structure of the MGR block. The MGR block is divided into two branches. The upper dashed box is the detailed branch used to extract multi-granularity detail features. The lower dashed box is the structure branch used to extract multiscale structural features.
Figure 3. Structure of the MGR block. The MGR block is divided into two branches. The upper dashed box is the detailed branch used to extract multi-granularity detail features. The lower dashed box is the structure branch used to extract multiscale structural features.
Entropy 25 00169 g003
Figure 4. Structure of the PSA module. The PSA module uses convolution kernels of different sizes and group convolution to obtain multiscale attention maps, and adaptive fusion of different modal images is achieved by adjusting the response between feature fusion channels.
Figure 4. Structure of the PSA module. The PSA module uses convolution kernels of different sizes and group convolution to obtain multiscale attention maps, and adaptive fusion of different modal images is achieved by adjusting the response between feature fusion channels.
Entropy 25 00169 g004
Figure 5. Representative results of the qualitative evaluation of the parameter sensitivity-analysis experiment. The scene on the left is Kaptain_1123, and that on the right is Bench. From top to bottom are the visible images, the infrared images, the fusion results based on the global average pooling method, and the fusion results based on the global level map method, respectively. The red boxes are used to mark key locations in the results.
Figure 5. Representative results of the qualitative evaluation of the parameter sensitivity-analysis experiment. The scene on the left is Kaptain_1123, and that on the right is Bench. From top to bottom are the visible images, the infrared images, the fusion results based on the global average pooling method, and the fusion results based on the global level map method, respectively. The red boxes are used to mark key locations in the results.
Entropy 25 00169 g005
Figure 6. Representative results of the qualitative evaluation of the network-architecture validation experiment. The scene at the top is soldier_behind_smoke, and the scene at the bottom is heather. From left to right are visible images, infrared images, fusion results without gradient residual, fusion results without the PSA module, and our GRPAFusion, respectively. The red boxes are used to mark key locations in the results.
Figure 6. Representative results of the qualitative evaluation of the network-architecture validation experiment. The scene at the top is soldier_behind_smoke, and the scene at the bottom is heather. From left to right are visible images, infrared images, fusion results without gradient residual, fusion results without the PSA module, and our GRPAFusion, respectively. The red boxes are used to mark key locations in the results.
Entropy 25 00169 g006
Figure 7. Subjective experimental results of comparison experiments on Street of the TNO dataset. Rectangular boxes were used to mark key locations in the results. The red boxes mark the infrared thermal information that requires attention, and the green boxes mark the texture details that require attention.
Figure 7. Subjective experimental results of comparison experiments on Street of the TNO dataset. Rectangular boxes were used to mark key locations in the results. The red boxes mark the infrared thermal information that requires attention, and the green boxes mark the texture details that require attention.
Entropy 25 00169 g007
Figure 8. Subjective experimental results of comparison experiments on Nato_camp, Ship, Meeting, and Sandpath of the TNO dataset. Rectangular boxes are used to mark key locations in the results. The red boxes mark the infrared thermal information that requires attention, and the green boxes mark the texture details that require attention.
Figure 8. Subjective experimental results of comparison experiments on Nato_camp, Ship, Meeting, and Sandpath of the TNO dataset. Rectangular boxes are used to mark key locations in the results. The red boxes mark the infrared thermal information that requires attention, and the green boxes mark the texture details that require attention.
Entropy 25 00169 g008
Figure 9. The results of eight objective evaluation metrics calculated on 21 pairs of representative images from the TNO dataset.
Figure 9. The results of eight objective evaluation metrics calculated on 21 pairs of representative images from the TNO dataset.
Entropy 25 00169 g009
Table 1. Objective experimental results for different parameter settings on the TNO dataset; the best results are presented in bold.
Table 1. Objective experimental results for different parameter settings on the TNO dataset; the best results are presented in bold.
ModelENSFAG Q abf FMI pixel FMI dct FMI w MS-SSIM
Global Average Pool6.64570.04704.66670.47780.90610.22080.25970.9289
Global Level Map6.81600.04644.67350.48180.90800.22040.26580.8880
Table 2. Objective experimental results for different network architectures on the TNO dataset; the best results are presented in bold.
Table 2. Objective experimental results for different network architectures on the TNO dataset; the best results are presented in bold.
ModelENSFAG Q abf FMI pixel FMI dct FMI w MS-SSIM
No-GradRes6.86770.04534.37470.46690.90480.21710.27440.8649
No-PSA6.84300.04604.58560.47560.90600.21530.26220.8730
Ours6.81600.04644.67350.48180.90800.22040.26580.8880
Table 3. Objective evaluation results of comparison experiments on the TNO dataset, with the best results shown in bold and the second-best underlined.
Table 3. Objective evaluation results of comparison experiments on the TNO dataset, with the best results shown in bold and the second-best underlined.
MethodENSFAG Q abf FMI pixel FMI dct FMI w MS-SSIM
ADF6.26910.03453.52170.41270.88290.22750.25950.8760
MSVD6.60880.03703.27040.30610.89600.17500.21910.8933
DLF6.18550.02442.35870.29650.90330.19750.22510.8655
DenseFuse6.17000.02192.21000.26270.89650.17670.21550.8603
FusionGAN6.36000.02152.11730.19050.88890.17270.19570.7245
U2Fusion6.24930.03133.33980.36550.88960.20540.23730.8931
SuperFusion6.61800.03163.03430.41380.89610.27010.30310.8566
SwinFusion6.70070.04083.96280.49870.91020.29450.32670.8844
Ours6.81600.04644.67350.48180.90800.22040.26580.8880
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Xi, X.; Li, D.; Li, F.; Zhang, G. GRPAFusion: A Gradient Residual and Pyramid Attention-Based Multiscale Network for Multimodal Image Fusion. Entropy 2023, 25, 169. https://doi.org/10.3390/e25010169

AMA Style

Wang J, Xi X, Li D, Li F, Zhang G. GRPAFusion: A Gradient Residual and Pyramid Attention-Based Multiscale Network for Multimodal Image Fusion. Entropy. 2023; 25(1):169. https://doi.org/10.3390/e25010169

Chicago/Turabian Style

Wang, Jinxin, Xiaoli Xi, Dongmei Li, Fang Li, and Guanxin Zhang. 2023. "GRPAFusion: A Gradient Residual and Pyramid Attention-Based Multiscale Network for Multimodal Image Fusion" Entropy 25, no. 1: 169. https://doi.org/10.3390/e25010169

APA Style

Wang, J., Xi, X., Li, D., Li, F., & Zhang, G. (2023). GRPAFusion: A Gradient Residual and Pyramid Attention-Based Multiscale Network for Multimodal Image Fusion. Entropy, 25(1), 169. https://doi.org/10.3390/e25010169

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop