Next Article in Journal
Fractional Calculus for Neutrosophic-Valued Functions and Its Application in an Inventory Lot-Sizing Problem
Previous Article in Journal
Pore Structure and Multifractal Characteristics of the Upper Lianggaoshan Formation in the Northeastern Sichuan Basin, China
Previous Article in Special Issue
Multi-Focus Image Fusion Based on Fractal Dimension and Parameter Adaptive Unit-Linking Dual-Channel PCNN in Curvelet Transform Domain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Focus Image Fusion Based on Dual-Channel Rybak Neural Network and Consistency Verification in NSCT Domain

1
School of Computer Science and Technology, Xinjiang University, Urumqi 830046, China
2
Key Laboratory of Signal Detection and Processing, Xinjiang University, Urumqi 830046, China
3
College of Mathematics and System Science, Xinjiang University, Urumqi 830046, China
4
School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
5
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Authors to whom correspondence should be addressed.
Fractal Fract. 2025, 9(7), 432; https://doi.org/10.3390/fractalfract9070432
Submission received: 6 May 2025 / Revised: 17 June 2025 / Accepted: 29 June 2025 / Published: 30 June 2025

Abstract

In multi-focus image fusion, accurately detecting and extracting focused regions remains a key challenge. Some existing methods suffer from misjudgment of focus areas, resulting in incorrect focus information or the unintended retention of blurred regions in the fused image. To address these issues, this paper proposes a novel multi-focus image fusion method that leverages a dual-channel Rybak neural network combined with consistency verification in the nonsubsampled contourlet transform (NSCT) domain. Specifically, the high-frequency sub-bands produced by NSCT decomposition are processed using the dual-channel Rybak neural network and a consistency verification strategy, allowing for more accurate extraction and integration of salient details. Meanwhile, the low-frequency sub-bands are fused using a simple averaging approach to preserve the overall structure and brightness information. The effectiveness of the proposed method has been thoroughly evaluated through comprehensive qualitative and quantitative experiments conducted on three widely used public datasets: Lytro, MFFW, and MFI-WHU. Experimental results show that our method consistently outperforms several state-of-the-art image fusion techniques, including both traditional algorithms and deep learning-based approaches, in terms of visual quality and objective performance metrics ( Q A B / F , Q C B , Q E , Q F M I , Q M I , Q M S E , Q N C I E , Q N M I , Q P , and Q P S N R ). These results clearly demonstrate the robustness and superiority of the proposed fusion framework in handling multi-focus image fusion tasks.

1. Introduction

Multi-source image fusion [1,2] refers to the technology of comprehensively processing multiple images acquired from different sensors, modalities, or time points to generate a fused image containing richer and more comprehensive information. This technique is widely applied in fields such as remote sensing [3,4], medical imaging [5,6,7], computer vision [8,9,10], and military surveillance [11,12,13].
In real-world imaging scenarios, optical devices often struggle to capture an entirely clear scene in a single shot due to inherent limitations in depth of field (DoF) [14]. This results in multi-focus images, where different regions of the same scene are selectively focused while others remain blurred. Multi-focus image fusion (MFIF) addresses this challenge by integrating complementary information from a series of partially focused images to synthesize a comprehensive, fully focused representation [15,16]. Such fused images are critical for enhancing visual quality and enabling high-level tasks, including medical diagnosis, microscopic imaging, surveillance systems, and autonomous robotics, where precision in detail is paramount [17,18].
Multi-focus image fusion algorithms can be categorized based on their core techniques and implementation approaches as follows: transform domain-based methods, spatial domain-based methods, deep learning-based methods, and hybrid methods [19,20]. In terms of transform domain, these methods transform input images into frequency or multi-scale domains, fuse coefficients (e.g., maximum selection, weighted averaging) to retain focused regions, and reconstruct the fused image; these methods include Laplacian pyramid, discrete wavelet transform, curvelet, contourlet, shearlet, etc. [21,22]. Li et al. [23] introduced the image fusion technique via sparse representation and guided filtering in the Laplacian pyramid domain, this algorithm demonstrates outstanding performance in infrared–visible image fusion and multi-focus image fusion. Li et al. [24] introduced the multi-focus image fusion approach using fractal dimension in curvelet domain. Lv et al. [25] introduced the image fusion via distance-weighted regional energy and structure tensor in the nonsubsampled contourlet transform (NSCT) domain; this algorithm has achieved remarkable results in both subjective and objective metrics for image fusion. Li et al. [26] proposed the image fusion approach via local energy in the shearlet domain; this algorithm demonstrates favorable performance in multi-focus image fusion. Nevertheless, the shearlet transform necessitates equal height and width for input images, thereby constraining its applicability to image fusion tasks involving varying sizes. The algorithms demonstrate robust multi-scale decomposition capabilities while achieving superior performance in edge and texture preservation. However, these methods are subject to certain limitations: they may introduce potential artifacts (particularly Gibbs oscillations), and their fusion processes often rely on heuristic rule design rather than theoretically optimized approaches [27,28,29].
Fractal and fractional methods have been widely applied in image fusion. For instance, Li et al. [15] developed a multi-focus image fusion method using fractal dimension within the NSCT domain, implemented via coupled neural P systems. Panigrahy et al. [30] presented a parameter-adaptive dual-channel PCNN model for multi-focus image fusion, incorporating fractal dimension. Joshua et al. [31] introduced an adaptive low-light image enhancement algorithm combining a novel intuitionistic fuzzy generator with fractal–fractional derivatives. Xian et al. [32] proposed a multi-focus fusion approach leveraging visual depth and fractional-order differentiation operators integrated with convolution norms. Zhang et al. [10] employed fractional-order differentiation and closed image matting for multi-focus fusion, while Zhang et al. [14] utilized fractional-order derivatives and intuitionistic fuzzy sets. Lu et al. [33] enhanced multi-focus fusion by incorporating residual removal and fractional-order differentiation focus measures. Additionally, Li et al. [34] introduced an adaptive fractional differential method with guided filtering for fusion, and Yu et al. [35] proposed a sparse representation framework based on fractional-order differentiation for multi-focus image fusion.
Spatial domain-based methods operate directly on pixel or local patch levels by selecting focused regions using sharpness metrics (e.g., gradient, energy, variance). These approaches offer computational simplicity, making them particularly suitable for real-time applications. However, they exhibit certain limitations: their performance is sensitive to block size selection and may introduce undesirable blocking artifacts during the fusion process [36].
Deep learning-based methods leverage neural networks to autonomously extract focus-related features and learn optimal fusion rules, encompassing both end-to-end and multi-stage architectures. Representative algorithms in this domain include CNNs, GANs, Transformers, Mamba, and their variants, which demonstrate remarkable adaptability and consistently deliver cutting-edge performance [37,38]. Zhang et al. [39] proposed the unsupervised generative adversarial network with adaptive and gradient joint constraints for image fusion. While these approaches have revolutionized image fusion through their data-driven nature, they inherently require extensive training datasets and demand significant computational resources for both training and deployment.
Hybrid methods strategically combine transform-domain, spatial-domain, and deep learning techniques to harness their complementary advantages. As exemplified by Avci et al.’s [40] innovative approach integrating discrete wavelet transform with deep convolutional neural networks for multi-focus image fusion, these hybrid architectures consistently demonstrate superior performance when benchmarked against conventional methods. While offering enhanced robustness for complex scene analysis, these methods inevitably introduce higher algorithmic complexity and present significant parameter optimization challenges due to their sophisticated multi-component nature.
In this paper, a novel multi-focus image fusion method based on the dual-channel Rybak neural network (DRYNN) and consistency verification (CV) in the NSCT domain is proposed. The dual-channel Rybak neural network and consistency verification models are used to process the high-frequency sub-bands generated by NSCT, and the average method is used to deal with the low-frequency sub-bands generated by NSCT. One of the key innovations of this paper is the application of consistency verification to the processing of high-frequency coefficients, which significantly improves the image fusion performance. The proposed algorithm was evaluated through a series of qualitative and quantitative analyses on three public datasets: Lytro, MFFW, and MFI-WHU [24]. Comparisons with state-of-the-art traditional and deep learning-based image fusion methods demonstrate the superiority of the proposed approach.
The remainder of this paper is organized as follows: Section 2 reviews the dual-channel Rybak neural network, Section 3 details the proposed methodology, Section 4 presents experimental results and analysis, and Section 5 concludes with future directions.

2. Dual-Channel Rybak Neural Network

The dual-channel Rybak neural network (DRYNN) [41], illustrated in Figure 1, processes two input images, A and B. After up to N iterations, it outputs a binary decision matrix D, which indicates the neuron with the higher internal activity between the corresponding neurons from the two inputs. The DRYNN model can be mathematically formulated as follows:
X S A i , j = V S x = 2 2 y = 2 2 F S x + 3 , y + 3 S A i + x , j + y
X S B i , j = V S x = 2 2 y = 2 2 F S x + 3 , y + 3 S B i + x , j + y
X I n A i , j = α x = 2 2 y = 2 2 F I x + 3 , y + 3 Z n 1 i + x , j + y + S A i , j
X I n B i , j = α x = 2 2 y = 2 2 F I x + 3 , y + 3 Z n 1 i + x , j + y + S B i , j
where X S X X A , B denotes the direct inputs, while X I X X A , B refers to the feedback inputs. S X X A , B represents the external stimulus associated with input X . Feedback inputs also receive their corresponding stimuli to ensure sufficient feature extraction. The variable n indicates the iteration number, whereas i , j represents the i , j - th neuron. The parameter α denotes the coefficient of backward inhibition, and V S represents the linking amplitude. F S and F I are the two 5 × 5 matrices that show the on-center/off-surround and oriented local connection receptive fields of the DRYNN model, respectively. The values of F S and F I are determined by Equations (5) and (6), respectively. Z is the binary output matrix, calculated using Equation (7).
F S = 0   0   0   0   0 0   1   1   1   0 0   1   0   1   0 0   1   1   1   0 0   0   0   0   0
F I = 1   1   1   0   0 1   1   0   0   1 1   1   0   1   1 1   0   0   1   1 0   0   1   1   1
Z n i , j = 1 ,   if   P n i , j > E n 1 i , j 0 ,   else
where P and E show the internal activity and threshold function of the DRYNN model, respectively. The internal activity is calculated using the following equation:
P n i , j = e α p P n 1 i , j + max P n A i , j , P n B i , j
where α p represents the decay constant governing the internal activity in the DRYNN model. The variables P n X i , j X A , B denote the internal states of the i , j - th neuron corresponding to the input of the DRYNN model at iteration n th . These internal states are computed as follows.
P n A i , j = X S A i , j 1 Φ X I n A i , j + h
P n B i , j = X S B i , j 1 Φ X I n B i , j + h
where Φ denotes the time constant associated with the internal blocks, while h serves as the threshold constant. The dynamic threshold is determined using Equation (11).
E n i , j = e α e E n 1 i , j + V E Z n i , j
where α e and V E denote the decay and amplitude constants for E , respectively. The constant parameters of the DRYNN model are determined according to Equation (12).
V S = 1 ,   α = 1.5 ,   α p = 0.7 ,   Φ = 1.2 , h = 0.28 ,   α e = 0.001 ,   V E = 30
The parameters V S , α , Φ , h , α e , and V E are empirically determined based on the literature by RYNN [42], while the parameter α p is set using the concept of decay constant for the internal activity of PCNN-based models [43]. The significance of a neuron is reflected in its activity, which represents its internal state.

3. Proposed Fusion Method

We present a multi-focus image fusion algorithm that combines a dual-channel Rybak neural network with consistency verification in the NSCT domain. The overall framework of the proposed method is illustrated in Figure 2. The fusion process consists of four main stages: (1) NSCT decomposition of the input images, (2) fusion of high-frequency sub-bands using the DRYNN model with consistency verification (CV), (3) fusion of low-frequency sub-bands via a simple averaging strategy, and (4) reconstruction of the final fused image through the inverse NSCT. The detailed steps of the algorithm are described as follows.

3.1. NSCT Decomposition

Owing to its shift-invariant, multi-scale, and multi-directional properties, NSCT exhibits excellent performance in tasks such as multi-focus image fusion [25]. In this section, the NSCT is utilized to perform multi-scale and multi-directional decomposition on the input source images A and B. Through this process, both low-pass L P A , L P B and high-pass H P A l , d , H P B l , d components are extracted from each image. Specifically, the low-frequency component, denoted as L P X X A , B , captures the coarse structural information, while the high-frequency component, denoted as H P X l , d X A , B , represents the detailed features of X at the d th direction corresponding to l th decomposition level. This decomposition enables the effective separation of image features across various scales and orientations, which is essential for high-quality image fusion.

3.2. High-Frequency Sub-Band Fusion

The DRYNN model is employed to extract and integrate fine details from the high-frequency sub-bands. To generate the fused high-frequency sub-bands, the model’s constant parameters are first configured according to Equation (12), followed by initialization using Equation (16). In this process, the absolute values (ABS) of the corresponding high-frequency sub-bands serve as external stimuli for the model, i.e., S X = H P X l , d , where X A , B .
X S A i , j = V S x = 2 2 y = 2 2 F S x + 3 , y + 3 H P A l , d i + x , j + y
X S B i , j = V S x = 2 2 y = 2 2 F S x + 3 , y + 3 H P B l , d i + x , j + y
P 0 i , j = 0
Z 0 i , j = 0
E 0 i , j = 1
The computational process begins by initializing parameter E to 1. The DRYNN model is then executed for N = 200 optimization cycles. Upon completion, the internal latent states derived from both input streams are aggregated to construct the decision matrix D , generated through the following procedure:
D i , j = 1 ,   if   P N H P A l . d i , j P N H P B l , d i , j 0 ,   else
where P N H P X l , d X A , B is the internal state of H P X l , d attained after the N th iteration.
In view of the integrity of the object, the decision map D i , j can be refined through a consistency verification (CV) operation [44]:
C V D i , j = 1   if   a , b θ   D i + a , j + b 0   else
where C V D i , j shows the final decision map at position i , j , and θ presents a square neighborhood centered at i , j with a size of 9 × 9 .
The fused high-frequency sub-band H P F l , d can be constructed from D as follows:
H P F l , d i , j = H P A l , d i , j ,   if   C V D i , j = 1 H P B l , d i , j ,   else

3.3. Low-Frequency Sub-Band Fusion

In this section, the fused low-frequency sub-band image L P F is generated using the averaging technique, which computes the mean of the corresponding low-frequency coefficients from the source images. This approach ensures a balanced preservation of overall brightness and contrast, and the mathematical formulation is provided as follows [45]:
L P F i , j = L P A i , j + L P B i , j 2

3.4. Inverse NSCT

The fused image F is reconstructed by applying the inverse NSCT to the fused low-frequency and high-frequency components. This process integrates the fused sub-bands into a single image, effectively preserving both coarse and detailed information, and the corresponding equation is defined as follows:
F i , j = I n v e r s e   N S C T L P F i , j , H P F l , d i , j

4. Experimental Results and Analysis

In this section, we validate the effectiveness of the proposed algorithm through comprehensive simulation experiments conducted on three publicly available benchmark datasets: Lytro [46], MFFW [47], and MFI-WHU [39]. The Lytro dataset comprises 20 multi-focus image pairs, the MFFW dataset includes 13 image pairs with challenging focus settings, and a total of 30 representative image pairs were selected from the MFI-WHU dataset for performance evaluation. Figure 3 illustrates sample image pairs drawn from each dataset, showcasing the diversity and complexity of the data used in our experiments.
To ensure a robust and fair comparison, we evaluated our method against eight state-of-the-art image fusion techniques: PMGI [48], MFFGAN [39], U2Fusion [49], XDoG [7], NSCTST [25], EgeFusion [50], FDFusion [51], and CVTFD [24]. These methods represent a range of classical, transform-based, and deep learning-based approaches.
Furthermore, to objectively assess the quality of the fused images, we employed a suite of widely recognized evaluation metrics. These metrics comprehensively measure various aspects of fusion performance, including Q A B / F [52], Q C B [53], Q E [53], Q F M I [54], Q M I [52], Q M S E [55,56,57,58], Q N C I E [53], Q N M I [53], Q P [53], and Q P S N R [59,60,61,62]. Except for the Q M S E metric, larger values of the other metrics indicate better image fusion performance. In our method, the NSCT decomposition level is 4, and the direction numbers are 4, 8, 8, and 16; the pyramidal and directional filters are defined as “9-7” and “pkva,” respectively.

4.1. Results on Lytro Dataset

Figure 4 and Figure 5 present the simulation results and difference maps of various algorithms applied to a sample dataset from the Lytro collection. The difference maps, generated by subtracting one source input image from the fused output, highlight the fusion performance. As illustrated in Figure 5i, the proposed method produces fusion results with sharp boundaries between focused and defocused regions while preserving color fidelity. In contrast, other methods exhibit limitations: NSCTST and CVTFD demonstrate boundary blurring (Figure 5e,h), while PMGI, MFFGAN, U2Fusion, XDoG, EdgeFusion, and FDFusion introduce significant artifacts (Figure 5a–d,f,g).
Table 1 provides a quantitative comparison of different methods on the Lytro-01 dataset (Figure 4). The proposed algorithm outperforms others, achieving optimal scores in 10 evaluation metrics. Both subjective and objective assessments confirm its superior performance.
Figure 6 provides insight into the performance of different fusion methods on the Lytro dataset. For each metric, the results for individual image pairs are connected into curves, with average scores shown in the legend. Most methods show consistent trends and stable performance, with few outliers. Therefore, the average values in Table 2 are representative. As shown in Table 2, our method achieves the highest scores across all metrics on the Lytro dataset.

4.2. Results on MFFW Dataset

Figure 7 and Figure 8 show that all methods are capable of generating fused images. We can focus on examining the fusion performance of different algorithms in the regions containing the boat and the bird. However, the quality of the difference images varies notably. Specifically, methods such as PMGI, MFFGAN, U2Fusion, XDoG, and EgeFusion struggle to accurately detect near-focus regions due to insufficient extraction of information from the source images. In contrast, NSCTST, FDFusion, and CVTFD perform better in identifying near-focused areas, although some unfocused artifacts remain within these regions. The proposed method, however, demonstrates superior performance in accurately capturing both the central and edge regions of the focus area.
Table 3 presents a quantitative comparison of various methods on the MFFW-01 dataset, as shown in Figure 7. The results indicate that our method achieves the highest performance across 10 evaluation metrics. Both subjective visual assessments and objective measurements validate the effectiveness of our approach for image fusion on the MFFW dataset. Furthermore, the average performance of each method is depicted in Figure 9 and summarized in Table 4, where our method consistently delivers the best overall results, with the exception of the Q P S N R metric.

4.3. Results on MFI-WHU Dataset

Figure 10 and Figure 11 display the fused images obtained by each method for the MFI-WHU-01 dataset image, along with the corresponding difference maps. From the results, the PMGI generates blurred output, which is clearly observable in the fused image and the associated difference map. The fused images produced by MFFGAN, U2Fusion, XDoG, EgeFusion, and FDFusion fail to adequately preserve background information, resulting in noticeable information loss or artifacts in the corresponding difference images, which indicates poor fusion performance. The NSCTST and CVTFD methods demonstrate relatively good fusion performance; however, the corresponding difference images reveal artifacts or block effects in certain regions, indicating that the fused images do not fully retain the information from the source images. Compared to other algorithms, our method produces fused images that preserve the complete information from the source images without introducing artifacts or noise.
Table 5 presents a quantitative comparison of various methods on the MFI-WHU-01 dataset illustrated in Figure 10. As shown in the table, our algorithm achieves the best performance across 10 evaluation metrics. Both subjective assessments and objective measurements confirm the effectiveness of our method for image fusion on the MFI-WHU dataset. The average performance metrics for each method are depicted in Figure 12 and summarized in Table 6. Except for the Q A B / F and Q P S N R metrics, our algorithm achieves the best average performance across the other eight metrics.

4.4. Ablation Experiment

In the ablation study section, we conducted a detailed analysis of the role of consistency verification (CV) in our method. We evaluated the effectiveness of the algorithm by comparing two models—one without and one with consistency verification—using the average values of 10 evaluation metrics across three datasets: Lytro, MFFW, and MFI-WHU. The corresponding experimental results are shown in Table 7. The data clearly indicate that incorporating consistency verification enhances the performance of multi-focus image fusion, achieving the best results across all evaluation metrics.

4.5. Extended Experiments

In this section, the proposed algorithm is further applied to a range of image fusion tasks, such as infrared and visible fusion [63,64,65], multi-exposure fusion [66,67], medical image fusion (CT/MRI/PET) [68,69,70], pan-sharpening (PAN/MS) [71,72,73,74,75], and SAR–optical image fusion [76]. For MRI and PET fusion as well as SAR–optical fusion, we apply RGB and YUV [68] color space transformations to the PET and optical images. For pan-sharpening, we apply RGB and IHS [71] color space transformations to the multispectral images. The fusion results of these extended applications are shown in Figure 13. As illustrated, the proposed algorithm also achieves a favorable performance in these multi-source image fusion tasks, demonstrating its good generalization ability and scalability.

5. Conclusions

In this study, we introduce a novel multi-focus image fusion algorithm that integrates the dual-channel Rybak neural network (DRYNN) with a consistency verification strategy in the NSCT domain. The primary objective of this method is to enhance the clarity and structural integrity of the fused images while effectively mitigating the impact of noise. To achieve this, the high-frequency sub-bands derived from NSCT decomposition are refined using the DRYNN model in conjunction with consistency verification, allowing for the precise retention of edge and texture details. Meanwhile, the low-frequency sub-bands, which primarily contain coarse information, are fused using a straightforward averaging technique to preserve overall image content. Consistency verification plays a significant role in our algorithm and serves as an innovative aspect of our approach. Ablation experiments have also confirmed the contribution of consistency verification in image fusion.
The proposed fusion framework is rigorously evaluated using three widely recognized public datasets, with performance assessed through both qualitative visual comparisons and quantitative metrics. To gain further insight into the fidelity of the fusion process, difference maps between the fused results and the original input images are generated, highlighting regions of information gain or loss. Additionally, we employ ten standard evaluation metrics to objectively compare our method against existing state-of-the-art fusion algorithms.
Our experimental results clearly demonstrate the superior performance of the proposed algorithm, both in terms of visual perception and numerical accuracy. Looking ahead, we intend to extend this fusion framework to broader application domains such as change detection [77,78,79,80,81], hyperspectral and multispectral fusion [82,83,84], and panchromatic and hyperspectral image fusion [85,86], where the need for robust, high-quality fusion techniques is critical for decision-making and analysis. In addition, the efficiency of our algorithm is relatively low and does not meet real-time requirements, which is also a problem we aim to address in our future work.

Author Contributions

The experiments and data collection were conducted by M.L., S.S., Z.J., L.L. and H.M. The manuscript was drafted by M.L., with contributions from the co-authors. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China Postdoctoral Science Foundation under Grant No. 2024M752692; the Natural Science Foundation of Xinjiang Uygur Autonomous Region under Grant No. 2024D01C240; the National Natural Science Foundation of China under Grant No. 62261053; the Tianshan Talent Training Project-Xinjiang Science and Technology Innovation Team Program (2023TSYCTD0012).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, W.; Deng, L.; Vivone, G. A general image fusion framework using multi-task semi-supervised learning. Inf. Fusion 2024, 108, 102414. [Google Scholar] [CrossRef]
  2. Wu, X.; Cao, Z.; Huang, T.; Deng, L.; Chanussot, J.; Vivone, G. Fully-connected transformer for multi-source image fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 2071–2088. [Google Scholar] [CrossRef] [PubMed]
  3. Vivone, G.; Deng, L. Deep learning in remote sensing image fusion: Methods, protocols, data, and future perspectives. IEEE Geosci. Remote Sens. Mag. 2025, 13, 269–310. [Google Scholar] [CrossRef]
  4. Matteo, C.; Giuseppe, G.; Gemine, V. Hyperspectral pansharpening: Critical review, tools, and future perspectives. IEEE Geosci. Remote Sens. Mag. 2025, 13, 311–338. [Google Scholar]
  5. Jie, Y.; Xu, Y.; Li, X.; Zhou, F.; Lv, J.; Li, H. FS-Diff: Semantic guidance and clarity-aware simultaneous multimodal image fusion and super-resolution. Inf. Fusion 2025, 121, 103146. [Google Scholar] [CrossRef]
  6. Zhang, X.; Yan, H. Medical image fusion and noise suppression with fractional-order total variation and multi-scale decomposition. IET Image Process. 2021, 15, 1688–1701. [Google Scholar] [CrossRef]
  7. Jie, Y.; Li, X.; Wang, M.; Zhou, F.; Tan, H. Medical image fusion based on extended difference-of-Gaussians and edge-preserving. Expert Syst. Appl. 2023, 227, 120301. [Google Scholar] [CrossRef]
  8. Zheng, K.; Cheng, J.; Liu, Y. Unfolding coupled convolutional sparse representation for multi-focus image fusion. Inf. Fusion 2025, 118, 102974. [Google Scholar] [CrossRef]
  9. Li, B.; Zhang, L.; Liu, J.; Peng, H. Multi-focus image fusion with parameter adaptive dual channel dynamic threshold neural P systems. Neural Netw. 2024, 179, 106603. [Google Scholar] [CrossRef]
  10. Zhang, X.; He, H.; Zhang, J. Multi-focus image fusion based on fractional order differentiation and closed image matting. ISA Trans. 2022, 129, 703–714. [Google Scholar] [CrossRef]
  11. Liu, J.; Wu, G.; Liu, Z.; Wang, D.; Jiang, Z.; Ma, L.; Zhong, W.; Fan, X.; Liu, R. Infrared and visible image fusion: From data compatibility to task adaption. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 2349–2369. [Google Scholar] [CrossRef] [PubMed]
  12. Yan, H.; Zhang, J.; Zhang, X. Injected infrared and visible image fusion via L1 decomposition model and guided filtering. IEEE Trans. Comput. Imaging 2022, 8, 162–173. [Google Scholar] [CrossRef]
  13. Yan, H.; Zhang, X. Adaptive fractional multi-scale edge-preserving decomposition and saliency detection fusion algorithm. ISA Trans. 2020, 107, 160–172. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, X.; Yan, H.; He, H. Multi-focus image fusion based on fractional-order derivative and intuitionistic fuzzy sets. Front. Inf. Technol. Electron. Eng. 2020, 21, 834–843. [Google Scholar] [CrossRef]
  15. Li, L.; Zhao, X.; Hou, H.; Zhang, X.; Lv, M.; Jia, Z.; Ma, H. Fractal dimension-based multi-focus image fusion via coupled neural P systems in NSCT domain. Fractal Fract. 2024, 8, 554. [Google Scholar] [CrossRef]
  16. Zhang, X. Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 4819–4838. [Google Scholar] [CrossRef]
  17. Luo, F.; Zhao, B. A review on multi-focus image fusion using deep learning. Neurocomputing 2025, 618, 129125. [Google Scholar] [CrossRef]
  18. Zhai, H.; Zhang, G.; Zeng, Z.; Xu, Z.; Fang, A. LSKN-MFIF: Large selective kernel network for multi-focus image fusion. Neurocomputing 2025, 635, 129984. [Google Scholar] [CrossRef]
  19. Quan, Y.; Wan, X.; Tang, Z.; Liang, J.; Ji, H. Multi-focus image fusion via explicit defocus blur modelling. Proc. AAAI Conf. Artif. Intell. 2025, 39, 6657–6665. [Google Scholar] [CrossRef]
  20. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Deep learning-based image fusion: A survey. J. Image Graph. 2023, 28, 3–36. [Google Scholar] [CrossRef]
  21. Li, L.; Ma, H.; Jia, Z.; Si, Y. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 12389–12409. [Google Scholar] [CrossRef]
  22. Li, L.; Si, Y.; Wang, L.; Jia, Z.; Ma, H. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  23. Li, L.; Shi, Y.; Lv, M.; Jia, Z.; Liu, M.; Zhao, X.; Zhang, X.; Ma, H. Infrared and visible image fusion via sparse representation and guided filtering in Laplacian pyramid domain. Remote Sens. 2024, 16, 3804. [Google Scholar] [CrossRef]
  24. Li, L.; Song, S.; Lv, M.; Jia, Z.; Ma, H. Multi-focus image fusion based on fractal dimension and parameter adaptive unit-linking dual-channel PCNN in curvelet transform domain. Fractal Fract. 2025, 9, 157. [Google Scholar] [CrossRef]
  25. Lv, M.; Li, L.; Jin, Q.; Jia, Z.; Chen, L.; Ma, H. Multi-focus image fusion via distance-weighted regional energy and structure tensor in NSCT domain. Sensors 2023, 23, 6135. [Google Scholar] [CrossRef]
  26. Li, L.; Lv, M.; Jia, Z.; Ma, H. Sparse representation-based multi-focus image fusion method via local energy in shearlet domain. Sensors 2023, 23, 2888. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Li, H.; Xu, T.; Wu, X.; Kittler, J. DDBFusion: An unified image decomposition and fusion framework based ondual decomposition and Bézier curves. Inf. Fusion 2025, 114, 102655. [Google Scholar] [CrossRef]
  28. Quan, Y.; Wan, X.; Zheng, T.; Huang, Y.; Ji, H. Dual-path deep unsupervised learning for multi-focus image fusion. IEEE Trans. Multimed. 2025, 27, 1165–1176. [Google Scholar] [CrossRef]
  29. Xie, X.; Jiang, Q.; Chen, D.; Guo, B.; Li, P.; Zhou, S. StackMFF: End-to-end multi-focus image stack fusion network. Appl. Intell. 2025, 55, 503. [Google Scholar] [CrossRef]
  30. Panigrahy, C.; Seal, A.; Mahato, N.K. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106141. [Google Scholar] [CrossRef]
  31. Joshua, A.; Balasubramaniam, P. An adaptive low-light image enhancement method via fusion of a new intuitionistic fuzzy generator and fractal-fractional derivative. Signal Image Video Process. 2025, 19, 233. [Google Scholar] [CrossRef]
  32. Xian, Y.; Zhao, G. Multi-focus image fusion based on visual depth and fractional-order differentiation operators embedding convolution norm. Signal Process. 2025, 233, 109955. [Google Scholar] [CrossRef]
  33. Lu, J.; Tan, K. Multi-focus image fusion using residual removal and fractional order differentiation focus measure. Signal Image Video Process. 2024, 18, 3395–3410. [Google Scholar] [CrossRef]
  34. Li, X.; Chen, H. Multi-focus image fusion via adaptive fractional differential and guided filtering. Multimed. Tools Appl. 2024, 83, 32923–32943. [Google Scholar] [CrossRef]
  35. Yu, L.; Zeng, Z. Fractional-order differentiation based sparse representation for multi-focus image fusion. Multimed. Tools Appl. 2022, 81, 4387–4411. [Google Scholar] [CrossRef]
  36. Ouyang, Y.; Zhai, H.; Hu, H. FusionGCN: Multi-focus image fusion using superpixel features generation GCN and pixel-level feature reconstruction CNN. Expert Syst. Appl. 2025, 262, 125665. [Google Scholar] [CrossRef]
  37. Li, S.; Huang, S. AFA–Mamba: Adaptive feature alignment with global–local Mamba for hyperspectral and LiDAR data classification. Remote Sens. 2024, 16, 4050. [Google Scholar] [CrossRef]
  38. Li, H.; Shen, T.; Zhang, Z.; Zhu, X.; Song, X. EDMF: A new benchmark for multi-focus images with the challenge of exposure difference. Sensors 2024, 24, 7287. [Google Scholar] [CrossRef]
  39. Zhang, H.; Le, Z. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  40. Avci, D.; Sert, E.; Özyurt, F.; Avci, E. MFIF-DWT-CNN: Multi-focus image fusion based on discrete wavelet transform with deep convolutional neural network. Multimed. Tools Appl. 2024, 83, 10951–10968. [Google Scholar] [CrossRef]
  41. Goyal, N.; Goyal, N. Dual-channel Rybak neural network based medical image fusion. Opt. Laser Technol. 2025, 181, 112018. [Google Scholar] [CrossRef]
  42. Qi, Y.; Yang, Z.; Lian, J.; Guo, Y.; Sun, W.; Liu, J.; Wang, R.; Ma, Y. A new heterogeneous neural network model and its application in image enhancement. Neurocomputing 2021, 440, 336–350. [Google Scholar] [CrossRef]
  43. Sinha, A.; Agarwal, R.; Kumar, V.; Garg, N.; Pundir, D.S.; Singh, H.; Rani, R.; Panigrahy, C. Multi-modal medical image fusion using improved dual-channel PCNN. Med. Biol. Eng. Comput. 2024, 62, 2629–2651. [Google Scholar] [CrossRef] [PubMed]
  44. Li, X.; Zhou, F.; Tan, H. Multi-focus image fusion based on nonsubsampled contourlet transform and residual removal. Signal Process. 2021, 184, 108062. [Google Scholar] [CrossRef]
  45. Zafar, R.; Farid, M.S.; Khan, M.H. Multi-focus image fusion: Algorithms, evaluation, and a library. J. Imaging 2020, 6, 60. [Google Scholar] [CrossRef]
  46. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  47. Xu, S.; Wei, X.; Zhang, C. MFFW: A newdataset for multi-focus image fusion. arXiv 2020, arXiv:2002.04780. [Google Scholar]
  48. Zhang, H.; Xu, H.; Xiao, Y. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar]
  49. Xu, H.; Ma, J.; Jiang, J. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  50. Tang, H.; Liu, G.; Qian, Y. EgeFusion: Towards edge gradient enhancement in infrared and visible image fusion with multi-scale transform. IEEE Trans. Comput. Imaging 2024, 10, 385–398. [Google Scholar] [CrossRef]
  51. Jie, Y.; Li, X.; Tan, T.; Yang, L.; Wang, M. Multi-modality image fusion using fuzzy set theory and compensation dictionary learning. Opt. Laser Technol. 2025, 181, 112001. [Google Scholar] [CrossRef]
  52. Qu, X.; Yan, J.; Xiao, H. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  53. Liu, Z.; Blasch, E.; Xue, Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 94–109. [Google Scholar] [CrossRef]
  54. Haghighat, M.; Razian, M. Fast-FMI: Non-reference image fusion metric. In Proceedings of the IEEE 8th International Conference on Application of Information and Communication Technologies, Astana, Kazakhstan, 15–17 October 2014; pp. 424–426. [Google Scholar]
  55. Chen, F.; Sha, Y.; Ji, H.; Peng, K.; Liang, X. Integrating multifractal features into machine learning for improved prediction. Fractal Fract. 2025, 9, 205. [Google Scholar] [CrossRef]
  56. Li, J.; Wang, C.; Su, W.; Ye, D.; Wang, Z. Uncertainty-aware self-attention model for time series prediction with missing values. Fractal Fract. 2025, 9, 181. [Google Scholar] [CrossRef]
  57. Bi, X.; Qie, R.; Tao, C.; Zhang, Z.; Xu, Y. Unsupervised multimodal UAV image registration via style transfer and cascade network. Remote Sens. 2025, 17, 2160. [Google Scholar] [CrossRef]
  58. Wadood, A.; Albalawi, H.; Alatwi, A.M.; Anwar, H.; Ali, T. Design of a novel fractional whale optimization-enhanced support vector regression (FWOA-SVR) model for accurate solar energy forecasting. Fractal Fract. 2025, 9, 35. [Google Scholar] [CrossRef]
  59. Zhang, X.; Dai, L. Image enhancement based on rough set and fractional order differentiator. Fractal Fract. 2022, 6, 214. [Google Scholar] [CrossRef]
  60. Zhang, X.; Liu, R.; Ren, J.; Gui, Q. Adaptive fractional image enhancement algorithm based on rough set and particle swarm optimization. Fractal Fract. 2022, 6, 100. [Google Scholar] [CrossRef]
  61. Zhang, X.; Boutat, D.; Liu, D. Applications of fractional operator in image processing and stability of control systems. Fractal Fract. 2023, 7, 359. [Google Scholar] [CrossRef]
  62. Tabirca, A.I.; Dumitrescu, C.; Radu, V. Enhancing banking transaction security with fractal-based image steganography using fibonacci sequences and discrete wavelet transform. Fractal Fract. 2025, 9, 95. [Google Scholar] [CrossRef]
  63. Li, L.; Lv, M.; Jia, Z.; Jin, Q.; Liu, M.; Chen, L.; Ma, H. An effective infrared and visible image fusion approach via rolling guidance filtering and gradient saliency map. Remote Sens. 2023, 15, 2486. [Google Scholar] [CrossRef]
  64. Heredia-Aguado, E.; Cabrera, J.J.; Jiménez, L.M.; Valiente, D.; Gil, A. Static early fusion techniques for visible and thermal images to enhance convolutional neural network detection: A performance analysis. Remote Sens. 2025, 17, 1060. [Google Scholar] [CrossRef]
  65. Li, L.; Ma, H. Saliency-guided nonsubsampled shearlet transform for multisource remote sensing image fusion. Sensors 2021, 21, 1756. [Google Scholar] [CrossRef] [PubMed]
  66. Zhang, X. Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion 2021, 74, 111–131. [Google Scholar] [CrossRef]
  67. Zhang, W.; Wang, C.; Zhu, J. MEF-CAAN: Multi-exposure image fusion based on a low-resolution context aggregation attention network. Sensors 2025, 25, 2500. [Google Scholar] [CrossRef]
  68. Yin, M.; Liu, X.; Liu, Y.; Chen, X. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans. Instrum. Meas. 2019, 68, 49–64. [Google Scholar] [CrossRef]
  69. Zhu, Z.; Wang, Z.; Qi, G.; Mazur, N.; Yang, P.; Liu, Y. Brain tumor segmentation in MRI with multi-modality spatial information enhancement and boundary shape correction. Pattern Recognit. 2024, 153, 110553. [Google Scholar] [CrossRef]
  70. Zhu, Z.; He, X.; Qi, G.; Li, Y.; Cong, B.; Liu, Y. Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI. Inf. Fusion 2023, 91, 376–387. [Google Scholar] [CrossRef]
  71. Liu, Y.; Wang, Z. A practical pan-sharpening method with wavelet transform and sparse representation. In Proceedings of the 2013 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 22–23 October 2013; pp. 288–293. [Google Scholar]
  72. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Restaino, R.; Scarpa, G. A new benchmark based on recent advances in multispectral pansharpening: Revisiting pansharpening with classical and emerging pansharpening methods. IEEE Geosci. Remote Sens. Mag. 2021, 9, 53–81. [Google Scholar] [CrossRef]
  73. Vivone, G. Robust band-dependent spatial-detail approaches for panchromatic sharpening. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  74. Wen, X.; Ma, H.; Li, L. A three-branch pansharpening network based on spatial and frequency domain interaction. Remote Sens. 2025, 17, 13. [Google Scholar] [CrossRef]
  75. Wen, X.; Ma, H.; Li, L. A multi-stage progressive pansharpening network based on detail injection with redundancy reduction. Sensors 2024, 24, 6039. [Google Scholar] [CrossRef] [PubMed]
  76. Li, J.; Zhang, J.; Yang, C.; Liu, H.; Zhao, Y.; Ye, Y. Comparative analysis of pixel-level fusion algorithms and a new high-resolution dataset for SAR and optical image fusion. Remote Sens. 2023, 15, 5514. [Google Scholar] [CrossRef]
  77. Li, L.; Ma, H.; Jia, Z. Multiscale geometric analysis fusion-based unsupervised change detection in remote sensing images via FLICM model. Entropy 2022, 24, 291. [Google Scholar] [CrossRef]
  78. Chen, Z.; Chen, H.; Leng, J.; Zhang, X.; Gao, Q.; Dong, W. VMMCD: VMamba-based multi-scale feature guiding fusion network for remote sensing change detection. Remote Sens. 2025, 17, 1840. [Google Scholar] [CrossRef]
  79. Li, L.; Ma, H.; Jia, Z. Change detection from SAR images based on convolutional neural networks guided by saliency enhancement. Remote Sens. 2021, 13, 3697. [Google Scholar] [CrossRef]
  80. Zhong, H.; Wu, C.; Xiao, Z. LRNet: Change detection in high-resolution remote sensing imagery via a localization-then-refinement strategy. Remote Sens. 2025, 17, 1849. [Google Scholar] [CrossRef]
  81. Li, L.; Ma, H.; Zhang, X.; Zhao, X.; Lv, M.; Jia, Z. Synthetic aperture radar image change detection based on principal component analysis and two-level clustering. Remote Sens. 2024, 16, 1861. [Google Scholar] [CrossRef]
  82. He, Y.; Li, H.; Zhang, M.; Liu, S.; Zhu, C.; Xin, B.; Wang, J.; Wu, Q. Hyperspectral and multispectral remote sensing image fusion based on a retractable spatial–spectral transformer network. Remote Sens. 2025, 17, 1973. [Google Scholar] [CrossRef]
  83. Li, J.; Zheng, K.; Gao, L.; Han, Z.; Li, Z.; Chanussot, J. Enhanced deep image prior for unsupervised hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5504218. [Google Scholar] [CrossRef]
  84. Vivone, G. Multispectral and hyperspectral image fusion in remote sensing: A survey. Inf. Fusion 2023, 89, 405–417. [Google Scholar] [CrossRef]
  85. Vivone, G.; Garzelli, A.; Xu, Y.; Liao, W.; Chanussot, J. Panchromatic and hyperspectral image fusion: Outcome of the 2022 WHISPERS hyperspectral pansharpening challenge. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 166–179. [Google Scholar] [CrossRef]
  86. Shen, X.; Chen, L.; Liu, H.; Zhou, X.; Vivione, G.; Chanussot, J. Iteratively regularizing hyperspectral and multispectral image fusion with framelets. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 5331–5346. [Google Scholar] [CrossRef]
Figure 1. The DRYNN model.
Figure 1. The DRYNN model.
Fractalfract 09 00432 g001
Figure 2. The framework of the proposed algorithm.
Figure 2. The framework of the proposed algorithm.
Fractalfract 09 00432 g002
Figure 3. The selection of image pairs from the three datasets. (a) Lytro; (b) MFFW; (c) MFI-WHU.
Figure 3. The selection of image pairs from the three datasets. (a) Lytro; (b) MFFW; (c) MFI-WHU.
Fractalfract 09 00432 g003
Figure 4. Results of Lytro-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 4. Results of Lytro-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g004
Figure 5. The difference maps of fused images and source image A in Figure 4. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 5. The difference maps of fused images and source image A in Figure 4. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g005
Figure 6. The line chart displays the evaluation metrics across various data from the Lytro dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Figure 6. The line chart displays the evaluation metrics across various data from the Lytro dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Fractalfract 09 00432 g006aFractalfract 09 00432 g006b
Figure 7. Results of MFFW-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 7. Results of MFFW-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g007
Figure 8. The difference maps of fused images and source image B in Figure 7. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 8. The difference maps of fused images and source image B in Figure 7. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g008
Figure 9. The line chart displays the evaluation metrics across various data from the MFFW dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Figure 9. The line chart displays the evaluation metrics across various data from the MFFW dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Fractalfract 09 00432 g009aFractalfract 09 00432 g009b
Figure 10. Results of MFI-WHU-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 10. Results of MFI-WHU-01. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g010
Figure 11. The difference maps of fused images and source image B in Figure 10. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Figure 11. The difference maps of fused images and source image B in Figure 10. (a) PMGI; (b) MFFGAN; (c) U2Fusion; (d) XDoG; (e) NSCTST; (f) EgeFusion; (g) FDFusion; (h) CVTFD; (i) Proposed.
Fractalfract 09 00432 g011
Figure 12. The line chart displays the evaluation metrics across various data from the MFI-WHU dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Figure 12. The line chart displays the evaluation metrics across various data from the MFI-WHU dataset. (a) Q A B / F ; (b) Q C B ; (c) Q E ; (d) Q F M I ; (e) Q M I ; (f) Q M S E ; (g) Q N C I E ; (h) Q N M I ; (i) Q P ; (j) Q P S N R .
Fractalfract 09 00432 g012aFractalfract 09 00432 g012b
Figure 13. Application of the proposed algorithm in other multi-source image fusion tasks. (a) Infrared and visible fusion; (b) multi-exposure fusion; (c) medical image fusion; (d) pan-sharpening; (e) SAR and optical fusion.
Figure 13. Application of the proposed algorithm in other multi-source image fusion tasks. (a) Infrared and visible fusion; (b) multi-exposure fusion; (c) medical image fusion; (d) pan-sharpening; (e) SAR and optical fusion.
Fractalfract 09 00432 g013
Table 1. Quantitative comparative analysis of different methods on the Lytro-01 dataset in Figure 4.
Table 1. Quantitative comparative analysis of different methods on the Lytro-01 dataset in Figure 4.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.57110.66490.64970.91426.0308133.23640.82310.80110.679626.8846
MFFGAN20210.68150.69000.83140.91715.603530.46300.82090.72760.811833.2931
U2Fusion20220.63140.54780.78410.90805.462819.59300.82020.70970.744335.2098
XDoG20230.68290.70040.80640.91475.346760.40930.81970.69350.820530.3198
NSCTST20230.76910.81130.87830.92207.141726.01300.83010.93020.899433.9789
EgeFusion20240.30430.37440.48310.86742.419993.07630.81000.31240.513428.4424
FDFusion20250.71680.76860.85460.91796.131533.04500.82360.79830.824032.9397
CVTFD20250.75970.80540.87930.92197.083119.69250.82960.92310.890235.1878
Proposed 0.77210.82680.87990.92297.425618.90230.83200.96750.904435.3657
Notes: The optimal values are in bold.
Table 2. Quantitative average comparative analysis of different methods on the Lytro dataset.
Table 2. Quantitative average comparative analysis of different methods on the Lytro dataset.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.39010.56560.47360.88155.864175.39560.82250.80040.462032.4782
MFFGAN20210.66420.64570.84090.89156.060434.37480.82370.80470.712533.5508
U2Fusion20220.61430.56820.78350.88445.776559.44240.82210.77250.665731.2098
XDoG20230.68850.64670.83490.89265.668561.24100.82160.75380.744430.5692
NSCTST20230.73880.72880.87450.89866.712831.39850.82790.89460.810033.6375
EgeFusion20240.35760.40340.50320.84723.219177.85970.81200.42480.540529.2757
FDFusion20250.65860.61270.80750.88986.183435.97130.82440.82520.617932.9763
CVTFD20250.72850.72130.87730.89886.729622.23360.82790.89640.798435.0684
Proposed 0.74790.75150.88200.90047.095021.44270.83040.94550.833835.2352
Table 3. Quantitative comparative analysis of different methods on the MFFW-01 dataset in Figure 6.
Table 3. Quantitative comparative analysis of different methods on the MFFW-01 dataset in Figure 6.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.56320.50180.61430.89174.036294.28250.81210.61430.274828.3865
MFFGAN20210.58320.57320.70640.90054.4393204.23170.81360.66310.366425.0296
U2Fusion20220.51120.53850.58400.89364.6178201.61280.81430.67690.378425.0856
XDoG20230.61860.57030.75720.90444.179247.72660.81270.62930.408731.3432
NSCTST20230.65580.63360.82270.90454.672225.02390.81470.70880.414134.1473
EgeFusion20240.27320.36430.32100.84932.444278.31060.80770.35650.325929.1926
FDFusion20250.60660.52640.79360.90334.677927.44250.81460.71360.351433.7466
CVTFD20250.63630.60070.79920.90194.621817.04830.81440.69940.387135.8140
Proposed 0.72720.65490.84220.90925.412312.18970.81830.82010.547237.2709
Table 4. Quantitative average comparative analysis of different methods on the MFFW dataset.
Table 4. Quantitative average comparative analysis of different methods on the MFFW dataset.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.38070.50570.42450.86755.047280.98290.81780.72440.355436.3612
MFFGAN20210.59050.58510.75570.87425.049882.95760.81790.70940.488729.8959
U2Fusion20220.55370.54990.70760.86904.889494.30350.81710.69920.478429.1532
XDoG20230.60900.58860.75120.87455.011986.12450.81780.70380.514629.2046
NSCTST20230.64080.63570.79760.87875.254865.31510.81890.74100.540230.6343
EgeFusion20240.35170.42130.45810.83803.378577.33970.81150.46850.437829.3955
FDFusion20250.58950.57100.70530.87355.314465.40620.81940.75040.444530.6006
CVTFD20250.62900.62760.80210.87965.102943.03550.81810.71990.526332.5370
Proposed 0.71530.67710.83270.88745.829640.05030.82210.82300.702933.0138
Table 5. Quantitative comparative analysis of different methods on the MFI-WHU-01 dataset in Figure 8.
Table 5. Quantitative comparative analysis of different methods on the MFI-WHU-01 dataset in Figure 8.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.38350.54480.45900.86585.6370212.21600.82090.76380.543424.8630
MFFGAN20210.63710.63360.81170.87485.710134.04670.82120.75560.796432.8101
U2Fusion20220.51670.49040.70570.86165.110426.71060.81840.69360.576733.8640
XDoG20230.64210.69130.80330.87285.7376 61.50200.82140.75630.816530.2419
NSCTST20230.71240.79450.84510.88147.667621.90730.83321.01370.878334.7249
EgeFusion20240.40400.42460.56460.83723.295868.77500.81190.44160.526629.7565
FDFusion20250.67150.70750.83440.87706.336236.57940.82460.83600.818632.4984
CVTFD20250.70400.78410.84420.88107.490120.85540.83190.99000.870834.9386
Proposed 0.71430.80050.84520.88158.017620.34410.83581.06070.881735.0464
Table 6. Quantitative average comparative analysis of different methods on the MFI-WHU dataset.
Table 6. Quantitative average comparative analysis of different methods on the MFI-WHU dataset.
Year Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
PMGI20200.42370.59330.50610.85585.488434.45730.82100.76140.475037.9714
MFFGAN20210.64270.63290.78260.86845.683251.51760.82220.77090.704131.6060
U2Fusion20220.55020.51560.69700.85655.149871.82140.81940.69910.621230.1022
XDoG20230.65630.67170.79680.86925.556457.11450.82150.75660.720930.9104
NSCTST20230.73010.80210.84540.87757.700122.74980.83631.05120.783334.8686
EgeFusion20240.28740.32770.37570.82552.805586.43810.81110.37610.519128.8418
FDFusion20250.67640.71040.82980.87546.249530.00260.82560.85240.679433.6057
CVTFD20250.71990.78750.84290.87727.521520.59810.83501.02700.774235.3640
Proposed 0.73000.80750.84590.87767.906320.30500.83841.07910.784835.4395
Table 7. Ablation study results on consistency verification.
Table 7. Ablation study results on consistency verification.
Q A B / F Q C B Q E Q F M I Q M I Q M S E Q N C I E Q N M I Q P Q P S N R
LytroW/o CV0.73900.74470.88150.89896.974321.78200.82960.92910.817335.1463
W/CV0.74790.75150.88200.90047.095021.44270.83040.94550.833835.2352
MFFWW/o CV0.64230.65410.81410.87965.268742.46450.81890.74360.546732.6365
W/CV0.71530.67710.83270.88745.829640.05030.82210.82300.702933.0138
MFI-WHUW/o CV0.72750.80460.84590.87717.731420.41730.83661.05550.783435.4132
W/CV0.73000.80750.84590.87767.906320.30500.83841.07910.784835.4395
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, M.; Song, S.; Jia, Z.; Li, L.; Ma, H. Multi-Focus Image Fusion Based on Dual-Channel Rybak Neural Network and Consistency Verification in NSCT Domain. Fractal Fract. 2025, 9, 432. https://doi.org/10.3390/fractalfract9070432

AMA Style

Lv M, Song S, Jia Z, Li L, Ma H. Multi-Focus Image Fusion Based on Dual-Channel Rybak Neural Network and Consistency Verification in NSCT Domain. Fractal and Fractional. 2025; 9(7):432. https://doi.org/10.3390/fractalfract9070432

Chicago/Turabian Style

Lv, Ming, Sensen Song, Zhenhong Jia, Liangliang Li, and Hongbing Ma. 2025. "Multi-Focus Image Fusion Based on Dual-Channel Rybak Neural Network and Consistency Verification in NSCT Domain" Fractal and Fractional 9, no. 7: 432. https://doi.org/10.3390/fractalfract9070432

APA Style

Lv, M., Song, S., Jia, Z., Li, L., & Ma, H. (2025). Multi-Focus Image Fusion Based on Dual-Channel Rybak Neural Network and Consistency Verification in NSCT Domain. Fractal and Fractional, 9(7), 432. https://doi.org/10.3390/fractalfract9070432

Article Metrics

Back to TopTop