Next Article in Journal
Factors Controlling Differences in Morphology and Fractal Characteristics of Organic Pores of Longmaxi Shale in Southern Sichuan Basin, China
Previous Article in Journal
Unsupervised Machine Learning-Based Singularity Models: A Case Study of the Taiwan Strait Basin
Previous Article in Special Issue
Estimation of Fractal Dimension and Segmentation of Body Regions for Deep Learning-Based Gender Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fractal Dimension-Based Multi-Focus Image Fusion via Coupled Neural P Systems in NSCT Domain

by
Liangliang Li
1,
Xiaobin Zhao
1,2,*,
Huayi Hou
3,
Xueyu Zhang
1,
Ming Lv
4,
Zhenhong Jia
4 and
Hongbing Ma
5
1
School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
2
Tianjin Key Laboratory of Autonomous Intelligence Technology and Systems, Tiangong University, Tianjin 300387, China
3
Hubei Key Laboratory of Optical Information and Pattern Recognition, Wuhan Institute of Technology, Wuhan 430205, China
4
School of Computer Science and Technology, Xinjiang University, Urumqi 830046, China
5
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Fractal Fract. 2024, 8(10), 554; https://doi.org/10.3390/fractalfract8100554
Submission received: 10 August 2024 / Revised: 5 September 2024 / Accepted: 7 September 2024 / Published: 25 September 2024

Abstract

:
In this paper, we introduce an innovative approach to multi-focus image fusion by leveraging the concepts of fractal dimension and coupled neural P (CNP) systems in nonsubsampled contourlet transform (NSCT) domain. This method is designed to overcome the challenges posed by the limitations of camera lenses and depth-of-field effects, which often prevent all parts of a scene from being simultaneously in focus. Our proposed fusion technique employs CNP systems with a local topology-based fusion model to merge the low-frequency components effectively. Meanwhile, for the high-frequency components, we utilize the spatial frequency and fractal dimension-based focus measure (FDFM) to achieve superior fusion performance. The effectiveness of the method is validated through extensive experiments conducted on three benchmark datasets: Lytro, MFI-WHU, and MFFW. The results demonstrate the superiority of our proposed multi-focus image fusion method, showcasing its potential to significantly enhance image clarity across the entire scene. Our algorithm has achieved advantageous values on metrics Q A B / F , Q C B , Q C V , Q E , Q F M I , Q G , Q M I , and Q N C I E .

1. Introduction

Multi-focus image fusion (MFIF) is a sophisticated process used in image processing that involves combining multiple images taken with different focus settings into a single image where all objects are in focus [1]. This technique is particularly useful in scenarios where the depth of field is limited and parts of the scene are out of focus in each image. By integrating these images, multi-focus image fusion produces a composite that is entirely sharp and clear [2].
The process of multi-focus image fusion involves several key techniques and steps: (1) Image registration: aligning the multiple images accurately is crucial, as even slight misalignments can lead to poor fusion results; (2) Focus measurement: this involves assessing the focus level of different parts of each image, often using clarity or sharpness metrics; (3) Fusion algorithm: the core of the process, where algorithms decide how to combine the sharp portions of each input image into the final composite. Techniques vary from simple averaging to complex wavelet-based methods; (4) Post-processing: enhancing the fused image to improve visual quality or to prepare it for analysis, including tasks like contrast adjustment or noise reduction [3,4].
MFIF is not without its challenges. These include handling misalignments, reducing artifacts that can arise during fusion, and dealing with variations in exposure and color balance among the source images. Advances in computational photography, machine learning, and deep learning have led to more sophisticated fusion algorithms that can more effectively address these challenges, resulting in higher-quality fused images. Various algorithms and techniques, including multi-resolution analysis, image decomposition, and feature-based methods, have been developed to effectively fuse multi-focus images. One of the key challenges in image fusion is to ensure that important information from both input images is preserved and enhanced in the fused image, without introducing artifacts or losing critical details [5,6,7].
In traditional image fusion algorithms, the multi-scale transforms such as nonsubsampled contourlet transform (NSCT) [8] and nonsubsampled shearlet transform (NSST) [9] are commonly employed. These are mathematical tools that are used for image analysis and processing. They offer an enhanced representation of local features and multi-scale analysis capabilities, making them particularly suitable for tasks like image fusion. Li et al. [10] introduced the medical image fusion approach using NSST. Lv et al. [11] proposed MFIF via parameter-adaptive pulse-coupled neural network and fractal dimension in NSST domain. Li et al. [12] introduced the image fusion algorithm based on spatial frequency and improved sum-modified-Laplacian in NSST domain. Coupled neural P (CNP) systems were proposed by Peng et al. [13] in 2019, and Li et al. [14] proposed the medical image fusion method based on coupled neural P systems (CNP) in NSST domain; the experiments show that the CNP systems achieve excellent results in image fusion. Although the algorithm achieves good image fusion results, it is only suitable for images of size m × m . For images of size m × n , due to the necessity of preprocessing the image with NSST to obtain an m × m or n × n image, followed by decomposition, applying different fusion rules to obtain the fused image, and finally resizing the image back to m × n , information loss and distortion occur to varying degrees during this process. Therefore, NSST is more suitable for the image fusion processing of m × m -sized images. NSCT can directly process images with size m × n , which can reduce image distortion and information loss. Li et al. [15] proposed the MFIF method via NSCT and achieved an excellent fusion effect.
To improve the clarity and information complementarity of the fused images, we propose a novel MFIF method based on fractal dimension and coupled neural P systems (CNP) in NSCT domain. The main contributions of our paper are as follows:
(1)
The coupled neural P systems (CNP) are used to process low-frequency components in order to obtain better background information;
(2)
A fractal dimension-based focus measure (FDFM) combined with spatial frequency (SF) is used to process high-frequency components, thereby obtaining more detailed image information;
(3)
Through extensive qualitative and quantitative experiments conducted on three datasets, our method consistently outperforms state-of-the-art (SOTA) techniques, demonstrating superior performance.
The remainder of the article is structured into six sections. Section 2 provides an overview of related works. In Section 3, the nonsubsampled contourlet transform is introduced. Section 4 presents the introduction of CNP systems. The proposed method is detailed in Section 5. Section 6 covers experimental results and discussions, while Section 7 delves into further discussions.

2. Related Works

In this section, we provide a concise overview of the existing literature on MFIF, categorizing it into three main approaches: spatial domain-based methods, transform domain-based methods, and deep learning-based methods [16].

2.1. Spatial Domain-Based Image Fusion

Spatial domain-based fusion methods involve the direct manipulation of pixel values using specific algorithms designed for image fusion. Among them, image fusion algorithms based on edge-preserving filtering are the most commonly used. These filters typically include guided image filtering [17], rolling guidance filtering [18], Gaussian curvature filter [19], etc. Fiza et al. [20] proposed a technique for MFIF specifically tailored for satellite images. It addresses issues such as visual distortion and spatial inconsistencies at sharp edges by introducing the edge discriminative diffusion filter (EDDF). EDDF combines anisotropic diffusion (AD) with guided filter (GF) to discriminate between local and global features, preserving edges while maintaining spatial consistency. The technique involves transforming source images into detail and base layers to extract low-pass and high-pass information, processing saliency maps through EDDF to generate weight maps, and, finally, combining fused detail and base layers to produce the desired fused image. Quantitative and qualitative tests demonstrate that the proposed technique outperforms some SOTA alternatives.
Yan et al. [21] introduced a novel MFIF approach, utilizing dictionary learning alongside a rolling guidance filter to accommodate both registered and mis-registered input images. Initially, a dictionary is learned from classical multi-focus images blurred by the rolling guidance filter. Then, a model is proposed for identifying focus regions by applying the learned dictionary to input images, generating focus feature maps. These maps are compared to derive an initial decision map, which is optimized and applied to the input images to produce fused images. Experimental results demonstrate the competitiveness of the proposed algorithm with the current state of the art, particularly excelling when handling both well-registered and mis-registered input images.
Adeel et al. [22] introduced a two-stage spatial domain framework for MFIF, which finds applications in computer vision. Initially, the salient features of focused regions are detected using Gaussian curvature filter (GCF) and range filtering. Subsequently, morphological filters refine the initial focus detection map. Experimental results demonstrate the method’s speed and robustness, outperforming recent multi-focus fusion schemes.
Tang et al. [23] introduced the image fusion technique that utilizes a simple weighted least squares filter. In their approach, source images are first decomposed into base and detail layers using this filter. The detail layers are then fused through a sub-window variance filter. For the base layer, they developed a fusion strategy that integrates visual saliency mapping with adaptive weight assignment techniques. Although the algorithm enhances the details, the fusion image as a whole suffers from some distortion.

2.2. Transform Domain-Based Image Fusion

Image fusion approaches via transform domains involve processing the transformed coefficients of source images using various transforms such as contourlet transform [24,25], shearlet transform [26,27], gradient domain [28], sparse representation [29,30], etc. Subsequently, these coefficients are inversely transformed back into the spatial domain. This approach typically comprises three primary stages: image transform, coefficient fusion, and inverse transform.
Jie et al. [31] proposed a method for enhancing full-field optical angiography (FFOA) images to achieve full focus, thereby improving its clinical utility. Existing FFOA techniques suffer from a limited depth of focus, resulting in partially unclear images. The method utilizes NSCT and contrast spatial frequency to fuse FFOA images effectively. Firstly, source images are decomposed into low-pass and bandpass images. Then, a sparse representation-based rule fuses low-pass images, while a contrast spatial frequency rule fuses bandpass images, considering pixel correlation and gradient relationships. Finally, a fully focused image is reconstructed. Experimental results demonstrate the superiority of the proposed method over SOTA approaches in both qualitative and quantitative evaluations, offering promising prospects for clinical applications in disease prevention and diagnosis.
Lu et al. [32] presented a novel focus measure called the sum of Gaussian-based fractional order differentiation (SGFD) to improve the accuracy of detecting low-frequency regions in MFIF. SGFD outperforms traditional focus measures by retaining more low-frequency information. The fusion process involves initial fusion using the NSST and SGFD, followed by refinement of the initial decision map through quadtree decomposition. Residual regions are identified and removed to generate the final fused image. Comparative experiments with SOTA approaches demonstrate that the SGFD-based approach excels in both subjective visual quality and objective metrics, offering promising advancements in MFIF.
Paul et al. [28] introduced a method for fusing color images that addresses both multi-exposure and multi-focus scenarios. It achieves this by blending the gradients of the luminance components of input images, prioritizing the maximum gradient magnitude at each pixel location. The fused luminance is then obtained through a Haar wavelet-based image reconstruction approach.
Tang et al. [33] proposed a sparse representation-based fusion approach for image fusion. Traditional methods often suffer from including redundant information during dictionary learning, leading to artifacts and increased computational time. To address this, the proposed method introduces a novel dictionary construction method based on joint patch grouping and informative sampling. Nonlocal similarity is utilized for joint patch grouping across all source images, simplifying the calculation by selecting only one class of informative image patches. Sparse coefficients are obtained using the orthogonal matching pursuit (OMP) algorithm, and a max-L1 fusion rule is employed for image reconstruction. Experimental results demonstrate the superiority of the proposed approach in effectively fusing multi-focus images.
Chen et al. [34] introduced an image fusion approach via a complex sparse representation (CSR) model. This model employs hypercomplex signal properties to derive directional information from real-valued signals by extending them into the complex domain. Following this, the directional aspects of the input signal are separated into sparse coefficients using corresponding directional dictionaries.

2.3. Deep Learning-Based Image Fusion

Deep learning has been widely applied in the field of image processing, including tasks such as image classification [35], image segmentation [36], object detection [37], image enhancement [38], image fusion [39], etc. The commonly used deep learning models have CNNs [40], GANs [41], autoencoders [42], etc. There has been extensive research into leveraging deep learning techniques for image fusion [43]. This has led to the emergence of numerous methods that utilize deep learning for MFIF. Zhang et al. [44] introduced the IFCNN fusion framework, a convolutional neural network (CNN)-based approach that offers a comprehensive fusion solution. Notably, it can undergo end-to-end training without necessitating preprocessing steps. Hu et al. [45] proposed a ZMFF method based on a deep prior network. Zhang et al. [46] introduced a fast unified image fusion network called PMGI. This network addresses various image fusion tasks like medical image fusion and MFIF. The key idea is to maintain the proportional relationship between texture and intensity information from source images. Xu et al. [47] introduced an innovative unified and unsupervised end-to-end image fusion network (U2Fusion).
Traditional algorithms have the following two advantages compared to deep learning: (1) Traditional algorithms typically operate based on mathematical models and specific rules, making their results easier to interpret and control; (2) Researchers can clearly understand how each step influences the final fusion outcome and can adjust parameters and methods according to their needs.

3. Nonsubsampled Contourlet Transform

The nonsubsampled contourlet transform (NSCT) is an advanced mathematical tool used for signal and image analysis, extending the principles of the contourlet transform but with significant improvements in handling images [48,49,50]. Developed to overcome some of the limitations of previous multi-scale and multi-directional transforms, the NSCT offers a flexible, multi-resolution, multi-direction, and shift-invariant framework for image decomposition. The NSCT decomposes images into components at multiple scales and orientations. It performs this through two main stages: a nonsubsampled pyramid (NSP) structure for capturing point discontinuities and a series of nonsubsampled directional filter banks (NSDFB) for linking point discontinuities into linear structures. This approach is adept at capturing edges and textures in an image, which are essential features in many image processing tasks. Figure 1a displays an overview of the proposed NSCT. The structure consists of a bank of filters that splits the 2-D frequency plane in the sub-bands illustrated in Figure 1b.
The NSCT requires selecting appropriate scale and directional parameters, which can significantly impact the final processing results and necessitate experience and experimentation to optimize. Due to its involvement with multi-scale and multidirectional data representations, NSCT requires high memory demands, especially when processing large-scale data.

4. Coupled Neural P Systems

For the fusion of multi-focus images, coupled neural P (CNP) systems are designed as an array of neurons with local topology, i.e., CNP systems with local topology [13,14].
A CNP system with local topology, of degree m × n , is delineated as follows:
Π = O , σ 11 , σ 12 , , σ 1 n , , σ m 1 , σ m 2 , , σ m n , syn
where
(1)
O = a is an alphabet (the objective a is known as the spike);
(2)
σ 11 , σ 12 , , σ m n are an array of m × n coupled neurons of the form
σ i j = x i j , y i j , z i j , R i j ,   1 i h ,   1 j w
where
(a)
x i j R is the value of spikes in feeding input unit in neuron σ i j ;
(b)
y i j R is the value of spikes in linking input unit in neuron σ i j ;
(c)
z i j R is the value of spikes in dynamic threshold unit in neuron σ i j ;
(d)
R i j denotes the finite set of spiking rules, of the form E / a x , a y , a z a p , where E is the firing condition, p 0 , τ 0 , and p u 1 + v .
(3)
syn = i j , k l | 1 i h , 1 j w , k i r , l j r , i k , j l , where r is the neighborhood radius.
Suppose that I is an input image with size h × w and I B C is the matrix containing the base components (BC) obtained from the image decomposition method. Figure 2 shows the relationship between CNP system Π and base components matrix I B C .
In CNP system Π , each coupled neuron only communicates with its neighboring neurons, i.e., local topology. An r-neighborhood of neurons σ i j is defined as follows, as shown in Figure 2:
δ r σ i j = σ k l k i r , l j r
The spiking rule of neuron σ i j is defined as follows [14]:
E / a x , a y , a z a p
where E is a firing condition denoted as follows [14]:
E n i t z i t x i t x y i t y z i t z
where n i j t = x i j t 1 + y i j t , which represents a nonlinear modulation mechanism. According to the spiking mechanism, the state equation for neuron σ i j can be given by
x i j t + 1 = x i j t x + C i j + σ k l δ r ω k l p k l t ,   if   σ i j   fires x i j t + C i j + σ k l δ r ω k l p k l t ,   otherwise
y i j t + 1 = y i j t y + σ k l δ r ω k l p k l t ,   if   σ i j   fires y i j t + σ k l δ r ω k l p k l t ,   otherwise
z i j t + 1 = z i j t z + p ,   if   σ i j   fires z i j t ,   otherwise
where p k l t is the value of the spikes received by neuron σ i j from neighboring neuron σ k l and ω k l t is the corresponding local weight, and C i j is an external stimulus. p is the value of spikes generated by neuron σ i j when it fires. Peng et al. [51] introduced the MFIF method based on CNP systems in NSCT domain, and the algorithm has achieved good fusion results.

5. The Proposed Method

We present a novel approach for MFIF utilizing fractal dimension and coupled neural P systems in NSCT domain. The main steps can be concluded as follows: NSCT decomposition, low-frequency coefficient fusion, high-frequency coefficient fusion, and inverse NSCT transform. Figure 3 illustrates the architecture of the proposed method.

5.1. NSCT Decomposition

Assume that A and B are the input images, the NSCT is used to decompose the two images, and the low- and high-frequency components are generated, named as L A , H A l , k and L B , H B l , k , respectively.

5.2. Low-Frequency Coefficient Fusion

The low-frequency components have the most brightness and energy information; the fusion strategy for low-frequency components has a significant impact on the final fusion quality. In this section, the CNP systems-based low-frequency fusion rule is designed. Suppose that A and B are two CNP systems with local topology; the low-frequency coefficients of two multi-focus images are regarded as the external inputs of A and B . Starting from the initial state, the two CNP systems work constantly until iteration number t m a x is reached. Then, they halt. Denoted by T A and T B , the excitation number matrixes are associated with A and B , i.e., T A = t i j A h × w and T B = t i j B h × w , where t i j A (or t i j B ) is the number of times that σ i j fires in A (or B ). The fusion rules for low-frequency NSCT coefficients are formulated based on the information provided by the two excitation number matrices as follows [14]:
L F i , j = L A i , j if   t i j A t i j B L B i , j i f   t i j A < t i j B
where L A i , j and L B i , j are the low-frequency coefficients of two source images at position i , j , respectively, and L F i , j is the fused low-frequency coefficients.

5.3. High-Frequency Coefficient Fusion

The high-frequency coefficients contain richer texture and detailed information, as well as some of the image’s noise. Fractal and fractional have extensive applications in image processing [52,53,54,55,56,57,58,59,60,61,62,63]. In this section, the spatial frequency (SF) [64] and fractal dimension-based focus measure (FDFM) [52] are defined as follows:
S F H A l , k i , j = i M , j N H A l , k i , j H A l , k i 1 , j 2 + H A l , k i , j H A l , k i , j 1 2
S F H B l , k i , j = i M , j N H B l , k i , j H B l , k i 1 , j 2 + H B l , k i , j H B l , k i , j 1 2
F D F M H A l , k i , j = g max H A l , k i , j g min H A l , k i , j
F D F M H B l , k i , j = g max H B l , k i , j g min H B l , k i , j
where g m a x X i , j and g min X i , j are the maximum and minimum intensities, respectively, over a 3 × 3 window centered at the i , j t h pixel of X H A l , k , H B l , k . FDFM can measure such small activity levels, which also affect the fusion results.
The fused high-frequency coefficients are generated by
H F l , k i , j = H A l , k i , j if   S F H A l , k i , j × F D F M H A l , k i , j   S F H B l , k i , j × F D F M H B l , k i , j H B l , k i , j e l s e
where H F l , k i , j shows the fused high-frequency coefficients.

5.4. Inverse NSCT Transform

The final fused image F can be generated by using inverse NSCT transform performed on the fused low- and high-frequency coefficients L F , H F l , k .
The main steps of the proposed method can be summarized as in Algorithm 1.
Algorithm 1 Proposed MFIF method
Input: the source images: A and B
Parameters: The number of NSCT decomposition levels: L , the number of directions at each decomposition level: K l , l 1 , L
Main step:
Step 1: NSCT decomposition
For each source image X A , B
   Perform NSCT decomposition on X to generate L X , H X l , k , l 1 , L , k 1 , K l ;
End
Step 2: Low-frequency components fusion
For each source image X A , B
   Calculate the CNP for L X using Equations (1)–(7);
End
Merge L A and L B using Equation (8) to generate L F ;
Step 3: High-frequency components fusion
For each level l = 1 : L
   For each direction k = 1 : K l
   For each source image X A , B
   Calculate the SF for H X l , k i , j using Equations (9) and (10);
   Calculate the FDFM for H X l , k i , j using Equations (11) and (12);
     End
     Merge H A l , k and H B l , k using Equation (13);
   End
End
Step 4: Inverse NSCT
Perform inverse NSCT on L F , H F l , k to generate F ;
Output: the fused image F .

6. Experimental Results and Discussion

6.1. Experimental Setup

In this section, the classical Lytro [65], MFI-WHU [66], and MFFW [67] datasets are used in the experiments, as shown in Figure 4. Eight image fusion methods, named GD [28], PMGI [46], MFFGAN [66], LEGFF [68], U2Fusion [47], CBFM [69], FUFusion [70], and EgeFusion [23], are used to compare. The metrics Q A B / F [64,71], Q C B [72], Q C V [72], Q E [72], Q F M I [73], Q G [72], Q M I [64], and Q N C I E [72] are used to evaluate the fusion results. The source code of these compared methods is available or shared by the authors, and the relevant parameter settings are set according to the original papers. In our method, the NSCT decomposition levels is 4, and the corresponding directions are 2, 4, 8, and 8. The parameters in CNP systems are set to i m a x , z 0 = 0.3 , r = 7 , and p = 1 . W 7 × 7 = ω i j 7 × 7 is determined as follows: (i) ω 44 = 0 ; (ii) ω i j = 1 / sqrt i 4 2 + j 4 2 for i 4 , j 4 , 1 i 7 , 1 j 7 .

6.2. Fusion Results and Discussion

(1)
Results on the Lytro Dataset
We conduct qualitative and quantitative evaluations to evaluate the performance of different image fusion methods. Figure 5 displays the fused images of Data 1 from the Lytro dataset. The fused image generated by GD exhibits over-brightness in certain areas, leading to the loss of some fine details and the presence of artifacts. The PMGI algorithm results in severe distortion, with the visual effect of the image appearing dim and blurry, and a significant loss of information. The brightness of the image generated by MFFGAN is insufficient, especially in areas such as the hair. The LEGFF algorithm achieved a relatively good fusion result, but there is slight darkness observed in the brightness of the hair section. The U2Fusion algorithm generated an image with some dim areas, particularly noticeable in regions like the hat and hair, as well as the watch section, where usable information is not observable. The CBFM generates a fused image that darkens the area around the arm. The FUFusion produces a blurry fused image, making it difficult to obtain some image information. The EgeFusion results in a significant distortion and loss of image information in the fused image. Compared to other algorithms, our method demonstrates a superior fusion result, featuring higher clarity and moderate brightness. Information pertaining to arm, watch, clothing, hat, golf club, and grass in the image is distinctly observable.
Table 1 presents the numerical values of various indicators corresponding to different algorithms depicted in Figure 5. These indicators serve as quantitative metrics for evaluating the performance of each algorithm in the context of the depicted data. By examining these values, we gain insights into the effectiveness of each algorithm in achieving the desired objectives. From Table 1, it is evident that our algorithm has achieved optimal values for the eight indicators, with respective values of Q A B / F (0.7524), Q C B (0.7745), Q C V (6.5508), Q E (0.8862), Q F M I (0.9380), Q G (0.7382), Q M I (6.5466), and Q N C I E (0.8245).
Figure 6 displays the fused images of Data 2 from the Lytro dataset. The image clarity obtained by the GD and PMGI algorithms is not high. The MFFGAN, LEGFF, and U2Fusion methods produce unevenly fused images, in which some regions are clear while others are dark, such as the neck area being too dim, obscuring detailed information. The fusion image generated by CBFM is dark and exhibits low brightness and clarity, resulting in a significant loss of information. The fused image produced by the FUFusion method is overall blurry, with a severe loss of detailed information. The EgeFusion algorithm produces an image with significant distortion, although some sharpening is applied to enhance detailed information; this sharpening also results in varying degrees of block effects in the image. Compared to other algorithms, our algorithm achieved the best fusion result, including image brightness, contrast, etc., allowing for a genuine observation of people and objects in the image.
Table 2 presents the numerical values of various indicators corresponding to different algorithms depicted in Figure 6. By analyzing the data in Table 2, we can conclude that our algorithm has achieved the optimal values in all eight indicators, with values as follows: Q A B / F (0.7445), Q C B (0.6870), Q C V (5.3153), Q E (0.8672), Q F M I (0.8732), Q G (0.7369), Q M I (7.6750), and Q N C I E (0.8339).
Figure 7 depicts a line chart of the metrics for different data in the Lytro dataset (20 sets of data were used in this experiment), allowing for an observation of fluctuations in the indicator data. Additionally, we have computed the average indicators, as shown in both Figure 7 and Table 3. Considering that a lower value of Q C V indicates better performance, we take its negative value (i.e., Q C V ) to illustrate the sub-figure of Q C V . The horizontal axis represents the number of image groups in the dataset, and the vertical axis represents the metric value. From the Figure 7 and Table 3, we can observe that our method has achieved the optimal average values for all the indicators, with values as follows: Q A B / F (0.7390), Q C B (0.7388), Q C V (31.4085), Q E (0.8772), Q F M I (0.8989), Q G (0.7362), Q M I (6.9683), and Q N C I E (0.8296).
(2)
Results on the MFI-WHU Dataset
Figure 8 depicts the fused images of Data 1 from the MFI-WHU dataset. The images produced by GD exhibit some slight pseudo-shadow artifacts. The PMGI algorithm generates blurry fused images; for instance, the details of the roof and mural cannot be accurately captured. The MFFGAN and U2Fusion algorithms achieve higher brightness fusion results in certain regions, such as the glass display area. However, there are also areas with lower brightness. For instance, MFFGAN exhibits shadows around the base of the display cabinet, while U2Fusion produces significant shadowing along the edges of the roof and the bench, hindering the retrieval of complete information. The fused image produced by LEGFF has a higher clarity. The fused images produced by the CBFM and FUFusion algorithms have some areas with shadows, such as on the side of a bench. Although the EgeFusion algorithm enhances the texture information in the image, it also introduces some degree of distortion. Upon comprehensive comparison, our algorithm outperforms others in the fusion experiment, enhancing the overall fusion effect by reducing pseudo-shadows and dark areas, while simultaneously improving the brightness and clarity of the fused image.
Table 4 presents the numerical values of various indicators corresponding to different algorithms depicted in Figure 8. From Table 4, it is evident that our algorithm has achieved optimal values for the eight indicators, with respective values of Q A B / F (0.7281), Q C B (0.8256), Q C V (6.4076), Q E (0.8534), Q F M I (0.9000), Q G (0.7288), Q M I (9.0964), and Q N C I E (0.8476).
Figure 9 depicts the fused images of Data 2 from the MFI-WHU dataset. The GD method produced a fused image with low clarity. The PMGI method generated a low-quality fused image with varying degrees of shadowing in objects such as the vehicle, windowsill, sky, and trees. The images produced by the MFFGAN and LEGFF algorithms cause the windowsill area to become darker. The U2Fusion method generated fused image with uneven brightness distribution, such as a darker area in the windowsill region and a brighter area in the vehicle region. The fused images produced by the CBFM and FUFusion algorithms have relatively low brightness. The EgeFusion algorithm causes distortion in the fused image. Through comprehensive comparison, our algorithm produces the optimal fusion result, with balanced brightness and high clarity, which are beneficial for information retrieval.
Table 5 presents the numerical values of various indicators corresponding to different algorithms depicted in Figure 9. From Table 5, it is evident that our algorithm has achieved optimal values for seven indicators, with respective values of Q A B / F (0.7408), Q C B (0.8250), Q E (0.8550), Q F M I (0.8795), Q G (0.7409), Q M I (7.8104), and Q N C I E (0.8348). The FUFusion algorithm achieved the optimal value on the Q C V metric, with a value of 30.6077. Our algorithm ranked second in the Q C V metric, with a value of 38.0683.
Figure 10 depicts a line chart of the metrics of different data in the MFI-WHU dataset (30 sets of data were used in this experiment), allowing for the observation of fluctuations in the indicator data. Additionally, we have computed the average indicators, as shown in both Figure 10 and Table 6. We can observe that our algorithm has achieved optimal values for seven indicators, with respective values of Q A B / F (0.7296), Q C B (0.8072), Q E (0.8453), Q F M I (0.8772), Q G (0.7254), Q M I (7.8107), and Q N C I E (0.8371). The FUFusion algorithm achieved the optimal value on the Q C V metric, with a value of 23.7140. Our algorithm ranked second in the Q C V metric, with a value of 36.4954.
(3)
Results on the MFFW Dataset
Figure 11 depicts the fused images of Data 1 from the MFFW dataset. The fused images produced by the GD, PMGI, and FUFusion algorithms are blurry and have lower clarity. The fused images produced by the MFFGAN and LEGFF methods cause the center of the rose to appear darker, with shadowing present. The U2Fusion causes the piano keys and the rose in the fused image to appear darker. The fused images produced by the CBFM and FUFusion algorithms exhibit varying degrees of artifacts. Although the EgeFusion algorithm enhances the texture information in the fused image, it also introduces a certain degree of distortion. By comparison with other algorithms, our method achieves a superior fusion result, attaining effective information complementarity.
Table 7 presents the numerical values of various indicators corresponding to different algorithms depicted in Figure 11. From Table 7, it is evident that our algorithm has achieved optimal values for six indicators, with respective values of Q C B (0.6155), Q C V (33.3908), Q E (0.8290), Q F M I (0.8987), Q M I (6.6801), and Q N C I E (0.8290). The GD algorithm achieved optimal values for the Q A B / F and Q G metrics, with values of 0.7297 and 0.6829, respectively.
Figure 12 depicts a line chart of the metrics of different data in the MFFW dataset (13 sets of data were used in this experiment), allowing for the observation of fluctuations in the indicator data. Additionally, we have computed the average indicators, as shown in both Figure 12 and Table 8. We can observe that our algorithm has achieved optimal values for five indicators, with respective values of Q A B / F (0.6377), Q C V (116.0756) Q E (0.8048), Q F M I (0.8799), and Q G (0.6224). The CBFM algorithm achieved the optimal value for the Q C B metric, with a value of 0.6408. Our algorithm ranked second in the Q C B metric, with a value of 0.6362. The FUFusion algorithm achieved the optimal value on the Q M I and Q N C I E metrics, with values of 5.3456 and 0.8196, respectively. Our method ranked second in these two metrics, with values of 5.0505 and 0.8180, respectively.

6.3. Application Extension

In this section, we extend the proposed algorithm to the application of multi-modal medical image fusion. We utilized two medical datasets from the Whole Brain Atlas [74]. When processing color medical images (such as magnetic resonance (MR) and positron emission tomography (PET)), the PET images need to be converted between the RGB and YUV color spaces. The Y channel of the PET image is fused with the MR image, and then the fused Y channel is converted back to the RGB space along with the U and V channels of the PET image to obtain the final color fusion image. The corresponding schematic diagram is shown in Figure 13. From the experimental results shown in Figure 14, it is evident that our algorithm performs exceptionally well for medical image fusion, achieving significantly enhanced information complementarity in the fused images.

7. Conclusions

This paper presents a novel approach to multi-focus image fusion that integrates fractal dimension and CNP systems in NSCT domain. Our fusion technique utilizes CNP systems with a local topology-based fusion model to effectively merge low-frequency components. Additionally, for high-frequency components, we employ a focus measure based on spatial frequency and FDFM to achieve superior fusion performance. Extensive experiments conducted on three classic datasets (Lytro, MFI-WHU, and MFFW), validate the effectiveness of the proposed method. In security surveillance, fusing images with different focal points can enhance the detail and clarity of the monitoring footage, improving target recognition capabilities. Furthermore, we have extended and applied this method to the field of medical image fusion, achieving the complementary integration of multi-modal medical image information. The number of decomposition levels in NSCT affects the effectiveness of image fusion. However, too many levels increase algorithm complexity. Therefore, choosing an optimal number of decomposition levels is our research focus. In future work, we will attempt to improve this algorithm and expand its application to the fusion of SAR (synthetic aperture radar) and optical images [48]. Additionally, the application of image fusion in multimodal finger knuckle print identification and image change detection are also worthwhile directions to explore [75,76,77,78,79].

Author Contributions

The experimental measurements and data collection were carried out by L.L., X.Z. (Xiaobin Zhao), H.H., X.Z. (Xueyu Zhang), M.L., Z.J., and H.M. The manuscript was written by L.L. with the assistance of X.Z. (Xiaobin Zhao), H.H., X.Z. (Xueyu Zhang), M.L., Z.J., and H.M. Writing—original draft, L.L. and H.M.; Writing—review & editing, X.Z. (Xiaobin Zhao), H.H., X.Z. (Xueyu Zhang), M.L. and H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 62261053; the Tianshan Talent Training Project-Xinjiang Science and Technology Innovation Team Program (2023TSYCTD0012); the Cross-Media Intelligent Technology Project of Beijing National Research Center for Information Science and Technology (BNRist) under Grant No. BNR2019TD01022; and the Hubei Key Laboratory of Optical Information and Pattern Recognition, Wuhan Institute of Technology under Grant No. 202305; and the Open Project of Tianjin Key Laboratory of Autonomous Intelligence Technology and Systems (No. AITS-20240001).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to thank the Tianjin Key Laboratory of Autonomous Intelligence Technology and Systems (Tiangong University) for the financial support (No. AITS-20240001).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, C.; Yan, K. Focus-aware and deep restoration network with transformer for multi-focus image fusion. Digit. Signal Process. 2024, 149, 104473. [Google Scholar] [CrossRef]
  2. Zhang, J.; Liao, Q.; Ma, H. Exploit the best of both end-to-end and map-based methods for multi-focus image fusion. IEEE Trans. Multimed. 2024, 26, 6411–6423. [Google Scholar] [CrossRef]
  3. Qiao, L.; Wu, S.; Xiao, B. Boosting robust multi-focus image fusion with frequency mask and hyperdimensional computing. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 3538–3550. [Google Scholar] [CrossRef]
  4. Li, B.; Zhang, L.; Liu, J.; Peng, H. Multi-focus image fusion with parameter adaptive dual channel dynamic threshold neural P systems. Neural Netw. 2024, 179, 106603. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  6. Lv, M.; Li, L.; Jin, Q.; Jia, Z.; Chen, L.; Ma, H. Multi-focus image fusion via distance-weighted regional energy and structure tensor in NSCT domain. Sensors 2023, 23, 6135. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, Y.; Qi, Z.; Cheng, J.; Chen, X. Rethinking the effectiveness of objective evaluation metrics in multi-focus image fusion: A statistic-based approach. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 5806–5819. [Google Scholar] [CrossRef] [PubMed]
  8. Wang, G.; Li, J. Fusion of full-field optical angiography images via gradient feature detection. Front. Phys. 2024, 12, 1397732. [Google Scholar] [CrossRef]
  9. Wu, M.; Yang, L.; Chai, R. Research on multi-scale fusion method for ancient bronze ware X-ray images in NSST domain. Appl. Sci. 2024, 14, 4166. [Google Scholar] [CrossRef]
  10. Li, L.; Wang, L.; Wang, Z.; Jia, Z. A novel medical image fusion approach based on nonsubsampled shearlet transform. J. Med. Imaging Health Inform. 2019, 9, 1815–1826. [Google Scholar]
  11. Lv, M.; Jia, Z.; Li, L.; Ma, H. Multi-focus image fusion via PAPCNN and fractal dimension in NSST domain. Mathematics 2023, 11, 3803. [Google Scholar] [CrossRef]
  12. Li, L.; Si, Y.; Wang, L.; Jia, Z.; Ma, H. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  13. Peng, H.; Wang, J. Coupled neural P systems. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 1672–1682. [Google Scholar] [CrossRef]
  14. Li, B.; Peng, H. Medical image fusion method based on coupled neural P systems in nonsubsampled shearlet transform domain. Int. J. Neural Syst. 2021, 31, 2050050. [Google Scholar] [CrossRef] [PubMed]
  15. Li, L.; Ma, H.; Jia, Z.; Si, Y. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 12389–12409. [Google Scholar] [CrossRef]
  16. Qi, Y.; Yang, Z. A multi-channel neural network model for multi-focus image fusion. Expert Syst. Appl. 2024, 247, 123244. [Google Scholar] [CrossRef]
  17. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  18. Li, L.; Lv, M.; Jia, Z.; Jin, Q.; Liu, M.; Chen, L.; Ma, H. An effective infrared and visible image fusion approach via rolling guidance filtering and gradient saliency map. Remote Sens. 2023, 15, 2486. [Google Scholar] [CrossRef]
  19. Huo, X.; Deng, Y.; Shao, K. Infrared and visible image fusion with significant target enhancement. Entropy 2022, 24, 1633. [Google Scholar] [CrossRef]
  20. Fiza, S.; Safinaz, S. Multi-focus image fusion using edge discriminative diffusion filter for satellite images. Multimed. Tools Appl. 2024, 83, 66087–66106. [Google Scholar] [CrossRef]
  21. Yan, X.; Qin, H.; Li, J. Multi-focus image fusion based on dictionary learning with rolling guidance filter. J. Opt. Soc. Am. A-Opt. Image Sci. Vis. 2017, 34, 432–440. [Google Scholar] [CrossRef]
  22. Adeel, H.; Riaz, M. Multi-focus image fusion using curvature minimization and morphological filtering. Multimed. Tools Appl. 2024, 83, 78625–78639. [Google Scholar] [CrossRef]
  23. Tang, H.; Liu, G.; Qian, Y. EgeFusion: Towards edge gradient enhancement in infrared and visible image fusion with multi-scale transform. IEEE Trans. Comput. Imaging 2024, 10, 385–398. [Google Scholar] [CrossRef]
  24. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [PubMed]
  25. Li, L.; Ma, H. Pulse coupled neural network-based multimodal medical image fusion via guided filtering and WSEML in NSCT domain. Entropy 2021, 23, 591. [Google Scholar] [CrossRef]
  26. Guo, K.; Labate, D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007, 39, 298–318. [Google Scholar] [CrossRef]
  27. Li, L.; Ma, H. Saliency-guided nonsubsampled shearlet transform for multisource remote sensing image fusion. Sensors 2021, 21, 1756. [Google Scholar] [CrossRef]
  28. Paul, S.; Sevcenco, I.; Agathoklis, P. Multi-exposure and multi-focus image fusion in gradient domain. J. Circuits Syst. Comput. 2016, 25, 1650123. [Google Scholar] [CrossRef]
  29. Li, L.; Lv, M.; Jia, Z.; Ma, H. Sparse representation-based multi-focus image fusion method via local energy in shearlet domain. Sensors 2023, 23, 2888. [Google Scholar] [CrossRef]
  30. Luo, Y.; Luo, Z. Infrared and visible image fusion: Methods, datasets, applications, and prospects. Appl. Sci. 2023, 13, 10891. [Google Scholar] [CrossRef]
  31. Jie, Y.; Li, X.; Wang, M.; Tan, H. Multi-focus image fusion for full-field optical angiography. Entropy 2023, 25, 951. [Google Scholar] [CrossRef] [PubMed]
  32. Lu, J.; Tan, K. Multi-focus image fusion using residual removal and fractional order differentiation focus measure. Signal Image Video Process. 2024, 18, 3395–3410. [Google Scholar] [CrossRef]
  33. Tang, D.; Xiong, Q. A novel sparse representation based fusion approach for multi-focus images. Expert Syst. Appl. 2022, 197, 116737. [Google Scholar] [CrossRef]
  34. Chen, Y.; Liu, Y.; Ward, R.K.; Chen, X. Multi-focus image fusion with complex sparse representation. IEEE Sens. J. 2024. Early Access. [Google Scholar]
  35. Shen, D.; Hu, H.; He, F.; Zhang, F.; Zhao, J.; Shen, X. Hierarchical prototype-aligned graph neural network for cross-scene hyperspectral image classification. Remote Sens. 2024, 16, 2464. [Google Scholar] [CrossRef]
  36. Akram, R.; Hong, J.S.; Kim, S.G. Crop and weed segmentation and fractal dimension estimation using small training data in heterogeneous data environment. Fractal Fract. 2024, 8, 285. [Google Scholar] [CrossRef]
  37. Zhou, M.; Li, B.; Wang, J. Optimization of hyperparameters in object detection models based on fractal loss function. Fractal Fract. 2022, 6, 706. [Google Scholar] [CrossRef]
  38. Zhao, P.; Zheng, H.; Tang, S. DAMNet: A dual adjacent indexing and multi-deraining network for real-time image deraining. Fractal Fract. 2023, 7, 24. [Google Scholar] [CrossRef]
  39. Fang, J.; Ning, X. A multi-focus image fusion network combining dilated convolution with learnable spacings and residual dense network. Comput. Electr. Eng. 2024, 117, 109299. [Google Scholar] [CrossRef]
  40. Wang, S.; Chen, Z.; Qi, F. Fractal geometry and convolutional neural networks for the characterization of thermal shock resistances of ultra-high temperature ceramics. Fractal Fract. 2022, 6, 605. [Google Scholar] [CrossRef]
  41. Sun, H.; Wu, S.; Ma, L. Adversarial attacks on GAN-based image fusion. Inf. Fusion 2024, 108, 102389. [Google Scholar] [CrossRef]
  42. Yu, Y.; Qin, C. An end-to-end underwater-image-enhancement framework based on fractional integral retinex and unsupervised autoencoder. Fractal Fract. 2023, 7, 70. [Google Scholar] [CrossRef]
  43. Zhang, X. Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 4819–4838. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Liu, Y.; Sun, P. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  45. Hu, X.; Jiang, J.; Liu, X.; Ma, J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion 2023, 92, 127–138. [Google Scholar] [CrossRef]
  46. Zhang, H.; Xu, H.; Xiao, Y. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar]
  47. Xu, H.; Ma, J.; Jiang, J. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  48. Li, J.; Zhang, J.; Yang, C.; Liu, H.; Zhao, Y.; Ye, Y. Comparative analysis of pixel-level fusion algorithms and a new high-resolution dataset for SAR and optical image fusion. Remote Sens. 2023, 15, 5514. [Google Scholar] [CrossRef]
  49. Li, L.; Si, Y.; Jia, Z. Remote sensing image enhancement based on non-local means filter in NSCT domain. Algorithms 2017, 10, 116. [Google Scholar] [CrossRef]
  50. Li, L.; Si, Y.; Jia, Z. A novel brain image enhancement method based on nonsubsampled contourlet transform. Int. J. Imaging Syst. Technol. 2018, 28, 124–131. [Google Scholar] [CrossRef]
  51. Peng, H.; Li, B. Multi-focus image fusion approach based on CNP systems in NSCT domain. Comput. Vis. Image Underst. 2021, 210, 103228. [Google Scholar] [CrossRef]
  52. Panigrahy, C.; Seal, A.; Mahato, N.K. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106141. [Google Scholar] [CrossRef]
  53. Zhang, X.; Boutat, D.; Liu, D. Applications of fractional operator in image processing and stability of control systems. Fractal Fract. 2023, 7, 359. [Google Scholar] [CrossRef]
  54. Zhang, X.; Dai, L. Image enhancement based on rough set and fractional order differentiator. Fractal Fract. 2022, 6, 214. [Google Scholar] [CrossRef]
  55. Zhang, X.; Chen, S.; Zhang, J. Adaptive sliding mode consensus control based on neural network for singular fractional order multi-agent systems. Appl. Math. Comput. 2022, 434, 127442. [Google Scholar] [CrossRef]
  56. Zhang, X.; Lin, C. A unified framework of stability theorems for LTI fractional order systems with 0 < α < 2. IEEE Trans. Circuit Syst. II-Express 2020, 67, 3237–3241. [Google Scholar]
  57. Di, Y.; Zhang, J.; Zhang, X. Robust stabilization of descriptor fractional-order interval systems with uncertain derivative matrices. Appl. Math. Comput. 2023, 453, 128076. [Google Scholar] [CrossRef]
  58. Zhang, X.; Chen, Y. Admissibility and robust stabilization of continuous linear singular fractional order systems with the fractional order α: The 0 < α < 1 case. ISA Trans. 2018, 82, 42–50. [Google Scholar]
  59. Zhang, J.; Yang, G. Low-complexity tracking control of strict-feedback systems with unknown control directions. IEEE Trans. Autom. Control 2019, 64, 5175–5182. [Google Scholar] [CrossRef]
  60. Zhang, J.; Wang, Q.; Ding, W. Global output-feedback prescribed performance control of nonlinear systems with unknown virtual control coefficients. IEEE Trans. Autom. Control 2022, 67, 6904–6911. [Google Scholar] [CrossRef]
  61. Zhang, J.; Ding, J.; Chai, T. Fault-tolerant prescribed performance control of wheeled mobile robots: A mixed-gain adaption approach. IEEE Trans. Autom. Control 2024, 69, 5500–5507. [Google Scholar] [CrossRef]
  62. Zhang, J.; Xu, K.; Wang, Q. Prescribed performance tracking control of time-delay nonlinear systems with output constraints. IEEE/CAA J. Autom. Sin. 2024, 11, 1557–1565. [Google Scholar] [CrossRef]
  63. Di, Y.; Zhang, J.-X.; Zhang, X. Alternate admissibility LMI criteria for descriptor fractional order systems with 0 < α < 2. Fractal Fract. 2023, 7, 577. [Google Scholar] [CrossRef]
  64. Qu, X.; Yan, J.; Xiao, H. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  65. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  66. Zhang, H.; Le, Z. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  67. Xu, S.; Wei, X.; Zhang, C. MFFW: A newdataset for multi-focus image fusion. arXiv 2020, arXiv:2002.04780. [Google Scholar]
  68. Zhang, Y.; Xiang, W. Local extreme map guided multi-modal brain image fusion. Front. Neurosci. 2022, 16, 1055451. [Google Scholar] [CrossRef] [PubMed]
  69. Li, X.; Li, X.; Liu, W. CBFM: Contrast balance infrared and visible image fusion based on contrast-preserving guided filter. Remote Sens. 2023, 15, 2969. [Google Scholar] [CrossRef]
  70. Jie, Y.; Chen, Y.; Li, X.; Yi, P.; Tan, H.; Cheng, X. FUFusion: Fuzzy sets theory for infrared and visible image fusion. Lect. Notes Comput. Sci. 2024, 14426, 466–478. [Google Scholar]
  71. Yang, H.; Zhang, J.; Zhang, X. Injected infrared and visible image fusion via L1 decomposition model and guided filtering. IEEE Trans. Comput. Imaging 2022, 8, 162–173. [Google Scholar]
  72. Liu, Z.; Blasch, E.; Xue, Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 94–109. [Google Scholar] [CrossRef]
  73. Haghighat, M.; Razian, M. Fast-FMI: Non-reference image fusion metric. In Proceedings of the IEEE 8th International Conference on Application of Information and Communication Technologies, Astana, Kazakhstan, 15–17 October 2014; pp. 424–426. [Google Scholar]
  74. Available online: http://www.med.harvard.edu/AANLIB/home.html (accessed on 1 March 2024).
  75. Aiadi, O.; Khaldi, B.; Korichi, A. Fusion of deep and local gradient-based features for multimodal finger knuckle print identification. Clust. Comput. 2024, 27, 7541–7557. [Google Scholar] [CrossRef]
  76. Li, L.; Ma, H.; Jia, Z. Multiscale geometric analysis fusion-based unsupervised change detection in remote sensing images via FLICM model. Entropy 2022, 24, 291. [Google Scholar] [CrossRef] [PubMed]
  77. Li, L.; Ma, H.; Zhang, X.; Zhao, X.; Lv, M.; Jia, Z. Synthetic aperture radar image change detection based on principal component analysis and two-level clustering. Remote Sens. 2024, 16, 1861. [Google Scholar] [CrossRef]
  78. Li, L.; Ma, H.; Jia, Z. Change detection from SAR images based on convolutional neural networks guided by saliency enhancement. Remote Sens. 2021, 13, 3697. [Google Scholar] [CrossRef]
  79. Li, L.; Ma, H.; Jia, Z. Gamma correction-based automatic unsupervised change detection in SAR images via FLICM model. J. Indian Soc. Remote Sens. 2023, 51, 1077–1088. [Google Scholar] [CrossRef]
Figure 1. Nonsubsampled contourlet transform. (a) Nonsubsampled filter bank structure that implements the NSCT; (b) Idealized frequency partitioning obtained with the proposed structure.
Figure 1. Nonsubsampled contourlet transform. (a) Nonsubsampled filter bank structure that implements the NSCT; (b) Idealized frequency partitioning obtained with the proposed structure.
Fractalfract 08 00554 g001
Figure 2. A CNP system Π and the corresponding base component matrix IBC.
Figure 2. A CNP system Π and the corresponding base component matrix IBC.
Fractalfract 08 00554 g002
Figure 3. The structure of the proposed method.
Figure 3. The structure of the proposed method.
Fractalfract 08 00554 g003
Figure 4. Examples of the Lytro, MFI-WHU, and MFFW datasets.
Figure 4. Examples of the Lytro, MFI-WHU, and MFFW datasets.
Fractalfract 08 00554 g004
Figure 5. Visual comparison for Data 1 in the Lytro dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Figure 5. Visual comparison for Data 1 in the Lytro dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Fractalfract 08 00554 g005aFractalfract 08 00554 g005b
Figure 6. Visual comparison for Data 2 in the Lytro dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Figure 6. Visual comparison for Data 2 in the Lytro dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Fractalfract 08 00554 g006
Figure 7. The line chart illustrates the metrics of various data in the Lytro dataset.
Figure 7. The line chart illustrates the metrics of various data in the Lytro dataset.
Fractalfract 08 00554 g007aFractalfract 08 00554 g007b
Figure 8. Visual comparison for Data 1 in the MFI-WHU dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Figure 8. Visual comparison for Data 1 in the MFI-WHU dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Fractalfract 08 00554 g008
Figure 9. Visual comparison for Data 2 in the MFI-WHU dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Figure 9. Visual comparison for Data 2 in the MFI-WHU dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Fractalfract 08 00554 g009
Figure 10. The line chart illustrates the metrics of various data in the MFI-WHU dataset.
Figure 10. The line chart illustrates the metrics of various data in the MFI-WHU dataset.
Fractalfract 08 00554 g010aFractalfract 08 00554 g010b
Figure 11. Visual comparison for Data 1 in the MFFW dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Figure 11. Visual comparison for Data 1 in the MFFW dataset. (a) GD; (b) PMGI; (c) MFFGAN; (d) LEGFF; (e) U2Fusion; (f) CBFM; (g) FUFusion; (h) EgeFusion; (i) Proposed.
Fractalfract 08 00554 g011
Figure 12. The line chart illustrates the metrics of various data in the MFFW dataset.
Figure 12. The line chart illustrates the metrics of various data in the MFFW dataset.
Fractalfract 08 00554 g012aFractalfract 08 00554 g012b
Figure 13. Structure of the proposed method for multi-modal medical images in the YUV color space.
Figure 13. Structure of the proposed method for multi-modal medical images in the YUV color space.
Fractalfract 08 00554 g013
Figure 14. The fusion results on medical images. (a) source A; (b) source B; (c) Proposed.
Figure 14. The fusion results on medical images. (a) source A; (b) source B; (c) Proposed.
Fractalfract 08 00554 g014
Table 1. Quantitative comparative analysis of different methods for Data 1 in the Lytro dataset.
Table 1. Quantitative comparative analysis of different methods for Data 1 in the Lytro dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.72200.668463.58140.81440.92220.69853.11610.8096
PMGI20200.54660.607070.27850.63160.91690.51565.13470.8169
MFFGAN20210.68600.702623.34390.84510.92960.65995.57830.8190
LEGFF20220.69230.685738.61560.82050.93060.66584.89190.8158
U2Fusion20220.65750.616456.58100.79520.92060.63385.28940.8176
CBFM20230.72010.740313.38630.87200.93340.69745.44620.8184
FUFusion20240.72020.665247.27510.81460.92590.69675.68560.8197
EgeFusion20240.31200.3356468.38960.43180.88920.30802.62940.8084
Proposed0.75240.77456.55080.88620.93800.73826.54660.8245
Table 2. Quantitative comparative analysis of different methods for Data 2 in the Lytro dataset.
Table 2. Quantitative comparative analysis of different methods for Data 2 in the Lytro dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.68230.613585.42170.75590.86450.66604.21160.8156
PMGI20200.47980.597753.32980.58160.85730.45926.30710.8251
MFFGAN20210.66090.629128.23930.79310.86510.64406.44910.8260
LEGFF20220.67700.646622.75960.79200.86800.66035.81730.8225
U2Fusion20220.59510.4969168.98200.68380.86190.57866.13250.8242
CBFM20230.71160.663431.86200.83480.86900.70195.97280.8234
FUFusion20240.72260.640016.71840.85110.86710.70887.02180.8294
EgeFusion20240.24920.3688490.22110.32100.84290.24193.29400.8127
Proposed0.74450.68705.31530.86720.87320.73697.67500.8339
Table 3. Quantitative average comparative analysis of different methods on the Lytro dataset.
Table 3. Quantitative average comparative analysis of different methods on the Lytro dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.70340.6115123.56910.78740.88870.69873.85210.8139
PMGI20200.39010.565698.34940.47360.88150.38575.86410.8225
MFFGAN20210.66420.645742.56550.84090.89150.65926.06040.8237
LEGFF20220.68100.675153.00730.81950.89370.67545.61380.8214
U2Fusion20220.61430.568297.59100.78350.88440.60935.77650.8221
CBFM20230.71370.701841.59590.86260.89560.71015.79670.8223
FUFusion20240.69060.6346128.58820.79720.88770.68626.47740.8263
EgeFusion20240.35760.4034340.41880.50320.84720.35413.21910.8120
Proposed0.73900.738831.40850.87720.89890.73626.96830.8296
Table 4. Quantitative comparative analysis of different methods for Data 1 in the MFI-WHU dataset.
Table 4. Quantitative comparative analysis of different methods for Data 1 in the MFI-WHU dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.67390.657182.36360.79530.89060.68294.18020.8180
PMGI20200.37690.652825.65140.44250.88830.37727.08430.8320
MFFGAN20210.63710.633637.05910.79360.89500.63376.99280.8315
LEGFF20220.62530.634837.47320.74510.89590.63456.22960.8271
U2Fusion20220.55710.4815140.72250.71730.88670.55746.28290.8273
CBFM20230.68460.701428.28990.82180.89630.68356.56580.8290
FUFusion20240.70240.754624.47000.84520.89700.70398.05550.8390
EgeFusion20240.26910.3308453.22950.39560.85570.27203.12270.8148
Proposed0.72810.82566.40760.85340.90000.72889.09640.8476
Table 5. Quantitative comparative analysis of different methods for Data 2 in the MFI-WHU dataset.
Table 5. Quantitative comparative analysis of different methods for Data 2 in the MFI-WHU dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.68760.5795137.27970.78620.86550.68423.92840.8140
PMGI20200.59060.464097.41590.65490.85590.58805.19750.8190
MFFGAN20210.65360.594564.38590.76610.86950.64975.37300.8198
LEGFF20220.65460.570958.22800.76600.87440.64624.80700.8173
U2Fusion20220.60530.528086.27560.72210.85780.60355.12870.8187
CBFM20230.70310.645859.43910.82120.87570.70155.33510.8196
FUFusion20240.70700.706330.60770.84110.87840.70615.90210.8224
EgeFusion20240.30310.2703653.97510.32960.83150.29532.87120.8110
Proposed0.74080.825038.06830.85500.87950.74097.81040.8348
Table 6. Quantitative average comparative analysis of different methods on the MFI-WHU dataset.
Table 6. Quantitative average comparative analysis of different methods on the MFI-WHU dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.67520.6301105.04180.77540.86480.66863.69400.8136
PMGI20200.42370.593362.81110.50610.85580.41775.48840.8210
MFFGAN20210.64270.632945.69600.78260.86840.63675.68320.8222
LEGFF20220.61900.606071.14620.70670.86920.61064.82910.8183
U2Fusion20220.55020.5156119.86390.69700.85650.54475.14980.8194
CBFM20230.69740.706459.45210.81740.87070.69225.67570.8224
FUFusion20240.70380.744323.71400.83370.87530.69926.69010.8284
EgeFusion20240.28740.3277537.72160.37570.82550.28412.80550.8111
Proposed0.72960.807236.49540.84530.87720.72547.81070.8371
Table 7. Quantitative comparative analysis of different methods for Data 1 in the MFFW dataset.
Table 7. Quantitative comparative analysis of different methods for Data 1 in the MFFW dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.72970.4875110.53920.69700.88820.68294.79580.8195
PMGI20200.40840.470961.04510.38630.88420.37666.58960.8282
MFFGAN20210.66260.567757.09830.77770.89230.61656.39930.8272
LEGFF20220.70010.594760.19300.78860.89680.64576.14990.8258
U2Fusion20220.59140.5325101.92270.68790.87900.55396.22970.8262
CBFM20230.70370.614651.85370.82370.89330.65886.05710.8254
FUFusion20240.69920.552276.27350.76520.88970.64796.63050.8285
EgeFusion20240.38770.4132365.61160.41420.84060.36114.21450.8173
Proposed0.70130.615533.39080.82900.89870.65716.68010.8290
Table 8. Quantitative average comparative analysis of different methods on the MFFW dataset.
Table 8. Quantitative average comparative analysis of different methods on the MFFW dataset.
Year Q A B / F Q C B Q C V Q E Q F M I Q G Q M I Q N C I E
GD20160.62790.5557217.99650.70110.87300.61143.61070.8122
PMGI20200.38070.5057275.92720.42450.86750.36755.04720.8178
MFFGAN20210.59050.5851138.11530.75570.87420.57445.04980.8179
LEGFF20220.62940.6032172.41730.73860.87750.61654.80880.8169
U2Fusion20220.55370.5499228.00640.70760.86900.53884.88940.8171
CBFM20230.63520.6408126.10580.79120.87480.62034.89050.8171
FUFusion20240.59870.5779226.83540.69740.87180.58175.34560.8196
EgeFusion20240.35170.4213443.44560.45810.83800.34393.37850.8115
Proposed0.63770.6362116.07560.80480.87990.62245.05050.8180
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, L.; Zhao, X.; Hou, H.; Zhang, X.; Lv, M.; Jia, Z.; Ma, H. Fractal Dimension-Based Multi-Focus Image Fusion via Coupled Neural P Systems in NSCT Domain. Fractal Fract. 2024, 8, 554. https://doi.org/10.3390/fractalfract8100554

AMA Style

Li L, Zhao X, Hou H, Zhang X, Lv M, Jia Z, Ma H. Fractal Dimension-Based Multi-Focus Image Fusion via Coupled Neural P Systems in NSCT Domain. Fractal and Fractional. 2024; 8(10):554. https://doi.org/10.3390/fractalfract8100554

Chicago/Turabian Style

Li, Liangliang, Xiaobin Zhao, Huayi Hou, Xueyu Zhang, Ming Lv, Zhenhong Jia, and Hongbing Ma. 2024. "Fractal Dimension-Based Multi-Focus Image Fusion via Coupled Neural P Systems in NSCT Domain" Fractal and Fractional 8, no. 10: 554. https://doi.org/10.3390/fractalfract8100554

Article Metrics

Back to TopTop