Next Article in Journal
AI-Enabled Vibrotactile Feedback-Based Condition Monitoring Framework for Outdoor Mobile Robots
Next Article in Special Issue
Plant Image Classification with Nonlinear Motion Deblurring Based on Deep Learning
Previous Article in Journal
Automatic Piston-Type Flow Standard Device Calibration System
Previous Article in Special Issue
ATC-YOLOv5: Fruit Appearance Quality Classification Algorithm Based on the Improved YOLOv5 Model for Passion Fruits
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Focus Image Fusion via PAPCNN and Fractal Dimension in NSST Domain

1
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
2
Key Laboratory of Signal Detection and Processing, Xinjiang University, Urumqi 830046, China
3
School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
4
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(18), 3803; https://doi.org/10.3390/math11183803
Submission received: 30 May 2023 / Revised: 27 August 2023 / Accepted: 29 August 2023 / Published: 5 September 2023

Abstract

:
Multi-focus image fusion is a popular technique for generating a full-focus image, where all objects in the scene are clear. In order to achieve a clearer and fully focused fusion effect, in this paper, the multi-focus image fusion method based on the parameter-adaptive pulse-coupled neural network and fractal dimension in the nonsubsampled shearlet transform domain was developed. The parameter-adaptive pulse coupled neural network-based fusion rule was used to merge the low-frequency sub-bands, and the fractal dimension-based fusion rule via the multi-scale morphological gradient was used to merge the high-frequency sub-bands. The inverse nonsubsampled shearlet transform was used to reconstruct the fused coefficients, and the final fused multi-focus image was generated. We conducted comprehensive evaluations of our algorithm using the public Lytro dataset. The proposed method was compared with state-of-the-art fusion algorithms, including traditional and deep-learning-based approaches. The quantitative and qualitative evaluations demonstrated that our method outperformed other fusion algorithms, as evidenced by the metrics data such as Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , Q N M I , Q Y , Q A G , Q P S N R , and Q M S E . These results highlight the clear advantages of our proposed technique in multi-focus image fusion, providing a significant contribution to the field.

1. Introduction

In real life, when people use cameras to shoot, they hope to obtain clear images of all the scenery in the same scene. However, the camera lens is limited by the depth of field and cannot focus on all targets simultaneously, so some areas in the captured photos are clear and some areas are blurry [1]. Multi-focus image fusion technology can fuse multiple images with different focus areas in the same scene into a fully clear image, effectively solving this problem and improving the information utilization rate of the image [2]. Accurately identifying and extracting the focus area in the source image is a challenge in multi-focus image fusion algorithms. Figure 1 shows the example of multi-focus image fusion; from the results, we can see that if the focus area is not fully extracted, it can lead to artifacts and loss of edge contours in the fusion results.
Image fusion technology has undergone rapid development in recent years, which can be divided into traditional-based and deep-learning-based algorithms. The traditional-based fusion methods include DWT [4], DTCWT [5], curvelet [6], contourlet [7], nonsubsampled contourlet transform [8], shearlet [9], and nonsubsampled shearlet transform (NSST) [10]. These wavelet-based algorithms and improved techniques for image fusion can typically be divided into three key steps, which are outlined below: image decomposition, coefficient fusion, and image reconstruction. Edge-preserving filtering algorithms have become an integral part of image fusion due to their high efficiency [11]. Their fusion structure shares similarities with wavelet-based methods, involving image decomposition into basic- and detail-layers, coefficient fusion, and eventual reconstruction. Additionally, methods based on sparse representation have also made significant strides in the field of image fusion.
Deep learning has proven to be highly successful in various image and visual tasks, mainly due to its remarkable feature learning capabilities [12]. In the field of multi-focus image fusion, deep-learning-based methods have rapidly become a prominent and popular research area since their initial proposal by Liu et al. in 2017 [13]. These methods leverage the power of convolutional neural networks (CNNs) [14] and generative adversarial networks (GANs) [15] to achieve impressive results in image fusion. The utilization of deep learning models such as CNNs and GANs in multi-focus image fusion has opened up new possibilities and achieved state-of-the-art results in the field. The ability of these models to automatically learn and extract relevant features from input images has led to significant advancements, providing researchers and practitioners with powerful tools to tackle complex image fusion challenges. As deep learning continues to evolve, it is expected to further drive the progress and innovation in multi-focus image fusion and other image-related tasks.
NSST, being a multi-scale transformation model, offers essential features such as translation invariance, making it highly effective in extracting precise edge details from images. Furthermore, its ability to correct registration errors to a certain extent plays a significant role in minimizing fusion errors. One of the remarkable advantages of NSST is that it imposes no restrictions on the number of directions after shearing and the inverse transformation does not require a synthetic direction filter. This unique property enables better sparse representation of images, ultimately reducing computational complexity and improving overall efficiency.
To achieve information complementarity and obtain crystal-clear fully focused images, we propose a novel multi-focus image fusion method based on a combination of the parameter-adaptive pulse coupled neural network and fractal dimension, all operating within the nonsubsampled shearlet transform domain. This method integrates the strengths of each component to deliver exceptional fusion results. The key contributions of our proposed method are as follows:
  • The nonsubsampled shearlet transform is applied to decompose and reconstruct the image.
  • The parameter-adaptive pulse coupled neural network is introduced to fuse the low-frequency bands.
  • The fractal dimension-based fusion rule via multi-scale morphological gradient is used to process the high-frequency bands.
  • The public Lytro dataset [16] is used to test the proposed method and the experiment also verified the effectiveness of our algorithm in multi-focus image fusion.
The rest of this paper is organized as follows. In Section 2, the related works are introduced. In Section 3, the background is described. In Section 4, the proposed multi-focus image fusion is constructed. In Section 5, the experimental results and discussions are described, and the advantages of the proposed method are verified. The conclusions with a brief summary are found in Section 6.

2. Related Works

2.1. Traditional-Based Fusion Methods

Zhang et al. [17] introduced an image fusion algorithm based on tensor decomposition integrating joint static and dynamic guidance in the nonsubsampled shearlet transform domain. Panigrahy et al. [18] introduced the image fusion using adaptive unit-linking pulse coupled neural network and distance-weighted regional energy in the nonsubsampled shearlet transform domain. Li et al. [19] introduced the sparse representation model into shearlet domain for multi-focus image fusion. Luo et al. [20] introduced the multi-state contextual hidden Markov model for image fusion via regional energy in the nonsubsampled shearlet transform domain, and this method has demonstrated superior performance compared to other state-of-the-art image fusion techniques, as evident from both subjective and objective assessments. Lu et al. [21] introduced the dual bilateral least squares hybrid filter for infrared and visible image fusion; the residual network ResNet50-based fusion rule and structure tensor-based fusion rule are used to process the base- and detail-layers, respectively. Zhao et al. [22] introduced the image fusion method using the fractional-order variational method and data-driven tight frame, and this method has been tested on medical images, infrared and visible images, and multi-focus images, wherein the effectiveness of the algorithm has been verified. Kong et al. [23] introduced the image fusion method based on side window filter and framelet transform. The methods mentioned above are all based on three steps: image decomposition, coefficient fusion, and image reconstruction. Indeed, the selection of rules for coefficients fusion is a crucial aspect of the image fusion process, as different fusion rules can have a significant impact on the final fusion results.

2.2. Deep-Learning-Based Fusion Methods

Bouzos et al. [24] introduced a convolutional-neural-network-based conditional random field model for structured multi-focus image fusion. Yang et al. [25] introduced a local binary pattern (LBP)-based proportional input generative adversarial network for image fusion. Jin et al. [26] introduced the unsupervised multi-focus image fusion algorithm based on Transformer and U-Net. Zhou et al. [27] introduced the unsupervised dense network with multi-scale convolutional block attention for multi-focus image fusion. Zhang et al. [28] proposed a fast unified image fusion network based on proportional maintenance of gradient and intensity (PMGI). Fang et al. [29] introduced the deep-learning-based threshold post-processing multi-focus image fusion method. Yang et al. [30] proposed an image fusion method based on latent low-rank representation and convolutional neural network. These deep learning-based methods have shown promising results in the field of image fusion, achieving relatively good performance compared to traditional techniques. However, it is important to acknowledge that these deep-learning-based methods often demand a substantial amount of computational time, mainly due to the requirement for large volumes of training data.

3. Background

3.1. Nonsubsampled Shearlet Transform

The nonsubsampled shearlet transform (NSST) is a multi-scale transform with translation invariance, which has significant effects in the field of image fusion [10,31]. The non-subsampled pyramid filter bank (NSP) and the shearlet filter (SF) are used for the multi-scale decomposition and multi-directional decomposition, respectively. Figure 2 shows an example of NSST decomposition performed on one source image with three levels [10,31]. NSST decomposes the image into high-frequency and low-frequency parts, and they are processed separately. Wavelet transform has limited ability in representing data features in the high-dimensional space of images; NSST solves this problem, and it can effectively extract edge details of images.

3.2. Parameter-Adaptive Pulse-Coupled Neural Network

The simplified pulse-coupled neural network (SPCNN) model was introduced by Chen et al. [32] in 2011 for image segmentation, and the SPCNN model is depicted as follows:
F i j n = S i j
L i j n = V L k l W i j k l Y k l n 1
U i j n = e α f U i j n 1 + F i j n 1 + β L i j n
Y i j n = 1     if   U i j n > E i j n 1 0     else
E i j n = e α e E i j n 1 + V E Y i j n
where F i j n and L i j n show the feeding input and linking input of the neuron at position i , j in iteration n , respectively. F i j n is fixed to the intensity of input image S i j during the whole iteration. L i j n is associated with the previous firing status of eight neighboring neurons through the synaptic weights.
W i j k l = 0.5 1 0.5 1 0 1 0.5 1 0.5
The structure of the SPCNN model is shown in Figure 3 [3]. The SPCNN model is initialized as Y i j 0 = 0 , U i j 0 = 0 and E i j 0 = 0 . In the improved model parameter-adaptive pulse-coupled neural network (PAPCNN), the parameters α f , λ , V E , and α e can be adaptively computed by
α f = log 1 / σ S
λ = S max / S 1 6
V E = e α f + 1 + 6 λ
α e = ln V E S 1 e 3 α f 1 e α f + 6 λ e α f
where σ S indicates the standard deviation of the input image S of range 0 ,   1 . S and S max show the normalized Otsu threshold and the maximum intensity of the input image, respectively. More details can be seen the references [3,32].

4. Proposed Fusion Method

In this section, the multi-focus image fusion method based on parameter-adaptive pulse-coupled neural network and fractal dimension in nonsubsampled shearlet transform domain is constructed. The proposed method consists of four steps: NSST decomposition, low-frequency sub-bands fusion, high-frequency sub-bands fusion, and inverse NSST. Figure 4 shows the structure of the proposed fusion method. More details can be concluded as follows:

4.1. NSST Decomposition

A L -level NSST decomposition is applied on source images A and B to generate the corresponding decomposed sub-bands L A , H A l , k and L B , H B l , k , respectively. Here, L A shows the low-frequency sub-band of A and H A l , k shows the high-frequency sub-band of A generated at the l -th decomposition level with the direction k . The meanings of L B and H B l , k are the same with respect to B.
A N S S T L A , H A l , k
B N S S T L B , H B l , k

4.2. Low-Frequency Sub-Bands Fusion

The low-frequency sub-bands contain the main energy information and background information. In this subsection, the PAPCNN model is used to process the low-frequency sub-bands, and the feeding input is F i j n = L S , S A , B . The activity level of the low-frequency coefficient is generated by the total firing times during the whole iteration. Based on the PAPCNN model, the firing times can be accumulated by adding the following step at the end of each iteration:
T i j n = T i j n 1 + Y i j n
Here, the firing times of each neuron is T i j N , where N depicts the total number of iterations. For the low-frequency sub-bands L A and L B of source images A and B, respectively, the PAPCNN firing times are computed and marked by T A , i j N and T B , i j N , respectively. The fused low-frequency sub-band L F i , j is computed by
L F i , j = L A i , j ,   if   T A , i j N T B , i j N L B i , j ,   else
which denotes that the coefficient with larger firing times is chosen as the fused coefficient.

4.3. High-Frequency Sub-Bands Fusion

The high-frequency sub-bands contain textures, details, and some noise. The fractal-dimension-based focus measure (FDFM) is constructed [33], and it is simple and easy to compute the clarity of a pixel; they are defined as follows:
F D F M H A l , k i , j = g max H A l , k i , j g min H A l , k i , j
F D F M H B l , k i , j = g max H B l , k i , j g min H B l , k i , j
where g max X i , j and g min X i , j are the maximum and minimum intensities, respectively, over a 3 × 3 window centered at the i , j t h pixel of X H A l , k , H B l , k .
The multi-scale morphological gradients (MSMG) of a pixel can be effectively used as a clarity measure of that pixel [34]. In the first step, the multi-scale structuring elements are constructed:
S E t = S E 1 S E 1 S E 1 t   times , t 1 ,   2 , ,   n
where n depicts the number of scales and S E 1 shows the basic structuring element at radius r . In this subsection, the radius r is 5.
In the next step, the morphological gradient operators are utilized to extract the gradient features G t from image X according to the following:
G t i , j = X i , j S E 1 X i , j Θ S E 1 , t 1 ,   2 , ,   n
where and Θ denote the morphological dilation and erosion operators, respectively.
In the third step, the gradients of all scales are merged into the multi-scale morphological gradients (MSMG):
M S M G i , j = t = 1 n w t G t i , j
w t = 1 2 × t + 1
In this subsection, the FDFM combined with MSMG is used to fuse the high-frequency sub-bands, and the corresponding equation is defined as follows:
H F l , k i , j = H A l . k i , j ,   if   F D F M H A l , k i , j × M S M G H A l . k i , j                                         F D F M H B l , k i , j × M S M G H B l , k i , j H B l , k i , j ,   else
where H F l , k i , j denotes the fused coefficient and M S M G X i , j | X H A l , k , H B l , k indicates the MSMG of X at the i , j th pixel.

4.4. Inverse NSST

The fused image F is generated by inverse NSST performed on the fused coefficients L F , H F l , k :
F I n v e r s e   N S S T L F , H F l , k
where F denotes the fused image.
The main steps of the proposed multi-focus image fusion approach are summarized in Algorithm 1.
Algorithm 1 Proposed multi-focus image fusion method
Input: the source images: A and B.
Parameters: the number of NSST decomposition levels: L, the number of directions at each decomposition level: K l , l 1 , L , the number of PAPCNN iterations: N, the radius r
 
Step 1: NSST decomposition
For each source image X A , B
Perform NSST on X to generate L X , H X l , k , l 1 , L , k 1 , K l ;
End
 
Step 2: Low-frequency sub-bands fusion
For each source image X A , B
Initialize the PAPCNN model: Y i j 0 = 0 ,   U i j 0 = 0 , E i j 0 = 0 , T i j 0 = 0 and F i j n = L X , n 1 , n ;
Estimate the PAPCNN parameters via Equations (7)–(10);
For each iteration n = 1 : N
Compute the PAPCNN model using Equations (2)–(5) and (13);
End
End
Merge L A and L B via Equation (14) to generate L F ;
 
Step 3: High-frequency sub-bands fusion
       For each level l = 1 : L
             For each direction k = 1 : K l
                    For each source image X A , B
       Estimate F D F M H X l , k i , j via Equations (15) and (16)
                    Compute M S M G H X l , k i , j via Equation (19)
       End
       Merge H A l , k and H B l , k via Equation (21) to generate H F l , k ;
             End
       End
 
Step 4: Inverse NSST
Perform inverse NSST on L F , H F l , k using Equation (22) to generate F ;
Output: the fused image  F .

5. Experimental Results and Discussion

5.1. Experimental Setting

In this subsection, we utilized the Lytro dataset [16] for evaluating the performance of the proposed method, along with several comparison algorithms. Figure 5 illustrates the examples of images from the Lytro dataset. These comparative algorithms include parameter-adaptive pulse coupled neural network and nonsubsampled shearlet transform (NSSTPA) [3], proportional maintenance of gradient and intensity (PMGI) [28], three-layer decomposition and sparse representation (TLDSR) [35], convolutional simultaneous sparse approximation (CSSA) [36], local extreme map guided multi-modal image fusion (LEGFF) [37], unified unsupervised image fusion network (U2Fusion) [38], distance-weighted regional energy and nonsubsampled shearlet transform (NSSTDW) [18], and zero-shot multi-focus image fusion (ZMFF) [2]. Qualitative and quantitative evaluations were used to evaluate the results of fusion; the quantitative evaluation indicators include the edge-based similarity measurement Q A B / F [39,40], the structural similarity based metric Q E [41], the feature mutual information metric Q F M I [42], the gradient based metric Q G [41], the nonlinear correlation information entropy Q N C I E [41], the phase-congruency-based metric Q P [41], the mutual information metric Q M I [39,40], the normalized mutual information Q N M I [41], the structural-similarity-based metric Q Y introduced by Yang et al. Q Y [41,43] and the average gradient metric Q A G [44,45], peak signal-to-noise ratio Q P S N R [46], and mean square error Q M S E [47]. The algorithms with larger indicator values (except for metric Q M S E ) are better. In our method, the “maxflat” filter for pyramidal decomposition was adopted in NSST transform; the parameter N was set to 110.

5.2. Analysis of NSST Decomposition Levels

In this subsection, the number of NSST decomposition levels L was set to 1–4, correspondingly. In this paper, the numbers of directions were empirically set to 16, 16, 8, and 8 for the first four scales from fine to coarse, and the detailed direction settings for different values of L can be shown in Table 1. The average metrics data of our method tested on the Lytro dataset with different NSST decomposition levels are shown in Table 2. From the results, we can denote that when the NSST decomposition levels L was set to 4 and the corresponding number of directions were 16, 16, 8, and 8 from fine to coarse, the proposed method can achieve the best values.

5.3. Qualitative and Quantitative Evaluation

In this subsection, five pairs of multi-focus images selected from the Lytro dataset were used as a qualitative evaluation display, and the fusion results are shown in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10, and the corresponding metric data can be found in Table 3, Table 4, Table 5, Table 6 and Table 7.
From Figure 6, we can observe that the NSSTPA method produced a relatively blurry image, failing to achieve a full focus effect; the PMGI method resulted in a dark image, leading to some loss of information in the fused image; the TLDSR method generated a fully focused fusion image, but the overall brightness of the image was slightly darker; both CSSA and LEGFF methods yielded similar fusion results, but the LEGFF method provided clearer details; the U2Fusion method enhanced sharpness across the entire image but caused darkening in some areas; the fused image generated by the NSSTDW method appeared blurry, with some details lost; the ZMFF method achieved a relatively good fusion effect, but when compared to our method, the fused image generated by our approach maintained a moderate brightness and full focus information. From Table 3, we can conclude that our method outperformed others, as the metrics data Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , Q N M I , and Q Y demonstrated superior results of 0.7524, 08849, 0.9379, 0.7386, 0.8239, 0.8000, 6.4258, 0.9227, and 0.9781, respectively.
From Figure 7, we can see that the NSSTPA algorithm generated blurry images, for example, the woman on the right side of the image was not in focus; the PMGI method generated a dark and fuzzy image; the TLDSR method generated a full-focused fusion image; the CSSA and LEGFF methods generated full-focused images, but a small portion of these images appeared darkened; the U2Fusion method produced a fused image with some areas darker, such as the glasses area; the NSSTDW method also resulted in a blurry image; the ZMFF method obtained a full-focused image; in contrast, the fused image computed by the proposed method exhibited full-focused and clear information. Based on the data in Table 4, we can conclude that the metrics Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , and Q N M I produced by the proposed method were the best among all the evaluated algorithms at 0.7223, 0.8621, 0.9146, 0.7236, 0.8382, 0.8404, 7.9702, and 1.0581, respectively. Comparatively, the metrics Q Y and Q A G obtained through the TLDSR and LEGFF algorithms yielded the best results, with values of 0.9639 and 9.4770, respectively. However, our corresponding metrics ranked as the second and third best performers, with values of 0.9623 and 8.0239, respectively. The metrics Q P S N R and Q M S E generated by the ZMFF method yielded the best results, with values of 37.7744 and 10.8553, respectively.
From Figure 8, it is evident that the NSSTPA and PMGI methods produced blurry images, particularly the PMGI algorithm, which exhibited the most severe information loss; the TLDSR algorithm introduced blurry artifacts into the fused image; the CSSA, LEGFF, and ZMFF methods had full-focused information in the fusion images, but the roof part of the building in these images appeared somewhat dark; the U2Fusion approach darkened certain areas in the fused image; the NSSTDW method resulted in a blurry image, particularly noticeable in the tree information in the lower left corner of the image. In contrast, our method generated a fused image with the best fusion effect, displaying moderate brightness and clear information. Referring to Table 5, it is evident that our method achieved the best performance based on the metrics Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , and Q N M I , achieving scores of 0.7199, 0.8405, 0.8974, 0.7206, 0.8321, 0.8283, 7.3742, and 0.9555, respectively.
From Figure 9, it is evident that the NSSTPA, TLDSR, and NSSTDW methods generated blurry images with unfocused background areas; for example, the building information on the left side of the images appeared blurry; the PMGI method caused distortion in the fused image and severe information loss; the CSSA method resulted in ambiguous building information on the right side of the fused image; the LEGFF and ZMFF methods generated fused images with all-focused information, and the LEGFF method produces a brighter fused image; the U2Fusion method led to some areas becoming darker and others brighter, resulting in potential information loss. However, the fused image achieved by our method exhibited the best performance, with fully focused information. Referring to Table 6, we can denote that the metrics data Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , Q N M I , and Q Y generated by our method were the best, with scores of 0.6995, 0.8413, 0.8852, 0.7032, 0.8295, 0.8760, 7.1167, 0.9244, and 0.9630, respectively.
From Figure 10, we can denote that artifacts were generated around the bottle in the fused images computed by the NSSTPA and TLDSR methods; the PMGI method caused distortion in the fused image; the fusion results achieved by the CSSA, LEGFF, and ZMFF methods looked similar, with the LEGFF method showing clearer details; the U2Fusion method created varying degrees of darkness on the bottle, cabin, and distant mountain scenery in the fused image; the fused image calculated by the NSSTDW method was blurry. However, the fused image generated by our method exhibited all-focused information without distortion and artifacts. Referring to Table 7, it is evident that the metrics Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , Q N M I , and Q Y obtained by our method were the best, with scores of 0.7869, 0.9161, 0.8914, 0.7829, 0.8204, 0.7979, 5.9240, 0.7794, and 0.9719, respectively.
Figure 11 presents a line chart illustrating the fluctuation trends of various algorithms on 20 data groups, along with the calculation of their respective average metrics. From Table 8, it becomes evident that our method outperformed the others, as it achieved the highest average metrics for Q A B / F , Q E , Q F M I , Q G , Q N C I E , Q P , Q M I , and Q N M I , with scores of 0.7384, 0.8756, 0.8989, 0.7359, 0.8289, 0.8114, 6.8598, and 0.9142, respectively. The average metrics Q Y and Q A G computed by the TLDSR and LEGFF methods were the best, with scores of 0.9563 and 14.8183, respectively, and the corresponding metrics computed by the proposed method ranked second and third, with scores of 0.9544 and 11.7205, respectively. The average metrics Q P S N R and Q M S E generated by the ZMFF method were the best, with scores of 36.4143 and 16.3895, respectively, and the corresponding metrics computed by the proposed method ranked in third, with scores of 34.2875 and 27.0097, respectively. This substantiates the superiority and universality of our fusion algorithm.
A thorough analysis led to the conclusion that our algorithm demonstrates superior fusion performance and effectively aligns subjective and objective analyses. The results indicate that our approach strikes a harmonious balance between the two, further affirming its efficacy and applicability.
For a more comprehensive experimental comparison, we added three fusion algorithms, namely, NSCTMSF [48], NSCTRPCNN [49], and GD [50], published in 2012, 2013, and 2016, respectively. The average metrics data comparison is provided in Table 9. From the metrics data, it can be seen that our algorithm had more advantages than previous fusion algorithms.

5.4. Sequence Multi-Focus Image Fusion

In this subsection, we proceed with the testing of our algorithm on four sets of sequence multi-focus images obtained from the Lytro dataset [16]. The results in Figure 12 clearly demonstrate the effectiveness of our approach in producing fully focused images. As a result, we can confidently assert that our algorithm is equally adept at handling sequence multi-focus image fusion.

5.5. Further Discussion

We conducted a comprehensive analysis of the impact of NSST decomposition levels on image fusion performance and found that increasing the number of decomposition levels appropriately significantly improved the image fusion results. Our proposed multi-focus image fusion model based on NSST was thoroughly validated through simulation experiments. When compared to traditional and deep-learning-based algorithms, our method exhibited substantial advantages in terms of preserving edge details and effectively utilizing complementary image information, as indicated by both subjective and objective evaluation criteria. However, despite these remarkable achievements, there are still certain limitations that warrant further investigation. Notably, the average gradient metric of our algorithm did not reach its optimal value, although it remains competitive with other approaches. As a result, our next step involves delving deeper into enhancing the average gradient information of the fused images. Figure 13 showcases the fusion results of our algorithm applied to medical images, infrared and visible images. For anatomical and functional image fusion, we utilized the RGB to YUV color space conversion, where the Y channel of the functional image was fused with the anatomical image using our proposed algorithm. Subsequently, the YUV to RGB space conversion was applied to obtain the final fusion result. The results demonstrate the effectiveness of our algorithm in achieving information complementarity.

6. Conclusions

To achieve clear and fully focused images while minimizing information loss, we proposed a novel parameter-adaptive pulse coupled neural network-based multi-focus image fusion method, operating in the nonsubsampled shearlet transform domain, with the incorporation of fractal dimension. This sophisticated approach allows for better preservation of background, brightness, and detail information in the fused images.
The fusion process involves two key components. First, we employed the parameter-adaptive pulse coupled neural network-based fusion rule to fuse the low-frequency sub-bands, which optimally combined information from different focus levels. Second, for the high-frequency sub-bands, we utilized the fractal dimension integrated multi-scale morphological gradient fusion rule, enabling us to effectively merge the fine details from the input images.
To validate the effectiveness of our proposed method, we conducted extensive testing and validation using the publicly available Lytro dataset, which comprises 20 pairs of images and four sets of sequence images. The results were compared with state-of-the-art fusion algorithms published both domestically and internationally from 2012 to 2023. Through this thorough evaluation, we demonstrated that our algorithm stands at the forefront of technology in the field of multi-focus image fusion.
Moreover, to enhance the practical usability and efficiency of the proposed method, the implementation of a Python or C++ version may significantly accelerate the fusion process.

Author Contributions

The experimental measurements and data collection were carried out by M.L. and Z.J. The manuscript was written by M.L. with the assistance of Z.J., L.L. and H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Science Foundation of China under grant no. 62261053.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cheng, C.; Xu, T.; Wu, X. MUFusion: A general unsupervised image fusion network based on memory unit. Inf. Fusion 2023, 92, 80–92. [Google Scholar] [CrossRef]
  2. Hu, X.; Jiang, J.; Liu, X.; Ma, J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion 2023, 92, 127–138. [Google Scholar] [CrossRef]
  3. Yin, M.; Liu, X.; Liu, Y. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans. Instrum. Meas. 2019, 68, 49–64. [Google Scholar] [CrossRef]
  4. Odusami, M.; Maskeliūnas, R.; Damaševičius, R. Pixel-level fusion approach with vision transformer for early detection of Alzheimer’s disease. Electronics 2023, 12, 1218. [Google Scholar] [CrossRef]
  5. Babu, B.; Narayana, M. Two stage multi-modal medical image fusion with marine predator algorithm-based cascaded optimal DTCWT and NSST with deep learning. Biomed. Signal Process. Control. 2023, 85, 104921. [Google Scholar] [CrossRef]
  6. Candes, E.; Demanet, L. Fast discrete curvelet transforms. Multiscale Model. Simul. 2006, 5, 861–899. [Google Scholar] [CrossRef]
  7. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef]
  8. Da, A.; Zhou, J.; Do, M. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar]
  9. Guo, K.; Labate, D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007, 39, 298–318. [Google Scholar] [CrossRef]
  10. Easley, G.; Labate, D.; Lim, W.Q. Sparse directional image representations using the discrete shearlet transform. Appl. Comput. Harmon. Anal. 2008, 25, 25–46. [Google Scholar] [CrossRef]
  11. Li, L.; Lv, M.; Jia, Z.; Jin, Q.; Liu, M.; Chen, L.; Ma, H. An effective infrared and visible image fusion approach via rolling guidance filtering and gradient saliency map. Remote Sens. 2023, 15, 2486. [Google Scholar] [CrossRef]
  12. Wang, J.; Xi, X.; Li, D.; Li, F.; Zhang, G. GRPAFusion: A gradient residual and pyramid attention-based multiscale network for multimodal image fusion. Entropy 2023, 25, 169. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, Y.; Chen, X.; Peng, H. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  14. Jiang, X.; Zhao, H.; Liu, J. Classification of mineral foam flotation conditions based on multi-modality image fusion. Appl. Sci. 2023, 13, 3512. [Google Scholar] [CrossRef]
  15. Xu, X.; Shen, Y.; Han, S. Dense-FG: A fusion GAN model by using densely connected blocks to fuse infrared and visible images. Appl. Sci. 2023, 13, 4684. [Google Scholar] [CrossRef]
  16. Nejati, M.; Samavi, S.; Shirani, S. Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 2015, 25, 72–84. [Google Scholar] [CrossRef]
  17. Zhang, R.; Wang, Z.; Sun, H.; Deng, L.; Zhu, H. TDFusion: When tensor decomposition meets medical image fusion in the nonsubsampled shearlet transform domain. Sensors 2023, 23, 6616. [Google Scholar] [CrossRef]
  18. Panigrahy, C.; Seal, A. Parameter adaptive unit-linking pulse coupled neural network based MRI–PET/SPECT image fusion. Biomed. Signal Process. Control 2023, 83, 104659. [Google Scholar] [CrossRef]
  19. Li, L.; Lv, M.; Jia, Z.; Ma, H. Sparse representation-based multi-focus image fusion method via local energy in shearlet domain. Sensors 2023, 23, 2888. [Google Scholar] [CrossRef]
  20. Luo, X.; Jiang, Y.; Wang, A. Infrared and visible image fusion based on multi-state contextual hidden Markov model. Pattern Recognit. 2023, 138, 109431. [Google Scholar] [CrossRef]
  21. Lu, Q.; Han, Z.; Hu, L.; Tian, F. An infrared and visible image fusion algorithm method based on a dual bilateral least squares hybrid filter. Electronics 2023, 12, 2292. [Google Scholar] [CrossRef]
  22. Zhao, R.; Liu, J. Fractional-order variational image fusion and denoising based on data-driven tight frame. Mathematics 2023, 11, 2260. [Google Scholar] [CrossRef]
  23. Kong, W.; Li, Y.; Lei, Y. Medical image fusion using SKWGF and SWF in framelet transform domain. Electronics 2023, 12, 2659. [Google Scholar] [CrossRef]
  24. Bouzos, O.; Andreadis, I.; Mitianoudis, N. A convolutional neural network-based conditional random field model for structured multi-focus image fusion robust to noise. IEEE Trans. Image Process. 2023. early access. [Google Scholar] [CrossRef]
  25. Yang, D.; Zheng, Y.; Xu, W.; Sun, P.; Zhu, D. LPGAN: A LBP-based proportional input generative adversarial network for image fusion. Remote Sens. 2023, 15, 2440. [Google Scholar] [CrossRef]
  26. Jin, X.; Xi, X. An unsupervised multi-focus image fusion method based on Transformer and U-Net. IET Image Process. 2023, 17, 733–746. [Google Scholar] [CrossRef]
  27. Zhou, D.; Jin, X. MCRD-Net: An unsupervised dense network with multi-scale convolutional block attention for multi-focus image fusion. IET Image Process. 2022, 16, 1558–1574. [Google Scholar] [CrossRef]
  28. Zhang, H.; Xu, H.; Xiao, Y. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar]
  29. Fang, L.; Zhao, J.; Pan, Z.; Li, Y. TPP: Deep learning based threshold post-processing multi-focus image fusion method. Comput. Electr. Eng. 2023, 110, 108736. [Google Scholar] [CrossRef]
  30. Yang, Y.; Gao, C. LatLRR-CNN: An infrared and visible image fusion method combining latent low-rank representation and CNN. Multimed. Tools Appl. 2023. early access. [Google Scholar] [CrossRef]
  31. Li, L.; Ma, H. Saliency-guided nonsubsampled shearlet transform for multisource remote sensing image fusion. Sensors 2021, 21, 1756. [Google Scholar] [CrossRef]
  32. Chen, Y.; Park, S.K.; Ma, Y. A new automatic parameter setting method of a simplified PCNN for image segmentation. IEEE Trans. Neural Netw. 2011, 22, 880–892. [Google Scholar] [CrossRef]
  33. Panigrahy, C.; Seal, A.; Mahato, N.K. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106141. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Bai, X.; Wang, T. Boundary finding based multi-focus image fusion through multi-scale morphological focus-measure. Inf. Fusion 2017, 35, 81–101. [Google Scholar] [CrossRef]
  35. Li, X.; Zhou, F.; Tan, H. Joint image fusion and deniosing via three-layer decomposition and sparse representation. Knowl. Based Syst. 2021, 224, 107087. [Google Scholar] [CrossRef]
  36. Veshki, F.; Vorobyov, S. Convolutional simultaneous sparse approximation with applications to RGB-NIR image fusion. In Proceedings of the 56th Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 31 October–2 November 2022; pp. 872–876. [Google Scholar]
  37. Zhang, Y.; Xiang, W. Local extreme map guided multi-modal brain image fusion. Front. Neurosci. 2022, 16, 1055451. [Google Scholar] [CrossRef] [PubMed]
  38. Xu, H.; Ma, J.; Jiang, J. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  39. Qu, X.; Yan, J.; Xiao, H. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  40. Li, L.; Ma, H. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 12389–12409. [Google Scholar] [CrossRef]
  41. Liu, Z.; Blasch, E.; Xue, Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 94–109. [Google Scholar] [CrossRef]
  42. Haghighat, M.; Razian, M. Fast-FMI: Non-reference image fusion metric. In Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan, 15–17 October 2014; pp. 424–426. [Google Scholar]
  43. Lv, M.; Li, L.; Jin, Q.; Jia, Z.; Chen, L.; Ma, H. Multi-focus image fusion via distance-weighted regional energy and structure tensor in NSCT domain. Sensors 2023, 23, 6135. [Google Scholar] [CrossRef]
  44. Li, L.; Si, Y.; Wang, L. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  45. Shreyamsha Kumar, B.K. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015, 9, 1193–1204. [Google Scholar] [CrossRef]
  46. Zhao, Y.; Zheng, Q. TUFusion: A transformer-based universal fusion algorithm for multimodal images. IEEE Trans. Circuits Syst. Video Technol. 2023. early access. [Google Scholar] [CrossRef]
  47. Lu, Z.; Zhu, T.; Zhou, H.; Zhang, L.; Jia, C. An image enhancement method for side-scan sonar images based on multi-stage repairing image fusion. Electronics 2023, 12, 3553. [Google Scholar] [CrossRef]
  48. Das, S.; Kundu, M.K. NSCT-based multimodal medical image fusion using pulse-coupled neural network and modified spatial frequency. Med. Biol. Eng. Comput. 2012, 50, 1105–1114. [Google Scholar] [CrossRef] [PubMed]
  49. Das, S.; Kundu, M.K. A neuro-fuzzy approach for medical image fusion. IEEE Trans. Biomed. Eng. 2013, 60, 3347–3353. [Google Scholar] [CrossRef]
  50. Paul, S.; Sevcenco, I.; Agathoklis, P. Multi-exposure and multi-focus image fusion in gradient domain. J. Circuits Syst. Comput. 2016, 25, 1650123. [Google Scholar] [CrossRef]
Figure 1. The example of multi-focus image fusion. (a) Right focus; (b) left focus; (c) NSSTPA [3]; (d) proposed.
Figure 1. The example of multi-focus image fusion. (a) Right focus; (b) left focus; (c) NSSTPA [3]; (d) proposed.
Mathematics 11 03803 g001
Figure 2. The example of NSST decomposition [31].
Figure 2. The example of NSST decomposition [31].
Mathematics 11 03803 g002
Figure 3. The structure of the SPCNN model [3].
Figure 3. The structure of the SPCNN model [3].
Mathematics 11 03803 g003
Figure 4. The structure of the proposed method.
Figure 4. The structure of the proposed method.
Mathematics 11 03803 g004
Figure 5. Lytro dataset.
Figure 5. Lytro dataset.
Mathematics 11 03803 g005aMathematics 11 03803 g005b
Figure 6. Results of Lytro-01. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Figure 6. Results of Lytro-01. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Mathematics 11 03803 g006
Figure 7. Results of Lytro-04. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Figure 7. Results of Lytro-04. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Mathematics 11 03803 g007
Figure 8. Results of Lytro-06. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Figure 8. Results of Lytro-06. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Mathematics 11 03803 g008
Figure 9. Results of Lytro-07. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Figure 9. Results of Lytro-07. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Mathematics 11 03803 g009aMathematics 11 03803 g009b
Figure 10. Results of Lytro-09. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Figure 10. Results of Lytro-09. (a) NSSTPA; (b) PMGI; (c) TLDSR; (d) CSSA; (e) LEGFF; (f) U2Fusion; (g) NSSTDW; (h) ZMFF; (i) proposed.
Mathematics 11 03803 g010
Figure 11. The line chart of metrics.
Figure 11. The line chart of metrics.
Mathematics 11 03803 g011aMathematics 11 03803 g011bMathematics 11 03803 g011c
Figure 12. The fusion results of sequence multi-focus image fusion.
Figure 12. The fusion results of sequence multi-focus image fusion.
Mathematics 11 03803 g012aMathematics 11 03803 g012b
Figure 13. Other multi-modal image fusion results. (a) Source 1; (b) source 2; (c) proposed.
Figure 13. Other multi-modal image fusion results. (a) Source 1; (b) source 2; (c) proposed.
Mathematics 11 03803 g013aMathematics 11 03803 g013b
Table 1. The number of NSST decomposition levels and the corresponding direction settings for parameter analysis.
Table 1. The number of NSST decomposition levels and the corresponding direction settings for parameter analysis.
Number of Decomposition LevelsNumber of Directions at Each Level
116
216, 16
316, 16, 8
416, 16, 8, 8
Table 2. The average metrics data of our method tested on the Lytro dataset with different NSST decomposition levels.
Table 2. The average metrics data of our method tested on the Lytro dataset with different NSST decomposition levels.
Levels Q A B / F Q E Q F M I Q G Q N C I E Q P Q M I Q N M I Q Y Q A G Q P S N R Q M S E
10.49860.48430.86550.49410.82790.48456.61890.88590.84638.928533.359233.8391
20.64150.75080.88630.63560.82570.64806.32870.84510.894311.021533.336533.9162
30.71850.86000.89720.71490.82720.77496.57380.87660.938111.632233.819830.3342
40.73840.87560.89890.73590.82890.81146.85980.91420.954411.720534.287527.0097
Table 3. The quantitative evaluation of the different methods shown in Figure 6.
Table 3. The quantitative evaluation of the different methods shown in Figure 6.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.65260.54660.75240.72050.69230.65750.64480.73420.7524
Q E 0.85430.63160.86040.88050.82050.79520.84880.8779 0.8849
Q F M I 0.93620.91690.93330.93410.93060.92060.93670.93120.9379
Q G 0.61670.51560.73840.69770.66580.63380.60950.71340.7386
Q N C I E 0.8180 0.81690.82320.82150.81580.81760.81800.82220.8239
Q P 0.48420.39250.79150.71870.69370.66400.47830.7673 0.8000
Q M I 5.36465.13476.25376.03854.89195.28945.34396.15056.4258
Q N M I 0.77170.74100.89950.86580.69710.74720.76960.88270.9227
Q Y 0.86800.76560.97390.94630.91640.88320.84730.96440.9781
Q A G 7.40285.94367.94078.0460 10.29199.45647.14698.01498.2592
Q P S N R 35.377644.784434.642537.100933.288031.866834.713337.403036.3369
Q M S E 18.85042.160922.327012.676330.498742.306021.965711.824515.1144
Table 4. The quantitative evaluation of the different methods shown in Figure 7.
Table 4. The quantitative evaluation of the different methods shown in Figure 7.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.59550.34910.71690.67290.66390.59880.60780.67800.7223
Q E 0.70670.39590.83870.8592 0.83270.78530.75990.85390.8621
Q F M I 0.89880.89470.91230.90890.91060.90070.90360.90650.9146
Q G 0.59440.34910.71790.67340.66490.59850.60550.67740.7236
Q N C I E 0.83080.82850.83560.83400.82790.82820.83050.83400.8382
Q P 0.61790.36400.82700.72610.72400.65730.60680.77620.8404
Q M I 7.03546.71407.72767.55626.59966.65137.00307.53597.9702
Q N M I 0.93550.91851.02651.00260.87540.88530.93061.00151.0581
Q Y 0.88370.67840.96390.91140.8700 0.77630.87950.9196 0.9623
Q A G 7.05443.84317.83697.62309.47708.36237.13437.59598.0239
Q P S N R 35.147433.662334.680537.418734.1485 29.444234.681837.774434.8978
Q M S E 19.876427.979922.132711.781925.016573.902322.125710.855321.0523
Table 5. The quantitative evaluation of the different methods shown in Figure 8.
Table 5. The quantitative evaluation of the different methods shown in Figure 8.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.66760.28190.71100.65800.67180.59190.64360.67020.7199
Q E 0.80930.34290.80250.83390.81590.76460.81860.82900.8405
Q F M I 0.89210.88350.89430.89210.89240.88630.89170.88980.8974
Q G 0.67140.28290.71110.66130.67570.59570.64850.67360.7206
Q N C I E 0.82800.82510.83110.82940.82520.82400.82800.82980.8321
Q P 0.76410.40290.82040.74760.78040.69070.71520.77630.8283
Q M I 6.77206.29357.26427.00106.31046.07346.71527.06757.3742
Q N M I 0.87800.84890.94140.90890.81680.81230.87010.91890.9555
Q Y 0.91460.60420.95660.90370.87680.68720.90210.91460.9463
Q A G 9.90094.324010.02269.684212.149310.07909.69119.850610.3422
Q P S N R 33.596828.063033.363435.859632.629630.429233.049636.340333.8703
Q M S E 28.4055101.574129.973616.870335.491458.905532.220015.102526.6719
Table 6. The quantitative evaluation of the different methods shown in Figure 9.
Table 6. The quantitative evaluation of the different methods shown in Figure 9.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.6164 0.38630.69440.64370.64150.56680.60360.65320.6995
Q E 0.8039 0.5027 0.82590.83220.80850.71620.80880.82970.8413
Q F M I 0.88390.86870.87950.87850.87900.8733 0.88480.87710.8852
Q G 0.6233 0.39590.6973 0.65000.64670.57410.61230.65880.7032
Q N C I E 0.8256 0.82420.82790.82600.82220.82300.82590.82680.8295
Q P 0.73140.49360.86450.80050.82080.71430.69860.82180.8760
Q M I 6.45116.23866.85086.55255.87416.02006.48536.69087.1167
Q N M I 0.83800.83060.89060.85040.76090.78180.8425 0.87010.9244
Q Y 0.90290.70670.96240.90790.88910.76380.8908 0.92110.9630
Q A G 9.26475.18239.89759.787511.871411.37239.07749.888710.2103
Q P S N R 33.971730.753733.513535.700532.860228.863233.708736.387635.0734
Q M S E 26.055954.665728.955417.499533.656184.480927.682614.939120.2181
Table 7. The quantitative evaluation of the different methods shown in Figure 10.
Table 7. The quantitative evaluation of the different methods shown in Figure 10.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.72480.24780.76190.73490.72100.66040.70420.74420.7869
Q E 0.90040.36560.84430.91430.83200.84840.89160.90880.9161
Q F M I 0.88920.87620.88340.88860.88260.87630.88940.88560.8914
Q G 0.71150.24100.75590.72570.71400.65360.69380.73550.7829
Q N C I E 0.81670.81550.82010.81980.81430.81660.81660.81910.8204
Q P 0.71270.43950.77130.74100.74940.66470.65700.76290.7979
Q M I 5.19544.94125.84855.76674.66685.17435.13165.67635.9240
Q N M I 0.68340.66720.76950.76050.61530.68560.67630.74870.7794
Q Y 0.91780.59500.96290.91790.89410.80390.89020.94330.9719
Q A G 14.69906.709714.766814.696320.524414.766214.214714.782615.5686
Q P S N R 31.384826.910031.063134.088630.528831.801931.052834.180032.4951
Q M S E 47.2717132.459950.906725.364457.570242.943151.027024.835736.6079
Table 8. The average metrics data of the different methods shown in Figure 11.
Table 8. The average metrics data of the different methods shown in Figure 11.
NSSTPAPMGITLDSRCSSALEGFFU2FusionNSSTDWZMFFProposed
Q A B / F 0.67200.39010.73200.68970.68100.61430.65540.70870.7384
Q E 0.82470.47360.84520.87060.81950.78350.83270.86870.8756
Q F M I 0.89310.88150.89470.89480.89370.88440.89430.89250.8989
Q G 0.66550.38570.72910.68530.67540.60930.64820.70300.7359
Q N C I E 0.82540.82250.82870.82640.82140.82210.82550.82710.8289
Q P 0.69320.46200.79950.73450.75650.66570.65190.78530.8114
Q M I 6.32125.86416.82266.53785.61385.77656.30236.62716.8598
Q N M I 0.84310.80040.91030.87160.74730.77250.84060.88380.9142
Q Y 0.89550.67380.95630.89110.88170.79120.87670.93130.9544
Q A G 10.84705.868411.382311.125214.818312.034310.566111.218311.7205
Q P S N R 34.032932.478233.578835.908632.616031.209833.570636.414334.2875
Q M S E 28.556775.395631.746318.017339.152359.442431.931516.389527.0097
Table 9. The average metrics data of other different methods.
Table 9. The average metrics data of other different methods.
NSCTMSFNSCTRPCNNGDProposed
Q A B / F 0.71310.71030.70340.7384
Q E 0.86150.86440.78740.8756
Q F M I 0.89710.89720.88870.8989
Q G 0.70870.70580.69870.7359
Q N C I E 0.82630.82800.81390.8289
Q P 0.76310.76160.74660.8114
Q M I 6.46606.70753.85216.8598
Q N M I 0.86160.89450.51130.9142
Q Y 0.92390.92080.86080.9544
Q A G 11.422511.347711.684411.7205
Q P S N R 33.513934.748626.574234.2875
Q M S E 32.625624.5159150.138227.0097
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, M.; Jia, Z.; Li, L.; Ma, H. Multi-Focus Image Fusion via PAPCNN and Fractal Dimension in NSST Domain. Mathematics 2023, 11, 3803. https://doi.org/10.3390/math11183803

AMA Style

Lv M, Jia Z, Li L, Ma H. Multi-Focus Image Fusion via PAPCNN and Fractal Dimension in NSST Domain. Mathematics. 2023; 11(18):3803. https://doi.org/10.3390/math11183803

Chicago/Turabian Style

Lv, Ming, Zhenhong Jia, Liangliang Li, and Hongbing Ma. 2023. "Multi-Focus Image Fusion via PAPCNN and Fractal Dimension in NSST Domain" Mathematics 11, no. 18: 3803. https://doi.org/10.3390/math11183803

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop