Next Article in Journal
Mapping Fire Severity in Southwest China Using the Combination of Sentinel 2 and GF Series Satellite Images
Next Article in Special Issue
Pansharpening Model of Transferable Remote Sensing Images Based on Feature Fusion and Attention Modules
Previous Article in Journal
Research on Two-Round Self-Balancing Robot SLAM Based on the Gmapping Algorithm
Previous Article in Special Issue
High-Spatial-Resolution NDVI Reconstruction with GA-ANN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PCNN Model Guided by Saliency Mechanism for Image Fusion in Transform Domain

1
College of Information Science and Technology, Gansu Agricultural University, Lanzhou 730070, China
2
School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(5), 2488; https://doi.org/10.3390/s23052488
Submission received: 4 February 2023 / Revised: 16 February 2023 / Accepted: 21 February 2023 / Published: 23 February 2023

Abstract

:
In heterogeneous image fusion problems, different imaging mechanisms have always existed between time-of-flight and visible light heterogeneous images which are collected by binocular acquisition systems in orchard environments. Determining how to enhance the fusion quality is key to the solution. A shortcoming of the pulse coupled neural network model is that parameters are limited by manual experience settings and cannot be terminated adaptively. The limitations are obvious during the ignition process, and include ignoring the impact of image changes and fluctuations on the results, pixel artifacts, area blurring, and the occurrence of unclear edges. Aiming at these problems, an image fusion method in a pulse coupled neural network transform domain guided by a saliency mechanism is proposed. A non-subsampled shearlet transform is used to decompose the accurately registered image; the time-of-flight low-frequency component, after multiple lighting segmentation using a pulse coupled neural network, is simplified to a first-order Markov situation. The significance function is defined as first-order Markov mutual information to measure the termination condition. A new momentum-driven multi-objective artificial bee colony algorithm is used to optimize the parameters of the link channel feedback term, link strength, and dynamic threshold attenuation factor. The low-frequency components of time-of-flight and color images, after multiple lighting segmentation using a pulse coupled neural network, are fused using the weighted average rule. The high-frequency components are fused using improved bilateral filters. The results show that the proposed algorithm has the best fusion effect on the time-of-flight confidence image and the corresponding visible light image collected in the natural scene, according to nine objective image evaluation indicators. It is suitable for the heterogeneous image fusion of complex orchard environments in natural landscapes.

1. Introduction

Automatic apple fruit picking in natural environments can reduce the intensity of heavy manual labor, which is an inevitable choice for modern agriculture [1]. The natural light in northwest China is strong, and the visible light images collected in the natural environment are vulnerable to the influences of changing light and complex backgrounds, so the recognition effect lacks some robustness [2]. In the complex environment of orchard operations, the most potential in vision research on picking robots lies in the technology of heterogeneous image fusion (IF) between time-of-flight (ToF) images and visible light images. The collected images have a variety of different attributes, including light invariance, spatial hierarchy, infrared perception, reliability of discrimination data, etc. [2]. The image is indirectly generated from the depth information, which can reflect the near–far relationship and infrared reflection characteristics of different objects in the scene, and the effect is not affected by light changes [2]. Image fusion generates a new information processing process that interprets the scene from a different source image which cannot be obtained from the information obtained by a single sensor [2,3]. Determining how to fuse ToF images and visible light images with different wavelength ranges and imaging mechanisms with high quality is currently a topic of great interest in image fusion research.
A non-subsampled shearlet transform (NSST) is a multi-scale, multi-directional, translation-invariant transform domain image decomposition method, which is widely used in image fusion [4]. An NSST shearlet wave transform avoids the down-sampling operation, and has the characteristics of translation invariance, simple operation, low time complexity, etc. [5]. Compared with wavelet transforms such as the discrete wavelet transform (DWT), stationary wavelet transform (SWT), discrete cosine transform (DCT), curvelet transform, and contourlet transform, an NSST has a good effect on searching for edges and contours. There are large numbers of deep neural network layers in deep learning methods. This characteristic could lead to low efficiency and a high cost. The advantage of an NSST is that it can fully fuse the source image information, and the fused image has good correlation coefficient and information entropy, which is more suitable for the situation where the image background in the natural orchard environment is complex, and the contour and image texture information need to be fused at the same time.
Related works are summarized as follows: A pulse coupled neural network (PCNN) is a neural network model established by simulating the activities of visual nerve cells in the cerebral cortex. Similar pattern features are classified into categories based on the principles of similarity clustering and capture characteristics [6]. In terms of image fusion in a transform domain, Cheng et al. used an adaptive dual-channel pulse coupled neural network with triple connection strength in the local non-down-sampled shear wave transform domain to solve the spectral difference between infrared and visible light [7]. Panigrahy et al. proposed a new medical fusion method in a non-down-sampled shear wave transform domain based on a weighted parameter adaptive dual channel PCNN [8]. In terms of image fusion in saliency attention models, Liu et al. proposed a saliency detection model that combines a global saliency map with a local saliency map [9]. Yang et al. designed a new fuzzy logic rule based on global saliency measurements to fuse the details extracted from panchromatic images with high spatial resolution and multispectral images with low spatial resolution [10]. Li et al. used the segmentation-driven low-rank matrix recovery model to detect the significance of each individual image in the image set to highlight the regions with sparse features in each image [11]. In terms of the optimization of image fusion parameters, Zhu et al. applied PCNN parameters to infrared and visible image fusion through quantum-behavior particle swarm optimization improvement [12]. Huang et al. used an NSCT to independently decompose the intensity hue saturation of the image, a PCNN to fuse high-frequency sub-band images and low-frequency images, and a hybrid leapfrog algorithm to optimize PCNN parameters [13]. Dharini et al. proposed a nature-inspired optimal feature selection method using ant colony optimization to reduce the complexity of the PCNN fusion of infrared and visible images [14]. In the research of overexposure problems with concern to the ongoing climate change-related environmental changes over mountainous areas, Muhuri et al. used polarization fraction variation with temporal RADARSAT-2 C-Band full-polarimetric to study SAR Data [15]. Raskar et al. introduced a novel technique to allow a user to interact with projected information and to update the projected information [16].
A PCNN classifies similar pattern features into categories based on the principles of similarity aggregation and capture characteristics. The segmentation combination has the advantages of the grayscale aggregation lighting mechanism and the same grayscale attribute priority lighting. This is consistent with the basic idea of cluster analysis. Qiu et al. proposed a new density peaks-based clustering method, called clustering with local density peaks-based minimum spanning tree [17]. Huang et al. proposed new adaptive spatial regularization for the representation coefficients to improve the robustness of the model to noise [18]. Huang et al. proposed ultra-scalable spectral clustering and ultra-scalable ensemble clustering methods [19].
Although scholars have studied the optimization and improvement of PCNN parameters, there are still cases of pixel artifacts, region blurring and unclear edges due to ignoring the impact of image changes and fluctuations on the results during the ignition process.
This paper introduces the concept of entropy [20] in information theory and proposes a PCNN model guided by a saliency mechanism (SMPCNN). The ToF low-frequency component after multiple lighting segmentation using PCNN is simplified into a first-order Markov situation, and the significance function is defined as first-order Markov mutual information. On this basis, a PCNN model guided by a saliency mechanism for image fusion in transform domain (NSST-SMPCNN) is proposed to fuse ToF and visible light heterogeneous images collected by a binocular acquisition system in an orchard environment.
We summarize our main contributions below.
First, we aim to solve the following existing problems:
  • The traditional method of space domain fusion is to create a fusion model in the image gray space, which has the disadvantage that it is not easy to find the source image texture and boundary features.
  • A PCNN model has the defects of parameter experience setting, unadaptive termination, and easy over-segmentation. In the ignition process, it ignores the impact of image change fluctuation on the results, resulting in pixel artifacts, area blurring, and unclear edges.
  • The differences in imaging mechanisms between ToF and visible light heterogeneous images collected by a binocular acquisition system in an orchard environment lead to the problem of low fusion quality.
Second, the innovations and novelties of this paper are as follows:
  • A PCNN model guided by a saliency mechanism is proposed and applied to the fusion of ToF and visible light heterogeneous images collected by a binocular acquisition system in an orchard environment.
  • The ToF low-frequency component is simplified after multiple lighting segmentation using a PCNN into a first-order Markov situation. The significance function is defined as first-order Markov mutual information.
  • The significance function is used as the termination condition of a PCNN model iteration, and Kullback–Leibler (KL) divergence is used to measure the dynamic threshold amplification coefficient of the PCNN model.
  • A new momentum-driven multi-objective artificial bee colony algorithm is proposed to optimize the parameters of link channel feedback, link strength, and dynamic threshold attenuation factor. The momentum update strategy of employing bees and observing bees is used. The grid density construction is used to ensure that the optimal solution distribution is not too dense. The absolute value of the difference between the grid index values of the same dimension of the nondominated solution is used as the deletion selection probability of the nondominated solution to construct the optimal solution set. Cross entropy (CE) and mutual information (MI), two image fusion quality evaluation functions, are selected as multi-objective fitness functions.
  • The low-frequency components of ToF and color image after multiple lighting segmentation using a PCNN are fused using the weighted average rule, and the high-frequency components are fused using improved bilateral filters.
Three, the advantage of our work is as follows: the proposed NSST-SMPCNN method combines the saliency mechanism, saliency function, and the PCNN clustering segmentation mechanism, and has the advantages of a grayscale clustering lighting mechanism and the same grayscale attribute first lighting, which is suitable for heterogeneous image fusion in complex orchard environments in Gansu.
The paper structure is summarized as follows: Section 1 contains the introduction and a description of the related works, as well as the highlights and contributions of this paper. Then, basic concept definitions of an NSST and a PCNN are introduced. The proposed definition of significance function is also defined in Section 2. In Section 3, a PCNN model guided by a saliency mechanism is proposed. Then, a PCNN transform domain image fusion method guided by a saliency mechanism is constructed in the Section 4. Lastly, the final section contains a description of the experiment and the conclusions.

2. Basic Concept Definition

2.1. NSST Transform Domain Decomposition Method

The traditional method of spatial domain fusion is to create a fusion model in the image gray space. The disadvantage of this is that it is difficult to find the texture and boundary features of the source image. The NSST transform domain decomposition method, which is proposed in reference [4], is used to perform the non-subsampled pyramid filter bank (NSP) u-level transformation on the two accurately registered heterogeneous images to obtain one low-frequency sub-band and u high-frequency sub-bands, realizing translation invariance. The high-frequency sub-band is then decomposed into 2v directional high-frequency sub-bands by shear filter bank (SF) v-level multi-directional decomposition, so as to effectively capture directional information and maintain anisotropy [4]. The decomposed sub-band is the same size as the source image, has high sparsity, and accurately represents the fusion information.

2.2. PCNN Lighting Segmentation Mechanism

The PCNN model proposed in reference [6] includes feedback input domain, coupling link domain, and pulse generation domain, which can be described by the mathematical equations shown in Formulas (1)–(5). A PCNN has a feature which classifies similar pattern features into categories based on the principles of similarity aggregation and capture characteristics, and has the mechanism of aggregation and illumination segmentation.
F i j ( n ) = I i j
L i j ( n ) = exp ( α L ) L i j ( n 1 ) + V L k l W i j , k l Y k l ( n 1 )
U i j ( n ) = F i j ( n ) ( 1 + β L i j ( n ) )
θ i j ( n ) = exp ( α θ ) θ i j ( n 1 ) + V θ Y i j ( n )
Y i j ( n ) = s t e p ( U i j ( n ) θ i j ( n ) )
In the formula, Iij is the external stimulation of neurons, represented by the gray value of the input image; Fij(n) is the feedback input field; Lij(n) is the link input field; Wij,kl refers to the link coefficient; β indicates the link strength, which determines the weight of the coupling link channel; Uij(n) is the internal state signal of the model; θij is the dynamic threshold of neurons, Vθ and VL are the dynamic threshold amplification coefficients, which control the threshold value increased after neuron activation; αL and αθ determine the decay rate of the feedback term and the dynamic threshold of the link channel, respectively; Yij(n) is the pulse output of the current neuron, which is the response result of the comparison between the internal active item and the dynamic threshold in the pulse generator. When Uij(n) > θij(n), the ignition condition will be reached and the output Yij(n) = 1. Step represents a step function, and its output is 0 or 1; n represents the nth neuron in the image.

2.3. Proposed Definition of Significance Function

The saliency mechanism originates from the visual attention mechanism (VAM) proposed by Itti and other scholars [21], inspired by the behavior and neuronal structure of early primate visual systems [22]. When the saliency mechanism processes a scene, it automatically processes the regions of interest, and selectively ignores the regions of noninterest. In this paper, a new saliency mechanism to define the significance function is proposed.
Definition 1. 
Significance first-order Markov situation.
The ToF low-frequency component after an NSST decomposition is sent into the PCNN model. The PCNN is divided iteratively many times, showing a dynamic ignition segmentation state. The two ignition segmentation diagrams at the time intervals t and t + 1 are correlated, but independent of the ignition segmentation diagram at the previous time. Therefore, the ignition segmentation diagram at the time interval 2 can be defined as a first-order Markov situation.
Definition 2. 
Significance one-step transition probability.
When the model is in the state s u after ignition segmentation at time t, the probability of model transition to the state s v after ignition segmentation at time t + 1 is defined as the significant one-step transition probability, which is expressed in Formula (6).
p u v = p ( S t + 1 = s v S t = s u ) = p ( s v s u ) , s u , s v S
Definition 3. 
Significant conditional entropy.
In the significance first-order Markov situation, the average uncertainty of the model when it is transferred to the state s v S under any state condition s u S is defined as the significance conditional entropy, which is expressed in Formula (7).
H ( U V ) = v = 0 N p ( s u , s v ) log p ( s v s u )
Definition 4. 
Significance first-order Markov information source entropy.
The overall uncertainty of the sequence formed by the ignition segmentation map in the significance first-order Markov situation is defined as the significance first-order Markov information source entropy, which is expressed in Formula (8).
H ( U ) = u = 0 N v = 0 N p ( x v , s u ) log p ( x v s u )
Definition 5. 
Significance first-order Markov mutual information.
The amount of information transmitted in the model state transition of the ignition segmentation map at different times is defined as the significant first-order Markov mutual information, which is expressed as Formula (9).
I ( U ; V ) = H ( U ) H ( U V )
Definition 6. 
Significance function.
The PCNN is divided by multiple iterations. The ignition segmentation graph with two time intervals has significant feature differences, representing the maximum information transmission rate and the maximum amount of mutual information in numerical terms. Because mutual information has a maximum under certain conditions, the significance function is numerically defined as significant first-order Markov mutual information. If Formula (7) and (8) are brought into Formula (9), Formula (10) is formed.
I s a l i e n c y ( u , v ) = I ( U ; V ) = u = 0 N v = 0 N p ( x v , s u ) log p ( x v s u ) + v = 0 N p ( s u , s v ) log p ( s v s u )

3. PCNN Model Guided by Saliency Mechanism

PCNN region segmentation and a saliency mechanism can locate the most interesting object region in the image well. Combining the saliency mechanism, saliency function and PCNN clustering segmentation mechanism, a PCNN model guided by a saliency mechanism is proposed, which has the advantages of a grayscale clustering lighting mechanism and the same grayscale attribute lighting priority, and is suitable for heterogeneous image fusion in complex orchard environments in Gansu.
The PCNN model has certain shortcomings, including that the parameters are limited by manual experience settings and cannot be terminated adaptively, and ignoring the impact of image changes and fluctuations on the results during the ignition process results in pixel artifacts, area blurring, and unclear edges. The iteration termination conditions and dynamic threshold amplification coefficients V θ and the feedback items of the link channel αL, link strength β, and dynamic threshold attenuation factor αθ are improved adaptively. A new momentum-driven multi-objective artificial bee colony algorithm (MMOABC) is used for parameter optimization and is applied to the proposed PCNN model guided by a saliency mechanism. The improved SMPCNN model has the characteristics of enhancing the same type of pulse connection, reducing the difficulty of parameter integration, and improving the performance of image segmentation.

3.1. Adaptive Iteration Termination Conditions

The traditional PCNN model has the defects of nonadaptive termination and over-segmentation. The authors of [23] used the maximum information entropy as the termination condition, but over-segmentation often occurs when the entropy is at its maximum, and the background with the same gray value will be mistaken for the target area and segmented together.
In this paper, the significance function is used as the criterion for model iteration termination, which is expressed as Equation (11). For a low-frequency ignition segmentation map, the greater the significance of the first-order Markov mutual information, the better the regional consistency.
I s a l i e n c y ( u , v ) > δ

3.2. Adaptive Dynamic Threshold Amplification Coefficient V θ

In the ToF image, the fruit target is often shown as a region with a high gray value and normal distribution. Two ignition segmentation images are used to measure the PCNN dynamic threshold amplification coefficient, which is expressed as Equation (12). The probability distribution p ( s u ) corresponding to the state s u S , as well as the probability distribution p ( s v ) corresponding to the state s v S , and the KL divergence of the two states are calculated. This formula is used to measure the similarity between the probability distributions of two ignition segmentation maps. The closer the probability distribution of the two ignition segmentation images is, the smaller the dynamic threshold amplification coefficient is, which will enable the PCNN model to ignite when the target region tends to be stable during continuous iteration.
V θ = D K L ( s v | | s u ) = v = 1 N [ p ( s v ) log p ( s v ) p ( s v ) log p ( s u ) ]

3.3. Parameter Optimization of Momentum Driven Multi-Objective Artificial Bee Colony Algorithm

An artificial bee colony algorithm (ABC) [24] is a swarm intelligence optimization algorithm proposed to simulate the characteristics of bee swarms. It has the advantages of strong global optimization ability, few parameters, high accuracy, and strong robustness. However, its optimization strategy has the defects of simplicity and randomness, which make the algorithm premature, cause convergence stagnation, and other problems. In order to accelerate the convergence rate of the artificial bee colony algorithm, the concept of momentum [25,26] in deep learning is introduced, and a new momentum-driven multi-objective artificial bee colony algorithm is proposed to optimize the three parameters. including the feedback from the link channel αL, link strength β, and dynamic threshold attenuation factor αθ.

3.3.1. Initial Population

The three parameters, including feedback from the link channel αL, link strength β, and dynamic threshold attenuation factor αθ, are used as the initial population of the momentum-driven multi-objective artificial bee colony algorithm. Random generation of NP food source information X = x i j x i j = x i 1 , x i 2 , , x i j , , x i d , i = 1 , 2 , , N P ; j = 1 , 2 , , d = 3 was performed according to Formula (13).
x i j = min j + r a n d ( 0 , 1 ) × ( max j min j )

3.3.2. Hiring Bees Momentum Updating Strategy

NP food source information was randomly generated. During a food update evolution, a randomly selected food source X k = ( x k 1 , x k 2 , , x k d ) was attached to a hired bee in the bee colony. In the d-dimensional space, the randomly selected jth dimension component x i j of each food source X i = ( x i 1 , x i 2 , x i j , , x i d ) in the food source information space database was evolved through the following hired bee momentum update strategy, as shown in Equations (14) and (15), to obtain a new food source X i n e w = ( x i 1 , x i 2 , , x i j n e w , , x i d ) . Among them, i , k [ 1 , 2 , , NP ] , i k , j [ 1 , 2 , , d ] , r [ 1 , 1 ] . In Equations (14) and (15), a i j represents the update step size of the previous update evolution, a i j n e w represents the update step size obtained after the current momentum update evolution, γ represents momentum, and the value is 0.9.
x i j n e w = x i j + a i j n e w
a i j n e w = r × ( x i j x k j ) + γ × a i j

3.3.3. Observation Bees Nesterov Momentum Updating Strategy

In a food update evolution, the selection probability of observation bees was calculated according to Formula (16), and a randomly selected food source X t = ( x t 1 , x t 2 , , x t d ) was attached to an observation bee in the bee colony. In the d-dimensional space, the randomly selected jth dimension component x i j of each food source X i = ( x i 1 , x i 2 , x i j , , x i d ) in the food source information space database was evolved through the following observation bees Nesterov momentum updating strategy, as shown in Formulas (17) and (18), to obtain a new food source X i n e w = ( x i 1 , x i 2 , , x i j n e w , , x i d ) . Among them, i , t [ 1 , 2 , , NP ] , i t , j [ 1 , 2 , , d ] , r [ 1 , 1 ] . In Formulas (17) and (18), b i j represents the update step size of the previous update evolution, b i j n e w represents the update step size obtained after the current Nesterov momentum update evolution, γ represents momentum, and the value is 0.9. T arg e t = 2 , j = 1 , 2 , , N P .
p r o b j = i = 1 T arg e t ( 0.9 × ( f i / max ( f i ) ) + 0.1 ) × ( 1 / T arg e t )
x i j n e w = x i j + b i j n e w
b i j n e w = r × ( x i j x t j γ × b i j ) ) + γ × b i j

3.3.4. Pareto Grid Density Construction Method

In multi-objective optimization problems, individuals are judged by dominance and dense information. In this paper, a grid density construction method is used to ensure that the distribution of optimal solutions in the Pareto optimal solution set (also known as Pareto) is not too dense. The grid is a dynamic, nGrid bisected interval within the range of (−inf, +inf). Here, nGrid is a variable, which represents the number of divided grids. The value inf represents a number, which is far less than infinity.
The maximum and minimum values of each dimension of the median value of the nondominated solution were determined. The predefined nGrid was used to divide the current interval, which was divided into nGrid + 1. The minimum interval starts from negative infinity inf, and the maximum interval ends at positive infinity + inf, to prevent the nondominated solutions from crossing the boundary, and make the nondominated solutions fall in the grid. The formula for solving the grid index value is shown in (19). The value lowi represents the minimum boundary value of the grid, and Target represents the number of objective functions. i = 1 , T arg e t , T arg e t = 2 , j = 1 , , n G r i d .
g r i d i j = inf l o w i + j × ( ( u p e r i l o w i ) / n G r i d ) inf

3.3.5. Pareto Optimal Solution Set Construction Method

First, constructing the optimal solution set requires a certain probability to randomly delete redundant nondominated solutions. The method to construct the deletion selection probability involves the use of the absolute value of the difference between the nondominated solution and the grid index value of the same dimension for operation. The formulas are shown in (20) and (21). The larger p o s s i the nondominated solution corresponding to Formula (20), the harder it will be to delete. The advantage of this is that the preference for a certain optimization objective brought by the nondominated solution interval is reduced, and the unified operation for all optimization objectives can be carried out fairly to obtain a relatively fair solution with the possibility of deletion.
p o s s i = i = 1 Re p j = 1 Re p k = 1 T arg e t g r i d i k g r i d j k
P o s s i = 1 / ( p o s s i + 1 )

3.3.6. Calculation Method of Multi-Objective Fitness

To solve the problem of the diversity of image fusion quality evaluation functions, two image fusion quality evaluation functions, cross entropy (CE) and mutual information (MI), are selected to form a multi-objective optimization problem for two objectives. The formula is shown in (22).
fitness _ pareto = max { CE , MI }

3.4. PCNN Model Structure Guided by Saliency Mechanism

The model structure is shown in Figure 1.

4. PCNN Transform Domain Image Fusion Method Guided by Saliency Mechanism

4.1. Fusion Rules

(1) Low Frequency Fusion Rules
In this paper, using the characteristics of a PCNN model’s clustering and lighting segmentation, the significance function is used as the criterion of a PCNN model’s iteration termination, and the ToF low-frequency component after an NSST decomposition is ignited and segmented. The component is recorded as C T o F L , and the low-frequency component of the color image after an NSST decomposition is recorded as C R G B L . According to the characteristics of the images collected by heterogeneous systems in the mountainous planting environment and the natural scenes of the disordered planting orchard picking operation in the Gansu Province, the low-frequency components of color images have sufficient detailed texture information, while the low-frequency components of ToF images have the characteristics of extracting targets at a certain distance and separating the background, but provide less detailed texture information. Therefore, the ToF low-frequency components and color image low-frequency components after multiple lighting segmentation using a PCNN are fused. The fusion rule uses weighted average, which is expressed as Formula (23), to highlight more foreground information belonging to the highlighted part of the ToF image.
C f u s e L ( m , n ) = 0.5 * C T o F L ( m , n ) + 0.5 * C R G B L ( m , n )
(2) High frequency fusion rules
Bilateral filtering is a local, nonlinear, and noniterative technology. High-frequency fusion rules are introduced to measure the similarity between the ToF image and color image at the corresponding position of the decomposed high-frequency component, as shown in Formula (24). Let the high-frequency component of the ToF image decomposed by an NSST be C T o F H , and the high-frequency component of the color image decomposed by an NSST be C R G B H . The spatial neighborhood Gaussian function w Neighborhood is shown in Equation (25), and the high-frequency component gray value similarity Gaussian function w Similarity is shown in Equation (26).
C f u s e H ( m , n ) = ( t , s ) A r e a ( i , j ) C R G B H ( t , s ) w Neighborhood w Similarity ( t , s ) A r e a ( i , j ) w Neighborhood w Similarity
w Neighborhood = e ( ( m t ) 2 + ( n s ) 2 2 σ 2 )
w Similarity = e ( C T o F H ( m , n ) C R G B H ( t , s ) 2 2 σ 2 )

4.2. Heterogeneous Image Fusion Process

The fusion process is shown in Figure 2.

4.3. NSST-SMPCNN Method Multi-Source Image Fusion Steps

NSST-SMPCNN algorithm is proposed, named as Algorithm 1. The fusion steps of NSST-SMPCNN algorithm for multi-source image are as follows.
Algorithm 1: NSST-SMPCNN.
Input: ToF confidence image and visible light image after registration.
M, N, u, v, n. limit, maxCycle, NP, d, γ , Rep, nGrid, and Target.
Output: Fusion image.
Step 1: NSST is performed, generate u low-frequency sub-band images and 2v high-frequency sub-band images.
Step 2:
n = 1
while (Equation (11))
Begin
Calculate the dynamic threshold amplification coefficient using Equation (12).
Use Equations (13)–(22) to construct MMOABC algorithm, optimize three parameters including αL, β and αθ.
The SMPCNN model is constructed using Equations (11)–(22).
Output ignition diagram.
n ++
End
Step 3: Use Formula (23) to fuse low-frequency components with the weighted average rule; The high-frequency components are fused using the improved bilateral filter of Equations (24)–(26).
Step 4: Perform NSST inverse transform.
Step 5: Stop running and output the fused image.
Note: M and N represent image size, u represents NSST decomposition level, v represents NSST decomposition direction number, and n represents current ignition number. The maximum number of food source stagnation is limit, the maximum number of iterations of algorithm evolution is maxCycle, the number of food sources is NP, and the dimension of bee individual component is d. γ represents momentum, Rep represents the number of nondominated solutions, nGrid represents the number of divided grids, and Target represents the number of objective functions. Where, d = 3, Target = 2, γ = 0.9. αL represents the link channel feedback term, β represents link strength, and αθ represents dynamic threshold attenuation factor.

5. Experiment

5.1. Image Fusion Evaluation Index

Six models were selected for testing to evaluate the image fusion performance of the heterogeneous vision system, including a non-subsampled contourlet transform (NSCT) model [27], a fusion method for infrared and visible light images based on an NSCT (ImNSCT) [28], a DWT model [29], a simplified pulse coupled neural network (SPCNN) model [30], a single target SPCNN fusion model (ST-SPCNN) [31] and the NSST-SMPCNN model described in this paper. Nine objective image evaluation indicators [32] were selected to objectively evaluate image quality, including average gradient (AG), edge strength (ES), information entropy (IE), standard deviation (SD), peak signal to noise ratio (PSNR), spatial frequency (SF), image clarity (IC), mutual information (MI), and structural similarity (SSI). The higher the values of these nine indicators, the better the fusion image quality.

5.2. Public Dataset Image Fusion Experiment

In this paper, three public datasets are used for the experimental testing of heterogeneous image fusion, namely, infrared and color vineyard heterogeneous public datasets taken in natural scenes [33] and apple RGB-D image datasets published by Universitat de Lleida in Spain named fuji_apple [34,35] and PApple_RGB-D-Size [36]. The above three datasets were recorded as dataset I, dataset II and dataset III, respectively, and four groups of data in each of the three datasets were selected for testing. The results are shown in Table 1, Table 2 and Table 3, respectively. The fusion effect is shown in Table 4. The data results show that the objective evaluation indexes of the NSST-SMPCNN method described in this paper, such as AG, ES, SF, IC, and MI are the best in dataset I. For dataset II and dataset III, AG, ES, IE, SF, IC, MI, and other objective evaluation indexes of the first and fourth groups of test data are the best. The values of SD and PSNR of the five other algorithms are better than those of the algorithm in this paper. The SSI value of the DWT algorithm is the best.

5.3. Heterogeneous Image Fusion Experiment of Natural Orchard

In this paper, a heterogeneous vision system is established using a ToF industrial depth camera (Basler AG, Ahrensburg, Germany) and a color camera (Canon Inc., Tokyo, Japan). The ToF camera can output four types of images, including a ToF intensity image, ToF range data, ToF confidence map, and ToF point cloud image [37]. The data collection site in the natural environment is located in the experimental base of the Fruit Research Institute, Qinzhou District, Tianshui City, Gansu Province, China. More than 1000 ToF intensity images, depth images, confidence images, and color images under different lighting conditions between 10:00 and 19:00 were collected using a heterogeneous vision system. The heterogeneous images collected from the natural scene of the orchard were recorded as dataset IV, and four groups of data were selected as samples, including ToF confidence images and corresponding visible light images for testing. The results are shown in Table 5, and the fusion effect is shown in Table 4. The data results show that the NSST-SMPCNN algorithm described in this paper has the best fusion effect on the ToF confidence image and the corresponding visible light image collected in the natural scene. The values of nine indicators, including AG, ES, IE, SD, PSNR, SF, IC, MI, and SSI indicated excellent performance.
In conclusion, the experimental results show that the NSST-SMPCNN algorithm presented in this paper performs well in a test using three common datasets, as indicated by AG, ES, SF, IC, MI, and other objective evaluation indicators. This is because the significance function is used as the iteration termination condition of the PCNN model described in this paper to realize adaptive ignition termination. A new momentum-driven multi-objective artificial bee colony algorithm is used to optimize the PCNN parameters, which enhances the mechanism of the PCNN model’s gray aggregation lighting and same gray attribute priority lighting. For the dataset IV established in this paper, the NSST-SMPCNN algorithm proposed in this paper performs well in nine indicators. This shows that the weighted average rule is used to fuse the low-frequency components, which can highlight more foreground information belonging to the highlighted part in the ToF image. The high-frequency components are fused by the improved bilateral filter, which strengthens the similarity between the ToF image and the color image. The proposed NSST-SMPCNN method is suitable for heterogeneous image fusion in complex orchard environments in Gansu.

6. Conclusions

The traditional method of spatial domain fusion is to create a fusion model in the image gray space, which has the disadvantage of not finding the texture and boundary characteristics of the source image easily. A PCNN model has the defects of parameter experience setting, nonadaptive termination, and easy over-segmentation. This paper proposes a PCNN model guided by the saliency mechanism and applies it to the fusion of ToF and visible light heterogeneous images collected by a binocular acquisition system in an orchard environment. The iteration termination conditions and dynamic threshold amplification coefficients Vθ, the feedback items of the link channel αL, link strength β, and dynamic threshold attenuation factor αθ are improved adaptively. A new momentum-driven multi-objective artificial bee colony algorithm (MMOABC) is used for parameter optimization. The proposed NSST-SMPCNN method combines the saliency mechanism, saliency function and PCNN clustering segmentation mechanism, and has the advantages of a grayscale clustering lighting mechanism and the same grayscale attribute first lighting, which is suitable for heterogeneous image fusion in complex orchard environments in Gansu. The data results show that the NSST-SMPCNN algorithm described in this paper has the best fusion effect on the ToF confidence image and the corresponding visible light image collected in the natural environment. The values of nine indicators, including AG, ES, IE, SD, PSNR, SF, IC, MI, and SSI, indicated excellent performance.
However, some data test results in the public dataset still have the disadvantage of a poor fusion effect, which needs further improvement. In future work, it is necessary to introduce a deep learning convolutional neural network to further explore the algorithm structure to capture better image features and improve the fusion effect.

Author Contributions

L.L. designed and wrote the paper. J.H. collected data and analyzed the experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Gansu Provincial University Teacher Innovation Fund Project [grant number 2023A-051], the Young Supervisor Fund of Gansu Agricultural University [grant number GAU-QDFC-2020-08] and the Gansu Science and Technology Plan [grant number 20JR5RA032].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors of this publication declare no conflict of interest.

References

  1. Bin, Y.A.N.; Pan, F.A.N.; Meirong, W.A.N.G.; Shuaiqi, S.H.I.; Xiaoyan, L.E.I.; Fuzeng, Y.A.N.G. Real-time Apple Picking Pattern Recognition for Picking Robot Based on Improved YOLOv5m. Trans. Chin. Soc. Agric. Mach. 2022, 53, 28–38, 59. [Google Scholar]
  2. Feng, J.; Zeng, L.; Liu, G.; Si, Y. Fruit Recognition Algorithm Based on Multi-source Images Fusion. Trans. Chin. Soc. Agric. Mach. 2014, 45, 73–80. [Google Scholar]
  3. Farina, A.; Morabito, F.C.; Serpico, S.; Simone, G. Fusion of radar images: State of art and perspective. In Proceedings of the International Conference on Radar, Beijing, China, 15–18 October 2001; pp. 9–15. [Google Scholar]
  4. Shilpa, S.; Rajan, M.R.; Asha, C.S.; Shyam, L. Enhanced JAYA optimization based medical image fusion in adaptive non subsampled shearlet transform domain. Eng. Sci. Technol. Int. J. 2022, 35, 101245. [Google Scholar] [CrossRef]
  5. Song, X.; Yang, C.; Han, K.; Ding, S. Robust JPEG steganography based on DCT and SVD in nonsubsampled shearlet transform domain. Multimed. Tools Appl. 2022, 81, 36453–36472. [Google Scholar] [CrossRef]
  6. Eckhorn, R.; Reitboeck, H.J.; Arndt, M. Feature linking via synchronization among distributed assemblies: Simulations of results from cat visual cortex. Neural Comput. 1999, 2, 293–307. [Google Scholar] [CrossRef]
  7. Cheng, B.; Jin, L.; Li, G. Infrared and visual image fusion using LNSST and an adaptive dual-channel PCNN with triple-linking strength. Neurocomputing 2018, 310, 135–147. [Google Scholar] [CrossRef]
  8. Panigrahy, C.; Seal, A.; Mahato, N.K. MRI and SPECT image fusion using a weighted parameter adaptive dual channel PCNN. IEEE Signal Process. Lett. 2020, 27, 690–694. [Google Scholar] [CrossRef]
  9. Liu, C.H.; Qi, Y.; Ding, W.R. Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 2017, 83, 94–102. [Google Scholar] [CrossRef]
  10. Yang, Y.; Lu, H.; Huang, S.; Tu, W. Remote sensing image fusion based on fuzzy logic and salience measure. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1943–1947. [Google Scholar] [CrossRef]
  11. Li, L.; Liu, Z.; Zou, W.; Zhang, X.; Le Meur, O. Co-saliency detection based on region-level fusion and pixel-level refinement. In Proceedings of the 2014 IEEE International Conference on Multimedia and Expo (ICME), Chengdu, China, 14–18 July 2014; pp. 1–6. [Google Scholar]
  12. Zhu, M.; Chen, C.; Wu, X. Infrared and Visible Image Fusion Using Improved QPSO-PCNN Algorithm. In Proceedings of the 2018 8th International Conference on Electronics Information and Emergency Communication (ICEIEC), Beijing, China, 15–17 June 2018; pp. 52–58. [Google Scholar]
  13. Huang, C.; Tian, G.; Lan, Y.; Peng, Y.; Ng, E.Y.K.; Hao, Y.; Cheng, Y.; Che, W. A new pulse coupled neural network (PCNN) for brain medical image fusion empowered by shuffled frog leaping algorithm. Front. Neurosci. 2019, 13, 210. [Google Scholar] [CrossRef] [Green Version]
  14. Dharini, S.; Jain, S. A novel metaheuristic optimal feature selection framework for object detection with improved detection accuracy based on pulse-coupled neural network. Soft Comput. 2022, 26, 12943–12955. [Google Scholar] [CrossRef]
  15. Muhuri, A.; Manickam, S.; Bhattacharya, A. Snow Cover Mapping Using Polarization Fraction Variation with Temporal RADARSAT-2 C-Band Full-Polarimetric SAR Data over the Indian Himalayas. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2192–2209. [Google Scholar] [CrossRef]
  16. Raskar, R.; Beardsley, P.; van Baar, J.; Wang, Y.; Dietz, P.; Lee, J.; Leigh, D.; Willwacher, T. RFIG lamps: Interacting with a self-describing world via photosensing wireless tags and projectors. ACM Trans. Graph. 2004, 23, 406–415. [Google Scholar] [CrossRef]
  17. Qiu, T.; Li, Y.J. Fast LDP-MST: An efficient density-peak-based clustering method for large-size datasets. IEEE Trans. Knowl. Data Eng. 2022, 1, 1. [Google Scholar] [CrossRef]
  18. Huang, S.; Zhang, H.; Pižurica, A. Subspace Clustering for Hyperspectral Images via Dictionary Learning with Adaptive Regularization. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5524017. [Google Scholar] [CrossRef]
  19. Huang, D.; Wang, C.D.; Wu, J.S.; Lai, J.H.; Kwoh, C.K. Ultra-Scalable Spectral Clustering and Ensemble Clustering. IEEE Trans. Knowl. Data Eng. 2020, 32, 1212–1226. [Google Scholar] [CrossRef] [Green Version]
  20. Santana-Carrillo, R.; González-Flores, J.S.; Magaña-Espinal, E.; Quezada, L.F.; Sun, G.H.; Dong, S.H. Quantum Information Entropy of Hyperbolic Potentials in Fractional Schrödinger Equation. Entropy 2022, 24, 1516. [Google Scholar] [CrossRef]
  21. Itti, L.; Koch, C. Computational modeling of visual attention. Nat. Rev. Neurosci. 2001, 2, 194–203. [Google Scholar] [CrossRef] [Green Version]
  22. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  23. Hao, J.; Wang, L. Infrared image segmentation of circuit board based on improved PCNN. Laser Infrared 2020, 50, 1410–1416. [Google Scholar]
  24. Karaboga, D. An Idea Based on Honey Bee Swarm for Numerical Optimization; Erciyes University: Kayseri, Turkey, 2005. [Google Scholar]
  25. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; Posts and Telecommunications Press: Beijing, China, 2017. [Google Scholar]
  26. Khan, S. A Guide to Convolutional Neural Networks for Computer Vision; China Machine Press: Beijing, China, 2019. [Google Scholar]
  27. Da, C.A.L.; Zhou, J.P.; Do, M.N. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar]
  28. Zhang, J.L.; Zhao, E. Fusion method for infrared and visible light images based on NSCT. Laser Infrared 2013, 43, 320–323. [Google Scholar]
  29. Chao, R.; Zhang, K.; Li, Y. An image fusion algorithm using wavelet transform. Acta Electron. Sin. 2004, 5, 750–753. [Google Scholar]
  30. Yang, Z.; Ma, Y.; Lian, J.; Zhu, L. Saliency motivated improved simplified PCNN model for object segmentation. Neurocomputing 2018, 275, 2179–2190. [Google Scholar]
  31. Liu, L.; Gu, R.; Zhou, Y.; Huo, J. Heterologous image fusion with multi-scale decomposition and dual optimization SPCNN in an apple orchard. Trans. Chin. Soc. Agric. Eng. (Trans. CSAE) 2022, 38, 158–167. [Google Scholar]
  32. Yang, Y.; Li, J.; Wang, Y. Review of image fusion quality evaluation methods. J. Front. Comput. Sci. Technol. 2018, 12, 1021–1035. [Google Scholar]
  33. Brown, M.; Süsstrunk, S. Multi-spectral SIFT for scene category recognition. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011. [Google Scholar]
  34. Gené-Mola, J.; Vilaplana, V.; Rosell-Polo, J.R.; Morros, J.R.; Ruiz-Hidalgo, J.; Gregorio, E. Multi-modal Deep Learning for Fruit Detection Using RGB-D Cameras and their Radiometric Capabilities. Comput. Electron. Agric. 2019, 162, 689–698. [Google Scholar] [CrossRef]
  35. Gené-Mola, J.; Vilaplana, V.; Rosell-Polo, J.R.; Morros, J.R.; Ruiz-Hidalgo, J.; Gregorio, E. KFuji RGB-DS database: Fuji apple multi-modal images for fruit detection with color, depth and range-corrected IR data. Data Brief 2019, 25, 104289. [Google Scholar] [CrossRef] [PubMed]
  36. Ferrer Ferrer, M.; Ruiz-Hidalgo, J.; Gregorio, E.; Vilaplana, V.; Morros, J.R.; Gené-Mola, J. Simultaneous Fruit Detection and Size Estimation Using Multitask Deep Neural Networks [EB/OL]. Available online: https://www.grap.udl.cat/en/publications/papple_rgb-d-size-dataset (accessed on 1 October 2022).
  37. Temer, A.M. Basler ToF Camera User’s Manual. Available online: https://www.baslerweb.com/cn/sales-support/downloads/document-downloads/basler-tof-camera-users-manual/ (accessed on 1 July 2020).
Figure 1. PCNN model structure guided by saliency mechanism.
Figure 1. PCNN model structure guided by saliency mechanism.
Sensors 23 02488 g001
Figure 2. Fusion Process.
Figure 2. Fusion Process.
Sensors 23 02488 g002
Table 1. Objective evaluation index results of dataset I.
Table 1. Objective evaluation index results of dataset I.
Test GroupingAlgorithmAGESIESDPSNRSFICMISSIRuntime/s
First groupNSCT8.0478.807.11108.3814.7419.3110.251.170.7742.94
ImNSCT7.4171.367.11110.7014.0117.959.761.050.7453.29
DWT6.5764.487.11146.0513.6315.978.431.250.791.10
SPCNN7.4175.697.15162.288.1916.348.750.020.32112.76
ST-SPCNN8.2380.767.08170.9115.4119.8010.610.730.7291.59
NSST-SMPCNN9.9896.837.19161.1314.1724.3813.072.530.6860.10
Second groupNSCT5.5654.656.58160.8815.0616.407.091.160.8543.11
ImNSCT5.2851.566.94159.3414.6616.876.851.230.8453.33
DWT4.1840.995.71199.1515.0113.255.371.490.870.87
SPCNN4.7148.286.65187.5710.8912.585.520.080.58114.17
ST-SPCNN5.4654.047.01196.3715.7915.986.960.760.8292.43
NSST-SMPCNN6.3162.026.78189.4615.5219.308.072.380.7959.37
Third groupNSCT8.4982.457.1098.4715.6420.5610.941.310.8043.16
ImNSCT7.1668.686.9991.2214.5018.409.541.170.7753.28
DWT7.3471.077.16137.4014.6818.259.611.400.820.86
SPCNN7.6477.767.45151.717.1216.959.050.100.28114.47
ST-SPCNN8.0878.107.25158.3515.5820.3310.620.800.7390.58
NSST-SMPCNN9.4191.467.42149.4914.5223.4312.312.720.7459.08
Fourth groupNSCT6.3562.866.69127.6014.2617.198.001.220.7843.20
ImNSCT6.3162.486.67105.3112.2818.408.041.030.7153.22
DWT5.0850.876.68156.0413.7113.736.351.140.780.87
SPCNN6.1764.087.21146.586.8214.957.110.110.38113.38
ST-SPCNN6.6466.047.15153.5015.2317.878.401.030.7294.33
NSST-SMPCNN7.0169.046.90176.6512.3020.058.942.600.7061.12
Table 2. Objective evaluation index results of dataset II.
Table 2. Objective evaluation index results of dataset II.
Test GroupingAlgorithmAGESIESDPSNRSFICMISSIRuntime/s
First groupNSCT8.8094.417.3196.3413.5919.619.531.150.5551.96
ImNSCT10.00107.217.48125.0412.2122.5610.861.000.4950.65
DWT8.1188.067.5490.3814.1217.298.581.340.630.94
SPCNN9.1499.346.4354.122.4520.989.680.080.11106.01
ST-SPCNN9.52102.866.3653.9012.3922.7410.220.620.4785.08
NSST-SMPCNN11.38122.067.80114.7614.2726.1812.422.480.5768.69
Second groupNSCT7.8183.897.3490.9413.9816.278.430.980.5251.01
ImNSCT7.1476.687.3089.1514.2115.547.760.900.4950.14
DWT5.4559.257.1561.8918.7610.975.751.140.660.9
SPCNN7.0576.646.8152.552.5214.697.430.130.13104.93
ST-SPCNN7.2378.306.8654.7617.5515.187.690.790.6184.15
NSST-SMPCNN7.4178.847.3584.9513.5116.568.182.050.5564.48
Third groupNSCT8.5491.397.55116.3911.8919.589.281.040.5652.29
ImNSCT7.5981.147.36119.1610.5318.028.290.880.5249.77
DWT6.0765.737.4687.3912.5513.406.451.170.640.93
SPCNN7.7884.386.9958.632.2917.718.220.040.14105.81
ST-SPCNN8.1988.547.2771.3311.0718.598.720.610.5585.24
NSST-SMPCNN8.2988.027.59126.498.9919.489.172.100.5064.03
Fourth groupNSCT11.93127.807.65130.2111.2725.1412.951.130.5351.34
ImNSCT11.17119.457.49132.189.6023.5612.150.940.5049.97
DWT8.4191.167.5191.9210.0217.158.951.350.560.91
SPCNN9.35101.566.5052.412.5020.539.900.080.11103.57
ST-SPCNN9.52103.076.4753.109.3021.1110.140.610.4483.84
NSST-SMPCNN12.06128.267.83127.8210.2126.2613.342.250.5062.62
Table 3. Objective evaluation index results of dataset III.
Table 3. Objective evaluation index results of dataset III.
Test GroupingAlgorithmAGESIESDPSNRSFICMISSIRuntime/s
First groupNSCT7.3671.327.04112.6215.4922.779.461.020.7243.89
ImNSCT6.3456.886.71119.2915.0523.829.230.780.7055.02
DWT4.7946.616.80122.3015.6016.556.301.770.730.96
SPCNN5.4155.795.97127.0715.2021.206.481.840.69117.29
ST-SPCNN7.6673.536.6889.0614.7327.1210.400.420.6793.79
NSST-SMPCNN7.9081.417.16116.074.9919.369.200.000.2662.61
Second groupNSCT9.0688.947.16143.2510.0025.2311.450.800.6343.66
ImNSCT7.6369.526.83136.049.3925.9010.830.570.6355.61
DWT5.2551.776.70158.349.9916.456.761.690.670.94
SPCNN7.0576.646.8152.552.5214.697.430.130.13117.76
ST-SPCNN8.5582.826.9298.3610.4027.7611.410.340.5994.81
NSST-SMPCNN7.1467.166.02211.496.3938.289.941.580.4763.42
Third groupNSCT8.0081.027.08106.8713.6020.549.600.860.6746.2
ImNSCT6.5764.206.69108.2112.4818.488.520.620.6655.79
DWT4.5145.926.65100.7512.8312.185.471.600.710.96
SPCNN6.2164.696.5566.752.5515.227.060.000.22116.82
ST-SPCNN6.8768.946.8280.9111.3919.648.590.430.6693.99
NSST-SMPCNN6.8267.305.82133.859.3929.888.861.730.5362.89
Fourth groupNSCT7.6478.407.34128.7410.6819.098.951.000.6544.07
ImNSCT6.3462.947.02118.238.5817.258.040.780.6555.75
DWT4.4546.057.00110.588.7511.575.231.910.700.97
SPCNN5.8861.876.7363.672.4113.896.560.000.22117.76
ST-SPCNN6.3864.846.7060.888.9617.197.750.330.5793.13
NSST-SMPCNN11.54107.645.88174.902.6563.3616.441.790.4162.43
Table 4. Fusion effect of four datasets.
Table 4. Fusion effect of four datasets.
DatasetTest GroupingNSCTImNSCTDWTSPCNNST-SPCNNNSST-SMPCNN
dataset IFirst groupSensors 23 02488 i001Sensors 23 02488 i002Sensors 23 02488 i003Sensors 23 02488 i004Sensors 23 02488 i005Sensors 23 02488 i006
Second groupSensors 23 02488 i007Sensors 23 02488 i008Sensors 23 02488 i009Sensors 23 02488 i010Sensors 23 02488 i011Sensors 23 02488 i012
Third groupSensors 23 02488 i013Sensors 23 02488 i014Sensors 23 02488 i015Sensors 23 02488 i016Sensors 23 02488 i017Sensors 23 02488 i018
Fourth groupSensors 23 02488 i019Sensors 23 02488 i020Sensors 23 02488 i021Sensors 23 02488 i022Sensors 23 02488 i023Sensors 23 02488 i024
dataset IIFirst groupSensors 23 02488 i025Sensors 23 02488 i026Sensors 23 02488 i027Sensors 23 02488 i028Sensors 23 02488 i029Sensors 23 02488 i030
Second groupSensors 23 02488 i031Sensors 23 02488 i032Sensors 23 02488 i033Sensors 23 02488 i034Sensors 23 02488 i035Sensors 23 02488 i036
Third groupSensors 23 02488 i037Sensors 23 02488 i038Sensors 23 02488 i039Sensors 23 02488 i040Sensors 23 02488 i041Sensors 23 02488 i042
Fourth groupSensors 23 02488 i043Sensors 23 02488 i044Sensors 23 02488 i045Sensors 23 02488 i046Sensors 23 02488 i047Sensors 23 02488 i048
dataset IIIFirst groupSensors 23 02488 i049Sensors 23 02488 i050Sensors 23 02488 i051Sensors 23 02488 i052Sensors 23 02488 i053Sensors 23 02488 i054
Second groupSensors 23 02488 i055Sensors 23 02488 i056Sensors 23 02488 i057Sensors 23 02488 i058Sensors 23 02488 i059Sensors 23 02488 i060
Third groupSensors 23 02488 i061Sensors 23 02488 i062Sensors 23 02488 i063Sensors 23 02488 i064Sensors 23 02488 i065Sensors 23 02488 i066
Fourth groupSensors 23 02488 i067Sensors 23 02488 i068Sensors 23 02488 i069Sensors 23 02488 i070Sensors 23 02488 i071Sensors 23 02488 i072
dataset IVFirst groupSensors 23 02488 i073Sensors 23 02488 i074Sensors 23 02488 i075Sensors 23 02488 i076Sensors 23 02488 i077Sensors 23 02488 i078
Second groupSensors 23 02488 i079Sensors 23 02488 i080Sensors 23 02488 i081Sensors 23 02488 i082Sensors 23 02488 i083Sensors 23 02488 i084
Third groupSensors 23 02488 i085Sensors 23 02488 i086Sensors 23 02488 i087Sensors 23 02488 i088Sensors 23 02488 i089Sensors 23 02488 i090
Fourth groupSensors 23 02488 i091Sensors 23 02488 i092Sensors 23 02488 i093Sensors 23 02488 i094Sensors 23 02488 i095Sensors 23 02488 i096
Table 5. Objective evaluation index results of dataset IV.
Table 5. Objective evaluation index results of dataset IV.
Test GroupingAlgorithmAGESIESDPSNRSFICMISSIRuntime/s
First groupNSCT8.8786.296.7870.6712.9022.7911.410.740.4942.92
ImNSCT10.3699.706.8991.0813.0227.7314.020.540.5052.60
DWT7.6674.246.8575.1813.0321.3510.221.320.491.05
SPCNN9.7498.906.8053.192.1524.5711.660.010.08111.08
ST-SPCNN9.7594.745.9932.6813.3028.7212.910.290.3390.74
NSST-SMPCNN12.05116.217.57111.2018.9432.6016.172.390.5461.39
Second groupNSCT9.0188.306.8870.6411.8823.3011.640.920.4642.96
ImNSCT10.96105.587.0294.9112.2029.3914.990.690.4853.08
DWT8.0778.126.9773.3912.0422.6910.911.770.450.91
SPCNN8.4486.345.4424.331.8023.4910.080.010.04115.60
ST-SPCNN9.7495.155.5825.7212.3129.5913.010.230.2690.66
NSST-SMPCNN13.16127.127.78116.5020.4735.3117.852.690.5262.22
Third groupNSCT7.9978.466.5881.3513.7120.009.560.480.5343.00
ImNSCT5.3249.746.2278.1513.7814.646.460.380.4552.86
DWT6.6365.816.76100.2714.1317.088.090.690.550.91
SPCNN8.4786.136.8252.131.8121.039.530.000.10114.72
ST-SPCNN8.8887.546.8954.0913.9022.1011.110.390.3988.77
NSST-SMPCNN8.9888.857.10129.2517.6622.0711.462.220.5360.59
Fourth groupNSCT7.0168.996.6173.4311.7318.978.810.780.4942.97
ImNSCT8.0978.216.7893.7611.7222.8510.700.610.4955.65
DWT6.3362.386.7488.3012.0018.638.191.120.510.91
SPCNN8.1483.606.5841.471.4121.139.550.000.08113.80
ST-SPCNN8.5884.806.3537.4312.8325.5011.060.370.3890.20
NSST-SMPCNN9.5193.207.05127.4916.3127.4312.362.290.5359.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, L.; Huo, J. PCNN Model Guided by Saliency Mechanism for Image Fusion in Transform Domain. Sensors 2023, 23, 2488. https://doi.org/10.3390/s23052488

AMA Style

Liu L, Huo J. PCNN Model Guided by Saliency Mechanism for Image Fusion in Transform Domain. Sensors. 2023; 23(5):2488. https://doi.org/10.3390/s23052488

Chicago/Turabian Style

Liu, Liqun, and Jiuyuan Huo. 2023. "PCNN Model Guided by Saliency Mechanism for Image Fusion in Transform Domain" Sensors 23, no. 5: 2488. https://doi.org/10.3390/s23052488

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop