Next Article in Journal
Statistical Significance Testing for Mixed Priors: A Combined Bayesian and Frequentist Analysis
Previous Article in Journal
An Information-Theoretic Bound on p-Values for Detecting Communities Shared between Weighted Labeled Graphs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception

Jiangsu Key Laboratory of Spectral Imaging and Intelligent Sense, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(10), 1327; https://doi.org/10.3390/e24101327
Submission received: 21 July 2022 / Revised: 19 September 2022 / Accepted: 19 September 2022 / Published: 21 September 2022
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Infrared-visible fusion has great potential in night-vision enhancement for intelligent vehicles. The fusion performance depends on fusion rules that balance target saliency and visual perception. However, most existing methods do not have explicit and effective rules, which leads to the poor contrast and saliency of the target. In this paper, we propose the SGVPGAN, an adversarial framework for high-quality infrared-visible image fusion, which consists of an infrared-visible image fusion network based on Adversarial Semantic Guidance (ASG) and Adversarial Visual Perception (AVP) modules. Specifically, the ASG module transfers the semantics of the target and background to the fusion process for target highlighting. The AVP module analyzes the visual features from the global structure and local details of the visible and fusion images and then guides the fusion network to adaptively generate a weight map of signal completion so that the resulting fusion images possess a natural and visible appearance. We construct a joint distribution function between the fusion images and the corresponding semantics and use the discriminator to improve the fusion performance in terms of natural appearance and target saliency. Experimental results demonstrate that our proposed ASG and AVP modules can effectively guide the image-fusion process by selectively preserving the details in visible images and the salient information of targets in infrared images. The SGVPGAN exhibits significant improvements over other fusion methods.

1. Introduction

Driving assistance at night has always been a focus for intelligent vehicles. With the development of sensor technology, long-wave infrared cameras have been applied in intelligent systems to make up for the poor night vision performance of visible-light cameras. Due to the different characteristics of the sensors, there exist significant differences between visible-light images and infrared images. Visible images usually possess rich textures with a higher resolution compared with infrared images. However, the image quality can be easily affected by the surrounding conditions. In contrast, infrared images are captured based on the temperature characteristics or emissivity of the object. Targets with significant thermal radiation characteristics in infrared images are more salient, which makes it much easier for humans to find the targets. Although infrared images provide relatively fewer details, the image quality is not in-line with the perception of human eyes. Therefore, combining two complementary images to obtain a fused image with high quality is strongly demanded by many practical applications.
Over the past decades, many image fusion methods have been proposed [1]. These methods can be roughly divided into two categories: traditional image-fusion algorithms and deep-learning-based image-fusion algorithms [2]. In particular, traditional methods consist of multi-scale transform image fusion [3], sparse representation image fusion [4], low-rank representation image fusion [5], subspace-based image fusion [6], and saliency-based image fusion [7]. However, traditional algorithms are mainly hand-crafted. To achieve a better fusion result, the traditional algorithms are becoming more and more complex, which leads to the difficulty of practical applications and tedious computing time. In addition, a large number of traditional algorithms ignore the semantic information of the image and the saliency of the targets, resulting in fuzzy and illegible targets in fusion images.
Therefore, with the rise of deep learning and neural networks over recent years, deep-learning techniques applied to infrared and visible image-fusion methods have also emerged [8]. Deep-learning-based methods can be broadly classified into two categories: neural-network-based image fusion with end-to-end training and generative adversarial-network-based image fusion. Generative adversarial networks (GAN) are a unsunpervised deep-learning method, where the generative network and the discriminator contest with each other. GAN-based methods do not need groundtruth and generate results with realistic characteristics.
From an information entropy perspective, GAN is a framework for estimating generative models through an adversarial process, which consists of two models: a generative model G that captures the distribution of the data, and a discriminative model D that estimates the probability that the samples came from the training data rather than G. The training procedure for G is to maximize the probability of D making a mistake. This framework corresponds to a minimax two-player game.
Although the existing deep-learning-based infrared and visible image-fusion methods perform well to some extent, these methods still suffer from several drawbacks. First, since IR and visible image fusion is an unsupervised task lacking labels for the fused images, existing fusion networks often directly use subjective losses [9]. Second, existing image-fusion networks ignore high-level semantic information and focus only on the fusion of global images. As a result, the local fusion results of the injected images are poor and the saliency of the targets is low [10].
Unlike image-fusion tasks, recognition tasks have a clear goal, such as a semantic-segmentation task, which relies on a large amount of labeled data [11]. The combination helps to improve the saliency and contrast of image fusion. Semantic-segmentation networks based on deep learning focus on mining the high-level semantic features of the image and restoring the resolution to the original image. The simple segmentation of the image also represents its target saliency. As a classical pixel-level task, semantic-segmentation techniques can extract semantics from the input data and play an important supporting role in many other tasks. The performance of many deep-learning-based tasks has been improved by exploiting the features of semantic segmentation. In image-fusion tasks, some methods claim to show that semantic segmentation is beneficial in guiding image fusion. However, these methods require obtaining segmentation labels as prior knowledge before image fusion, which is costly in terms of time and manpower during the testing phase.
To address the above issues, we propose Semantic Guided Visual Perception GAN (SGVPGAN), an adversarial framework for high-quality IR-visible image fusion. SGVPGAN is based on an Adversarial Semantic Guidance (ASG) and an Adversarial Visual Perception (AVP) module for visible-like and target-highlighted appearances. In addition, we constructed a joint distribution function between the fused images and the corresponding semantics and used the discriminator to improve the fusion performance of natural appearance and targets. Through qualitative subjective evaluation and quantitative objective evaluation, we find that SGVPGAN offers superior image-fusion performance compared to other infrared and visible image-fusion methods.
Our primary contributions are summarized as follows:
  • We propose SGVPGAN, an adversarial framework for IR-visible image fusion, which has a visible-like and target-highlighted appearance.
  • We develop the ASG and AVP for global and local enhancement during the training phase.
  • The proposed SGVPGAN achieves significant improvements in terms of both visual appearance and quantification results.

2. Related Work

In this section, we briefly describe the background and related work, including generative adversarial networks, semantic segmentation, and the deep-learning-based fusion of infrared and visible images.

2.1. Generative Adversarial Network

The concept of generative adversarial networks was first introduced by Goodfellow et al. [12]. It has profound and wide applications in the field of image generation [13]. From the perspective of data distribution in image generation, GAN is an implicit generative model and the training-data distribution is contained in the trained generator G. Therefore, given any target domain A, a sample point similar to the distribution of domain A can be obtained through the GAN.
To achieve a better quality of generated images and a more stable adversarial training process, GAN has derived a series of variants during development. DCGAN replaces the multilayer perceptron (MLB) in the generator and discriminator of the original GAN with a convolutional neural network (CNN) for feature extraction. LSGAN uses least square loss instead of the cross-entropy loss of the initial GAN, which improves the performance of image generation and makes training more stable. CGAN adds additional conditions to the GAN, making the GAN generation process controllable [14]. WGAN introduces the Wasserstein distance and simply adapts the original GAN and achieves a surprising performance. To some extent, it solves the problems of pattern collapse, training difficulties, and instability, and the loss of generators can account for problems in the training process [15]. BigGAN improves the performance of GAN by increasing the number of parameters and scaling up the batch size. It uses the truncation technique to increase the stability of the training process. At the same time, it offers a certain balance between the stability of training and the performance of the network [16].
U-Net GAN is a recently developed GAN with excellent image-generation performance, which achieves state-of-the-art performance on several datasets [17]. Based on BigGAN, U-Net GAN changes the discriminator to a U-Net structure. The encoder discriminates the input image globally, while the decoder discriminates the image pixels locally. As a result, the generated images have higher quality and more realistic texture details.
The discriminator of our proposed network structure adopts the idea of U-Net GAN. We design the discriminator as a simple U-shaped network that classifies the global fusion results and improves the texture details of the fused images to some extent.

2.2. Semantic Segmentation

Semantic segmentation is a fundamental topic in computer vision, which recognizes targets at the pixel level. FCN was the first network that made a major breakthrough in the field of image segmentation using deep-learning methods. FCN replaced the fully connected layer of a neural network with a convolutional layer, thus creating a fully convolutional neural-network design. The U-Net structure [18] was proposed by Olaf et al. It was the first neural network for medical-image segmentation, and its network structure was more or less adopted by later semantic segmentation networks. Semantic-segmentation networks usually have a large number of parameters but often require real-time performance in their applications. Many semantic segmentation networks for real-time requirements have been proposed, such as ENet [19], ERFNet [20], ICNet [21], and BiseNet [22]. RPNet proposes a novel feature residual pyramid, which can help improve detail and edge information [23].
Since semantic segmentation not only extracts semantic features of images but also classifies images at the pixel level, it is often used in detection, tracking, and image-fusion tasks to provide target information and improve models. In detection tasks, DES uses segmentation modules to obtain attention masks and uses advanced semantic features to improve detection performance for small targets [24]. In the tracking task, Xue et al. use a segmentation network to obtain semantic information. The segmentation network provides an ROI mask to suppress background interference and improve the performance of the tracking network [25].
In an image-fusion task, the segmentation method preserves more semantic information for the image-fusion process. Hou et al. used the semantic-segmentation mask as prior knowledge for the image-fusion network. The mask divides the image into foreground and background and performs high-quality fusion so that the fused image retains more information [26]. SGVPGAN uses the semantic information extracted by the semantic segmentation network to guide the image-fusion network to generate fused images. Tang et al. cascade the image-fusion module and semantic-segmentation module and leverage the semantic loss to guide high-level semantic information to flow back to the image-fusion module, which effectively boosts the performance of high-level vision tasks on fused images [27]. Zhou et al. design an information quantity discrimination (IQD) block to guide fusion progress and identify the distribution of infrared and visible information in the fused image [28]. Wu et al. employ the semantic segmentation technology to obtain the semantic region of each object, and generate special weight maps for the infrared and visible image via pre-designed fusion weights [29].

2.3. Deep-Learning-Based Infrared and Visible Image Fusion

Deep-learning-based image-fusion methods mainly consist of supervised methods and GAN methods. Liu et al. used CNN to fuse IR and visible images, but they still used some traditional algorithms. The network is used to generate a weighted Gaussian pyramid and fuse each layer based on the weights. Deepfuse designed the network structure for image fusion, including feature-fusion and image-reconstruction modules. This network sets an objective human constraint that is determined by the SSIM loss between the fused image and the original image [30]. Densefuse improves the network structure by adding dense connections based on Deepfuse and adds L2 regularization to the SSMI loss [31]. Based on the physical prior, DRF innovatively decomposes the image into a physical property representation vector and a scene-representation feature map after being constrained by certain criteria and trained by the network. The fused image is reconstructed by fusing the two physical attributes [32].
Recently, GAN-based approaches have emerged and demonstrated significant potential. The FusionGAN proposed by Ma et al. is the first GAN-based network for fusing infrared and visible images [9]. As an end-to-end model, FusionGAN does not require the manual design of complex fusion rules. However, the infrared component of the fused images gradually diminishes during the training of FusionGAN. To address this problem, the authors propose DDcGAN with dual discriminators. Two discriminators can discriminate the fused image from the input infrared image and visible image, which enhances the infrared component of the fused image to some extent. Meanwhile, the authors tried to add detail loss and target enhancement loss to FusionGAN to solve the problem of the low saliency and coarse details of targets in FusionGAN [33]. Hou et al. proposed a semantic-segmentation-based fusion network for visible and infrared images [26]. The network adds image segmentation labels as prior knowledge to the network to utilize this information for image fusion. However, this also means that the labels need to be prepared in advance for testing.
Recently, a unified framework to solve different fusion problems, including multi-modal, multi-exposure, and multi-focus cases has became a research hot-spot. U2Fusion automatically estimates the importance of adaptive preservation degrees for effective and universal image fusion [34]. SwinFusion combines cross-domain long-range learning and Swin Transformer to integrate complementary information and global interaction [35]. GIDGAN designs a powerful gradient and intensity-discriminator generative adversarial network for gradient and intensity retention in image-fusion tasks [36].
In practical applications, labels for image segmentation are not always easy to obtain. Our proposed network architecture only uses labels as input during training, and does not require segmentation labels during testing. This avoids the extra work of manual labeling in real applications.

3. Method

3.1. Overall Network Architecture

The SGVPGAN framework consists of three components: a generator, a perceptual discriminator, and a semantic auxiliary module. The overall structure of SGVPGAN is shown in Figure 1. The RGB image I v i s and the infrared image I i r are passed through two encoders, and the outputs of the two encoders are fused with the decoder of the single luminance channel of the image, which is used to reconstruct the fused RGB image I f with the color components of the original visible image I v i s . The perceptual discriminator is used to discriminate the true RGB image from the fused image to obtain a true fusion result, and the semantic auxiliary module is used to extract the semantics of the target and background from the fused image I f to guide the fusion network to generate a fused image with great target saliency. The generator and the discriminator establish an adversarial structure for SGVPGAN network training. In the testing phase, the generator fuses IR-visual image pairs to obtain realistic fused images without segmentation labels.

3.1.1. Generator Network Architecture

The generator structure is shown in Figure 2. This generator adopts two encoders for both infrared and visual image extraction and a decoder for the feature fusion.
We feed the visible image I v i s and the infrared image I i r into the dual encoder individually. These two encoders share the same network structure. Each convolution layer uses 3 × 3 convolution and the scale of the feature map is kept constant. In both decoders, the number of channels of the feature map is increased. There is no pooling layer during the process to prevent the loss of image information. Referring to DenseNet [37], we use dense concatenation to continuously replenish the information of the forward features. Dense concatenation ensures that the shallow features can be reused and efficiently utilized in deep convolution, which can effectively help our fused images retain more detailed information.
The output feature maps from the two encoders are fused in a cascaded manner and then fed to the decoder. The number of channels of the feature maps is gradually reduced during the decoder process. Finally, we use the sigmoid activation function to obtain a probability map with the number of channels as 2. The two channels are used as the probability distribution of the visible and infrared components in the fused image. The output feature maps of the two channels are points multiplied with the luminance channels of the visible and infrared images, respectively, and then are added together. After the tanh activation function, the fused single-channel image is produced. We confirmed through experiments that this operation can prevent the target from darkening as a function of the segmentation network. Throughout the architecture of the generator, we used spectral normalization (SN) [38], after each convolutional layer. To prevent gradient explosion or gradient disappearance and to speed up the convergence of the network, we added batch normalization (BN) to the network. We set Leaky Relu as the activation function. On the contrary, Relu loses negative values of the feature map during fusion, and information is lost for the fusion task. Leaky Relu can solve this problem.

3.1.2. Perceptual Discriminator

The luminance channels of the visible image I v i s _ y and the fused single image I f _ y are fed into the encoder in an unpaired manner to build a generative and adversarial structure. To obtain the global and local discrimination between the visible image and the fused image, and enhance the performance of the fused images on structure and textures, the perceptual discriminator employs a conditional U-shaped network with attention to detail and local structural information. The conditional discriminator utilizes the segmentation label as the condition in the discrimination of the visible image and the fused image to ensure the correspondence of the semantics. In addition, the U-shaped discriminator makes strong spatial discrimination [17]. The structure of the Perceptual discriminator is shown in Figure 3.
To construct the joint distribution function for perceptual and semantic enhancement, the input image and the corresponding segmentation label are concatenated together before being fed to the encoder. The label is an auxiliary condition. By adding segmentation labels, the discriminator can judge the fused image with higher quality based on this. It is helpful to optimize the details of the fused image, to discriminate the fused image spatially based on the high-level semantics in a reasonable way, and to impose certain constraints on the pixel-level fusion of the fused image. In other words, some high-level semantic information is sent to the U-discriminator, which drives the image fusion based on the semantic information and increases the amount of information in the fused image.
After being fed to the encoder, the number of channels of the feature map increases in full convolution. For each convolution, the size of the feature map is halved. The encoder extracts the global features throughout the process and, finally, obtains the global discriminant result through the global pooling layer and fully connected layer. Our global discriminant result is to discriminate the overall appearance of the visual image and the fused image. This improves the overall image perception of the fused image and makes it more natural.
In the decoder, we adopt transpose convolution for the output high-level feature map of the encoder. For each transpose convolution, the number of channels decreases while the size of the feature map doubles, creating an asymmetric relationship with the encoder. We continuously supplement the forward information with skipped connections, which effectively reuses the lost information in the encoder. After restoring the feature map to the original image size, we obtain a discriminant of the original image size by using convolution. This decision can be interpreted as a pixel-level decision in the image space. It allows to make a judgment on the local texture structure of the fused image and gives feedback to the generator in space. In the fusion task, this spatial decision can give more visibility to the local texture details of the fused image and improve the naturalness of the fused image from a local perspective.
After each layer of the discriminator, spectrum normalization is employed to increase the stability of the GAN in the training process. Similar to the generator, each layer uses batch normalization and Leaky Relu as the activation function.

3.1.3. Semantic Auxiliary Module

We employ the RPNet segmentation network as a semantic auxiliary module. RPNet segmentation networks are based on residual pyramids, which have fewer parameters, fast inference and good segmentation performance. The color channel of the visible image is connected to the fused luminance channel, and then the YUV image is converted to an RGB image I f and fed into RPNet. Finally, we obtain a probability map and compute the segmentation loss with labels.
The segmentation network works as an auxiliary module. On the one hand, the network enhances the ability to mine the semantic features of the fused images. On the other hand, semantic information guides the generator to fuse images with better target saliency. The constraint of segmentation loss motivates the segmentation network to learn the semantics of the fused image, which, in turn, guides the fused image to make a suitable fusion in space to achieve high-quality image fusion. In contrast to the direct MSE loss between the fused and IR images, our semantic auxiliary module parses the fused images in a high-level manner to guide image fusion. This approach considers the spatial distribution of the fused images instead of the coarse global MSE loss.

3.2. Loss Function

The discriminator D and the generator G play the following two-player minimax game. It is expressed in the form of the sum of two minus cross-entropy losses (real samples and generated samples). This means that the discriminator maximizes the distribution gap between generated and training data, while the generator minimizes the distance between generated and training data.
min θ ( G ) max θ ( D ) E x P d a t a log D ( x ) + E x P G log ( 1 D ( G ( x ) ) )
where E expresses the expectation of the data distribution. The complete cross-entropy of the training sample is 1 log D x 1 + 0 log D x 0 , where D x 0 and D x 1 are the discriminator judgment probability (whether the training sample is a training sample (1) or a generated sample (0)). The full cross-entropy of the generated samples is 1 log D G z 1 + 0 log D G z 0 , where D G z 0 and D G z 1 represent the discriminator judgment probability (whether the generated sample is a training sample (1) or a generated sample (0)). D G z needs to be as small as possible. However, in order to align (maximize) the direction of the cross-entropy function with the training sample, it is represented by log ( 1 D ( G ( x ) ) ) .
In our network, there are three loss functions: visual perceptual loss (AVP loss), semantic guidance loss (ASG loss), and generator loss in the generative adversarial structure, which is used to train perceptual discriminator, semantic discriminator, and generator, respectively. The generator loss is composed of the perceptual adversarial loss, the semantic adversarial loss, and the detail loss.

3.2.1. Visual Perceptual Loss

The perceptual discriminator is continuously enhanced to distinguish the visible image from the fused image during the training process and continuously gives feedback to the generator. The discriminator will discriminate it as true when the luminance channel I v i s _ y of the visible image is input, and it will discriminate it as false when the fused image I f _ y is input as the fused image. The visual perceptual loss is conducted at the end of the encoder and decoder for structure and detail discrimination. We use D U to denote the discriminator, which consists of an encoder and a decoder, representing D e n c U and D d e c U , respectively. The discriminator’s encoder outputs global information loss L D e n c U , and the discriminator’s decoder outputs local information loss L D d e c U . We use the following formula,
L D e n c U U = E I v i s _ y min 0 , 1 + D e n c U I v i s _ y E I v i s , I i r min 0 , 1 D e n c U G I v i s , I i r
L D d e c U U = E I v i s _ y min 0 , 1 + D d e c U I v i s _ y E I v i s , I i r min 0 , 1 D d e c U G I v i s , I i r
where [ D d e c U ( I v i s _ y ) ] i , j and [ D d e c U ( G ( I v i s , I i r ) ) ] i , j refer to the discriminator decision at pixel (i,j). E ( · ) indicates the mathematical expectation of the decision distribution. We follow the hinge loss proposed in U-Net GAN [17]. Our two loss functions represent global and local decision distances, respectively. Thus, in the process of the continuous strengthening of the discriminator, the discriminator cannot only make global decisions but also local decisions.

3.2.2. Semantic Guidance Loss

The semantic guidance loss is based on the semantic auxiliary module. We adopt the RPNet segmentation network; however, we do not use the auxiliary loss in RPNet for simplicity. After inputting RPNet the fusion image I f , we obtain the output segmentation result I p r e d from RPNet. Then we calculate cross-entropy loss between I p r e d and I l a b e l . Both of them are normalized to [−1, 1]. The semantic guidance loss formula is shown as follows:
L s a d v = L s e g = 1 W H i , j c = 0 N 1 y c ( i , j ) l o g [ p c ( i , j ) ]
where y c ( i , j ) denotes the value of channel c of I l a b e l one-hot vector at pixel ( i , j ) . A one-hot vector here refers to the classification of whether a pixel is foreground or background, with foreground as vector [1, 0] and background as vector [0, 1]. p c ( i , j ) indicates the prediction value of the channel c of I p r e d at pixel (i,j). N refers to the number of channels. W and H refer to the width and height of the image.

3.2.3. Generator Loss

The loss function of the generator consists of three main components: perceptual adversarial loss L p a d v , semantic adversarial loss L s a d v , and detail loss L d e t a i l . The perceptual adversarial loss L p a d v is used to guide the fused image being recognized as real in the discrimination, and the global and local details of the fused image tend to be perceptual to visible light. The semantic adversarial loss L s a d v guides the fused image towards easy image segmentation. Since IR images contain richer semantic information, the semantic adversarial loss L s a d v is equivalent to adding target saliency information from IR images to the fused images, thus improving the target saliency of the fused images. The detail loss L d e t a i l is used to enhance the visible detail information of the fused image.
When computing the perceptual adversarial loss L p a d v , the parameters of the discriminator are fixed and the parameters of the generator are trained. The purpose of the generator is to train the fused image that can cheat the discriminator. In addition, the perceptual adversarial loss L p a d v is calculated as follows:
L p a d v = E I v i s , I i r D e n c U ( G ( I v i s , I i r ) ) + [ D d e c U ( G ( I v i s , I i r ) ) ] i , j
When computing the semantic adversarial loss L s a d v , the parameters of the semantic auxiliary module are fixed and the parameters of the generator are trained. The generator is continuously adjusted during the training process and gradually outputs fused images with better segmentation performance. The target saliency of the image will be improved and the effective information of the IR image will gradually increase in the fused image. The segmentation loss in the generator is also the cross-entropy loss. In FusionGAN, the content loss is the average difference between the pixel values of the fused image and the IR image. This results in the whole fused image becoming blurred due to too many infrared components. In contrast to this approach, we use adversarial segmentation loss to regionally increase the infrared component to make the target of the fused image more salient.
By using the adversarial loss to train the generator, the distance between the fused image and the visible light is gradually reduced so that the fused image has visible-light image perception. The output of encoder and decoder losses make the fused image globally and locally constrained by the visible image. Specifically, we follow the hinge loss used in the training of the U-Net GAN.
The detail loss L d e t a i l is the distance between the gradient of the fused image and the visible image. We calculate the gradient of both and find the average of the L2 criterion of the difference between the two gradients. Its formula is as follows.
L d e t a i l = 1 W H i , j ( I f _ y ( i , j ) I v i s _ y ( i , j ) ) 2
where ∇ denotes the gradient of the image, ( i , j ) is the position of the pixel, W and H represent the width and height of the image, respectively. Detail loss is used to close the gradient of the fused image towards the gradient of visible light, which allows the fused image to have more detailed information.
Combining the three losses above, we obtain the loss function of the generator. The formula is as follows:
L G = L p a d v + α L s a d v + ε L d e t a i l
where α and ε are hyperparameters balancing the weights of the three losses.

4. Experimental Results

4.1. Dataset

We evaluated our algorithm on the MF datase t [39]. The dataset is designed for semantic segmentation of visible and infrared images, which contains images of street scenes for autonomous vehicles. There are 1569 sets of RGB-T image pairs in the dataset, with 820 for daytime scenes and 749 for nighttime scenes. We trained the nighttime RGB-T dataset with 374 pairs of RGB-T images in the training set, 187 in the validation set, and 188 in the test set. The sizes of the images are 640 × 480 . Although this dataset has been used as a dataset for semantic segmentation, the RGB-T images in this dataset are mostly pixel-aligned and, thus, can be used for image fusion. The dataset is also rich in night scenes, often with dim lights, glare, and other common driving scenes. Each pair of RGB-T images has a corresponding label. In addition to the unlabeled classes, the labels contain eight classes. Here, we use only three classes that are common in driving: cars, pedestrians, and bicycles. The pedestrian class is more prominent in the IR images.

4.2. Training and Test Details

In the training phase, data augmentation was adopted. We randomly cropped the images with a crop size of 400 × 400. Meanwhile, we used random shifting and random horizontal flipping for data augmentation, the detailed steps in the training phase is shown in Algorithm 1.
Algorithm 1: Training  SGVPGAN.
Entropy 24 01327 i001
The tasks of GAN and semantic segmentation are very different. In practice, the semantic loss is easily converged, but the perceptual loss fluctuates as the generator and discriminator slowly converge in an adversarial way. That means that after several epochs, the perceptual loss will be much bigger than the semantic loss. Therefore, in the setting of the network loss hyperparameters, the semantic adversarial loss hyperparameter was set to 10,000 and the detail loss hyperparameter was set to 100. During the training phase, we set batchsize to B s . For each epoch, we first trained the perceptual discriminator M times, then trained the semantic auxiliary module N times, and, finally, trained the generator once. The optimizer was Adam and the training epoch number was K. In our experiments, we set the parameters as B s = 4 , M = 2 , N = 2 , and K = 300 , and the number of images in the training set to N U M = 374 . Regarding the loss parameters, we chose α = 10,000, ε = 100.
In the test phase, we kept only the generator. We removed the data-augmentation methods and input the RGB-T test images with the original size to obtain the fusion result. The NVIDIA TITAN RTX and 32 GB of memory were used for training and testing.

4.3. Evaluation Metrics

Based on the benchmark VIFB [40] for visible and infrared image fusion, we conclude that the various evaluations of image fusion can hardly reflect the quality of image fusion, indeed; however, some objective evaluations can reflect the quality of image fusion to some extent. Therefore, we will demonstrate the superiority of our fusion method by both objective and subjective evaluations in this section.

4.3.1. Objective Evaluation

We used the following metrics to evaluate the performance: AG (average gradient), EI (edge intensity), SF (spatial frequency), EN (entropy), mutual information (MI), and multi-exposure fusion structural similarity (MEF-SSIM) [41]. These evaluations can quantitatively evaluate our image quality based on image features and information theory, and can comprehensively evaluate our image-fusion quality. Our fusion image is based on target saliency, so we compared our image quality with other methods under the mask of segmentation labels.
AG (average gradient) is employed to measure the clarity of the fused image. The greater the average gradient, the better the image clarity and the better the quality of image fusion [42]. The formula of A G is as follows:
A G = 1 ( M 1 ) ( N 1 ) i = 1 M 1 j = 1 N 1 ( I f ( i + 1 , j ) I f ( i , j ) ) 2 + ( I f ( i , j + 1 ) I f ( i , j ) ) 2 2
where M and N denote the width and height of the image, respectively.
EI (edge intensity) calculates the edge intensity of the fused image. Similarly, there is no need for the original reference images. The greater the edge intensity, the better the quality of the fused image [43]. The formula of AG is shown as follows:
E I ( I f ) = i = 1 M j = 1 N ( s x ( i , j ) 2 + s y ( i , j ) 2 ) M N s x = I f h x , s y = I f h y
where h x and h y are the Sobel operator in x and y directions. s x and s y are the results after employing Sobel operator to the fused image.
SF (spatial frequency) was used to calculate the change rate of the image gray level. The larger the spatial frequency, the clearer the image and the better the fusion effect [44]. The formula of AG is as follows:
S F = R F 2 + C F 2 R F = 1 M N i = 1 M j = 2 N ( I ( i , j ) I ( i , j 1 ) ) 2 C F = 1 M N i = 2 M j = 1 N ( I ( i , j ) I ( i 1 , j ) ) 2
The larger the S F , the higher the image quality.
EN (entropy) calculates the amount of information contained in the image. A larger value indicates more information in the image, indicating more details [45]. The formula of E N is as follows: where p I f ( i ) is the statistical probability of the gray histogram.
E N ( I f ) = i = 0 255 p I f ( i ) l o g 2 ( p I f )
MI measures the degree of dependency between visible and infrared images; the calculation method refers to [46]. MEF-SSIM is an advanced MEF image-quality model, and the calculation method refers to [47].
We chose the typical traditional methods and deep-learning methods developed recently to compare the performance, including ADF [48], CNN [49], DLF [50], FPDE [8], GFF [6], Hybrid-MSD [49], MGFF [51], MST-SR [52], ResNet [53], RP-SR [54] TIF [55], VSMWLS [56] and FusionGAN [9]. The results are shown in Table 1:
The above methods include excellent traditional methods and deep-learning methods in recent years. Bold and underlined numbers indicate that the method ranks first in the fusion evaluation index, and only bold numbers indicate that the method ranks second in the fusion evaluation index. From the comparison of these objective evaluations, our fusion method has advantages over these methods.
As the measures used mainly evaluate the richness of textures, they do not directly indicate the fusion performance, e.g., the noise in the dark visible images will interfere with the quantitative evaluation. The proposed method highlights the target in the fusion result and maintains the textures at the same time. In addition, the inference time per image of the proposed method is competitive.

4.3.2. Subjective Evaluation

As there are many evaluations for image-fusion tasks, it is difficult to evaluate them in a uniform manner when comparing various methods in practice. Furthermore, although they are highly evaluated quantitatively, the perception of image quality may not be satisfactory. Therefore, here, we evaluate the fused images qualitatively and compare our fused images with other fusion methods.
As can be seen from the sample 1 in Figure 4, our image basically preserves the details of the billboard of the visible image. In addition, the infrared radiation details of the building can be retained in our image. However, other methods retain less such information. More importantly, it is difficult for other methods to observe the faint IR information of distant pedestrians, while in our method we can see the distant people more clearly, which proves that pedestrians with distinct IR radiation features will be more salient in our fusion method.
From the sample 2 in Figure 4, we can see that our method still retains the details such as leaves in the visible image and the infrared details of buildings next to leaves are also evident. The bicycles with more salient detail information in the infrared image can also be displayed in the fusion image. Furthermore, our method clearly retains human information and has a high contrast ratio. Other methods clearly lose target saliency information in the infrared image.
The sample 3 in Figure 4 shows a strong light scene at night. In visible light, pedestrians, cars and other targets cannot be seen due to the presence of glare. In our fusion image, the car and the person in the car are more salient than in other fusion methods, and the person on the right is also more salient. In other fusion images, it is almost difficult to distinguish the car, the person in the car, and the person on the right. It can be seen that the significant advantage of our target prominence in the fused image is obvious in the strong light scene. On the contrary, in other methods, the target in the glare scene loses some infrared information, and less target information is retained, and it is even difficult to identify the target.

4.4. Ablation Study

4.4.1. U-Shaped Discriminator Experiment

We design a U-shaped discriminator in the AVP module to distinguish the fused image globally and locally, so as to enhance the local detail and perception of the image. To validate the effectiveness of our method, we removed the decoder of the U-shaped discriminator under the condition that the overall framework of the original network remains unchanged.
The comparison of the objective evaluation results of the U-shaped ablation experiment is as in Table 2 and the subjective evaluation results is as in Figure 5.
After the U-shaped discriminator is added, the objective indicators A G and E I are improved, which means that the fused image contains more information and the image fusion effect is improved.
From Figure 5, we observe that our fused image offers more visible detail information when using a U-shaped discriminator. For example, the details of the billboard and light in the visible image are well-retained when we use the U-shaped discriminator. This shows that our U-shaped discriminator enhances the visible detail of fused images and the visible perception of the image to a certain extent. Obviously, the detail of the pedestrian is also rich with a U-shaped discriminator.

4.4.2. Conditional Discriminator Experiment

At the input of the perceptual, we also input the segmentation label as the prior knowledge of the discriminator, so that the discriminator can obtain the target prior information. We maDe the network automatically select the fusion rules spatially. The discriminator has the ability to perceive high-level semantics and spatial information. In order to demonstrate the effectiveness of our method, we conducted comparison experiments. The results are shown in Table 3 and Figure 6:
After the conditional discriminator is added, the objective indicators A G , E I , S F , and E N are improved, which means that the fused image contains more information and the image fusion effect is improved. From Figure 6, it is clear to see that the contrast of the ground is significantly enhanced and the detail information is richer when the conditional discriminator is used. This indicates that the segmentation labels added as a prior knowledge can optimize the spatial distribution and improve the perceptual capability of the fusion network.

4.4.3. Semantic Auxiliary Module Experiment

To show the role of the semantic auxiliary module, we adjust the hyperparameters of the semantic adversarial loss of the generators to show the effect of the semantic auxiliary module on the fused images.
We adjusted the hyperparameters of the semantic adversarial loss to 1000, 5000, 10,000, and 20,000. The comparison of the objective evaluation results of the different losses experiment is as in Table 4 and the objective evaluation results are as in Figure 7.
It is obvious that when the hyperparameters are small, there are more visible details in the fused images, but the targets with distinct IR thermal features are significantly blurred, which means less IR salient information. As α gradually increases, the semantics of the segmentation network gradually guides the progress of fusion, and more IR image components containing salient semantic information are added to the fused image. However, when α is larger, the detail loss of the visible image is larger, and more infrared information makes the fused image blurred. Therefore, we need to reasonably adjust the hyperparameter α to keep a balance between the detailed information of the visible image and the prominent information of the infrared image. Based on the experimental results, we choose α = 10,000 to balance this relationship.

5. Conclusions

In this paper, we propose a novel infrared and visible image-fusion framework SGVPGAN for night vision enhancement, which has a visible-like and target-highlighting appearance. The pixel-level ASG and AVP in the framework combine semantic segmentation and visual enhancement, and significantly improve the fused images. Experiments on the MF dataset demonstrate that our fusion method outperforms other popular fusion methods with qualitative and quantitative evaluations. In the future, we will optimize our network to achieve better performance and improve the training strategy of the GAN architecture. The proposed method has a wide range of applications in intelligent vehicles and great potential in scene perception.

Author Contributions

Conceptualization, X.C.; methodology, X.C. and Z.T.; software, Z.T.; validation, X.C. and Z.T.; formal analysis, Z.T.; investigation, Y.L.; resources, Y.L.; data curation, Y.L.; writing—original draft preparation, J.L.; writing—review and editing, X.C. and J.H.; visualization, J.L.; supervision, L.B. and J.H.; project administration, L.B.; funding acquisition, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China (62101256) and China Postdoctoral Science Foundation (2021M691591).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.mi.t.u-tokyo.ac.jp/static/projects/mil_multispectral/, accessed on 18 September 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, S.; Kang, X.; Fang, L.; Hu, J.; Yin, H. Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 2017, 33, 100–112. [Google Scholar] [CrossRef]
  2. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  3. Chen, J.; Li, X.; Luo, L.; Mei, X.; Ma, J. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 2020, 508, 64–78. [Google Scholar] [CrossRef]
  4. Zhang, Q.; Liu, Y.; Blum, R.S.; Han, J.; Tao, D. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Inf. Fusion 2018, 40, 57–75. [Google Scholar] [CrossRef]
  5. Li, H.; Wu, X.J.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef]
  6. Bavirisetti, D.P.; Xiao, G.; Liu, G. Multi-sensor image fusion based on fourth order partial differential equations. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–9. [Google Scholar]
  7. Zhang, X.; Ma, Y.; Fan, F.; Zhang, Y.; Huang, J. Infrared and visible image fusion via saliency analysis and local edge-preserving multi-scale decomposition. JOSA A 2017, 34, 1400–1410. [Google Scholar] [CrossRef]
  8. Li, H.; Wu, X.J.; Kittler, J. Infrared and visible image fusion using a deep learning framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2705–2710. [Google Scholar]
  9. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  10. Fu, Y.; Wu, X.J.; Durrani, T. Image fusion based on generative adversarial network consistent with perception. Inf. Fusion 2021, 72, 110–125. [Google Scholar] [CrossRef]
  11. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef]
  12. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar]
  13. Gui, J.; Sun, Z.; Wen, Y.; Tao, D.; Ye, J. A review on generative adversarial networks: Algorithms, theory, and applications. arXiv 2020, arXiv:2001.06937. [Google Scholar] [CrossRef]
  14. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  15. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  16. Brock, A.; Donahue, J.; Simonyan, K. Large scale GAN training for high fidelity natural image synthesis. arXiv 2018, arXiv:1809.11096. [Google Scholar]
  17. Schonfeld, E.; Schiele, B.; Khoreva, A. A u-net based discriminator for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8207–8216. [Google Scholar]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar]
  20. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 19, 263–272. [Google Scholar] [CrossRef]
  21. Zhao, H.; Qi, X.; Shen, X.; Shi, J.; Jia, J. Icnet for real-time semantic segmentation on high-resolution images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 405–420. [Google Scholar]
  22. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Bisenet: Bilateral segmentation network for real-time semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  23. Chen, X.; Lou, X.; Bai, L.; Han, J. Residual pyramid learning for single-shot semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2019, 21, 2990–3000. [Google Scholar] [CrossRef]
  24. Zhang, Z.; Qiao, S.; Xie, C.; Shen, W.; Wang, B.; Yuille, A.L. Single-shot object detection with enriched semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5813–5821. [Google Scholar]
  25. Xue, X.; Li, Y.; Yin, X.; Shang, C.; Peng, T.; Shen, Q. Semantic-Aware Real-Time Correlation Tracking Framework for UAV Videos. IEEE Trans. Cybern. 2020, 52, 2418–2429. [Google Scholar] [CrossRef] [PubMed]
  26. Hou, J.; Zhang, D.; Wu, W.; Ma, J.; Zhou, H. A generative adversarial network for infrared and visible image fusion based on semantic segmentation. Entropy 2021, 23, 376. [Google Scholar] [CrossRef]
  27. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  28. Zhou, H.; Wu, W.; Zhang, Y.; Ma, J.; Ling, H. Semantic-supervised Infrared and Visible Image Fusion via a Dual-discriminator Generative Adversarial Network. IEEE Trans. Multimed. 2021. [Google Scholar] [CrossRef]
  29. WU, W.; ZHANG, D.; HOU, J.; WANG, Y.; LU, T.; ZHOU, H. Semantic Guided Infrared and Visible Image Fusion. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2021, E104.A, 1733–1738. [Google Scholar] [CrossRef]
  30. Ram Prabhakar, K.; Sai Srikar, V.; Venkatesh Babu, R. Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4714–4722. [Google Scholar]
  31. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  32. Xu, H.; Wang, X.; Ma, J. DRF: Disentangled representation for visible and infrared image fusion. IEEE Trans. Instrum. Meas. 2021, 70, 5006713. [Google Scholar] [CrossRef]
  33. Ma, J.; Liang, P.; Yu, W.; Chen, C.; Guo, X.; Wu, J.; Jiang, J. Infrared and visible image fusion via detail preserving adversarial learning. Inf. Fusion 2020, 54, 85–98. [Google Scholar] [CrossRef]
  34. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  35. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  36. Zhou, H.; Hou, J.; Zhang, Y.; Ma, J.; Ling, H. Unified gradient- and intensity-discriminator generative adversarial network for image fusion. Inf. Fusion 2022, 88, 184–201. [Google Scholar] [CrossRef]
  37. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. Densenet: Implementing efficient convnet descriptor pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  38. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral normalization for generative adversarial networks. arXiv 2018, arXiv:1802.05957. [Google Scholar]
  39. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar]
  40. Zhang, X.; Ye, P.; Xiao, G. VIFB: A visible and infrared image fusion benchmark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 104–105. [Google Scholar]
  41. Ma, K.; Duanmu, Z.; Yeganeh, H.; Wang, Z. Multi-Exposure Image Fusion by Optimizing A Structural Similarity Index. IEEE Trans. Comput. Imaging 2018, 4, 60–72. [Google Scholar] [CrossRef]
  42. Cui, G.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt. Commun. 2015, 341, 199–209. [Google Scholar] [CrossRef]
  43. Rajalingam, B.; Priya, R. Hybrid multimodality medical image fusion technique for feature enhancement in medical diagnosis. Int. J. Eng. Sci. Invent. 2018, 2, 52–60. [Google Scholar]
  44. Eskicioglu, A.M.; Fisher, P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995, 43, 2959–2965. [Google Scholar] [CrossRef]
  45. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar]
  46. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 1. [Google Scholar] [CrossRef]
  47. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  48. Bavirisetti, D.P.; Dhuli, R. Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sensors J. 2015, 16, 203–209. [Google Scholar] [CrossRef]
  49. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  50. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresoluti. Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  51. Zhou, Z.; Wang, B.; Li, S.; Dong, M. Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion 2016, 30, 15–26. [Google Scholar] [CrossRef]
  52. Bavirisetti, D.P.; Xiao, G.; Zhao, J.; Dhuli, R.; Liu, G. Multi-scale guided image and video fusion: A fast and efficient approach. Circuits, Syst. Signal Process. 2019, 38, 5576–5605. [Google Scholar] [CrossRef]
  53. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  54. Li, H.; Wu, X.J.; Durrani, T.S. Infrared and visible image fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 2019, 102, 103039. [Google Scholar] [CrossRef]
  55. Bavirisetti, D.P.; Dhuli, R. Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 2016, 76, 52–64. [Google Scholar] [CrossRef]
  56. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
Figure 1. SGVPGAN network architecture. Solid arrows indicate the forward pass and dashed arrows indicate the backward pass.
Figure 1. SGVPGAN network architecture. Solid arrows indicate the forward pass and dashed arrows indicate the backward pass.
Entropy 24 01327 g001
Figure 2. Inference flowchart of the generator.
Figure 2. Inference flowchart of the generator.
Entropy 24 01327 g002
Figure 3. The structure of the perceptual discriminator.
Figure 3. The structure of the perceptual discriminator.
Entropy 24 01327 g003
Figure 4. The sample image of subjective experiments of various methods on the MF dataset.
Figure 4. The sample image of subjective experiments of various methods on the MF dataset.
Entropy 24 01327 g004
Figure 5. Experiment on ablation of the U-shaped discriminator.
Figure 5. Experiment on ablation of the U-shaped discriminator.
Entropy 24 01327 g005
Figure 6. Experiment on ablation of the segmentation label.
Figure 6. Experiment on ablation of the segmentation label.
Entropy 24 01327 g006
Figure 7. Experiment on the hyperparameters of the semantic auxiliary module.
Figure 7. Experiment on the hyperparameters of the semantic auxiliary module.
Entropy 24 01327 g007
Table 1. Objective evaluation of fusion algorithms. The bold-underlined indicates the best, and the bold indicates the second best.
Table 1. Objective evaluation of fusion algorithms. The bold-underlined indicates the best, and the bold indicates the second best.
AGEISFENMIMEF-SSIMInference Time (s)
ADF0.1661.772.580.3781.700.9190.619
CNN0.2072.223.050.3822.170.9046.448
DLF0.1621.742.550.3771.780.9326.726
FPDE0.1691.812.590.3791.700.9151.450
GFF0.2132.283.110.3921.640.9070.393
Hybrid-MSD0.2272.423.260.3912.270.9044.929
MGFF0.2192.343.110.3851.450.9021.455
MST-SR0.2092.243.050.3862.140.9030.709
ResNet0.1621.742.540.3781.710.9320.532
RP-SR0.2002.133.070.3941.600.9050.369
TIF0.2162.323.140.3881.520.9040.061
VSMWLS0.2222.363.190.3871.780.9052.351
FusionGAN0.1701.832.670.3771.490.9140.369
Ours0.2492.693.250.3941.950.9220.121
Table 2. Objective evaluation results of the U-shaped discriminator ablation experiment. The bold indicate the best.
Table 2. Objective evaluation results of the U-shaped discriminator ablation experiment. The bold indicate the best.
AGEI
Without a U-shaped discriminator0.2362.553
With a U-shaped discriminator0.2482.689
Table 3. Objective evaluation results of the conditional discriminator ablation experiment. The bold indicate the best.
Table 3. Objective evaluation results of the conditional discriminator ablation experiment. The bold indicate the best.
AGEISFEN
Without a segmentation label0.2172.3510.3893.106
With a segmentation label0.2482.6890.3943.246
Table 4. Objective evaluation results of different hyperparameters of the semantic adversarial loss experiment. The bold indicate the best.
Table 4. Objective evaluation results of different hyperparameters of the semantic adversarial loss experiment. The bold indicate the best.
AGEISFEN
α = 10000.1391.5040.3652.246
α = 50000.1982.1280.3792.700
α = 10,0000.2482.6890.3943.246
α = 20,0000.2352.5400.3953.243
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, X.; Teng, Z.; Liu, Y.; Lu, J.; Bai, L.; Han, J. Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception. Entropy 2022, 24, 1327. https://doi.org/10.3390/e24101327

AMA Style

Chen X, Teng Z, Liu Y, Lu J, Bai L, Han J. Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception. Entropy. 2022; 24(10):1327. https://doi.org/10.3390/e24101327

Chicago/Turabian Style

Chen, Xiaoyu, Zhijie Teng, Yingqi Liu, Jun Lu, Lianfa Bai, and Jing Han. 2022. "Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception" Entropy 24, no. 10: 1327. https://doi.org/10.3390/e24101327

APA Style

Chen, X., Teng, Z., Liu, Y., Lu, J., Bai, L., & Han, J. (2022). Infrared-Visible Image Fusion Based on Semantic Guidance and Visual Perception. Entropy, 24(10), 1327. https://doi.org/10.3390/e24101327

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop