Next Article in Journal
Effects of High-Pressure Processing (HPP) on Antioxidant Vitamins (A, C, and E) and Antioxidant Activity in Fruit and Vegetable Preparations: A Review
Previous Article in Journal
Contextual Real-Time Optimization on FPGA by Dynamic Selection of Chaotic Maps and Adaptive Metaheuristics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Two-Stage Approach for Infrared and Visible Image Fusion and Segmentation

by
Wang Ren
1,
Lanhua Luo
1,2 and
Jia Ren
3,*
1
Faculty of Data Science, City University of Macau, Macao 999078, China
2
School of Artificial Intelligence, Hezhou University, Hezhou 542899, China
3
School of Information and Communication Engineering, Hainan University, Haikou 570100, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(19), 10698; https://doi.org/10.3390/app151910698
Submission received: 10 August 2025 / Revised: 18 September 2025 / Accepted: 29 September 2025 / Published: 3 October 2025

Abstract

Early studies rarely considered cascades among multiple tasks, such as image fusion and semantic segmentation tasks. Most image fusion methods fail to consider the interrelationship between image fusion and segmentation. We propose a new two-stage infrared and visible image fusion and segmentation method called TSFS. By cascading the fusion module and the segmentation module in the first stage, we obtain better fusion results and enhance the semantic information transferred to the second stage. The fusion module in the first stage of this method uses a feature extraction module (FEM) to extract deep features and then fuses these features through a feature mixture fusion module (FMFM). To enhance the fusion performance of multimodal data in the second-stage segmentation network, we propose a Cross-Semantic Fusion Attention Module (CSFAM) to cross-fuse these features. Compared to state-of-the-art methods, experimental evaluations on public datasets reveal that our proposed TSFS improves mIoU segmentation by 1.5% and 3.3% on the FMB and MFNet datasets, respectively, and produces visually better fused images.

1. Introduction

With the rapid pace at which that technology is advancing, image fusion displays enormous potential for application in computer vision, enabling the development of many sectors and having an important position. Image fusion aims to fuse the images obtained by different imaging systems to generate an image that is rich in detailed texture information and has a friendly visual effect. Multimodal images can be made more usable and of higher quality through image fusion, which also serves as a strong basis for further processing and analysis.
Visible sensors are prone to being affected by bad environments, such as low light or nighttime. However, infrared sensors offer special benefits at night or in low light and can detect thermal radiation information from objects with sensitivity. Visible sensors can capture the light that a surface of an object reflects. Visible sensors not only can obtain visible images that offer extensive texture and detailed information but also can obtain visible images that are closer to human visual perception in environments with better lighting conditions. Visible sensors cannot achieve all-day and all-weather imaging at night or under partial occlusion. They also cannot penetrate the obstructed area to monitor the target object. The resulting visible image is shown in Figure 1a,d. As shown in Figure 1b,e, infrared sensors are easily interfered with by background noise, and the imaging results lack texture detail information.
Imaging sensors have enhanced the information richness in the process of continuous development. The fusion results generated by imaging from different sensors through image fusion technology can not only help provide more efficient technical support for subsequent advanced vision tasks but also effectively enhance the performance of subsequent tasks. Considering the difference in complementary characteristics between visible sensor and infrared sensor imaging, fusion results are generated based on features containing complementary information extracted from infrared and visible images. As shown in Figure 1, the generated fusion result not only effectively integrates the rich detailed texture information in the visible image but also fuses the significant infrared information of the person in the infrared image. The fusion of infrared and visible images can specifically handle feature information with complementary characteristics. By integrating multimodal information, their advantages can be combined, which is crucial in areas such as target detection [1], target recognition [2], and target tracking [3].
Generally, image fusion methods can be classified into two main categories: traditional fusion methods and deep learning-based methods [4,5]. Since the CNN-based multi-focus fusion method was first proposed by Liu et al. [6] and achieved better performance than the most advanced multi-focus image fusion methods, subsequent research works have proposed numerous CNN-based infrared and visible image fusion (IVIF) to further improve the fusion quality. TC-GAN [7] proposed to construct a single-channel generator consisting of an encoder, a squeeze-and-excitation (SE-Net) [8] attention module, and a decoder to generate a texture map, which is input into adaptive guided filtering to obtain a decision map to guide image fusion. NestFuse proposed to fuse the encoder output and input it into the decoder through a spatial attention model based on regional energy and a channel attention model based on regional pooling [9].
Although the above-mentioned image fusion method based on a deep learning method has certain advantages in feature extraction, due to the lack of constrained semantic information, the semantic information that the merged image contains is still insufficient, and the semantic information in both the fused and source images cannot be well processed. Advanced computer vision tasks can benefit from cascaded image fusion in a mutually reinforcing manner.
Slight misalignment of the source images may cause serious artifacts in the fusion results. Wang et al. proposed the UMF-CMGR [10], which made a preliminary attempt to address this issue through multi-task mixing and aimed to unify the data into the infrared domain for image registration using cycleGAN [11]. UMF-CMGR alleviates the impact of small deformations in the image based on unimodal registration, but the fusion result will still be affected by artifacts when the source images are significantly misaligned. Although UMF has made preliminary attempts at multi-task mixing, combining image style transfer with image registration, it has only performed sequential training of image registration and image fusion, without cascading registration and without fusion for feedback-guided training between the networks. In the research work at the same stage as UMF-CMGR, SeAFusion is a cascaded fusion and segmentation network proposed by Tang et al. to calculate semantic loss to provide feedback to guide the training of the fusion network [12]. SeAFusion is a further attempt at multi-task hybrid training of fusion and other tasks, but it only implements segmentation by inputting the fusion result into a unimodal segmentation network, without considering inputting the multimodal source images and the fusion results into the multimodal segmentation network to implement segmentation. The SpuerFusion [13] proposed by Tang et al. extended SeAFusion by combining registration, fusion, and segmentation into one network. The network obtained by cascading registration and fusion was then input into the segmentation for network fine-tuning. The segmentation network used adopted a pre-trained segmentation network, but the pre-trained segmentation model was obtained through the unimodal segmentation of SeAFusion, and the segmentation results of the multimodal source images were still not added to the network training. Although UMF, SeAFusion, and SuperFusion have initially explored the need to consider high-level visual tasks with image fusion, they only improve the associated training of two tasks through loss functions. We should explore a more self-consistent approach to study the impact of semantic information in multimodal source images and fusion results on multimodal segmentation. Not only can a multi-stage fusion strategy be used to focus on the training of the fusion network [14,15], but a two-stage training strategy of reconstruction–fusion can also be used to focus on the reconstruction and fusion of the fusion network respectively [16,17,18]. A two-stage reconstruction–fusion training strategy is adopted in which the first stage focuses on the encoding reconstruction performance of the model’s encoder and decoder, and the second stage inserts the fusion part to train the feature fusion performance of the network. This paper adopts a two-stage training approach, training the fusion module, segmentation module, and segmentation network separately.
However, despite the significant progress made in image fusion, it still faces many severe challenges. It is impossible to overlook the problem of information loss during the fusion process. Due to the differences between multimodal images and the limitations of the fusion method, the fused results will lose some important information in the source image, thus impacting the fusion results’ visual quality. Some crucial detail information may be lost during the feature extraction and fusion process, leaving the fused image with insufficient semantic information and affecting the analysis and processing of subsequent high-level vision tasks. By cascading two different task modules in the first stage, we not only improve the quality of the fusion results but also pass the segmentation results to the multimodal segmentation model, which improves the model’s capabilities for extracting features from multimodal source images and for feature fusion. TSFS effectively integrates multimodal data, effectively combines image fusion and segmentation tasks, and achieves mutual optimization between the two tasks.
The main contributions of this paper are as follows:
1.
We propose a new two-stage image fusion and segmentation method called TSFS, which effectively combines image fusion and segmentation. In addition to considering the semantic information of the enhanced fused image, the semantic segmentation of the multimodal image is also guided by parameter transfer.
2.
We designed the FEM and FMFM in the fusion module. In order to improve the performance of the model in feature fusion, MFB using multiple convolutional layers and CAB using the channel attention mechanism were designed.
3.
We use the cross semantics fusion attention module (CSFAM) in the second stage to fuse features through cross fusion, thereby improving the segmentation accuracy.

2. Materials and Methods

We first briefly introduce the overall method of our method in Section 2, then describe the fusion module in the first stage S1 in detail in Section 2.1, then describe the semantic segmentation modules in the first stage S1 and the second stage S2 in Section 2.2, and, finally, describe the loss function in the entire method in Section 2.3.
Our main goal is to explore a new method based on two-stage image fusion and image segmentation. To this end, we propose several key components, including FEM, FMFM in the fusion module in the first stage S1, and CSFAM in the second stage S2. This design is designed to more effectively utilize multimodal features and enhance the quality and accuracy of fusion and semantic segmentation results. Figure 2 displays the TSFS network frame. The fusion result output by the fusion module in S1 is input into the segmentation module in S1, and then the parameters of the segmentation result of the fusion result are passed to the multimodal segmentation model. The following will describe in detail the fusion module in the first stage S1, the segmentation in the first and second stage, and the loss function.

2.1. The Fuse Module in the First Stage S1

TSFS consists of the S1 and S2. As shown in Figure 3, the first stage S1 consists of a fusion module and a segmentation module. In this section, we will explain the fusion module in the first stage S1, in detail.
Feature Extraction Module (FEM): We use the Residual in Residual Dense Block (RRDB) [19] to design the FEM in the fusion module in S1. We use RRDB to design the FEM, which can enhance the network architecture’s ability to extract features and utilize them more effectively. The infrared image and visible image are input into their respective FEM to obtain the deep feature information in the latent space and then input into FMFM for fusion. For visible images, we use the FEM we designed to extract features from their rich details and texture information to obtain deep features so that subsequent processing can better utilize this information. The FEM structure of the input infrared image is the same as that of the visible image. The depth features of the infrared image are extracted through the same feature extraction process. The infrared image and the visible image are input into the FEM through separate input branches. The separate input branch allows FEM to focus on extracting the unique feature characteristics of the input image. The FEM consists of six RRDBs, each containing three residual dense units, and the residual dense flow’s final output is the feature vector.
Feature Mix Fusion Module (FMFM): We concatenate the separately extracted deep features of the infrared image and the visible image and input them into the FMFM, which consists of the Mix Fuse Branch (MFB) and the Channel Attention Branch (CAB). FMFM is like a powerful information processor that can perform mixed fusion operations on the deep features of infrared images and visible images, thereby producing fusion in the deep feature space. We designed seven convolutional layers in MFB, with convolution kernel sizes of 3, 1, 3, 5, 7, 3, 1, and 3, respectively. The first convolution layer performs a preliminary fusion of deep features, which are then input into convolution layers with convolution kernels of 1, 3, 5, and 7 for multi-scale mixed fusion, thereby achieving more comprehensive feature fusion. GeLU activates the deep features that are aggregated and summed from the first four layers, and they are then coupled to the last convolutional layer for a 3 × 3 convolution process. The final output of CAB are activation vectors used to guide the fusion of deep spatial features. Subsequently, we multiply the input CAB features with the activation vector through channel multiplication to generate the fused features after CAB processing. Different convolution operations in MFB are not processed by pooling layers to decrease information loss in FMFM.
In summary, the processing flow of the fusion module in the S1 is as follows: First, two FEM are used to extract deep features from infrared images and visible images. Subsequently, the deep features are interactively fused through MFB and CAB in FMFM, respectively. Finally, two convolution operations are performed to obtain the fusion result. Image fusion not only utilizes the complementary characteristics between infrared images and visible images but also utilizes the unique characteristic information in single-modal images. In the TSFS method, feature extraction and interactive fusion are critical processes. FEM is a key component in the fusion module. FCFM can perform more effective fusion through feature mix fusion so that more semantic information can be passed from the fusion module to the segmentation in the first and second stages.

2.2. Segmentation in the S1 and S2

The semantic module in S1 plays a key role in connecting the previous and the next stages. Since the parameters of the segmentation module in S1 are transferred to S2, it not only completes the initial segmentation but also comprehensively guides the semantic segmentation of multimodal images in S2.
Therefore, the cascaded optimization of image fusion and semantic segmentation is of great significance. By cascading these two tasks into a unified method, the mutual reinforcement effect between them can be exploited to its potential. Therefore, we propose the TSFS method, which aims to cascade image fusion and segmentation, breaking through the limitations of existing single-task models and providing new insights and methods for the development of computer vision. The structure of the segmentation in the S2 is shown in Figure 4. This method effectively extracts, fuses, and segments multimodal information by designing a two-stage training method: FEM, FMFM, and CSFAM. The fusion module and segmentation module in S1 and the segmentation module in S2 are three important parts of TSFS. Through this design, TSFS effectively combines image fusion and semantic segmentation tasks, achieving cascade optimization of the two.
Cross Semantics Fusion Attention Module (CSFAM): As the key module in the second stage S2 of TSFS, CSFAM is important for the cross-modal semantic features fusion. From the perspective of intra-modality enhancement and inter-modality complementarity, the self-cross-attention module first enhances local visual features within the modality. By emphasizing the refinement of local features, the importance of local features in the fusion process is increased, and effective feature areas can be focused and processed more efficiently. In CSFAM, the calculation process of CSFAM is processed in the order of unimodal self-attention, cross-modal cross-attention, cross-modal self-attention, and further cross-modal cross-attention, and the effective fusion of semantic information is achieved through cross fusion. Unimodal self-attention captures the image’s unimodal features. Unimodal self-attention can be expressed as follows:
A I X = V X S K X , Q X
where S ( · ) represents the similarity operation. Unimodal self-attention, cross-modal cross-attention, cross-modal self-attention, and further cross-modal cross-attention are all calculated by S ( · ) . Here, we take unimodal self-attention as an example to illustrate the generation of Value, Key, and Query [20] from the feature X of the visible image, denoted as V, K, and Q, respectively, and the feature of the infrared image is denoted as Y. S ( · ) is expressed as follows:
S K , Q = s o f t m a x K X T Q X d n
A I X enhances the aggregation of the same classification within the modality by performing self-attention calculation on the deep features of the image itself, making the unimodal semantic information of the image obtained by the model more stable and accurate. Cross-modal cross-attention enhances semantic information by integrating fine-grained information extracted from cross-modal attention. A I X Y is obtained through cross-modal cross-attention. Cross-modal self-attention further strengthens the cross-modal features captured by cross-modal cross-attention. Finally, further input into further cross-modal cross-attention to obtain the fused features.

2.3. Loss Function

In the first-stage loss function L S 1 of TSFS, we adopted pixel intensity distribution loss L p i x e l , comparative loss L c r , structural similarity loss L s s i m , and conventional cross-entropy loss L c e . The loss function in this paper includes the loss function L S 1 of the first stage S1 and the loss function L S 2 of the second stage S2. The loss function is expressed as follows:
L S 1 = L f u s + L s e g S 1 = L p i x e l + L c r + L s s i m + L c e
L S 2 = L c e
L t o t l e = L S 1 + L S 2
Infrared images and visible images are paired multimodal images with complementary characteristics, but neither the FMB dataset nor the MFNet dataset provides officially released fusion images as a reference standard. So to address this problem, we need to consider a key issue in L c r in image fusion methods: how to generate high-quality positive samples and paired matching negative samples. We manually generated masks for the FMB dataset and MFNet dataset. We use the intermediate weight maps a 1 = S ( I v ) S ( I r ) and a 2 = S ( I r ) S ( I v ) to calculate the visual saliency maps of visible images and infrared images, denoted as m v and m r , respectively. m v and m r are complementary, and m v + m r = 1 . The expression of m v is as follows:
m v = e a 1 e a 1 + e a 2
We adopted L c r in the loss function L S 1 of the first stage S1 and used the pre-trained model G (VGG-19 [21]) with fixed weights commonly used in the field of image processing, aiming to make the fusion result as close to the positive sample as possible while staying away from the corresponding negative sample. The definition of L c r is as follows:
L c = i = 1 n w i G i I f G i I v · m v + I r · m r G i I f G i I v · m r + I r · m v
The four loss functions in the first stage, S1, constrain the model from different perspectives and together guide the fusion results generated by TSFS to include more semantic information. This improves the quality of the fusion results and provides strong support for segmentation that is mutually reinforcing with image fusion.

3. Experiments

Our method is implemented in PyTorch 1.11, and both training and testing are performed on a machine with an NVIDIA GeForce RTX 4090 (NVIDIA, Santa Clara, CA, USA) with 24 GB of video memory and a Xeon Platinum 8352V processor (Intel, Santa Clara, CA, USA) with 50 GB of memory. During the training process, the initial learning rate of the fusion module is set to 1 × 10 3 , and the initial learning rate of the segmentation module and the segmentation network in the second stage is set to 1 × 10 2 . We use the Adam optimizer and stochastic gradient descent to optimize the segmentation module and the segmentation network in the S2.
All the datasets we used in this study are divided into training and test sets, where the FMB dataset [22] and MFNet [23] dataset contain 1220/1176 pairs and 280/393 pairs of infrared and visible images in the training and test sets, respectively. In comparative experiments, we present the evaluation results of the proposed TSFS on the FMB and MFNet datasets, conducting extensive qualitative and quantitative experiments against our method and seven state-of-the-art methods. These seven methods were tested according to the public codes and parameter settings of the original papers, namely MDLatLRR [24], DenseFuse [25], LTS [26], MetaLearning [27], SeAFusion, TarDAL [28], Diff-if [29], EGFNet [30], and LASNet [31].
For the fusion results in the first stage, we selected the following five objective evaluation indicators to evaluate, namely spatial frequency (SF), mutual information (MI), visual information fidelity (VIF) [32], nonlinear correlation information entropy (NCIE), and normalized mutual information ( Q M I ) [33]. For the evaluation of segmentation, we chose the main metric used by most methods for evaluating semantic understanding: mean Intersection over Union (mIoU) [12].

3.1. Image Fusion Results Analysis

3.1.1. Subjective Analysis

We chose two daylight and two nighttime photos—images 00085 and 00381 from the FMB dataset and images 01314 and 01538 from the MFNet dataset—to perform the subjective analysis of the experimental findings.
For the pair of night images in the FMB dataset, the infrared targets are obvious in LTS, MetaLearning, TarDAL, Diff, and TSFS, among which the infrared target of TSFS is the most obvious, with high contrast. The ones that are difficult to distinguish from daytime due to the dark sky are MDLatLRR, DenseFuse, LTS, MetaLearning, and TarDAL. From the results of the seven comparison algorithms and the fusion results of TSFS, it can be clearly seen that, in a daytime environment with good lighting, TSFS effectively integrates the information of this area, especially the area marked by the blue box. The infrared target person in TSFS not only integrates the information of the infrared target person area in the infrared image but also integrates some texture detail information in the visible image, as shown in the yellow box of the fusion result of nighttime infrared and visible images in Figure 5. Furthermore, in the fusion results of the infrared and visible images during the daytime in Figure 5, the infrared target person in the yellow frame is enhanced in both visual quality and clarity by incorporating the texture details in the visible image. In night vision conditions, objects in visible images are dim and have low recognition. In infrared images, not only are infrared target people highly recognizable, but thermal radiation information of objects, including cars, is also captured. This phenomenon can be seen in the two pairs of infrared and visible images in the dark environment in Figure 5 and Figure 6. In the images in the dark environment of Figure 6, it can be seen from the blue box that although the car in the visible image is dim and has low recognition, the thermal radiation information in the infrared image can appropriately compensate for this. TSFS improves the recognition of the car in the fused image.
The fusion of infrared images and visible images requires the effective fusion of complementary information. Due to the imaging characteristics of infrared sensors and visible light sensors in different environments, combined with Figure 5 and Figure 6, it is possible to distinguish between night and day. In the dark, the visibility of objects in visible images is significantly reduced, while the recognition of objects in infrared images is significantly improved. Effective image fusion can achieve an overall improvement in the recognition of background information and objects. In the daytime, due to the single-channel imaging characteristics of infrared images, if the effective feature information cannot be correctly extracted in the feature information extraction stage, the clarity of the clear background information obtained from the visible image will be reduced. In summary, in the task of image fusion, we not only well fuse the salient infrared target information from the infrared image but also well fuse the detailed texture information from the visible image, obtaining fusion results that are friendly to human visual perception.

3.1.2. Objective Analysis

The results in Table 1 show that our results achieved consistent advantages in all five evaluation indicators except for MI and Q M I , which achieved the second-best performance. From the perspective of judging the degree of information transfer, the highest NCIE, and the second-best MI, Q M I show that there is significant transfer of source images. In addition, based on the statistical characteristics of the fused image itself, the best SF indicates that our approach has higher image contrast, a clearer image, and richer image details. In terms of image quality relative to natural scenes, the highest VIF shows that our approach is better with the visual system of humans, which means it is more friendly to the human eye. In summary, the fusion image has richer detailed textures and transfers more information from the source image, which proves that TSFS is effective and has accomplished the target task of image fusion well, especially in harsh environments such as night.

3.2. Segmentation Results Analysis

3.2.1. Subjective Analysis

To confirm the efficacy of TSFS in the segmentation experiment, we also selected a pair night image and a pair day image, which are the images named 00634 and 00359 in the FMB dataset and the images named 01342 and 01477 in the MFNet dataset. Light in visible images has a halo, which affects the quality of the segmentation results. As shown in Figure 7, the segmentation results of the seven methods except TSFS are all affected by this factor, with MetaLearning being the most affected. Compared with the MFNet dataset, the scene annotations in the FMB dataset are more detailed. At the same time, since MFNet has more unlabeled information than FMB, people and cars occupy a large proportion of the MFNet annotations. As can be seen from Figure 8, although DenseFuse, LTS, MetaLearning, SeAFusion, and TarDAL all achieved relatively good segmentation results for the car and infrared target person, none of the curves on the left achieved correct segmentation results. Although EGFNet and LAS obtain the correct segmentation results of the roadside on the left, they obtain redundant, incorrect segmentation results compared to the right area of the infrared target person in Ground Truth. Comparing the segmentation results in Figure 8 with other methods, especially compared with EGFNet and LAS, the semantic segmentation results of TSFS show a more comprehensive target outline, and the posture of the infrared target person is also more vivid.

3.2.2. Objective Analysis

The results in Table 2 show that our results have achieved consistent advantages in the car and building categories, and achieved the second best ranking in the person category. Comparative analysis of segmentation results on the FMB dataset shows that TSFS improves mIoU by 3.3%. The results in Table 3 show that our results have achieved consistent advantages in the five categories except people and cones, and ranked second in the people category. For the MFNet dataset, TSFS outperforms all other methods, with a 1.5% improvement in mIoU over LAS. Cars and people occupy a large proportion in both the FMB dataset and the MFNet dataset. TSFS has shown relatively outstanding performance in categories such as cars and people. Compared with the categories of segmentation results of the FMB dataset, the improvement effect of different instance category on the MFNet dataset is more obvious. In summary, TSFS demonstrates superior segmentation performance, not only obtaining the optimal mIoU but also presenting a comprehensive target outline.

3.3. Ablation Studies

To assess the necessity for the two stages in our method TSFS, we performed ablation experiments on the MFNet dataset and examined six variants. In experiment 1, we replaced MFB with concatenate. In experiment 2, we removed the negative samples in L c r . In experiment 3, we removed L c r . In experiment 4, we retained the S1 and removed the S2. In experiment 5, we removed the S2 and retained the S1. In experiment 6, we replaced CSFAM with local attention.
Figure 9 and Figure 10 respectively show the effects of different designs in the two stages on image fusion and semantic segmentation. Removing any module will lead to a performance decline. The results in Table 4 show that, except for the MI indicator, which achieved the second best ranking, the other four indicators, MI, NCIE, QY, and Q M I , all achieved significant and consistent advantages. The experimental data indicate that, if L c r is missing, the image contrast and clarity will decrease, and the mIoU will also drop. If we eliminate the negative samples based on contrast loss, we will lose some of the image’s textures and details, as shown in Figure 9. The human behind the yellow warning sign in the yellow frame in Figure 9 can also be easily identified in TSFS. If the second stage S2 is missing, that is, only the first stage S1 is present, the mIoU drops by 1.3%. If the first stage S1 is missing, that is, there is only S2, the mIoU drops by 1.7%. A comparative analysis of Experiments 4, 5, and 6 with TSFS shows that the target person’s posture is more vivid and the edge contours are clearer. Only our method can obtain relatively accurate segmentation results for the target person. These results all emphasize the positive impact of different designs in these two stages on the fusion results and segmentation results.

4. Discussion

We designed a new two-stage image fusion and segmentation method in TSFS using FEM, FMFM, and CSFAM. We fuse multimodal information with a two-stage approach to enhance the accuracy of segmentation and the quality of fused results. In the fusion module, we design FEM and FMFM to extract multimodal image information and fuse features. The proposed FMFM can fuse features at multiple levels through MFB and CAB. Compared with existing methods, TSFS shows significant advantages in all indicators. According to experimental data, TSFS is effective in image fusion and segmentation since it considerably enhances the fused image’s visual effect, evaluation index values, and segmentation mIoU. Although the experimental results are excellent, we only conducted comparative experiments on FMB and MFNet. Further verification can be done on other infrared and visible image datasets. In the following studies, we want to enhance the generalization ability of the model and further consider its practical application significance. We will continue to improve the model’s generalization and robustness across other datasets.

5. Conclusions

To explore the intrinsic connection between the image fusion problem and segmentation in a more self-consistent way, we propose a new method called TSFS. The semantic information in the fused image is enhanced by cascading the fusion module and the segmentation module. The segmentation results of the fusion result are passed to the multimodal segmentation model in the form of pre-trained model parameters, reducing the problem of semantic information loss caused by the lack of multimodal source image information. We designed the FEM and FMFM in the fusion module and also input the fusion results output by the fusion module in S1 into the segmentation module to obtain semantic loss, guiding the fusion image generated by the fusion module in S1 to have more semantic information. We use CSFAM in the second stage S2 to fuse features in a cross-fusion manner in different directions to improve the segmentation accuracy. Multiple comparisons and ablation studies on the FMB dataset and MFNet dataset show that our proposed TSFS not only improves the visual effect of the fused image but also improves the segmentation mIoU. Future research will continue to investigate joint training of image fusion with other high-level computer tasks and improve the generalization ability of the model.

Author Contributions

Conceptualization, W.R.; methodology, W.R.; software, W.R.; validation, W.R. and L.L.; formal analysis, W.R.; data curation, W.R.; writing—original draft preparation, W.R.; writing—review and editing, W.R., J.R. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (62262016, 62302132), 14th Five-Year Plan Civil Aerospace Technology Preliminary Research Project (D040405).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  1. Han, J.; Bhanu, B. Fusion of color and infrared video for moving human detection. Pattern Recognit. 2007, 40, 1771–1784. [Google Scholar] [CrossRef]
  2. Sun, Y.; Yan, K.; Li, W. CycleGAN-based SAR-optical image fusion for target recognition. Remote Sens. 2023, 15, 5569. [Google Scholar] [CrossRef]
  3. Chandrakanth, V.; Murthy, V.; Channappayya, S. Siamese cross-domain tracker design for seamless tracking of targets in RGB and thermal videos. IEEE Trans. Artif. Intell. 2022, 4, 161–172. [Google Scholar]
  4. Maqsood, S.; Javed, U.; Riaz, M.; Muzammil, M.; Muhammad, F.; Kim, S. Multiscale image matting based multi-focus image fusion technique. Electronics 2020, 9, 472. [Google Scholar] [CrossRef]
  5. Li, X.; Li, Y.; Chen, H.; Peng, Y.; Chen, L.; Wang, M. RITFusion: Reinforced interactive transformer network for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 2023, 73, 1–16. [Google Scholar] [CrossRef]
  6. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion. 2017, 36, 191–207. [Google Scholar] [CrossRef]
  7. Yang, Y.; Liu, J.; Huang, S.; Wan, W.; Wen, W.; Guan, J. Infrared and visible image fusion via texture conditional generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4771–4783. [Google Scholar] [CrossRef]
  8. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  9. Li, H.; Wu, X.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  10. Wang, D.; Liu, J.; Fan, X.; Liu, R. Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv 2022, arXiv:2205.11876. [Google Scholar] [CrossRef]
  11. Zhu, J.; Park, T.; Isola, P.; Efros, A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  12. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion. 2022, 82, 28–42. [Google Scholar] [CrossRef]
  13. Tang, L.; Deng, Y.; Ma, Y.; Huang, J.; Ma, J. SuperFusion: A versatile image registration and fusion network with semantic awareness. IEEE/CAA J. Autom. Sin. 2022, 9, 2121–2137. [Google Scholar] [CrossRef]
  14. Chang, K.; Huang, J.; Sun, X.; Luo, J.; Bao, S.; Huang, H. Infrared and visible image fusion network based on multistage progressive injection. Complex Intell. Syst. 2025, 11, 367. [Google Scholar] [CrossRef]
  15. Wang, C.; Wu, J.; Zhu, Z.; Chen, H. MSFNet: MultiStage Fusion Network for infrared and visible image fusion. Neurocomputing 2022, 507, 26–39. [Google Scholar] [CrossRef]
  16. Cao, Y.; Luo, X.; Tong, X.; Yang, J.; Cao, Y. Infrared and visible image fusion based on a two-stage class conditioned auto-encoder network. Neurocomputing 2023, 544, 126248. [Google Scholar] [CrossRef]
  17. Zheng, X.; Yang, Q.; Si, P.; Wu, Q. A Multi-Stage Visible and Infrared Image Fusion Network Based on Attention Mechanism. Sensors 2022, 22, 3651. [Google Scholar] [CrossRef]
  18. Huang, S.; Kong, X.; Yang, Y.; Wan, W.; Song, Z. FTSFN: A Two-Stage Feature Transfer and Supplement Fusion Network for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2025, 74, 1–15. [Google Scholar] [CrossRef]
  19. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  20. Wang, J.; Mao, Y.; Ma, X.; Guo, S.; Shao, Y.; Lv, X.; Han, W.; Christopher, M.; Zangwill, L.M.; Bi, Y. ODFormer: Semantic Fundus Image Segmentation Using Transformer for Optic Nerve Head Detection. Inf. Fusion. 2024, 112, 102533. [Google Scholar] [CrossRef]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Liu, J.; Liu, Z.; Wu, G.; Ma, L.; Liu, R.; Zhong, W.; Luo, Z.; Fan, X. Multi-interactive feature learning and a full-time multi-modality benchmark for image fusion and segmentation. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–8 October 2023; pp. 8115–8124. [Google Scholar]
  23. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar]
  24. Li, H.; Wu, X.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef] [PubMed]
  25. Li, H.; Wu, X. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef]
  26. Liu, J.; Wu, Y.; Wu, G.; Liu, R.; Fan, X. Learn to search a lightweight architecture for target-aware infrared and visible image fusion. IEEE Signal Process. Lett. 2022, 29, 1614–1618. [Google Scholar] [CrossRef]
  27. Li, H.; Cen, Y.; Liu, Y.; Chen, X.; Yu, Z. Different input resolutions and arbitrary output resolution: A meta learning-based deep framework for infrared and visible image fusion. IEEE Trans. Image Process. 2021, 30, 4070–4083. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 5802–5811. [Google Scholar]
  29. Li, H.; Cen, Y.; Liu, Y.; Chen, X.; Yu, Z. Diff-IF: Multi-modality image fusion via diffusion model with fusion knowledge prior. Inf. Fusion. 2024, 110, 102450. [Google Scholar]
  30. Zhou, W.; Dong, S.; Xu, C.; Qian, Y. Edge-aware guidance fusion network for rgb–thermal scene parsing. In Proceedings of the 36th AAAI Conference on Artificial Intelligence (AAAI), Virtual, 22–28 February 2022; pp. 3571–3579. [Google Scholar]
  31. Li, G.; Wang, Y.; Liu, Z.; Zhang, X.; Zeng, D. RGB-T semantic segmentation with location, activation, and sharpening. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1223–1235. [Google Scholar] [CrossRef]
  32. Han, Y.; Cai, Y.; Cao, Y.; Xu, X. A new image fusion performance metric based on visual information fidelity. Inf. Fusion. 2013, 14, 127–135. [Google Scholar] [CrossRef]
  33. Liu, Z.; Blasch, E.; Xue, Z.; Zhao, J.; Laganiere, R.; Wu, W. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 94–109. [Google Scholar] [CrossRef]
Figure 1. Examples of visible images, infrared images, and fusion results.
Figure 1. Examples of visible images, infrared images, and fusion results.
Applsci 15 10698 g001
Figure 2. The network frame of TSFS.
Figure 2. The network frame of TSFS.
Applsci 15 10698 g002
Figure 3. The network frame of S1.
Figure 3. The network frame of S1.
Applsci 15 10698 g003
Figure 4. The network frame of S2.
Figure 4. The network frame of S2.
Applsci 15 10698 g004
Figure 5. Fusion results on the FMB dataset.
Figure 5. Fusion results on the FMB dataset.
Applsci 15 10698 g005
Figure 6. Fusion results on the MFNet dataset.
Figure 6. Fusion results on the MFNet dataset.
Applsci 15 10698 g006
Figure 7. Segmentation results on the FMB dataset.
Figure 7. Segmentation results on the FMB dataset.
Applsci 15 10698 g007
Figure 8. Segmentation results on the MFNet dataset.
Figure 8. Segmentation results on the MFNet dataset.
Applsci 15 10698 g008
Figure 9. Fusion results of ablation studies on the MFNet dataset. (a) Visible image. (b) Infrared image. (c) TSFS. (d) Replaced MFB with concatenate. (e) w/o Negative sample. (f) w/o L c r . (g) Only S1.
Figure 9. Fusion results of ablation studies on the MFNet dataset. (a) Visible image. (b) Infrared image. (c) TSFS. (d) Replaced MFB with concatenate. (e) w/o Negative sample. (f) w/o L c r . (g) Only S1.
Applsci 15 10698 g009
Figure 10. Segmentation results of ablation studies on the MFNet dataset. (a) Visible image. (b) Infrared image. (c) Ground Truth (d) Replaced MFB with concatenate. (e) w/o Negative sample. (f) w/o L c r . (g) Only S1. (h) Only S2. (i) Replaced CSFAM with Local Attention. (j) TSFS.
Figure 10. Segmentation results of ablation studies on the MFNet dataset. (a) Visible image. (b) Infrared image. (c) Ground Truth (d) Replaced MFB with concatenate. (e) w/o Negative sample. (f) w/o L c r . (g) Only S1. (h) Only S2. (i) Replaced CSFAM with Local Attention. (j) TSFS.
Applsci 15 10698 g010
Table 1. The FMB and MFNet datasets’ objective evaluation parameter outcomes of the fusion results.
Table 1. The FMB and MFNet datasets’ objective evaluation parameter outcomes of the fusion results.
DatasetsMethodsSFMIVIFNCIE Q MI
FMBSeAFusion0.0562.8550.8730.8070.425
DenseFuse0.0333.1750.7790.8080.472
LTS0.0493.2870.9850.8080.468
MetaLearning0.0474.1110.8540.8120.604
SeAFusion0.0533.8330.9530.8100.569
TarDAL0.0463.3840.8500.8090.492
Diff-If0.0544.5090.8820.8150.679
TSFS0.0664.4430.9970.8150.647
MFNetMDLatLRR0.0572.1680.8590.8050.336
DenseFuse0.0292.2770.7610.8050.368
LTS0.0462.6170.8870.8060.411
MetaLearning0.0463.1990.7630.8090.522
SeAFusion0.0522.5590.9020.8060.400
TarDAL0.0392.4390.770.8060.387
Diff-If0.0513.6790.9260.8110.597
TSFS0.0583.4960.9890.8120.535
Table 2. The FMB datasets’ objective evaluation parameter outcomes of the segmentation results.
Table 2. The FMB datasets’ objective evaluation parameter outcomes of the segmentation results.
MethodsCarPersonT-LampT-SignBuildingVegetablePolemIoU
DenseFuse72.854.240.371.078.683.640.652.2
LTS72.558.540.969.776.382.840.551.7
MetaLearning70.050.234.862.876.675.836.448.1
SeAFusion74.059.340.4 73.481.284.843.354.3
TarDAL72.060.039.959.876.982.734.149.3
Diff-If72.058.338.371.978.783.341.751.2
TSFS81.159.339.170.083.983.230.357.6
Table 3. The MFNet dataset’s objective evaluation parameter outcomes of the segmentation results.
Table 3. The MFNet dataset’s objective evaluation parameter outcomes of the segmentation results.
MethodsCarPersonBikeCurveCar StopConeCumpmIoU
EGFNet87.669.858.842.833.87.047.154.8
LASNet84.267.156.941.139.618.940.154.9
DenseFuse83.469.960.829.528.451.745.951.9
LTS83.468.759.827.524.750.644.750.8
MetaLearning77.934.556.021.123.545.938.343.9
SeAFusion85.571.360.035.429.951.146.953.1
TarDAL82.169.256.534.127.349.842.850.5
Diff-If85.472.2 59.632.425.848.948.652.3
TSFS88.071.861.244.345.046.353.156.4
Table 4. We have obtained the objective evaluation parameter outcomes of the ablation studies from the MFNet dataset.
Table 4. We have obtained the objective evaluation parameter outcomes of the ablation studies from the MFNet dataset.
SFMIVIFNCIE Q M I mIoU
Concatenate0.0483.3830.9460.8090.51455.2
w/o Neg0.0513.499 0.9070.8110.52855.2
w/o L c r 0.0292.5490.8140.8060.41255.6
Only S10.0583.4960.9890.8120.53555.1
Only S2 54.7
Exp 60.0583.4960.9890.8120.53552.3
TSFS0.0583.4960.9890.8120.53556.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, W.; Luo, L.; Ren, J. A Two-Stage Approach for Infrared and Visible Image Fusion and Segmentation. Appl. Sci. 2025, 15, 10698. https://doi.org/10.3390/app151910698

AMA Style

Ren W, Luo L, Ren J. A Two-Stage Approach for Infrared and Visible Image Fusion and Segmentation. Applied Sciences. 2025; 15(19):10698. https://doi.org/10.3390/app151910698

Chicago/Turabian Style

Ren, Wang, Lanhua Luo, and Jia Ren. 2025. "A Two-Stage Approach for Infrared and Visible Image Fusion and Segmentation" Applied Sciences 15, no. 19: 10698. https://doi.org/10.3390/app151910698

APA Style

Ren, W., Luo, L., & Ren, J. (2025). A Two-Stage Approach for Infrared and Visible Image Fusion and Segmentation. Applied Sciences, 15(19), 10698. https://doi.org/10.3390/app151910698

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop