Author Contributions
Conceptualization , R.C. and S.Z.; Methodology, J.D., S.H., L.X. and C.Z.; Software, J.D. and C.Z.; Validation, N.Z.; Formal analysis, J.D., S.H. and L.X.; Investigation, R.C., N.Z., H.B., C.Z. and Y.Y.; Resources, L.X.; Data curation, R.C., S.H., H.B. and Y.Y.; Writing—original draft, N.Z., S.Z., S.H. and C.Z.; Writing—review & editing, S.Z.; Supervision, S.Z. and H.B.; Funding acquisition, R.C., J.D., N.Z. and Y.Y. All authors have read and agreed to the published version of the manuscript.
Figure 1.
Structure of two-branch-feature-complementarity-based fusion method. All source images are transferred to respective branches for feature extraction and feature fusion in generator. Then, fusion results are obtained through feature reconstruction. Finally, fusion images, visible images and infrared images are used as the input of discriminator respectively, and identification results are guided by loss functions to optimize fusion network.
Figure 1.
Structure of two-branch-feature-complementarity-based fusion method. All source images are transferred to respective branches for feature extraction and feature fusion in generator. Then, fusion results are obtained through feature reconstruction. Finally, fusion images, visible images and infrared images are used as the input of discriminator respectively, and identification results are guided by loss functions to optimize fusion network.
Figure 2.
The generator structure based on two-branch feature complementarity. Visible images are fed into one branch, and infrared images are fed into another branch in the proposed generator. Feature complementation is carried out by pixel superposition in the process of feature extraction of different branches. Different features obtained from both branches are fused and reconstructed to obtain final fused images.
Figure 2.
The generator structure based on two-branch feature complementarity. Visible images are fed into one branch, and infrared images are fed into another branch in the proposed generator. Feature complementation is carried out by pixel superposition in the process of feature extraction of different branches. Different features obtained from both branches are fused and reconstructed to obtain final fused images.
Figure 3.
Structure of double-classification discriminator based on layer-hopping connections. Both source and fused images are fed to discriminator containing two skip connection layers, respectively, to calculate the discriminant results of all images.
Figure 3.
Structure of double-classification discriminator based on layer-hopping connections. Both source and fused images are fed to discriminator containing two skip connection layers, respectively, to calculate the discriminant results of all images.
Figure 4.
The Fused results of the proposed fusion method and twelve other comparison methods used on the M3FD dataset. Three different groups of contrast images are listed as a, b, c. Distinct features are marked with GREEN boxes , and enlarged features are in the corner of the images.
Figure 4.
The Fused results of the proposed fusion method and twelve other comparison methods used on the M3FD dataset. Three different groups of contrast images are listed as a, b, c. Distinct features are marked with GREEN boxes , and enlarged features are in the corner of the images.
Figure 5.
The fused results of the proposed fusion method and twelve other comparison methods used on the MFNet dataset. Distinct features are marked with GREEN boxes, and enlarged features are in the lower left corner of images.
Figure 5.
The fused results of the proposed fusion method and twelve other comparison methods used on the MFNet dataset. Distinct features are marked with GREEN boxes, and enlarged features are in the lower left corner of images.
Figure 6.
The fused results of the proposed fusion method and twelve other comparison methods used on TNO dataset. Distinct features are marked with GREEN boxes.
Figure 6.
The fused results of the proposed fusion method and twelve other comparison methods used on TNO dataset. Distinct features are marked with GREEN boxes.
Figure 7.
The fused results of the proposed fusion method and twelve other comparison methods used on RoadScene dataset. Distinct features are marked with GREEN boxes, and enlarged features are in upper corner of images.
Figure 7.
The fused results of the proposed fusion method and twelve other comparison methods used on RoadScene dataset. Distinct features are marked with GREEN boxes, and enlarged features are in upper corner of images.
Figure 8.
Visual comparison of different fusion results based on the MFNet dataset in target detection task.
Figure 8.
Visual comparison of different fusion results based on the MFNet dataset in target detection task.
Figure 9.
The visualization results of different modules in ablation experiment conducted on the M3FD dataset. The fusion model with double-classification discriminator added is Module 1, the fusion model with double-classification discriminator and two-branch generator is Module 2, and Module 3 is the fusion model with information interactions added in two-branch generator. Areas with large differences are highlighted by GREEN boxes.
Figure 9.
The visualization results of different modules in ablation experiment conducted on the M3FD dataset. The fusion model with double-classification discriminator added is Module 1, the fusion model with double-classification discriminator and two-branch generator is Module 2, and Module 3 is the fusion model with information interactions added in two-branch generator. Areas with large differences are highlighted by GREEN boxes.
Table 1.
The average values of the proposed and twelve other comparison methods for six evaluation indicators evaluated using the M3FD dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Table 1.
The average values of the proposed and twelve other comparison methods for six evaluation indicators evaluated using the M3FD dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Methods Name | EN↑ | SF↑ | PSNR↑ | VIF↑ | AG↑ | MSE↓ |
---|
GTF | 7.334 | 0.059 | 62.886 | 0.836 | 6.505 | 0.039 |
MDLatLRR | 6.362 | 0.031 | 62.183 | 0.658 | 2.560 | 0.041 |
FusionGAN | 6.737 | 0.036 | 63.810 | 0.589 | 4.155 | 0.031 |
GANMcC | 6.860 | 0.034 | 63.861 | 0.809 | 3.924 | 0.030 |
IFCNN | 6.794 | 0.061 | 65.707 | 0.792 | 6.683 | 0.020 |
RFN | 7.343 | 0.021 | 61.612 | 1.090 | 2.244 | 0.052 |
U2Fusion | 7.051 | 0.065 | 64.795 | 0.978 | 7.923 | 0.024 |
DDcGAN | 7.473 | 0.043 | 57.656 | 0.087 | 4.049 | 0.112 |
DIDFuse | 7.066 | 0.047 | 61.132 | 0.789 | 3.917 | 0.051 |
MFEIF | 6.526 | 0.029 | 61.505 | 0.688 | 2.561 | 0.048 |
SeAFusion | 6.649 | 0.048 | 59.165 | 0.765 | 3.896 | 0.083 |
TarDAL | 7.028 | 0.045 | 60.823 | 0.734 | 3.550 | 0.057 |
OURS | 7.170 | 0.072 | 64.553 | 0.993 | 8.878 | 0.025 |
Table 2.
The average values of the proposed and twelve other comparison methods for six evaluation indicators evaluated using the MFNet dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Table 2.
The average values of the proposed and twelve other comparison methods for six evaluation indicators evaluated using the MFNet dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Methods Name | EN↑ | SF↑ | PSNR↑ | VIF↑ | AG↑ | MSE↓ |
---|
GTF | 6.053 | 0.033 | 63.190 | 0.708 | 2.569 | 0.034 |
MDLatLRR | 6.262 | 0.028 | 65.253 | 0.793 | 2.389 | 0.021 |
FusionGAN | 5.879 | 0.019 | 63.703 | 0.724 | 1.652 | 0.030 |
GANMcC | 6.463 | 0.023 | 64.049 | 0.794 | 2.055 | 0.031 |
IFCNN | 6.423 | 0.037 | 65.945 | 0.786 | 3.119 | 0.018 |
RFN | 6.019 | 0.014 | 62.590 | 0.848 | 1.315 | 0.040 |
U2Fusion | 6.698 | 0.044 | 63.624 | 0.664 | 3.898 | 0.031 |
DDcGAN | 7.269 | 0.042 | 57.866 | 0.100 | 4.038 | 0.108 |
DIDFuse | 5.915 | 0.040 | 62.193 | 0.680 | 2.886 | 0.039 |
MFEIF | 6.373 | 0.026 | 64.732 | 0.860 | 2.248 | 0.025 |
SeAFusion | 6.693 | 0.045 | 63.598 | 0.920 | 3.758 | 0.031 |
TarDAL | 6.615 | 0.038 | 62.961 | 0.830 | 3.004 | 0.035 |
OURS | 6.943 | 0.048 | 64.459 | 0.940 | 4.936 | 0.024 |
Table 3.
The average values of the proposed method and twelve other comparison methods for six evaluation indicators evaluated using the RoadScene dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Table 3.
The average values of the proposed method and twelve other comparison methods for six evaluation indicators evaluated using the RoadScene dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Methods Name | EN↑ | SF↑ | PSNR↑ | VIF↑ | AG↑ | MSE↓ |
---|
GTF | 7.654 | 0.035 | 59.796 | 0.718 | 3.728 | 0.074 |
MDLatLRR | 6.752 | 0.041 | 64.255 | 0.738 | 4.183 | 0.035 |
FusionGAN | 7.131 | 0.035 | 59.796 | 0.609 | 3.729 | 0.074 |
GANMcC | 7.299 | 0.038 | 60.424 | 0.754 | 4.242 | 0.069 |
IFCNN | 7.043 | 0.062 | 64.195 | 0.769 | 6.054 | 0.035 |
RFN | 7.316 | 0.036 | 60.757 | 0.787 | 4.172 | 0.067 |
U2Fusion | 7.095 | 0.069 | 62.720 | 0.754 | 7.172 | 0.043 |
DDcGAN | 7.628 | 0.049 | 56.243 | 0.078 | 5.313 | 0.161 |
DIDFuse | 7.250 | 0.063 | 62.010 | 0.805 | 6.540 | 0.049 |
MFEIF | 6.882 | 0.041 | 63.109 | 0.789 | 4.236 | 0.044 |
SeAFusion | 7.288 | 0.077 | 61.357 | 0.855 | 7.991 | 0.063 |
TarDAL | 7.225 | 0.055 | 62.187 | 0.761 | 5.195 | 0.050 |
OURS | 7.414 | 0.095 | 62.304 | 0.671 | 10.592 | 0.046 |
Table 4.
The average values of the proposed method and twelve other comparison methods for six evaluation indicators evaluated using the TNO dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Table 4.
The average values of the proposed method and twelve other comparison methods for six evaluation indicators evaluated using the TNO dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Methods Name | EN↑ | SF↑ | PSNR↑ | VIF↑ | AG↑ | MSE↓ |
---|
GTF | 6.863 | 0.041 | 62.043 | 0.682 | 3.892 | 0.042 |
MDLatLRR | 6.298 | 0.028 | 63.333 | 0.731 | 2.901 | 0.033 |
FusionGAN | 6.651 | 0.026 | 61.058 | 0.712 | 2.587 | 0.054 |
GANMcC | 6.803 | 0.026 | 61.869 | 0.784 | 2.858 | 0.049 |
IFCNN | 6.590 | 0.047 | 63.774 | 0.736 | 4.551 | 0.022 |
RFN | 7.020 | 0.025 | 62.362 | 0.903 | 3.016 | 0.038 |
U2Fusion | 7.060 | 0.051 | 63.072 | 0.913 | 5.523 | 0.030 |
DDcGAN | 7.490 | 0.043 | 57.090 | 0.071 | 5.063 | 0.122 |
DIDFuse | 6.922 | 0.042 | 61.445 | 0.811 | 4.253 | 0.048 |
MFEIF | 6.578 | 0.027 | 62.577 | 0.817 | 3.012 | 0.040 |
SeAFusion | 7.196 | 0.054 | 62.193 | 1.063 | 5.716 | 0.039 |
TarDAL | 7.095 | 0.045 | 61.183 | 0.838 | 4.225 | 0.050 |
OURS | 7.072 | 0.061 | 63.479 | 0.819 | 6.845 | 0.029 |
Table 5.
Fusion results of this study were compared quantitatively with fusion results of twelve other advanced methods in object detection. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Table 5.
Fusion results of this study were compared quantitatively with fusion results of twelve other advanced methods in object detection. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result.
Methods Name | Recall | [email protected] | mAP@[0.5:0.95] |
---|
Visible images | 0.287 | 0.326 | 0.153 |
Infrared images | 0.386 | 0.436 | 0.243 |
GTF | 0.336 | 0.373 | 0.208 |
MDLatLRR | 0.362 | 0.408 | 0.223 |
FusionGAN | 0.297 | 0.376 | 0.205 |
GANMcC | 0.353 | 0.403 | 0.223 |
IFCNN | 0.353 | 0.409 | 0.229 |
RFN | 0.251 | 0.22 | 0.126 |
U2Fusion | 0.358 | 0.404 | 0.220 |
DDcGAN | 0.239 | 0.300 | 0.145 |
DIDFuse | 0.326 | 0.380 | 0.206 |
MFEIF | 0.355 | 0.410 | 0.228 |
SeAFusion | 0.358 | 0.399 | 0.223 |
TarDAL | 0.353 | 0.404 | 0.229 |
OURS | 0.355 | 0.417 | 0.237 |
Table 6.
The running time of thirteen comparison methods on M3FD dataset, MFNet dataset, RoadScene dataset, and TNO dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result (unit: s).
Table 6.
The running time of thirteen comparison methods on M3FD dataset, MFNet dataset, RoadScene dataset, and TNO dataset. GREEN highlights the third best result, BLUE highlights the second best result, and RED highlights the best result (unit: s).
Methods Name | M3FD | MFNet | RoadScene | TNO |
---|
GTF | 14.921 | 14.015 | 6.174 | 4.450 |
MDLatLRR | 115.347 | 35.710 | 16.908 | 37.584 |
FusionGAN | 0.401 | 0.165 | 1.058 | 0.660 |
GANMcC | 0.672 | 0.301 | 2.023 | 1.278 |
IFCNN | 3.507 | 45.131 | 38.934 | 25.620 |
RFN | 26.730 | 11.691 | 6.823 | 10.206 |
U2Fusion | 9.081 | 3.240 | 1.717 | 3.039 |
DDcGAN | 1.069 | 0.659 | 3.453 | 2.218 |
DIDFuse | 1.045 | 0.382 | 0.110 | 0.406 |
MFEIF | 0.364
| 0.165 | 0.092 | 0.182 |
SeAFusion | 6.039 | 0.235 | 0.058 | 0.175 |
TarDAL | 0.144 | 0.086 | 0.056 | 0.121 |
OURS | 0.211 | 0.096 | 0.052 | 0.125 |
Table 7.
The mean value of each module in ablation experiment under six different evaluation indices conducted on M3FD dataset. The fusion model with double-classification discriminator added is Module 1, the fusion model with double-classification discriminator and two-branch generator is Module 2, and Module 3 is the fusion model with information interaction added in two-branch generator. RED indicates the best result, and BLUE represents the second best result.
Table 7.
The mean value of each module in ablation experiment under six different evaluation indices conducted on M3FD dataset. The fusion model with double-classification discriminator added is Module 1, the fusion model with double-classification discriminator and two-branch generator is Module 2, and Module 3 is the fusion model with information interaction added in two-branch generator. RED indicates the best result, and BLUE represents the second best result.
Module | EN↑ | SF↑ | PSNR↑ | VIF↑ | AG↑ | MSE↓ |
---|
NO | 6.737 | 0.036 | 63.810 | 0.589 | 4.155 | 0.031 |
Module 1 | 6.668 | 0.026 | 59.592 | 1.035 | 3.364 | 0.047 |
Module 2 | 6.443 | 0.042 | 61.132 | 0.796 | 5.960 | 0.143 |
Module 3 | 7.170 | 0.072 | 64.553 | 0.993 | 8.878 | 0.025 |