Next Article in Journal
Semiconductor Optical Amplifier (SOA)-Driven Reservoir Computing for Dense Wavelength-Division Multiplexing (DWDM) Signal Compensation
Next Article in Special Issue
YOLOv5 Drone Detection Using Multimodal Data Registered by the Vicon System
Previous Article in Journal
Exploring the Limits of Early Predictive Maintenance in Wind Turbines Applying an Anomaly Detection Technique
Previous Article in Special Issue
Dynamic Weighting Network for Person Re-Identification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Fusion and Detection via Deep Learning in UAV-Borne Multispectral Sensing of Scatterable Landmine †

1
School of Electronics and Communication Engineering, Sun Yat-sen University, Shenzhen 518107, China
2
Science and Technology on Near-Surface Detection Laboratory, Wuxi 214035, China
*
Authors to whom correspondence should be addressed.
This paper is an extended version of our paper published in Guo, H.; Jiang, H.; Hu, J.; Luo, C. “UAV-Borne Landmine Detection via Intelligent Multispectral Fusion”, 2022 4th International Conference on Applied Machine Learning (ICAML) , 23–25 July 2022.
Sensors 2023, 23(12), 5693; https://doi.org/10.3390/s23125693
Submission received: 3 May 2023 / Revised: 4 June 2023 / Accepted: 14 June 2023 / Published: 18 June 2023

Abstract

:
Compared with traditional mine detection methods, UAV-based measures are more suitable for the rapid detection of large areas of scatterable landmines, and a multispectral fusion strategy based on a deep learning model is proposed to facilitate mine detection. Using the UAV-borne multispectral cruise platform, we establish a multispectral dataset of scatterable mines, with mine-spreading areas of the ground vegetation considered. In order to achieve the robust detection of occluded landmines, first, we employ an active learning strategy to refine the labeling of the multispectral dataset. Then, we propose an image fusion architecture driven by detection, in which we use YOLOv5 for the detection part, to improve the detection performance instructively while enhancing the quality of the fused image. Specifically, a simple and lightweight fusion network is designed to sufficiently aggregate texture details and semantic information of the source images and obtain a higher fusion speed. Moreover, we leverage detection loss as well as a joint-training algorithm to allow the semantic information to dynamically flow back into the fusion network. Extensive qualitative and quantitative experiments demonstrate that the detection-driven fusion (DDF) that we propose can effectively increase the recall rate, especially for occluded landmines, and verify the feasibility of multispectral data through reasonable processing.

1. Introduction

Landmines have caused great harm to human life and health and greatly hindered the harmonious development of human society [1]. According to incomplete statistics, more than 110 million landmines are still present in post-war areas. To date, many methods have been developed to detect and remove landmines during war or post-war reconstruction. They can be divided into two categories based on their fundamental principles [2]. One employs the geophysical differences between certain parts of landmines and their surroundings, including ground-penetrating radar, low-frequency electromagnetic induction, metal detectors, ultrasonic devices, infrared imaging, and other combinations of detection techniques. The other has a wide range of research dimensions based on particular techniques, such as acoustics, biological sensing, and molecular tracking.
With the development of military technologies, scatterable landmines have emerged and are attracted increasing attention, as opposed to landmines covered with vegetation. The scatterable landmines naturally possess the advantages of speed, flexibility, and high mobility. Consequently, large-scale and large-area minefields can be deployed quickly during war. Moreover, aerial minelaying or projection from armored vehicles results in the randomized distribution of landmines in diverse topographies and complex vegetation environments. These pose new challenges for minefield clearance and make the long-distance and rapid detection of scatterable landmines increasingly necessary.
Meanwhile, massive advances in unmanned aerial vehicles (UAVs) have presented enormous potential [3]. Specifically, UAVs with various on-board electromagnetic sensors have yielded a considerable number of applications, including a rise in low-altitude landmine detection, as they can provide better safety performance and faster measurements than other mine detection platforms. Many investigations aiming at the combination of drones and predominantly ground-penetrating radar (GPR) have been conducted [4,5,6,7,8,9]. The authors in [4,6] aimed to achieve a precise and steady flight over the surface, to increase the positioning accuracy and reliability of GPR measurements, while on-board software-defined radio (SDR)-based radars were developed in [6,10] to detect landmines. Other customized sensors mounted on drones have also emerged. Among them, synthetic aperture radar (SAR) [9,11] and hyper-spectral [12] imaging techniques have shown unprecedented potential with advanced image processing and computer vision algorithms. In [11], the authors reported the capability of high-resolution 3D SAR images for the detection of buried hazards, while the hyper-spectral imaging of landmines was highlighted with different signal processing techniques in [12,13].
Numerous factors, such as the materials and shapes of the mines, terrain complexity, vegetation humidity, flight altitude, etc., limit the detection of scatterable mines. It is challenging for a single sensor to solve all of the aforementioned issues. It is well known that objects exhibit different features in different spectra, and multispectral remote sensing that obtains and comprehensively exploits different spectra information of the same target by means of photography or scanning is appealing. For instance, a particular spectrum may offer a unique advantage in certain terrains or settings, whereas the proper fusion of multispectral remote sensing might expand the observation range and reveal more information. Multispectral fusion greatly outperforms a single spectrum in pedestrian recognition, especially in challenging situations such as occluded appearances and background clutter [14]. However, the development of multispectral techniques in the detection of scattered landmines is still in a preliminary exploration stage. Research in this area mainly concentrates on the analysis of different spectral signatures and the simulation of ground environmental influences [15,16,17,18,19]. Differences in spectral signatures between soils and landmines are essential for detection, and a study has been performed regarding this [15]. The utilization of remote sensing thermal infrared technology has led to satisfactory results in PFM-1 minefield detection [16], and plastic mines such as PFM-1 have higher thermal conductivity when the heat flux is high in the early morning [18]. In addition, ref. [17] simulated the influence of environmental forces on the orientation of PFM-1, complicating the minefield environment.
In the past few years, deep learning, based on the cognitive structure of the human brain, has achieved great success in many areas. Compared with the traditional machine learning algorithm (shallow learning), it can extract more abstract features through multi-layer nonlinear transformation. As one of the fundamental visual tasks, object detection based on deep learning has outperformed traditional methods over the last few years. Ideas and implementations based on convolutional neural networks (CNNs) have facilitated the success of deep learning in detection [20]. Widely acclaimed deep learning models, such as ResNet [21], regions with CNN features [22] (R-CNN), Faster R-CNN [23], etc., show outstanding transferability in mine detection scenes [19,24,25,26]. Some customized deep learning models also show powerful feature extraction abilities and adaptability in varied scenes [27,28]. It is worth noting that different types of data were used in the above studies, such as thermal infrared images, GPR B-scans, and radar signals, but deep learning methods still achieve excellent performance.
The Binghamton University team [16,17,18,19] is conducting ongoing research into multispectral landmine detection. Comparative experiments indicate that thermal cameras can detect temperature variations between plastic PFM-1 landmines and host geology during the early-morning hours, when the thermal flux is high [18]. Their preliminary studies in [16,17] demonstrate that a thermal camera mounted on a low-cost UAV can successfully detect PFM-1 mines at the elevation of 10 meters above the ground, regardless of their orientations. In 2020, they incorporated an analysis of multispectral and thermal datasets and discovered that different bands are beneficial in different situations [19]. Furthermore, they employed the Faster R-CNN [23] model and yielded satisfactory accuracy for a partially or wholly withheld testing set. The employed data only consist of RGB visible images, without fusing other bands for detection, but their solid work gives confidence and inspiration for multispectral detection. In our past work [29], we combined the detection results of RGB and NIR images and finally achieved excellent performance, demonstrating the advantages of multispectral over monospectral images in detection. In this paper, to further investigate the application of multispectral imaging, we explore in depth a new approach as an extension of the aforementioned work.
Most of the existing research and trials only consider an extremely elementary simulation environment and landmine setting. However, high-quality data from the real environment or diversified and high-fidelity experimental scenes are the foundation for the sufficient generalization of supervised deep learning. Therefore, in this study, we were motivated to set up complex minefield conditions to simulate real scenes and collect raw high-quality multispectral data acquired via a drone with an integrated multispectral camera. One RGB sensor and four monochrome sensors collected data in discrete spectral bands—green, red, red edge, and near-infrared. More specifically, our dataset contains four types of scatterable landmines, four scenes with different ground vegetation, and various flight altitudes. To overcome the labeling issues introduced by the severe occlusion and the tiny size of the landmines and improve the labeling efficiency, we utilize the idea of active learning to ensure the complete and accurate labeling of the dataset. To improve the detection performance for the obtained high-quality multispectral images, we present a detection-driven image fusion framework with YOLOv5 [30] plugged in. When compared to the original YOLOv5 [30] benchmark or other fusion detection cascade models, our proposed fusion detection enhances the recall markedly. The RGB and NIR bands are used to demonstrate this in the study. There are 789 pairs of images in our dataset, which are partitioned into training and test sets in an 8:2 ratio. A total of 991 mines are labeled in the test dataset, and a recall rate of 0.848 is achieved when using RGB alone for prediction, while 0.906 is achieved with the fusion of NIR and RGB driven by the landmine detection task. As a plug-and-play module, it can be integrated into most of the current state-of-the-art object detection models.
The rest of the paper is organized as follows. Section 2 illustrates the details of our dataset, including the instruments and the data processing. Section 3 introduces the adopted deep learning method and proposed fusion module. Section 4 presents the experimental results and comparative analysis. Finally, conclusions are drawn in Section 5.

2. Multispectral Dataset

In order to address the shortage of public multispectral landmine datasets, we constructed a small landmine dataset, including labeled RGB and near-infrared (NIR) image pairs.

2.1. Landmine Samples

In the dataset, 40 mines with five different sizes and materials were selected in total, three of which were anti-personnel mines, while the others were anti-vehicle mines. Their primary materials were metal, and the shell parts contained a small amount of plastic and rubber. The specifications of the four types of mines are given in Figure 1.
With the exception of the M93 landmine, they are all olive-drab-covered and cylindrical in shape. Specifically, the smallest anti-personnel mine M14 is less than 6 cm in diameter and primarily non-metallic. In comparison, the anti-vehicle M93 Hornet mine possesses a relatively large body size and is plastic-cased with a rubber top. Before the explosion, they are air-delivered and hidden well on the ground or in vegetation, remaining active without being noticed for long periods.

2.2. Equipment

We acquired the multispectral images using the Sequoia multispectral kit. The Sequoia sensor comprises four monochrome cameras with specific narrow-band filters and high-resolution RGB lenses in the NIR and visible domains, respectively, as shown in Table 1.
The camera is equipped with a “sunlight sensor” that measures sky irradiance. The built-in spectroradiometer has precision of 1.5 nm to measure irradiance of 2048 pixels in the visible and NIR spectral ranges (350–1000 nm). The drone equipped with the multispectral sensor (Figure 2) can sustain a continuous flight of an hour approximately.
The preparations before take-off include the calibration of the drones and multispectral cameras and radiation correction through the sunshine sensor. The matching sunshine sensor can monitor radiation fluctuations to obtain multispectral images with stable quality under different lighting and natural conditions. The flying height changes according to the host environment and whether sparse, tall plants are present. In this work, the flying heights are fixed at approximately 3 m and 8 m, respectively, to obtain images with a similar ground resolution. During the data acquisition process, the drone cruises at a constant speed with an area overlap rate of 80% to ensure complete coverage of the minefields of the given range. Moreover, the built-in GPS/GNSS of the multispectral camera can determine the device’s location when capturing images, which provides good feasibility to convert the image coordinate system and the geographic coordinate system.

2.3. Experimental Scenes

We conducted data collection experiments and placed dozens of mines on grassland, instead of experimental scenes such as sand land, cement roads, and gravel roads, etc. A more complex environment with vegetation is considered to validate the feasibility of our method. We selected several low-vegetation environments to simulate minefields. As displayed in Figure 3, mines are easily obscured by wild vegetation. The mines were scattered randomly, with approximately 60% obscured by vegetation; in other words, only part of the mine body was visible. Moreover, a small number of them were entirely covered by ground vegetation, which caused extremely low visibility. In addition, there were many leaves, stones, and plant rhizomes that were similar in color and shape to the mines, which may have caused false alarms.

2.4. Acquisition and Registration

With a constant capturing interval of 2 s, UAVs may complete the minefield cruise in a very short amount of time. Sequoia’s monochrome and RGB cameras are not calibrated throughout production, resulting in a lack of rigorous time synchronization due to the various shutters. As a result, the collected multispectral raw images have a spatio-temporal shift. Thus, in addition to the typical nonlinear intensity differences, there are also spatio-temporal mismatches between the multi-modal data. For instance, the vegetation in the simulated landscape is mostly filled with dense grasses that sway in the wind in different directions. Completely accurate image registration is almost impossible. This can be a significant problem for actual minefield detection, but not for host environments without vegetation. Registration allows image pixels to be aligned for more precise fusion and also eases subsequent labeling. The registered infrared and visible image pair that we obtained is shown in Figure 4. We also refer to the channel feature of oriented gradients [31] (CFOG) method to register the multispectral images, which could effectively further improve the precision of the registration result. By comparing the positions of the corresponding landmines, we could directly assess the quality of the registration result.

2.5. Labeling

Supervised learning strongly relies on high-quality labeled data, which usually comes with high costs. Manually labeled data are time-consuming to obtain and sometimes unreliable. Even in the well-known public datasets in the academic community, there are still many instances of mislabeling. In this paper, we labeled our dataset using the following steps: detection equipment assembly, data collection, data cleaning, image alignment, and labeling.
Firstly, random landmine placement and extreme occlusion substantially increase the difficulty of labeling. Moreover, nonsalient objects such as landmines usually blend in easily with their surroundings. In a 1280 × 960 resolution image, the target is estimated to occupy 100 pixels, while, for the smallest mine, M14, it occupies only approximately 10 pixels. It is not feasible to label each image manually. Therefore, during the labeling precess, we referred to the strategy described in [14]. We compared and labeled images in parallel after aligning the image pairs. Moreover, to accelerate the labeling, we introduced an active learning technique, which is shown in Figure 5.
In real-world circumstances, it is worth considering how to collect valuable labeled data for a lower cost and subsequently increase the algorithm’s performance. Active learning focuses on reducing the burden of labeling by incorporating human experience into machine learning models. Firstly, the query function determines the candidate set of data to be labeled. Then, professionals with business experience annotate the selected datasets. With continual manual annotation to be added, the model is therefore incrementally updated. The core principle behind active learning is to manually resolve the “hard sample” via human interference. In our situation, the term “hard sample” refers to landmines that are well hidden and difficult to detect. Furthermore, the training results assist in the improvement of the annotation. After three rounds of the iterative process, a dataset with reliable annotations was created.

3. Methods

Existing image fusion algorithms tend to focus on the visual effect of the fused image itself and the pursuit of quantitative evaluation, while ignoring the promotion ability of image fusion as a low-level vision task for the subsequent high-level vision task. In fact, the ideal image fusion method should not only provide rich texture details and accord with human visual perception, but also help to conduct high-quality downstream tasks.
Compared with the traditional architectures, the fusion architecture driven by high-level vision tasks can better conserve semantic information as well as exclude external noise. Liu in [32] designed a cascade structure to connect the image denoising module with the high-level vision network and jointly trained them to minimize the image reconstruction loss. Under the guidance of semantic information, a denoising network could further improve the visual effect and help to complete the semantic task. Similarly, Guo in [33] designed a residual network for semantic refinement and jointly trained the “drainage” network and the segmentation network in two stages, which was successfully utilized to overcome the impact of rainwater on sensors during automatic driving. Object detection, as one of the high-level vision tasks, was also used to guide image super-resolution reconstruction. In the research of [34], the authors built a novel end-to-end super-resolution framework, in which the updating of the super-resolution network was influenced by the detection loss. The proposed task-driven super-resolution algorithm was verified to improve the detection accuracy on the low-resolution image under various scale factors or conditions.

3.1. Detection-Driven Fusion Framework

By focusing on landmine detection, we introduce a joint training framework based on detection-driven fusion (DDF), as shown in Figure 6, in which we select YOLOv5 [30] for detection.
The image fusion network is not limited to the preservation of low-level semantic information but is driven by the detection task to gather more valuable high-level semantic information. Meanwhile, the fusion network is partially influenced by the detection result, which can preserve fine-grained details while providing more semantic information for computer perception.

3.2. Fusion Network

On the basis of auto-encoder, we design a lightweight fusion network (denoted as DDFNet) with two inputs and one output, in which the registered images include the visible image I r g b R H W 3 and the near-infrared image I n i r R H W 1 as the original inputs. The images pass through the encoder, fusion layer, and decoder, respectively, as shown in Figure 7, where the encoder is used for feature extraction and the decoder for image reconstruction, which can be expressed as follows:
F r g b , F n i r = EN ( I r g b ) , EN ( I n i r ) .
F f = Conv ( Concat ( F r g b , F n i r ) ) .
I f = DE ( F f )
where F r g b and F n i r are the depth features extracted by two different encoders EN , respectively. F f denotes the fused features of the visible and the infrared images through concatenation, and  I f is the fused image reconstructed by decoder DE. In Figure 7, the encoder part is mainly composed of two general convolutional blocks as well as two dense residual blocks (DRB), as shown in Figure 8, which refers to the combination of the residual and dense blocks from [35]. The main branch of DRB aggregates shallow and deep semantic information by dense connection as the shortcut branch uses the classic 1 × 1 convolutional layer in the residual block to achieve identity mapping. Next, the output of the main branch and the shortcut are added at the element level to integrate deep features with fine-grained features. Thus, the superposition of two DRBs in the encoder ensures that the abundant semantic features are fully utilized. Then, the fusion layer fuses the rich features extracted from the infrared image and the visible image via concatenation. Finally, the fused features are fed back into the decoder for image reconstruction. With the dimensions of the feature map shrinking, the decoder outputs a three-channel fused image. In general, the design of our fusion network is simple and lightweight, which contributes to a higher fusion speed, and it is easy to facilitate subsequent deployments. Moreover, the introduction of the auto-encoder architecture as well as the DRB makes it possible for the fused image to fully restore the details and semantic information of the source images.

3.3. Loss Function

The design of the loss function starts with two aspects. Firstly, the fusion network needs to fully integrate the complementary information of the source images, such as the salient objects in the infrared image and the texture details in the visible image. For this reason, the content loss L c o n t e n t is constructed, which aims to ensure the visual fidelity of the fused image. Secondly, the fused image should effectively promote the detection effect. Therefore, the semantic loss L d e t e c t is introduced to indicate the contribution of the fused image to the detection task. In summary, the joint loss function is designed as follows:
L a l l = L c o n t e n t + λ L d e t e c t .
where the joint loss L a l l of the fusion network comprises the content loss L c o n t e n t and the semantic loss L d e t e c t , and  λ is set as the weight of the semantic loss.
Meanwhile, the content loss is composed of three types of loss, image intensity loss ( L i n t e n s i t y ), texture loss( L t e x t u r e ) [35], and structural similarity loss ( L s s i m ), formatted as
L c o n t e n t = L i n t e n s i t y + β 1 L t e x t u r e + β 2 L s s i m .
where β 1 and β 2 are the hyperparameters to define the weights of texture loss and structural similarity loss, respectively. Image intensity loss L i n t e n s i t y assesses the difference in pixel intensity between a fused image and a source image, which can be more precisely formulated as
L i n t e n s i t y = 1 H W I f max ( I r g b , I n i r ) 1
where H and W are, respectively, the height and width of the fused image I f . · 1 stands for the l 1 -norm. m a x ( · ) ensures that the fused image maintains the intensity sufficiently, which contributes to the positioning of landmines. The texture loss L t e x t u r e is introduced to preserve the details. It can be calculated as
L t e x t u r e = 1 H W | I f | max ( | I r g b | , | I n i r | ) 1
where the absolute value operation is denoted by | · | and the Sobel gradient operator is denoted by ∇. Although the standard L 2 distance can maintain edges well in image fusion, it cannot effectively remove noise. Moreover, it lacks the capability to determine how structurally similar the images are, and it is highly sensitive to changes in lighting, color, etc., which poses a challenge in maintaining specificity in multispectral images. The Structure Similarity Index Measure (SSIM) proposed by [36] is used to compare the structural similarity of an image before and after alteration. SSIM is sensitive to local structural changes, as with the human visual system (HVS). More importantly, SSIM can better tolerate the influence of outside information on the source image. The SSIM between two images can be more precisely expressed as follows:
S S I M x , y = x i , y i 2 μ x i μ y i + C 1 μ x i 2 + μ y i 2 + C 1 × 2 σ x i σ y i + C 2 σ x i 2 + σ y i 2 + C 2 × σ x i y i + C 3 σ x i σ y i + C 3
where μ and σ represent the mean and the covariance of images, respectively, and  C 1 , C 2 , C 3 are the parameters used to stabilize the measurement. Therefore, the SSIM loss L s s i m can be formulated as
L s s i m = w 1 ( 1 S S I M I f , I r g b ) + w 2 ( 1 S S I M I f , I n i r )
where w 1 and w 2 denote the weights of structural loss between the fused image I f and the source image I r g b and I n i r , respectively. Although content loss can maintain information in the fused image and minimize distortion to some extent, it is challenging to measure the promotion impact of the fusion result on high-level vision tasks. As a consequence, we redefine the semantic loss using the standard loss function based on YOLOv5 [30] to guide the training of the fusion network and ultimately enhance the detection performance.
The object detection task normally requires us to refine the target size as well as identify and categorize the detected item. In particular, the accuracy of the test results depends on three optimization directions:
  • All targets in the image should be identified, with a low missed and false alarm rate;
  • The bounding boxes should completely and accurately enclose the target;
  • The categorization of the detected object should be consistent with its label.
Therefore, the semantic loss (marked L d e t e c t [30]) is defined by the confidence loss (marked l o s s c o n f ), the bounding box loss (marked l o s s r e c t ), and the categorized loss (marked l o s s c l s ) as follows:
L d e t e c t = a × l o s s r e c t + b × l o s s c o n f + c × l o s s c l s .
where a, b, and c are their weights, respectively. The confidence loss generally weighs the most, followed by the bounding box loss and the categorized loss.

3.4. Joint Training Algorithm

There are primarily two approaches to low-level vision model training that is motivated by high-level visual tasks. The first one is to train the low-level vision model under the guidance of the pre-trained high-level vision model, but the training result cannot be effectively fed back to the high-level vision model. The second one is to concurrently train the low-level and high-level vision models in one stage, and this strategy may bring difficulties in maintaining a balance between the performance of vision models at various levels. Therefore, we introduce a joint training method based on an iterative strategy to train our fusion network.
More specifically, we train the fusion and detection networks repeatedly, with the number of iterations set to n. Firstly, the fusion and the detection network employ the Adaptive Moment Estimation (ADAM) optimizer to maximize each of their individual parameters under the supervision of the joint loss L a l l and the detection loss L d e t e c t , respectively. Secondly, the parameters of the detection model are fixed to generate real-time semantic loss when the fusion network is updated. It should be noted that the hyperparameter λ set for the joint loss is dynamically adjusted in an iterative process, which is expressed as follows:
λ n = n × λ n 1
This option is primarily intended to guarantee that, as the fusion network has been trained to a certain extent, the fraction of semantic loss steadily increases, and the visual effect becomes a secondary priority.
During the training process, we adopt the periodic joint training strategy demonstrated in Algorithm 1, which is iterated for 5 rounds.
Algorithm 1: Multi-stage joint training algorithm.
Sensors 23 05693 i001
In the first round, the fusion network only utilizes the content loss for backward propagation, while the semantic loss is set to 0. Then, in the next four rounds, both the content and semantic loss contribute to the updating of the fusion network. In addition, the detection network is trained through the updated fused image in each round.

4. Experiment and Results

4.1. Fusion Result

4.1.1. Comparative Experiment

In our experiments, the landmine dataset is randomly divided into a training part and a testing part at a ratio of 8:2. The hyperparameters involved are all set to the default values, with no validation set for adjustment. This experimental setup makes it easier to demonstrate the feasibility of our landmine detection method.
To begin with, we conducted fusion comparative experiments for DDF and other algorithms. Figure 9 shows fused results of the selected deep learning fusion models and some conventional algorithms in the Visible and Infrared Image Fusion Benchmark [37,38] (VIFB). It should be noted that since FusionGAN [39] and RFN-Nest [40] only work for single-channel images with a depth of 8 bits, the fusion result is also a single-channel grayscale image. The testing images are randomly selected from the infrared and visible image pairings from the multispectral landmine dataset, and part of the region is intercepted to illustrate the fusion effect. There are two types of scatterable mines (M93 and T125) in the interception area, of which the rose-red mine (M93) is easy to identify due to its great size and distinguishing characteristics. T125 is partially covered by vegetation, and therefore it is difficult to detect in visible images but it reveals strong visual features in near-infrared images. It can be observed from Figure 9 that M93 is still clearly identifiable in the fusion results of most algorithms, while the distortion is severe in certain fused images. For example, testing images employing Densefuse [41], IFCNN [42], LatLRR [43], and GFCE [44] perform poorly in terms of pixel intensity and color. Moreover, the fused image generated by FusionGAN has the lowest image quality, and the target characteristics are nearly lost. Compared to the source images (RGB and NIR), RFN-Nest [40] could successfully integrate the complementary features of the two spectra. However, although the significant features of landmines in the two images are likewise preserved in the fused image, the color information is lost and is visibly different. In contrast, U2Fusion [45] well maintains the color feature distribution, but it introduces some noise in the white area. As for MST-SR [46], NSCT-SR [46], and the DDF that we propose, all of them enhance the visual saliency of T125 while preserving the original color.
Next, we randomly chose 20 pairs of near-infrared and visible images from the multispectral landmine dataset and processed them using different fusion algorithms. We performed a comprehensive fusion evaluation using 13 statistical metrics [37,38], which is shown in Table 2.
To some extent, the evaluation metrics are designed to quantitatively simulate the human visual system (HVS). Some of them, such as the Root Mean Square Error (RMSE), Peak Signal to Noise Ratio (PSNR), and Structural Similarity Index Measure (SSIM), are measured based on the original RGB and NIR images as benchmarks, which can reflect the amount of complementary information retained from the source images in the fused image. Others, such as the Average Gradient (AG), Entropy Index (EI), and Entropy (EN), can be calculated with the fused image itself and be used for the evaluation of the image quality, such as edges, saliency, and other features, which may be helpful for the subsequent detection.
In Table 2, blue represents the best-performing model for the corresponding metric. For the quantitative metrics designated as ↓, a lower value is better. The last row of Table 2 displays the evaluation metrics of DDF, where red, orange, and green, respectively, represent excellent, moderate, and poor performance of the fused image based on DDF compared with other fusion algorithms. It can be seen that LatLRR [43] ranks the highest for approximately half of the metrics, but it is not satisfactory regarding the visual effect shown in the aforementioned Figure 9. NSCT-SR [46] and MST-SR [46] perform better in the metrics of Q C B [48] and Q C V [49], as it can be seen in Figure 9 that both of them provide fusion results consistent with human vision. The DDF that we propose demonstrates satisfactory performance in most of the statistical indicators, while only providing unsatisfactory assessment results for Cross-Entropy (CE) and the Peak Signal to Noise Ratio (PSNR).

4.1.2. Generalization Experiment

In addition, in order to verify the generalization performance of the proposed method, we also chose 21 pairs of infrared and visible images from VIFB [37,38] as a further evaluation. Figure 10 offers a qualitative comparison between DDF and the other 10 fusion models. It can be observed that the fusion results of Densefuse [41], U2Fusion [45], FusionGAN [39], RFN-Nest [40], GFCE [44], MST-SR [46], and NSCT-SR [46] are significantly affected by radiation or lose a lot of color information; among them, RFN-Nest [40] shows blurred edges and a loss of detail. IFCNN [42] and DDF achieve satisfactory fusion results, presenting excellent visual effects. For example, the textural details of the distant buildings in the image are greatly enhanced compared with the original visible image.
Table 3 lists the mean values of the 21 pairs of image fusion metrics for the VIFB [37,38] dataset. It can be observed that the performance of DDF is mostly moderate. The reason is that the DDF is trained with the scatterable mine images—more specifically, the near-infrared and visible images—while the infrared images in VIFB [37,38] are mainly thermal infrared. The contributions of the different statistical metrics to the final advanced visual task are still difficult to analyze and determine, but in the experimental study of this section, DDF shows better performance in the evaluation of some metrics and has an excellent visual effect.

4.2. Detection Performance

Furthermore, we conducted comparisons of the detection performance for the aforementioned fusion models. During the training and testing process, we employed two NVIDIA GeForce RTX 3090 graphics cards. We randomly shuffled and divided the collected data into a training set and a testing set at a ratio of 8:2, with quantities of 3964 and 991, respectively.

4.2.1. Landmine Detection Evaluation

To further validate the DDF that we propose under the condition of vegetation, we compared the detection performance of DDF with other fusion methods. We randomly selected an image pair from the testing set for landmine detection. Figure 11 shows the detection results based on different fusion methods. Except for the source images RGB and NIR, the first two rows are based on deep learning fusion methods, while the last row relates to conventional fusion methods. As can be seen from the source images, most of the landmines in RGB can be distinguished more clearly. However, a small number of landmines, such as T125 and M93, magnified to be displayed in ground truth, show clearer outlines in the NIR image, which demonstrates the complementary information of RGB and NIR images. Therefore, it can be observed that most of the landmines, besides the aforementioned T125, can be directly detected in RGB, outperforming most of the fusion-based algorithms. DDF successfully detects all landmines, even when some of them are covered with vegetation. It is worth noting that the detection of T125 is not so precise, as the main body of its long strip is obscured by weeds. MST-SR [46] can also obtain a high recall rate, but it may cause false alarms. NSCT-SR [46] is able to detect most of the landmines, with the exception of B91, which is occluded severely.
Then, we produced a statistical table according to the detection results of all the images in the testing set. The detection evaluation metrics of various fusion models are shown in Table 4, in which red and blue, respectively, represent the best result and the second-best result. It can be observed that the DDF has the best performance in terms of precision, recall, and [email protected]. In terms of the recall rate, the detection results of most fusion models are even worse than those of single-spectrum visible images, which also indicates that the existing fusion algorithms do not focus sufficiently on high-level visual tasks. The superior performance of DDF demonstrates the validity and feasibility of our methodology.

4.2.2. Efficiency Evaluation

Next, we evaluated the inference efficiency of different fusion models, which is shown in Table 5. The experiment was conducted twice, one on 20 pairs of randomly selected landmine images and the other on the VIFB [37,38] dataset, including 21 pairs of visible and infrared images. Only one NVIDIA GeForce RTX 3090 graphics card was utilized for inference. Table 5 shows the time consumption of the different models—more specifically, the average time consumption of the fusion process for each pair of images. The calculation covers the process of data loading and image storage, while the time for model loading is not included. Since the efficiency of the traditional algorithm is generally much lower than that of the deep learning model with the forward computation, the table is separated into two parts: fusion models based on either the traditional or deep learning methods. Green highlights the highest time cost in the current column, while red indicates the lowest. On one hand, we can observe that the well-performing algorithm LatLRR [43], based on the traditional method, is much more time-consuming than the others. On the other hand, except for CNN [52] used as a reference, the distinction between the models based on deep learning is not obvious. Meanwhile, the inference efficiency of our model is moderate, whereas the IFCNN [42] model, which performs poorly in terms of the visual effect or the detection task, shows slight advantages in inference efficiency.

5. Conclusions

In recent years, the application of deep-learning-based techniques in landmine detection has progressed at a rapid pace. This paper investigates recent advances and suggests some opportunities for future research in this field. Specifically, we propose a UAV-based solution for the detection of scatterable landmines. We establish a dataset of scatterable landmines, including RGB and NIR image pairs collected by a multispectral sensor, with complex scenes covered. In addition, to fully utilize the multispectral potential in detection, we employ a training framework for image fusion, which is task-oriented and allows high-level semantic information to flow back into the fusion network upstream, so as to boost the quality of the fusion image as well as enhance the detection performance purposefully. We have carried out extensive comparative experiments to verify the effectiveness of the fusion and detection methods that we propose. Our attempt at multispectral utilization for landmine detection paves the way for future advancement.
Overall, the obscuration and burial of landmines remain the most challenging problems that we confront. We have focused on the concealment and occlusion of landmines by vegetation when constructing the dataset. However, the simulated settings are still far from the natural minefield environment. The aging, corrosion, and burial of mines in real situations are time-consuming and complex to reproduce. Furthermore, the range of host conditions, such as snow, gravel pavement, and so on, needs to be greatly enlarged. In the future, we will supplement the dataset with more hard samples (primarily obscured by weeds, fallen leaves, sand, snow, and gravel or buried under shallow soils). We expect that the proposed dataset will encourage the development of better landmine detection methods, and our efforts will be of assistance and inspiration to demining organizations.

Author Contributions

Methodology, H.G.; validation, J.H.; resources, C.L.; data curation, H.J.; writing—original draft preparation, Z.Q. and H.G.; writing—review and editing, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the National Key R&D Program of China under Grant 2021YFA0716600, the Science and Technology on Near-Surface Detection Laboratory Pre-Research Foundation under Grant No. 6142414200607, the Shenzhen Fundamental Research Program under Grant No. JCYJ20180307151430655, and the Shenzhen Science and Technology Program under Grant No. KQTD20190929172704911.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. ICBL-CMC. Landmine Monitor 2021. 2021. Available online: http://www.the-monitor.org/media/3318354/Landmine-Monitor-2021-Web.pdf/ (accessed on 13 June 2023).
  2. Hussein, E.; Waller, E. Landmine detection: The problem and the challenge. Appl. Radiat. Isot. 2000, 53, 557–563. [Google Scholar] [CrossRef] [PubMed]
  3. Miccinesi, L.; Beni, A.; Pieraccini, M. UAS-Borne Radar for Remote Sensing: A Review. Electronics 2022, 11, 3324. [Google Scholar] [CrossRef]
  4. Colorado, J.; Devia, C.; Perez, M.; Mondragon, I.; Mendez, D.; Parra, C. Low-altitude autonomous drone navigation for landmine detection purposes. In Proceedings of the 2017 International Conference on Unmanned Aircraft Systems (ICUAS), Miami, FL, USA, 13–16 June 2017; pp. 540–546. [Google Scholar]
  5. Šipoš, D.; Gleich, D. A lightweight and low-power UAV-borne ground penetrating radar design for landmine detection. Sensors 2020, 20, 2234. [Google Scholar] [CrossRef] [Green Version]
  6. Colorado, J.; Perez, M.; Mondragon, I.; Mendez, D.; Parra, C.; Devia, C.; Martinez-Moritz, J.; Neira, L. An integrated aerial system for landmine detection: SDR-based Ground Penetrating Radar onboard an autonomous drone. Adv. Robot. 2017, 31, 791–808. [Google Scholar] [CrossRef]
  7. Sipos, D.; Planinsic, P.; Gleich, D. On drone ground penetrating radar for landmine detection. In Proceedings of the 2017 First International Conference on Landmine: Detection, Clearance and Legislations (LDCL), Beirut, Lebanon, 26–28 April 2017; pp. 1–4. [Google Scholar]
  8. García-Fernández, M.; López, Y.Á.; Andrés, F.L.H. Airborne multi-channel ground penetrating radar for improvised explosive devices and landmine detection. IEEE Access 2020, 8, 165927–165943. [Google Scholar] [CrossRef]
  9. Schreiber, E.; Heinzel, A.; Peichl, M.; Engel, M.; Wiesbeck, W. Advanced buried object detection by multichannel, UAV/drone carried synthetic aperture radar. In Proceedings of the 2019 13th European Conference on Antennas and Propagation (EuCAP), Krakow, Poland, 31 March–5 April 2019; pp. 1–5. [Google Scholar]
  10. Bossi, L.; Falorni, P.; Capineri, L. Versatile Electronics for Microwave Holographic RADAR Based on Software Defined Radio Technology. Electronics 2022, 11, 2883. [Google Scholar] [CrossRef]
  11. Garcia-Fernandez, M.; Alvarez-Lopez, Y.; Las Heras, F. Autonomous airborne 3D SAR imaging system for subsurface sensing: UWB-GPR on board a UAV for landmine and IED detection. Remote Sens. 2019, 11, 2357. [Google Scholar] [CrossRef] [Green Version]
  12. Makki, I.; Younes, R.; Francis, C.; Bianchi, T.; Zucchetti, M. A survey of landmine detection using hyperspectral imaging. ISPRS J. Photogramm. Remote Sens. 2017, 124, 40–53. [Google Scholar] [CrossRef]
  13. Khodor, M.; Makki, I.; Younes, R.; Bianchi, T.; Khoder, J.; Francis, C.; Zucchetti, M. Landmine detection in hyperspectral images based on pixel intensity. Remote Sens. Appl. Soc. Environ. 2021, 21, 100468. [Google Scholar] [CrossRef]
  14. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; So Kweon, I. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar]
  15. Silva, J.S.; Guerra, I.F.L.; Bioucas-Dias, J.; Gasche, T. Landmine detection using multispectral images. IEEE Sens. J. 2019, 19, 9341–9351. [Google Scholar] [CrossRef]
  16. DeSmet, T.; Nikulin, A.; Frazer, W.; Baur, J.; Abramowitz, J.; Finan, D.; Denara, S.; Aglietti, N.; Campos, G. Drones and “Butterflies”: A Low-Cost UAV System for Rapid Detection and Identification of Unconventional Minefields. J. Conv. Weapons Destr. 2018, 22, 10. [Google Scholar]
  17. Nikulin, A.; De Smet, T.S.; Baur, J.; Frazer, W.D.; Abramowitz, J.C. Detection and identification of remnant PFM-1 ‘Butterfly Mines’ with a UAV-Based thermal-imaging protocol. Remote Sens. 2018, 10, 1672. [Google Scholar] [CrossRef] [Green Version]
  18. de Smet, T.S.; Nikulin, A. Catching “butterflies” in the morning: A new methodology for rapid detection of aerially deployed plastic land mines from UAVs. Lead. Edge 2018, 37, 367–371. [Google Scholar] [CrossRef]
  19. Baur, J.; Steinberg, G.; Nikulin, A.; Chiu, K.; de Smet, T.S. Applying deep learning to automate UAV-based detection of scatterable landmines. Remote Sens. 2020, 12, 859. [Google Scholar] [CrossRef] [Green Version]
  20. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object Detection in 20 Years: A Survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  23. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Ward, C.M.; Harguess, J.; Hilton, C.; Mediavilla, C.; Sullivan, K.; Watkins, R. Deep learning for automatic ordnance recognition. In Geospatial Informatics IX; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10992, p. 109920H. [Google Scholar]
  25. Priya, C.N.; Ashok, S.D.; Maji, B.; Kumaran, K.S. Deep Learning Based Thermal Image Processing Approach for Detection of Buried Objects and Mines. Eng. J. 2021, 25, 61–67. [Google Scholar] [CrossRef]
  26. Kafedziski, V.; Pecov, S.; Tanevski, D. Detection and classification of land mines from ground penetrating radar data using faster R-CNN. In Proceedings of the 2018 26th Telecommunications Forum (TELFOR), Belgrade, Serbia, 20–21 November 2018; pp. 1–4. [Google Scholar]
  27. Picetti, F.; Testa, G.; Lombardi, F.; Bestagini, P.; Lualdi, M.; Tubaro, S. Convolutional autoencoder for landmine detection on GPR scans. In Proceedings of the 2018 41st International Conference on Telecommunications and Signal Processing (TSP), Athens, Greece, 4–6 July 2018; pp. 1–4. [Google Scholar]
  28. Lameri, S.; Lombardi, F.; Bestagini, P.; Lualdi, M.; Tubaro, S. Landmine detection from GPR data using convolutional neural networks. In Proceedings of the 2017 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 508–512. [Google Scholar]
  29. Guo, H.; Jiang, H.; Hu, J.; Luo, C. UAV-Borne Landmine Detection via Intelligent Multispectral Fusion. In Proceedings of the 2022 4th International Conference on Applied Machine Learning (ICAML), Changsha, China, 23–25 July 2022; IEEE Computer Society: Washington, DC, USA, 2022; pp. 179–183. [Google Scholar]
  30. Jocher, G. yolov5. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 13 June 2023).
  31. Ye, Y.; Bruzzone, L.; Shan, J.; Bovolo, F.; Zhu, Q. Fast and Robust Matching for Multimodal Remote Sensing Image Registration. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9059–9070. [Google Scholar] [CrossRef] [Green Version]
  32. Liu, D.; Wen, B.; Jiao, J.; Liu, X.; Wang, Z.; Huang, T.S. Connecting image denoising and high-level vision tasks via deep learning. IEEE Trans. Image Process. 2020, 29, 3695–3706. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Guo, M.; Chen, M.; Ma, C.; Li, Y.; Li, X.; Xie, X. High-level task-driven single image deraining: Segmentation in rainy days. In Proceedings of the International Conference on Neural Information Processing, Bangkok, Thailand, 23–27 November 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 350–362. [Google Scholar]
  34. Haris, M.; Shakhnarovich, G.; Ukita, N. Task-driven super resolution: Object detection in low-resolution images. In Proceedings of the International Conference on Neural Information Processing: 28th International Conference, ICONIP 2021, Sanur, Bali, Indonesia, 8–12 December 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 387–395. [Google Scholar]
  35. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  36. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Zhang, X.; Ye, P.; Xiao, G. VIFB: A Visible and Infrared Image Fusion Benchmark. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 468–478. [Google Scholar]
  38. Zhang, X.; Demiris, Y. Visible and Infrared Image Fusion Using Deep Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 1, 1–20. [Google Scholar] [CrossRef]
  39. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  40. Li, H.; Wu, X.J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  41. Li, H.; Wu, X.J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [Green Version]
  42. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  43. Yu, S.; Chen, X. Infrared and Visible Image Fusion Based on a Latent Low-Rank Representation Nested With Multiscale Geometric Transform. IEEE Access 2020, 8, 110214–110226. [Google Scholar] [CrossRef]
  44. Zhou, Z.; Dong, M.; Xie, X.; Gao, Z. Fusion of infrared and visible images for night-vision context enhancement. Appl. Opt. 2016, 55, 6480–6490. [Google Scholar] [CrossRef] [PubMed]
  45. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  46. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  47. Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429), Barcelona, Spain, 14–17 September 2003; Volume 3, p. III–173. [Google Scholar]
  48. Chen, Y.; Blum, R.S. A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 2009, 27, 1421–1432. [Google Scholar] [CrossRef]
  49. Chen, H.; Varshney, P.K. A human perception inspired quality metric for image fusion based on regional information. Inf. Fusion 2007, 8, 193–207. [Google Scholar] [CrossRef]
  50. Bavirisetti, D.P.; Dhuli, R. Fusion of Infrared and Visible Sensor Images Based on Anisotropic Diffusion and Karhunen-Loeve Transform. IEEE Sens. J. 2016, 16, 203–209. [Google Scholar] [CrossRef]
  51. Shreyamsha Kumar, B. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015, 9, 1193–1204. [Google Scholar] [CrossRef]
  52. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolution Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  53. Li, H.; Wu, X.J.; Kittler, J. Infrared and visible image fusion using a deep learning framework. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2705–2710. [Google Scholar]
  54. Bavirisetti, D.P.; Xiao, G.; Liu, G. Multi-sensor image fusion based on fourth order partial differential equations. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017; pp. 1–9. [Google Scholar]
  55. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  56. Zhang, Y.; Zhang, L.; Bai, X.; Zhang, L. Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol. 2017, 83, 227–237. [Google Scholar] [CrossRef]
  57. Bavirisetti, D.P.; Xiao, G.; Zhao, J.; Dhuli, R.; Liu, G. Multi-scale guided image and video fusion: A fast and efficient approach. Circuits Syst. Signal Process. 2019, 38, 5576–5605. [Google Scholar] [CrossRef]
  58. Naidu, V. Image fusion technique using multi-resolution singular value decomposition. Def. Sci. J. 2011, 61, 479. [Google Scholar] [CrossRef] [Green Version]
  59. Li, H.; Wu, X.j.; Durrani, T.S. Infrared and visible image fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 2019, 102, 103039. [Google Scholar] [CrossRef] [Green Version]
  60. Bavirisetti, D.P.; Dhuli, R. Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. 2016, 76, 52–64. [Google Scholar] [CrossRef]
  61. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  62. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  63. Zhou, Z.; Wang, B.; Li, S.; Dong, M. Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion 2016, 30, 15–26. [Google Scholar] [CrossRef]
Figure 1. Landmine samples. (a) Type: M14; size: diameter 58 mm × height 38 mm; (b) Type: T125; size: diameter 50 mm × height 135 mm; (c) Type: B91; size: diameter 147 mm × height 66 mm; (d) Type: M93; size: diameter 122 mm × height 168 mm.
Figure 1. Landmine samples. (a) Type: M14; size: diameter 58 mm × height 38 mm; (b) Type: T125; size: diameter 50 mm × height 135 mm; (c) Type: B91; size: diameter 147 mm × height 66 mm; (d) Type: M93; size: diameter 122 mm × height 168 mm.
Sensors 23 05693 g001
Figure 2. Equipment: Left—UAV; Right—Sequoia kit (with a sunshine sensor and a multispectral camera).
Figure 2. Equipment: Left—UAV; Right—Sequoia kit (with a sunshine sensor and a multispectral camera).
Sensors 23 05693 g002
Figure 3. Mines on different backgrounds.
Figure 3. Mines on different backgrounds.
Sensors 23 05693 g003
Figure 4. Registration, showing that some mines are clearer in NIR.
Figure 4. Registration, showing that some mines are clearer in NIR.
Sensors 23 05693 g004
Figure 5. The workflow of active learning.
Figure 5. The workflow of active learning.
Sensors 23 05693 g005
Figure 6. The architecture of the joint training framework based on detection-driven fusion (DDF).
Figure 6. The architecture of the joint training framework based on detection-driven fusion (DDF).
Sensors 23 05693 g006
Figure 7. The architecture of the visible and infrared image fusion network (termed DDFNet), which is based on auto-encoder. The convolution leaky ReLu block (CLB 3 × 3 ) is composed of a 3 × 3 convolutional layer and a leaky ReLu layer. The convolution tanh block (CTB 1 × 1 ) is composed of a 1 × 1 convolutional layer and a tanh layer.
Figure 7. The architecture of the visible and infrared image fusion network (termed DDFNet), which is based on auto-encoder. The convolution leaky ReLu block (CLB 3 × 3 ) is composed of a 3 × 3 convolutional layer and a leaky ReLu layer. The convolution tanh block (CTB 1 × 1 ) is composed of a 1 × 1 convolutional layer and a tanh layer.
Sensors 23 05693 g007
Figure 8. The structure of the dense residual block (DRB). Convolution leaky ReLu block (CLB) is composed of a convolutional layer and a leaky ReLu layer.
Figure 8. The structure of the dense residual block (DRB). Convolution leaky ReLu block (CLB) is composed of a convolutional layer and a leaky ReLu layer.
Sensors 23 05693 g008
Figure 9. Qualitative comparisons of fusion results for a visible (RGB) and near-infrared (NIR) image pair from landmine dataset.
Figure 9. Qualitative comparisons of fusion results for a visible (RGB) and near-infrared (NIR) image pair from landmine dataset.
Sensors 23 05693 g009
Figure 10. Qualitative comparisons of fusion results for a visible (RGB) and near-infrared (NIR) image pair obtained from [37,38].
Figure 10. Qualitative comparisons of fusion results for a visible (RGB) and near-infrared (NIR) image pair obtained from [37,38].
Sensors 23 05693 g010
Figure 11. Qualitative comparisons of detection performance based on different fusion models and the same detection model (YOLOv5 [30]). The source image NIR is not considered in the comparison as it lacks quantities of color information for detection, so it is shown as ground truth.
Figure 11. Qualitative comparisons of detection performance based on different fusion models and the same detection model (YOLOv5 [30]). The source image NIR is not considered in the comparison as it lacks quantities of color information for detection, so it is shown as ground truth.
Sensors 23 05693 g011
Table 1. Details of Sequoia bands.
Table 1. Details of Sequoia bands.
Band NameWavelengthBandwidthDefinition
Green550 nm40 nm1.2 Mpx
Red660 nm40 nm1.2 Mpx
Red-edge735 nm10 nm1.2 Mpx
Near-infrared790 nm40 nm1.2 Mpx
RGB 16 Mpx
Table 2. Quantitative comparisons of 13 metrics [37,38], i.e., Average Gradient (AG), Cross-Entropy (CE), Entropy Index (EI), Entropy (EN), Mutual Information (MI), Peak Signal to Noise Ratio (PSNR), Q A B / F [47], Q C B [48], Q C V [49], Root Mean Square Error (RMSE), Spatial Frequency (SF), Structural Similarity Index Measure (SSIM), and Standard Deviation (SD), on 20 pairs of near-infrared and visible images from the landmine dataset. Blue indicates the best result. Red, orange, and green represent excellent, moderate, and poor performance, respectively, of the fused image based on DDF compared with other fusion algorithms. ↓ indicates that a lower result is better; otherwise, a higher value is better.
Table 2. Quantitative comparisons of 13 metrics [37,38], i.e., Average Gradient (AG), Cross-Entropy (CE), Entropy Index (EI), Entropy (EN), Mutual Information (MI), Peak Signal to Noise Ratio (PSNR), Q A B / F [47], Q C B [48], Q C V [49], Root Mean Square Error (RMSE), Spatial Frequency (SF), Structural Similarity Index Measure (SSIM), and Standard Deviation (SD), on 20 pairs of near-infrared and visible images from the landmine dataset. Blue indicates the best result. Red, orange, and green represent excellent, moderate, and poor performance, respectively, of the fused image based on DDF compared with other fusion algorithms. ↓ indicates that a lower result is better; otherwise, a higher value is better.
AGCE↓EIENMIPSNR Q AB / F [47] Q CB [48] Q CV ↓ [49]RMSE↓SFSSIMSD
NIR7.770.3876.856.734.7263.360.360.76258.120.0316.841.1626.32
RGB19.190.25172.837.355.1563.360.720.6750.260.0346.561.1639.91
ADF [50]15.330.23135.176.960.9462.980.550.5367.190.0337.071.1430.18
CBF [51]17.430.22157.787.231.4162.870.600.5844.020.0342.171.1536.40
CNN [52]19.190.26173.177.351.1962.520.620.6049.310.0446.291.1339.88
DLF [53]11.570.32107.216.790.9263.200.390.5667.180.0327.251.2426.83
FPDE [54]10.930.36102.276.740.9063.210.360.5669.680.0325.231.2425.98
GFCE [44]19.740.52178.807.371.0561.360.510.5754.730.0547.531.1940.43
GFF [55]19.130.26173.627.351.5662.640.630.6250.900.0445.821.1139.88
IFEVIP [56]14.950.96139.827.231.0660.550.390.57188.010.0634.991.2137.23
LatLRR [43]32.011.0792.687.501.0359.110.380.59142.030.0877.570.9462.47
MGFF [57]18.800.19173.107.410.9562.480.470.61100.800.0444.611.2341.67
MST_SR [46]19.160.21173.357.351.2662.580.620.6148.880.0446.051.1339.68
MSVD [58]15.000.20132.867.000.8262.880.420.5072.120.0339.021.1730.98
NSCT_SR [46]19.130.26173.367.351.5562.640.630.6149.990.0445.911.1239.84
ResNet [59]10.720.36100.726.730.9063.200.340.5671.610.0324.721.2525.83
RP_SR [46]18.840.24167.857.321.1262.430.580.5660.420.0446.311.1438.92
TIF [60]15.570.17144.097.180.9362.740.460.6060.560.0337.351.1835.25
VSMWLS [61]18.930.16164.557.230.9362.700.510.5566.870.0347.411.1936.70
Densefuse [41]10.970.35101.206.741.0063.320.390.5768.920.0326.051.3026.01
IFCNN [42]18.140.17161.987.181.0662.970.550.5368.250.0344.461.2735.22
FusionGAN [39]15.260.23136.937.260.2560.880.440.43485.240.0537.390.9937.40
RFN-Nest [40]12.110.25119.237.110.8962.710.480.5997.020.0425.751.2233.52
U2Fusion [45]12.120.25114.427.000.3361.300.290.46337.340.0526.740.8931.53
DDF18.920.44173.877.331.4561.960.620.5845.140.0444.531.2039.31
Table 3. Quantitative comparisons of 13 metrics [37,38], i.e., Average Gradient (AG), Cross-Entropy (CE), Entropy Index (EI), Entropy (EN), Mutual Information (MI), Peak Signal to Noise Ratio (PSNR), Q A B / F [47], Q C B [48], Q C V [49], Root Mean Square Error (RMSE), Spatial Frequency (SF), Structural Similarity Index Measure (SSIM), and Standard Deviation (SD), on 21 pairs of near-infrared and visible images from the VIFB [37,38] dataset. Blue indicates the best result. Red, orange, and green represent excellent, moderate, and poor performance, respectively, of the fused image based on DDF compared with other fusion algorithms. ↓ indicates that a lower result is better; otherwise, a higher value is better.
Table 3. Quantitative comparisons of 13 metrics [37,38], i.e., Average Gradient (AG), Cross-Entropy (CE), Entropy Index (EI), Entropy (EN), Mutual Information (MI), Peak Signal to Noise Ratio (PSNR), Q A B / F [47], Q C B [48], Q C V [49], Root Mean Square Error (RMSE), Spatial Frequency (SF), Structural Similarity Index Measure (SSIM), and Standard Deviation (SD), on 21 pairs of near-infrared and visible images from the VIFB [37,38] dataset. Blue indicates the best result. Red, orange, and green represent excellent, moderate, and poor performance, respectively, of the fused image based on DDF compared with other fusion algorithms. ↓ indicates that a lower result is better; otherwise, a higher value is better.
AGCE↓EIENMIPSNR Q AB / F [47] Q CB [48] Q CV ↓ [49]RMSE↓SFSSIMSD
ADF [50]4.581.4646.536.791.9258.410.520.47777.820.1014.131.4035.19
CBF [51]7.150.9974.597.322.1657.590.580.531575.150.1320.381.1748.54
CNN [52]5.811.0360.247.322.6557.930.660.62512.570.1218.811.3960.08
DLF [53]3.821.4138.576.722.0358.440.430.45759.810.1012.491.4634.72
FPDE [54]4.541.3746.026.771.9258.400.480.46780.110.1013.471.3934.93
GFCE [44]7.501.9377.477.271.8455.940.470.53898.950.1722.461.1351.56
GFF [55]5.331.1955.207.212.6458.100.620.62881.620.1117.271.4050.06
GTF [62]4.301.2943.666.511.9957.860.440.412138.370.1214.741.3735.13
HMSD_GF [63]6.251.1665.037.272.4757.940.620.60532.960.1219.901.3957.62
Hybrid_MSD [63]6.131.2663.497.302.6258.170.640.62510.870.1119.661.4154.92
IFEVIP [56]4.981.3451.786.942.2557.170.490.46573.770.1415.851.3948.49
LatLRR [43]8.961.6892.816.911.6556.180.440.50697.290.1729.541.1857.13
MGFF [57]5.841.2960.617.111.7758.210.570.54676.890.1117.921.4144.29
MST_SR [46]5.850.9660.787.342.8157.950.660.64522.690.1218.811.3957.31
MSVD [58]3.541.4636.206.711.9558.410.330.43808.990.1012.531.4334.37
NSCT_SR [46]6.490.9067.967.402.9957.430.650.621447.340.1319.391.2852.47
ResNet [59]3.671.3637.266.731.9958.440.410.44724.830.1011.741.4634.94
RP_SR [46]6.360.9965.227.352.3457.780.570.61888.850.1221.171.3355.81
TIF [60]5.561.3757.847.081.7758.230.580.54613.000.1117.741.4042.64
VSMWLS [61]5.611.4157.257.032.0358.190.550.50754.700.1117.661.4246.25
Densefuse-L1 [41]3.541.3436.176.702.0358.440.370.44762.800.1011.021.4634.24
Densefuse-Add [41]3.541.3436.176.702.0358.440.370.44762.800.1011.021.4634.24
IFCNN-Max [42]5.851.5660.396.912.0058.040.580.47470.480.1118.671.4144.15
IFCNN-Sum [42]5.321.6254.436.841.9258.390.570.47757.320.1017.611.4536.99
IFCNN-Mean [42]5.031.6450.966.771.9158.390.530.45742.890.1016.641.4536.37
FusionGAN [39]4.251.5643.516.771.9257.970.480.47825.040.1114.561.3839.50
RFN-Nest [40]3.661.5039.427.152.0858.100.410.48829.630.1110.031.4045.36
U2Fusion [45]3.571.1238.006.922.1458.280.410.50739.270.1110.161.4640.26
DDF5.631.5258.616.962.1257.280.590.46417.230.1317.641.4549.40
Table 4. Quantitative comparisons of detection performance based on different fusion models and the same detection model (YOLOv5 [30]). Red indicates the best result. Blue represents the second-best result.
Table 4. Quantitative comparisons of detection performance based on different fusion models and the same detection model (YOLOv5 [30]). Red indicates the best result. Blue represents the second-best result.
All
TPFPPrecisionRecall[email protected]
Labels991\\\\
NIR [29]6742290.7460.6800.632
RGB [29]8401230.8720.8480.848
Decision Fuse [29]8591960.8140.8670.841
DDF898640.9330.9060.922
IFCNN [42]835930.90.8430.884
Densefuse [41]856840.9110.8640.884
U2Fusion [45]833840.9080.8410.87
RFN-Nest [40]7131340.8420.7190.728
FusionGAN [39]585590.9080.5900.611
Table 5. Time consumption of 25 typical fusion models (Unit: seconds per image pair). Red indicates the fastest. Green indicates the slowest.
Table 5. Time consumption of 25 typical fusion models (Unit: seconds per image pair). Red indicates the fastest. Green indicates the slowest.
Conventional Fusion ModelsDeep-Learning-Based Fusion Models
VIFBLandmine VIFBLandmine
ADF [50]1.002.89CNN [52]31.76117.82
CBF [51]22.9780.29DLF [53]18.6236.68
FPDE [54]2.7210.12ResNet [59]4.810.76
GFCE [44]2.136.62Densefuse [41]0.030.51
GFF [55]0.411.02IFCNN-Max [42]0.030.21
IFEVIP [56]0.170.35IFCNN-Sum [42]0.020.20
LatLRR [43]271.04910.48IFCNN-Mean [42]0.030.24
MGFF [57]1.083.33FusionGAN [39]0.380.65
MST_SR [46]0.762.20RFN-Nest [40]0.080.44
MSVD [58]1.062.27U2Fusion [45]0.040.28
NSCT_SR [46]94.65500.29DDF0.080.24
RP_SR [46]0.862.91
TIF [60]0.13 0.37
VSMWLS [61]3.5112.87
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qiu, Z.; Guo, H.; Hu, J.; Jiang, H.; Luo, C. Joint Fusion and Detection via Deep Learning in UAV-Borne Multispectral Sensing of Scatterable Landmine. Sensors 2023, 23, 5693. https://doi.org/10.3390/s23125693

AMA Style

Qiu Z, Guo H, Hu J, Jiang H, Luo C. Joint Fusion and Detection via Deep Learning in UAV-Borne Multispectral Sensing of Scatterable Landmine. Sensors. 2023; 23(12):5693. https://doi.org/10.3390/s23125693

Chicago/Turabian Style

Qiu, Zhongze, Hangfu Guo, Jun Hu, Hejun Jiang, and Chaopeng Luo. 2023. "Joint Fusion and Detection via Deep Learning in UAV-Borne Multispectral Sensing of Scatterable Landmine" Sensors 23, no. 12: 5693. https://doi.org/10.3390/s23125693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop