Next Article in Journal
Contrasting Effect of Thinning and Understory Removal on Soil Microbial Communities in a Subtropical Moso Bamboo Plantation
Next Article in Special Issue
Tropical Forest Disturbance Monitoring Based on Multi-Source Time Series Satellite Images and the LandTrendr Algorithm
Previous Article in Journal
Dynamic Evolution and Future Prediction of Land Use Patterns in the Arid Desert Region of Northwest China from 1990 to 2020
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-Supervised Learning for Forest Fire Segmentation Using UAV Imagery

1
College of Information Science and Technology, Nanjing Forestry University, Nanjing 210037, China
2
School of Engineering, University of Warwick, Coventry CV4 7AL, UK
*
Author to whom correspondence should be addressed.
Forests 2022, 13(10), 1573; https://doi.org/10.3390/f13101573
Submission received: 1 September 2022 / Revised: 18 September 2022 / Accepted: 22 September 2022 / Published: 26 September 2022

Abstract

:
Unmanned aerial vehicles (UAVs) are an efficient tool for monitoring forest fire due to its advantages, e.g., cost-saving, lightweight, flexible, etc. Semantic segmentation can provide a model aircraft to rapidly and accurately determine the location of a forest fire. However, training a semantic segmentation model requires a large number of labeled images, which is labor-intensive and time-consuming to generate. To address the lack of labeled images, we propose, in this paper, a semi-supervised learning-based segmentation network, SemiFSNet. By taking into account the unique characteristics of UAV-acquired imagery of forest fire, the proposed method first uses occlusion-aware data augmentation for labeled data to increase the robustness of the trained model. In SemiFSNet, a dynamic encoder network replaces the ordinary convolution with dynamic convolution, thus enabling the learned feature to better represent the fire feature with varying size and shape. To mitigate the impact of complex scene background, we also propose a feature refinement module by integrating an attention mechanism to highlight the salient feature information, thus improving the performance of the segmentation network. Additionally, consistency regularization is introduced to exploit the rich information that unlabeled data contain, thus aiding the semi-supervised learning. To validate the effectiveness of the proposed method, extensive experiments were conducted on the Flame dataset and Corsican dataset. The experimental results show that the proposed model outperforms state-of-the-art methods and is competitive to its fully supervised learning counterpart.

1. Introduction

Forest fire causes devastating disasters that are paroxysmal and uncontrollable. It not only reduces the forest stock, but also has a severe impact on forest growth. In addition, it is challenging to restore forests after a fire to their original state where the vegetation structure, species diversity and the ecosystem require a long time to recover [1]. Thus, it is of vital importance to monitor and detect forest fire accurately and rapidly [2].
There are currently four approaches to forest fire monitoring [3]: (1) ground patrols; (2) lookout station spotting; (3) satellite monitoring; and (4) unmanned aerial vehicle (UAV) patrols. Ground patrols are usually undertaken by forest rangers on foot or motorbike to check and monitor the implementation of rules and regulations for the prevention of forest fire, and to put out fire promptly when detected. This approach requires a large number of specialist staff and yet it is difficult to prevent forest fire. Lookout station spotting is usually located on the summit of a mountain or other high vantage points to spot a fire and determine its location. However, vegetation with varying size often obscures the spotting view. Satellite-based monitoring provides an efficient and large-scale wildfire assessment due to its wide range, time frequency and accuracy. However, this approach suffers from aerosol and cloud occlusions, which renders forest fire hardly detectable. Compared to satellite monitoring, UAV monitoring is undertaken at low-to-medium altitudes, which effectively avoids the cloud interference [4]. It also has some advantages, e.g., cost-saving, lightweight, flexible and easily manipulated. Therefore, UAV has widely been used to monitor forest fire in recent years [5].
To address the challenges (e.g., varying scene illumination and complex back-ground) in monitoring forest fire using UAV imagery, a few studies involving machine learning have been undertaken. Ko et al. introduced a support vector machine (SVM)-based fire detection method, which removes non-fire regions based on brightness maps and designed a two-class SVM classifier with a radial basis function kernel [6]. Hossain et al. attempted to solve the problem of fire point monitoring via UAV images by using a single artificial neural network (ANN) [7]. Pérez-Rodríguez et al. used ANN-based classifiers to assess the feasibility of using multispectral images acquired on drones to estimate the severity of vegetation and soil after a forest fire [8]. However, traditional machine learning methods heavily rely on the effectiveness of feature extraction and the classifier, which lack the sufficient robustness and adaptability when applied to different environments.
In recent years, convolution neural network (CNN)-based methods have been introduced for monitoring forest fire due to its superior performance on computer vision tasks [9]. Most related studies leveraged the semantic segmentation methods to accurately identify the fire spots as well as its boundary [10,11] with promising results. However, all of these methods use fully supervised learning semantic segmentation and train the segmentation model with a large amount of pixel-wise annotated data. This approach is labor intensive and time-consuming. In fact, it is not trivial to obtain the pixel-wise annotation data for training CNN-based methods, especially in UAV images of forest fire. This is because forest fire tends to be of irregular shape and varying size, rendering it difficult to identify. The complex scene background may additionally confuse the forest fire recognition system, resulting in inaccurate annotation of the forest fire regions. Furthermore, manual annotation suffers from degradation due to artificial noise, which leads to the introduction of more useless information in the training model and thus decreases the performance of the model. Semi-supervised learning [12,13,14] is regarded as one of the promising techniques to deal with the above-mentioned problems, where a small amount of labeled data with a large amount of unlabeled data are mixed for model training, thus significantly reducing the manual annotations while increasing the performance of the model.
In this paper, we propose to segment forest fire in UAV-acquired images by using semi-supervised learning, where only a few pixel-wise labeled images are required for training the model. Our application of semi-supervised learning to forest fire segmentation provides a promising direction for the development of forest fire monitoring techniques. In addition, we take into account the three challenges exclusively present in fire segmentation in UAV imagery where the target fire could: (1) be partly occluded by vegetation; (2) have varying size and shape as well as boundary that is not easy to identify; (3) be distracted by complex scene background (e.g., soil, trees, snow, etc.). To this end, we focus on addressing the challenges by proposing a semi-supervised learning-based semantic segmentation network, SemiFSNet, for remote sensing forest fire monitoring. The major contributions of this paper are summarized as follows:
(1) A semi-supervised segmentation network specifically for forest fire segmentation in UAV optical imagery.
(2) A data augmentation strategy of random dropping a grid of image pixels, effectively alleviating the impact of occlusions caused by vegetation.
(3) A dynamic convolution module to encode multi-scale features in forest fire images extracted by fusing convolution kernels of different sizes, thus increasing the segmentation accuracy.
(4) A feature refinement module that cascades an attention module to encoded feature, thus highlighting the significant information of the target forest fire. This module further enhances the feature aggregation via two different types of attentions, thus enabling the model to focus more on forest fire while mitigating the influence of complex scene background.
(5) Extensive experiments are conducted on two publicly available datasets with different environments, demonstrating the promising performance of the model and the effectiveness of each contributing module of the proposed network.
The remainder of this paper is structured as follows. Section 2 reviews the related works and presents the proposed method. The experiment results are provided in Section 3. The discussion is given in Section 4 and the conclusions are drawn in Section 5.

2. Materials and Methods

2.1. Related Works

2.1.1. Semantic Segmentation

Semantic segmentation aims to label each pixel of an image with a corresponding classification. Due to its superb performance, CNN-based image segmentation [15,16,17,18] has become the mainstream method. Long et al. [17] proposed full convolutional networks (FCN) by up-sampling (de-convolving) the output activation maps and fusing the output with the output of shallower layers. Badrinarayanan et al. [15] proposed an encoder-decoder segmentation network, SegNet, where the decoder is used to map the low-resolution feature representations to full-resolution feature maps for pixel-wise classification. As another encoder-decoder network, UNet was proposed by Ronneberger et al. [18], which involves a contracting path to extract context information and a symmetric expanding path to allow for precise localization. Chen et al. [16] proposed a segmentation architecture, DeepLabv3, which combines dilated convolutions with feature pyramid pooling. Chen2 et al. [19] extended DeepLabv3 by adding a simple yet effective refinement module in the decoder, namely, DeepLabv3+, which increases its capability in identifying object boundaries. In this paper, we use DeepLabv3+ as our basic encoder-decoder framework due to its two competitive advantages: (1) enabling to depict the multi-scale feature, which is widely existing in UAV-acquired forest fire images; and (2) significantly reducing the computational complexity, which is suitable for real applications.

2.1.2. Semi-Supervised Semantic Segmentation

Traditional semantic segmentation methods have yielded promising performances, where the data used for their training are all pixel-wise annotated (referred to as full supervised learning). The acquisition of a large amount of annotated data tends to be time-consuming and labor-intensive, and the outcome of the training heavily relies on the quality of the labels used. To deal with the above-mentioned problems, semi-supervised learning has been proposed as an effective solution [20], where only a few labeled data are needed. Alternatively, many more unlabeled data are fully exploited to optimize the learning process as a complement, where unlabeled data are much easier to obtain when compared with the labeled ones [21]. The key problems in semi-supervised semantic segmentation are: (1) how to use the few labeled images; and (2) how to make full use of the many unlabeled images. Hong et al. [22] proposed to decouple classification from segmentation and used two separate networks for each task. The label of an image is predicted using the classification network, and each predicted label is binarily segmented using the segmentation network. The decoupled structure allows one to learn, classify and segment separately using the training data with image-level and pixel-wise class labels. Olsson et al. [23] proposed ClassMix data augmentation by blending unlabeled images, where the network predictions corresponding to target boundaries are leveraged to generate labeled bounding boxes. ClassMix data augmentation enriches the dataset by making full use of the unlabeled data. Adversarial learning-based semi-supervised segmentation is proposed in [24], where a full convolutional discriminator is used to distinguish the predicted probability maps from the ground truth distribution. Mondal et al. [25] proposed another adversarial learning-based method, where cycle consistency is used to learn bidirectional mapping between unpaired images generated by CycleGAN and segmentation masks. Yang et al. [26] proposed a self-training semi-supervised semantic segmentation by integrating data augmentation on unlabeled images. Lai et al. [27] proposed to retain the consistency related to context information between features of the selected identity and accomplish consistency using directional contrastive loss, which achieves state-of-the-art performance. Hu et al. [28] proposed a new image synthesis and semi-supervised learning pipeline to train a site-specific weed detection model, enabling color matching between training and testing images, color enhancement during training, and iterative semi-supervised learning to greatly improve the performance of the model. Ke et al. [29] proposed guided collaborative training (GCT) for pixel-wise semi-supervised learning to learn additional information from the pseudo-segmentation generated by the model. However, they do not utilize generative models of the image itself, which limits the performance for the tasks requiring highly accurate pseudo-labels. French et al. [30] proposed that the network uses CutMix for semi-supervised learning and segmentation to achieve superior results on the PASCAL VOC 2012, CITYSCAPES, and ISIC 2017 datasets. However, for the aerial forest fire image, when the proportion of fire points in the whole image is small, CutMix will randomly erase a part of the pixel information on the image, which causes the reduction in the proportion of pixels containing information on the training image, thus resulting in the underfitting of the model. Although the semi-supervised segmentation has made substantive progress in recent years, only a few studies have investigated its application to remote sensing.

2.1.3. Semi-Supervised Semantic Segmentation for RS

There are a few studies exploring the application of semi-supervised learning to remote sensing. Yan et al. designed a semi-supervised method, which uses the generative adversarial network (GAN) [31] to enrich the training samples. With regard to semi-supervised semantic segmentation, Song and Yang [32] proposed the use of transfer learning and clustering for scene classification. Wang et al. [33] proposed the combination of consistency regularization and pseudo label to enable the use of unlabeled images, where thresholds are utilized to gradually [34] refine the model performance. Desai [35] proposed an active learning-based sampling strategy to select a set of highly representative labeled images for training and demonstrated the effectiveness of the proposed network on two publicly available satellite datasets. The BAS4Net model proposed by Sun et al. [36] uses the channel weighted multi-scale feature (CMF) module to balance the semantic and spatial information, and uses the boundary attention module (BAM) to weight the features with rich semantic boundary information to alleviate the boundary ambiguity. In the PiCoCo model [37], the semi-supervised segmentation of building footprint using pixel contrast and consistency learning has achieved considerable results. However, using semi-supervised learning for forest fire monitoring has rarely been reported. In this paper, we aim to bridge the gap, and address the problems (e.g., arbitrary shape and varying size of forest fire, and complex scene background) when applying semi-supervised leaning to forest fire monitoring.

2.2. Proposed Method

In this paper, we propose a semi-supervised learning network, SemiFSNet, for forest fire segmentation in UAV-acquired RGB imagery. Figure 1 shows the overall framework, which consists of supervised and unsupervised branches, where the encoder is shared for extracting the features from labeled and unlabeled inputs. First, an information deletion data enhancement strategy is performed on the labeled dataset by choosing different scales of grid to mask the images. This effectively alleviates the impact of the occlusions of fire caused by vegetation. For unlabeled data, a simple random cropping operation is applied for data augmentation. The augmented data including the labeled and unlabeled images are then input to the encoder network. To better extract the meaningful information from the forest fire region with varying shape and size, the encoder structure is constructed by replacing traditional convolution with dynamic convolution to enable the model to be automatically aware of the features at each scale of the forest fire image, thus facilitating the extraction of the multi-scale information. The encoder features are then refined by an attention module to suppress the interference caused by complex scene background. Finally, the labeled data are fed to the classifier layer and the unlabeled data are non-linearly projected to the embedded feature space, where cross entropy (CE) loss and improved directional contrastive (DC) loss are used for training, respectively.

2.2.1. Data Augmentation

Generally, the richer in information the labeled samples are, the stronger the generalization ability of the trained model. In our case, due to the limited number of labeled samples, it is essential to select an appropriate means to augment the data. Moreover, the potential occlusion of fire caused by vegetation needs to be considered. There are a few commonly used data augmentation techniques for remote sensing, such as CutOut [34], RErase [38] and HaS [39]. CutOut is performed by removing a square region of an image. RErase is used for the random deletion of a rectangle region in the image. HaS enhances the image by dividing it evenly into small chunks and then deleting them randomly. However, there is a significant risk that these data augmentation methods might remove parts of the forest fire (as illustrated in Figure 2), which degrades the model training. Inspired by the work in [40], we introduce Gridmask strategy for labeled data augmentation. The Gridmask strategy is one of the information-dropping-based data augmentation methods, which avoids excessive deletion while maintaining continuous regions.
Specifically, Gridmask generates a grid-like mask of the same size as the original image. Gridmask has only two values, 0 and 1, where 0 denotes the black-masked area and 1 denotes the transparent area. The greater the number of 1s in mask, the greater the proportion of original information that will be retained. The smaller the number of 1s in the mask, the lower the proportion of information in the original image but the higher the risk of underfitting the model. There are, in total, four parameters (i.e., r, d, δ x , δ y ) for the mesh (as shown in Figure 3) that affect the performance of extracted feature, where δ x and δ y , respectively, denote the width and height of a dropped square with value 0, r denotes the retained proportion of the input image, and d denotes the length of a square.
Given the input images X R H × W × C , M { 0 , 1 } H × W denoting the binary mask, which stores pixels to be removed, the Gridmask data augmentation multiplies X and M to obtain the augmented image Y , i.e.,
Y = X × M
According to [40], another parameter k denotes the keep ratio of a mask M, which is defined as:
k = sum ( M ) H × W .    
Ignoring incomplete units in the given mask, k is related to r by:
k = 1 ( 1 r ) 2 = 2 r r 2 .
The choice of d decides the size of a dropped square, where a small d helps to avoid failure cases. δ x and δ y control the shift of the mask. When parameters r and d are given, the change of δ x and δ y enables the mask to cover all possible situations.
In the original Gridmask [40], when the value of d is between 96 to 224, the feature accuracy extracted by the model is higher, while the value of r is generally 0.3, 0.4 and 0.5. Considering that the fire point is small in the early stage of the fire, setting r to 0.5 may lead to the phenomenon that the small fire point is blocked, so we set r to 0.1, 0.3 and 0.4, respectively. In our case, to capture the occluded forest fire with varying size and shape, we design three different scales of Gridmask, where dropped squares with different scales are generated. We randomly select δ x and δ y , ranging from 0 to d 1 , and set (r, d) for different scales of masks as (120, 0.1), (100, 0.4) and (50, 0.3). The generated Gridmask maps are then multiplied with the original image to obtain the augmented sample sets. Figure 4 illustrates the use of Gridmask in data augmentation. For the same forest fire image, Gridmask maps with three scales (r, d) of (120, 0.1), (100, 0.4) and (50, 0.3) are used to cover it, and the occlusion parts and areas of different scales of Gridmask maps in the original image are different.

2.2.2. Dynamic Encoder

Traditional convolution tends to use a single convolution kernel per layer with the kernel size fixed, leading to limited representation ability. In this paper, to increase the capability in representing forest fire, we propose the dynamic encoder module for replacing the typical fixed-size convolution with dynamic convolution, where a set of L parallel convolution kernels are introduced [41]. Dynamic convolution computes multiple attention weights for each kernel and aggregates all L kernels based on the computed weights, where the weights are calculated based on the attention operation. The size of the dynamic convolution kernel varies with the size of the feature map, and this property makes dynamic convolution suitable for the task of identifying the target with varying scales, i.e., forest fire.
Specifically, given an input x and multiple (L) linear functions { W L T x + b L } , the squeeze-and-excitation attention [42] is applied to compute the kernel attention { π L ( x ) } . Unlike the work in [42], the attentions are computed over convolution kernels instead of over output channels. After obtaining { π L ( x ) } , the aggregated weight W ( x ) and bias b ( x ) are computed based on attention weighed as follows:
W ( x ) = W 1 * π 1 ( x ) + W 2 * π 2 ( x ) + + W l * π k ( x ) = l = 1 k ( π l ( x ) W l ) ,
b ( x ) = b 1 * π 1 + b 2 * π 2 + + b k * π k = l = 1 k ( π l ( x ) b l ) ,
where W l ' and b l ' , respectively, represent the weight and bias of the dynamic perceptron, and π l ( x ) denotes the attention weight of the l th linear function { W l T x + b l } . The constraints of π l ( x ) are
0 π k ( x ) 1 ,   l = 1 k π l ( x ) = 1 .
The output of dynamic convolution y is obtained by:
y = g ( W T ( x ) x + b ( x ) ) .
where g denotes the activation function, and RELU activation function is used in g in the experiment.
This is followed by batch normalization and an activation function (ReLU in our case). Dynamical convolution automatically assembles the different kernels, increasing the model capability to extract useful information. A dynamic convolution layer is shown in Figure 5.
In our encoder network, the 3 × 3 convolutions are replaced with dynamic convolutions over every block of the backbone, i.e., Resnet50, as shown in Figure 6. By using the dynamic convolution kernel, the dynamic encoder f D E is generated and the output features F D E are defined as:
F D E = f D E ( M i x ( x l + x u ) ,
where M i x ( ) denotes the batches comprising labeled images x l and unlabeled images x u .

2.2.3. Feature Refinement Module

Forest fire segmentation using remote sensing imagery tend to be more difficult compared with other common images due to the more complex scene background. The attention mechanism is essential to enable the model to focus more on the meaningful information and ignore irrelevant information [43]. Thus, we propose a feature refinement module by integrating an attention mechanism to the output features of the encoder, which aids the model to highlight the useful fire features while reducing the interference of background information. According to [44], to obtain the more discriminative and powerful feature, the spatial and channel-wise attention is incorporated in a CNN. To this end, we propose to integrate the channel attention module (CAM) with the spatial attention module (SAM) in a concatenating way (as shown in Figure 7), where the output features are processed along two independent dimensions and thus resulting in a more discriminative representation.
Specifically, the output feature   F D E of the dynamic encoder are first aggregated using max pooling and average pooling to generate spatial descriptors F m a x C and F a v g C . The two descriptors are then fed to a shared multi-layer perceptron (MLP) network and merged using element-wise summation. The channel attention is computed by:
C A ( x ) = σ ( M L P ( A v g P o o l ( F D E ) ) + M L P ( M a x P o o l (   F D E ) ) ) ,
where σ denotes the sigmoid function, and W 1 and W 0 are the MLP. The generated channel attention is multiplied by F D E to obtain the CAM feature maps F D E , i.e.,
F D E = C A (   F D E )   F D E
After obtaining F D E , the spatial attention is computed by applying max pooling and average pooling, resulting in a descriptor F S . Such pooling operations highlight the informative forest fire regions. The output is then fed to a 7 × 7 convolution layer followed by a sigmoid function to generate spatial attention maps. The spatial attention is computed by:
S A ( F D E ' ) = σ ( Conv ( F S ) ) .
The feature maps and F D E are element-wise multiplied to obtain the SAM feature maps, i.e., the final refined features are computed by:
F = S A ( F D E ' ) F D E '

2.2.4. Consistency Regularization for Unsupervised Learning

There are two batches of inputs in our semi-supervised segmentation framework, i.e., x l and x u , denoting labeled and unlabeled images, respectively. The dynamic encoder architecture with the feature refinement module f D E + R embeds the labeled images in the feature maps F l = f D E + R ( x l ) , and the decoder (classifier) makes predictions p l = f C ( F l ) . The learning process is provided by ground-truth labels y l using the standard cross entropy loss c e . For the unlabeled images, we randomly crop the two patches x u 1 and x u 2 with an overlapping region x o , and then augment x u 1 and x u 2 using low-level augmentation. The two augmented patches are then fed to the encoder model f D E + R to obtain the feature map F u 1 and F u 2 . Following the work in [27], the two features are embedded using non-linear projection as Φ , i.e.,
ϕ u 1 = Φ ( F u 1 ) ,
ϕ u 2 = Φ ( F u 2 ) ,
As illustrated in Figure 1. The features from the overlapping areas in ϕ u 1 and ϕ u 2 are, respectively, referred to as ϕ o 1 and ϕ o 2 , where the ϕ o 1 and ϕ o 2 should remain consistent under different contexts.
A context-ware consistency constraint, i.e., DC loss [27], is used to enable the features from the overlapping areas to remain consistent with each other. The DC loss is inspired by the contrastive loss, which pulls the positive samples closer while separating the negative samples belonging to other classes. In our case, the features from overlapping locations ϕ u 1 and ϕ u 2 are regarded as a positive pair, as they share the same pixels despite being under different contexts, and any two features in ϕ u 1 and ϕ u 2 from different locations are regarded as a negative pair. Unlike contrastive loss, the DC loss further exploits a directional alignment for the positive pairs, which effectively prevents the high confident feature from suppressing the low confident one. This is necessary, as the prediction with higher confidence tends to be more accurate, and features with lower confidence need to be aligned to their counterpart with higher confident. The confidence of each feature ϕ u 1 is measured using maximum probability among all classes, i.e., max ( C ( f i ) ) . For the t-th unlabeled image, the DC loss d c t is computed as:
l d c t ϕ o 1 , ϕ o 2 = 1 N h , w M d h , w · log r ϕ o 1 h , w , ϕ o 2 h , w r ϕ o 1 h , w , ϕ o 2 h , w + ϕ n F u r ϕ o 1 h , w , ϕ n = 1 { m a x C ( f o 1 h , w ) < m a x C ( f o 2 h , w ) } ,
d c t = l d c t ( ϕ o 1 , ϕ o 2 ) + l d c t ( ϕ o 2 , ϕ o 1 ) ,  
where N is the number of spatial locations of overlapping area, h and w represent the 2-D spatial locations, ϕ n denotes the negative counterpart of feature ϕ o 1 h , w and r represents the exponential function of the cosine similarity s between two features with a temperature τ , i.e., r ( ϕ 1 , ϕ 2 ) = exp ( s ( ϕ 1 , ϕ 2 ) τ ) , and u denotes the set of negative samples. Since more negative samples result in better performance, a memory bank is used to store the features from the last few batches to acquire more negative samples [27]. The final loss is then computed by summing the loss of each image, i.e.,
d c = 1 T t = 1 T d c t ,
where T denotes the number of unlabeled images within a training batch.

2.2.5. Joint Loss Function

The joint loss function of the proposed semi-supervised-based method comprises two parts: cross entropy loss c e for supervised learning and consistency constraint loss d c for unsupervised learning. It is defined as:
= w l c e + w u n d c
where w l and w u n , respectively, denote the supervised loss weight and unsupervised loss weight, which balance the contributions of two loss parts. In our experiments, numerous experiments have shown that w l and w u n are set to 0.7 and 0.4, respectively.

3. Results

To verify the effectiveness of the proposed method SemiFSNet for forest fire segmentation in UAV images, we conducted intensive experiments on two publicly available datasets, the Flame dataset [45] and Corsican dataset [46].

3.1. Datasets

The Flame dataset [45] is created by researchers from Northern Arizona University, where the samples are collected using UAVs during the prescribed burning of piles of debris in an Arizona pine forest. The dataset includes video recordings and heat maps taken by infrared cameras. The captured videos and images are labeled by frame and aid researchers to easily build models. The dataset contains a total of 2003 forest fire images, each of which is 3840 × 2160 in size. The fire points are generally freshly started or extinguished, with the fire points taking up a relatively small proportion of the whole image.
The Corsican dataset [46] is created by the Environmental Sciences Laboratory of the University of Corsica, which includes 1136 forest fire images, each with the size of 1024 × 729 captured using an RGB camera. The background of the images varies considerably, making the segmentation difficult. To evaluate the performance in segmenting remote small fire, we selected 654 images with smaller fire spots for our experiment. Note that it is more challenging to segment small-sized fire due to its low resolution and less spatial information. Figure 8 shows some samples from the Flame dataset and Corsican dataset.

3.2. Implementation Details

DeepLabV3+ is employed as the encoder-decoder network of our proposed method due to its effectiveness in extracting multi-scale information, where ResNet50 is used as the backbone. The labeled and unlabeled images are first resized to 224 × 224, and then augmented using Gridmask and random crop. During model training, SGD optimizer is used and the learning rate, weight decay and momentum are set to 0.01, 0.0001 and 0.9, respectively. The values are gradually decreased to zero following the polynomial decay schedule. The training batch size is set to 4 and the model is trained for 100 epochs. In total, 20% of the samples from two datasets are randomly selected as the testing set, and the remaining samples as the training set. All experiments were performed on an NVIDIA RTX3090.
The intersection-over-union (IoU) for each class is employed as the evaluation metric. IoU is also known as the Jaccard Index, and is a statistic indicating the similarity and diversity of samples. In semantic segmentation, IoU denotes the ratio of the intersection of the pixel-wise classification results and ground truth, and is used to determine the spatial overlap between the prediction and ground truth, i.e.,
I o U = T P T P + F P + F N ,
where TP denotes true positives, FP denotes the false positives, TN denotes the true negatives and FN denotes false negatives. TP refers to positive samples predicted by the model as positive class, FP refers to negative samples predicted by the model as positive class, TN refers to negative samples predicted by the model as negative class, and TN refers to positive samples predicted by the model as negative class.
In multiclass semantic segmentation, mean IoU (mIoU) is mainly used as an evaluation metric. There are only two class categories in our case, i.e., forest fire and background. Since the proportion of fire area is small, and the scene background occupies the large portion of an entire image, if mIoU is used as the evaluation metric then the IoU of the background has a significant influence on the overall mIoU, which have no value. Therefore, this paper only uses the IoU of forest fire as our evaluation indicator.

3.3. Comparison with the State-of-the-Art Methods

Flame dataset. To assess the effectiveness of the proposed method, we made comparisons with state-of-the-art methods, including Adv-Semi [24], CycleGAN [25], ST++ [26], CAC [27], CCT [47], and ECS [48]. We implemented these methods within a unified framework following their official code, where the same base backbone, i.e., ResNet, is used and the same data lists are used for training and testing. We compared the proposed method under the setting with various labeled data proportions of the training set, i.e., 2/8, 3/7, 5/5 and full labeled data. In the full data setting, images fed to the unsupervised branch are simply collected from the labeled set. We ensured that all other parameters are consistent during the experiment. The experimental results on the Flame dataset are shown in Table 1.
Table 1 shows that the segmentation results using the proposed method on the Flame dataset consistently outperforms the state-of-the-art methods on all data pro-portions. This shows the superiority of the proposed method in identifying small objects. Under 2/8 data proportion, the IoU performance of the proposed method is 11.5% higher than the Adv-Semi model and 3.8% higher than the ST++. These demonstrate that the proposed method effectively addresses the situation with limited labeled data. Compared with the SupOnly (i.e., using only supervised loss), the proposed method using the supervised loss and unsupervised loss achieves a constant increase in terms of IoU under all data proportions. The idea of self-correction has been exploited in CCT and ECS by creating the Correction Network and Flaw Detector, respectively, to amend the defects in predictions. Our method not only uses self-correction, but also optimizes the entire model in the feature extraction stage to strengthen the extraction of small-target fire features. As can be seen in Table 1, our method outperforms CCT by 3.1%, 1.3% and 0.6%, and outperforms ECS by 4.1%, 2.5% and 2.5%, under three different proportions.
To visually illustrate the segmentation performance of the proposed method, we also present a visual comparison under the 2/8 proportion with the state-of-the-art methods, as shown in Figure 9.
As the target forest fire in the Flame dataset is small, we provide the enlarged comparison for clearer illustration in Figure 10. As shown in Figure 9 and Figure 10, the proposed method accurately localizes and segments the small fire spots, and the segmentation results are almost consistent with the Ground truth mask both in terms of size and boundary. The prediction results using ST++, Adv-Semi and CycleGAN have relatively large errors, and their predicted fire points are smaller than the Ground truth. The CAC predictions are closer to the Ground truth, but there are more dispersed irrelevant points. The results of CCT and ECS are close to the ground truth, but the segmentation results of our method are more refined, and the contour features are more obvious. These points are caused by the similarity between smaller fire and dead twigs or bare soil. Overall, the experiment results show that SemiFSNet improves the accuracy for the segmentation of forest fire, especially with small-sized fire.
Corsican dataset. Compared with the Flame dataset, the images in the Corsican dataset have relatively larger fire targets. We use this dataset to further evaluate the effectiveness of SemiFSNet on aerial images with large fire points by comparing its performance with state-of-the-art methods quantitatively and qualitatively. The results are shown in Table 2 and Figure 11. Once again, the proposed method consistently outperforms the other methods under all data proportion except for CycleGAN. The performance of CycleGAN is higher than the proposed method by 1.6% when the ratio of labeled and unlabeled is 2/8. However, CycleGAN training was unstable, and IoU fluctuated greatly under different proportions. In addition, the IoU of the proposed method under a different semi-supervised learning setting (i.e., 80.7%, 75.2% and 80.3%) is even far better than using a full supervised setting (i.e., 65.4%). This shows that replacing labeled images with unlabeled images has significant benefits for model learning, alleviating the negative impact of noise caused by manual labeling on model training. At the same time, the requirement of manual annotation is reduced. Figure 11 visually shows that the results using the proposed method outperform those of other methods in terms of the boundary and shape of the forest fire.

3.4. Ablation Study

To better demonstrate the efficacy of each component of SemiFSNet, an ablation study was performed on the same two datasets. There are three key components in SemiFSNet, i.e., Gridmask, Dynamic Encoder with Dynamic Convolution, and the attention-based feature refinement module. Intensive experiments were conducted using the baseline method [27] with the individual components and their combinations under different data proportions. The results are shown in Table 3 and Table 4.
As can be seen from Table 3 and Table 4, the labeled dataset is expanded by using Gridmask data augmentation so that the model achieves better segmentation results with a small amount of labeled data. Dynamic convolution leads to the increased performance on both datasets, where the features of fire with small or multiple scales are enhanced. Attention-based feature refinement significantly improves the IoU by focusing on the salient information, i.e., fire while suppressing the complex scene background. By combining the baseline with all three components, our method improves the segmentation of forest fire on the Flame dataset by up to 4.1% (for 2/8 data proportion). This demonstrates the efficacy of our method on the segmentation of small fire with a limited number of samples. Similarly, for the Corsican dataset, our SemiFSNet yields a much higher performance than the method without Gridmask, Dynamic Convolution and Feature Refinement for all data proportions. Overall, SemiFSNet using Gridmask, Dynamic Convolution and Feature Refinement considerably improves the performance of forest fire segmentation in aerial images, where a limited number of labeled samples are used for model training, thus significantly reducing the labor cost.

4. Discussion

To better show the performance of the three key components on different numbers of labeled data, we generated the line graphs in Figure 12. The figure shows that SemiFSNet integrates all three components to consistently outperform the other methods without or with the use of individual component, regardless of the number of labeled data.
The loss reflects the error between the actual and predicted values, and the loss trend reflects the state of the model learning. It is expected that both the training loss and the validation (testing) loss will gradually decrease and converge, indicating the good feature learning by the model. The trend of the training loss and the validation loss can be used to determine whether the model is overfitting or underfitting [49]. The training loss of a general semi-supervised model comprises supervised and unsupervised, i.e., supervised loss and unsupervised loss. We visualize the supervised loss, unsupervised loss and validation loss of our model under different semi-supervised allocation strategies and determine whether there are any anomalies during training and testing. Figure 13 and Figure 14 show the visualization of the loss under all allocation strategies on the Flame dataset and Corsican dataset, respectively.
Both Figure 13 and Figure 14 show that the supervised loss, unsupervised loss and validation loss display a trend of decreasing and then converging during the model training. This indicates that our model keeps adjusting the direction of convergence, making the model adjust to the best state. Judging from the convergence of the loss, the model is trustworthy and reliable.
We also compared the total parameters of our model with those of other models, and the total parameters of each model are shown in Table 5. The number of parameters of our model is 2.22 M less than that of CycleGAN’s model. Considering the stability and number of parameters of our model in the segmentation of forest fire datasets at various proportions, our model is significantly better than CycleGAN’s model. The number of parameters in our model is only 0.1 M higher than that in the CAC model and 0.2 M higher than that in the ST++ model. The whole experimental results show that the performance of our model in forest fire image segmentation is far better than those two models.

5. Conclusions

This paper focuses on solving the problem of forest fire segmentation in UAV-acquired RGB images, where semi-supervised learning techniques are applied to save human resources. We propose an effective SemiFSNet, which aims to address the challenges faced by existing remote sensing fire segmentation where the fire could: (1) be partly occluded by vegetation; (2) have varying size and shape as well as a boundary that is not easy to identify; and (3) be distracted by complex scene background. First, Gridmask data augmentation is used to increase the number of labeled images, where potential occlusions are considered by using random information deletion. The augmented images are then fed to the improved encoder, where the traditional 3 × 3 convolution is replaced with dynamic convolution to boost the model ability in extracting the fire feature with varying size and shape. The feature refinement module refines the encoder feature by integrating the channel and spatial attention and increase the robustness to complex scene background in remote sensing images. Extensive experiments conducted on two publicly available datasets show that the proposed method achieves superior performance for forest fire segmentation compared with state-of-the-art semi-supervised segmentation networks. Our method also yields a competitive result in comparison with fully supervised semantic segmentation, showing that the use of unlabeled data allows the model to automatically extract the meaningful information of the fire, effectively mitigating the influence of artificial annotation noise.
In this paper, we explore the application of semi-supervised learning to remote-sensing-based forest monitoring, which provides an efficient solution for designing a model with limited labeled data. In the future, we will continue to explore the semi-supervised learning method and increase the accuracy of the model.

Author Contributions

J.W. contributed to writing the draft, conducting experiments and visualizing the results; X.F. contributed to revising the draft and methodology; X.Y. and T.T. contributed to reviewing and editing the draft; Y.W. contributed to processing the data. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the joint fund of Science & Technology Department of Liaoning Province and State Key Laboratory of Robotics (No. 2020-KF-22-04), the National Natural Science Foundation of China (No. 61902187), and the Postgraduate Research & Practice Innovation Program of Jiangsu Province (KYCX22_1105).

Data Availability Statement

The Flame dataset is available at https://ieee-dataport.org/open-access/flame-dataset-aerial-imagery-pile-burn-detection-using-drones-uavs. The Corsican dataset is available at http://cfdb.univ-corse.fr/index.php?menu=1 (Accessed on 19 January 2022).

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Petrov, V.; Katkova, T.; Vinogradova, E. Risk Management of Forest Fire Occurrence. In IOP Conference Series: Earth and Environmental Science; IOP Publishing: Bristol, Britain, 2019; Volume 316, p. 012050. [Google Scholar]
  2. Zhan, J.; Hu, Y.; Zhou, G.; Wang, Y.; Cai, W.; Li, L. A High-Precision Forest Fire Smoke Detection Approach Based on ARGNet. Comput. Electron. Agric. 2022, 196, 106874. [Google Scholar] [CrossRef]
  3. Casbeer, D.W.; Beard, R.W.; McLain, T.W.; Li, S.-M.; Mehra, R.K. Forest Fire Monitoring with Multiple Small UAVs. In Proceedings of the 2005 American Control Conference, Portland, OR, USA, 8–10 June 2005; pp. 3530–3535. [Google Scholar]
  4. Sun, C.; Huang, C.; Zhang, H.; Chen, B.; An, F.; Wang, L.; Yun, T. Individual Tree Crown Segmentation and Crown Width Extraction from a Heightmap Derived from Aerial Laser Scanning Data Using a Deep Learning Framework. Front. Plant Sci. 2022, 13, 914974. [Google Scholar] [CrossRef] [PubMed]
  5. Guan, Z.; Miao, X.; Mu, Y.; Sun, Q.; Ye, Q.; Gao, D. Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model. Remote Sens. 2022, 14, 3159. [Google Scholar] [CrossRef]
  6. Ko, B.C.; Cheong, K.-H.; Nam, J.-Y. Fire Detection Based on Vision Sensor and Support Vector Machines. Fire Saf. J. 2009, 44, 322–329. [Google Scholar] [CrossRef]
  7. Hossain, F.A.; Zhang, Y.M.; Tonima, M.A. Forest Fire Flame and Smoke Detection from UAV-Captured Images Using Fire-Specific Color Features and Multi-Color Space Local Binary Pattern. J. Unmanned Veh. Syst. 2020, 8, 285–309. [Google Scholar] [CrossRef]
  8. Pérez-Rodríguez, L.A.; Quintano, C.; Marcos, E.; Suarez-Seoane, S.; Calvo, L.; Fernández-Manso, A. Evaluation of Prescribed Fires from Unmanned Aerial Vehicles (UAVs) Imagery and Machine Learning Algorithms. Remote Sens. 2020, 12, 1295. [Google Scholar] [CrossRef]
  9. Ghali, R.; Akhloufi, M.A.; Mseddi, W.S. Deep Learning and Transformer Approaches for UAV-Based Wildfire Detection and Segmentation. Sensors 2022, 22, 1977. [Google Scholar] [CrossRef]
  10. Barmpoutis, P.; Stathaki, T.; Dimitropoulos, K.; Grammalidis, N. Early Fire Detection Based on Aerial 360-Degree Sensors, Deep Convolution Neural Networks and Exploitation of Fire Dynamic Textures. Remote Sens. 2020, 12, 3177. [Google Scholar] [CrossRef]
  11. Ghali, R.; Akhloufi, M.A.; Jmal, M.; Mseddi, W.S.; Attia, R. Forest Fires Segmentation Using Deep Convolutional Neural Networks. In Proceedings of the 2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Melbourne, Australia, 17–20 October 2021; pp. 2109–2114. [Google Scholar]
  12. Mittal, S.; Tatarchenko, M.; Brox, T. Semi-Supervised Semantic Segmentation with High-and Low-Level Consistency. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1369–1379. [Google Scholar] [CrossRef]
  13. Wang, Z.; Liu, R. Semi-Supervised Learning for Large Scale Image Cosegmentation. In Proceedings of the 2013 IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013; pp. 393–400. [Google Scholar]
  14. Zeng, X.; Huang, R.; Zhong, Y.; Sun, D.; Han, C.; Lin, D.; Ni, D.; Wang, Y. Reciprocal Learning for Semi-Supervised Segmentation; Springer: Berlin/Heidelberg, Germany, 2021; pp. 352–361. [Google Scholar]
  15. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  16. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  17. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  20. Papandreou, G.; Chen, L.-C.; Murphy, K.P.; Yuille, A.L. Weakly-and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1742–1750. [Google Scholar]
  21. Güldenring, R.; Nalpantidis, L. Self-Supervised Contrastive Learning on Agricultural Images. Comput. Electron. Agric. 2021, 191, 106510. [Google Scholar] [CrossRef]
  22. Hong, S.; Noh, H.; Han, B. Decoupled Deep Neural Network for Semi-Supervised Semantic Segmentation. Adv. Neural Inf. Processing Syst. 2015, 28, 1495–1503. [Google Scholar]
  23. Olsson, V.; Tranheden, W.; Pinto, J.; Svensson, L. Classmix: Segmentation-Based Data Augmentation for Semi-Supervised Learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2021, Waikoloa, HI, USA, 3–8 January 2021; pp. 1369–1378. [Google Scholar]
  24. Hung, W.-C.; Tsai, Y.-H.; Liou, Y.-T.; Lin, Y.-Y.; Yang, M.-H. Adversarial Learning for Semi-Supervised Semantic Segmentation. arXiv 2018, arXiv:1802.07934. [Google Scholar]
  25. Mondal, A.K.; Agarwal, A.; Dolz, J.; Desrosiers, C. Revisiting CycleGAN for Semi-Supervised Segmentation. arXiv 2019, arXiv:1908.11569. [Google Scholar]
  26. Yang, L.; Zhuo, W.; Qi, L.; Shi, Y.; Gao, Y. ST++: Make Self-Training Work Better for Semi-Supervised Semantic Segmentation. arXiv 2021, arXiv:2106.05095. [Google Scholar]
  27. Lai, X.; Tian, Z.; Jiang, L.; Liu, S.; Zhao, H.; Wang, L.; Jia, J. Semi-Supervised Semantic Segmentation with Directional Context-Aware Consistency. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 1205–1214. [Google Scholar]
  28. Hu, C.; Thomasson, J.A.; Bagavathiannan, M.V. A Powerful Image Synthesis and Semi-Supervised Learning Pipeline for Site-Specific Weed Detection. Comput. Electron. Agric. 2021, 190, 106423. [Google Scholar] [CrossRef]
  29. Ke, Z.; Qiu, D.; Li, K.; Yan, Q.; Lau, R.W. Guided Collaborative Training for Pixel-Wise Semi-Supervised Learning; Springer: Berlin/Heidelberg, Germany, 2020; pp. 429–445. [Google Scholar]
  30. French, G.; Laine, S.; Aila, T.; Mackiewicz, M.; Finlayson, G. Semi-Supervised Semantic Segmentation Needs Strong, Varied Perturbations. arXiv 2019, arXiv:1906.01916. [Google Scholar]
  31. Yan, P.; He, F.; Yang, Y.; Hu, F. Semi-Supervised Representation Learning for Remote Sensing Image Classification Based on Generative Adversarial Networks. IEEE Access 2020, 8, 54135–54144. [Google Scholar] [CrossRef]
  32. Song, H.; Yang, W. GSCCTL: A General Semi-Supervised Scene Classification Method for Remote Sensing Images Based on Clustering and Transfer Learning. Int. J. Remote Sens. 2022, 1–25. [Google Scholar] [CrossRef]
  33. Wang, J.-X.; Chen, S.-B.; Ding, C.H.; Tang, J.; Luo, B. RanPaste: Paste Consistency and Pseudo Label for Semisupervised Remote Sensing Image Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 2002916. [Google Scholar] [CrossRef]
  34. DeVries, T.; Taylor, G.W. Improved Regularization of Convolutional Neural Networks with Cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  35. Desai, S.; Ghose, D. Active Learning for Improved Semi-Supervised Semantic Segmentation in Satellite Images. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 4–8 January 2022; pp. 553–563. [Google Scholar]
  36. Sun, X.; Shi, A.; Huang, H.; Mayer, H. BAS4Net: Boundary-Aware Semi-Supervised Semantic Segmentation Network for Very High Resolution Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5398–5413. [Google Scholar] [CrossRef]
  37. Kang, J.; Wang, Z.; Zhu, R.; Sun, X.; Fernandez-Beltran, R.; Plaza, A. PiCoCo: Pixelwise Contrast and Consistency Learning for Semisupervised Building Footprint Segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10548–10559. [Google Scholar] [CrossRef]
  38. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random Erasing Data Augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13001–13008. [Google Scholar]
  39. Singh, K.K.; Yu, H.; Sarmasi, A.; Pradeep, G.; Lee, Y.J. Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization and Beyond. arXiv 2018, arXiv:1811.02545. [Google Scholar]
  40. Chen, P.; Liu, S.; Zhao, H.; Jia, J. Gridmask data augmentation. arXiv 2020, arXiv:2001.04086. [Google Scholar]
  41. Chen, Y.; Dai, X.; Liu, M.; Chen, D.; Yuan, L.; Liu, Z. Dynamic Convolution: Attention over Convolution Kernels. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11030–11039. [Google Scholar]
  42. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  43. Kang, J.; Liu, L.; Zhang, F.; Shen, C.; Wang, N.; Shao, L. Semantic Segmentation Model of Cotton Roots In-Situ Image Based on Attention Mechanism. Comput. Electron. Agric. 2021, 189, 106370. [Google Scholar] [CrossRef]
  44. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional Block Attention Module. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  45. Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P.Z.; Blasch, E. Aerial Imagery Pile Burn Detection Using Deep Learning: The FLAME Dataset. Comput. Netw. 2021, 193, 108001. [Google Scholar] [CrossRef]
  46. Toulouse, T.; Rossi, L.; Campana, A.; Celik, T.; Akhloufi, M.A. Computer Vision for Wildfire Research: An Evolving Image Dataset for Processing and Analysis. Fire Saf. J. 2017, 92, 188–194. [Google Scholar] [CrossRef] [Green Version]
  47. Ouali, Y.; Hudelot, C.; Tami, M. Semi-supervised semantic segmentation with cross-consistency training. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12674–12684. [Google Scholar]
  48. Mendel, R.; de Souza, L.A.; Rauber, D.; Papa, J.P.; Palm, C. Semi-supervised segmentation based on error-correcting supervision. In Computer Vision—ECCV 2020, Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020; pp. 141–157. [Google Scholar]
  49. Jadon, S. A Survey of Loss Functions for Semantic Segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Via del Mar, Chile, 27–29 October 2020; pp. 1–7. [Google Scholar]
Figure 1. Overall structure of the proposed SemiFSNet.
Figure 1. Overall structure of the proposed SemiFSNet.
Forests 13 01573 g001
Figure 2. Images using different data augmentation techniques.
Figure 2. Images using different data augmentation techniques.
Forests 13 01573 g002
Figure 3. The dotted square denotes one unit of the mask.
Figure 3. The dotted square denotes one unit of the mask.
Forests 13 01573 g003
Figure 4. Gridmask with different scales for data augmentation.
Figure 4. Gridmask with different scales for data augmentation.
Forests 13 01573 g004
Figure 5. Dynamic convolution structure diagram.
Figure 5. Dynamic convolution structure diagram.
Forests 13 01573 g005
Figure 6. Replacing a 3 × 3 convolution with a dynamic convolution in Resnet50.
Figure 6. Replacing a 3 × 3 convolution with a dynamic convolution in Resnet50.
Forests 13 01573 g006
Figure 7. Feature refinement module with CAM and SAM.
Figure 7. Feature refinement module with CAM and SAM.
Forests 13 01573 g007
Figure 8. Selected samples from Flame dataset and Corsican dataset.
Figure 8. Selected samples from Flame dataset and Corsican dataset.
Forests 13 01573 g008
Figure 9. Visualization of segmentation results using the proposed method and state-of-art methods on the Flame dataset, where subfigure (a) shows the image of a single fire point taken by UAV from high altitude, subfigure (b) shows the image of fire point taken by UAV from low altitude, and subfigure (c) shows the image of multiple fire points taken by UAV.
Figure 9. Visualization of segmentation results using the proposed method and state-of-art methods on the Flame dataset, where subfigure (a) shows the image of a single fire point taken by UAV from high altitude, subfigure (b) shows the image of fire point taken by UAV from low altitude, and subfigure (c) shows the image of multiple fire points taken by UAV.
Forests 13 01573 g009
Figure 10. Enlarged visualization of segmentation results using the proposed method and state-of-art methods on the Flame dataset.
Figure 10. Enlarged visualization of segmentation results using the proposed method and state-of-art methods on the Flame dataset.
Forests 13 01573 g010
Figure 11. Visualization of segmentation results using the proposed method and state-of-art methods on the Corsican dataset, where (ac) shows the samples with different background.
Figure 11. Visualization of segmentation results using the proposed method and state-of-art methods on the Corsican dataset, where (ac) shows the samples with different background.
Forests 13 01573 g011
Figure 12. IoU for forest fire segmentation under various numbers of labeled images.
Figure 12. IoU for forest fire segmentation under various numbers of labeled images.
Forests 13 01573 g012
Figure 13. Loss curve for supervise training loss, unsupervised training loss and validation loss on the Flame dataset.
Figure 13. Loss curve for supervise training loss, unsupervised training loss and validation loss on the Flame dataset.
Forests 13 01573 g013
Figure 14. Loss curve for supervise training loss, unsupervised training loss and validation loss on the Corsican dataset.
Figure 14. Loss curve for supervise training loss, unsupervised training loss and validation loss on the Corsican dataset.
Forests 13 01573 g014
Table 1. Comparison with the baseline (SupOnly, i.e., using only supervised loss) and other state-of-the-art methods on Flame dataset with 2/8, 3/7, 5/5 and full labeled data.
Table 1. Comparison with the baseline (SupOnly, i.e., using only supervised loss) and other state-of-the-art methods on Flame dataset with 2/8, 3/7, 5/5 and full labeled data.
NetworksThe Ratio of Labeled and Unlabeled for Training
2/83/75/5Full
CAC62.5%65.4%67.7%69.8%
ST++60.6%67.6%67.9%N/A
Adv-semi52.9%52.6%49.7%N/A
CycleGAN49.7%55.4%63.6%N/A
CCT61.3%64.1%68.4%N/A
ECS60.3%62.8%66.5%N/A
Ours64.4%65.4%69.0%N/A
Table 2. Comparison with the baseline and other state-of-the-art methods on Corsican dataset with 2/8, 3/7, 5/5 and full labeled data.
Table 2. Comparison with the baseline and other state-of-the-art methods on Corsican dataset with 2/8, 3/7, 5/5 and full labeled data.
NetworksThe Ratio of Labeled and Unlabeled for Training
2/83/75/5Full
CAC70.5%75.9%76.2%65.4%
ST++63.1%78.6%77.1%N/A
Adv-semi73.9%76.9%80.1%N/A
CycleGAN69.7%73.9%80.0%N/A
CCT68.3%70.4%75.3%N/A
ECS70.2%72.7%75.4%N/A
Ours80.7%75.2%80.3%N/A
Table 3. Results of ablation study on the Flame dataset. (✔ denotes the module is included in the method.)
Table 3. Results of ablation study on the Flame dataset. (✔ denotes the module is included in the method.)
Allocation StrategyBaselineGridmaskDynamic-ConvolutionCAM + SAMIOU
2/8 60.3%
61.1%
62.5%
63.1%
63.3%
63.2%
64.2%
64.4%
3/7 61.5%
62.5%
61.9%
62.3%
63.7%
64.2%
65.2%
65.4%
5/5 64.7%
64.9%
65.6%
67.3%
66.4%
67.3%
68.7%
69.0%
Table 4. Results of ablation study on the Corsican dataset. (✔ denotes the module is included in the method.)
Table 4. Results of ablation study on the Corsican dataset. (✔ denotes the module is included in the method.)
Allocation StrategyBaselineGridmaskDynamic-ConvolutionCAM + SAMIOU
2/8 78.0%
78.6%
79.2%
79.4%
79.3%
79.4%
80.1%
80.7%
3/7 73.2%
74.0%
73.8%
73.6%
74.2%
74.7%
74.8%
75.2%
5/5 77.9%
78.2%
78.8%
78.0%
79.2%
79.6%
80.0%
80.3%
Table 5. Parameters of each model.
Table 5. Parameters of each model.
CACST++Adv-SemiCycle-GANOurs
Parameters40.4 M40.3 M44.05 M42.72 M40.5 M
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Fan, X.; Yang, X.; Tjahjadi, T.; Wang, Y. Semi-Supervised Learning for Forest Fire Segmentation Using UAV Imagery. Forests 2022, 13, 1573. https://doi.org/10.3390/f13101573

AMA Style

Wang J, Fan X, Yang X, Tjahjadi T, Wang Y. Semi-Supervised Learning for Forest Fire Segmentation Using UAV Imagery. Forests. 2022; 13(10):1573. https://doi.org/10.3390/f13101573

Chicago/Turabian Style

Wang, Junling, Xijian Fan, Xubing Yang, Tardi Tjahjadi, and Yupeng Wang. 2022. "Semi-Supervised Learning for Forest Fire Segmentation Using UAV Imagery" Forests 13, no. 10: 1573. https://doi.org/10.3390/f13101573

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop