Next Article in Journal
Multi-Featured Sea Ice Classification with SAR Image Based on Convolutional Neural Network
Previous Article in Journal
Monitoring and Comparative Analysis of Hohhot Subway Subsidence Using StaMPS-PS Based on Two DEMS
Previous Article in Special Issue
Attention-Aware Spectral Difference Representation for Hyperspectral Anomaly Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-FCMNet: Semi-Supervised Learning for Forest Cover Mapping from Satellite Imagery via Ensemble Self-Training and Perturbation

1
College of Information Science and Technology, Nanjing Forestry University, Nanjing 210037, China
2
School of Engineering, University of Warwick, Coventry CV4 7AL, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2023, 15(16), 4012; https://doi.org/10.3390/rs15164012
Submission received: 5 July 2023 / Revised: 5 August 2023 / Accepted: 9 August 2023 / Published: 13 August 2023
(This article belongs to the Special Issue Machine Learning with Extremely Few Annotations for Remote Sensing)

Abstract

:
Forest cover mapping is of paramount importance for environmental monitoring, biodiversity assessment, and forest resource management. In the realm of forest cover mapping, significant advancements have been made by leveraging fully supervised semantic segmentation models. However, the process of acquiring a substantial quantity of pixel-level labelled data is prone to time-consuming and labour-intensive procedures. To address this issue, this paper proposes a novel semi-supervised-learning-based semantic segmentation framework that leverages limited labelled and numerous unlabelled data, integrating multi-level perturbations and model ensembles. Our framework incorporates a multi-level perturbation module that integrates input-level, feature-level, and model-level perturbations. This module aids in effectively emphasising salient features from remote sensing (RS) images during different training stages and facilitates the stability of model learning, thereby effectively preventing overfitting. We also propose an ensemble-voting-based label generation strategy that enhances the reliability of model-generated labels, achieving smooth label predictions for challenging boundary regions. Additionally, we designed an adaptive loss function that dynamically adjusts the focus on poorly learned categories and dynamically adapts the attention towards labels generated during both the student and teacher stages. The proposed framework was comprehensively evaluated using two satellite RS datasets, showcasing its competitive performance in semi-supervised forest-cover-mapping scenarios. Notably, the method outperforms the fully supervised approach by 1–3% across diverse partitions, as quantified by metrics including mIoU, accuracy, and mPrecision. Furthermore, it exhibits superiority over other state-of-the-art semi-supervised methods. These results indicate the practical significance of our solution in various domains, including environmental monitoring, forest management, and conservation decision-making processes.

1. Introduction

Forests play a vital role in the land ecosystem of the Earth. They are indispensable for conserving biodiversity, protecting watersheds, capturing carbon, mitigating climate change effects [1,2], maintaining ecological balance, regulating rainfall patterns, and ensuring the stability of large-scale climate systems [3,4]. As a result, the timely and precise monitoring and mapping of forest cover has emerged as a vital aspect of sustainable forest management and the monitoring of ecosystem transformations [5].
Traditionally, the monitoring and mapping of forest cover has primarily relied on field research and photo-interpretation techniques. However, these methods are limited by the extensive manpower required. With the advancements in remote sensing (RS) technology, the acquisition of large-scale, high-resolution forest imagery data has become possible without the need for physical contact and without causing harm to the forest environment. Taking advantage of RS imagery, numerous studies have proposed various methods for forest cover mapping, including decision trees [6], regression trees [7], maximum likelihood classifiers [8], random forest classification algorithms [9,10], support vector machines, spatio-temporal Markov random-field super-resolution mapping [11], and multi-scale spectral–spatial–temporal super-resolution mapping [12].
Recently, due to the growing prevalence of deep Convolutional Neural Networks (CNNs) [13] and semantic segmentation [14,15,16], there has been a notable shift in the research community towards utilising these techniques for forest cover mapping with RS imagery. CNNs have emerged as powerful tools for analysing two-dimensional images, employing their multi-layered convolution operations to effectively capture low-level spatial patterns (such as edges, textures, and shapes) and extract high-level semantic information. Meanwhile, semantic segmentation techniques enable the precise identification and extraction of different objects/regions in an image by classifying each image pixel into a specific semantic category, achieving pixel-level image segmentation. In the context of forest cover mapping, several existing methods have demonstrated the effectiveness of using semantic segmentation techniques. Bragagnolo et al. [17] proposed to integrate an attention block into the basic UNet network to segment the forest area using satellite imagery from South America. Flood et al. [18] also proposed a UNet-based network and achieved promising results in mapping the presence or absence of trees and shrubs in Queensland, Australia. Isaienkov et al. [19] directly employed the baseline U-Net model combined with Sentinel-2 satellite data to detect changes in Ukrainian forests. However, all these methods rely on fully supervised learning for semantic segmentation, which necessitates a substantial amount of labelled pixel data, resulting in a significant labelling expense. Semi-supervised learning [20,21,22] has emerged as a promising approach to address the aforementioned challenges. It involves training models using a combination of limited labelled data and a substantial amount of unlabelled data, which reduces the need for manual annotations while still improving the performance of the model. Several research studies [23,24] have explored the application of semi-supervised learning in semantic segmentation for land-cover-mapping tasks in RS. While these studies have assessed the segmentation of forests to some extent, their focus has predominantly been on forests situated in urban or semi-natural areas, limiting their performance in densely forested natural areas. Moreover, there are unique challenges associated with utilising satellite RS imagery specifically for forest cover mapping:
Challenge 1: As illustrated in Figure 1a, satellite remote sensing (RS) forest images often face problems such as variations in scene illumination and atmospheric interference during the image acquisition process. These factors can lead to colour deviations and distortions, resulting in poor colour fidelity and low contrast. Therefore, it becomes essential to employ image enhancement techniques to improve visualisation and reveal more details. This enhancement facilitates the ability of CNNs to effectively capture spatial patterns.
Challenge 2: Due to the high density of natural forest cover and the similar reflectance characteristics between forest targets and other non-forest targets, e.g., grass and the shadow of vegetation, the boundaries between forest and non-forest areas often become unclear. As a result, it becomes challenging to accurately distinguish and delineate the details and edges of the regions of interest, as depicted in Figure 2a.
Challenge 3: For unknown forest distributions, there are two scenarios: imbalanced (as illustrated in Figure 2) or balanced datasets. Current methods face challenges in effectively handling datasets with different distributions, resulting in poor model generalisation.
In this study, we have undertaken a pioneering endeavour to integrate semi-supervised learning techniques into forest cover mapping using satellite RS imagery. We propose a novel semi-supervised semantic segmentation network called Semi-FCMNet, designed to effectively tackle the associated challenges in this task. To tackle challenge 1, which encompasses image distortion issues in RS images, we employ the concept of multi-level perturbations. Perturbations are designed at different stages, namely, at the input level, feature level, and model level. At the input level, perturbations utilising mixed augmentation techniques are employed to enhance the representation of forest features within the image. The feature-level and model-level perturbations facilitate model learning and capture valuable information during training. Importantly, these perturbations effectively counteract the overfitting of the model to noise. For challenge 2, we combine the auxiliary teacher module with the Test-Time Augmentation (TTA) approach to integrate the generated pseudo-labels through voting and multi-scale fusion, enhancing the reliability and clarity of edge information. To address challenge 3, we introduce an adaptive loss function that automatically focuses on under-learned classes and adjusts the attention towards labels generated by both the student and teacher models. This approach enables our model to achieve excellent performance on both balanced and imbalanced datasets. The adaptive loss function effectively addresses the issue of insufficient learning in certain classes and improves the capability of the model to handle diverse data distributions. Furthermore, we adopt a progressive learning approach and design a data augmentation strategy from easy to difficult, employing different intensity levels of data augmentation for models at different stages. The code is publicly available at https://github.com/baizegugugu/Semi-FCMNet.
The primary contributions of this paper are summarised as follows:
  • We have designed the multi-level perturbation (MP) module, including input-, feature- and model-level perturbations at different module stages. The proposed approach incorporates perturbations at the input stage to enhance forest representation features by using mixed augmentation. Additionally, the auxiliary teacher module introduces perturbations at both the feature and model levels, allowing the model to concentrate on feature disparities and proficiently learn forest characteristics while effectively mitigating the overfitting problem to noise.
  • By integrating auxiliary teachers with the student model, the basic self-training method was enhanced. To generate more stable and reliable pseudo-labels during the pseudo-labelling phase, we introduced a novel ensemble voting (EV) module, smoothing the decision-making process for challenging boundary regions. This module leverages a combination of multiple models and adopts a strategy based on TTA and multi-model voting.
  • We have developed a simple yet effective adaptive loss (AL) that enables the model to adapt to both balanced and imbalanced data distributions while also increasing its focus on labels generated by the teacher. By incorporating AL into the training process, our model demonstrates robust performance across different data scenarios.
This paper is organised as follows. Section 2 reviews semi-supervised semantic segmentation, and Section 3 provides a detailed description of the proposed framework. Section 4.1 presents detailed information on the data distribution of the two datasets we used (Atlantic Forest and Amazon Forest), while Section 4.2 and Section 4.3 introduce the relevant parameters set in our experiments and the metrics used to verify the experimental results, respectively. Section 4.4 compares and analyses our methods with the SOTA methods on the two datasets, while Section 4.5 presents the ablation experimental results for our method to validate its effectiveness. Finally, Section 5 outlines the limitations of our method, future research directions, and application prospects.

2. Related Work

2.1. Semi-Supervised Semantic Segmentation

Consistency regularisation and pseudo-labelling are two main categories of methods in the field of semi-supervised semantic segmentation [21,22].
Consistency regularisation methods aim to improve model performance by promoting consistency among diverse predictions for the same image. This is achieved by introducing perturbations to either the input images or the models themselves. For instance, the CCT approach [25] utilises an auxiliary decoder structure that incorporates multiple robust perturbations at both the feature level and decoder output stage. These perturbations are strategically employed to enforce consistency in the model predictions, ensuring that the predictions remain consistent even in the presence of perturbations. Furthermore, Liu et al. proposed PS-MT [26], which introduces innovative extensions to improve upon the mean-teacher (MT) model. These extensions include the introduction of an auxiliary teacher and the replacement of the MT’s mean square error (MSE) loss with a more stringent confidence-weighted cross-entropy (Conf-CE) loss. These enhancements greatly enhance the accuracy and consistency of the predictions in the MT model. Building upon these advancements, Abulikemu Abuduweili et al. [27] proposed a novel method that leverages adaptive consistency regularisation to effectively combine pre-trained models and unlabelled data, improving model performance. However, consistency regularisation typically relies on perturbation techniques that impose certain requirements on dataset quality and distribution, as well as numerous challenging hyperparameters to fine-tune.
Contrary to the consistency regularisation methods, pseudo-labelling techniques, exemplified by self-training [28], leverage predictions from unlabelled data to generate pseudo-labels, which are then incorporated into the training process, effectively expanding the training set and enriching the model with more information. In this context, Yi Zhu et al. [29] proposed a self-training framework for semantic segmentation that utilises pseudo-labels from unlabelled data, addressing data imbalance through centroid sampling and focusing on optimising computational efficiency. However, pseudo-labelling methods offer more stable training but have limitations in achieving substantial improvements in model performance, leading to under-utilisation of the potential of unlabelled data.
In the task of forest cover mapping using RS imagery, we propose an integrated framework that combines self-training and consistency regularisation methods. This approach effectively enhances the performance of the model by effectively utilising the abundant unlabelled data available in the context of forest RS.

2.2. Semi-Supervised Semantic Segmentation in RS

Several studies have investigated the application of semi-supervised learning in RS. Lucas et al. [23] proposed Sourcerer, a deep-learning-based technique for semi-supervised domain adaptation in land cover mapping from satellite image time-series data. Sourcerer surpasses existing methods by effectively leveraging labelled data from a source domain and adapting the model to the target domain using a novel regulariser, even with limited labelled target data. Chen et al. [30] introduced SemiRoadExNet, a novel semi-supervised road extraction network based on a Generative Adversarial Network (GAN). The network efficiently utilises both labelled and unlabelled data by generating road segmentation results and entropy maps. Zou et al. [31] introduced a novel pseudo-labelling approach for semantic segmentation, improving the training process using unlabelled or weakly labelled data. Through the intelligent combination of diverse sources and robust data augmentation, the proposed strategy demonstrates effective consistency training, showing its effectiveness for data of low or high density. On the other hand, Zhang et al. [32] proposed a semi-supervised deep learning framework for the semantic segmentation of high-resolution RS images. The framework utilises transformation consistency regularisation to make the most of limited labelled samples and abundant unlabelled data. However, the application of semi-supervised semantic segmentation methods in forest cover mapping has not been explored. In this paper, we aim to address the challenges of forest cover mapping in high-density RS imagery and investigate the application of semi-supervised semantic segmentation methods in this context.

3. Method

3.1. Problem Definition

Semi-supervised semantic segmentation is a method of semantic segmentation that uses labelled and unlabelled data. Compared to traditional fully supervised semantic segmentation methods, it does not require a large amount of labelled data to train the model, making computation more economical and efficient. The main idea of semi-supervised semantic segmentation is to enhance the generalisation ability of the model by utilising unlabelled data. Specifically, semi-supervised semantic segmentation seeks to generalise from a combined dataset consisting of pixel-wise labelled images D l = { ( x i , y i ) } i = 1 M unlabelled images D u = { u i } i = 1 N , where, typically, N > > M . In the majority of studies, the overall optimisation objective is formulated as
L = L s + λ L u ,
where λ serves as a tradeoff between labelled and unlabelled data. The parameter λ can either be a fixed value or be scheduled during training. The unsupervised loss L u is a crucial aspect that distinguishes various semi-supervised methods, whereas the supervised loss L s typically refers to the cross-entropy loss between predictions and manually annotated masks.

3.2. Auxiliary Mean Teachers and Student Models

Although many more advanced models and methods have emerged, classical methods such as self-training and mean teachers can also perform well with improvements. To further improve the performance of the model based on the self-training method, we incorporate an improved mean-teacher mechanism. Figure 3 shows the architecture of our model. The model uses the classical encoder–decoder model, using ResNet-101 as the encoder for better pixel information extraction and restoration and using DeeplabV3+ as the decoder. All auxiliary teachers and students share the same structure, and both of the two auxiliary teachers receive exponential moving average (EMA) transfers of parameters from different epochs of the student, as shown in Figure 4, i.e.,
θ k = γ · θ k + ( 1 γ ) · θ s ,
where k t a 1 , t a 2 , γ ( 0 , 1 ) . For the training of teacher models, we update the parameters of only one of the two teachers at each training epoch.

3.3. Training Strategy: Self-Training (ST)

Currently, the perturbation method based on consistency regularisation is widely considered the most effective approach to semi-supervised learning. This method involves perturbing the image during data augmentation and improving the performance of the model by constraining the similarity between its final predicted outputs. These perturbations mostly focus on perturbing the image representation, such as adding noise, random dropout, cutout, and CutMix. However, the unprocessed RGB colour features of the forest satellite RS dataset are not obvious, and the data distribution is imbalanced. Random perturbations at the image level make it difficult for the classes with few samples to be fully learned, while the classes with many samples are constrained by the loss function, making it difficult for the model to fully learn from those samples, resulting in poor model performance. At the same time, the perturbation method based on consistency regularisation requires manually adjusting the weights of unsupervised loss and supervised loss, and it is difficult to adjust the proportion for different weights, leading to a further decline in model performance.
Therefore, in order to fully utilise the information in the dataset and reduce the settings of hyperparameters, we primarily adopt a method based on self-training, as shown in Figure 5. Firstly, we train the teacher model t l on the labelled dataset D l , and then t l is used to assign pseudo-labels to the remaining unlabelled samples in D u . Finally, the pseudo-labels are used as the labels for the unlabelled images, and the student model s is trained on the entire dataset D a . During the training process, auxiliary teachers are introduced to more stably evaluate and predict unlabelled images.
The pseudocode of our ST framework is shown in Algorithm 1.
Algorithm 1 ST with perturbation
Require: Labelled training set D l = { ( x i , y i ) } i = 1 M ,
    Unlabelled training set D u = { u i } i = 1 N ,
    Strong/VAT/weak augmentations A s / A V / A w ,
    Teacher/student model t l , t a 1 , t a 2 / s
Ensure: Student model s
  1:
for minibatch { ( x i , y i ) } i = 1 M D l  do
  2:
for k ∈ { 1 , . . . , M }  do
  3:
   x k , y k A V ( A w ( x k , y k ) )
  4:
   y k ^ = t l ( x k )
  5:
  Update t l to minimise L c e and L f o c a l of { ( y k ^ , y k ) } k = 1 M
  6:
  Update t a 1 , t a 2 with EMA of s
  7:
   end for
  8:
end for
  9:
D u ^ = Label( D u )
10:
for minibatch { ( x i , y i ) } i = 1 B ( D l D u ^ )  do
11:
for k ∈ { 1 , . . . , B }  do
12:
   x k , y k A V ( A s ( A w ( x k , y k ) ) )
13:
   y k ^ = s ( x k )
14:
   y k 2 ^ = 0.5 ( t a 1 ( x k ) + t a 2 ( x k ) )
15:
  Update s to minimise L c e and L f o c a l of { ( y k ^ , y k ) } k = 1 B
16:
  Update s to minimise L c e and L f o c a l of { ( y k 2 ^ , y k ^ ) } k = 1 B
17:
  Update t a 1 , t a 2 with EMA of s
18:
end for
19:
end for
20:
return s
In this training strategy, strong augmentation refers to data transformations that alter the colour, contrast, and other properties of the image, such as colorJitter, random greyscale, blur, etc. On the other hand, weak augmentation pertains to transformations like resizing, cropping, and horizontal flipping that do not modify the main features of the original image. It is important to note that we employ weak augmentation during the supervised phase, whereas in the unsupervised phase, we adopt a combination of strong and weak augmentations. Both of these types of augmentation fall under input-level image perturbation and, together with subsequent transformations like VAT (feature-level perturbation), constitute multi-level perturbation.

3.4. Multi-Level Perturbation (MP)

In the process of model learning, the most basic self-training method will overfit the errors during iteration and reduce the performance of the student model. To better capture the intrinsic information of forest images and mitigate the risk of overfitting incorrect labels, we propose a multi-level perturbation strategy including input-level image perturbation, feature-level perturbation, and model-level perturbation.

3.4.1. Input-Level Image Perturbation

Given the limited prominence of colour features and the difficulty in learning image features, we opted for a mixed-image augmentation as input-level perturbation for fully supervised and unsupervised learning, which allows the model to prioritise the overall image rather than focusing solely on partial regions. Specifically (as shown in Figure 3), in the fully supervised stage, we applied weak augmentations, including resize, crop, and horizontal flip transformations, to the input images. In the unsupervised stage, we employed strong augmentations, including colorJitter, random greyscale, and blur, as well as random cutout and contrast/colour-filtering techniques, on the input images.

3.4.2. Feature-Level Perturbation VAT

Feature perturbation involves the creation of adversarial perturbations that pose a challenge by intentionally disrupting the cluster or low-density assumption. This is achieved by transforming the image features, computed from the model encoder, towards the classification boundaries within the feature space. One effective approach to generating such adversarial feature noise is through virtual adversarial training (VAT), which optimises a perturbation vector to maximise the divergence between correct and adversarial classifications. However, current methods estimate the adversarial noise by using the same single network where the consistency loss is applied. Thus, we suggest estimating the adversarial noise using the more accurate teachers and then applying this estimated noise to the feature of the student model, which we call VAT feature perturbation. In particular, we used the VAT feature enhancement method in both the fully supervised and semi-supervised stages and achieved significant improvements. VAT is used in the student model output, i.e.,
p s ( x ) = d e c o d e r s ( e n c o d e r s ( x ) + r a d v ) ,
where p s is the prediction result of the student model for each pixel, and e n c o d e r s and d e c o d e r s are, respectively, the encoder and decoder of the student model. The adversarial feature perturbation r a d v Z is estimated from the response of the ensemble of teacher models using
max d 0.5 · ( d e c o d e r a 1 ( z s ) + d e c o d e r a 2 ( z s ) ) , 0.5 · ( d e c o d e r a 1 ( z s + r a d v ) + d e c o d e r a 2 ( z s + r a d v ) ) , subject to | | r a d v | | 2 ξ ,
where z s = s ( x ) , and d(.) is the sum of the pixel-wise Kullback–Leibler (KL) divergence between the original and perturbed pixel predictions. Figure 6 illustrates feature-level perturbation.

3.4.3. Model-Level Perturbation

To enhance the perturbations of the self-training method, we introduced model-level perturbations through auxiliary teachers. In the unsupervised stage, we utilised the teacher model, which received parameters from the student model with EMA, to predict the input images. Based on the assumption of consistency regularisation, multiple models trained at different stages should have similar predictions for the same image. By measuring the differences between the labels generated by teachers and the student model using a loss function, we provided feedback to the student model and updated its parameters, thus improving the generalisation performance of the model.
It is worth noting that, based on the idea of gradually strengthening model learning, we used weak data augmentation methods and feature-level VAT perturbation instead of directly incorporating strong data augmentation methods into the supervised learning. We found from experiments that training the model from weak to strong effectively improved the performance of the supervised learning as well. Using strong data augmentation throughout all epochs may lead to a decrease in the performance of the model. However, if strong data augmentation is selectively applied in the latter half of the epochs, there is potential for improvement in the performance. Since VAT perturbation searches for pixels with more noise in the current features, it is related to the performance and is considered a weak-to-strong data augmentation method. Meanwhile, based on the idea of progressive learning, we gradually shifted the focus of the loss function towards the labels generated by the teacher model by dynamically adjusting its attention. This adjustment allowed the loss function to increasingly prioritise the labels provided by the teacher model as the training progressed, leading to more reasonable model learning.

3.5. Ensemble Voting: Pseudo-Label Generation Strategy (EV)

In a model based on the self-training paradigm, the assignment of pseudo-labels to unlabelled data using a trained teacher model plays a crucial role. However, past self-training methods have faced limitations when manually setting confidence thresholds and filtering the softmax probability results from the model. Although this approach has contributed to some improvement in the confidence of the model, it suffers from issues such as the inefficient utilisation of all pixels, heavy reliance on artificially set hyperparameters, and the presence of fuzzy segmentation boundaries.
To address these challenges and enhance the model predictions, we integrated TTA technology. TTA leverages multiple scales of images, and we uniformly resized them using bilinear interpolation to ensure consistent dimensions for predictions. In our experiments, we explored different scaling factors, including 0.5, 0.75, 1.0, 1.5, and 2.0, and weighted the predictions obtained at each scale. Furthermore, we horizontally flipped the images at each scale to augment the ability of the model to recognise objects in the image. By aggregating the predictions from all scales and applying softmax, we obtained the final prediction.
Additionally, we introduce auxiliary teachers to further support the decision-making process of the model. Following the ensemble learning vote concept, we utilise TTA technology to predict labels using the student model. Subsequently, we combine the TTA-augmented predicted labels from the auxiliary teachers and the student model, assigning appropriate weights to achieve more reliable and smoother labelling results.
Overall, our labelling method, outlined in Algorithm 2, effectively leverages TTA to enhance the model predictions and achieve improved performance in forest cover mapping. The incorporation of TTA for both auxiliary teachers and the student model contributes to better decision making and yields superior segmentation results.
Algorithm 2 Labelling
Require: Unlabelled training set D u = { u i } i = 1 N ,
    Test-Time Augmentation A t ,
    Teacher/Auxiliary teacher model t l / t a 1 , t a 2
Ensure: pseudo-label D u ^
  1:
for minibatch { ( x i ) } i = 1 B D u  do
  2:
for k ∈ { 1 , . . . , B }  do
  3:
   y k 1 = A t ( t l ( x k ) )
  4:
   y k 2 = A t ( t a 1 ( x k ) )
  5:
   y k 3 = A t ( t a 2 ( x k ) )
  6:
   y k = 0.5 ( 0.5 ( y k 2 + y k 3 ) + y k 1 )
  7:
   D u ^ y k
  8:
end for
  9:
end for
10:
return  D u ^

3.6. Adaptive Loss (AL)

To address the issue of dataset imbalance, we trained the model using a combination of cross-entropy loss and focal loss. Cross-entropy is given by
L c e ( y , y ^ ) = c = 1 C y c log ( y ^ c ) ,
where y is a one-hot vector with length C, representing the true class; y c is the cth element in the vector y; y ^ is the predicted probability distribution vector of the model; and y ^ c represents the probability that the model predicts the cth class. Focal loss is given by
L f o c a l ( y , y ^ ) = c = 1 C y c ( 1 y ^ c ) γ log ( y ^ c ) ,
where y is a one-hot vector with length C, representing the true class; y c is the cth element in the vector y; y ^ is the predicted probability distribution vector of the model; y ^ c represents the probability that the model predicts the cth class; and γ is a hyperparameter called the focusing parameter, which is used to adjust the degree of attention that the loss function pays to the predicted probability of different classes. When γ > 0 , focal loss pays more attention to the mispredicted samples, thus reducing the problem of class imbalance.
It is worth noting that focal loss was originally designed to solve the problem of extremely imbalanced positive and negative samples, as well as samples that are difficult to classify. Given that the focal loss adjusts the ratio of positive and negative sample loss adaptively based on the difficulty of the dataset, we did not remove focal loss when conducting experiments on balanced datasets. At the same time, to minimise problems caused by the manual setting of hyperparameters, we introduced an automatic coefficient,
L u = t o t a l i t e r i t e r t o t a l i t e r ( L c e ( y k ^ , y k ) + L f o c a l ( y k ^ , y k ) ) + i t e r t o t a l i t e r ( L c e ( y k 2 ^ , y k ^ ) + L f o c a l ( y k 2 ^ , y k ^ ) ) ,
where y k 2 ^ represents the output of the auxiliary teachers, y k ^ represents the pseudo-labels generated by the previous stage model, and y k represents the predicted results of the student model. As training progresses, the reliability of the pseudo-labels generated during the initial training phase decreases, and the predictions from the auxiliary teachers become more accurate. Therefore, the loss for the original labelled data gradually decreases as the training progresses, while the loss for the predictions between the auxiliary teachers and the student model gradually increases. At the same time, the loss between the predicted results of the model and the pseudo-labels generated by the previous stage of the model and the auxiliary teachers can also be seen as a consistency regularisation method. Currently, we have only tried linear transformations and have not explored whether there are better adaptive adjustment coefficients.

4. Data and Experiments

4.1. Data Description

In evaluating the performance of forest cover mapping, we utilised two datasets sourced from the SentinelHub satellite image database. The details regarding the number of images and the distribution between forest and non-forest categories are presented in Table 1. Both datasets consist of four-band data, with one originating from the Amazon Rainforest and the other from the Atlantic Forest (Mata Atlantica) [33]. The geographical distribution of these biomes can be observed in Figure 7, accompanied by sample images highlighting the dataset concentration in two distinct regions.
To streamline the training process, we adopted an approach similar to a related work [34] by training solely on the RGB channels extracted from the four-band dataset. Notably, the model demonstrated favourable performance. Each image in the datasets has dimensions of (512, 512, 3), while each forest cover mapping mask is represented by (512, 512, 1).
These datasets provide insights into two real-world scenarios: one involving an imbalanced class distribution and complex images (Atlantic Forest), posing a challenging learning task, and the other involving a balanced class distribution and relatively simple images (Amazon Forest).

4.2. Experimental Settings

To ensure a fair comparison with most existing works, we maintained consistent hyperparameters between the supervised pre-training of the teacher models and the semi-supervised re-training of the student model. Specifically, we set the batch size to 8 during training with a V100-SXM2-32GB GPU. For optimisation, we used the SGD optimiser with an initial base learning rate of 0.001 for the backbones. The learning rate of the randomly initialised segmentation head was 10 times larger than that of the backbones. Additionally, we adopted poly scheduling to decay the learning rate during the training process, i.e.,
l r = b a s e l r · ( 1 i t e r t o t a l i t e r ) 0.9 .
The model was trained for 80 epochs using weak data augmentations, which includes the random flipping and resizing of training images between 0.5 and 2.0. For strong data augmentations on unlabelled images, we used colorJitter with the same intensity as in [31], greyscale, blur (same as in [35]), and cutout with random values filled. The cutout regions were ignored in loss computation. During the pseudo-labelling phase, all unlabelled images underwent TTA, which involved five scales and horizontal flipping. The testing images were evaluated at their original resolution, and no post-processing techniques were employed. It is worth noting that to enable a fair comparison with most existing works, we have not incorporated any advanced optimisation strategies, such as OHEM in [36], auxiliary supervision in [36,37], or SyncBN, into our method.

4.3. Evaluation Metrics

To evaluate the performance of our proposed method in forest cover mapping, we measured several evaluation metrics, including IoU, mean IoU (mIoU), mean precision, mean recall, mean F1-score, and accuracy for both forest and non-forest classes, i.e.,
I o U = T P T P + F P + F N
m I o U = 1 c i = 1 c T P i T P i + F P i + F N i
m e a n p r e c i s i o n = 1 c i = 1 c T P i T P i + F P i
m e a n r e c a l l = 1 c i = 1 c T P i T P i + F N i
m e a n F 1 = 1 c i = 1 c 2 T P i 2 T P i + F P i + F N i
a c c u r a c y = T P + T N T P + T N + F P + F N ,
where c represents the number of shared classes between the benchmark datasets ( c = 2 ) . Figure 8 shows the class distribution of the two datasets used.
The evaluation metrics were computed using the confusion matrix generated by our semi-supervised segmentation framework, which contains the pixel numbers of true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). Specifically, mIoU and IoU measure the similarity between predicted and ground-truth forest/non-forest areas, while accuracy measures the overall percentage of correctly classified pixels. The mean precision, mean recall, and mean F1-score consider both precision and recall, which are suitable for multi-class and pixel-level classification tasks. Our results demonstrate the superior performance of our proposed method in accurately mapping forest cover, as evidenced by the higher values of these evaluation metrics.

4.4. Results and Analysis

Based on the aforementioned forest RS datasets, we compared our proposed method with several SOTA semi-supervised semantic segmentation frameworks, including the feature-perturbation-based CCT [25], the multi-perturbation-based PS-MT [26], and baseline ST with TTA. We also included the supervised DeeplabV3+ [38] model using ResNet-101 [39] trained only with labelled data as a baseline. All semi-supervised methods were implemented with identical experimental conditions and settings to ensure fairness. The results demonstrate the effectiveness and superiority of our proposed method in accurately mapping forest cover. Additionally, our method has huge potential for practical applications in forest monitoring and management.

4.4.1. Comparison Results on the Atlantic Forest Dataset

Table 2 and Table 3 show comparative results with other SOTA methods on the validation and test sets of the Atlantic dataset. As introduced in Section 2, the CCT method based on image surface perturbation cannot effectively learn the information of satellite RS data, resulting in poor performance with inadequate fitting during the training process. Similarly, other semi-supervised models (e.g., PS-MT) also suffer from this problem and the instability caused by numerous hyperparameters, performing far worse than our proposed method on multiple metrics after our adjustment. Methods based on self-training and data augmentation using colour transformations exhibit stronger performance, and our method, which is an improvement on self-training, outperforms them. Meanwhile, compared with fully supervised methods, our approach also shows significant superiority.
Figure 9 graphically shows the performance of mIoU for different models. Specifically, on the validation set, except for a slightly lower forest segmentation IoU compared to the ST method at the 1/16 split, our model outperforms the other SOTA methods in all metrics. Similarly, on the test set, our method surpasses the majority of the SOTA methods in most metrics. For the important metric mIoU in forest cover segmentation, our model demonstrates superior performance across different splits, as well as on the validation and test sets. This indicates that our model is capable of effectively learning information from satellite RS images and exhibits strong robustness.
We present partial visual comparison results of all the methods in Figure 10. On the Atlantic Forest dataset, due to the data imbalance and the difficulty of image learning, it can be observed that our proposed method produces smoother boundaries and highlights more details compared to the fully supervised methods. Moreover, our method yields predictions that are closer to the ground truth compared to other semi-supervised methods.

4.4.2. Comparison Results on the Amazon Forest Dataset

Table 4 and Table 5 present the comparison results on the Amazon Forest validation and test sets using our evaluation metrics of accuracy, mRecall, mPrecision, mF1, and mIoU. The tables show that on the datasets with class balance and lower image difficulty, the supervised Deeplabv3+ achieves better segmentation accuracy than the results obtained on the Atlantic Forest dataset. Similarly, the performance of PS-MT gradually improves, but due to the existence of manually adjusted semi-supervised loss ratio coefficients, the overall performance remains unstable. Additionally, the CCT method fails to learn image information due to the challenges presented by satellite RS datasets. Consistent with our expectations, the self-training method still exhibits strong stability and significantly improves performance across all evaluation metrics compared to fully supervised methods. However, as the amount of labelled data increases, the performance of the model decreases. Furthermore, the performance of the supervised method varies due to varying image difficulties, with its performance at the 1/8 partition being inferior to that achieved at the 1/16 partition. In contrast, our proposed Semi-FCMNet, which enhances the perturbation of the model, leads to performance improvements. Moreover, our method also outperforms the supervised method and other SOTA methods on the test set, indicating its strong generalisation ability.
Figure 11 graphically shows the performance of mIoU for different models.
We also present partial visual comparison results of all methods on the Amazon Forest dataset in Figure 12. Due to the lower sample complexity of this dataset, most methods achieve good prediction results. However, it is worth noting that our proposed method generates accurate predictions at different partition ratios, as shown in the figure.

4.5. Ablation Experiments

Ablation studies were conducted to validate the effectiveness of each key component of our proposed method. Our method mainly consists of the following four core components: hybrid perturbations (including input image representation level (MP), feature level, and model level); AL; and a pseudo-label generation strategy based on TTA and multi-model voting (PGS). We present specific metric data in Table 6 and Table 7, along with the visualisation results of the ablation experiments in Figure 13 and Figure 14.
Base: When employing the basic self-training method on the imbalanced Atlantic Forest dataset with high sample learning difficulty, the performance of the model at the 1/16 partition is slightly inferior to that of the fully supervised method. However, in other partitions, the semi-supervised method showcases its superiority by exhibiting better performance on the validation and test sets compared to the fully supervised method. This outcome fully demonstrates the advantages of the semi-supervised approach. However, on the Amazon Forest dataset with lower sample learning difficulty, the basic self-training method is prone to overfitting to the noise and cannot exceed the performance of the fully supervised method.
MP: After adding the MP method, the various indicators of the model are further improved, which proves that adding perturbations to the self-training paradigm, i.e., integrating the consistency regularisation method with the self-training method, effectively improves model performance.
EV: When the pseudo-label generation strategy based on TTA and multi-model voting (EV) is added, the scores of the model on important indicators, e.g., mIoU and mF1-score, are improved in various partitions of datasets with different data distributions. It can be seen that the method of generating pseudo-labels has a significant impact on the performance of the self-training paradigm, and our pseudo-label generation method effectively improves the performance of the model.
AL: After adding the adaptive loss, the model dynamically adjusts its focus on the loss. At the beginning of the training, the model pays more attention to the predicted results of the teacher model in the previous round. However, due to the limited performance of the teacher in the previous round, the current student model cannot be effectively improved in the later stage of training. The introduction of the adaptive loss makes the model focus more on the difference between its multiple prediction results in the later stage and improve its own performance through consistency regularisation. The improvement in various indicators reflects the correctness of our method.

5. Conclusions

In this study, we propose the integration of the self-training paradigm with the mean-teacher paradigm based on multiple perturbations and EV. We apply consistency regularisation and self-training pseudo-labelling techniques in the forest-cover-mapping scenario. By introducing auxiliary teachers and establishing more detailed pseudo-label generation strategies, as well as enhancing the perturbations in our model, we demonstrate excellent segmentation results. Our method demonstrates robust performance on datasets exhibiting both class imbalance and class balance. This is accomplished through the dynamic adjustment of the loss function and straightforward hyperparameter settings. Furthermore, extensive ablation experiments conducted on the validation and test sets confirm the effectiveness and robustness of our proposed method. Finally, considering the performance of our method across various metrics and other objectives in RS-image-processing tasks, we have identified potential avenues for further research and development.
Although our proposed method has achieved promising results, we believe that there are several directions for further research in the future:
  • In the pseudo-label generation strategy based on TTA and EV, we used averaging to generate pseudo-labels with higher confidence using multi-model votes. In the future, adaptive weighting generation methods will be explored.
  • The adaptive loss is set for linear adjustment, and thus, the non-linear setting will be investigated.
  • In view of model limitations, the easy accessibility of RGB remote sensing images, and equipment cost, this study only explored the RGB channel combination. In future work, more effective channel fusion methods can be investigated, or alternative approaches, like using the NIR band instead of the blue band, can be explored for forest cover mapping.

Author Contributions

Methodology, B.C., L.W. and X.F.; Validation, W.B.; Investigation, X.Y.; Data curation, L.W.; Writing—original draft, B.C.; Writing—review & editing, X.F. and T.T.; Visualization, X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Program (No. 2022YFD2201005) and Natural Science Foundation of China (No. 61902187).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Van Mantgem, P.J.; Stephenson, N.L.; Byrne, J.C.; Daniels, L.D.; Franklin, J.F.; Fulé, P.Z.; Harmon, M.E.; Larson, A.J.; Smith, J.M.; Taylor, A.H.; et al. Widespread increase of tree mortality rates in the western United States. Science 2009, 323, 521–524. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Zhu, Y.; Li, D.; Fan, J.; Zhang, H.; Eichhorn, M.P.; Wang, X.; Yun, T. A reinterpretation of the gap fraction of tree crowns from the perspectives of computer graphics and porous media theory. Front. Plant Sci. 2023, 14, 1109443. [Google Scholar] [CrossRef] [PubMed]
  3. Boers, N.; Marwan, N.; Barbosa, H.M.; Kurths, J. A deforestation-induced tipping point for the South American monsoon system. Sci. Rep. 2017, 7, 41489. [Google Scholar] [CrossRef] [Green Version]
  4. Li, X.; Wang, X.; Gao, Y.; Wu, J.; Cheng, R.; Ren, D.; Bao, Q.; Yun, T.; Wu, Z.; Xie, G.; et al. Comparison of Different Important Predictors and Models for Estimating Large-Scale Biomass of Rubber Plantations in Hainan Island, China. Remote Sens. 2023, 15, 3447. [Google Scholar] [CrossRef]
  5. Lewis, S.L.; Lopez-Gonzalez, G.; Sonké, B.; Affum-Baffoe, K.; Baker, T.R.; Ojo, L.O.; Phillips, O.L.; Reitsma, J.M.; White, L.; Comiskey, J.A.; et al. Increasing carbon storage in intact African tropical forests. Nature 2009, 457, 1003–1006. [Google Scholar] [CrossRef] [PubMed]
  6. Hansen, M.C.; Stehman, S.V.; Potapov, P.V.; Loveland, T.R.; Townshend, J.R.; DeFries, R.S.; Pittman, K.W.; Arunarwati, B.; Stolle, F.; Steininger, M.K.; et al. Humid tropical forest clearing from 2000 to 2005 quantified by using multitemporal and multiresolution remotely sensed data. Proc. Natl. Acad. Sci. USA 2008, 105, 9439–9444. [Google Scholar] [CrossRef]
  7. Sexton, J.O.; Song, X.P.; Feng, M.; Noojipady, P.; Anand, A.; Huang, C.; Kim, D.H.; Collins, K.M.; Channan, S.; DiMiceli, C.; et al. Global, 30-m resolution continuous fields of tree cover: Landsat-based rescaling of MODIS vegetation continuous fields with lidar-based estimates of error. Int. J. Digit. Earth 2013, 6, 427–448. [Google Scholar] [CrossRef] [Green Version]
  8. Hamunyela, E.; Reiche, J.; Verbesselt, J.; Herold, M. Using space-time features to improve detection of forest disturbances from Landsat time series. Remote Sens. 2017, 9, 515. [Google Scholar] [CrossRef] [Green Version]
  9. Yin, H.; Khamzina, A.; Pflugmacher, D.; Martius, C. Forest cover mapping in post-Soviet Central Asia using multi-resolution remote sensing imagery. Sci. Rep. 2017, 7, 1375. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, P.; Ke, Y.; Zhang, Z.; Wang, M.; Li, P.; Zhang, S. Urban land use and land cover classification using novel deep learning models based on high spatial resolution satellite imagery. Sensors 2018, 18, 3717. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Zhang, Y.; Li, X.; Ling, F.; Atkinson, P.M.; Ge, Y.; Shi, L.; Du, Y. Updating Landsat-based forest cover maps with MODIS images using multiscale spectral-spatial-temporal superresolution mapping. Int. J. Appl. Earth Obs. Geoinf. 2017, 63, 129–142. [Google Scholar] [CrossRef] [Green Version]
  12. Flores, E.; Zortea, M.; Scharcanski, J. Dictionaries of deep features for land-use scene classification of very high spatial resolution images. Pattern Recognit. 2019, 89, 32–44. [Google Scholar] [CrossRef]
  13. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  14. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  15. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  16. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Computer Vision—ECCV 2018; Lecture Notes in Computer Science; Springer International Publishing: Cham, Germany, 2018; pp. 833–851. [Google Scholar] [CrossRef] [Green Version]
  17. Bragagnolo, L.; da Silva, R.; Grzybowski, J. Amazon forest cover change mapping based on semantic segmentation by U-Nets. Ecol. Inform. 2021, 62, 101279. [Google Scholar] [CrossRef]
  18. Flood, N.; Watson, F.; Collett, L. Using a U-net convolutional neural network to map woody vegetation extent from high resolution satellite imagery across Queensland, Australia. Int. J. Appl. Earth Obs. Geoinf. 2019, 82, 101897. [Google Scholar] [CrossRef]
  19. Isaienkov, K.; Yushchuk, M.; Khramtsov, V.; Seliverstov, O. Deep Learning for Regular Change Detection in Ukrainian Forest Ecosystem with Sentinel-2. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 364–376. [Google Scholar] [CrossRef]
  20. Papandreou, G.; Chen, L.C.; Murphy, K.P.; Yuille, A.L. Weakly-and semi-supervised learning of a deep convolutional network for semantic image segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1742–1750. [Google Scholar]
  21. Peláez-Vegas, A.; Mesejo, P.; Luengo, J. A Survey on Semi-Supervised Semantic Segmentation. arXiv 2023, arXiv:cs.CV/2302.09899. [Google Scholar]
  22. Van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef] [Green Version]
  23. Lucas, B.; Pelletier, C.; Schmidt, D.; Webb, G.I.; Petitjean, F. A bayesian-inspired, deep learning-based, semi-supervised domain adaptation technique for land cover mapping. Mach. Learn. 2021, 112, 1941–1973. [Google Scholar] [CrossRef]
  24. Hong, D.; Yokoya, N.; Ge, N.; Chanussot, J.; Zhu, X.X. Learnable manifold alignment (LeMA): A semi-supervised cross-modality learning framework for land cover and land use classification. ISPRS J. Photogramm. Remote Sens. 2019, 147, 193–205. [Google Scholar] [CrossRef]
  25. Ouali, Y.; Hudelot, C.; Tami, M. Semi-supervised semantic segmentation with cross-consistency training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12674–12684. [Google Scholar]
  26. Liu, Y.; Tian, Y.; Chen, Y.; Liu, F.; Belagiannis, V.; Carneiro, G. Perturbed and strict mean teachers for semi-supervised semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4258–4267. [Google Scholar]
  27. Abuduweili, A.; Li, X.; Shi, H.; Xu, C.Z.; Dou, D. Adaptive Consistency Regularization for Semi-Supervised Transfer Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6923–6932. [Google Scholar]
  28. Zoph, B.; Ghiasi, G.; Lin, T.Y.; Cui, Y.; Liu, H.; Cubuk, E.D.; Le, Q. Rethinking Pre-training and Self-training. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: San Francisco, CA, USA, 2020; Volume 33, pp. 3833–3845. [Google Scholar]
  29. Zhu, Y.; Zhang, Z.; Wu, C.; Zhang, Z.; He, T.; Zhang, H.; Manmatha, R.; Li, M.; Smola, A.J. Improving Semantic Segmentation via Efficient Self-Training. IEEE Trans. Pattern Anal. Mach. Intell. 2021, early access. 1. [Google Scholar] [CrossRef]
  30. SemiRoadExNet: A semi-supervised network for road extraction from remote sensing imagery via adversarial learning. ISPRS J. Photogramm. Remote Sens. 2023, 198, 169–183. [CrossRef]
  31. Zou, Y.; Zhang, Z.; Zhang, H.; Li, C.L.; Bian, X.; Huang, J.B.; Pfister, T. Pseudoseg: Designing pseudo labels for semantic segmentation. arXiv 2020, arXiv:2010.09713. [Google Scholar]
  32. Zhang, B.; Zhang, Y.; Li, Y.; Wan, Y.; Guo, H.; Zheng, Z.; Yang, K. Semi-supervised Deep Learning via Transformation Consistency Regularization for Remote Sensing Image Semantic Segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 5782–5796. [Google Scholar] [CrossRef]
  33. Bragagnolo, L.; da Silva, R.V.; Grzybowski, J.M.V. Amazon and Atlantic Forest image datasets for semantic segmentation. 2021. [Google Scholar] [CrossRef]
  34. John, D.; Zhang, C. An attention-based U-Net for detecting deforestation within satellite sensor imagery. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102685. [Google Scholar] [CrossRef]
  35. Chen, X.; Fan, H.; Girshick, R.; He, K. Improved baselines with momentum contrastive learning. arXiv 2020, arXiv:2003.04297. [Google Scholar]
  36. Hu, H.; Wei, F.; Hu, H.; Ye, Q.; Cui, J.; Wang, L. Semi-supervised semantic segmentation via adaptive equalization learning. Adv. Neural Inf. Process. Syst. 2021, 34, 22106–22118. [Google Scholar]
  37. Chen, X.; Yuan, Y.; Zeng, G.; Wang, J. Semi-supervised semantic segmentation with cross pseudo supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2021; pp. 2613–2622. [Google Scholar]
  38. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
Figure 1. Visualisation of images after application of different processing methods: (a) original image; and (b) image enhanced using contrast enhancement.
Figure 1. Visualisation of images after application of different processing methods: (a) original image; and (b) image enhanced using contrast enhancement.
Remotesensing 15 04012 g001
Figure 2. Visualisation of balanced and imbalanced samples (where black represents non-forest, and white represents forest): (a) balanced sample and (b) imbalanced sample.
Figure 2. Visualisation of balanced and imbalanced samples (where black represents non-forest, and white represents forest): (a) balanced sample and (b) imbalanced sample.
Remotesensing 15 04012 g002
Figure 3. Overview of the proposed network pipeline.
Figure 3. Overview of the proposed network pipeline.
Remotesensing 15 04012 g003
Figure 4. Overview of the proposed auxiliary teacher module.
Figure 4. Overview of the proposed auxiliary teacher module.
Remotesensing 15 04012 g004
Figure 5. Overview of the proposed training strategy.
Figure 5. Overview of the proposed training strategy.
Remotesensing 15 04012 g005
Figure 6. Visualisation results of feature maps before and after feature-level perturbation: feature maps (extracted features of the image after input perturbation by the encoder) and VAT disturbance (image with injected VAT perturbation after feature extraction).
Figure 6. Visualisation results of feature maps before and after feature-level perturbation: feature maps (extracted features of the image after input perturbation by the encoder) and VAT disturbance (image with injected VAT perturbation after feature extraction).
Remotesensing 15 04012 g006
Figure 7. Map of biomes in Amazon Rainforest and Atlantic Forest.
Figure 7. Map of biomes in Amazon Rainforest and Atlantic Forest.
Remotesensing 15 04012 g007
Figure 8. Class distribution of the datasets: (a) class distribution of the Atlantic dataset and (b) class distribution of the Amazon Forest dataset.
Figure 8. Class distribution of the datasets: (a) class distribution of the Atlantic dataset and (b) class distribution of the Amazon Forest dataset.
Remotesensing 15 04012 g008
Figure 9. Performance of mIoU metric for different models on the Atlantic Forest dataset: (a) performance on the validation set of the Atlantic Forest and (b) performance on the test set of the Atlantic Forest.
Figure 9. Performance of mIoU metric for different models on the Atlantic Forest dataset: (a) performance on the validation set of the Atlantic Forest and (b) performance on the test set of the Atlantic Forest.
Remotesensing 15 04012 g009
Figure 10. Visualisation of model predictions on different partitions on the Atlantic Forest dataset: (a) supervised; (b) CCT; (c) PS-MT; (d) ST; and (e) Semi-FCMNet.
Figure 10. Visualisation of model predictions on different partitions on the Atlantic Forest dataset: (a) supervised; (b) CCT; (c) PS-MT; (d) ST; and (e) Semi-FCMNet.
Remotesensing 15 04012 g010
Figure 11. Performance of mIoU metric for different models on the Amazon Forest dataset: (a) performance on the validation set of the Amazon Forest and (b) performance on the test set of the Amazon Forest.
Figure 11. Performance of mIoU metric for different models on the Amazon Forest dataset: (a) performance on the validation set of the Amazon Forest and (b) performance on the test set of the Amazon Forest.
Remotesensing 15 04012 g011
Figure 12. Visualisation of model predictions under different partitions on the Amazon Forest dataset: (a) supervised; (b) CCT; (c) PS-MT; (d) ST; and (e) Semi-FCMNet.
Figure 12. Visualisation of model predictions under different partitions on the Amazon Forest dataset: (a) supervised; (b) CCT; (c) PS-MT; (d) ST; and (e) Semi-FCMNet.
Remotesensing 15 04012 g012
Figure 13. Visualisation of ablation experiment results on the Atlantic Forest dataset: (a) input image; (b) ground truth; (c) Base; (d) EV; (e) EV + MP; and (f) all.
Figure 13. Visualisation of ablation experiment results on the Atlantic Forest dataset: (a) input image; (b) ground truth; (c) Base; (d) EV; (e) EV + MP; and (f) all.
Remotesensing 15 04012 g013
Figure 14. Visualisation of ablation experiment results on the Amazon Forest dataset: (a) input image; (b) ground truth; (c) Base; (d) EV; (e) EV + MP; and (f) all.
Figure 14. Visualisation of ablation experiment results on the Amazon Forest dataset: (a) input image; (b) ground truth; (c) Base; (d) EV; (e) EV + MP; and (f) all.
Remotesensing 15 04012 g014
Table 1. Number of images within each dataset, as well as the forest (F) and non-forest (NF) class balance within each dataset.
Table 1. Number of images within each dataset, as well as the forest (F) and non-forest (NF) class balance within each dataset.
DatasetNumber of Images (F-NF Class Balance)
TrainingTestingValidation
RGB 3-band Amazon Forest250 (50.0–50.0%)20 (49.8–50.2%)100 (47.8–52.2%)
RGB 3-band Atlantic Forest250 (33.3–66.7%)20 (31.5–68.5%)100 (33.8–66.2%)
Table 2. Comparison results with other SOTA methods on Atlantic Forest validation dataset using evaluation metrics (%). (* Denotes that we have made appropriate modifications to the model, alleviating the issue of unsuccessful training).
Table 2. Comparison results with other SOTA methods on Atlantic Forest validation dataset using evaluation metrics (%). (* Denotes that we have made appropriate modifications to the model, alleviating the issue of unsuccessful training).
Validation
Labelled DataMetricSupOnlyCCTPS-MT *ST (TTA)Semi-FCMNet
1/16(16)mIoU69.8635.5062.0970.3572.93
IoU(NF)76.3454.2273.0172.8579.85
IoU(F)63.3716.7951.1867.8465.99
Accuracy76.3058.0978.9680.7981.48
Mean precision81.3249.9976.6383.5483.58
Mean recall83.8649.9975.5983.3384.92
Mean F1-score82.0849.5376.0582.5784.15
1/8(32)mIoU71.1635.5063.3072.3173.74
IoU(NF)78.8554.2277.1379.9680.67
IoU(F)63.4616.7949.4764.6566.79
Accuracy81.5458.0981.3281.1082.67
Mean precision82.5849.9982.7583.5484.23
Mean recall83.2849.9974.6583.8685.27
Mean F1-score82.9149.5376.6483.6984.69
1/4(63)mIoU70.2035.5067.9572.7774.07
IoU(NF)77.7554.2280.1980.2181.21
IoU(F)62.6416.7955.7065.3166.91
Accuracy79.7758.0982.1483.0683.96
Mean precision81.7349.9983.5983.7884.66
Mean recall82.9549.9977.9884.2885.16
Mean F1-score82.2549.5380.2784.3084.90
1/2(125)mIoU70.8735.5045.0973.3573.95
IoU(NF)77.5054.2270.4179.3380.51
IoU(F)64.2316.7919.7767.3667.38
Accuracy77.9058.0972.4278.9581.65
Mean precision82.0049.9981.6383.6584.19
Mean recall84.2449.9959.6284.3285.80
Mean F1-score82.7749.5357.8284.4884.85
Table 3. Comparison results with other SOTA methods on Atlantic Forest test dataset using evaluation metrics (%). (* Denotes that we have made appropriate modifications to the model, alleviating the issue of unsuccessful training).
Table 3. Comparison results with other SOTA methods on Atlantic Forest test dataset using evaluation metrics (%). (* Denotes that we have made appropriate modifications to the model, alleviating the issue of unsuccessful training).
Test
Labelled DataMetricSupOnlyCCTPS-MT *ST (TTA)Semi-FCMNet
1/16(16)mIoU69.3635.9861.1770.0271.65
IoU(NF)77.3755.7773.7778.9780.60
IoU(F)61.3316.1848.5861.0662.70
Accuracy76.3259.2578.9879.8982.38
Mean precision80.6749.9975.7281.4982.94
Mean recall83.4949.9974.6882.7283.41
Mean F1-score81.6449.7375.1482.0483.16
1/8(32)mIoU69.9535.9856.6770.1773.61
IoU(NF)79.7355.7776.2179.1982.45
IoU(F)60.1516.1837.1361.1464.76
Accuracy82.4759.2579.1480.3285.40
Mean precision82.0949.9982.8181.6684.79
Mean recall81.7649.9968.3282.7084.22
Mean F1-score81.9249.7370.3282.1384.49
1/4(63)mIoU70.2035.9862.1671.8774.51
IoU(NF)79.8155.7778.6080.8383.26
IoU(F)61.0116.1845.7362.9065.76
Accuracy81.9359.2581.8782.8186.75
Mean precision82.1849.9984.6183.1685.67
Mean recall82.3849.9972.8583.4784.61
Mean F1-score82.2849.7375.3883.3185.10
1/2(125)mIoU72.7835.9844.3175.2375.62
IoU(NF)81.2255.7772.1583.0283.24
IoU(F)64.3316.1816.4767.4368.00
Accuracy82.4359.2573.6084.1284.11
Mean precision83.5449.9984.9685.2085.39
Mean recall84.4649.9958.1786.1386.49
Mean F1-score83.9649.7356.0485.6385.90
Table 4. Comparison results with other SOTA methods on the Amazon Forest validation dataset using evaluation metrics (%).
Table 4. Comparison results with other SOTA methods on the Amazon Forest validation dataset using evaluation metrics (%).
Validation
Labelled DataMetricSupOnlyCCTPS-MTST (TTA)Semi-FCMNet
1/16(16)mIoU88.0730.7954.3187.6191.70
IoU(NF)87.6341.2443.6687.0591.69
IoU(F)88.5120.3464.9688.1691.69
Accuracy94.1148.9172.4493.0096.73
Mean precision93.6450.0080.4893.4195.79
Mean recall93.6750.0071.2793.3795.85
Mean F1-score93.6546.1069.7693.3995.81
1/8(32)mIoU87.9630.7977.1089.1492.02
IoU(NF)87.4841.2474.1888.7091.76
IoU(F)88.4320.3480.0389.5792.28
Accuracy93.7848.9187.3194.4096.78
Mean precision93.5850.0089.1194.2595.82
Mean recall93.6050.0086.8494.2695.88
Mean F1-score93.5946.1087.0494.2595.84
1/4(63)mIoU90.3230.7980.4689.9592.48
IoU(NF)90.0441.2479.8189.6392.19
IoU(F)90.5920.3481.1190.2692.77
Accuracy96.2148.9189.1895.7196.45
Mean precision94.8950.0089.1594.6896.08
Mean recall94.9650.0089.1994.7596.10
Mean F1-score94.9146.1089.1794.7096.09
1/2(125)mIoU90.8130.7983.4290.6991.71
IoU(NF)90.6641.2483.3290.5591.49
IoU(F)90.9620.3483.5290.8291.93
Accuracy97.7648.9190.9697.8997.28
Mean precision95.2150.0091.0695.1595.66
Mean recall95.2950.0091.1095.2395.74
Mean F1-score95.1846.1090.9695.1195.67
Table 5. Comparison results with other SOTA methods on the Amazon Forest test dataset using evaluation metrics (%).
Table 5. Comparison results with other SOTA methods on the Amazon Forest test dataset using evaluation metrics (%).
Test
Labelled DataMetricSupOnlyCCTPS-MTST (TTA)Semi-FCMNet
1/16(16)mIoU89.4431.3855.2589.8492.64
IoU(NF)89.5342.7246.0489.8592.65
IoU(F)89.3320.0464.4689.8292.62
Accuracy95.7549.9172.7395.1796.76
Mean precision94.4550.0081.0094.6496.18
Mean recall94.4250.0072.6294.6496.18
Mean F1-score94.4246.6270.7194.6496.18
1/8(32)mIoU82.0931.3874.7190.8192.80
IoU(NF)81.4542.7272.0190.8492.84
IoU(F)82.7120.0477.4290.7692.79
Accuracy87.1449.9185.7296.0396.95
Mean precision90.3750.0087.8695.1996.28
Mean recall90.1650.0085.6695.1896.27
Mean F1-score90.1646.6285.4995.1896.27
1/4(63)mIoU91.0031.3883.1591.4293.02
IoU(NF)91.0942.7283.6391.4993.00
IoU(F)90.9020.0482.6691.3493.02
Accuracy96.7349.9190.8196.7996.62
Mean precision95.3250.0091.0095.5496.38
Mean recall95.2950.0090.8295.5296.38
Mean F1-score95.2846.6290.7995.5196.38
1/2(125)mIoU90.9431.3883.0491.0692.12
IoU(NF)91.0842.7283.8691.2592.18
IoU(F)90.7920.0482.2390.8692.05
Accuracy97.2349.9190.7697.6097.15
Mean precision95.3250.0091.2795.4495.92
Mean recall95.2650.0090.7895.3395.90
Mean F1-score95.2546.6290.7395.3295.89
Table 6. Quantitative evaluation of method contributions on Atlantic Forest dataset. (✓ denotes the incorporation of this module).
Table 6. Quantitative evaluation of method contributions on Atlantic Forest dataset. (✓ denotes the incorporation of this module).
RatioMethodValidationTest
BaseMPEVALmIoUAccuracymPrecisionmRecallmF1-ScoremIoUAccuracymPrecisionmRecallmF1-Score
1/16 68.8777.7080.6882.3781.3369.5978.2880.9782.9081.76
70.3580.7983.5483.3382.5770.0279.8981.4982.7282.04
72.3480.6183.1284.6883.7671.1580.6182.3883.3582.83
72.9381.4883.5884.9284.1571.6582.3882.9483.4183.16
1/8 72.0283.5083.4983.4883.4970.1481.1281.8282.3982.09
72.3181.1083.5483.8683.6970.1780.3281.6682.7082.13
72.8181.7983.5684.7284.0772.9184.0784.0484.0184.02
73.7482.6784.2385.2784.6973.6185.4084.7984.2284.49
1/4 70.3381.9182.2182.4182.3169.8682.3982.0281.7081.86
72.7783.0683.7884.2889.0271.8782.8183.1683.4783.31
73.6082.2284.0885.3084.6173.8584.8484.7384.6284.67
74.0783.9684.6685.1684.9074.5186.7585.6784.6185.10
1/2 71.8777.6182.6785.3283.4874.1781.7384.1586.0484.95
73.5280.2083.8085.9184.5875.3384.1885.2786.2085.70
73.3578.9583.6586.2284.4875.2384.1285.2086.1385.63
73.9581.6684.2085.8084.8575.6284.1185.3986.4985.90
Table 7. Quantitative evaluation of method contributions on Amazon Forest dataset. (✓ denotes the incorporation of this module).
Table 7. Quantitative evaluation of method contributions on Amazon Forest dataset. (✓ denotes the incorporation of this module).
RatioMethodValidationTest
BaseMPEVALmIoUAccuracymPrecisionmRecallmF1-ScoremIoUAccuracymPrecisionmRecallmF1-Score
1/16 87.0492.7793.0893.0593.0689.2995.2294.3594.3494.34
87.6193.0093.4193.3793.3989.8495.1794.6494.6494.64
91.6095.9995.6095.6395.6192.2796.4695.9895.9895.98
91.7096.7395.7995.8595.8192.6496.7696.1896.1896.18
1/8 88.7394.8694.0094.0694.0290.0996.0694.8194.7994.78
89.1494.4094.2594.2694.2590.8196.0395.1995.1895.18
91.9095.9495.7795.7895.7792.6096.2796.1596.1596.15
92.0296.7895.8295.8895.8492.8296.9596.2896.2796.27
1/4 88.6995.0393.9894.0494.0088.6194.8493.9793.9693.96
89.9595.7194.6894.7594.7091.4296.7995.5495.5295.51
92.1496.0995.9095.9195.9192.8596.5896.2996.2996.29
92.4896.4596.0896.1096.0993.0296.6296.3896.3896.38
1/2 89.9697.8894.7894.8494.7190.2098.0495.0394.8694.84
90.6997.8995.1595.2395.1191.0697.9795.4495.3395.32
91.5297.6295.5695.6595.5791.9797.5695.8795.8295.81
91.7197.2895.6695.7495.6792.1297.1595.9295.9095.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, B.; Wang, L.; Fan, X.; Bo, W.; Yang, X.; Tjahjadi, T. Semi-FCMNet: Semi-Supervised Learning for Forest Cover Mapping from Satellite Imagery via Ensemble Self-Training and Perturbation. Remote Sens. 2023, 15, 4012. https://doi.org/10.3390/rs15164012

AMA Style

Chen B, Wang L, Fan X, Bo W, Yang X, Tjahjadi T. Semi-FCMNet: Semi-Supervised Learning for Forest Cover Mapping from Satellite Imagery via Ensemble Self-Training and Perturbation. Remote Sensing. 2023; 15(16):4012. https://doi.org/10.3390/rs15164012

Chicago/Turabian Style

Chen, Beiqi, Liangjing Wang, Xijian Fan, Weihao Bo, Xubing Yang, and Tardi Tjahjadi. 2023. "Semi-FCMNet: Semi-Supervised Learning for Forest Cover Mapping from Satellite Imagery via Ensemble Self-Training and Perturbation" Remote Sensing 15, no. 16: 4012. https://doi.org/10.3390/rs15164012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop