Next Article in Journal
RS-DARTS: A Convolutional Neural Architecture Search for Remote Sensing Image Scene Classification
Previous Article in Journal
Evaluation of BRDF Information Retrieved from Time-Series Multiangle Data of the Himawari-8 AHI
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Positive Unlabelled Learning for Satellite Images’Time Series Analysis: An Application to Cereal and Forest Mapping

1
INRAE, UMR TETIS, University of Montpellier, 34000 Montpellier, France
2
Syngenta Seeds, 31790 Saint-Sauveur, France
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(1), 140; https://doi.org/10.3390/rs14010140
Submission received: 25 October 2021 / Revised: 17 December 2021 / Accepted: 25 December 2021 / Published: 29 December 2021
(This article belongs to the Topic Big Data and Artificial Intelligence)

Abstract

:
Applications in which researchers aim to extract a single land type from remotely sensed data are quite common in practical scenarios: extract the urban footprint to make connections with socio-economic factors; map the forest extent to subsequently retrieve biophysical variables and detect a particular crop type to successively calibrate and deploy yield prediction models. In this scenario, the (positive) targeted class is well defined, while the negative class is difficult to describe. This one-class classification setting is also referred to as positive unlabelled learning (PUL) in the general field of machine learning. To deal with this challenging setting, when satellite image time series data are available, we propose a new framework named positive and unlabelled learning of satellite image time series (PUL-SITS). PUL-SITS involves two different stages: In the first one, a recurrent neural network autoencoder is trained to reconstruct only positive samples with the aim to higight reliable negative ones. In the second stage, both labelled and unlabelled samples are exploited in a semi-supervised manner to build the final binary classification model. To assess the quality of our approach, experiments were carried out on a real-world benchmark, namely Haute-Garonne, located in the southwest area of France. From this study site, we considered two different scenarios: a first one in which the process has the objective to map Cereals/Oilseeds cover versus the rest of the land cover classes and a second one in which the class of interest is the Forest land cover. The evaluation was carried out by comparing the proposed approach with recent competitors to deal with the considered positive and unlabelled learning scenarios.

Graphical Abstract

1. Introduction

Currently, modern remote sensing systems provide image acquisitions describing the Earth’s surface at a high spatial resolution and revisit time period. A notable example is the Copernicus programme, which, through the Sentinel-2 mission, provides optical multi-spectral images within the visible and near-infrared regions (electromagnetic spectrum) with a spatial resolution between 10 m and 60 m and a revisit time period of approximately 5 days [1]. Such a stream of information can be profitably organised as satellite image time series (SITS) and can support a wide range of application domains, such as ecology [2], agriculture [3], mobility, health, risk assessment [4], land management planning [5], and forest [6], and natural habitat monitoring [7]. Due to the large panel of applications that can be supported by such data, it constitutes a valuable source of information to monitor the dynamics of the Earth’s surface.
One of the main tasks related to SITS data analysis is associated with land cover mapping, where a classification model is learned to make the connection between satellite data (i.e., SITS) and the associated land cover classes [5]. SITS data capture the temporal dynamics exhibited by land cover classes, thus resulting in a more effective discrimination among them [8].
Land cover maps can be the final outcome of the analysis or it can constitutes an intermediate step of a more general pipeline [9]. Examples of the latter case are the mapping of a specific culture versus the rest of the land cover classes to successively calibrate and deploy a yield prediction model [9], distinguish between forest areas and the rest of the land cover classes to subsequently retrieve biophysical variables such as canopy height or biomass volume [10], or extract the settlement footprint detecting urban vs. non-urban land covers to make a connection with socio-economic factors [11]. In these particular applications, the objective is to obtain binary land cover maps where one category (the positive one) is well defined since it is associated with the land cover of interest, while the other category (the negative one) is difficult to describe since it gathers together the rest of the possible land cover classes present in a particular study area. This one-class classification setting is also referred to as positive unlabelled learning (PUL) [12] in the general field of machine learning. Different from the standard supervised classification context, in the positive unlabelled learning setting, training data are composed of a set of positive samples and a set of unlabelled ones with the latter involving both positive and negative samples. The objective of a PUL framework is to build a binary classification model that can make predictions on unseen data distinguishing between positive and negative samples.
With the aim to focus on the analysis of SITS data, in this work, we propose a new framework, named PUL-SITS, to deal with positive and unlabelled learning for land cover mapping from satellite image time series. PUL-SITS involves two different stages: In the first one, a recurrent neural network autoencoder is trained only on positive samples. Successively, the same autoencoder model is employed to filter out reliable negative samples from the unlabelled ones. In the second stage, both labelled (positive and reliable negative) and unlabelled samples are exploited in a semi-supervised manner to build the final binary classification model. In this stage, the unlabelled samples are employed to force consistency between the prediction of the classification model and a variant of the same network trained on the SITS data reconstructed by the recurrent neural network autoencoder model.
To assess the effectiveness of PUL-SITS, we considered two different scenarios involving data coming from a particular study area (Haute-Garonne) located in the southwest area of France. More precisely, the Haute-Garonne study area is heavily characterized by the Cereals/Oilseeds and Forest land cover classes. For this reason, we designed a first scenario in which the Cereals/Oilseeds land cover class has the role of the positive class, while the rest of the land cover classes play the role of the negative class. In the second scenario, the Forest land cover class is targeted as the positive class. Experiments were carried out comparing the proposed approach w.r.t. well-suited competitors to deal with the considered positive and unlabelled learning scenarios. The remainder of this paper is structured as follows: First, Section 2 reviews related work in literature at the intersection of positive and unlabelled learning and remote sensing. The proposed framework is described in Section 3. The study site and the associated data are presented in Section 4. The experimental settings, as well as the evaluation are reported and discussed in Section 5. Finally, Section 6 draws the conclusion.

2. Related Work

Regarding the remote sensing domain, Reference [13] investigated the use of positive and unlabelled learning to cope with urban footprint mapping considering high-spatial-resolution aerial photographs. The PUL algorithm trains a classifier on positive and unlabelled data, estimating the probability that a positive training sample has been labelled, and finally, it generates binary predictions for test samples using an adjusted threshold. Reference [14] proposed an ensemble approach to deal with one-class classification under a PUL setting. Here, the class of interest was still the urban land cover, and the approach was evaluated considering different sensors (Landsat 8, WorldView-3, and Gaofen-1). Similar to [13], also in this case, the classification method was deployed on single-date images. In [15], the authors proposed to employ positive and unlabelled learning to characterize landslide susceptibility from several features extracted from multi-source data (elevation, slope angle, curvature, slope aspect, normalized difference vegetation index, soil type, loess erosion intensity, and topographic wetness index). Recently, Reference [16] evaluated PUL-based classifiers on different land cover classes (i.e., trees, bare soil), highlighting that the proposed strategy is well adapted for a plethora of different application contexts.
Unfortunately, despite the fact that the positive unlabelled learning setting naturally fits a wide range of remote sensing applications in which researchers aim to extract a binary land cover map with a focus on a particular class of interest, all the proposed strategies were developed considering mono-date images [13,14,16], while, today, SITS data are ubiquitous in the remote sensing community. According to our literature survey, no research study has yet proposed dealing with SITS data, explicitly with the temporal dynamics carried out by such a rich source of information, under the positive and unlabelled learning setting.

3. Method

In this section, we briefly introduce the positive and unlabelled learning setting, and successively, we describe PUL-SITS, a new framework to deal with land cover mapping from satellite image time series under the PUL setting.
We remind the reader that we considered satellite image time series at the pixel level as the unit of analysis. We refer to a generic SITS (at the pixel level) as T S = ( x 1 , , x t , , x T ) where T is the length of the time series and a generic element x t is a multi-variate (or multidimensional) vector with the subscript t, the corresponding timestamp. The PU learning setting considers a scenario in which we dispose of D = P U a set of samples { T S i } i = 1 | D | composed by a set P of positive samples and a set U of unlabelled ones. The set U contains both positive and negative samples, but their label information is not accessible. The task of learning from both positive and unlabelled data consists of exploiting both labelled P and unlabelled U samples to learn a binary classification model, allowing the assignment of a binary label (positive or negative) to new, previously unseen, satellite image time series.
Regarding PUL-SITS, our framework is constituted of two different stages. The first stage is devoted to selecting a set of reliable negative samples that highly differ from the positive set. This step is addressed via a variational recurrent neural network autoencoder (VAERNN), which is used to model the set of positive samples.
In the second stage, PUL-SITS learns a satellite image time series classifier, based on the recurrent neural network (RNN) model proposed in [17], to cope with the underlying binary classification task. In this stage, a semi-supervised training procedure is designed in order to take advantage of both labelled (positive and reliable negative) and unlabelled samples when the binary classification model is trained.

3.1. Selection of Reliable Negative Samples

The first stage of our framework addresses the identification of a set of reliable negative ( R N ) samples. A sketch of this procedure is depicted in Figure 1. To this end, we first modelled the positive set of satellite image time series (at the pixel level) via a variational RNN autoencoder (VAERNN), then we used the trained neural network to select reliable negative R N samples from the set of unlabelled satellite image time series data. To this end, the reconstruction error was used as the evaluation metric. Samples with a small reconstruction error are likely to belong to the positive class, while samples on which the reconstruction error is high are probably those representing the negative class. The use of recurrent neural networks (RNNs) to model the data distribution naturally arises due to the fact that such a model is especially tailored to managing multi-variate sequential data [18] and, in particular, compressing sequential information to extract useful multi-variate time series representations [19,20].
In our case, we used a variational autoencoder [21] based on a two-level gated recurrent unit [22]. Gated recurrent units are preferred to long-short term memory [23] since the former model has demonstrated better performance than the latter one in several remote-sensing-related applications [24,25]. The gated recurrent units have fewer parameters than the LSTM cells, and this makes the learned cell more robust and easy to train; finally, GRU cells have also exhibited good performance for reconstruction tasks and, more generally, for unsupervised analysis [20,21].
The gated recurrent unit (GRU) is formally defined as follows:
z t = σ ( W z x x t + W z h h t 1 + b z )
r t = σ ( W r x x t + W r h h t 1 + b r )
h t = z t h t 1 + ( 1 z t ) tanh ( W h x x t + W h r ( r t h t 1 ) + b h )
The ⊙ symbol indicates an elementwise multiplication, while σ and tanh represent the sigmoid and hyperbolic tangent functions, respectively. x t is the timestamp input vector, and h t 1 is the hidden state of the recurrent unit at time t 1 . The different weight matrices W * * and bias vectors b * are parameters learned during the training of the model.
This unit follows the general philosophy of modern recurrent neural network models implementing gates and cell states. The GRU cell has two gates, update ( z t ) and reset ( r t ), and one cell state, the hidden state ( h t ). Moreover, the two gates combine the current input ( x t ) with the information coming from the previous timestamps ( h t 1 ). The update gate effectively controls the tradeoff between how much information from the previous hidden state will carry over to the current hidden state and how much information of the current timestamp needs to be kept. On the other hand, the reset gate monitors how much information of the previous timestamps needs to be integrated with the current information. As each hidden cell has separate reset and update gates, they are able to capture dependencies over different time scales. Cells more prone to capturing short-term dependencies will tend to have a frequently activated reset gate, but those that capture longer-term dependencies will have update gates that remain mostly active [22]. Figure 2 visually depicts the GRU cell.
The encoder involves two GRU cells, while the decoder is symmetric to the encoder. Variational autoencoders, as opposed to standard ones, model the latent space to match a Gaussian distribution [26] (mean and variance). Such a way of compressing the input information can be more appropriate to manage complex data distributions such as in the case of satellite image time series, as recently highlighted in [27].
Furthermore, for an element at timestamps t output by the autoencoder, we performed an additional linear transformation to obtain the reconstructed element x ^ t , since, considering regression tasks, the softmax layer is commonly replaced by a fully connected layer with no activation (linear transformation) [28]. Finally, the variational autoencoder model was trained end-to-end by means of a reconstruction loss function on the set of positively labelled satellite image time series P:
L O S S V A E R N N = T S P H L o s s ( T S , L i n e a r ( VAERNN ( T S ) ) ) | P |
where ( L i n e a r ( VAERNN ( T S ) ) ) is the variational recurrent neural network autoencoder followed by a linear transformation and HLoss is the Huber loss [28], defined as follows:
H L o s s ( y , y ^ ) = 1 2 ( y y ^ ) 2 if | y y ^ | < δ δ | y y ^ | 1 2 δ 2 otherwise
Basically, the Huber loss behaves as the mean squared error (MSE) loss when the error is lower than a threshold δ , while it mimics the mean absolute error (MAE) loss otherwise. In addition, it is also differentiable at 0. The Huber loss combines the good properties from both the MAE and MSE, and it avoids their limitations. As opposed to the MAE, it bypasses large gradient back propagation when the estimated quantity becomes closer to the real value, and different from the MSE loss, it is more robust to outliers. To manage the balance between the two components of the Huber loss function, we relied on the default parameter setting [29] in which the σ value is set to 1.
Once the VAERNN is trained on the set of positive SITS data, it is successively deployed on the set of unlabelled time series U. For each time series T S U , a reconstruction error is computed via the Huber loss by comparing the original multi-variate time series with the reconstructed one.
Then, we computed the average reconstruction error μ r e c (the mean reconstruction error value on the set of unlabelled time series). After that, we considered multi-variate time series (belonging to the unlabelled set) with a reconstruction error greater than μ r e c , and we named such a set U μ r e c . Finally, we randomly selected elements from U μ r e c to build the set of reliable negative samples R N with a size equal to the size of the positive set P such that | P | = = | R N | .

3.2. Semi-Supervised Training Procedure

Once both P and R N sets were available, we trained a classification model to deal with the underlying binary classification task. As a classification backbone for our PUL framework, we considered the RNN architecture recently proposed in [17] and referred to as the fully connected gated recurrent unit (FCGRU). We chose such a model due to the interesting performance it has exhibited w.r.t. other state-of-the-art approaches for land cover mapping from SITS data. The FCGRU cell extends the standard GRU cell with two additional fully connected layers that are applied to the input multi-variate time series, with the aim to enrich the initial representation before deploying the GRU cell, as illustrated in Figure 2. More formally, the FCGRU cell is defined as follows:
x t = tanh ( W 2 tanh ( W 1 x t + b 1 ) + b 2 )
z t = σ ( W z x x t + W z h h t 1 + b z )
r t = σ ( W r x x t + W r h h t 1 + b r )
h t = z t h t 1 + ( 1 z t ) tanh ( W h x x t + W h r ( r t h t 1 ) + b h )
where the ⊙ symbol indicates an elementwise multiplication, while σ and tanh represent the sigmoid and hyperbolic tangent function, respectively. x t is the timestamp input vector, and h t 1 is the hidden state of the recurrent unit at time t 1 . The different weight matrices W * * and bias vectors b * are the parameters learned during the training of the model.
The classification model is trained via a semi-supervised procedure with the aim to leverage as much as possible the information available considering both labelled and unlabelled samples [30,31]. The procedure is depicted in Figure 3. More precisely, we considered a consistency-regularization-based strategy where the binary classification model is forced to behave similarly to an auxiliary model trained on the perturbed data. Such a consistency regularization strategy was deployed considering a reduced set of unlabelled samples U ˜ = U \ U μ r e c .
We denote by f θ the binary classification model. When only labelled data P and R N are considered, the model is trained via the standard binary cross-entropy (BCE) loss, defined as:
L o s s B C E ( Y , P R N , f θ ) = 1 | P R N | i = 1 | P R N | y i log ( f θ ( T S i ) ) + ( 1 y i ) log ( f θ ( 1 T S i ) )
where T S i P R N is a generic satellite image time series sample with an associated label y i Y and Y is the set of labels.
With the aim to exploit available unlabelled data U ˜ in the learning process, we added an extra term to the main loss function as follows:
L o s s = L o s s B C E ( Y , P R N , f θ ) + λ 1 | U ˜ | j = 1 | U ˜ | K L ( f θ ^ ( T S j ^ ) | | f θ ( T S j ) )
where T S j U ˜ and f θ ^ is an auxiliary classification model that has the same network structure as f θ , but trained on P ^ and R N ^ , which are the reconstructed (via the VAERNN autoencoder) positive and reliable negative samples, respectively. K L ( X | | Z ) is the standard Kullback–Leibler divergence, which quantifies how much the distribution Z differs from the distribution X. The second term of the semi-supervised loss function has the goal of leveraging the unlabelled set of SITS data, U ˜ , regularizing the classification behaviour via consistency assessment [31]. This is achieved by imposing that the output distribution of the f θ model should behave similarly to the output distribution of a perturbed version of itself, named f θ ^ , trained on the set of reconstructed positive and reliable negative samples. Such a strategy implicitly enforces the classifier f θ to produce similar output distributions considering a specific sample and its perturbed version, thus stressing smoothness on the manifold on which the data are projected by the neural network model [31]. The trade-off hyper-parameter λ is equal to 2.
To implement such a training strategy, at each iteration of an epoch, the classification models f θ and f θ ^ were trained simultaneously. Firstly, f θ ^ was trained on a batch of samples coming from the perturbed data, and successively, the weight parameters of the f θ model were updated considering the binary cross-entropy loss on a batch of labelled samples from P R N plus the consistency regularization loss on a batch of unlabelled samples from the set U ˜ .
At the end of the training procedure, the classifier f θ is the model that is employed to perform inference on new unseen samples.

4. Data

The study area we considered in our research study, namely Haute-Garonne, is located in the southwest area of France (Figure 4a), and it covers a surface area of 4146.2 km2. The satellite data describing the study site consist of a Sentinel-2 (S2) time series involving 31 images acquired over the year 2019 from January to December. The temporal distribution of the satellite image time series is depicted in Figure 5.
The Sentinel-2 images were downloaded from the THEIA pole platform (http://theia.cnes.fr, accessed on 10 November 2020) at level-2A in top-of-canopy reflectance values and provided with cloud masks. Only 10 m spatial resolution bands (blue, green, red, and near-infrared spectrum) were considered in this analysis. Additionally, two common spectral indices were then extracted: the NDVI [32] and the NDWI [33]. A preprocessing step was performed over each band to replace cloudy pixel values as detected by the available cloud masks through a linear multi-temporal interpolation (cf. temporal gap-filling [5]). The entire study site is enclosed in the Sentinel-2 tile T31TCJ.
Since we are addressing a positive and unlabelled learning setting, we considered two different scenarios where each scenario involves a particular land cover class as the positive class and all the other land cover classes as negative. As positive classes, we considered the Cereals/Oilseeds and the Forest land covers.
The statistics about the Cereals/Oilseeds and Forest land cover classes, in the Haute-Garonne study site, are reported in Table 1. We also report the statistics for the number of labelled samples belonging to the other land cover classes present in the Haute-Garonne study area (i.e., Water, Urban, Orchards, etc.) that are neither Cereals/Oilseeds nor Forest samples.
The ground truth (GT) was obtained from various sources: (i) the Registre Parcellaire Graphique (RPG) reference data for 2019 (https://www.data.gouv.fr/en/datasets, accessed on 10 November 2020), which is part of the European Land Parcel Identification System (LPIS) provided by the French Agency for services and payment; (ii) the French Topographic Database (BD TOPO) (https://geoservices.ign.fr/blog/2020/10/26/Grand_angle_diffusion_BDTOPO.html, accessed on 10 November 2020), which contains geographical and administrative entities of the national territory; (iii) the French National Forest inventory (BD FORET), which includes information about forest cover at the national scale. The GT was assembled in a Geographic Information System (GIS) vector file, containing a collection of polygons, each attributed to a land cover category (see Table 1). Finally, the polygons were rasterized at the Sentinel-2 spatial resolution (10 m).

5. Experimental Evaluation

In this section, we describe the experimental evaluation we conducted with the aim to assess the behaviour of our framework. To this end, we firstly compared PUL-SITS with different competing approaches on the study site introduced in Section 4. Secondly, we evaluated the different components on which PUL-SITS is based on through an ablation study. Finally, we visually inspected the land cover maps produced by the different competing approaches.

5.1. Competing Methods and Ablation

To assess the behaviour of PUL-SITS, we considered the following competitors:
  • A random forest approach based on the positive unlabelled learning framework introduced in [34]. Here, the authors proposed to exploit probability theory to weight unlabelled samples with the aim to bias the learning stage of standard supervised classification methods. We named such a competitor RFPUL;
  • A one-class support vector machine approach [35]. In this case, the learning approach discards unlabelled samples, and it only leverages positive samples to build the final binary classification model. We named such a competing approach OCSVM;
  • An ensemble-based approach for the positive unlabelled learning scenario recently introduced in the field of remote sensing analysis [14]. The ensemble approach combines the prediction of different base learners via a mean-squared-error-based aggregation function. Our ensemble-based competitor involves the following base learners: logistic ridge regression, logistic elastic-net regression, random forest, multilayer perceptron and linear discriminant analysis. We named such competitors ENSEMBLEPUL.
In order to disentangle the contributions of the different stages on which PUL-SITS is based, we considered the following ablations of our framework:
  • The FCGRU [17] classification model only trained on the positive and reliable negative samples obtained from the first stage of our framework. This ablation did not involve the consistency loss. We named such an ablation PUL-SITSnoReg;
  • The FCGRU classification model trained on the positive and reliable negative samples obtained from the first stage of our framework. Here, the classifier was not trained on the original data, but on the reconstructed data output by the autoencoder model. We named such an ablation PUL-SITSreco.

5.2. Experimental Settings

We designed two different positive unlabelled learning experiments according to the data presented in Section 4. We remind the reader that in the first (resp. second) experiment, we considered as the positive class the Cereals/Oilseeds (resp. Forest) land cover class, while the negative class was constituted by all the samples not belonging to the land cover class representing the positive class.
With the purpose of setting up a fair evaluation protocol to assess the quality of the different competitors, for each experiment, we divided the data (both positive and negative classes) into two sets: training and test. We considered 50% of the positive (resp. negative) class as the test data. The remaining 50% of the data were considered as the training data. Then, the training set was split again into two parts: the positive and the unlabelled set. While the former contained only positive samples, the latter consisted of samples from both the positive and negative classes. Whereas the amount of positive samples can influence the model behaviour, we considered increasing the quantity of positive samples ranging in the set { 20 , 40 , 60 , 80 , 100 } in terms of objects. Once the set of positive samples was fixed, the rest of the training data were considered as unlabelled data. The number of objects and pixels for the positive, unlabelled, and negative samples in the training and test sets for the first (Cereals/Oilseeds land cover as the positive class) and second (Forest land cover as the positive class) experiments is reported in Table 2 (resp. Table 3).
We remind the reader that, despite having defined the size of the positive samples in terms of objects, the classification procedure was performed at the pixel level (all the pixels belonging to the object belong exclusively to the training or the test set). We organized the training/test data split procedures in terms of objects with the aim to avoid possible spatial bias in the evaluation procedure [5], since, in this way, pixels belonging to the same ground truth object were assigned exclusively to one of the data partitions. In addition, for time series data, all values were normalized per band in the interval [ 0 , 1 ] using 2 % (resp. 98 % ), rather than the minimum (resp. maximum) value to rescale the data [8].
Regarding our framework, PUL-SITS, the VAERNN involves an encoder with two levels with 64 and 16 units, respectively, while the decoder is symmetric to the encoder. The FCGRU [17] recurrent cell involved in the binary classification process has a number of units equal to 32. Both models were trained for 50 epochs with a batch size equal to 32 and a learning rate of 10 e 4 through the Adam optimizer. For the latter component, we also employed a dropout rate equal to 0.2. The same learning procedure was employed for the PUL-SITSnoReg and PUL-SITSreco ablations. Regarding the competing methods, RFPUL, OCSVM, and ENSEMBLEPUL, they were employed with the default settings.
The assessment of the quantitative performances was performed considering the Accuracy, F-Measure, Sensitivity, Specificity, and Kappa metrics [36]. To reduce the bias induced by the training/test data split, all the results were averaged over ten different random splits. Finally, both the average and standard deviation are reported for each evaluation metric.
The experiments were carried out on a workstation with an AMD Ryzen 7 3700X 8-core processor [email protected] with 64 GB of RAM and a GeForce RTX 2080 Ti GPU. All the deep-learning methods were implemented using the Python Tensorflow library.

5.3. Quantitative Results

As described before, we compared PUL-SITS with respect to the competing methods on two different experiments: the first one involved the Cereals/Oilseeds land cover as the class of interest, while the second experiment considered the Forest land cover as the positive class. In both cases, we firstly assessed the performances of the competing strategies in terms of the standard evaluation metrics, then we discuss the associate confusion matrices; finally, we provide a qualitative analysis, inspecting some extracts of the produced land cover maps.

5.3.1. Experiments with the Cereals/Oilseeds Land Cover as the Positive Class

Table 4, Table 5, Table 6, Table 7 and Table 8 summarise the results in terms of the F-Measure, Accuracy, Sensitivity, Specificity, and Kappa measure, respectively. Generally, we can observe that no matter the size of the training data, our framework always outperformed all the competing approaches considering almost all the evaluation metrics. We can also note that all the approaches exhibited better behaviours as the amount of available positive labelled data increased.
Regarding the competing methods, we can observe that OCSVM achieved the best performance. Such a method, as opposed to RFPUL and ENSEMBLEPUL, only exploited the positive samples to build its classification model. This fact underlines that RFPUL and ENSEMBLEPUL are probably impacted by the fact that the negative class (all the land cover classes except Cereals/Oilseeds) is characterised by a high intra-heterogeneity, thus influencing the ability of the algorithm to model such a wide range of sample as one class in the original feature space. This fact is illustrated by high Sensitivity (Table 6) and low Specificity (Table 7) values, which highlight that RFPUL and ENSEMBLEPUL simply label the majority of the data points with the negative class. Concerning the ablations associated with our framework, PUL-SITSnoReg and PUL-SITSreco, we can see that both strategies achieved better performances than the competing methods, with PUL-SITSnoReg as a clear winner between them, but the PUL-SITSreco component played an important role in supporting the performance of PUL-SITS, underlying the quality of our two-stage methodology. Figure 6 shows the confusion matrices associated with the three direct competing approaches, as well as PUL-SITS. A closer look at these statistics confirms that PUL-SITS was more precise than the competitors. This consideration emerges from the fact that the corresponding heat map (Figure 6d) has a more visible diagonal structure (the dark green blocks concentrated on the diagonal). This is not the case for the other methods, where the majority of the samples are associated with the negative (zero) class.

5.3.2. Experiments with the Forest Land Cover as the Positive Class

Table 9, Table 10, Table 11, Table 12 and Table 13 resume respectively the results in terms of the F-Measure, Accuracy, Sensitivity, Specificity, and Kappa of the different competing methods when the Forest land cover was considered as the positive class. In this case also, PUL-SITS outperformed all the other approaches, while RFPUL and ENSEMBLEPUL simply labelled all of the data points as the negative class (as highlighted by the high values of the Sensitivity and the low values of the Sensibility they achieved). OCSVM exhibited the best performance among the state-of-the-art competitors. Concerning the ablations, PUL-SITSnoReg did not always achieve better performance over OCSVM, while PUL-SITSreco did. This phenomena clearly underlines the importance of exploiting the available unlabelled data and the appropriateness of the consistency-based procedure in order to regularize the training stage of the final classification model. Figure 7 shows the confusion matrices, which confirm that PUL-SITS achieved a better predictive performance w.r.t. the other approaches, i.e., it has a more visible diagonal structure.

5.4. Visual Inspection of the Binary Classification Maps

Here, we performed a visual inspection of the binary classification maps generated via the different competing approaches regarding the two scenarios we previously described.
Figure 8 and Figure 9 report the results of the mapping process on two different extracts of the study site when the Cereals/Oilseeds land cover class was considered as the positive class. The first extract (Figure 8) depicts an agricultural area mainly characterized by cereal and oilseed parcels. We can observe that PUL-SITS clearly recovered the parcel structure, while the other approaches had serious issues in detecting pixels belonging to the Cereals/Oilseeds class. While RFPUL and ENSEMBLEPUL exhibited high false-negative rates, the OCSVM method showed reasonable performance on the considered extract, but still (visually) inferior w.r.t. our framework. A similar behaviour was exhibited by all the competing methods on the second extract depicted in Figure 9. In this case also, PUL-SITS provided a more coherent land cover map that clearly discriminated between positive (Cereals/Oilseeds) and negative pixels.
Figure 10 and Figure 11 depict the results of the mapping process on two different extracts of the study site when the Forest land cover class was considered as the positive class. The first extract (Figure 8) shows a portion of the study site where clear forest cover is present. Here, we can note that PUL-SITS well identified the large forest area in the bottom left part of the extract. All the other methods were not capable of recovering all the forest extent with, also in this case, issues related to false-negative detection. The second extract (Figure 11) describes a zone on which the forest cover is not concentrated, but, conversely, spread over all the area. Similar to what happened in the previous case, PUL-SITS produced a binary land cover map that was consistent with what we can observe in the RGB composite on the left side of the figure. To sum up, we can state that the visual inspection of the binary land cover maps, considering both scenarios, is coherent with the quantitative results we obtained in the previous stage of the evaluation analysis.

6. Conclusions

In this research work, we introduced a novel framework, namely PUL-SITS, to deal with the positive unlabelled learning analysis of satellite image time series data for the task of land cover mapping. Our approach was based on a two-step strategy: in the first one, a recurrent neural network autoencoder was trained only on positive samples. Successively, the same autoencoder model was employed to filter out reliable negative samples. In the second stage, both labelled (positive and reliable negative) and unlabelled samples were exploited in a semi-supervised manner to build the final binary classification model. To assess the performances of PUL-SITS, we compared it with state-of-the-art methods for positive unlabelled learning considering two different scenarios: a first one in which the Cereal/Oilseeds land cover class has the role of the positive class while the rest of the land cover classes play the role of the negative class and a second scenario where the Forest land cover class was targeted as the positive class. Both scenarios underlined the quality of our approach. As future work, we plan to extend our framework towards multi-source analysis and introduce, for instance, satellite image time series of radar sensors (i.e., Sentinel-1) in order to exploit the complementarity between different available sources/modalities.

Author Contributions

D.I. and J.D. conceived and designed the experiments; J.D. performed the experiments; J.D. and D.I. analyzed the results; D.I. and J.D. wrote the article; D.I., A.B. and N.R. revised the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by ANRT, grant number CIFRE 2019/1993.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors wish also to thank the European Space Agency (ESA) for the Sentinel-2 data and Theia pole for the calibration of the Sentinel-2 images.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Berger, M.; Moreno, J.; Johannessen, J.; Levelt, P.; Hanssen, R. ESA’s sentinel missions in support of Earth system science Keywords: Sentinel ESA Earth system science Global Monitoring for Environment and Security (GMES). Remote Sens. Environ. 2012, 120, 84–90. [Google Scholar] [CrossRef]
  2. Chen, L.; Jin, Z.; Michishita, R.; Cai, J.; Yue, T.; Chen, B.; Xu, B. Dynamic monitoring of wetland cover changes using time-series remote sensing imagery. Ecol. Inform. 2014, 24, 17–26. [Google Scholar] [CrossRef]
  3. Bellón, B.; Bégué, A.; Seen, D.L.; de Almeida, C.A.; Simões, M. A Remote Sensing Approach for Regional-Scale Mapping of Agricultural Land-Use Systems Based on NDVI Time Series. Remote Sens. 2017, 9, 600. [Google Scholar] [CrossRef] [Green Version]
  4. Olen, S.; Bookhagen, B. Mapping Damage-Affected Areas after Natural Hazard Events Using Sentinel-1 Coherence Time Series. Remote Sens. 2018, 10, 1272. [Google Scholar] [CrossRef] [Green Version]
  5. Inglada, J.; Vincent, A.; Arias, M.; Tardy, B.; Morin, D.; Rodes, I. Operational High Resolution Land Cover Map Production at the Country Scale Using Satellite Image Time Series. Remote Sens. 2017, 9, 95. [Google Scholar] [CrossRef] [Green Version]
  6. Wulder, M.A.; Masek, J.G.; Cohen, W.B.; Loveland, T.R.; Woodcock, C.E. Opening the archive: How free data has enabled the science and monitoring promise of Landsat Author links open overlay panel. Remote Sens. Environ. 2012, 122, 2–10. [Google Scholar] [CrossRef]
  7. Khiali, L.; Ienco, D.; Teisseire, M. Object-oriented satellite image time series analysis using a graph-based representation. Ecol. Inform. 2018, 43, 52–64. [Google Scholar] [CrossRef] [Green Version]
  8. Pelletier, C.; Webb, G.; Petitjean, F. Temporal Convolutional Neural Network for the Classification of Satellite Image Time Series. Remote Sens. 2019, 11, 523. [Google Scholar] [CrossRef] [Green Version]
  9. Azzari, G.; Jain, M.; Lobell, D. Towards fine resolution global maps of crop yields: Testing multiple methods and satellites in three countries. Remote Sens. Environ. 2017, 202, 129–141. [Google Scholar] [CrossRef]
  10. Li, W.; Niu, Z.; Shang, R.; Qin, Y.; Wang, L.; Chen, H. High-resolution mapping of forest canopy height using machine learning by coupling ICESat-2 LiDAR with Sentinel-1, Sentinel-2 and Landsat-8 data. Int. J. Appl. Earth Obs. Geoinf. 2020, 92, 102163. [Google Scholar] [CrossRef]
  11. Xie, S.M.; Jean, N.; Burke, M.; Lobell, D.B.; Ermon, S. Transfer Learning from Deep Features for Remote Sensing and Poverty Mapping. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3929–3935. [Google Scholar]
  12. Bekker, J.; Davis, J. Learning from positive and unlabelled data: A survey. Mach. Learn. 2020, 109, 719–760. [Google Scholar] [CrossRef] [Green Version]
  13. Li, W.; Guo, Q.; Elkan, C. A Positive and Unlabeled Learning Algorithm for One-Class Classification of Remote-Sensing Data. IEEE Trans. Geosci. Remote Sens. 2011, 49, 717–725. [Google Scholar] [CrossRef]
  14. Ran, L.; Li, W.; Liu, X.; Lu, X.; Li, T.; Guo, Q. An Ensemble of Classifiers Based on Positive and Unlabeled Data in One-Class Remote Sensing Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 572–584. [Google Scholar]
  15. Wu, B.; Qiu, W.; Jia, J.; Liu, N. Landslide Susceptibility Modeling Using Bagging-Based Positive-Unlabeled Learning. IEEE Geosci. Remote Sens. Lett. 2021, 18, 766–770. [Google Scholar] [CrossRef]
  16. Li, W.; Guo, Q.; Elkan, C. One-Class Remote Sensing Classification From Positive and Unlabeled Background Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 730–746. [Google Scholar] [CrossRef]
  17. Gbodjo, Y.J.E.; Ienco, D.; Leroux, L.; Interdonato, R.; Gaetano, R.; Ndao, B. Object-Based Multi-Temporal and Multi-Source Land Cover Mapping Leveraging Hierarchical Class Relationships. Remote Sens. 2020, 12, 2814. [Google Scholar] [CrossRef]
  18. Linzen, T.; Dupoux, E.; Goldberg, Y. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies. Trans. Assoc. Comput. Linguist. 2016, 4, 521–535. [Google Scholar] [CrossRef]
  19. Kieu, T.; Yang, B.; Guo, C.; Jensen, C.S. Outlier Detection for Time Series with Recurrent Autoencoder Ensembles. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 2725–2732. [Google Scholar]
  20. Ienco, D.; Interdonato, R. Deep Multivariate Time Series Embedding Clustering via Attentive-Gated Autoencoder. Adv. Knowl. Discov. Data Min. 2020, 12084, 318–329. [Google Scholar]
  21. Guo, Y.; Liao, W.; Wang, Q.; Yu, L.; Ji, T.; Li, P. Multidimensional Time Series Anomaly Detection: A GRU-based Gaussian Mixture Variational Autoencoder Approach. In Proceedings of the 10th Asian Conference on Machine Learning, ACML, PMLR, Beijing, China, 14–16 November 2018; Volume 95, pp. 97–112. [Google Scholar]
  22. Cho, K.; van Merrienboer, B.; Gülçehre, Ç.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  23. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A Search Space Odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Ndikumana, E.; Ho Tong Minh, D.; Baghdadi, N.; Courault, D.; Hossard, L. Deep Recurrent Neural Network for Agricultural Classification using multitemporal SAR Sentinel-1 for Camargue, France. Remote Sens. 2018, 10, 1217. [Google Scholar] [CrossRef] [Green Version]
  25. Bakhti, K.; Larabi, M.E.A. Comparing Deep Recurrent Learning and Convolutional Learning for Multi-Temporal Vegetation Classification. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 4392–4395. [Google Scholar]
  26. Doersch, C. Tutorial on Variational Autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  27. Sanchez, E.H.; Serrurier, M.; Ortner, M. Learning Disentangled Representations of Satellite Image Time Series. arXiv 2019, arXiv:1903.08863. [Google Scholar]
  28. Lathuilière, S.; Mesejo, P.; Alameda-Pineda, X.; Horaud, R. A Comprehensive Analysis of Deep Regression. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2065–2081. [Google Scholar] [CrossRef] [Green Version]
  29. Huber, P.J. Robust Estimation of a Location Parameter. Ann. Stat. 1964, 53, 73–101. [Google Scholar] [CrossRef]
  30. van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef] [Green Version]
  31. Yang, X.; Song, Z.; King, I.; Xu, Z. A Survey on Deep Semi-supervised Learning. arXiv 2021, arXiv:2103.00550. [Google Scholar]
  32. Rouse, J.W.; Hass, R.H.; Schell, J.; Deering, D. Monitoring vegetation systems in the great plains with ERTS. NASA Spec. Publ. 1973, 1, 309–317. [Google Scholar]
  33. cai Gao, B. NDWI—A normalized difference water index for remote sensing of vegetation liquid water from space. Remote Sens. Environ. 1996, 58, 257–266. [Google Scholar]
  34. Elkan, C.; Noto, K. Learning Classifiers from Only Positive and Unlabeled Data. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 213–220. [Google Scholar]
  35. Tran, Q.; Li, X.; Duan, H. Efficient performance estimate for one-class support vector machine. Pattern Recognit. Lett. 2005, 26, 1174–1182. [Google Scholar] [CrossRef]
  36. Tan, P.N.; Steinbach, M.; Kumar, V. Introduction to Data Mining; Addison Wesley: Boston, MA, USA, 2005. [Google Scholar]
Figure 1. The VAERNN is trained on the set of positively labelled samples P. Then, the model is applied on the set of unlabelled samples U. Successively, the reconstruction error for each T S U is derived, and the set of reliable negative samples R N is identified based on the reconstruction error. The vertical red line indicates the average reconstruction error ( μ r e c ) on the set of unlabelled samples U. Samples with a reconstruction error bigger than μ r e c are possible candidates for the reliable negative set R N .
Figure 1. The VAERNN is trained on the set of positively labelled samples P. Then, the model is applied on the set of unlabelled samples U. Successively, the reconstruction error for each T S U is derived, and the set of reliable negative samples R N is identified based on the reconstruction error. The vertical red line indicates the average reconstruction error ( μ r e c ) on the set of unlabelled samples U. Samples with a reconstruction error bigger than μ r e c are possible candidates for the reliable negative set R N .
Remotesensing 14 00140 g001
Figure 2. Visual representation of the GRU and FCGRU cells [17].
Figure 2. Visual representation of the GRU and FCGRU cells [17].
Remotesensing 14 00140 g002
Figure 3. The training strategy involves two models trained simultaneously. The first model f θ ^ is trained on the set of perturbed positive labelled P and reliable negative R N samples by the autoencoder with the supervised BCE loss on true labels y i . The second f θ is the final classification model trained on P and R N samples, with a semi-supervised process involving two losses: the binary cross entropy loss on the true labels y i ( L S ) and the KL divergence between the generated soft labels f θ ( T S j ) and f θ ^ ( T S ^ j ) to enforce consistency ( L U ). The λ hyper-parameter weights the importance of the unsupervised term.
Figure 3. The training strategy involves two models trained simultaneously. The first model f θ ^ is trained on the set of perturbed positive labelled P and reliable negative R N samples by the autoencoder with the supervised BCE loss on true labels y i . The second f θ is the final classification model trained on P and R N samples, with a semi-supervised process involving two losses: the binary cross entropy loss on the true labels y i ( L S ) and the KL divergence between the generated soft labels f θ ( T S j ) and f θ ^ ( T S ^ j ) to enforce consistency ( L U ). The λ hyper-parameter weights the importance of the unsupervised term.
Remotesensing 14 00140 g003
Figure 4. Haute-Garonne is located in the southwest of France (a). The spatial distribution of the land covers (Cereals/Oilseeds and Forest) considered as positive classes (b). The Sentinel-2 RGB composite obtained over the year 2019 (c).
Figure 4. Haute-Garonne is located in the southwest of France (a). The spatial distribution of the land covers (Cereals/Oilseeds and Forest) considered as positive classes (b). The Sentinel-2 RGB composite obtained over the year 2019 (c).
Remotesensing 14 00140 g004
Figure 5. Overview of the acquisition dates of Sentinel-2 (S2) images over Haute-Garonne. S2 acquisitions are sparsely distributed due to the ubiquitous cloudiness.
Figure 5. Overview of the acquisition dates of Sentinel-2 (S2) images over Haute-Garonne. S2 acquisitions are sparsely distributed due to the ubiquitous cloudiness.
Remotesensing 14 00140 g005
Figure 6. Confusion matrices of (a) RFPUL, (b) ENSEMBLEPUL, (c) OCSVM, and (d) PUL-SITS considering the case in which the land cover class of interest (the positive class) is Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Figure 6. Confusion matrices of (a) RFPUL, (b) ENSEMBLEPUL, (c) OCSVM, and (d) PUL-SITS considering the case in which the land cover class of interest (the positive class) is Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Remotesensing 14 00140 g006
Figure 7. Confusion matrices of (a) RFPUL, (b) ENSEMBLEPUL, (c) OCSVM, and (d) PUL-SITS considering the case in which the land cover class of interest (the positive class) is Forest with the number of positive training objects (n) equal to 80.
Figure 7. Confusion matrices of (a) RFPUL, (b) ENSEMBLEPUL, (c) OCSVM, and (d) PUL-SITS considering the case in which the land cover class of interest (the positive class) is Forest with the number of positive training objects (n) equal to 80.
Remotesensing 14 00140 g007
Figure 8. A first visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Figure 8. A first visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Remotesensing 14 00140 g008
Figure 9. A second visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Figure 9. A second visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Cereals/Oilseeds with the number of positive training data objects (n) equal to 80.
Remotesensing 14 00140 g009
Figure 10. A first visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Forest with the number of positive training data objects (n) equal to 80.
Figure 10. A first visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Forest with the number of positive training data objects (n) equal to 80.
Remotesensing 14 00140 g010
Figure 11. A second visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Forest with the number of positive training data objects (n) equal to 80.
Figure 11. A second visual evaluation zone with (a) its RGB composite and the corresponding binary classification maps for (b) RFPUL, (c) ENSEMBLEPUL, (d) OCSVM, and (e) PUL-SITS for the positive class Forest with the number of positive training data objects (n) equal to 80.
Remotesensing 14 00140 g011
Table 1. Haute-Garonne ground truth characteristics.
Table 1. Haute-Garonne ground truth characteristics.
Class# of Objects# of Pixels
Cereals/Oilseeds898336,044
Forest846186,870
Other land cover classes6460323,925
Total7358846,838
Table 2. Number of positive, unlabelled, and negative objects/pixels in the training and test set when the positive samples belong to the Cereals/Oilseeds land cover class.
Table 2. Number of positive, unlabelled, and negative objects/pixels in the training and test set when the positive samples belong to the Cereals/Oilseeds land cover class.
# of Objects Training (P/U)# of Pixels Training (P/U)# of Objects Test (P/N)# of Pixels Test (P/N)
20/36597010/410,712451/3228168,404/260,711
40/363914,618/403,104
60/361921,001/403,822
80/359928,955/388,767
100/357936,442/381,280
Table 3. Number of positive, unlabelled, and negative objects/pixels in the training and test set when the positive samples belong to the Forest land cover class.
Table 3. Number of positive, unlabelled, and negative objects/pixels in the training and test set when the positive samples belong to the Forest land cover class.
# of Objects Train (P/U)# of Pixels Train (P/U)# of Objects Test (P/N)# of Pixels Test (P/N)
20/36584206/413,517428/325095,725/333,390
40/36389565/408,157
60/361813,900/403,822
80/359817,865/399,857
100/357822,143/395,579
Table 4. Performances in terms of the F-Measure of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
Table 4. Performances in terms of the F-Measure of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
2046.2 ± 1.747.3 ± 1.561.6 ± 4.080.3± 2.375.5 ± 3.578.9 ± 4.0
4046.7 ± 1.749.3 ± 1.763.8 ± 2.784.9 ± 1.982.0 ± 2.385.3± 2.8
6047.2 ± 1.852.4 ± 1.664.7 ± 2.086.6 ± 2.884.2 ± 2.087.3± 2.2
8048.3 ± 2.055.7 ± 2.164.7 ± 2.386.5 ± 2.385.0 ± 2.588.0± 2.6
10050.2 ± 1.660.1 ± 1.864.3 ± 2.086.2 ± 3.085.5 ± 1.888.6± 1.3
Table 5. Performances in terms of the Accuracy of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
Table 5. Performances in terms of the Accuracy of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
2060.9 ± 1.461.3 ± 1.363.9 ± 2.180.9± 2.076.3 ± 3.179.5 ± 3.6
4061.1 ± 1.462.1 ± 1.265.1 ± 2.285.0 ± 1.982.3 ± 2.285.4± 2.8
6061.3 ± 1.463.5 ± 1.266.0 ± 1.686.6 ± 2.884.2 ± 2.087.3± 2.1
8061.8 ± 1.565.1 ± 1.365.4 ± 1.986.5 ± 2.385.1 ± 2.487.9± 2.5
10062.6 ± 1.367.5 ± 1.265.2 ± 1.786.1 ± 2.985.5 ± 1.888.6± 1.3
Table 6. Performances in terms of the Sensitivity of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
Table 6. Performances in terms of the Sensitivity of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
20100± 0.099.8 ± 0.181.5 ± 5.091.7 ± 4.188.1 ± 3.588.6 ± 8.9
40100± 0.099.6 ± 0.279.6 ± 3.988.4 ± 2.889.0 ± 1.890.3 ± 2.0
60100± 0.099.3 ± 0.377.3 ± 3.886.7 ± 5.887.3 ± 3.688.3 ± 2.5
80100± 0.099.1 ± 0.277.0 ± 3.686.6 ± 5.388.9 ± 2.588.7 ± 2.0
10099.9± 0.198.9 ± 0.277.7 ± 3.285.1 ± 5.987.4 ± 2.889.4 ± 2.0
Table 7. Performances in term of the Specificity of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
Table 7. Performances in term of the Specificity of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
200.3 ± 0.31.6 ± 0.836.9 ± 11.265.3 ± 9.859.3 ± 10.066.6± 13.6
400.9 ± 0.64.0 ± 2.242.7 ± 7.780.0± 4.772.4 ± 5.278.2 ± 5.3
601.4 ± 0.78.1 ± 2.147.4 ± 6.885.9 ± 2.979.6 ± 5.085.9± 5.0
802.7 ± 0.912.4 ± 3.047.7 ± 6.986.1 ± 4.279.4 ± 5.586.8± 4.5
1005.0 ± 1.218.8 ± 2.845.9 ± 6.087.6± 3.182.7 ± 3.787.3 ± 4.3
Table 8. Performances in term of the Kappa measure of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
Table 8. Performances in term of the Kappa measure of the different competing strategies when the land cover class of interest (the positive class) is Cereals/Oilseeds.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
200.00 ± 0.000.02 ± 0.010.19 ± 0.070.59± 0.050.49 ± 0.070.56 ± 0.07
400.01 ± 0.010.04 ± 0.020.23 ± 0.050.68 ± 0.040.62 ± 0.050.69± 0.06
600.02 ± 0.010.09 ± 0.020.25 ± 0.040.72 ± 0.050.67 ± 0.040.74± 0.04
800.03 ± 0.010.14 ± 0.030.25 ± 0.040.72 ± 0.040.69 ± 0.050.75± 0.05
1000.06 ± 0.010.21 ± 0.030.24 ± 0.040.72 ± 0.050.70 ± 0.030.76± 0.03
Table 9. Performances in terms of the F-Measure of the different competing strategies when the land cover class of interest (the positive class) is Forest.
Table 9. Performances in terms of the F-Measure of the different competing strategies when the land cover class of interest (the positive class) is Forest.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
2068.4 ± 2.769.8 ± 2.482.7 ± 3.086.1 ± 1.889.7 ± 2.091.4± 1.9
4069.7 ± 2.772.8 ± 3.183.1 ± 2.685.1 ± 3.489.8 ± 3.792.9± 2.3
6071.6 ± 2.976.4 ± 3.184.4 ± 2.586.5 ± 2.590.1 ± 2.292.6± 2.5
8073.9 ± 379.5 ± 3.184.7 ± 2.486.6 ± 2.988.8 ± 2.991.4± 3.5
10076.8 ± 2.783.1 ± 2.885.2 ± 2.485.0 ± 1.788.7 ± 2.891.9± 1.8
Table 10. Performances in terms of the Accuracy of the different competing strategies when the land cover class of interest (the positive class) is Forest.
Table 10. Performances in terms of the Accuracy of the different competing strategies when the land cover class of interest (the positive class) is Forest.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
2077.9 ± 1.978.4 ± 1.885.0 ± 2.183.2 ± 2.387.9 ± 2.690.1± 2.7
4078.4 ± 1.979.8 ± 2.185.8 ± 1.981.9 ± 4.688.0 ± 4.892.1± 2.8
6079.3 ± 2.081.7 ± 2.186.7 ± 1.883.8 ± 3.888.4 ± 2.991.5± 3.1
8080.4 ± 2.083.5 ± 2.286.9 ± 1.883.7 ± 3.986.7 ± 3.790.1± 4.3
10081.9 ± 1.985.8 ± 2.187.2 ± 1.881.8 ± 2.386.7 ± 3.790.8± 2.2
Table 11. Performances in terms of the Sensitivity of the different competing strategies when the land cover class of interest (the positive class) is Forest.
Table 11. Performances in terms of the Sensitivity of the different competing strategies when the land cover class of interest (the positive class) is Forest.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
20100± 0.099.9 ± 0.199.7 ± 0.281.6 ± 2.587.2 ± 3.089.7 ± 2.9
40100± 0.099.8 ± 0.199.7 ± 0.380.1 ± 5.187.1 ± 5.591.9 ± 3.3
6099.9± 0.099.8 ± 0.299.7 ± 0.282.0 ± 3.787.6 ± 3.491.2 ± 3.6
8099.9± 0.199.8 ± 0.199.7 ± 0.282.2 ± 4.385.7 ± 4.389.5 ± 5.0
10099.8± 0.199.7 ± 0.199.7 ± 0.179.9 ± 2.685.6 ± 4.390.2 ± 2.6
Table 12. Performances in terms of the Specificity of the different competing strategies when the land cover class of interest (the positive class) is Forest.
Table 12. Performances in terms of the Specificity of the different competing strategies when the land cover class of interest (the positive class) is Forest.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
200.8 ± 0.43.7 ± 1.336.0 ± 7.897.3± 1.0194.1 ± 1.594.0 ± 1.6
403.4 ± 2.410.1 ± 3.737.2 ± 7.297.6± 1.2695.2 ± 1.893.6 ± 2.5
607.5 ± 3.218.7 ± 5.241.1 ± 6.997.9± 1.095.5 ± 1.794.3 ± 2.9
8012.5 ± 4.226.8 ± 6.342.0 ± 6.597.9± 1.096.2 ± 1.595.2 ± 2.2
10019.6 ± 4.737.3 ± 6.843.6 ± 6.598.3± 0.796.3 ± 1.494.9 ± 2.0
Table 13. Performances in terms of the Kappa measure of the different competing strategies when the land cover class of interest (the positive class) is Forest.
Table 13. Performances in terms of the Kappa measure of the different competing strategies when the land cover class of interest (the positive class) is Forest.
# of Obj SamplesRFPULENSEMBLEPULOCSVMPUL-SITSnoRegPUL-SITSrecoPUL-SITS
200.01 ± 0.010.05 ± 0.020.46 ± 0.090.46 ± 0.050.55 ± 0.070.61± 0.08
400.05 ± 0.040.14 ± 0.050.47 ± 0.070.44 ± 0.080.57 ± 0.130.67± 0.09
600.11 ± 0.050.26 ± 0.070.51 ± 0.070.47 ± 0.080.57 ± 0.080.66± 0.1
800.18 ± 0.060.36 ± 0.080.52 ± 0.070.48 ± 0.080.54 ± 0.080.62± 0.1
1000.27 ± 0.060.47 ± 0.080.54 ± 0.070.44 ± 0.050.53 ± 0.070.63± 0.05
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Desloires, J.; Ienco, D.; Botrel, A.; Ranc, N. Positive Unlabelled Learning for Satellite Images’Time Series Analysis: An Application to Cereal and Forest Mapping. Remote Sens. 2022, 14, 140. https://doi.org/10.3390/rs14010140

AMA Style

Desloires J, Ienco D, Botrel A, Ranc N. Positive Unlabelled Learning for Satellite Images’Time Series Analysis: An Application to Cereal and Forest Mapping. Remote Sensing. 2022; 14(1):140. https://doi.org/10.3390/rs14010140

Chicago/Turabian Style

Desloires, Johann, Dino Ienco, Antoine Botrel, and Nicolas Ranc. 2022. "Positive Unlabelled Learning for Satellite Images’Time Series Analysis: An Application to Cereal and Forest Mapping" Remote Sensing 14, no. 1: 140. https://doi.org/10.3390/rs14010140

APA Style

Desloires, J., Ienco, D., Botrel, A., & Ranc, N. (2022). Positive Unlabelled Learning for Satellite Images’Time Series Analysis: An Application to Cereal and Forest Mapping. Remote Sensing, 14(1), 140. https://doi.org/10.3390/rs14010140

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop