Next Article in Journal
Satellite Jitter Estimation and Validation Using Parallax Images
Previous Article in Journal
High Sensitivity Detection of CdSe/ZnS Quantum Dot-Labeled DNA Based on N-type Porous Silicon Microcavities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Evaluation of Machine Learning Methods for Leaf Area Index Retrieval from Time-Series MODIS Reflectance Data

State Key Laboratory of Remote Sensing Science, Beijing Normal University, Beijing 100875, China
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(1), 81; https://doi.org/10.3390/s17010081
Submission received: 1 September 2016 / Accepted: 28 December 2016 / Published: 1 January 2017
(This article belongs to the Section Remote Sensors)

Abstract

:
Leaf area index (LAI) is an important biophysical parameter and the retrieval of LAI from remote sensing data is the only feasible method for generating LAI products at regional and global scales. However, most LAI retrieval methods use satellite observations at a specific time to retrieve LAI. Because of the impacts of clouds and aerosols, the LAI products generated by these methods are spatially incomplete and temporally discontinuous, and thus they cannot meet the needs of practical applications. To generate high-quality LAI products, four machine learning algorithms, including back-propagation neutral network (BPNN), radial basis function networks (RBFNs), general regression neutral networks (GRNNs), and multi-output support vector regression (MSVR) are proposed to retrieve LAI from time-series Moderate Resolution Imaging Spectroradiometer (MODIS) reflectance data in this study and performance of these machine learning algorithms is evaluated. The results demonstrated that GRNNs, RBFNs, and MSVR exhibited low sensitivity to training sample size, whereas BPNN had high sensitivity. The four algorithms performed slightly better with red, near infrared (NIR), and short wave infrared (SWIR) bands than red and NIR bands, and the results were significantly better than those obtained using single band reflectance data (red or NIR). Regardless of band composition, GRNNs performed better than the other three methods. Among the four algorithms, BPNN required the least training time, whereas MSVR needed the most for any sample size.

1. Introduction

Leaf area index (LAI) is defined as one half the total green leaf area per unit horizontal ground surface area [1]. It is an important biophysical parameter, and is widely used in crop growth monitoring, yield estimation, and global change studies. Currently, many methods have been developed to retrieve LAI from remote sensing data. These inversion methods can be generally divided into two categories: statistical methods and physical methods [2]. The statistical methods employ the statistical relationships between vegetation indices (VIs) and LAI to estimate LAI [3,4,5]. The physical methods are based on the inversion of radiative transfer models that clearly describe the connection between biophysical variables and canopy reflectance [6,7,8]. Because the physical methods can be adjusted for a wide range of situations [9], radiative transfer models are increasingly used in the inverse mode to estimate LAI from remotely sensed data. However, the commonly used inversion techniques based on iterative minimization of a cost function require excessively computationally demanding in order to achieve high retrieval accuracy [9]. For operational applications, machine learning methods based on precomputed radiative transfer simulations [10] or existing LAI products [11,12] are popular inversion techniques.
Among the machine learning methods, artificial neural networks (ANNs) are the most widely used techniques to retrieve LAI from remote sensing data [13,14]. Danson et al. [15] used a canopy radiative transfer model to simulate reflectance data for a wide range of biophysical conditions and these data were used to train back-propagation neural networks (BPNN) to retrieve LAI from measured reflectance data. The results showed that the estimation of sugar beet LAI using the trained neural network is more reliable than the LAI values retrieved using statistical relationships between VIs and LAI. Fang et al. [16] used ANN methods to retrieve LAI from the Landsat-7 Enhanced Thematic Mapper Plus surface reflectance and top of atmosphere radiance, respectively, and the results showed that the ANN methods can estimate LAI accurately. Bacour et al. [17] designed a multilayer perceptron to estimate LAI, the fraction of absorbed photosynthetically active radiation (FAPAR), the fraction of vegetation cover (FCover), and the canopy chlorophyll content (LAI × Cab) from Medium Resolution Imaging Spectrometer (MERIS) surface reflectance data. Baret et al. [10] developed an algorithm to generate global LAI, FAPAR, and FCover products from VEGETATION observations based on training BPNN with SAIL + PROSPECT radiative transfer model simulations for each biophysical variable. The GEOV1 LAI retrieval algorithm relies on BPNNs trained with the “best estimates” of LAI obtained by fusing and scaling the Moderate Resolution Imaging Spectroradiometer (MODIS) and CYCLOPES LAI products and the VEGETATION nadir surface reflectance values over the Benchmark Land Multisite Analysis and Intercomparison of Products (BELMANIP) sites [11]. Zhu et al. [18] used feedforward neural networks trained with the best-quality MODIS LAI data to generate LAI products from the third-generation Global Inventory Modeling and Mapping Studies (GIMMS3g) normalized difference vegetation index (NDVI) data.
Support vector machines (SVMs) are another type of popular machine learning method and are well known for their good performance in classification and function approximation. In recent years, SVMs are increasingly being used for the estimation of biophysical variables from satellite observations. Durbha et al. [19] proposed a method for LAI retrieval using support vector regression (SVR) from multiangle imaging spectroradiometer data. The method used a single-output SVR and could retrieve only one parameter at a time. Multi-output support vector regression (MSVR) [20,21] expanded single-output SVR into multiple-output and used ε-insensitive cost function to maintain simple and sparse solution. Tuia et al. [22] used MSVR for estimation of LAI × Cab, LAI and FCover from hyperspectral compact high-resolution imaging spectrometer images.
Indeed, many machine learning methods have been used to estimate LAI from remotely sensed data. On the one hand, most machine learning methods use satellite observations at the specific time to retrieve LAI, a result of which is that the LAI products generated by these methods are spatially incomplete and temporally discontinuous [23]. To generate a long-time series of high-quality global LAI product, Xiao et al. [12] developed an operational algorithm for estimating LAI from time-series MODIS surface reflectance data using general regression neural networks (GRNNs) with a multi-input-multi-output architecture. This method achieved excellent performance and was used to produce the Global Land Surface Satellite (GLASS) LAI product from time-series MODIS and AVHRR surface reflectance data [24]. On the other hand, each machine learning method has its own pros and cons, and there is no performance comparison between the machine learning methods for LAI retrieval from time-series remote sensing data.
In this study, machine learning algorithms were developed to retrieve LAI from time-series MODIS surface reflectance data. The reprocessed MODIS reflectance data for a one-year period were entered into the trained machine learning algorithms to estimate LAI profiles for the one-year period. The performance of the machine learning algorithms was evaluated by retrieving LAI from time-series MODIS reflectance data with different band combinations. The sensitivity to the training sample size and the computational efficiency of the machine learning algorithms are also evaluated. This study may provide guidance for the future work of LAI inversion.
The organization of this paper is as follows. Section 2 introduces the method to retrieve LAI from time-series MODIS reflectance data. This includes a framework for LAI inversion from time-series data using machine learning algorithms, the method to generate the training and testing samples, the machine learning algorithms used in this study and their training. The statistical parameters used to evaluate the performance of the machine learning algorithms are also briefly described in this section. Section 3 presents the performance comparison of the machine learning algorithms. The conclusions are given in Section 4.

2. Methodology

2.1. A Framework for LAI Inversion from Time-Series Data Using Machine Learning Algorithms

In this study, machine learning algorithms are used to retrieve LAI from time-series MODIS surface reflectance data (MOD09A1). All the machine learning algorithms employ multi-input-multi-output architectures. In contrast to existing methods that only use remote sensing data acquired at a specific time to retrieve LAI, the reprocessed MODIS surface reflectance data for a one-year period were inputted into the machine learning algorithms to estimate one-year LAI profiles (Figure 1).
The MOD09A1 product has an eight-day temporal resolution. To retrieve LAI profiles from time-series MODIS surface reflectance data using the machine learning algorithms, the input vector X of the machine learning algorithms comprised the reprocessed MODIS time-series reflectance values (for a one-year period); that is, X = ( R 1 1 , R 2 1 , ...... , R 46 1 , R 1 2 , R 2 2 , ...... , R 46 2 , ...... , R 1 m , R 2 m , ...... , R 46 m ) T contains 46 × m components, where m is the number of bands. The output vector Y = ( L A I 1 , L A I 2 , ...... , L A I 46 ) T is the corresponding LAI time series for the year, which contains 46 components.

2.2. Generating the Training and Testing Samples

To produce GLASS LAI product from time-series MODIS/AVHRR reflectance data using GRNNs, Xiao et al. [12,24] generated a training dataset from MODIS and CYCLOPES LAI products and MODIS surface reflectance product of the BELMANIP sites during the period from 2001–2003. The effective CYCLOPES LAI was first converted to the true LAI, which was then combined with the MODIS LAI. The MODIS surface reflectance was reprocessed to remove remaining effects of cloud contamination and other factors. GRNNs were then trained using the fused time-series LAI values and reprocessed MODIS surface reflectance data over the BELMANIP sites.
The training dataset is globally representative of surface types and conditions. In this study, the dataset was used to training and testing the machine learning algorithms and evaluate their performances. Twenty groups of training samples with different sample sizes and the corresponding testing samples were randomly generated from the dataset to evaluate the sensitivity of the machine learning algorithms to the size of the training samples. Table 1 shows the numbers of training and testing samples for each group (denoted as G1 to G20). To avoid attributes in greater numeric ranges dominating those in smaller numeric ranges and numerical difficulties during the calculation, the training and testing samples were scaled before applying machine learning algorithms [25,26]. In this study, the following formula was used to normalize the values of the training and testing samples [12]:
X n o r m = 2.0 × ( X X min ) / ( X max X min ) 1 ,
where X max and X min denote the maximum and minimum values for variable X , respectively, and X n o r m is the normalized value corresponding to the variable X .

2.3. Machine Learning Algorithms and Their Training

Four machine learning algorithms, including BPNN, radial basis function network (RBFNs), GRNNs, and MSVR, were used to retrieve LAI from time-series MODIS surface reflectance data and their performance was evaluated. The architecture of these machine learning algorithms and their training processes are as follows.

2.3.1. BPNN

BPNN is named for the use of error back-propagation algorithm [27]. It is a multi-layer feedforward neutral network that is currently one of the most widely used neural network algorithms. BPNN has a simple structure and it is used mainly for pattern recognition and classification, function approximation, data compression and prediction, and other applications. This method overcomes the limitations of the perceptron and linear neural network, so it can achieve any linear or nonlinear function mapping. But in the actual operation process, it often requires repeated trials to determine the training parameters and continued training to obtain satisfactory results [27].
Some studies have shown that one-hidden-layer network is sufficient to model any complex system with any desired accuracy [28,29,30,31,32], so the BPNN used in this study had three layers (Figure 2a): input layer, one hidden layer, and output layer. Many investigations have analyzed the determination of the number of hidden nodes and several empirical equations have been proposed [33,34,35]. In this study, the number of neurons in the hidden layer was determined by the following equation [33]:
n h = n i + n o + a ,
where n h is the number of hidden layer nodes, n i and n o are the numbers of input and output samples, respectively, and a is a constant between 1 and 10. The input layer contained 46 × m nodes and the output layer had 46 nodes, which corresponded to the one-year LAI values. Changes of learning parameters mainly affect the speed of convergence of the learning procedures, but they do not affect the BPNN structure and its overall performance [36], so the learning rate was set to 0.01. Back-propagation has disadvantages in terms of its slow convergence speed and a high likelihood of being trapped by local minima, but some improved training algorithms are available for BPNN. The Levenberg-Marquardt algorithm [37] is used widely in these training algorithms, although it occupies a large memory when applied to big training samples. The scaled conjugate gradient algorithm [38] is a normalized conjugate gradient method with low memory requirements and it is the only method that does not need linear search. Thus, the scaled conjugate gradient algorithm was used as the training algorithm in this study. In addition, the convergence criteria used for training the BPNN model comprised a mean squared error (MSE) less than or equal to 0.001 or a maximum of 5000 iterations.
The initial weights of the BPNN were set with randomly selected values. The training samples were then passed through the network and the result was outputted. Next, the network output was compared with the desired output and the error was calculated. This process is known as feed-forward. The error was back-propagated through the network and the weights of the connections were adjusted in a back-propagation process. The process of feeding forward the training samples and back-propagating the error continued until the error was minimized in the network or it reached an acceptable magnitude [39,40].

2.3.2. RBFNs

RBFNs, developed by Broomhead and Lowe [41], are also feedforward neutral networks, and they have been used widely for function approximation and pattern recognition in recent years. Figure 2b shows the architecture of the RBFN used in this study, which comprises three layers: input layer, output layer, and one hidden layer. The output of a RBFN can be described by the following equation [42]:
Y = w 0 + i = 1 n h w i f ( X c i ) ,
where f ( ) represents a radial-basis function (RBF), w i are the output layer weights, w 0 is an offset, X are the inputs to the network, c i are the centers of the basis functions, n h is the number of nodes in the hidden layer, and is the Euclidean norm. RBFN has a similar architecture to the BPNN, but it only has one hidden layer. The input into RBFN is nonlinear, whereas the output is linear. The weights between the input and hidden layers are 1, and only the weights between the hidden and output layers can be tuned.
The basis function used in this study is the following Gaussian basis function:
f ( X ) = exp ( X c ) 2 r 2 ,
where r is the basis function width, which is the only free parameter. More details of basis functions can be found in Haykin (1999) [43].
There are two stages during the training of RBFNs. First, the basis functions are determined by unsupervised analysis, and the weights connecting the nodes in the input and hidden layers are then defined effectively. For a given input, the Euclidean norm computes the distance from the center of its RBF. If the input vector is closer to the function center, then the RBF will yield a larger number. Second, the weights connecting the hidden and output layers are derived by a linear supervised method. For a given input vector, one or more nodes will provide an output because local data are employed for modeling the RBFNs. The outputs are then linearly combined to form the network output [40,42].

2.3.3. GRNNs

GRNNs, developed by Specht [44], are a generalization of RBFNs and probabilistic neural networks. At present, GRNNs are used widely for system identification, adaptive control fields, pattern recognition, and time series prediction. GRNNs comprise four layers, including the input layer, pattern layer, summation layer, and output layer. Figure 2c shows a GRNN with a multi-input-multi-output architecture. The kernel function of the GRNN used in this study is Gaussian and its fundamental formulation is as follows:
Y ( X ) = i = 1 N Y i exp ( D i 2 2 σ 2 ) i = 1 N exp ( D i 2 2 σ 2 ) ,
where D i 2 = ( X X i ) T ( X X i ) is the squared Euclidean distance between the input vector X and the ith training input vector X i , Y i is the output vector corresponding to the vector X i , Y ( X ) is the estimate corresponding to the vector X , N is the number of samples, and σ is a smoothing parameter for controlling the size of the receptive region. Thus, the smoothing parameter is the only free parameter.
In this study, the shuffled complex evolution method developed at the University of Arizona (SCE-UA) was used to obtain the optimal smoothing parameter when the following cost function of the smoothing parameter was minimized [12]:
f ( σ ) = 1 N i = 1 N ( Y ^ i ( X i ) Y i ) 2 ,
where Y ^ i ( X i ) is the estimate corresponding to X i using the GRNN trained over all of the training samples except the ith sample.

2.3.4. MSVR

The support vector machine (SVM) was proposed by Vapnik [45] and has become a very popular machine learning method. It is a learning systems based on Vapnik statistical learning theory and the structural risk minimization principle, avoiding local minima and over learning problems. Support vector regression (SVR) is the application of support vector machine in function regression area. The traditional SVR algorithm only applies to single-output systems. MSVR is a multivariate regression estimation algorithm based on the SVR [20,21,22]. It expands single-output SVR into multiple-output and uses-insensitive cost function to maintain simple and sparse solution.
In the multidimensional regression estimation problems, the expected output y is a vector with Q variables to be predicted, i.e., y R Q . Thus, we have to find a regressor w j and b j ( j = 1 , , Q ) for every output in the multidimensional regression estimation problem. We can directly generalize the 1-D SVR to solve the multidimensional case by minimizing the following formulation:
L P ( W , b ) = 1 2 j = 1 Q w j 2 + C i = 1 l L ( u i ) ,
where u i = e i = e i T e i , e i T = y i T ϕ T ( x i ) W b T , W = [ w 1 , ... , w Q ] , b = [ b 1 , ... , b Q ] T , and ϕ ( ) is a transformation to a higher-dimensional space. L ( u ) is the Vapnik ε-insensitive loss-function and can be expressed as:
L ( u ) = { 0 , u < ε u 2 2 u ε + ε 2 , u ε .
It is usual to work with the feature space kernel (inner product of the transformed vectors k ( x i , x j ) = ϕ T ( x i ) ϕ ( x j ) ), instead of the whole nonlinear mapping. In this study, the RBF kernel is used because it can handle the case when the relationship between class labels and attributes is nonlinear with mapping nonlinearly samples into a higher dimensional space. There are three parameters for MSVR with a RBF kernel: the penalty coefficient C , loss function ε and kernel width parameter σ . The SCE-UA algorithm combined with exponential transformation was used to search the optimal values of these parameters [26].

2.4. Evaluation Statistics

After the machine learning algorithms were trained using the training data, the testing data were used to evaluate the performance of these algorithms. The reprocessed MODIS reflectance data for a one-year period were entered into the trained machine learning algorithms and the outputs of the trained machine learning algorithms represented the one-year LAI profile for each pixel.
The MODIS surface reflectance data provide seven bands with the corresponding center wavelengths of 648 nm, 858 nm, 470 nm, 555 nm, 1240 nm, 1640 nm and 2130 nm. In this study, the MODIS reflectance in the red band (620–670 nm), near infrared band (841–876 nm) and shortwave infrared band (1628–1652 nm) were used to evaluate the performance of the four machine learning algorithms through different band combinations. The band combinations comprised one band (only red band or only NIR band), two bands (red and NIR bands), and all three bands. Meanwhile, the performance of the four machine learning methods to retrieve LAI values from NDVI was also evaluated in this study. The LAI values derived from the MODIS surface reflectance data with different band combinations using the machine learning algorithms were evaluated with the collocated testing LAI values using the following statistical parameters: the coefficient of determination (R2), root mean squared error (RMSE), and mean absolute error (MAE):
RMSE = 1 N i = 1 N ( L A I i e L A I i t ) 2 ,
MAE = 1 N i = 1 N | L A I i e L A I i t | ,
where L A I t e and L A I i t represent the estimated and testing LAI values, respectively, and N is the total number of data points. Indicators similar to these are often used in performance evaluation studies [11,12]. In addition, the computational efficiency of the four machine learning algorithms was also evaluated in this study.

3. Results and Discussion

3.1. Sensitivity of the Machine Learning Algorithms to Sample Size

Figure 3 shows the changes of R2, RMSE, and MAE as the sample size varied. In the left panel, the LAI values were derived from the MODIS reflectance data in the red and NIR bands, whereas in the right panel, the LAI values were derived from the MODIS reflectance data in the red, NIR, and SWIR bands. With the increase of the sample size, R2 of the RBFNs, GRNNs and MSVR increased gradually, which indicates that the performance of the three machine learning algorithms has a slight improvement with the sample size increased. For BPNN, R2 increased rapidly with increasing sample size up to 2000. When the sample size was larger than 3000, R2 of BPNN was almost unchanged and slightly smaller than that of GRNNs. Among the four machine learning algorithms, MSVR has the smallest R2 for any sample size. As the sample size increased, RMSE and MAE of these four methods decreased. BPNN and GRNNs achieved significantly better performance than RBFNs and MSVR in terms of the RMSE and MAE. For BPNN, RMSE and MAE decreased rapidly with increasing sample size up to 2000. RBFNs just had smaller RMSE and MAE values than MSVR, whose RMSE and MAE values were the largest among the four machine learning algorithms.
The changes of R2, RMSE, and MAE for the LAI values derived from the MODIS reflectance data in the red and NIR bands were similar to those in the red, NIR, and SWIR bands. As the sample size increased, the metrics changed with slight fluctuations and become stable gradually. However, there are also some differences between the results shown in the left and right panels. First, R2 of all these methods using MODIS reflectance data in the red and NIR bands were smaller than those in the red, NIR, and SWIR bands, whereas RMSE and MAE of all these methods using MODIS reflectance data in the red and NIR bands were larger than those in the red, NIR, and SWIR bands. Second, BPNN and GRNNs had more similar performance in red and NIR bands than those in red, NIR, and SWIR bands.
Figure 3 shows that all of the algorithms obtained rather good performance (R2 > 0.61, RMSE < 1.19, and MAE < 0.95). The results show that GRNNs had the least sensitivity to the sample size and BPNN had the most sensitivity. RBFNs and MSVR were also less sensitive to the sample size, but they performed worse than the other methods.

3.2. Comparisons of the Retrieved LAI from MODIS Reflectance Data with Different Band Combinations

The above results showed that the retrieved LAI was more accurate when the sample size was larger. Thus, the training data with the maximum sample size was used to evaluate performance of the four machine learning methods to retrieve LAI values from MODIS reflectance data in different band combinations.
Figure 4 shows the scatter density plots with different band combinations and NDVI between the testing LAI values and the retrieved LAI values using the machine learning methods. The scatter density plots between the testing LAI values and the LAI values retrieved from the MODIS reflectance data in the NIR band using the four machine learning methods are shown in the first panel in Figure 4. GRNNs performed better than the other three methods, although all four algorithms did not obtain the desired results. The second panel in Figure 4 shows the scatter density plots between the testing LAI values and the LAI values retrieved from the MODIS reflectance data in the red band. It is clear that the four machine learning methods performed better compared with the LAI values retrieved from the MODIS reflectance data in the NIR band. Among the four machine learning methods, GRNNs delivered the best performance, with the highest R2 and lowest RMSE values, whereas MSVR performed the worst. The third panel in Figure 4 shows the scatter density plots for the four methods between the testing LAI values and the LAI values retrieved from the MODIS reflectance data in the red and NIR bands. These results show that all of the methods performed better compared with using only one band. The fourth panel in Figure 4 shows the scatter density plots for the four methods between the testing LAI values and the LAI values retrieved from NDVI values. Performance of BPNN and GRNNs was just slightly worse than that with red and NIR bands, while the performance of RBFNs was obviously worse than that with red and NIR bands. The performance of MSVR was different with the other four methods, which performed better than that with red and NIR bands. The final panel in Figure 4 shows the scatter density plots for the four methods between the testing LAI values and the LAI values retrieved from the MODIS reflectance data in the red, NIR, and SWIR bands. The LAI values retrieved from the three bands reflectance data were slightly better compared with the LAI values retrieved from the red and NIR bands reflectance data. The LAI values retrieved from reflectance data in the red, NIR, and SWIR bands were almost the same using BPNN and GRNNs. Among the four machine learning methods, MSVR had the worst performance, with the lowest R2 and largest RMSE values.
Figure 4 demonstrates that all the methods performed better, with higher R2 and lower RMSE and MAE values, as the number of bands increased. Using only one band reflectance data to train the methods for LAI retrieval, the LAI values retrieved from surface reflectance in the red band were better than those obtained from the NIR band reflectance. All the methods performed well when MODIS surface reflectance data in the red and NIR bands were used together to train the methods. All the methods had a worse performance with the NDVI values than that with red and NIR bands, except MSVR. When the reflectance data in the red, NIR and SWIR bands were used to train the methods, the retrieved LAI values using the four machine learning methods improved further. Regardless of the band combinations employed, GRNNs outperformed the other three methods. BPNN obtained almost the same performance, whereas the performance of RBFNs and MSVR was worse than that of the other two methods.
In addition, we also note that RBFNs are affected by the problem of negative values during LAI inversion. When the testing data are at the edge of the training data range, the estimates of the RBFNs may appear the mutation. Furthermore, RBFNs based on the empirical risk minimization principle have a simple structure with only one hidden layer, which may lead to overfitting. These reasons may explain the negative values obtained by RBFNs.

3.3. Comparisons of Computational Efficiency

Computational efficiency is an important evaluation metric for operational algorithms to generate global LAI products. Thus, the time required by the four methods for the training and testing processes was computed to further assess the performance of these methods. The MODIS surface reflectance data in the red, NIR, and SWIR bands were used to train the machine learning methods. The computer used to execute these programs had an Intel(R) Core(TM) i7-3770K CPU @ 3.40 GHz processor and installed memory (RAM) of 8.00 GB.
Figure 5 shows histograms of the time required under different sample sizes by these methods during the training process. BPNNs required the least time for the training process regardless of the sample size. RBFNs had a relatively fast convergence speed when the training sample size was small. RBFNs only has one hidden layer, so it needed more nodes compared with BPNN when the number of training samples was large, thereby highlighting its disadvantages in terms of efficiency. GRNNs and MSVR consumed much more time than the other two methods. MSVR consumed slightly more time than GRNNs for any sample size.
Figure 6 shows the time required by the four machine learning methods to retrieve LAI values from MODIS surface reflectance data with red, NIR, and SWIR bands over a MODIS tile in 1200 rows by 1200 columns. These results demonstrate BPNN, RBFNs, and MSVR required nearly the same time for LAI retrieval over a MODIS tile. GRNNs required the most time compared with the other three methods.

4. Conclusions

The currently LAI inversion methods use remote sensing data at the specific time for LAI retrieval, which yields spatially incomplete and temporally discontinuous LAI products. In this study, four machine learning algorithms were proposed to retrieve LAI from time-series MODIS reflectance data, and performance of these machine learning algorithms was evaluated. The results showed that RBFNs, GRNNs and MSVR were less sensitive to the training sample size, whereas BPNNs exhibited high sensitivity. The BPNNs produced poor LAI estimates with small training samples. Comparison of the LAI values retrieved from MODIS surface reflectance data in different band combinations using the four machine learning methods demonstrated that all of the methods performed better as the number of bands in the samples increased. Among the four machine learning methods, GRNNs achieved the best performance irrespective of the training sample size and it also retrieved the best results regardless of the band combinations employed. When the training samples were sufficiently large, all four machine learning methods delivered improvement in LAI retrieval and BPNN performed almost the same as GRNNs. Comparison of the training time required by the four methods demonstrated that BPNN required the least time regardless of the sample size. RBFNs had a relatively fast convergence speed when the training sample size was small. GRNNs and MSVR consumed much more training time than the other two methods for any sample size. BPNN and RBFNs required nearly the same time and slightly more than MSVR for LAI retrieval over a MODIS tile. GRNNs required the most time compared with the other three methods. Considering the performance of the four machine learning methods, BPNN and GRNNs can be recommended as operational methods for LAI retrieval from time series satellite observations.

Acknowledgments

This work was supported by the Chinese 973 Program under grant No. 2013CB733403, and the National Natural Science Foundation of China under grant No. 41171264, 41571409.

Author Contributions

Zhiqiang Xiao proposed the idea of the study. Tongtong Wang implemented the comparison work of the study, performed the data analysis and wrote the paper. Zhigang Liu helped to solve the problems of the study. Zhiqiang Xiao and Zhigang Liu reviewed and edited the manuscript. All authors read and approved the manuscript.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Chen, J.M.; Black, T. Defining leaf area index for non-flat leaves. Plant Cell Environ. 1992, 15, 421–429. [Google Scholar] [CrossRef]
  2. Liang, S.; Li, X.; Wang, J. Advanced Remote Sensing: Terrestrial Information Extraction and Applications; Academic Press: London, UK, 2012. [Google Scholar]
  3. Turner, D.P.; Cohen, W.B.; Kennedy, R.E.; Fassnacht, K.S.; Briggs, J.M. Relationships between leaf area index and Landsat TM spectral vegetation indices across three temperate zone sites. Remote Sens. Environ. 1999, 70, 52–68. [Google Scholar] [CrossRef]
  4. Wang, Q.; Adiku, S.; Tenhunen, J.; Granier, A. On the relationship of NDVI with leaf area index in a deciduous forest site. Remote Sens. Environ. 2005, 94, 244–255. [Google Scholar] [CrossRef]
  5. Potithep, S.; Nagai, S.; Nasahara, K.N.; Muraoka, H.; Suzuki, R. Two separate periods of the LAI-VIs relationships using in situ measurements in a deciduous broadleaf forest. Agric. For. Meteorol. 2013, 169, 148–155. [Google Scholar] [CrossRef]
  6. Myneni, R.B.; Ramakrishna, R.; Nemani, R.; Running, S.W. Estimation of global leaf area index and absorbed PAR using radiative transfer models. IEEE Trans. Geosci. Remote Sens. 1997, 35, 1380–1393. [Google Scholar] [CrossRef]
  7. Jacquemoud, S.; Bacour, C.; Poilve, H.; Frangi, J.-P. Comparison of four radiative transfer models to simulate plant canopies reflectance: Direct and inverse mode. Remote Sens. Environ. 2000, 74, 471–481. [Google Scholar] [CrossRef]
  8. Darvishzadeh, R.; Skidmore, A.; Schlerf, M.; Atzberger, C. Inversion of a radiative transfer model for estimating vegetation LAI and chlorophyll in a heterogeneous grassland. Remote Sens. Environ. 2008, 112, 2592–2604. [Google Scholar] [CrossRef]
  9. Kimes, D.; Knyazikhin, Y.; Privette, J.; Abuelgasim, A.; Gao, F. Inversion methods for physically-based models. Remote Sens. Rev. 2000, 18, 381–439. [Google Scholar] [CrossRef]
  10. Baret, F.; Hagolle, O.; Geiger, B.; Bicheron, P.; Miras, B.; Huc, M.; Berthelot, B.; Niño, F.; Weiss, M.; Samain, O. LAI, fAPAR and fCover CYCLOPES global products derived from VEGETATION: Part 1: Principles of the algorithm. Remote Sens. Environ. 2007, 110, 275–286. [Google Scholar] [CrossRef] [Green Version]
  11. Baret, F.; Weiss, M.; Lacaze, R.; Camacho, F.; Makhmara, H.; Pacholcyzk, P.; Smets, B. GEOV1: LAI and FAPAR essential climate variables and FCOVER global time series capitalizing over existing products. Part 1: Principles of development and production. Remote Sens. Environ. 2013, 137, 299–309. [Google Scholar] [CrossRef]
  12. Xiao, Z.; Liang, S.; Wang, J.; Chen, P.; Yin, X.; Zhang, L.; Song, J. Use of general regression neural networks for generating the GLASS leaf area index product from time-series MODIS surface reflectance. IEEE Trans. Geosci. Remote Sens. 2014, 52, 209–223. [Google Scholar] [CrossRef]
  13. Liang, S. Quantitative Remote Sensing of Land Surfaces; Wiley: Weinheim, Germany, 2005. [Google Scholar]
  14. Yang, G.; Zhao, C.; Liu, Q.; Huang, W.; Wang, J. Inversion of a radiative transfer model for estimating forest LAI from multisource and multiangular optical remote sensing data. IEEE Trans. Geosci. Remote Sens. 2011, 49, 988–1000. [Google Scholar] [CrossRef]
  15. Danson, F.; Rowland, C.; Baret, F. Training a neural network with a canopy reflectance model to estimate crop leaf area index. Int. J. Remote Sens. 2003, 24, 4891–4905. [Google Scholar] [CrossRef]
  16. Fang, H.; Liang, S. Retrieving leaf area index with a neural network method: Simulation and validation. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2052–2062. [Google Scholar] [CrossRef]
  17. Bacour, C.; Baret, F.; Béal, D.; Weiss, M.; Pavageau, K. Neural network estimation of LAI, fAPAR, fCover and LAI × Cab, from top of canopy MERIS reflectance data: Principles and validation. Remote Sens. Environ. 2006, 105, 313–325. [Google Scholar] [CrossRef]
  18. Zhu, Z.; Bi, J.; Pan, Y.; Ganguly, S.; Anav, A.; Xu, L.; Samanta, A.; Piao, S.; Nemani, R.R.; Myneni, R.B. Global data sets of vegetation leaf area index (LAI)3g and Fraction of Photosynthetically Active Radiation (FPAR)3g derived from Global Inventory Modeling and Mapping Studies (GIMMS) Normalized Difference Vegetation Index (NDVI3g) for the period 1981 to 2011. Remote Sens. 2013, 5, 927–948. [Google Scholar]
  19. Durbha, S.S.; King, R.L.; Younan, N.H. Support vector machines regression for retrieval of leaf area index from multiangle imaging spectroradiometer. Remote Sens. Environ. 2007, 107, 348–361. [Google Scholar] [CrossRef]
  20. Pérez-Cruz, F.; Camps-Valls, G.; Soria-Olivas, E.; Pérez-Ruixo, J.J.; Figueiras-Vidal, A.R.; Artés-Rodríguez, A. Multi-dimensional function approximation and regression estimation. In Artificial Neural Networks—ICANN 2002; Springer: Heidelberg, Germany, 2002; pp. 757–762. [Google Scholar]
  21. Sánchez-Fernández, M.; de-Prado-Cumplido, M.; Arenas-García, J.; Pérez-Cruz, F. SVM multiregression for nonlinear channel estimation in multiple-input multiple-output systems. IEEE Trans. Signal Process. 2004, 52, 2298–2307. [Google Scholar] [CrossRef]
  22. Tuia, D.; Verrelst, J.; Alonso, L.; Pérez-Cruz, F.; Camps-Valls, G. Multioutput support vector regression for remote sensing biophysical parameter estimation. IEEE Geosci. Remote Sens. Lett. 2011, 8, 804–808. [Google Scholar] [CrossRef]
  23. Fang, H.; Liang, S.; Townshend, J.R.; Dickinson, R.E. Spatially and temporally continuous LAI data sets based on an integrated filtering method: Examples from North America. Remote Sens. Environ. 2008, 112, 75–93. [Google Scholar] [CrossRef]
  24. Xiao, Z.; Liang, S.; Wang, J.; Xiang, Y.; Zhao, X. Long Time Series Global Land Surface Satellite (GLASS) Leaf Area Index Product Derived from MODIS and AVHRR Data. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5301–5318. [Google Scholar] [CrossRef]
  25. Hsu, C.-W.; Chang, C.-C.; Lin, C.-J. A Practical Guide to Support Vector Classification. Available online: http://mbi.ee.ncku.edu.tw/mlb2007/res/MLB08_1.pdf (accessed on 29 December 2016).
  26. Lin, J.-Y.; Cheng, C.-T.; Chau, K.-W. Using support vector machines for long-term discharge prediction. Hydrol. Sci. J. 2006, 51, 599–612. [Google Scholar] [CrossRef]
  27. Rumelhart, D.E.; Hintont, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  28. Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  29. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  30. Dai, H.; MacBeth, C. Automatic picking of seismic arrivals in local earthquake data using an artificial neural network. Geophys. J. Int. 1995, 120, 758–774. [Google Scholar] [CrossRef]
  31. Zhang, G.; Patuwo, B.E.; Hu, M.Y. Forecasting with artificial neural networks: The state of the art. Int. J. Forecast. 1998, 14, 35–62. [Google Scholar] [CrossRef]
  32. Chou, S.-M.; Lee, T.-S.; Shao, Y.E.; Chen, I.-F. Mining the breast cancer pattern using artificial neural networks and multivariate adaptive regression splines. Exp. Syst. Appl. 2004, 27, 133–142. [Google Scholar] [CrossRef]
  33. Zhou, K.; Kang, Y. Neural Network Model and its MATLAB Simulation Program Design; Tsinghua University Press: Beijing, China, 2005. (In Chinese) [Google Scholar]
  34. Sonmez, H.; Gokceoglu, C.; Nefeslioglu, H.; Kayabasi, A. Estimation of rock modulus: For intact rocks with an artificial neural network and for rock masses with a new empirical equation. Int. J. Rock Mech. Min. Sci. 2006, 43, 224–235. [Google Scholar] [CrossRef]
  35. Karsoliya, S. Approximating number of hidden layer neurons in multiple hidden layer BPNN architecture. Int. J. Eng. Trends Technol. 2012, 3, 713–717. [Google Scholar]
  36. Dai, H.; MacBeth, C. Effects of learning parameters on learning procedure and performance of a BPNN. Neural Netw. 1997, 10, 1505–1521. [Google Scholar] [CrossRef]
  37. Moré, J.J. The Levenberg-Marquardt algorithm: Implementation and theory. In Numerical Analysis; Springer: Berlin, Germany, 1978; pp. 105–116. [Google Scholar]
  38. Møller, M.F. A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw. 1993, 6, 525–533. [Google Scholar] [CrossRef]
  39. Atkinson, P.M.; Tatnall, A. Introduction neural networks in remote sensing. Int. J. Remote Sens. 1997, 18, 699–709. [Google Scholar] [CrossRef]
  40. Foody, G.M. Thematic mapping from remotely sensed data with neural networks: MLP, RBF and PNN based approaches. J. Geogr. Syst. 2001, 3, 217–232. [Google Scholar] [CrossRef]
  41. Broomhead, D.S.; Lowe, D. Radial Basis Functions, Multi-Variable Functional Interpolation and Adaptive Networks; Controller HMSO: London, UK, 1988. [Google Scholar]
  42. Mas, J.; Flores, J. The application of artificial neural networks to the analysis of remotely sensed data. Int. J. Remote Sens. 2008, 29, 617–663. [Google Scholar] [CrossRef]
  43. Haykin, S. Neural Networks A Comprehensive Introduction; Springer: Prentice Hall, NJ, USA, 1999. [Google Scholar]
  44. Specht, D.F. A general regression neural network. IEEE Trans. Neural Netw. 1991, 2, 568–576. [Google Scholar] [CrossRef] [PubMed]
  45. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic description of the method to retrieve leaf area index (LAI) values from time-series Moderate Resolution Imaging Spectroradiometer (MODIS) surface reflectance data.
Figure 1. Schematic description of the method to retrieve leaf area index (LAI) values from time-series Moderate Resolution Imaging Spectroradiometer (MODIS) surface reflectance data.
Sensors 17 00081 g001
Figure 2. Multi-input-multi-output architectures of: (a) back-propagation neural networks (BPNN); (b) radial basis function networks (RBFNs); and (c) general regression neutral networks (GRNNs).
Figure 2. Multi-input-multi-output architectures of: (a) back-propagation neural networks (BPNN); (b) radial basis function networks (RBFNs); and (c) general regression neutral networks (GRNNs).
Sensors 17 00081 g002aSensors 17 00081 g002b
Figure 3. Sensitivity of the machine learning algorithms to sample size. Changes of R2 with training sample sizes for (a) the LAI values derived from reflectance data in the red and NIR bands and (b) the LAI values derived from reflectance data in the red, NIR, and SWIR bands; changes of RMSE with training sample sizes for (c) the LAI values derived from reflectance data in the red and NIR bands and (d) the LAI values derived from reflectance data in the red, NIR, and SWIR bands; changes of MAE with training sample sizes for (e) the LAI values derived from reflectance data in the red and NIR bands and (f) the LAI values derived from reflectance data in the red, NIR, and SWIR bands.
Figure 3. Sensitivity of the machine learning algorithms to sample size. Changes of R2 with training sample sizes for (a) the LAI values derived from reflectance data in the red and NIR bands and (b) the LAI values derived from reflectance data in the red, NIR, and SWIR bands; changes of RMSE with training sample sizes for (c) the LAI values derived from reflectance data in the red and NIR bands and (d) the LAI values derived from reflectance data in the red, NIR, and SWIR bands; changes of MAE with training sample sizes for (e) the LAI values derived from reflectance data in the red and NIR bands and (f) the LAI values derived from reflectance data in the red, NIR, and SWIR bands.
Sensors 17 00081 g003
Figure 4. Scatter density plots of four machine learning methods between the testing LAI values and the LAI values retrieved from MODIS surface reflectance data in different band combinations and NDVI values.
Figure 4. Scatter density plots of four machine learning methods between the testing LAI values and the LAI values retrieved from MODIS surface reflectance data in different band combinations and NDVI values.
Sensors 17 00081 g004
Figure 5. Time required for the training process by the machine learning methods with different sample sizes.
Figure 5. Time required for the training process by the machine learning methods with different sample sizes.
Sensors 17 00081 g005
Figure 6. Time required using the four machine learning methods to retrieve LAI values over a MODIS tile in 1200 rows by 1200 columns.
Figure 6. Time required using the four machine learning methods to retrieve LAI values over a MODIS tile in 1200 rows by 1200 columns.
Sensors 17 00081 g006
Table 1. Numbers of training and testing samples in each group.
Table 1. Numbers of training and testing samples in each group.
Sample NameG1G2G3G4G5G6G7G8G9G10
Training Datasets3326649961263153017972064227224812689
Testing Datasets332332332332332332332332331331
Sample NameG11G12G13G14G15G16G17G18G19G20
Training Datasets2896306232283394356036933826395940914187
Testing Datasets332332332332332332332332332332

Share and Cite

MDPI and ACS Style

Wang, T.; Xiao, Z.; Liu, Z. Performance Evaluation of Machine Learning Methods for Leaf Area Index Retrieval from Time-Series MODIS Reflectance Data. Sensors 2017, 17, 81. https://doi.org/10.3390/s17010081

AMA Style

Wang T, Xiao Z, Liu Z. Performance Evaluation of Machine Learning Methods for Leaf Area Index Retrieval from Time-Series MODIS Reflectance Data. Sensors. 2017; 17(1):81. https://doi.org/10.3390/s17010081

Chicago/Turabian Style

Wang, Tongtong, Zhiqiang Xiao, and Zhigang Liu. 2017. "Performance Evaluation of Machine Learning Methods for Leaf Area Index Retrieval from Time-Series MODIS Reflectance Data" Sensors 17, no. 1: 81. https://doi.org/10.3390/s17010081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop