Next Article in Journal
Effect of Pulsed Electromagnetic Field on Growth, Physiology and Postharvest Quality of Kale (Brassica oleracea), Wheat (Triticum durum) and Spinach (Spinacia oleracea) Microgreens
Next Article in Special Issue
Lychee Surface Defect Detection Based on Deep Convolutional Neural Networks with GAN-Based Data Augmentation
Previous Article in Journal
Agronomic Evaluation of the Results of Selection within Late-Maturing Dactylis glomerata Populations
Previous Article in Special Issue
A Densely Connected GRU Neural Network Based on Coattention Mechanism for Chinese Rice-Related Question Similarity Matching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations

1
School of Surveying and Geospatial Engineering, University of Tehran, Tehran 6791986363, Iran
2
Centre Eau Terre Environnement, Institut National de la Recherche Scientifique, Québec, QC G1K 9A9, Canada
3
School of Engineering Science, College of Engineering, University of Tehran, Tehran 6215845762, Iran
4
C-CORE, St. John’s, NL A1B 3X5, Canada
5
Department of Electrical and Computer Engineering, Memorial University of Newfoundland, St. John’s, NL A1C 3S7, Canada
6
Department of Geomatics Science, Laval University, Québec, QC G1V 0A6, Canada
*
Author to whom correspondence should be addressed.
Agronomy 2021, 11(7), 1363; https://doi.org/10.3390/agronomy11071363
Submission received: 21 May 2021 / Revised: 28 June 2021 / Accepted: 30 June 2021 / Published: 3 July 2021
(This article belongs to the Special Issue Applications of Deep Learning in Smart Agriculture)

Abstract

:
Remote sensing data are considered as one of the primary data sources for precise agriculture. Several studies have demonstrated the excellent capability of radar and optical imagery for crop mapping and biophysical parameter estimation. This paper aims at modeling the crop biophysical parameters, e.g., Leaf Area Index (LAI) and biomass, using a combination of radar and optical Earth observations. We extracted several radar features from polarimetric Synthetic Aperture Radar (SAR) data and Vegetation Indices (VIs) from optical images to model crops’ LAI and dry biomass. Then, the mutual correlations between these features and Random Forest feature importance were calculated. We considered two scenarios to estimate crop parameters. First, Machine Learning (ML) algorithms, e.g., Support Vector Regression (SVR), Random Forest (RF), Gradient Boosting (GB), and Extreme Gradient Boosting (XGB), were utilized to estimate two crop biophysical parameters. To this end, crops’ dry biomass and LAI were estimated using three input data; (1) SAR polarimetric features; (2) spectral VIs; (3) integrating both SAR and optical features. Second, a deep artificial neural network was created. These input data were fed to the mentioned algorithms and evaluated using the in-situ measurements. These observations of three cash crops, including soybean, corn, and canola, have been collected over Manitoba, Canada, during the Soil Moisture Active Validation Experimental 2012 (SMAPVEX-12) campaign. The results showed that GB and XGB have great potential in parameter estimation and remarkably improved accuracy. Our results also demonstrated a significant improvement in the dry biomass and LAI estimation compared to the previous studies. For LAI, the validation Root Mean Square Error (RMSE) was reported as 0.557 m2/m2 for canola using GB, and 0.298 m2/m2 for corn using GB, 0.233 m2/m2 for soybean using XGB. RMSE was reported for dry biomass as 26.29 g/m2 for canola utilizing SVR, 57.97 g/m2 for corn using RF, and 5.00 g/m2 for soybean using GB. The results revealed that the deep artificial neural network had a better potential to estimate crop parameters than the ML algorithms.

1. Introduction

Due to rapid population growth and climate changes, global food security and agriculture production risks have been increased [1]. Information about annual crop production is vital for global and local food security. In particular, measuring and monitoring crop biophysical parameters, including dry biomass, crop height, crop density, and LAI, during the crop growing season are essential for improving crop growth models and yield estimation [1,2]. Biomass and LAI are two widely used crop parameters in crop monitoring and growth models [3,4]. As the input data in crop models, crop biophysical parameters are estimated using direct and indirect methods. The direct method consists of a ground measuring of the plant’s parameters. These methods are usually destructive, costly, time-consuming, and complicated [5]. The information extracted from remote sensing data is non-destructive and significantly reduces time and cost. Remote sensing provides vital information on crop growth conditions over agricultural areas due to its extensive coverage and spatio-temporal resolution [6]. To this end, remote sensing imagery could be suitable for accurate crop monitoring.
Both SAR and optical data have been used to estimate crop parameters. Substantial studies have been carried out to investigate satellite optical data’s potential to estimate various crop parameters. VIs extracted from optic bands are widely used to estimate crop parameters and monitor crop conditions. However, when the crop canopy is dense, optical data tend to be saturated [7]. In addition, since optical data in cloudy conditions are not helpful, SAR sensors use microwave wavelengths that can penetrate clouds and haze [1,8,9,10,11,12,13].
SAR sensors can also provide data in day and night without considering sun illumination, with suitable temporal coverage and sufficient spatial resolution [12,13]. Furthermore, soil and surface parameters and the crop canopy state can easily affect radar backscattering [14]. Moreover, the SAR backscattering coefficient is affected by crop and soil parameters. However, these effects have changed by various sensor parameters (i.e., wavelength, incidence angles, and polarization), different target parameters (i.e., canopy structure, water content, soil moisture, and soil roughness), and crop type and growth stage [1,11,15,16,17,18]. Thus, the combination of optical and SAR data has a great ability in crop monitoring.
Considerable researches have been conducted to estimate various crop parameters using satellite Earth observations, including RADARSAT-2 [1,17,19,20,21], RapidEye [5,19,20,22,23], Sentinel-1 [7,17,24,25,26,27,28], Sentinel-2 [7,25,29,30,31,32], Landsat-5 Thematic Mapper (TM) [33,34], Landsat-7 Enhanced Thematic Mapper Plus (ETM+) [34,35,36], Landsat-8 Operational Land Imager (OLI) [7,17,31,32,35,36,37], Worldview-2/3 [17,27,28,38,39], and MODIS [40,41].
The crop parameters estimation methods in the literature can be generally categorized into three groups: (1) parametric models, (2) non-parametric models, and (3) physically-based models [42]. Parametric models assume a clear relationship between input and output variables. In contrast, there is no assumption for the statistical distribution of input data in non-parametric models. Finally, physically-based models use physical laws, and model variables are frequently obtained from Radiative Transfer Models (RTMs) [42].
The new generation of satellite sensors coupled with an increasing need for big data mining has increased the essential need to use artificial intelligence (AI) for Earth observation data analysis. Machine learning (ML), a subset of AI, is learning algorithms by using training data. ML algorithms rapidly process a large amount of data and give helpful insight into the information leads to astonishing output. Another advantage of ML algorithms is that any apriori assumption is needed about data distribution [43]. Non-parametric MLAs, without any assumption for the statistical distribution of input data, have successfully been applied to remote sensing data to retrieve crop biophysical parameters, yield estimation, and crop mapping. Reisi Gahrouei, et al. [22] used an artificial neural network (ANN) and SVR to estimate LAI and dry biomass of three crops, including soybean, corn, and canola high-resolution RapidEye data. Reisi-Gahrouei, et al. [44] also used MLR and ANN to estimate crop biomass using UAVSAR data. Sharifi and Hosseingholizadeh [45] have investigated the potential of MLR, relevance vector regression (RVR), and SVR to estimate cereal height and biomass. Zhu, et al. [46] utilized unmanned aerial vehicles (UAV) data to assess the ability of four MLAs, including MLR, RF, ANN, and SVR, to estimate the above-ground biomass (AGB). Luo, et al. [7] utilized MLR and SVR to estimate corn LAI and biomass. Deb, et al. [37] used parametric regression models and SVR to estimate agro-pastoral AGB. The excellent generalize capability of ML methods and their robustness to the noise in the case of low samples data makes them excellent tools to process remote sensing data and provide smart solutions in the field of precision agriculture.
In this study, we considered two scenarios to estimate crop parameters. First, we used four MLAs, including SVR, RF, GB, and XGB, in estimating two crop biophysical parameters. GB and XGB are the novel machine learning algorithms that received less attention in the crop parameters estimation method. This scenario was performed through three steps: (1) using polarimetric SAR features, (2) using optical VIs, and (3) using the integration of SAR and VIs features. These three steps could clearly show that the radar or optical remote sensing data or their combination in estimating crop parameters has excellent potential. Also, we used a deep artificial neural network to model the crop’s LAI and dry biomass in the following scenario. In addition, the deep neural network received less attention in crop parameters estimation. Therefore, we tried to utilize the best performance from these algorithms using feature engineering and suitable parameters tuning. Several features counting VIs extracted from RapidEye spectral reflectance and polarimetric SAR decomposition feature extracted from UAVSAR data were selected to utilize as the input of mentioned algorithms. Moreover, the importance of each feature is investigated attentively. The results were compared, and the best method for estimating crop parameters was determined.

2. Materials and Methods

2.1. Study Area

The SMAPVEX12 campaign has been planned to support the calibration and validation of NASA’s Soil Moisture Active Passive (SMAP) satellite mission products. This campaign was conducted over an agricultural area near Winnipeg, MB, Canada (Figure 1). The study area covers about 12.8 km by 70 km [47]. Five forested sites and 55 agriculture fields were selected for measuring dynamic and static crop (e.g., dry and wet biomass, LAI, crop height, crop density, etc.) and soil parameters (e.g., soil moisture, soil roughness, etc.), including 19 soybeans, eight corn and seven canola fields.

2.2. Sampling Strategy

In situ measurements from crop and soil parameters were collected over six weeks, from 7 June 2012 to 19 July 2012. The selected fields are relatively large, often 800 m × 800 m in size. In each field, 16 sampling sites were designed for sampling. However, crop parameters were measured only at three of these 16 sampling points. For canola, soybean, and corn, ground data collection was done through five plants along two rows (a total of ten plants). The wet biomass weight was immediately determined using a portable scale. Following wet weighing, wet biomass samples are placed in drying facilities for one week at 30 °C. After that, the weight of dry samples was determined. Crop’s LAI was measured through the processing of hemispherical digital photos. Along two transacts in each field, seven photos were captured (a total of 14 photos). The CanEye software was then used to process these photos. The complete information about sampling strategy and collecting ground observations can be found in [47].

2.3. SAR Earth Observations

UAVSAR is a NASA L-band sensor that acquires SAR data in quad polarization mode. During SMAPVEX12, UAVSAR collected multitemporal high-resolution image data over the whole study area of the campaign. The detail of the UAVSAR sensor configurations and the acquisition date are shown in Table 1.
UAVSAR is a fully polarimetric SAR system that can collect a single-look complex (SLC) with 0.6 m × 1.6 m pixel spacing. Near to far range incidence angles are between 21.01° and 64.11°. UAVSAR is typically flown at 41,000 altitudes over the cross-track with a swath of approximately 20 km. The data were calibrated to complex cross-product format, multi-looked, and projected to ground range in simple geographic coordinates by NASA JPL and publicly available [44]. A 3 × 3 boxcar filter was applied to the data for speckle noise reduction. In addition to linear intensities (e.g., VV, HH, and HV), several polarimetric features were extracted from SAR data, including Cloude–Pottier decomposition components, i.e., entropy (H), anisotropy (A), and alfa angle (α) [48,49], Freeman-Durden decomposition components [50], i.e., Surface (OF), Double (DF), and Volume (VF) Scattering, and Yamaguchi decomposition components [51], Surface (OY), Double (DY), Volume (VY), and Helix (HY) Scattering. Besides, Radar Vegetation Index (RVI) was extracted and used as another model-based radar vegetation index using Equation (1):
RVI = 8 σ HV ° σ HH ° + σ HV ° + σ VV °

2.4. Optical Earth Observations

Multi-temporal and multispectral RapidEye images were also used to estimate crop parameters. RapidEye constellation of five identical satellites acquires data in five bands, including blue, green, red, red-edge, and Near-Infrared. Each sensor’s swath width is 77 km, and the ground sampling distance (GSD) at nadir is 6.5 m. The RapidEye images were atmospherically and geometrically corrected using PCI Geomatica’s ATCOR2 and PCI Geomatica’s OrthoEngine, respectively. The geometric correction was done using a rational function, satellite orbital information, ground control points (GCPs) collected from Canada’s National Road Network vector data, and the Canadian Digital Elevation Data (CDED). Several VIs were extracted from RapidEye images. The detail of the 11 VIs used in this study is shown in Table 2.

3. Methodology

MLAs are recently used in classification and regression problems in many areas. In this study, regression models, e.g., RF, GB, XGB, and SVR, were used to estimate crop’s LAI and dry biomass. MLAs were implemented using the open-source Python Scikit-learn package. Besides, deep ANN was implemented using the Keras package. The data were divided into train and test data. Two-thirds (i.e., ~66.7%) of the data were selected to train the models, and the remaining data (i.e., ~33.3%) were used as test data. In this study, first, we calculate the correlation between SAR and optical VIs.
Then, RF feature importance was calculated for each crop and crop parameter separately. The less important features were removed by considering the absolute value of feature importance greater than 0.9 between features. Finally, the remaining features were combined to estimate crop parameters. The selected SAR and VIs features were separately fed into the model as the input data in the first and second steps. Then, the combining of SAR and VIs features was used for modeling. The results of the three separate input data were compared to each other. Furthermore, a deep artificial neural network based on the selected feature was designed and implemented. Furthermore, Grid Search Cross-Validation (GridSearchCV) was used to tune the hyper-parameters of all the ML algorithms.

3.1. Random Forest Regression

RF is a robust ensemble learning method, which is widely used in classification and regression problems. Ensemble learning is the process in which multiple models are produced and combined to solve a particular task. Two common types of ensemble learning are boosting and bagging. Bagging is made up of fitting several models that train independently to reduce variance to avoid overfitting while improving combined models’ stability and accuracy [63]. RF is a successful bagging approach made up of a substantial number of individual decision trees. Each tree makes its prediction. Finally, the model combines all predictions to obtain a better performance [64]. Each tree grows independently using a bootstrap sampling of the training data [29].
In contrast to the linear regression model, an RF regressor model cannot predict outlier data, e,g, predicting the data from training samples. Various researches have used RF regression and classification models to estimate crop parameters or map croplands [31,40,46,65]. The GridSearchCV parameters used in the RF are shown in Table 3.

3.2. Support Vector Regression

The support vector machines (SVMs) algorithm, developed by Vapnik and his colleagues [66], is one of the most widely used kernel-based machine learning algorithms, which is used in a variety of problems, especially in classification tasks [63]. Maintaining all the algorithm’s main features, like maximal margin, SVM can also be used in regression problems. SVR, firstly introduced by Drucker, et al. [67], has several minor differences from SVM. The regression model’s output has infinite numbers, but in SVM, the output is finite numbers.
In regression models, a margin of tolerance (epsilon) is set in approximation. There will be various reasons that make regression models more complicated than the SVM model. SVR gives us the flexibility to define how much error is acceptable in our model and find an appropriate line (or hyperplane in higher dimensions) to fit the data. In this manner, the tube’s points, the points outside the tube, receive penalization; however, the prediction function receives no penalization either above or below. SVR and SVM are widely used in recent researches to estimate crop parameters and cropland mapping [7,26,31,37,45,46,65,68,69]. The Grid Search parameters used for the SVR model are shown in Table 4.

3.3. Gradient Boosting and Extreme Gradient Boosting

GB regression algorithms were subsequently developed by Friedman [70,71]. As we said in part2, two common types of ensemble learning are boosting and bagging. GB, a machine learning method, is an extension of the boosting method. GB, like RF, is used in regression and classification tasks. GB method is based on minimizing a loss function, and various types of loss functions can be used. The regularization techniques are customarily used to reduce overfitting effects. GB negligibly has been used in crop biomass estimation [31,65]. One of the most attractive gradients boosting implementations is XGB [72], first started by Tianqi Chen (Tianqi Chen on http://datascience.la/xgboost-workshop-and-meetup-talk-with-tianqi-chen/, accessed on 3 July 2021) as a research project. It is an ensemble machine learning algorithm that uses a gradient boosting framework. XGB is designed to enhance a machine learning model’s performance, speed, flexibility, and efficiency. The Grid Search parameters used for the GB and XGB algorithms are shown in Table 5.

3.4. Deep Artificial Neural Network Regression

ANNs are popular machine learning algorithms inspired by the human brain [64]. A simplified model of the brain shows a considerable number of primary computing devices called neurons. Through these substantially connected neurons, highly complex computations can be carried out. ANN consists of interconnected neurons that learn by adopting and modifying the weights [29]. This model typically includes one input layer, more than two hidden layers, and one output layer. In the ANN model, neurons of one layer can be connected to all other layers’ neurons but not to the same layer’s neurons. Each neuron is connected to all neurons in the previous and following layers in a fully connected ANN [73].
In this study, we used a dense, deep ANN. The primary considerations for tuning hyper-parameters of ANN are the number of neurons and hidden layers. Several empirical methods can determine the number of neurons in each layer [74,75]. In this study, we have determined the number of the neurons using Equation (2) [74]:
N n = ( m + 2 ) N + 2 N m + 2
In this equation, Nn is the number of neurons in each layer, N is the number of input neurons, and m is the number of layers. We examined various activation functions for the deep ANN model, including ReLU, Tanh, Sigmoid, and Linear. Adam’s optimization method, an extension of Stochastic Gradient Descent (SGD), was used to update the network’s weight iteratively. The early stopping approach was used to avoid overfitting. Furthermore, 20% of the training sample was selected as the validation data.

3.5. Evaluation Criteria

Several criteria were used to evaluate prediction performance, including RMSE, mean absolute error (MAE), and Pearson correlation coefficient (R). The formula of the RMSE is as follows:
RMSE = i = 1 N ( y ^ i y i ) 2 N
where N is the number of data.
In addition, the normalized RMSE (nRMSE) ( RMSE related   crop   parameter   average ) is presented in one figure for better and accurate visualization. MAE is calculated as the following equation:
MAE = i = 1 N | y ^ i y i | N
R is used in statistics problems to measure how the relationship between predicted and observed data is robust:
R = ( N i = 1 N ( y ^ i y i ) ) ( i = 1 N y ^ i ) ( i = 1 N y i ) ( N i = 1 N y ^ i 2 ( i = 1 N y ^ i ) 2 ) ( N i = 1 N y i 2 ( i = 1 N y i ) 2 )

4. Results and Discussion

Several optical VIs and polarimetric SAR data were extracted from RapidEye and UAVSAR data to explore satellite data’s potential to evaluate and estimate crop parameters. The results showed acceptable agreement with the researches had done before by Hosseini, et al. [9] and Reisi Gahrouei, et al. [22]. The impact of optical VIs, UAVSAR polarimetric features, and integrating them on the accuracy of retrieving dry biomass and LAI using four machine learning regression models is assessed in the following sections.

4.1. Time Series Analysis of Radar Backscattering

Figure 2 presents the temporal profiles of the three crops. The left axes represent three SAR backscattering, including VV, HH, and HV. The right axes in the left images are regarding dry biomass, while the right images are regarding LAI. For canola, all three intensities, including VV, HH, and HV, generally decrease from 17 June to 14 July 2012. As expected, the amount of dry biomass during the campaign increased, while the LAI reduced from start to middle of the campaign. Generally, all three SAR backscattering coefficients from 17 June to 14 July 2012, are rising for corn. Also, the amount of dry biomass and LAI increased during the campaign. For soybean, in total, the HV and HH backscattering coefficient is rising, but the VV behavior is irregular. For soybean, similar to corn, the amount of dry biomass and LAI increased during the campaign.

4.2. Correlation and Features Importance

Correlation coefficients between all features extracted for each crop have shown in Figure 3. For canola, the correlation between DF and DY, OF and OY, and VF and VY is high. The absolute correlation between DF and most of the other radar features is generally higher than 0.9. Overall, the correlation between OF and OY with other decompositions is relatively low. The correlation between DF and DY with the other radar features is also high. As well, the correlation between radar features with VIs is low. Between VIs, approximately in most cases, correlation is high. Apart from A, H, and Alpha, the correlation between other SAR features is relatively high for corn. However, the number of radar features with an absolute correlation exceeded 0.9 is negligible. The correlation between DF and DY, OF and OY, and VF and VY is higher than 0.9. Besides, a high correlation can be seen between VIs. Compared to corn and canola, the correlation between SAR features is relatively low.
For corn’s dry biomass, the higher importance is related to MCTI VI. However, between 5 high important features, four of them are SAR parametric features. For corn LAI, similar to dry corn biomass, the higher feature importance is MCTI. Nevertheless, in contrast to corn’s dry biomass, from 5 higher importance features, four of them are related to spectral VI. For canola’s dry biomass, the higher importance is related to DF.
Of five higher importances, three are related to SAR parameters, and the remaining are regarding spectral VIs. For canola LAI, the higher importance is related to RVI. Also, between five high importance features, four of them are regarding SAR parameters. For soybean’s dry biomass, the higher importance is related to CL-EDGE. Besides, four out of five higher importance are related to spectral VIs. Finally, for soybean LAI, similar to soybean’s dry biomass, CL-EDGE has higher importance. In addition, between 5 higher importance, four of them are related to spectral VIs.
The details of the RF feature importance are listed in Table 6. The color of the feature cell with higher importance tends to be green, and the low important feature’s color tend to be yellow. Details of the features used in each crop parameter can be seen in Table 7.

4.3. Sensitivity Analysis

Complete information over validation and calibration accuracies for retrieving dry biomass and LAI for each crop is shown in Table 8 and Table 9, respectively. The following details are for validation data. As maturity methods, canola builds up appreciable plant material and vegetation water. This considerable water volume might lead to a greater tendency towards saturation of signals from canola canopies, especially for SAR backscatter. For canola’ dry biomass, generally, the accuracy of integrated input data was better than the SAR parameters or spectral VIs, separately. The prediction performance of optical VIs was slightly better than the SAR features using. The best performance was related to SVR MLA with RMSE = 26.29 g/m2, MAE = 20.72 g/m2, and R = 0.95 using a combination of SAR and optical data (Figure 4a). A low amount of overestimated value can be seen in the early growth stage. Moreover, SVR underestimated dry biomass for canola at advanced development stages. Low error among dry biomass estimation using VIs spectral data provided by GB method with RMSE of 38.87 g/m2 and MAE of 28.34 g/m2, considerably higher than integrated estimation error. Using SAR polarimetric features, high accuracy was delivered by RF algorithms (RMSE = 46.67 g/m2, 33.45 g/m2, and R = 0.83). The results showed no saturation in the high and low amount of dry biomass prediction using a combination of optic and SAR data.
LAI is indicative of the crop structure and affects both reflectance and backscatter at canopy scales. For canola LAI, the integration of SAR polarimetric data and spectral VIs in GB and XGB improved performance. However, optical VIs outperforms the integration of SAR and optical features in SVR and RF. Besides, the accuracy of LAI estimation with VIs spectral data was better than SAR polarimetric data. As demonstrated by Figure 4b, LAI for canola estimated by GB using both SAR and VIs feature was highly correlated with in situ measured LAI (0.557 m2/m2, 0.399 m2/m2, and R = 0.95). However, GB slightly underestimates the LAI in the mid-growth stage. The higher accuracy with spectral VIs regarded to GB and XGB, with approximately the same RMSE, but better MAE in XGB. A minimal amount of saturation could be seen in high values of LAI.
For corn’s dry biomass, like canola’s dry biomass, integrated features have higher accuracy. The performance of optical VIs data was better than SAR polarimetric parameters in RF, GB, and XGB, while in SVR, the accuracy of SAR polarimetric features is higher than VIs spectral data. The prediction accuracy delivered by RF, SVR, and GB has the same R, but RF has a lower error than SVR and GB. The higher accuracy was related to the RF regression model with RMSE = 57.97 g/m2, MAE = 41.15 g/m2, and R = 0.96 (Figure 4c). Although a few ground measurements are available at high dry biomass, observed and estimated, dry biomass values are well distributed about the 1:1 line. The best performance of optical VIs regarded to GB regression model with RMSE = 69.85 g/m2, MAE = 44.37 g/m2, R = 0.94. The higher accuracy with SAR polarimetric data was related to the SVR model with RMSE = 60.2 g/m2, MAE = 46.94 g/m2, and R = 0.94. The result of RF show low saturation in the high amount of corn’s dry biomass.
For corn’s LAI, the best performance was related to integrating SAR polarimetric data and spectral VIs. Besides, the estimation accuracy of spectral VIs is slightly worse than SAR polarimetric data. The best accuracy was regarding the GB regression method with RMSE = 0.298 m2/m2, MAE = 0.219 m2/m2, and R = 0.96 using a combination of SAR and optical VIs features (Figure 4d). Early in the season, the GB overestimates LAI. This may be due to a more open canopy in the early growth stages, leaving more soil exposed to soil properties, contributing significantly to reflectance and backscatter. The RMSE of SVR, RF, and XGB is nearly equal. The best performance for spectral VIs was provided by the XGB model with RMSE = 0.399 m2/m2, MAE = 0.273 m2/m2, and R = 0.92. The best performance for SAR data was delivered by the XGB model with RMSE = 0.321 m2/m2, MAE = 0.221 m2/m2, and R = 0.95.
For soybean’s dry biomass, generally, the integration of SAR and optical data had better performance. The estimation performance of SAR polarimetric data was slightly better than optic data in all cases. The best performance among MLAs was related to GB regression model with RMSE = 5.00 g/m2, MAE = 3.5 g/m2, and R = 0.94 (Figure 4e). SVR had the lower MAE among all algorithms; however, RF has the lower RMSE comparison to SVR. Also, RF and GB had a similar MAE, but RF had the lower RMSE. ML algorithms had a significant saturation in the high value of dry biomass (higher than 60 g/m2).
For soybean’s LAI, like corn and canola, the best performance belongs to integrating SAR and optic data. Also, the accuracy of SAR polarimetric data, compared to spectral VIs, was better. The XGB and GB had the same RMSE, however, XGB had a better MAE (RMSE = 0.233 m2/m2, MAE = 0.164 m2/m2, and R = 0.94 (Figure 4f). The higher accuracy using SAR data was related to SVR with RMSE = 0.291 m2/m2, MAE = 0.209 m2/m2, and R = 0.9. Using optical VIs data XGB provided better results with RMSE = 0.36 m2/m2, MAE = 0.247 m2/m2, and R = 0.84. The best performance in estimating dry biomass and LAI for each crop is shown in Figure 4.
Figure 5 shows the results of four MLAs’ nRMSE for three crops shown in the boxplot. The quartile of distribution shows in the box. The rest of the dataset showed whiskers. Each boxplot’s data consisted of each method’s calibration and validation data, including three various input data (SAR polarimetric features, VIs spectral data, and integration of SAR and optical features). The results showed that MLAs performed better in canola rather than corn and soybean. For corn’s dry biomass, SVR has better accuracy rather than the other methods. In addition, the median for XGB in corn dry biomass and LAI are lower, which means better accuracy.

4.4. The Results of Deep Neural Network

The results of Deep ANN can be seen in Figure 6. The results showed that Deep ANN, in all cases, improved the accuracy of estimation. For canola dry biomass and LAI, the model delivered the RMSE of 25.8 g/m2 and 0.525 m2/m2, respectively (Figure 6a,b). The results of deep ANN clearly showed improved canola dry biomass and LAI. For corn dry biomass and LAI, the deep ANN provided the RMSE of 54.43 g/m2 and 0.273 m2/m2, respectively (Figure 6c,d). For both corn LAI and dry biomass, deep ANN improved the retrieval accuracy. Finally, for soybean dry biomass and LAI, the deep ANN provided the RMSE of 4.95 g/m2 and 0.211 m2/m2, respectively (Figure 6e,f). Besides, the deep ANN slightly improved the estimation’s accuracy for both LAI and dry biomass. A minimal amount of saturation can be seen in the high biomass for soybean’ dry biomass (approximately higher than 60 g/m2).

4.5. Discussion

For the last decades, remote sensing satellite SAR and optic data’s progress provides an environment for further research on crop biophysical parameters. MLAs showed significant potential in broad areas; utilizing these methods recently has grown to solve remote sensing problems. Crop biophysical parameters are vital parameters for crop monitoring, crop stress assessments, crop growth model, to name but a few. Identify the number of train and test samples, the best value for each parameter in tuning MLAs’ hyperparameters, and many other things that are not mentioned here, are the reasons that we need to compare several MLAs to determine the best approach to estimate target parameters. In this study, we focused on the potential of four MLAs to assess two crop biophysical parameters. XGB is a new method that is used in the fields related to crop parameter estimation. Information during crop growth duration is available from UAVSAR data. In general, for all three crops, a combination of UAVSAR polarimetric features and spectral VIs have a better performance to estimate crop biomass and LAI. the estimation accuracy of regression models UAVSAR L-band polarimetric features showed great potential in retrieving soybean dry biomass and LAI. For canola and corn LAI and dry biomass, generally, the accuracy of estimation using optical VIs was better than SAR polarimetric features in each regression model.
Considering other research works, Reisi-Gahrouei, et al. [44] achieved RMSE of 56.55 g/m2 and R = 0.72 for canola’s dry biomass using decomposition UAVSAR L-band data. Besides, they achieved an RMSE of 13.48 g/m2 and R = 0.82 for soybean’s dry biomass. In another study, Reisi Gahrouei, et al. [22] achieved 25.22 g/m2 for canola, 88.13 g/m2 for corn, 5.91 g/m2 for soybean using spectral VIs extracted from RapidEye optical data. Their model delivered RMSE of 0.59 m2/m2 for canola, 0.27 m2/m2 for corn and 0.21 m2/m2 for soybean, a combination of UAVSAR L-band data and spectral VIs improved the soybean and canola dry biomass estimation in our study. Mandal, et al. [76] used various methods to estimate wet biomass and PAI of soybean and wheat. They achieved RMSE between 0.73 to 1.21 g/m2 for wheat’s wet biomass. As well, their results for wheat PAI were between 0.83 to 1.48 m2/m2. The best soybean wet biomass and PAI results were 0.34 g/m2 and 0.72 m2/m2, respectively. We achieved the RMSE of 4.95 g/m2 and 25.80 g/m2 for soybean and canola dry biomass, respectively. Our model also provided the RMSE of 0.211 m2/m2 for soybean LAI, 0.273 m2/m2 for corn LAI, and 0.525 m2/m2 for canola LAI. Our model amazingly improved the accuracy of LAI estimation, especially for corn.

5. Conclusions

Biomass and LAI are two critical parameters in the crop growth model and crop monitoring. This paper assessed four MLAs’ potential to estimate dry biomass and LAI of three crops, including soybean, corn, and canola. In situ measurements have been collected during the SMAPVEX-12 campaign over Manitoba, Canada. Several polarimetric features were extracted from UAVSAR data. Besides, various spectral VIs were extracted from RapidEye optical data. Correlation for all features was calculated; also, RF feature importance for each feature was obtained. Finally, the correlation with an absolute value of 0.9 was considered, and the feature with low importance and high correlation was removed. The remaining features were incorporated into machine learning regression models. The results showed that the integration of SAR polarimetric and spectral VIs better estimate dry biomass and LAI. Besides, XGB showed great potential in assessing crop biophysical parameters. For LAI, RMSE was reported as 0.557 m2/m2 for canola, 0.298 m2/m2 for corn, and 0.233 m2/m2 for soybean. Also, RMSE was reported for dry biomass as 29.45 g/m2 for canola, 26.29 g/m2 for corn, 5.00 g/m2 for soybean. In addition, the results of deep neural networks were 0.525 m2/m2, 0.273 m2/m2, and 0.211 m2/m2 for canola, corn, and soybean LAI, respectively. The results of deep neural networks were 25.80 g/m2, 57.97 g/m2, 5.00 g/m2 for canola, corn, and soybean dry biomass, respectively.

Author Contributions

H.B., S.H., S.M., and M.M.; methodology, H.B. and O.R.-G.; data processing, H.B. and M.M.; programming and implementation; H.B., S.H. and M.M.; validation, H.B., S.H., M.M., A.S. and S.M.; formal analysis: H.B., S.H., A.S., S.M. and M.M.; investigation, H.B. and S.H.; resources, S.H., M.M.; data curation, H.B., S.H. and M.M.; writing—original draft preparation, H.B., S.H., A.S., S.M. and M.M.; writing—review and editing, H.B.; visualization. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The SMAPVEX12 ground sampling data used in this study is available on https://smapvex12.espaceweb.usherbrooke.ca/ (access on 13 June 2021).

Acknowledgments

The authors thank the SMAPVEX12 team, Agriculture and Agri-Food Canada, for providing the in-situ measurements. Moreover, we truly appreciate the NASA Jet Propulsion Laboratory (JPL) for providing UAVSAR data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiao, X.; McNairn, H.; Shang, J.; Pattey, E.; Liu, J.; Champagne, C. The sensitivity of RADARSAT-2 polarimetric SAR data to corn and soybean leaf area index. Can. J. Remote Sens. 2011, 37, 69–81. [Google Scholar] [CrossRef]
  2. Pichierri, M.; Hajnsek, I.; Zwieback, S.; Rabus, B. On the potential of Polarimetric SAR Interferometry to characterize the biomass, moisture and structure of agricultural crops at L-, C- and X-Bands. Remote Sens. Environ. 2018, 204, 596–616. [Google Scholar] [CrossRef]
  3. Kross, A.; McNairn, H.; Lapen, D.; Sunohara, M.; Champagne, C. Assessment of RapidEye vegetation indices for estimation of leaf area index and biomass in corn and soybean crops. Int. J. Appl. Earth Obs. Geoinf. 2015, 34, 235–248. [Google Scholar] [CrossRef] [Green Version]
  4. Asrar, G.; Kanemasu, E.; Jackson, R.; Pinter, P., Jr. Estimation of Total above-Ground Phyto-mass Production Using Remotely Sensed Data. Remote Sens. Environ. 1985, 17, 211–220. [Google Scholar] [CrossRef]
  5. Asad, M.H.; Bais, A. Crop and Weed Leaf Area Index Mapping Using Multi-Source Remote and Proximal Sensing. IEEE Access 2020, 8, 138179–138190. [Google Scholar] [CrossRef]
  6. Canisius, F.; Shang, J.; Liu, J.; Huang, X.; Ma, B.; Jiao, X.; Geng, X.; Kovacs, J.M.; Walters, D. Tracking crop phenological development using multi-temporal polarimetric Radarsat-2 data. Remote Sens. Environ. 2018, 210, 508–518. [Google Scholar] [CrossRef]
  7. Luo, P.; Liao, J.; Shen, G. Combining Spectral and Texture Features for Estimating Leaf Area Index and Biomass of Maize Using Sentinel-1/2, and Landsat-8 Data. IEEE Access 2020, 8, 53614–53626. [Google Scholar] [CrossRef]
  8. McNairn, H.; Champagne, C.; Shang, J.; Holmstrom, D.; Reichert, G. Integration of optical and Synthetic Aperture Radar (SAR) imagery for delivering operational annual crop inventories. ISPRS J. Photogramm. Remote Sens. 2009, 64, 434–449. [Google Scholar] [CrossRef]
  9. Hosseini, M.; McNairn, H.; Merzouki, A.; Pacheco, A. Estimation of Leaf Area Index (LAI) in corn and soybeans using multi-polarization C- and L-band radar data. Remote Sens. Environ. 2015, 170, 77–89. [Google Scholar] [CrossRef]
  10. Huang, Y.; Walker, J.P.; Gao, Y.; Wu, X.; Monerris, A. Estimation of Vegetation Water Content from the Radar Vegetation Index at L-Band. IEEE Trans. Geosci. Remote Sens. 2015, 54, 981–989. [Google Scholar] [CrossRef]
  11. Inoue, Y.; Kurosu, T.; Maeno, H.; Uratsuka, S.; Kozu, T.; Dabrowska-Zielinska, K.; Qi, J. Season-Long Daily Measurements of Multifrequency (Ka, Ku, X, C, and L) and Full-Polarization Backscatter Signatures over Paddy Rice Field and Their Relationship with Bi-ological Variables. Remote Sens. Environ. 2002, 81, 194–204. [Google Scholar] [CrossRef]
  12. Jia, M.; Tong, L.; Zhang, Y.; Chen, Y. Rice Biomass Estimation Using Radar Backscattering Data at S-band. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 469–479. [Google Scholar] [CrossRef]
  13. Wu, F.; Wang, C.; Zhang, H.; Zhang, B.; Tang, Y. Rice Crop Monitoring in South China With RADARSAT-2 Quad-Polarization SAR Data. IEEE Geosci. Remote Sens. Lett. 2011, 8, 196–200. [Google Scholar] [CrossRef]
  14. Brown, S.; Quegan, S.; Morrison, K.; Bennett, J.; Cookmartin, G. High-resolution measurements of scattering in wheat canopies-implications for crop parameter retrieval. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1602–1610. [Google Scholar] [CrossRef] [Green Version]
  15. Wu, S.-T. Potential Application of Multipolarization SAR for Pine-Plantation Biomass Estimation. IEEE Trans. Geosci. Remote Sens. 1987, GE-25, 403–409. [Google Scholar] [CrossRef]
  16. Karjalainen, M.; Kaartinen, H.; Hyyppä, J. Agricultural Monitoring Using Envisat Alternating Polarization SAR Images. Photogramm. Eng. Remote Sens. 2008, 74, 117–126. [Google Scholar] [CrossRef]
  17. Molijn, R.A.; Iannini, L.; Rocha, J.V.; Hanssen, R.F. Sugarcane Productivity Mapping through C-Band and L-Band SAR and Optical Satellite Imagery. Remote Sens. 2019, 11, 1109. [Google Scholar] [CrossRef] [Green Version]
  18. Wiseman, G.; McNairn, H.; Homayouni, S.; Shang, J. RADARSAT-2 Polarimetric SAR Response to Crop Biomass for Agricultural Production Monitoring. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4461–4471. [Google Scholar] [CrossRef]
  19. Hosseini, M.; McNairn, H.; Mitchell, S.; Robertson, L.D.; Davidson, A.; Homayouni, S. Synthetic aperture radar and optical satellite data for estimating the biomass of corn. Int. J. Appl. Earth Obs. Geoinf. 2019, 83, 101933. [Google Scholar] [CrossRef]
  20. Homayouni, S.; McNairn, H.; Hosseini, M.; Jiao, X.; Powers, J. Quad and compact multitemporal C-band PolSAR observations for crop characterization and monitoring. Int. J. Appl. Earth Obs. Geoinf. 2019, 74, 78–87. [Google Scholar] [CrossRef]
  21. Moran, M.S.; Alonso, L.; Moreno, J.F.; Mateo, M.P.C.; De La Cruz, D.F.; Montoro, A. A RADARSAT-2 Quad-Polarized Time Series for Monitoring Crop and Soil Conditions in Barrax, Spain. IEEE Trans. Geosci. Remote Sens. 2012, 50, 1057–1070. [Google Scholar] [CrossRef]
  22. Gahrouei, O.R.; McNairn, H.; Hosseini, M.; Homayouni, S. Estimation of Crop Biomass and Leaf Area Index from Multitemporal and Multispectral Imagery Using Machine Learning Approaches. Can. J. Remote Sens. 2020, 46, 84–99. [Google Scholar] [CrossRef]
  23. Baghdadi, N.; Boyer, N.; Todoroff, P.; El Hajj, M.; Begue, A. Potential of SAR sensors TerraSAR-X, ASAR/ENVISAT and PALSAR/ALOS for monitoring sugarcane crops on Reunion Island. Remote Sens. Environ. 2009, 113, 1724–1738. [Google Scholar] [CrossRef]
  24. Wali, E.; Tasumi, M.; Moriyama, M. Combination of Linear Regression Lines to Understand the Response of Sentinel-1 Dual Polarization SAR Data with Crop Phenology—Case Study in Miyazaki, Japan. Remote Sens. 2020, 12, 189. [Google Scholar] [CrossRef] [Green Version]
  25. Clerici, N.; Calderón, C.A.V.; Posada, J.M. Fusion of Sentinel-1A and Sentinel-2A data for land cover mapping: A case study in the lower Magdalena region, Colombia. J. Maps 2017, 13, 718–726. [Google Scholar] [CrossRef] [Green Version]
  26. Mandal, D.; Kumar, V.; Lopez-Sanchez, J.M.; Bhattacharya, A.; McNairn, H.; Rao, Y.S. Crop biophysical parameter retrieval from Sentinel-1 SAR data with a multi-target inversion of Water Cloud Model. Int. J. Remote Sens. 2020, 41, 5503–5524. [Google Scholar] [CrossRef]
  27. Bao, N.; Li, W.; Gu, X.; Liu, Y. Biomass Estimation for Semiarid Vegetation and Mine Rehabilitation Using Worldview-3 and Sentinel-1 SAR Imagery. Remote Sens. 2019, 11, 2855. [Google Scholar] [CrossRef] [Green Version]
  28. Lu, B.; He, Y. Leaf Area Index Estimation in a Heterogeneous Grassland Using Optical, SAR, and DEM Data. Can. J. Remote Sens. 2019, 45, 618–633. [Google Scholar] [CrossRef]
  29. Mao, H.; Meng, J.; Ji, F.; Zhang, Q.; Fang, H. Comparison of Machine Learning Regression Algorithms for Cotton Leaf Area Index Retrieval Using Sentinel-2 Spectral Bands. Appl. Sci. 2019, 9, 1459. [Google Scholar] [CrossRef] [Green Version]
  30. Filho, M.G.; Kuplich, T.M.; De Quadros, F.L.F. Estimating natural grassland biomass by vegetation indices using Sentinel 2 remote sensing data. Int. J. Remote Sens. 2019, 41, 2861–2876. [Google Scholar] [CrossRef]
  31. Mansaray, L.R.; Kanu, A.S.; Yang, L.; Huang, J.; Wang, F. Evaluation of machine learning models for rice dry biomass estimation and mapping using quad-source optical imagery. GISci. Remote Sens. 2020, 57, 785–796. [Google Scholar] [CrossRef]
  32. Dong, T.; Liu, J.; Qian, B.; He, L.; Liu, J.; Wang, R.; Jing, Q.; Champagne, C.; McNairn, H.; Powers, J.; et al. Estimating crop biomass using leaf area index derived from Landsat 8 and Sentinel-2 data. ISPRS J. Photogramm. Remote Sens. 2020, 168, 236–250. [Google Scholar] [CrossRef]
  33. Wu, C.; Shen, H.; Shen, A.; Deng, J.; Gan, M.; Zhu, J.; Xu, H.; Wang, K. Comparison of machine-learning methods for above-ground biomass estimation based on Landsat imagery. J. Appl. Remote Sens. 2016, 10, 35010. [Google Scholar] [CrossRef]
  34. Porter, T.F.; Chen, C.; Long, J.A.; Lawrence, R.L.; Sowell, B.F. Estimating biomass on CRP pastureland: A comparison of remote sensing techniques. Biomass Bioenergy 2014, 66, 268–274. [Google Scholar] [CrossRef]
  35. Blinn, C.E.; House, M.N.; Wynne, R.H.; Thomas, V.A.; Fox, T.R.; Sumnall, M. Landsat 8 Based Leaf Area Index Estimation in Loblolly Pine Plantations. Forest 2019, 10, 222. [Google Scholar] [CrossRef] [Green Version]
  36. Gaso, D.V.; Berger, A.G.; Ciganda, V.S. Predicting wheat grain yield and spatial variability at field scale using a simple regression or a crop model in conjunction with Landsat images. Comput. Electron. Agric. 2019, 159, 75–83. [Google Scholar] [CrossRef]
  37. Deb, D.; Deb, S.; Chakraborty, D.; Singh, J.P.; Singh, A.K.; Dutta, P.; Choudhury, A. Aboveground biomass estimation of an agro-pastoral ecology in semi-arid Bundelkhand region of India from Landsat data: A comparison of support vector machine and traditional regression models. Geocarto Int. 2020, 1–16. [Google Scholar] [CrossRef]
  38. Räsänen, A.; Juutinen, S.; Aurela, M.; Virtanen, T. Predicting aboveground biomass in Arctic landscapes using very high spatial resolution satellite imagery and field sampling. Int. J. Remote Sens. 2018, 40, 1175–1199. [Google Scholar] [CrossRef] [Green Version]
  39. Han, J.; Wei, C.; Chen, Y.; Liu, W.; Song, P.; Zhang, D.; Wang, A.; Song, X.; Wang, X.; Huang, J. Mapping Above-Ground Biomass of Winter Oilseed Rape Using High Spatial Resolution Satellite Data at Parcel Scale under Waterlogging Conditions. Remote Sens. 2017, 9, 238. [Google Scholar] [CrossRef] [Green Version]
  40. Sakamoto, T. Incorporating environmental variables into a MODIS-based crop yield estimation method for United States corn and soybeans through the use of a random forest regression algorithm. ISPRS J. Photogramm. Remote Sens. 2020, 160, 208–228. [Google Scholar] [CrossRef]
  41. Ali, I.; Cawkwell, F.; Dwyer, E.; Green, S. Modeling Managed Grassland Biomass Estimation by Using Multitemporal Remote Sensing Data—A Machine Learning Approach. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 3254–3264. [Google Scholar] [CrossRef]
  42. Verrelst, J.; Camps-Valls, G.; Muñoz-Marí, J.; Rivera, J.P.; Veroustraete, F.; Clevers, J.G.; Moreno, J. Optical remote sensing and the retrieval of terrestrial vegetation bio-geophysical properties—A review. ISPRS J. Photogramm. Remote Sens. 2015, 108, 273–290. [Google Scholar] [CrossRef]
  43. Ali, I.; Greifeneder, F.; Stamenkovic, J.; Neumann, M.; Notarnicola, C. Review of Machine Learning Approaches for Biomass and Soil Moisture Retrievals from Remote Sensing Data. Remote Sens. 2015, 7, 16398–16421. [Google Scholar] [CrossRef] [Green Version]
  44. Reisi-Gahrouei, O.; Homayouni, S.; McNairn, H.; Hosseini, M.; Safari, A. Crop biomass estimation using multi regression analysis and neural networks from multitemporal L-band polarimetric synthetic aperture radar data. Int. J. Remote Sens. 2019, 40, 6822–6840. [Google Scholar] [CrossRef]
  45. Sharifi, A.; Hosseingholizadeh, M. Application of Sentinel-1 Data to Estimate Height and Biomass of Rice Crop in Astaneh-ye Ashrafiyeh, Iran. J. Indian Soc. Remote Sens. 2019, 48, 11–19. [Google Scholar] [CrossRef]
  46. Zhu, W.; Sun, Z.; Peng, J.; Huang, Y.; Li, J.; Zhang, J.; Yang, B.; Liao, X. Estimating Maize Above-Ground Biomass Using 3D Point Clouds of Multi-Source Unmanned Aerial Vehicle Data at Multi-Spatial Scales. Remote Sens. 2019, 11, 2678. [Google Scholar] [CrossRef] [Green Version]
  47. McNairn, H.; Jackson, T.J.; Wiseman, G.; Belair, S.; Berg, A.; Bullock, P.; Colliander, A.; Cosh, M.H.; Kim, S.-B.; Magagi, R.; et al. The Soil Moisture Active Passive Validation Experiment 2012 (SMAPVEX12): Prelaunch Calibration and Validation of the SMAP Soil Moisture Algorithms. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2784–2801. [Google Scholar] [CrossRef]
  48. Cloude, S.; Pottier, E. A review of target decomposition theorems in radar polarimetry. IEEE Trans. Geosci. Remote Sens. 1996, 34, 498–518. [Google Scholar] [CrossRef]
  49. Cloude, S.; Pottier, E. An entropy based classification scheme for land applications of polarimetric SAR. IEEE Trans. Geosci. Remote Sens. 1997, 35, 68–78. [Google Scholar] [CrossRef]
  50. Freeman, A.; Durden, S.L. Three-component scattering model to describe polarimetric SAR data. In Proceedings of the Radar Polarimetry, San Diego, CA, USA, 22–22 July 1992. [Google Scholar]
  51. Yamaguchi, Y.; Moriyama, T.; Ishido, M.; Yamada, H. Four-component scattering model for polarimetric SAR image decomposition. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1699–1706. [Google Scholar] [CrossRef]
  52. Rousel, J.; Haas, R.; Schell, J.; Deering, D. Monitoring Vegetation Systems in the Great Plains with Erts. In Paper Presented at the Proceedings of the Third Earth Resources Technology Satellite—1 Symposium; NASA: Washington, DC, USA, 1973. [Google Scholar]
  53. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a green channel in remote sensing of global vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  54. Jordan, C.F. Derivation of Leaf-Area Index from Quality of Light on the Forest Floor. Ecology 1969, 50, 663–666. [Google Scholar] [CrossRef]
  55. Gitelson, A.; Merzlyak, M.N. Spectral Reflectance Changes Associated with Autumn Senescence of Aesculus hippocastanum L. and Acer platanoides L. Leaves. Spectral Features and Relation to Chlorophyll Estimation. J. Plant Physiol. 1994, 143, 286–292. [Google Scholar] [CrossRef]
  56. Haboudane, D.; Miller, J.R.; Pattey, E.; Zarco-Tejada, P.J.; Strachan, I.B. Hyper-spectral Vegetation Indices and Novel Algorithms for Predicting Green Lai of Crop Canopies: Modeling and Validation in the Context of Precision Agriculture. Remote Sens. Environ. 2004, 90, 337–352. [Google Scholar] [CrossRef]
  57. Huete, A.; Didan, K.; Miura, T.; Rodriguez, E.P.; Gao, X.; Ferreira, L.G. Overview of the radiometric and biophysical performance of the MODIS vegetation indices. Remote Sens. Environ. 2002, 83, 195–213. [Google Scholar] [CrossRef]
  58. Gitelson, A.A.; Verma, S.B.; Vina, A.; Rundquist, D.C.; Keydan, G.; Leavitt, B.; Arkebauer, T.J.; Burba, G.G.; Suyker, A.E. Novel Technique for Remote Estimation of CO2 Flux in Maize. Geophys. Res. Lett. 2003, 30. [Google Scholar] [CrossRef] [Green Version]
  59. Gitelson, A.A.; Viña, A.; Ciganda, V.; Rundquist, D.C.; Arkebauer, T.J. Remote estimation of canopy chlorophyll content in crops. Geophys. Res. Lett. 2005, 32. [Google Scholar] [CrossRef] [Green Version]
  60. Gitelson, A.A.; Viña, A.; Arkebauer, T.J.; Rundquist, D.C.; Keydan, G.; Leavitt, B. Remote estimation of leaf area index and green leaf biomass in maize canopies. Geophys. Res. Lett. 2003, 30. [Google Scholar] [CrossRef] [Green Version]
  61. Dash, J.; Curran, P.J.; Tallis, M.J.; Llewellyn, G.M.; Taylor, G.; Snoeij, P. Validating the MERIS Terrestrial Chlorophyll Index (MTCI) with ground chlorophyll content data at MERIS spatial resolution. Int. J. Remote Sens. 2010, 31, 5513–5532. [Google Scholar] [CrossRef] [Green Version]
  62. Huete, A. Huete, Ar a Soil-Adjusted Vegetation Index (Savi). Remote Sensing of Envi-ronment. Remote Sens. Environ. 1998, 25, 295–309. [Google Scholar] [CrossRef]
  63. Sheykhmousa, M.; Mahdianpari, M.; Ghanbari, H.; Mohammadimanesh, F.; Ghamisi, P.; Homayouni, S. Support Vector Machine Versus Random Forest for Remote Sensing Image Classification: A Meta-Analysis and Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6308–6325. [Google Scholar] [CrossRef]
  64. Dangeti, P. Statistics for Machine Learning; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  65. Mansaray, L.R.; Zhang, K.; Kanu, A.S. Dry biomass estimation of paddy rice with Sentinel-1A satellite data using machine learning regression algorithms. Comput. Electron. Agric. 2020, 176, 105674. [Google Scholar] [CrossRef]
  66. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  67. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.J.; Vapnik, V. Support Vector Regression Machines. Paper Presented at the Advances in Neural Information Processing Systems 1997; Available online: https://www.researchgate.net/publication/2880432_Support_Vector_Regression_Machines (accessed on 3 July 2021).
  68. Mandal, D.; Rao, Y. SASYA: An integrated framework for crop biophysical parameter retrieval and within-season crop yield prediction with SAR remote sensing data. Remote Sens. Appl. Soc. Environ. 2020, 20, 100366. [Google Scholar] [CrossRef]
  69. Duan, B.; Liu, Y.; Gong, Y.; Peng, Y.; Wu, X.; Zhu, R.; Fang, S. Remote Estimation of Rice Lai Based on Fourier Spectrum Texture from Uav Image. Plant Methods 2019, 15, 124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  71. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  72. Brownlee, J. Xgboost with Python. Mach. Learn. Mastery 2018. Available online: https://docplayer.net/132742546-Jason-brownlee-xgboost-with-python-7-day-mini-course.html (accessed on 3 July 2021).
  73. Vasilev, I.; Slater, D.; Spacagna, G.; Roelants, P.; Zocca, V. Python Deep Learning: Exploring Deep Learning Techniques and Neural Network Architectures with Pytorch, Keras, and Tensorflow; Packt Publishing Ltd.: Birmingham, UK, 2019. [Google Scholar]
  74. Huang, G.-B. Learning capability and storage capacity of two-hidden-layer feedforward networks. IEEE Trans. Neural Netw. 2003, 14, 274–281. [Google Scholar] [CrossRef] [Green Version]
  75. Madhiarasan, M.; Deepa, S.N. Comparative analysis on hidden neurons estimation in multi layer perceptron neural networks for wind speed forecasting. Artif. Intell. Rev. 2017, 48, 449–471. [Google Scholar] [CrossRef]
  76. Mandal, D.; Kumar, V.; McNairn, H.; Bhattacharya, A.; Rao, Y. Joint estimation of Plant Area Index (PAI) and wet biomass in wheat and soybean from C-band polarimetric SAR data. Int. J. Appl. Earth Obs. Geoinf. 2019, 79, 24–34. [Google Scholar] [CrossRef]
Figure 1. Location of study area ding SMAPVEX-12 campaign, the identified fields have been used in this research.
Figure 1. Location of study area ding SMAPVEX-12 campaign, the identified fields have been used in this research.
Agronomy 11 01363 g001
Figure 2. Temporal profile analysis for SAR backscatter coefficient of dry biomass and LAI for (a) canola, (b) corn, (c) soybean.
Figure 2. Temporal profile analysis for SAR backscatter coefficient of dry biomass and LAI for (a) canola, (b) corn, (c) soybean.
Agronomy 11 01363 g002
Figure 3. Correlation analysis calculated for SAR and optic features for (a) Canola, (b) Corn, (c) Soybean.
Figure 3. Correlation analysis calculated for SAR and optic features for (a) Canola, (b) Corn, (c) Soybean.
Agronomy 11 01363 g003
Figure 4. The best performance of machine learning regression models in estimating (a) canola dry biomass, (b) canola LAI, (c) corn dry biomass, (d) corn LAI, (e) soybean dry biomass, and (f) soybean LAI.
Figure 4. The best performance of machine learning regression models in estimating (a) canola dry biomass, (b) canola LAI, (c) corn dry biomass, (d) corn LAI, (e) soybean dry biomass, and (f) soybean LAI.
Agronomy 11 01363 g004
Figure 5. Comparing four MLAs’ nRMSE for three crops (a) dry biomass and (b) LAI.
Figure 5. Comparing four MLAs’ nRMSE for three crops (a) dry biomass and (b) LAI.
Agronomy 11 01363 g005
Figure 6. The results of Deep ANN for (a) canola’s dry biomass, (b) canola’s LAI, (c) corn’s dry biomass, (d) corn’s LAI, (e) soybean’s dry biomass, and (f) soybean’s LAI.
Figure 6. The results of Deep ANN for (a) canola’s dry biomass, (b) canola’s LAI, (c) corn’s dry biomass, (d) corn’s LAI, (e) soybean’s dry biomass, and (f) soybean’s LAI.
Agronomy 11 01363 g006
Table 1. UAVSAR sensor configuration and acquisition date.
Table 1. UAVSAR sensor configuration and acquisition date.
Data CharacteristicsUAVSAR
FrequencyL-band (1.26 GHz)
PolarizationsQuad-Pol (HH, HV, VH, VV)
Spatial Resolution1.66 m range × 0.8 m azimuth SLC
Incident angle25 to 65 degrees (relevant: 35 to 45 degrees)
Acquisition dates17, 19, 22, 23, 25, 27, 29 June and 5, 8, 10, 13, 14, 17 July
Table 2. Detail of the VIs extracted from RapidEye optic sensor used in this study.
Table 2. Detail of the VIs extracted from RapidEye optic sensor used in this study.
Index (Abbreviation)FormulaRef.
Normalized Difference Vegetation Index (NDVI) R N I R     R R R N I R   +   R R [52]
Green NDVI (GNDVI) R N I R     R G R N I R   +   R G [53]
Simple Ratio (SR) R N I R R R [54]
Red-Edge Normalized Difference Vegetation Index (NDVI-RE) R N I R     R R E R N I R   +   R R E [55]
Red-Edge Simple Ratio(SR-RE) R N I R R R E [55]
Modified Triangular Vegetation Index (MTVI2) 1.5 { 1.2 ( R N I R     R G )     2.5 ( R R     R G ) } ( 2 R N I R   +   1 ) 2     ( 6 R N I R     5 R R )     0.5 [56]
Enhanced Vegetation Index (EVI) 2.5 ( R N I R     R R ) 1   +   R N I R   +   6 R R     7.5 R B [57]
Green Chlorophyll Index (CL-G) R N I R R G 1 [58,59,60]
Red-Edge Chlorophyll Index (CL-RE) R N I R R R E 1 [58,59,60]
MERIS Trrestrial Chlorophyll Index (MTCI) R N I R     R R E R R   +   R R E [61]
Soil-Adjusted Vegetation Index (SAVI) ( 1   +   0.5 ) ( R N I R     R R ) R N I R   +   R R   +   0.5 [62]
Table 3. Grid Search parameters used in the RF model.
Table 3. Grid Search parameters used in the RF model.
ParametersDescriptionGrid Search Values
n_estimatorsNo. of trees in the forest5, 10, 25, 50, 100
max_depthMaximum depth of the trees2, 3, 5, 8, 10
min_samples_splitMinimum number of samples required to split an internal node2, 3, 5, 10
min_samples_leafMinimum number of samples required to be at a leaf node1, 2, 3, 5, 10
Table 4. GridSearchCV parameters for SVR.
Table 4. GridSearchCV parameters for SVR.
ParametersDescriptionGrid Search Values
KernelSpecifies the kernel type to be used in the algorithm‘linear’, ‘rbf’, ‘poly’
GammaKernel coefficient for ‘rbf’, ‘poly’ and ‘sigmoid’0.001, 0.01, 0.1, 0.5, 0.8, 1, 3
Degreedegree of polynomial2, 3
CPenalty parameter5, 10, 20, 50, 100, 200, 500, 1000
Table 5. GridSearchCV parameters for the GB and XGB algorithms.
Table 5. GridSearchCV parameters for the GB and XGB algorithms.
ParametersDescriptionGrid Search Values
learning_rateShrinks the contribution of each tree0.001, 0.005, 0.01, 0.05, 0.1, 0.15, 0.3
n_estimatorsThe number of boosting stages to conduct.10, 25, 50, 70, 100
max_depthLimits the number of nodes in the tree.2, 3, 5, 7, 10
max_featuresThe number of features to consider when searching for the best split.‘auto’, ‘sqrt’, ‘log2’
Table 6. RF feature importance results.
Table 6. RF feature importance results.
Dry BiomassLAI
CornCanolaSoybeanCornCanolaSoybean
HH0.2350.0320.0240.1910.0130.006
HV0.0060.0090.0050.0380.0050.004
VV0.0040.0090.0050.0020.3550.004
ALPHA0.0200.0880.0370.0380.2300.012
A0.0110.0520.0150.0070.0110.005
H0.0090.0570.0080.0070.0190.025
F_ODD0.0120.0280.0070.0020.0170.010
F_DBL0.0070.0860.0050.0070.0840.005
F_VOL0.0410.0110.0040.0430.0080.007
Y_ODD0.0030.0080.0030.0020.0090.007
Y_DBL0.0020.0060.0120.0030.0150.007
Y_VO0.0150.0220.0210.0350.0130.013
Y_HIX0.0100.0190.0110.0130.0130.004
RVI0.0090.0570.0070.0040.0210.008
NDVI0.0090.0030.0920.0250.0130.083
GNDVI0.0200.0050.0430.0090.0200.009
SR0.0050.0180.0170.0580.0010.007
NDVI_EDGE0.0050.1020.2180.0510.0050.288
SR_EDGE0.0100.1030.0120.0340.0070.038
MTVI20.0070.0260.0110.0170.0260.005
EVI0.0050.0050.0110.0060.0150.007
CL_G0.0040.0050.0350.0080.0020.009
CL_EDGE0.0180.1270.2960.0960.0330.409
MCTI0.5230.0810.0850.2890.0100.024
SAVI0.0100.0400.0170.0170.0040.004
Table 7. SAR and Optical features used for different crop parameter modeling.
Table 7. SAR and Optical features used for different crop parameter modeling.
Dry BiomassLAI
CornCanolaSoybeanCornCanolaSoybean
HH
HV🗴🗴
VV
ALPHA
A
H
F_ODD🗴🗴🗴🗴
F_DBL🗴🗴🗴🗴🗴🗴
F_VOL🗴🗴🗴🗴
Y_ODD🗴🗴
Y_DBL
Y_VO🗴🗴
Y_HIX
RVI
NDVI🗴
GNDVI🗴🗴🗴🗴🗴🗴
SR
NDVI_EDGE🗴🗴🗴🗴🗴🗴
SR_EDGE🗴🗴
MTVI2🗴🗴
EVI🗴🗴🗴🗴🗴
CL_G
CL_EDGE🗴🗴
MCTI
SAVI🗴🗴🗴🗴🗴
Table 8. Results of four MLAs in estimating dry biomass.
Table 8. Results of four MLAs in estimating dry biomass.
Dry Biomass
CanolaCornSoybean
RRMSEMAERRMSEMAERRMSEMAE
RFCalibrationOptical0.9429.4921.330.9278.6849.870.926.184.25
SAR0.9233.13723.580.9294.2558.60.945.053.75
Integrated0.9724.1918.140.9663.7339.150.964.353.15
ValidationOptical0.8543.3535.560.9477.549.190.887.074.69
SAR0.8346.6733.450.9198.0460.830.96.794.87
Integrated0.9136.7430.020.9657.9741.150.945.53.7
SVRCalibrationOptical0.9331.2618.550.9472.5342.560.858.015.04
SAR0.9134.5418.610.996.8460.890.877.735.02
Integrated0.9721.3610.780.9476.5347.680.964.42.18
ValidationOptical0.8642.0332.530.9176.5551.750.857.765.2
SAR0.849.4634.460.9560.246.940.896.884.43
Integrated0.9526.2920.720.9658.6242.340.925.793.75
GBCalibrationOptical0.9328.6118.310.9471.2844.970.926.124.17
SAR0.9134.1418.250.8793.5861.20.935.43.81
Integrated0.9524.6111.840.9475.2545.230.963.562.19
ValidationOptical0.8838.8728.340.9469.8544.370.798.646.13
SAR0.8147.8632.190.9296.463.950.867.224.87
Integrated0.9330.7724.30.9663.0240.280.9453.5
XGBCalibrationOptical0.9130.2519.560.9472.3845.760.945.223.46
SAR0.9134.8418.490.8991.1561.680.935.63.97
Integrated0.9524.68712.150.9472.0245.010.973.672.41
ValidationOptical0.8640.5330.380.9377.8748.030.847.414.93
SAR0.7950.2633.620.9192.1462.390.896.984.99
Integrated0.9331.7923.180.9568.7342.10.945.253.58
Table 9. Results of four MLAs in estimating LAI.
Table 9. Results of four MLAs in estimating LAI.
LAI
CanolaCornSoybean
RRMSEMAERRMSEMAERRMSEMAE
RFCalibrationOptical0.950.6020.40.920.3960.2520.890.3510.258
SAR0.890.8560.550.940.3390.2380.950.2360.18
Integrated0.960.5550.3680.960.3020.2030.970.1740.115
ValidationOptical0.920.7710.5770.90.4420.2870.830.3680.262
SAR0.711.2040.7980.940.3550.2530.90.2920.218
Integrated0.930.6990.5150.950.3210.2040.930.2390.167
SVRCalibrationOptical0.970.4320.2760.950.4140.2490.880.3450.228
SAR0.880.8910.5650.940.3580.2060.920.2950.209
Integrated0.970.3710.2410.960.3060.1790.960.1940.129
ValidationOptical0.930.6740.5620.890.4540.2830.840.3640.254
SAR0.781.2440.7970.910.3990.2340.90.2910.209
Integrated0.940.5790.4590.950.3160.2240.920.2610.183
GBCalibrationOptical0.960.4510.2920.930.4570.2860.920.2790.199
SAR0.890.830.4640.930.3610.2120.940.2360.167
Integrated0.950.5640.250.950.3290.2190.960.1960.131
ValidationOptical0.930.6450.5420.920.4240.3020.840.3840.283
SAR0.781.0820.6410.940.3710.2660.890.3220.222
Integrated0.950.5570.3990.960.2980.2190.940.2330.167
XGBCalibrationOptical0.970.4050.2730.930.380.2680.90.3180.222
SAR0.890.8120.4560.970.2510.1730.940.2680.202
Integrated0.920.5910.3430.950.310.2080.980.1330.055
ValidationOptical0.930.6460.5270.920.3990.2730.840.360.247
SAR0.781.0930.6450.950.3210.2240.90.2930.213
Integrated0.930.6350.4990.950.3150.2220.940.2330.164
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bahrami, H.; Homayouni, S.; Safari, A.; Mirzaei, S.; Mahdianpari, M.; Reisi-Gahrouei, O. Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations. Agronomy 2021, 11, 1363. https://doi.org/10.3390/agronomy11071363

AMA Style

Bahrami H, Homayouni S, Safari A, Mirzaei S, Mahdianpari M, Reisi-Gahrouei O. Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations. Agronomy. 2021; 11(7):1363. https://doi.org/10.3390/agronomy11071363

Chicago/Turabian Style

Bahrami, Hazhir, Saeid Homayouni, Abdolreza Safari, Sayeh Mirzaei, Masoud Mahdianpari, and Omid Reisi-Gahrouei. 2021. "Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations" Agronomy 11, no. 7: 1363. https://doi.org/10.3390/agronomy11071363

APA Style

Bahrami, H., Homayouni, S., Safari, A., Mirzaei, S., Mahdianpari, M., & Reisi-Gahrouei, O. (2021). Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations. Agronomy, 11(7), 1363. https://doi.org/10.3390/agronomy11071363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop