Next Article in Journal
Stand Structure and Abiotic Factors Modulate Karst Forest Biomass in Southwest China
Previous Article in Journal
Drought Hardening Contributes to the Maintenance of Proportions of Non-Embolized Xylem and Cambium Status during Consecutive Dry Treatment in Container-Grown Seedling of Japanese Cedar (Cryptomeria japonica)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling Height–Diameter Relationship for Poplar Plantations Using Combined-Optimization Multiple Hidden Layer Back Propagation Neural Network

1
Academy of Agricultural Planning and Engineering, Ministry of Agriculture and Rural Affairs, Beijing 100125, China
2
Research Institute of Forest Resource Information Techniques, Chinese Academy of Forestry, Beijing 100091, China
3
Key Laboratory of Forest Management and Growth Modeling, National Forestry and Grassland Administration, Beijing 100091, China
4
College of Resources, Sichuan Agricultural University, Chengdu 611130, China
5
Institute of Forestry, Tribhuwan University, Kritipur, Kathmandu 44600, Nepal
6
Institute of Biophysics, Chinese Academy of Sciences, Beijing 100101, China
7
Forestry Surveying and Designing Institute of Guangdong Province, Guangzhou 510520, China
*
Author to whom correspondence should be addressed.
Forests 2020, 11(4), 442; https://doi.org/10.3390/f11040442
Submission received: 8 March 2020 / Revised: 8 April 2020 / Accepted: 12 April 2020 / Published: 14 April 2020
(This article belongs to the Section Forest Ecology and Management)

Abstract

:
Relationship of total height and diameter at breast height (hereafter diameter) of the trees is generally nonlinear, and therefore has complex characteristics, which can be accurately described by the height-diameter model developed using the back propagation (BP) neural network approach. The multiple hidden layered-BP neural network has several hidden layers and neurons, and is therefore considered more appropriate modeling approach compared to the single hidden layered-BP neural network approach. However, the former approach is not widely applied for tree height prediction due to absence of the effective optimization method, but it can be done using the BP neural network modeling approach. The poplar (Populus spp. L.) plantation data acquired from Guangdong province of China were used for evaluating the BP neural network modeling approach and compared its results with those obtained from the traditional regression modeling and mixed-effects modeling approaches. We determined the best BP neural network structure with two hidden layers and five neurons in each layer, and logistic sigmoid transfer functions. Relative to the Mitscherlich height-diameter model that had the highest fitting precision among the six traditional height-diameter models evaluated, coefficient of determination (R2) of the neural network height-diameter model increased by 10.3%, root mean squares error (RMSE) and mean absolute error (MAE) decreased by 12% and 13.5%, respectively. The BP neural network height-diameter model also appeared more accurate than the mixed-effects height-diameter model. Our study proposes the method of determining the optimal numbers of hidden layers, neurons of each layer, and transfer functions in the BP neural network structure. This method can be useful for other modeling studies of similar or different types, such as tree crown modeling, height, and diameter increments modeling, and so on.

1. Introduction

Tree height is one of the most important tree characteristics and measurement of which is used as a fundamental basis for evaluating forest growth and biomass, site quality, and classifying the vertical structures of a forest [1,2]. Direct measurement of tree height is generally difficult and time consuming. However, due to a strong relationship between tree height and diameter at breast height (DBH), height can be predicted using DBH as a predictor in the height-diameter model [3,4,5,6,7,8,9,10,11,12]. This method uses the measurements of tree height and DBH to fit the mathematical functions with different forms and number of parameters, and the optimal one is determined based on the standard statistical indices. This modeling approach is generally known as traditional modeling, and its main theme is to establish the mathematical equations and get the tree prediction by solving them [3,4,5,6,7,8,9,10,11,12]. However, tree height growth is substantially affected by various factors whose relationships may be in the nonlinear forms, which may pose the difficulty in describing wider variations of the tree height with a single height-diameter equation.
As mentioned above, generally, tree height growth has nonlinear characteristics, and is strongly correlated to various factors, such as tree size, site quality, stand density or competition and climate factors. Site factors consist of slope, altitude, soil depth, soil texture, humus layer, and soil chemical constituents. Competition is attributed to the stand crowding and density, such as number of trees, stand basal area, and canopy density. Climate factors include solar radiation, temperature, and precipitation. Because of the difficulty in acquiring the accurate information for all these factors and easy-to-apply-purpose, height prediction models are usually developed using DBH as a single predictor (simple model) or stand variables, such as basal area and number of trees per hectare are used in the models in addition to DBH (generalized model) or model incorporating DBH, stand variables and random effects (generalized mixed-effects model). In order to develop these simple or complex types of the height-diameter models, some versatile growth functions [1,9,10,11,12] are used and fitting of these functions to data using the least square regression is generally known as traditional modeling approach. However, in recent years, there has been an increasing trend of applying the mixed-effects modeling approach to account for larger variability of tree height at the subject-level (e.g., sample plot level) and increase the model’s prediction accuracy [4,6,7].
The back propagation (BP) neural network is one of the machine learning methods, which is a multiple layer feed-forward network trained by an error inverse propagation algorithm. The BP neural network is a modern modeling approach and can be used to develop various forest models. The BP neural network is often composed of the input layer, hidden layer and output layer. The BP neural networks can realize the mapping function from input to output, and can approximate any nonlinear continuous function with high precisions. The BP neural network is characterized with transfer functions that can be selected between the input layer and hidden layers, and hidden layers have different functions, such as logistic sigmoid and tangent sigmoid functions. The transfer functions selected between the layers are also different, and therefore their expressions are different. Thus, the neural network properly represents the various forms of nonlinear effects. In recent years, the neural network is increasingly applied for predicting forest dynamics with precise results. The neural network modeling approach was used to predict different stand and individual tree characteristics, such as height [13], diameter distribution [14,15] and stem volume [16], and to establish the models of height-diameter relationships [17], growth and yield [18], inside-bark diameter and heartwood relationships [19], and to assess forest biomass [20]. These studies compared the fitting precisions of the traditional regression models of different forms with the neural network models and showed higher precisions of the neural network models. Özçelik et al. [17] and Castaño-Santamaría et al. [13] compared the mixed-effects models with neural network and their results showed higher precisions of the mixed-effects models than those of the neural network models. All these studies were based on a single hidden layer neural network, and therefore their analyses lack sufficient performance analyses and comparisons of the multiple hidden layered-neural networks. Furthermore, none of the previously applied neural network modeling approach has proposed the methods for determining the optimal structure of neural network (determining optimal number of hidden layers and number of neurons in each hidden layer of neural network).
Since tree height-diameter relationship is substantially affected by various factors that may be nonlinearly related, the traditional height-diameter equation cannot accurately simulate growth and development of the tree height. This method has other shortcomings, such as low fitting precision and complex operational steps associated with the fitting procedures. However, the BP neural network modeling approach has both the higher fitting efficiency and higher precision, and therefore has a great usefulness in the forest modeling researches. However, current application of the neural network in the tree height-diameter modeling is limited to a single hidden layer neural network, which lacks the sufficient performance analyses and comparison of the differences in optimizing the neural network structure. This study thus intends to solve this problem, mainly improving the performance of the neural network structure through height-diameter modeling.
Using the poplar tree height and DBH data collected from Guangdong Province in China, this study establishes the multiple hidden layered-BP neural network height-diameter model and analyzes the difference of a single hidden layered- and multiple hidden layered-BP neural network modeling approaches using the MatLab 2016b software (MathWorks, Natick, MA, USA). This study also compares the fitting precision of the optimal BP neural network height-diameter model with that of the traditional regression height-diameter models and mixed-effects height-diameter models. The presented result will be important basis for developing height-diameter models using the BP neural network and predicting tree height. The proposed methods can be useful for other modeling studies of similar or different types, such as tree crown modeling, height and diameter increment modeling, and so on.

2. Materials and Methods

2.1. Data Materials

The data we used came from the sample plots that were established on the poplar plantations in the Guangdong province of China to develop the height-diameter models. The square-shaped sample plots with an area of 666.67 m2 were established in the plantations. We only used the sample plots with a stand density of more than 300 trees per hectare and normal records in the tree height. A total of 9659 trees in 112 sample plots (of which 20 sample plots were measured in 1997 and 92 sample plots in 2002) were utilized for modeling height-diameter relationship. We calculated the means of height (hereafter height) and means of DBH (hereafter DBH) by sample plots for easy-to-fit purpose, especially for neural network fitting. We divided the sample plots randomly into two parts: one for training the model (80 sample plots, also defined as a fitting data set) and another for testing the model (32 sample plots, also defined as a validation data set) by application of the k-fold cross-validation method. Summary statistics of both fitting and validation datasets are presented in Table 1 and scattered graph of tree height against DBH is presented in Appendix A (Figure A1).

2.2. Modelling Approach

We developed height-diameter models using three different modeling approaches: traditional least squares regression, mixed-effects modeling, and artificial neural network approach. We focused more on modeling height-diameter relationship using the last approach, for example, BP neural network. The optimal BP neural network height-diameter model obtained from several alternative models was compared against the height-diameter models fitted using the traditional regression and mixed-effects modeling approaches.

2.2.1. Traditional Approach

This involves fitting of the traditional height-diameter functions using ordinary least square regression implemented by the nls function in R software (version 3.2.2) based on fitting data set [21]. We considered six commonly used versatile height-diameter equations (Table 2) for the purpose. Since all these are the power exponential equations, they are more complex in fitting compared to other forms of the equations (e.g., linear and fractional forms) and choosing the best performing one would be more difficult also.

2.2.2. Mixed-Effects Modeling Approach

We considered the sample plot-level effect as a random effect to establish the mixed-effects height-diameter model. In order to get the convergence with the global minimum, a relatively less complex function (Schumacher function, Table 2) was chosen to include the random effect. Our main intention of developing mixed-effects height-diameter model was to compare its performance against the model obtained from the BP neural network modeling approach.
We evaluated three different variance-stabilizing functions (exponential function, power function and power function with constant) for their effectiveness in removing the heteroskedasticity problem. Akaike’s information criterion (AIC), Bayesian information criterion (BIC), and log likelihood (logLik) criteria were used to select the most effective variance-stabilizing function.
The parameters in the developed mixed-effects height-diameter model were estimated by maximum likelihood using the Lindstrom and Bates (LB) algorithm implemented in the R software (version 3.2.2) nlme function based on fitting dataset [27]. Detailed descriptions of the mixed-effects modeling are presented in the references [28,29,30,31].

2.2.3. BP Neural Network

As pointed out in the introduction section, artificial neural network has been increasingly applied to forest growth and yield modeling in recent years [13,14,15,16,17,18,19,20]. It has the tremendous advantages on nonlinear mapping, adaptive generalization and fault tolerance, which can make up of the shortcomings of traditional modeling approaches. However, most of the existing modeling studies are based on the single hidden layer neural network, and application of the multiple hidden- layered-neural network, e.g., BP neural network in forestry has been rarely reported. The main reason for this is due to the absence of optimization method used to predict the hidden layer numbers, the number of nodes and the transfer functions of the neural network. The neural network modeling studies have shown that more the complex problem, the higher would be usefulness of the multiple hidden layers [32]. The BP neural network is suitable for function approximation, pattern recognition and classification [33]. Considering the above-mentioned advantages of the BP neural network, we developed the neural network height-diameter models in this study, which were expected to be more accurate than those obtained from traditional regression and mixed-effects modeling approaches. The structural parameters of the BP neural network include the number of hidden layers, number of nodes in each layer, and transfer functions between the layers [34].
  • Setting up of the BP Neural Network Structure
We established the tree height-diameter model based on the multiple hidden-layered BP neural network. For this, firstly, we set the range and step size of the hidden layers, number of nodes and transfer functions, and secondly, the values in a reasonable range were obtained, so as to generate a series of neural network height-diameter models. Finally, we used the k-fold cross-validation to identify the best performing one among several height-diameter alternatives.
The number of nodes in each hidden layer needs to be determined according to fitting precision. Equation (1) represents a commonly used determination method [35].
S = n + o + m
where S is the number of hidden layer nodes, n is the number of nodes in the input layer, and o is the number of nodes in the output layer, and m is an integer (m = 1,2,…..,10).
The transfer functions, which occur between the hidden layers or between the input layer and hidden layer, are S-shaped logistic sigmoid and tangent sigmoid functions. The former is a unipolar S-function and the latter is a bipolar S-function. The expressions of logistic sigmoid function and tangent sigmoid function are represented by Equation (2) and Equation (3), respectively. The transfer function occurring between the hidden layer and output layer is a linear function, and its expression is represented by Equation (4):
f ( x ) = 1 / ( 1 + e x )
f ( x ) = 2 / ( 1 + e 2 x ) 1
f ( x ) = a x + b
In order to get the best BP neural network structure for a given data, we set the selection range of the number of hidden layers, the number of nodes in each layer and the transfer function, and then generated several height-diameter neural network models. Based on the mean squared error and the number of iterations, the best performing model was identified. The exhaustive analyses of all the network models would not be appropriate. Thus, we applied the “trial and error approach” [36] to reduce the number of tests and applied the k-fold cross-validation [37] to improve the test results of the BP neural network structure.
  • Normalizing Input and Output Factors
As the input factors may have different measurement units, the existence of singular samples in data would cause an increased network training time and it may also lead to the non-convergence of neural network. In response to this problem, we used the mapminmax function to normalize the input and output factors and mapped them to a scale between −1 and 1. The anti-normalization approach was used to program the results of the operation from the interval [−1,1] mapped to an actual prediction.
  • Training the Model
We trained the BP neural network applying the Levenberg-Marquatdt (L-M) algorithm [38,39]. This algorithm does not follow a single negative gradient direction for each iteration, but allows the errors to be searched in the direction of deteriorating. At the same time, through the adaptively adjustment of the steepest gradient descent method and the Gaussian–Newton method optimizes the network weight, so that the neural network can effectively converge. Equation (5) is used for adjusting the weights and thresholds.
Δ w = ( J T J + μ I ) 1 J T e
where Δw is the adjusted weights and thresholds, I is unit matrix, J is the Jacobian matrix of the error-weight differential, and e is the vector of errors, μ is an adaptively adjusted scalar that increases as it approaches the steepest descent method with small learning rate, and when it descends to 0, the algorithm becomes a smooth harmonic between the Gauss-Newton methods.
While training the model, the parameters were set as follows: learning rate 0.01, maximum number of iterations 1000, target precision 0.001, maximum number of verification failures 20, and minimum performance gradient 0.000001.
  • Model Evaluation
We used the coefficient of determination (R2), root mean square error (RMSE) and mean absolute error (MAE) as evaluation indices to compare the models based on validation data set. These indices were calculated using formulae (6), (7), and (8), respectively.
R 2 = 1 - i = 1 n ( Y i Y ^ i ) 2 ( Y i Y ¯ ) 2
R M S E = 1 n i = 1 n ( Y i Y ^ i ) 2
M A E = i = 1 n | Y i Y ^ i | n
where n is the number of samples; Y ¯ , Y i , and Y ^ are mean value, measured value and predicted value of a response variable in the model (tree height, in our case), respectively.
Theoretically, the closer the determination coefficient to 1, the smaller the root mean square error and the mean absolute error, and the higher would be the model’s fitting precision.
  • Model Selection
We first obtained a series of the BP neural network height-diameter models by setting different values of the structural parameters. Then after, the k-fold cross-validation method was used to select the most suitable model [37]. When there is a sample set S containing m data records, and t models to be chosen are M1, M2, …, Mt, k-fold (k = 5) cross-validation procedures would be as follow:
Step 1.
A sample set S is randomly divided into k disjoint subsets, the number of samples in each subset is m/k, and these subsets are denoted by S1, S2, ……, Sk.
Step 2.
For each model M j ( j = 1 , 2 , , t ) , following is done:
For n = 1 to k
{
Take S 1 S n 1 S n + 1 S k as a training set;
Train the model M j , and get the corresponding hypothetical function H j n ;
Take S n as a verification set, and calculate the model M j generalization error ε S n H j n .
}
Calculate the average of ε S ( H j n ) , n = 1 , 2 , , k , and get the average generalization error of model M j .
Step 3.
Calculate the average generalization error of all the models, and select the model M p with the smallest average generalization error, which is the best model.
It is noted that, as Arlot and Lerasle [40] recommended, five-fold (k = 5) cross-validation was applied in this study.
In general, the mean square errors are used to represent the generalized errors ε S n H j n , as shown in Equation (9).
M S E = i = 1 n ( Y i Y ^ i ) 2 / n
where, n is the number of samples, Y i , Y ^ i , respectively, for all observed height values and model predicted height values.
In addition to the generalized error ε S n H j n , the number of iterations, running time, and other criteria are also used to select the best performing model.
Tree height-diameter modeling process using the BP neural networks is shown in Figure 1. According to data situation and actual demand, we first set the implied layer number, the number of hidden layer and the number of nodes, range of values of the transfer function. Then, after we used the “trial and error approach” to determine the actual value of these structural parameters and generate the N number of height-diameter models. Finally, the optimum BP neural network height-diameter model was selected through the k-fold cross-validation.
The computations including the developed BP neural network height-diameter models and 5-fold cross-validation were implemented in the MatLab 2016b software (MathWorks, Natick, MA, USA).

3. Results

3.1. Model Generation and Performance

We used MatLab 2016b (MathWorks, Natick, MA, USA) to build the neural network height-diameter models through writing M program. Several neural network height-diameter model alternatives were generated when we set the number of hidden layers from 1 to 3, the number of neurons in each hidden layer from 2 to 11 with the step size of 3, the neurons adapted through “trial and error approach”, and logistic sigmoid and tangent sigmoid transfer functions employed with DBH and height as input layer and output layer. The model with the smallest RMSE and MAE and the largest R2 was then identified using the k-fold cross-validation with k = 5. The iterative results are listed in Table 3 and Table 4, and more results are in Appendix A (Table A1).
The MSE and the number of iterations were used as evaluation indices in screening the models and the performance statistics of different numbers of the hidden layers corresponding to the neural networks are presented in Table 5. When the number of hidden layer was 1, there were 8-networks structure combinations. When the number of hidden layers was 2, there were 64 network structure combinations. When the number of hidden layers was 3, there were 512 network structure combinations. Difference of the average MSE of the double hidden layers from that of the triple hidden layers was not substantially large even though the neural network with the double hidden layers, the structure of which is 1:5:5:1, had the smallest MSE. There was an indication that this structure, which provided the best precision, could be used as an optimal neural network structure of the height-diameter model.
There was a slight difference between the double-hidden layer and single hidden layer in the number of iterations, but it was much lower than the three-hidden layer, which showed that the convergence rate of the double-hidden layer and single hidden layer were almost similar but slightly higher than the three-hidden layers. Taken together, double hidden layer had a higher precision. We selected the double hidden layers (Table 4). When the number of neurons in each layer was 1:5:5:1, and the logistic sigmoid functions were all selected, the minimum value of MSE was 0.0416. The neural network height-diameter models generated with structure had the best fitting performance. The number of iterations required to generate the best neural network model was not necessarily the minimum number of iterations corresponding to the hidden layer (Table 5). The reason for this is that the number of iterations required to obtain the minimum increase of MSE.

3.2. Transfer Function

We generated several neural network models (Figure 2) and selected the best performing one. This figure shows that the number of neurons of both the first hidden layer and second hidden layer were 5; the first hidden layer neurons were H11, H12, H13, H14, H15, the second hidden layer neurons were H21, H22, H23, H24, H25. Different layers would have different biases, namely Bias1, Bias2, Bias3, respectively. The transfer functions of the input layer and the first hidden layer was the logistic sigmoid function, the transfer function of the first hidden layer and the second hidden layer was also the logistic sigmoid function, and the output layer was the purelin function.

3.3. Comparison with Traditional Model Accuracy

We compared the height-diameter models (Table 2) fitted using ordinary least squares regression and BP neural network height-diameter model based on three evaluation indices (Table 6). Parameter estimates of all the traditional height-diameter models were significant (p < 0.05) and they are presented in the Appendix A (Table A2).
The BP neural network modeling approach appeared better than the traditional regression approach for establishing the tree height-diameter models. Among the traditional models evaluated (Table 2), the highest fitting precision was found with the Mitscherlich model, and the lowest fitting precision was with the Logistic model. The BP neural network height-diameter model had larger R2 (by 10.3%), and smaller RMSE (by 12%) and MAE (by 13.51%) than those of the Mitscherlich height-diameter model. The tree heights predicted from the best BP neural network height-diameter model and the Mitscherlich height-diameter model were compared against the observed height (Figure 3). The prediction accuracy of the former model shows a substantially higher accuracy than the latter model.
The residuals for the BP neural network model are concentrated around 0 (Figure 4), indicating that this model has better fitting performance. Using the two intervals of (−1,1) and (−2,2) in which our data points scattered mostly as examples, Figure 4 shows that using the BP neural networks height-diameter model, 20 points of the residual are within the range of (−1,1), accounting for 62.5%, and range of (−2,2) has 30 points, accounting for 93.75%. However, using the Mitscherlich height-diameter model, 15 points are within the range of (–1,1), accounting for 50%, and range of (−2,2) has 28 points, accounting for 87.5%. Thus, the precision of the BP neural network model appeared higher than the Mitscherlich model.

3.4. Comparison with Mixed-Effects Model Accuracy

We included random effects at sample plot-level into the simplest traditional height-diameter model (Schumacher model) among the six models presented in Table 2. The random effect added to the parameter b of this model converged with the smallest AIC and BIC, and the largest log likelihood among the alternative mixed-effects models formulated through addition of the random effect to the fixed-effect parameters. Exponential function showed the most powerful ability to account for the heteroskedasticity and thus was used to develop the final mixed-effects height-diameter model (Table 7).
The mixed-effects height-diameter models with estimated parameters are presented in Equation (10).
H i = 1.3 + 17.8522 exp ( ( 10.5868 + u i ) / D B H i ) + ε i u i N ( 0 , ψ = 0.3966 ) ε i N ( 0 , R i = 8 . 4698 exp ( 0.1422 D B H i ) )
where H i and D B H i are the stand mean height and stand mean diameter at breast height of the i th sample plot, u i is the random effects generated by the i th sample plot and assumed to be distributed normally with zero expectation and a variance-covariance matrix ψ , and ε i is the error term of i th sample plot and also assumed to be distributed normally with zero expectation and a variance-covariance matrix R i .
The fitting precision of the mixed effects height-diameter model (R2 = 0.7179, RMSE = 1.1926, and MAE = 0.9888) was substantially lower than that of the neural network height-diameter model, but higher than that of the traditional height-diameter model (Equation 10, Table 6). We considered that when a mixed effects model was used, it would be equivalent to adding more input factors. Then, it was necessary to change the structure of the neural networks and added random effect factor as the input to the neural network for comparison purpose, that is, it would be more meaningful to compare the same number of input factors.

4. Discussion

We established the modeling method that could generate several BP neural network height-diameter models based on the combinatorics mathematics. Among these models, we selected the best model through the comparison of the fitting and prediction accuracies, and convergence rates. The BP neural network optimization method was employed to establish the optimal height-diameter model for poplar plantations in the Guangdong Province in China. The poplar tree species has been becoming one of the main plantation tree species in China in recent decades due to its faster growing characteristics, and is recognized as a focus for research of the woody plants and ideal materials for bioenergy research. Also, it is of great importance for taking poplar as a research objective in our study. This study compares the fitting performances of a single hidden layered- and multiple hidden layered-BP neural network approaches. When there were two hidden layers, the higher performance was obtained, i.e., the neural network with double hidden layers is attributed to a higher fitting precision, higher estimating efficiency, better acceptable time of the iterations. This is because that, increasing the number of hidden layers may not only result in a longer computational time, but also increases the likelihood of over fitting, which results in the model’s non-optimal prediction performance. In the past, most studies focused on single hidden layer neural networks [13,14,15,16,17,18,19,20], but none of them investigated the effects due to more numbers of hidden layers, neurons and transfer functions of the neural networks. In this context, our study, which focused on these features of the neural network modeling, may be interesting and useful to other researchers.
The traditional Mitscherlich height-diameter model with the highest fitting precision was compared with the optimized multiple hidden layered-BP neural network height-diameter model. The BP neural network model appears substantially superior to the Mitscherlich model in predicting tree height (Table 6, Figure 3 and Figure 4). Our results are also consistent with those from Castaño-Santamaría et al. [13] and Özçelik et al. [17], which predict tree height of the uneven-aged beech forests in northwestern Spain and Crimean juniper in southwestern region of Turkey. These studies compared the neural network models against the nonlinear regression models. Although our study is based on the different tree species from those studied by Castaño-Santamaría et al. [13] and Özçelik et al. [17], comparison results are almost identical, meaning that neural networks can be the best alternative of modeling on any tree data regardless of species. Özçelik et al. [17] used the single hidden layer with only one or two hidden nodes in the neural network and they did not investigate the effects of multiple hidden layers on the precision of the neural network model and determination of appropriate forms of the transfer functions. Our study is substantially different from the previous studies [13,14,15,16,17,18,19,20], because we proposed the method of selecting the optimal model through application of the “trial and error approach” [30], k-fold cross-validation approach [31] and combinatorial optimization approach. It can help determining the structure of the neural network, such as the hidden layer nodes, transfer functions and the number of hidden layers. The hidden layers, number of neurons in each layer, and transfer functions can have substantial effects on the precision of the neural network model. Castaño-Santamaría [13] considered the change of input factors, but did not take into account other factors, such as different transfer functions of the neural network, and did not introduce the process of determining the optimal neural network model. Castro et al. [18] established the multi-layer perceptron neural network growth model for Eucalyptus. They estimated the annual mortality with the best structure associated with three neurons in the input layer, four neurons in the hidden layer, and one neuron in the output layer. All these studies [13,17,18] compared the precisions using different input variables, but none of them compared different numbers of transfer functions and hidden layers, and neurons of neural networks.
In our study, the neural network modeling produced the highest fitting precision and prediction accuracy, followed by mixed-effects models, and finally non-linear traditional regression models (Table 6, Equation (10)). This might be related to our employed methods of the structural optimization of the BP neural network. Fitting precisions of three different modeling approaches, such as traditional regression, mixed-effects modeling and neural network modeling were also compared in the previous modeling studies [13,17]. These studies showed the highest precisions of mixed-effects models, followed by neural network models, and traditional regression models. This may be due to the weaknesses of their modeling techniques, whereby they did not investigate the effects of multiple hidden layers on the precisions of the neural network models they developed. The results obtained from different modeling approaches including ours thus indicate the inconsistent ranking of approaches on the basis their fitting and prediction accuracies. Further investigation on the models, especially those to be developed with mixed-effects modeling and neural network modeling approaches using large datasets collected from extensive forest areas is necessary to better explore their differences. Generally, the neural network models have strong robustness, but traditional regression models and mixed-effects models have biological significance, for example, they have the parameters describing growth rates and growth patterns. The neural network modeling approach has tremendous advantages, such as avoiding complex selection procedures for the best performing model and obtaining higher precision. Furthermore, the BP neural network modeling has a better generalization ability, and therefore can approximate any nonlinear continuous function with a high precision. The BP neural network modeling approach is suitable for describing plant growth, which generally follows nonlinear patterns, and making it suitable for solving the problems caused by interrelated factors affecting plant growth.
The process of determining the selection of the traditional height-diameter models and neural network height-diameter models was also evaluated in this study. The traditional regression modeling needs the evaluation and comparison of differences among the fitting precisions of the candidate models considered, and the fitted model with the highest precision could be selected as the final model. The neural network modeling, on the other hand, needs the determination of the number of hidden layers and the number of neurons in each layer, and the numbers and forms of the transfer functions, and this modeling approach determines the best structure of the BP neural network. There are none of the well-organized robust methods, which can determine the number of hidden layers and the number of transfer functions, and the number of neurons based on the combinatorial mathematics that could help obtain the high precisions. We applied this method, and thus the height-diameter model proposed in this article is based on the multiple hidden layered-BP neural network. This method can help modelers to find the best neural network structure, and thus provide the best performance. However, neural network weights and thresholds are not easy to explain and determining the optimal structure of neural network is more tedious, and this can be done into a friendly interface program module, which can help future modelers to quickly determine the best structure of the BP neural network.
Our modeling system introduces the method of determining the best neural network structure with optimal numbers of hidden layers and neurons in each layer, and optimal number and forms of transfer functions. The model comparison would make the senses when traditional regression models, mixed-effects models, and neural network models have the same input factor, such as DBH in our case. Because DBH is the main factor influencing tree height, in this study, only the effect of DBH on the tree height was considered for both modeling approaches. In our subsequent studies, in addition to DBH, the effects of other factors on the tree height growth may be considered and models can be made more comprehensive and complex. However, introducing many variables does not guarantee the high accuracies of the models that are developed using any modeling approach. In the neural network modeling, all potential interconnected factors that affect the response variable (tree height, in our case) of the model are assumed to be properly described, and thus this approach can be considered more appropriate than other modeling approaches, such as ordinary least square regression and mixed-effects modeling approaches. This is because
The neural network is able to optimize the model efficiently through the combinatorial optimization process.

5. Conclusions

We proposed the modeling method which can generate several BP neural network models based on the combinatorics mathematics, and among them, the model with the best structure was selected through comparison of the fitting and prediction precisions, and convergence rate. In the process of determining the structure of the neural network, both the number of hidden layers, the numbers of neurons and the number of transfer functions of the BP neural network were considered. We developed the BP neural network optimization method to establish the optimal tree height-diameter model for poplar plantations in the Guangdong Province of China. The optimal BP neural network structure was 1:5:5:1 and transfer functions determined were the logistic sigmoid functions. The optimal structure of the BP neural network height-diameter accounted for 75% variations of tree height-diameter relationship, which is higher (by 10.3%) than the best fitted traditional regression height-diameter model. The BP neural network height-diameter model also outperformed the mixed-effects height-diameter model. In addition to diameter at breast height, tree height growth is also substantially affected by other several factors, such as site and climate factors, and stand conditions, which may be introduced in the height-diameter model for gaining a higher prediction accuracy. The proposed method of the neural network modeling can be suitable for other forest modeling studies of similar or different types, such as tree crown modeling, height and diameter increments modeling, and so on.

Author Contributions

J.S., G.W. and R.P.S. conceived the idea and designed methodology; J.S., L.F., X.M. and M.W. collected data; J.S., Z.H., G.W., L.F. and Q.W. analyzed data; J.S., L.F., R.P.S. and G.W. wrote manuscript and contributed critically to improve the manuscript, and gave a final approval for publication. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to thank the Forestry Public Welfare Scientific Research Project of China (Grant No. 201504303), the Central Public-interest Scientific Institution Basal Research Fund (Grant No. CAFYBB2019QD003) and the National Natural Science Foundations of China (Nos. 31570628 and 31570627) for the financial support of this study.

Acknowledgments

We are grateful to four anonymous reviewers, who provided the constructive comments and suggestions for improvement of the article.

Data Accessibility

Data used in this study are available from the Chinese Forestry Science Data Center (http://www.cfsdc.org/, accessed on 12 April 2020).

Conflicts of Interest

The authors declare no competing financial interests.

Appendix A

Table A1. Performance of neural network with three hidden layers.
Table A1. Performance of neural network with three hidden layers.
Neurons of Each LayerMSEIterationsMSEIterationsMSEIterationsMSEIterations
tan:tan:tantan:tan:logtan:log:tantan:log:log
1:2:2:2:10.088426.20.055631.00.094033.60.101423.2
1:2:2:5:10.102018.20.077226.80.050519.40.057221.0
1:2:2:8:10.046124.00.079514.60.087817.80.071015.8
1:2:2:11:10.388023.80.056317.80.094521.60.070818.8
1:2:5:2:10.067445.00.086322.20.080017.20.103916.6
1:2:5:5:10.102814.60.122813.60.175925.40.080116.2
1:2:5:8:10.119424.20.080318.60.051815.40.117917.6
1:2:5:11:10.076118.40.091837.40.121320.40.078513.4
1:2:8:2:10.077624.80.119518.60.051121.80.058531.6
1:2:8:5:10.106616.40.093616.80.104917.20.107121.8
1:2:8:8:10.056019.20.098613.60.054515.80.054814.8
1:2:8:11:10.078843.20.067319.00.089915.80.096516.4
1:2:11:2:10.135119.40.103715.20.080325.60.084319.2
1:2:11:5:10.077215.00.057816.20.062119.00.071120.4
1:2:11:8:10.140340.20.103714.80.141816.00.116722.4
1:2:11:11:10.118025.40.083322.20.120025.40.100715.4
1:5:2:2:10.066018.40.132320.00.044623.20.071115.0
1:5:2:5:10.062416.40.076326.80.119225.00.067918.2
1:5:2:8:10.106527.80.105621.20.358117.80.073013.6
1:5:2:11:10.089219.20.081319.60.079622.00.118315.2
1:5:5:2:10.063020.20.098922.40.073316.40.110618.6
1:5:5:5:10.062514.60.086422.00.098117.00.067814.0
1:5:5:8:10.108719.40.102013.60.050113.80.052814.6
1:5:5:11:10.053720.20.059515.20.063414.20.105919.0
1:5:8:2:10.071613.60.080218.20.075220.60.076216.6
1:5:8:5:10.140913.40.100616.80.067618.20.062613.4
1:5:8:8:10.110118.80.125813.20.065215.60.098014.4
1:5:8:11:10.059016.80.103513.00.177516.60.102814.4
1:5:11:2:10.066517.00.081517.80.050916.00.097522.4
1:5:11:5:10.138714.20.068520.40.079614.40.130215.4
1:5:11:8:10.114512.80.087315.60.091917.80.096512.4
1:5:11:11:10.145115.20.075622.80.085612.80.072620.0
1:8:2:2:10.070915.80.078016.40.068915.00.066828.2
1:8:2:5:10.081434.80.083119.60.069816.40.068715.0
1:8:2:8:10.117819.80.075220.60.059316.80.060326.0
1:8:2:11:10.081417.60.083716.80.058019.40.079816.0
1:8:5:2:10.081115.60.108034.00.153018.40.095817.8
1:8:5:5:10.091615.40.089218.00.065313.60.119213.8
1:8:5:8:10.111013.80.095114.80.081413.60.112313.0
1:8:5:11:10.089915.80.079314.00.107612.60.067915.0
1:8:8:2:10.093915.60.101118.40.085914.80.139122.8
1:8:8:5:10.077218.00.104916.40.135314.60.190715.8
1:8:8:8:10.185914.00.042816.60.079313.20.115721.0
1:8:8:11:10.081216.20.183819.20.160714.20.092219.0
1:8:11:2:10.095722.80.065413.00.057821.60.109616.4
1:8:11:5:10.054413.00.071118.20.057916.20.067213.8
1:8:11:8:10.093712.40.127420.20.146817.00.065213.6
1:8:11:11:10.119323.20.154716.20.065215.40.074712.6
1:11:2:2:10.076716.60.073420.40.095317.20.070417.2
1:11:2:5:10.156413.80.076316.41.644521.20.064018.2
1:11:2:8:10.100917.40.046915.20.136912.40.147422.0
1:11:2:11:10.239242.40.120618.80.091918.60.065024.2
1:11:5:2:10.078026.40.089616.60.071116.60.108314.2
1:11:5:5:10.155113.80.145414.80.084613.80.095817.6
1:11:5:8:10.081415.60.060718.00.162015.80.081316.6
1:11:5:11:10.098715.00.075218.20.106913.40.101212.6
1:11:8:2:10.090116.00.059815.60.106914.60.091723.4
1:11:8:5:10.083817.00.109712.40.057015.80.095615.2
1:11:8:8:10.104114.40.109415.40.147415.20.114716.2
1:11:8:11:10.110521.80.122018.60.150815.00.050414.2
1:11:11:2:10.075518.40.056024.20.135012.00.085415.0
1:11:11:5:10.078518.60.087215.60.085411.80.164118.8
1:11:11:8:10.112916.00.088322.00.094613.20.144412.4
1:11:11:11:10.100414.60.125213.60.153813.20.089321.0
1:2:2:2:10.098520.60.085223.40.084042.40.121817.6
1:2:2:5:10.086166.20.106322.20.123018.40.085318.0
1:2:2:8:10.115719.20.068420.00.081240.20.108915.0
1:2:2:11:10.354026.60.073027.40.061322.00.094115.4
1:2:5:2:10.106629.60.093117.80.106217.40.327827.6
1:2:5:5:10.115117.20.090517.40.204316.80.069014.4
1:2:5:8:10.074918.20.085923.80.068617.00.078214.4
1:2:5:11:10.057815.00.046917.40.196618.00.124821.0
1:2:8:2:10.117621.60.045819.20.082719.20.069121.0
1:2:8:5:10.088517.60.071917.20.153916.80.082120.6
1:2:8:8:10.082425.00.118214.60.080421.40.068118.0
1:2:8:11:10.079719.60.080717.00.058013.20.098317.2
1:2:11:2:10.102415.60.107418.20.072732.40.082917.2
1:2:11:5:10.096115.40.069616.60.078515.00.081923.0
1:2:11:8:10.092219.60.072319.00.092115.60.079730.0
1:2:11:11:10.073319.80.087213.80.073415.20.101614.4
1:5:2:2:10.087413.40.086423.80.067324.00.078216.6
1:5:2:5:10.076917.80.071519.20.066815.60.117015.8
1:5:2:8:10.065114.00.071820.80.072919.80.123818.8
1:5:2:11:10.066119.00.089414.60.079314.40.132715.0
1:5:5:2:10.126015.40.168215.60.062717.80.183014.8
1:5:5:5:10.103813.40.075015.80.074818.80.099116.8
1:5:5:8:10.068413.00.052233.80.078617.60.099915.2
1:5:5:11:10.156817.40.092018.00.090018.80.116619.2
1:5:8:2:10.079323.40.094717.40.052114.00.125719.4
1:5:8:5:10.123421.60.105019.00.071619.20.133221.6
1:5:8:8:10.090314.00.079411.60.137919.00.069517.4
1:5:8:11:10.175117.40.107318.20.163021.20.089513.2
1:5:11:2:10.092718.20.106515.40.073932.40.117724.8
1:5:11:5:10.094113.20.084716.20.082121.20.070213.8
1:5:11:8:10.208420.60.070313.60.099414.00.112022.0
1:5:11:11:10.089622.00.083221.20.089514.20.085313.2
1:8:2:2:10.132214.80.124416.00.094816.00.094017.6
1:8:2:5:10.096313.00.098116.00.127821.80.071716.8
1:8:2:8:10.120415.00.096115.80.090029.20.082319.8
1:8:2:11:10.055022.40.133512.80.119713.20.069414.2
1:8:5:2:10.082417.20.080423.60.189922.60.067118.4
1:8:5:5:10.069317.60.130320.40.082015.60.086715.4
1:8:5:8:10.093014.80.101113.20.110912.60.141515.0
1:8:5:11:10.249517.40.085013.20.067615.20.160713.0
1:8:8:2:10.053815.20.082521.60.111518.00.103825.6
1:8:8:5:10.082712.40.051315.40.067814.20.151716.4
1:8:8:8:10.112017.00.099415.60.156015.00.073312.2
1:8:8:11:10.102717.40.133113.40.104014.60.076321.0
1:8:11:2:10.049915.00.064818.60.095520.00.076229.6
1:8:11:5:10.168516.40.108213.00.098215.60.078612.4
1:8:11:8:10.070013.20.140912.40.104619.00.105322.6
1:8:11:11:10.123418.20.076718.20.068616.60.078919.2
1:11:2:2:10.086419.20.058520.20.065520.40.092015.0
1:11:2:5:10.245617.20.111916.40.088123.60.103814.8
1:11:2:8:10.057019.20.085221.80.072417.80.149326.4
1:11:2:11:10.134518.20.140719.20.139214.40.083616.2
1:11:5:2:10.120518.60.065820.60.056116.80.116316.8
1:11:5:5:10.073617.20.089617.20.086717.00.120323.6
1:11:5:8:10.074615.60.132716.40.062520.00.137117.2
1:11:5:11:10.063414.60.099918.40.086215.60.183019.8
1:11:8:2:10.091819.40.112425.20.160913.80.070416.4
1:11:8:5:10.218419.40.070914.00.102514.60.135615.6
1:11:8:8:10.107817.80.087314.40.080615.40.109113.0
1:11:8:11:10.069315.40.135716.60.081214.20.116416.0
1:11:11:2:10.099215.20.091914.80.120216.40.097314.0
1:11:11:5:10.099216.80.109513.60.099017.20.065213.6
1:11:11:8:10.133519.40.079817.20.092815.80.099117.0
1:11:11:11:10.103013.80.109222.00.102514.80.114012.6
Table A2. Parameter estimates of traditional height-diameter models.
Table A2. Parameter estimates of traditional height-diameter models.
EquationsParameter Estimates
abc
Richards13.23930.11191.9216
Logistic12.31677.05940.1959
Gompertz12.64563.04140.1447
Korf16.986411.95391.0749
Mitscherlich22.21020.0331
Schumacher18.009110.6440
Figure A1. Scattered diagram of stand mean height and stand mean diameter at breast height (DBH) for both model fitting and validation datasets.
Figure A1. Scattered diagram of stand mean height and stand mean diameter at breast height (DBH) for both model fitting and validation datasets.
Forests 11 00442 g0a1

References

  1. Burkhart, H.E.; Tomé, M. Modeling Forest Trees and Stands; Springer: Noordwijkerhout, The Netherlands, 2016. [Google Scholar]
  2. Watt, M.S.; Dash, J.P.; Bhandari, S.; Watt, P. Comparing parametric and non-parametric methods of predicting Site Index for radiata pine using combinations of data derived from environmental surfaces, satellite imagery and airborne laser scanning. For. Ecol. Manag. 2015, 357, 1–9. [Google Scholar] [CrossRef]
  3. Sharma, M. Comparing Height-diameter Relationships of Boreal Tree Species Grown in Plantations and Natural Stands. For. Sci. 2016, 62, 70–77. [Google Scholar] [CrossRef]
  4. Sharma, R.P.; Breidenbach, J. Modeling height-diameter relationships for Norway spruce, Scots pine, and downy birch using Norwegian national forest inventory data. For. Sci. Technol. 2015, 11, 44–53. [Google Scholar] [CrossRef]
  5. Sánchez, C.A.L.; Varela, J.G.; Dorado, F.C.; Alboreca, A.R.; Soalleiro, R.R.; González, J.G.Á.; Rodríguez, F.S. A height-diameter model for Pinus radiata D. Don in Galicia (Northwest Spain). Ann. For. Sci. 2003, 60, 237–245. [Google Scholar]
  6. Castedo, D.F.; Diéguez-Aranda, U.; Barrio, A.M.; Rodríguez, M.S.; Von Gadow, K. A generalized height-diameter model including random components for radiata pine plantations in northwestern Spain. For. Ecol. Manag. 2006, 229, 202–213. [Google Scholar] [CrossRef]
  7. Schmidt, M.; Kiviste, A.; Gadow, K.V. A spatially explicit height-diameter model for Scots pine in Estonia. Eur. J. For. Res. 2011, 130, 303–315. [Google Scholar] [CrossRef] [Green Version]
  8. Kearsley, E.; Moonen, P.C.; Hufkens, K.; Doetterl, S.; Lisingo, J.; Bosela, F.B.; Boeckx, P.; Beeckman, H.; Verbeeck, H. Model performance of tree height-diameter relationships in the central Congo Basin. Ann. For. Sci. 2017, 74, 7. [Google Scholar] [CrossRef] [Green Version]
  9. Richards, F.J. A Flexible Growth Function for Empirical Use. J. Exp. Bot. 1959, 10, 290–301. [Google Scholar] [CrossRef]
  10. Temesgen, H.; Hann, D.W.; Monleon, V.J. Regional Height-diameter Equations for Major Tree Species of Southwest Oregon. West. J. Appl. For. 2007, 22, 213–219. [Google Scholar] [CrossRef] [Green Version]
  11. Zang, H.; Lei, X.; Zeng, W. Height-diameter equations for larch plantations in northern and northeastern China: A comparison of the mixed-effects, quantile regression and generalized additive models. Forestry 2016, 89, 434–445. [Google Scholar] [CrossRef]
  12. Koirala, A.; Kizha, A.R.; Baral, S. Modeling Height-diameter Relationship and Volume of Teak (Tectona grandis L. F.) in Central Lowlands of Nepal. J. Trop. For. Environ. 2017, 7, 28–42. [Google Scholar] [CrossRef]
  13. Castaño-Santamaría, J.; Crecente Campo, F.; Fernández-Martínez, J.L.; Barrio-Anta, M.; Obeso, J.R. Tree height prediction approaches for uneven-aged beech forests in northwestern Spain. For. Ecol. Manag. 2013, 307, 63–73. [Google Scholar] [CrossRef]
  14. Scrinzi, G.; Marzullo, L.; Galvagni, D. Development of a neural network model to update forest distribution data for managed alpine stands. Ecol. Model. 2007, 206, 331–346. [Google Scholar] [CrossRef]
  15. Leduc, D.J.; Matney, T.G.; Belli, K.L.; Baldwin, V.C. Predicting Diameter Distributions of Longleaf Pine Plantations: A Comparison between Artifical Neural Networks and Other Accepted Methodologies; USDA Forest Service, Southern Research Station: Asheville, NC, USA, 2001. [Google Scholar]
  16. Diamantopoulou, M.J.; Milios, E. Modelling total volume of dominant pine trees in reforestations via multivariate analysis and artificial neural network models. Biosyst. Eng. 2010, 105, 306–315. [Google Scholar] [CrossRef]
  17. Özçelik, R.; Diamantopoulou, M.J.; Crecente-Campo, F.; Eler, U. Estimating Crimean juniper tree height using nonlinear regression and artificial neural network models. For. Ecol. Manag. 2013, 306, 52–60. [Google Scholar] [CrossRef]
  18. Castro, R.V.O.; Soares, C.P.B.; Leite, H.G.; Lopes de Souza, A.; Saraiva Nogueira, G.; Bolzan Martins, F. Individual growth model for Eucalyptus stands in Brazil using artificial neural network. Forestry 2013, 1–12. [Google Scholar]
  19. Leite, H.G.; Silva, M.L.M.D.; Binoti, D.H.B.; Fardin, L.; Takizawa, F.H. Estimation of inside-bark diameter and heartwood diameter for Tectona grandis, Linn. trees using artificial neural networks. Eur. J. For. Res. 2011, 130, 263–269. [Google Scholar] [CrossRef]
  20. Nandy., S.; Singh, R.; Ghosh, S.; Watham, T.; Kushwaha, S.P.S.; Kumar, A.S.; Dadhwal, V.K. Neural network-based modelling for forest biomass assessment. Carbon Manag. 2017, 8, 305–317. [Google Scholar] [CrossRef]
  21. Venables, W.N.; Ripley, B.D. Modern Applied Statistics with S-PLUS, 3rd ed.; Springer: New York, NY, USA, 1999. [Google Scholar]
  22. Pearl, R.; Reed, L.J. On the rate of growth of the population of the United States since 1790 and its mathematical representation. Proc. Natl. Acad. Sci. USA 1920, 6, 275–288. [Google Scholar] [CrossRef] [Green Version]
  23. Winsor, C.P. The Gompertz Curve as a Growth Curve. Proc. Natl. Acad. Sci. USA 1932, 18, 1–8. [Google Scholar] [CrossRef] [Green Version]
  24. Lundqvist, B. On the height growth in cultivated stands of pine and spruce in Northern Sweden. Medd. Fran Statens Skogforsk. Band 1957, 47, 1–64. [Google Scholar]
  25. Levin, I.; Nitsan, J. Use of the Mitscherlich equation in designing factorial fertilizer field experiments to reduce the number of treatments. Plant Soil 1964, 21, 249–252. [Google Scholar] [CrossRef]
  26. Alder, D.; Montenegro, F. A yield model for Cordia alliodora plantations in Ecuador. Int. For. Rev. 1999, 1, 242–250. [Google Scholar]
  27. Pinheiro, J.C.; Bates, D.M. Mixed-Effects Models in S and S-PLUS; Springer: New York, NY, USA, 2000. [Google Scholar]
  28. Fu., L.; Sun, H.; Sharma, R.P.; Lei, Y.; Zhang, H.; Tang, S. Nonlinear mixed-effects crown width models for individual trees of Chinese fir (Cunninghamia lanceolata) in South-Central China. For. Ecol. Manag. 2013, 302, 210–220. [Google Scholar] [CrossRef]
  29. Fu, L.; Zhang, H.; Sharma, R.P.; Pang, L.; Wang, G. A generalized nonlinear mixed-effects height to crown base model for Mongolianoak in northeast China. For. Ecol. Manag. 2017, 384, 34–43. [Google Scholar] [CrossRef]
  30. Fu, L.; Wang, M.; Wang, Z.; Song, X.; Tang, S. Maximum likelihood estimation of nonlinear mixed-effects models with crossed random effects by combining first order conditional linearization and sequential quadratic programming. Int. J. Biomath. 2019. [Google Scholar] [CrossRef]
  31. Fu, L.; Duan, G.; Ye, Q.; Meng, X.; Luo, P.; Sharma, R.P.; Sun, H.; Wang, G.; Liu, Q. Prediction of Individual Tree Diameter Using a Nonlinear Mixed-Effects Modeling Approach and Airborne LiDAR Data. Remote Sens. 2020, 12, 1066. [Google Scholar] [CrossRef] [Green Version]
  32. Hecht-Nielsen, R. Komogorovs Mapping Neural Network Existence Theorem. In Proceedings of the International Conference on Neural Networks; MIT Press: Cambridge, MA, USA, 1987; pp. 11–13. [Google Scholar]
  33. Ding, S.; Su, C.; Yu, J. An optimizing BP neural network algorithm based on genetic algorithm. Artif. Intell. Rev. 2011, 36, 153–162. [Google Scholar] [CrossRef]
  34. Neog, D.K. Microstrip Antenna and Artificial Neural Network; LAP LAMBERT Academic Publishing: Saarbrücken, Germany, 2010. [Google Scholar]
  35. Han, W.; Nan, L.; Su, M.; Chen, Y.; Li, R.; Zhang, X. Research on the Prediction Method of Centrifugal Pump Performance Based on a Double Hidden Layer BP Neural Network. Energies 2019, 12, 2709. [Google Scholar] [CrossRef] [Green Version]
  36. Liu, C.; Ling, J.; Kou, L. Performance Comparison between GA-BP Neural Network and BP Neural Network. Chin. J. Health Stat. 2013, 30, 173–494. [Google Scholar]
  37. Zhang, Y.; Yang, Y. Cross-validation for selecting a model selection procedure. J. Econom. 2015, 187, 95–112. [Google Scholar] [CrossRef]
  38. Moré, J.J. The levenberg-marquardt algorithm, implementation and theory. In Numerical Analysis; Lecture Notes in Mathematics, 630; Watson, G.A., Ed.; Springer: Berlin, Germany, 1977. [Google Scholar]
  39. Ding, S.; Chang, X.H. Application of Improved BP Neural Networks Based on LM Algorithm in Characteristic Curve Fitting of Fiber-Optic Micro-Bend Sensor. Adv. Mater. Res. 2014, 889, 825–828. [Google Scholar]
  40. Arlot, S.; Lerasle, M. Why V=5 Is Enough in V-Fold Cross-Validation. 2014. Available online: https://hal.archives-ouvertes.fr/hal-00743931v2 (accessed on 14 April 2020).
Figure 1. Flow chart for selecting an optimal tree height-diameter model.
Figure 1. Flow chart for selecting an optimal tree height-diameter model.
Forests 11 00442 g001
Figure 2. Height-diameter model of BP neural networks.
Figure 2. Height-diameter model of BP neural networks.
Forests 11 00442 g002
Figure 3. Comparison of height prediction by two different modeling approaches: left graph is based on the back propagation (BP) neural network approach, and right is based on the traditional approach (Mitscherlich model). Fitting equation based on the observed data and predicted data, y = 1.0927x − 0.2554 for left graph, and y = 1.1433x − 1.306 for right graph.
Figure 3. Comparison of height prediction by two different modeling approaches: left graph is based on the back propagation (BP) neural network approach, and right is based on the traditional approach (Mitscherlich model). Fitting equation based on the observed data and predicted data, y = 1.0927x − 0.2554 for left graph, and y = 1.1433x − 1.306 for right graph.
Forests 11 00442 g003
Figure 4. Residual graphs: left is based on the BP neural network height-diameter model and right is based on the Mitscherlich height-diameter model.
Figure 4. Residual graphs: left is based on the BP neural network height-diameter model and right is based on the Mitscherlich height-diameter model.
Forests 11 00442 g004
Table 1. Summary of the statistics for the model fitting and validation datasets (Min, minimum; Max, maximum; Mean, average value; Std = standard deviation).
Table 1. Summary of the statistics for the model fitting and validation datasets (Min, minimum; Max, maximum; Mean, average value; Std = standard deviation).
DataVariableMinMaxMeanStd
Fitting data setDBH (cm)5.925.711.53.0
Height (m)2.513.08.31.9
Validation data setDBH (cm)7.022.010.43.4
Height (m)3.714.17.52.2
Table 2. Traditional height-diameter equations (H, height (m); DBH, diameter at breast height (cm); a, b, and c are parameters to be estimated).
Table 2. Traditional height-diameter equations (H, height (m); DBH, diameter at breast height (cm); a, b, and c are parameters to be estimated).
Name of EquationForm of EquationSource
RichardsH = 1.3 + a(1 − exp(−bDBH))c[9]
LogisticH= 1.3 + a/(1 + bexp(−cDBH))[22]
GompertzH = 1.3 + aexp(−bexp(−cDBH))[23]
KorfH = 1.3 + aexp(−bDBH(−c))[24]
MitscherlichH = 1.3 + a(1−exp(−bDBH))[25]
SchumacherH = 1.3 + aexp(−b/DBH)[26]
Table 3. Performance of neural network with single hidden layer, MSE: mean square errors, logsig: logistic sigmoid function, tansig: Tangent sigmoid function.
Table 3. Performance of neural network with single hidden layer, MSE: mean square errors, logsig: logistic sigmoid function, tansig: Tangent sigmoid function.
Neurons in Each LayerMSEIterationsMSEIterations
logsigtansig
1: 2: 10.110724.00.305716.0
1: 5: 10.082714.30.066915.0
1: 8: 10.056813.10.114317.0
1:11: 10.055611.00.043713.1
Table 4. Performance of neural network with two hidden layers, MSE: Mean square errors.
Table 4. Performance of neural network with two hidden layers, MSE: Mean square errors.
Neurons of Each LayerMSEIterationsMSEIterationsMSEIterationsMSEIterations
log:logtan:tantan:loglog:tan
1:2:2:10.123715.20.088139.60.079022.60.107116.6
1:2:5:10.076415.20.175119.40.044621.40.108723.0
1:2:8:10.074317.60.081316.60.088423.40.146718.6
1:2:11:10.083919.80.093521.60.080115.80.094820.6
1:5:2:10.062617.00.093613.00.055814.20.075715.0
1:5:5:10.041616.20.085419.40.101715.00.083222.4
1:5:8:10.089714.20.058417.20.109116.20.097112.8
1:5:11:10.076417.60.132320.60.095014.40.121412.0
1:8:2:10.061014.60.098617.00.064413.80.158214.4
1:8:5:10.097113.60.094813.40.086914.00.130418.8
1:8:8:10.059921.00.084112.20.097813.40.058613.2
1:8:11:10.106013.60.144117.40.100413.00.098313.4
1:11:2:10.075022.40.085217.80.053015.00.127216.4
1:11:5:10.157312.80.253917.60.140012.20.088718.8
1:11:8:10.105823.40.172017.00.102015.80.091712.2
1:11:11:10.140212.80.272623.40.076614.40.165113.0
Table 5. Statistics for performance of three neural networks, MSE: Mean square errors.
Table 5. Statistics for performance of three neural networks, MSE: Mean square errors.
CategoryNumber of Neural NetworkBest Neural Network Structure
Average MSEMSE (min)Ratio of MSE < 0.1Average IterationsIterations (min)Neurons of Each LayerTransfer FunctionIterations
Single hidden layer80.10450.043762.50%1511.01:11:1tansig13.0
Double hidden layer640.10270.041662.50%1612.01:5:5:1logsig:logsig16.2
Triple hidden layer5120.10100.042862.89%2111.61:8:8:8:1tansig:tansig:logsig16.6
Table 6. Evaluation indices of the BP neural network height-diameter models and traditional height-diameter model (R2, coefficient of determination; RMSE, root mean square error; MAE, mean absolute error).
Table 6. Evaluation indices of the BP neural network height-diameter models and traditional height-diameter model (R2, coefficient of determination; RMSE, root mean square error; MAE, mean absolute error).
Evaluation IndexRichardsLogisticGompertzKorfMitscherlichSchumacherNeural Network
R20.67940.67740.67710.67570.68370.67880.7541
RMSE1.27131.27521.27581.27861.26281.27241.1133
MAE1.08871.10411.09721.08581.09631.08090.9482
Table 7. Evaluation indices of three variance-stabilizing functions employed to the nonlinear mixed-effects height-diameter model; AIC, Akaike information criterion; BIC, Bayesian information criterion; loglik, log likelihood.
Table 7. Evaluation indices of three variance-stabilizing functions employed to the nonlinear mixed-effects height-diameter model; AIC, Akaike information criterion; BIC, Bayesian information criterion; loglik, log likelihood.
Variance FunctionFormulaAICBICloglik
Exponential function var ( ε ) = σ 2 exp ( β D B H ) 283.8982295.8084−136.9491
Power function var ( ε ) = σ 2 D B H β 283.9949295.9051−136.9975
Power function with constant var ( ε ) = σ 2 ( β 1 + D B H β 2 ) 2 285.9949300.2871−136.9975

Share and Cite

MDPI and ACS Style

Shen, J.; Hu, Z.; Sharma, R.P.; Wang, G.; Meng, X.; Wang, M.; Wang, Q.; Fu, L. Modeling Height–Diameter Relationship for Poplar Plantations Using Combined-Optimization Multiple Hidden Layer Back Propagation Neural Network. Forests 2020, 11, 442. https://doi.org/10.3390/f11040442

AMA Style

Shen J, Hu Z, Sharma RP, Wang G, Meng X, Wang M, Wang Q, Fu L. Modeling Height–Diameter Relationship for Poplar Plantations Using Combined-Optimization Multiple Hidden Layer Back Propagation Neural Network. Forests. 2020; 11(4):442. https://doi.org/10.3390/f11040442

Chicago/Turabian Style

Shen, Jianbo, Zongda Hu, Ram P. Sharma, Gongming Wang, Xiang Meng, Mengxi Wang, Qiulai Wang, and Liyong Fu. 2020. "Modeling Height–Diameter Relationship for Poplar Plantations Using Combined-Optimization Multiple Hidden Layer Back Propagation Neural Network" Forests 11, no. 4: 442. https://doi.org/10.3390/f11040442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop