Next Article in Journal
Current Status and Potential Invasiveness Evaluation of an Exotic Mangrove Species, Laguncularia racemosa (L.) C.F. Gaertn, on Hainan Island, China
Previous Article in Journal
Comparative Physiological, Transcriptomic, and Metabolomic Analyses of Acacia mangium Provide New Insights into Its Molecular Mechanism of Self-Incompatibility
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stem Detection from Terrestrial Laser Scanning Data with Features Selected via Stem-Based Evaluation

1
School of Smart City, Chongqing Jiaotong University, No. 66, Xuefu Avenue, Nan’an District, Chongqing 400074, China
2
Key Laboratory of Urban Resources Monitoring and Simulation, Ministry of Natural Resources, Shenzhen 518000, China
*
Author to whom correspondence should be addressed.
Forests 2023, 14(10), 2035; https://doi.org/10.3390/f14102035
Submission received: 4 August 2023 / Revised: 8 October 2023 / Accepted: 10 October 2023 / Published: 11 October 2023

Abstract

:
Terrestrial laser scanning (TLS) is an effective tool for extracting stem distribution, providing essential information for forest inventory and ecological studies while also assisting forest managers in monitoring and controlling forest stand density. A feature-based method is commonly integrated into the pipelines of stem detection, facilitating the transition from stem point to stem instance, but most studies focus on feature effectiveness from the point level, neglecting the relationship between stem point extraction and stem detection. In this paper, a feature-based method is proposed to identify stems from TLS data, with features selected from stem levels. Firstly, we propose a series of voxel-based features considering the stem characteristics under the forest. Then, based on the evaluation of some commonly used and proposed features, a stem-based feature selection method is proposed to select a suitable feature combination for stem detection by constructing and evaluating different combinations. Experiments are carried out on three plots with different terrain slopes and tree characteristics, each having a sample plot size of about 8000 m2. The results show that the voxel-based features can supplement the basic features, which improve the average accuracy of stem point extraction and stem detection by 9.5% and 1.2%, respectively. The feature set obtained by the proposed feature selection method achieves a better balance between accuracy and feature number compared with the point-based feature selection method and the features used in previous studies. Moreover, the accuracies of the proposed stem detection methods are also comparable to the three methods evaluated in the international TLS benchmarking project.

1. Introduction

Terrestrial laser scanning (TLS) is an effective tool for observing vegetation objects (e.g., tree stems) at the forest holding level because it can acquire detailed three-dimensional (3D) point clouds in a nondestructive way [1,2]. For forest resources inventory, stem detection is the foundation of many applications such as stem mapping [3,4], the estimation of the diameter at breast height (DBH) [5], stem volume prediction [6], and stem reconstruction [7].
There have been many studies retrieving stem information from the TLS point cloud, which can be categorized into model-fitting and feature-based methods. In the model-fitting methods, the raw point clouds are divided into slices along the vertical direction, and model fitting (e.g., circle fitting [8,9,10], circle-ellipse fitting [11], and cylinder fitting [3,12]) is applied to the sliced points. The point clouds that are successfully fitted to the model will be considered as stem points. However, those methods are easily disturbed by a high stem density and rich understory vegetation, leading to model fitting failure. Therefore, the height and thickness of slices play crucial roles in model-fitting methods [13]. Moreover, before model fitting, image-based methods [14,15,16] and top-based methods [17,18] were used to extract tree locations.
Feature-based methods are commonly used to directly identify all potential wood or stem points from TLS data. Béland et al. [19] classified wood and leaf points based on intensity information because the intensity of wood points is usually higher than that of leaves. Ma et al. [20] separated photosynthetic and non-photosynthetic components using geometric information and two additional enhanced ground filters, which improved the overall classification accuracy of forest canopies. Zhang et al. [21] proposed a method to extract stem points using the curvature feature of the points and connected component segmentation. A part of the branches and foliage points were removed by using the difference of the surface curvature, and subsequently, the stem points were identified effectively via connected component segmentation. Moreover, the deviation between the z-component of normal vectors is commonly used in the first step of stem and leaf separation because it represents the spatial distribution relative to the vertical direction (i.e., verticality) [22,23]. Wang et al. [24] used the median z-normal value and the planar density to detect the stem locations. Liang et al. [12] used the local geometrical features (i.e., flatness and normal direction) to identify the stem points. Xia et al. [25] adopted two-scale classification based on the dimensionality features to extract the candidate stem points. Additionally, the point cloud is translated into a voxel structure and the flatness saliency feature is calculated to recognize the stem points [26].
Moreover, machine learning algorithms are confirmed for leaf–wood classification by using abundant geometric information. Zhu et al. [27] extracted a series of local radiometric and geometric features derived from TLS point clouds and separated wood and leaf points using the Random Forest (RF) model. The overall classification accuracy is between 80% and 90%. Chen et al. [28] also applied the Support Vector Machine (SVM) classifier and combined multiple features, such as dimensionality features, normal vectors, and intensity values. Hackel et al. [29] and Becker et al. [30] constructed 16- and 15-geometric feature Random Forest training models, with classification accuracies of 90% and 84%, respectively. Moorthy et al. [31] used radially bounded nearest neighbors at multiple spatial scales to classify wood and foliage in an RF model. There were 30 features applied, with an overall average accuracy of 94.2%. However, research shows that more features do not mean a higher accuracy in the classification [32].
In summary, most previous feature-based methods have achieved satisfactory point-based results. Then, the individual stem can be detected using segmentation and clustering based on prior knowledge (e.g., tree growth direction) [4,33]. However, there is still room for improvement and discussion. Firstly, although voxel-based features have been proven to be effective for object recognition in urban scenes [34,35,36,37] and have been used to extract stem points with a series of predefined segmentation rules [38,39], their effectiveness has been discussed less under a classification framework in terms of stem extraction results. Secondly, the combination of features, rather than simply integrating features, plays a key role in classification [38,40]. In a forest environment, particularly, factors such as variations in the point cloud density and occlusion can render the features ineffective. Therefore, constructing reasonable feature combinations is a critical aspect of improving the stem detection result. Thirdly, the performance of different feature combinations is commonly evaluated based on point-based results in urban scenes [32,41,42]. However, the utilization of a forest point cloud is typically centered around stem instances, and there has been limited discussion in previous studies regarding the assessment of various feature combinations in the context of stem-based results.
Considering the points mentioned above, the target of this study is to detect stems using a feature-based method and analyze the potential relationship between different feature combinations and stem-based results. Our two main contributions are (1) proposing a series of voxel-based features to identify stem points and detect stems from TLS data considering the stem characteristics under a forest and (2) developing a feature selection method by analyzing the effectiveness of different feature combinations in terms of stem-based metrics.
This article is structured as follows: In Section 2, we provide a comprehensive introduction of the data. Section 3 describes the proposed approach. Following that, Section 4 presents the experimental results. In Section 5, we engage in a discussion of the approach. Lastly, in Section 6, we conclude this article.

2. Data Acquisition and Introduction

Two groups of datasets are used in this study. The first dataset was obtained from three plots with different topographic and tree characteristics (Figure 1). Plot 1 was collected in a camphor forest at Wuhan university (30.54° N, 114.36° E), China, characterized by a flat terrain, straight tree stems, and sparse understory vegetation. Plot 2 was obtained from a sloping mixed forest in Luojia hill (30.59° N, 114.29° E), China, primarily consisting of Masson’s pine, oak and cypress trees. It has a higher density with smaller trees, denser understory vegetation, and more irregular and diverse stem forms. Plot 3 was collected in a mountainous region in Shennongjia (30.07° N, 112.32° E) in China, where the terrain fluctuates greatly and is mainly covered by Euptelea pleiospermum. In this plot, the trees start branching from a height close to the ground and have diverse growth directions. All data were acquired in single-scan mode using Riegl VZ-400 (Riegl GmbH, Horn, Austria) with an angular density of 0.04°, and the largest distance to scanner position was more than 100 m. Detailed information of the three plots is shown in Table 1. The second dataset was obtained from the international TLS benchmarking project organized by the Finish Geospatial Research Institute, including six forest plots obtained using both single- and multi-scan format. These plots are situated in the southern boreal forests of Evo, Finland, covering various tree species, stem densities, developmental stages, and understory vegetation richness. Based on these characteristics, the six plots were categorized into three complexity levels (i.e., easy, medium, and difficult), each with a fixed size of 32 × 32 m2. Further details regarding the comprehensive description of the benchmark datasets can be found in [43]. All experiments were performed with MATLAB R2018b on a computer with Intel Core i7-10750H CPU and 32 GB RAM.
But as point-level accuracy is also validated in this study, point-level reference data are also required. Below is the process of manually labeling each point in this study. We manually divided all the points into three classes: foliage, stem, and ground. Initially, the ground points were extracted via CSF filtering, and the misclassified points were then refined through manual inspection. Subsequently, we separated the stem points from the remaining points, with only the trunk, primary, and secondary branches labeled as stem points in this study, as shown in Figure 2.

3. Methods

The proposed method consists of four major steps (Figure 3): feature extraction, point classification, stem detection, and feature selection. The goal of the first two stages is to determine the subset of probable candidate stem points from the entire input. This is achieved via supervised classification with a series of features, including some commonly used features and the proposed features. Section 3.1 and Section 3.2 describe these procedures in detail. As outlined in Section 3.3, a clustering-based point cloud segmentation framework is applied to individual stem extraction. In the final section of the method, the features are evaluated and selected via the stem-based results after the feature importance ranking based on the point classification results.

3.1. Feature Extraction

We extract candidate stem points using the classification method in this study. Constructing reasonable features is crucial for obtaining satisfying classification results [44,45], and we select the features proposed in previous studies [27,32,46,47] as basic features in this study. As the aim is to provide a group of basic features for the following proposed feature selection strategy, this study does not attempt to cover all ever-used point features. Additionally, feature correlation can be recognized partially by the Recursive Feature Elimination of Random Forest [48] in our strategy. Then, we follow the feature construction and classification framework in [41,49], classifying the basic features (BFs) into 3D features, 2D features, grid features, and intensity-based features, and voxel-based features (VFs) are proposed.

3.1.1. Basic Features (BFs)

  • Three-dimensional features
Each 3D point and its k nearest points form a local neighborhood, and we optimize k for each point to obtain an optimal neighborhood according to the principle of the minimum eigenentropy [41]. In the search process for the optimal neighborhood, the three eigenvalues, λ 1 , λ 2 , and λ 3   ( λ 1 > λ 2 > λ 3 ) , are obtained by performing a principal component analysis (PCA) [50], and the eigenentropy E λ is calculated using Equation (1). Subsequently, by comparing the eigenentropy across different neighborhoods, the neighborhood associated with the minimum eigenentropy is chosen as the optimal neighborhood for each point. Figure 4 illustrates the process of neighborhood selection. Based on the eigenvalues of the optimal neighborhood, a variety of geometric 3D features are defined to represent the spatial distribution of the 3D point cloud within the neighborhood. Table 2 summarizes the first set of 3D features.
E λ = λ 1 ln ( λ 1 ) λ 2 ln ( λ 2 ) λ 3 ln ( λ 3 )
2.
Two-dimensional features
The extraction process of 2D features is similar to that of 3D features. First, the 2D point and its k closest neighbors are obtained based on the cylinder-based neighborhood. Then, the eigenvalues, λ 1 , 2 D and λ 2 , 2 D   ( λ 1 , 2 D > λ 2 , 2 D ) , are calculated by constructing the second-order covariance matrix. From its eigenvalues, the sum of eigenvalues and some basic properties of local 2D neighborhood can be calculated and exploited as 2D features, as shown in Table 3.
3.
Grid features
The grid feature is formed by projecting the scanning area into a regular grid in the XY plane. For the points in the XY plane, the regular grid is used as a special neighborhood for feature extraction, which can simplify the calculation related to the spatial relationship of the point cloud. The grid size is set to 1 m by considering the stem spacing in the sample plot, and then the following features are calculated according to the points in each grid:
  • Ngrid: the point number in each grid.
  • H: the maximum height difference of point cloud in each grid.
  • σH: the standard deviation of point height of each grid.
  • Norz: the z-value difference between each point and the lowest point in each grid.
4.
Intensity-based feature
The intensity information can be used to differentiate wood and leaf points because the intensity of leaf points is usually darker than that of wood. In this paper, the intensity-based feature is labeled as I.

3.1.2. Voxel-Based Features (VF)

Voxelization is an effective way to organize and structure 3D point clouds. In our study, we treat a voxel as a form of a neighborhood instead of directly using it to detect stem points. Firstly, we establish cubical voxels over the whole point cloud, as shown in Figure 5b, and only retain voxels containing at least one point to reduce memory consumption and improve traversal speed, as shown in Figure 5c. Then, the voxel is converted to the axis-aligned bounding box (AABB), as shown in Figure 5d, and the voxel-based features (VFs) are extracted based on both points in each voxel and the voxel size change.
Each retained voxel is recoded as V k , containing m points { p n = ( x n , y n , z n ) | p n V k , n = 1 ,   2 , , m } . Similar to the calculation of basic features, we use the covariance matrix to calculate the normal vector of points within each voxel, which is defined as
= 1 m n = 1 m ( p n p ¯ ) ( p n p ¯ ) T
where p ¯ = ( x ¯ , y ¯ , z ¯ ) is the coordinate of the center of the point within voxels.
After performing the eigenvalue decomposition of matrix in Equation (2), the process yields three distinct eigenvectors along with their corresponding eigenvalues. Among these eigenvectors, the one associated with the smallest eigenvalue is referred to as the normal vector. The different distributions of leaf, stem, and ground points can be captured by the three components, v x , v y , v z , of normal vector based on voxel space, as shown in Figure 6b–d. This difference in spatial distribution makes it possible to separate stem points from other points. As shown in Figure 6c, the stem points have a dominant direction of normal vector, which is indicated by the v x or v y components of the XY plane rather than v z . Similarly, the ground point has a prominent value of v z (Figure 6d). Leaf points, on the other hand, have similar values for both v y and v z (Figure 6b).
In addition, VFs are extracted based on the size difference between the initial voxel and the formed AABB. The bottom area SAABB and the height hAABB are calculated according to the points of each voxel V k as follows:
S A A B B = ( m a x ( x n ) m i n ( x n ) ) × ( m a x ( y n ) m i n ( y n ) )
h A A B B = m a x ( z n ) m i n ( z n )
where xn, yn, and zn are the x-, y-, and z-axis coordinate values of point pn, p n V k , n = 1 , 2 , , m . Then, the bottom area ratio RS and the height ratio Rh can be calculated.
R S = S A A B B / S v o x e l
R h = h A A B B / h v o x e l
where Svoxel and hvoxel indicate the bottom area and height of the initial voxel, which can be calculated using our predefined voxel size.
Figure 6e–g show the differences between the foliage, stem, and ground points. Since the foliage points are usually scattered and evenly distributed in different directions, the size change is relatively small, and both RS and Rh are almost equal to 1 (Figure 6e). For voxels containing stem points, the height ratio Rh is close to 1, while the bottom area ratio RS is much smaller (Figure 6f), because the stem is shaped like a slender pole. Conversely, the ground points cover almost the whole bottom part of the voxel, while the height ratio Rh is much smaller (Figure 6g). Moreover, we also calculate the mean intensity and variance of intensity values of the 3D points within each voxel. Finally, the VF is assigned to each 3D point in the voxel. Table 4 summarizes the VFs and depicts the difference between each class using a radar chart.

3.2. Classification and Feature Importance Evaluation

In this paper, the Random Forest (RF) classifier is used for the stem point identification based on the aforementioned features [50]. The accuracy of stem point extraction is evaluated using Equations (7)–(9).
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where TP is the number of correctly classified stem points, FP is the number of wrongly identified stem points, and FN is the number of stem points that are wrongly labeled as other classes. RF is chosen as it gives a good trade-off between accuracy and computational cost. In addition, it has also been proven to be successfully applied to wood–leaf classification from forest TLS point clouds [31,51]. Meanwhile, RF can also be used for feature selection in the workflow of point classification based on the Recursive Feature Elimination (RFE) algorithm [52]. All features are ranked by the decision tree-based strategies, and the Gini index of each feature can be calculated using the average of all decision trees in the RF classifier. Through this process, a Random Forest can estimate the importance of each feature for the classification task.
To train the RF model, 1000 training samples are randomly selected from three classes, respectively. Moreover, all features are used as inputs in the RF classifier. Note that a unity-based normalization is used to bring the values of each dimension into the range [0, 1] before being applied to the Random Forest classifier.

3.3. Stem Detection

3.3.1. Candidate Stem Recognition

Candidate stem points can be obtained after the classification. Due to the similarity of point features and the uncertainty of noise, there are still non-stem points (mainly leaf points). However, the residual non-stem points are sparser than tree stem points after classification. Thus, the Euclidean clustering algorithm is used to generate stem point clusters and exclude non-stem points [25,28]. If the distance between any two points is less than d, they are labeled as the same cluster; otherwise, they are divided into two different clusters. Generally, the point number of stem clusters is much larger than that of the remaining non-stem clusters. Therefore, we remove the clusters with point number less than Nc.
Our situation is similar to that in [28], in which the maximum scanning distance of the scanner is more than 100 m, and the point density decreases with the increase in distance. In this case, a fixed threshold may not be able to generate appropriate stem clusters and filter small clusters. Therefore, according to the method used in [28], the thresholds of d and Nc are generated adaptively, based on the distance. The threshold d can be calculated as
d = N d p ρ a n g l e
where d p represents the distance from the current point to the scanner position, and ρ a n g l e represents the angular resolution in radians. Therefore, the threshold d is expressed as N times the arc length formed by two adjacent scanning beams.
The adaptive filtering threshold can be obtained using the following equation:
N c = ( D m i n γ H m i n ) / ( ρ a n g l e d c ) 2
where D m i n and H m i n are the preset minimum stem diameter and height, respectively. d c is the average distance from the point in the target cluster to the scanner position, and γ is the occlusion rate, which represents the invisibility or incompleteness of the stem, expressed as a percentage and the value between 0 and 1. According to the characteristics of the trees in different plots, D m i n and H m i n are set to 0.2 and 10 m (plot 1), 0.05 and 8 m (plot 2), and 0.05 and 4 m (plot 3), respectively.

3.3.2. Calculation of Stem Position

In order to calculate the stem position, we use a cylinder model based on the Random Sample Consensus algorithm (RANSAC) [53] to fit the lowest slice of clusters. The cylindrical fitting needs to meet the following two conditions. First, the diameter of a cylinder should be within the range of [0.05, 0.5]. This range is intuitively related with the size of the trees in a specific plot and it will work well as long as the distribution range of the diameter is covered for most trees. Thus, it can be determined based on prior knowledge about the tree species, a random single tree inspection from the point cloud, and on-site sampling. Second, the angle between the cylindrical axis and the z-axis should be less than 8°. It is suitable for most cases, as most trees grow vertically. The intersection point of the cylindrical axis vector with the triangle irregular network (TIN) model constructed from ground points is calculated as the stem position, as shown in Figure 7.

3.4. Stem-Based Feature Selection

Feature combination is commonly evaluated based on point-based classification results, but the basic application is usually based on stems in forest TLS data. Thus, we propose a feature selection method based on stem detection results in this study. The main steps are as follows:
(1)
Data preparation. A sample scan is picked from a plot, and sample points are manually selected for RF training, as described in Section 3.2. Feature importance ranking can be obtained using an RF classifier with the Recursive Feature Elimination (RFE) algorithm. Then, stem positions are marked manually as reference for stem-based result evaluation.
(2)
Feature combination construction. We sort all features in a decreasing order of feature importance and iteratively add them to form different feature combinations. Each combination is used for stem point extraction and stem detection.
(3)
Feature combination evaluation. Stem-based accuracy is calculated for each combination based on the reference data obtained in (1).
C o r r e c t n e s s = n m a t c h n e x t r
C o m p l e t e n e s s = n m a t c h n r e f
I o U = n m a t c h n r e f + n e x t r n m a t c h
where nmatch is the number of found reference stems, nextr is the number of extracted stems, and nref is the number of reference stems. The feature combination with the highest IoU is regarded as the optimal, as shown in Figure 8.

4. Results

4.1. Performance of Voxel-Based Features

In order to analyze the performance of voxel-based features, we compare the results with and without voxel-based features, as shown in Table 5 and Table 6. Thus, two groups of feature combinations are used in this test. The first only consists of the basic features described in Section 3.1.1, labeled as BF in Table 5 and Table 6. The second consists of both basic and voxel-based features, labeled as BF-VF in Table 5 and Table 6. The voxel size for BF-VF is set as x { 0.2   m , 0.4   m , 0.6   m , 0.8   m , 1.0   m , 1.2   m , 1.4   m } , according to the distance of the adjacent stems and the average DBH of the stem in the scene. The result of BF-VF is averaged by different voxel sizes.
The point-based results are shown in Table 5. In order to facilitate an objective comparison, the accuracy of each feature combination is averaged over 10 loops since the result of the RF classifier may vary slightly for each run. It shows that the voxel-based features significantly improve the result of the stem point extraction, with the F1-score increasing by 7.3%, 11.6%, and 9.5% for plots 1, 2, and 3, respectively. In addition, with the increase in the complexity of the sample plot, the accuracy of stem point extraction gradually decreases. Figure 9 shows the detailed comparison of the classification results. It can be seen that more continuous stem (the red boxes in Figure 9a,c) and foliage points (the red box in Figure 9b) can be obtained by including voxel-based features. However, within the blue boxes in Figure 9b, we observe less satisfactory classification results, particularly in areas characterized by thin stems and the presence of foliage. The occlusion of the foliage in these areas diminishes the number of identifiable stem points, which, in turn, affects the effectiveness of voxel-based features.
Table 6 shows the results of stem detection generated on the basis of extracted stem points. The IoU increases by 0.4%, 2.6%, and 0.7% for plots 1, 2, and 3, respectively, when voxel-based features are involved. This shows that the voxel-based features are also helpful in stem detection. It should be noticed that, although more stem points are extracted by BF-VF in plot 1 (recall of 91.81% versus 88.61% in Table 5), fewer stems are detected with the completeness of 0.8656 versus 0.8711 in Table 6. This is mainly because a certain stem segment can be sufficient for stem detection, even the points on the stem surface are not completely extracted.
In order to evaluate the effect of the voxel size, we make an insight into the results in terms of different voxel sizes, as shown in Figure 10. It shows that including voxel-based features can improve the result of stem point extraction with all tested parameters, while the improvement in stem-based results differs among different voxel sizes (Figure 10a). A voxel size of [0.2 m, 0.8 m], including voxel-based features, can improve the stem-based results stably, but both the best and worst results occur when the voxel size is larger than 1 m (Figure 10b). Additionally, a better stem point extraction result may not achieve a better stem detection result, which is similar to the results in Table 5 and Table 6. In total, the proposed voxel-based features are beneficial to improve both point- and stem-based results and the improvement is stable in different parameter settings.

4.2. Effects of Feature Combination on Stem Detection

As shown in Figure 11, we use the Random Forest method to evaluate the importance of each feature for different plots, and the importance value is averaged over different voxel sizes. It is noted that there are 2–4 voxel-based features among the top 5 features ranked by importance in different plots, which helps to explain why voxel-based features can improve the accuracy of stem point extraction. It is worth mentioning that the feature importance ranking in plot 3 is different from those in plots 1 and 2, which is mainly reflected in the features related to vertical distribution, including nz, Vp, Rh, and vz. The probable reason is the large slope in plot 3 and the diverse stem direction.
The point-based feature selection method is commonly used to evaluate features [54,55], so it is used for comparison with the proposed method. To generate different feature combinations, the order of feature importance in Figure 11 is used to successively generate stem-based results with one additional feature added per iteration. The results of stem point extraction and stem detection under different feature combinations are depicted in Figure 12. It shows that with the increase in the feature number, the accuracy of stem point extraction gradually improves at the beginning and finally tends to be stable, which is similar to the conclusion in many previous studies [32,38,42]. And the stem-based accuracy shows the similar trend as a point-based result. However, the result of stem detection is not fully synchronized with the point-based result. Therefore, according to the point-based feature selection method, the optimal feature subsets in the three plots are composed of 26-, 22-, and 19 best-ranked features, respectively, while the first 9, 12, and 15 features will be selected using the proposed method. Compared to the proposed method, the features selected using the point-based feature selection method did not yield higher stem detection results, which further proves that extracting more stem points may not be a necessary condition to stem detection.
A detailed comparison of the two feature selection methods is shown in Table 7. The relation between the point-based and stem-based results in terms of recall (completeness) and precision (correctness) is not absolute either, which is caused by three probable reasons. First, it is important to note that detecting more stem points may not necessarily lead to the detection of more tree stems, as the additionally detected points may belong to the stems that have already been covered (Figure 13a–c). Second, even more non-stem points labeled as stem points can lower the precision significantly, and they can be removed easily because of obviously different characteristics. Third, the addition of features enhances the ability to discriminate details (e.g., branches), but it can lead to adjacent tree stems being unable to separate (Figure 13d–f). Thus, the key to stem detection may be preserving points that can cover stem objects as more as possible. In summary, compared with the point-based feature selection method, the proposed feature selection method can improve the accuracy of stem detection and reduce the feature dimension. In addition, due to distinct optimal feature sets containing different features and numbers, the universality of features is restricted. To address this issue, the union of optimal feature sets from the three plots is regarded as the final feature set Ffinal, consisting of 16 features (Figure 14), and representing the most effective components of optimal feature combinations in each plot. The results are presented in Table 8. Despite a slight decrease in the stem detection accuracy, it still surpasses the results obtained using the point-based method, thereby validating the portability of this feature set. Finally, Figure 15 illustrates the stem detection result using the proposed feature selection method in each plot, with the slope of different plots and the distance to the scanner position. It can be seen that most undetected trees are distributed far away from the scanner as the stems at a large distance are prone to be obscured by the occlusion of the foreground objects.

4.3. Comparison with Feature-Based Methods

The final feature set obtained by using the feature selection method in this paper is compared with four feature combinations used in previous studies [28,31,33,38], as shown in Table 9. The experimental data were collected from different scans in the three research areas outlined in Section 2 to verify the effectiveness of the proposed method in the plot with similar structural distributions and stem characteristics.
Compared to the four feature combinations in previous studies, the feature number selected using the proposed method and the accuracy of the stem point extraction reach middle to upper levels, while the accuracy of stem detection is the highest in plots 1 and 2, and only slightly lower than that of [31] in plot 3. First, 30 features were calculated at five scales for stem–leaf separation in [31], and the highest accuracy of stem point extraction was obtained, particularly in plot 3. However, the stem detection accuracy did not improve significantly, which aligned with the conclusions of [21] and this paper; although most stem points are extracted, stem detection accuracy is still influenced by the scene’s complexity. Secondly, the feature sets used in [28,33] resulted in lower accuracies of stem point extraction and stem detection compared with our method. However, it is worth noting that the feature set in [38] produced a stem detection accuracy that was nearly identical to that obtained by the feature set selected in this paper. This is possibly due to the use of voxel-based features, which are advantageous for stem detection. Additionally, even though the stem detection accuracy of the feature sets used in [31,38] is similar to this paper, they use more features that require more processing time and memory. Therefore, the stem-based feature selection method has certain advantages, which can ensure the accuracy of stem detection and avoid the effective feature loss and the effect of redundant features.

4.4. Experiments on Benchmark Datasets

To further validate the effectiveness of the proposed method in stem detection, we conducted verification using the second dataset, the benchmark dataset. We selected points from the multi-scan TLS datasets at a regular interval of 0.01 m to reduce the number of points and calculated each feature of the final feature set Ffinal. Before computing the voxel features, a voxel size of 0.6 m was set. Table 10 and Table 11 show the accuracy of stem detection of the six plots using both single- and multi-scan TLS datasets. The mean accuracy measured in there is more suitable for comparison with other methods.
From the stem detection results of the single-scan scenarios, the proposed method achieves a superior balance between completeness and correctness, while also attaining the highest mean accuracy. Although correctness reached 100% in plots 2, 3, and 6 in [21], it was achieved at the cost of reduced completeness, indicating that a considerable number of stems in the plots remained undiscovered. Similarly, in the multi-scan scenarios, the proposed method exhibits lower correctness compared to [21] but offers higher completeness. The mean accuracy of the proposed method across all plots is slightly lower than that achieved using the deep learning approaches [56]. However, it is worth noting that, whether in single- or multi-scan mode, the proposed method detects more trees in difficult plots, and the trade-off in correctness remains acceptable.

5. Discussion

In previous studies, feature combinations are commonly evaluated from the view of point-based results [40,47]. Here, we focus on stem-based results, as applications of the point cloud under a forest scene are usually based on stem instance, e.g., stem mapping [3], DBH estimation [10], and biomass estimation [57]. For stem detection, we only require partial stem points to fit a stem model (e.g., cylinder); thus, missing some stem points may not cause stem misdetection. On the other side, as non-stem points can be filtered via clustering [25] or cylindrical fitting [12] in the process of stem detection, the points falsely labeled as stem in classification may not cause a false detection. This also explains the above experimental results that the improving stem point extraction result may not improve the stem detection result. Considering the above points, we try to construct feature combination from both point-based and stem-based evaluations. Although feature combination is not directly related to stem-based results, an obvious correlation between them can be observed (Figure 12). The comparison of different feature combinations also shows that the stem-based feature selection method is effective. Compared with point-based feature selection, less features are used to obtain similar or better stem-based results in our method, which significantly reduces the computational burden with respect to processing time and memory consumption. Moreover, compared to constructing feature combination using prior knowledge, the feature set selected in this study performed well and struck a good balance between stem detection in various forest scenes and feature redundancy. In terms of the scanning mode, we first tested and validated the effectiveness of the proposed method on single-scan TLS data. The single-scan mode serves as the foundation for the multi-scan mode, and the stem positions obtained from the single-scan mode can be used for TLS point cloud registration [58,59,60] and even assist in the fusion of multi-sensor data [61,62]. Subsequent testing on the benchmark dataset indicated the advantages of the proposed method in all plots under the single-scan mode and in two different plots under the multi-scan mode. However, the difference in accuracy is related to the plot size and forest stand complexity [63,64]. In regions with dense vegetation within the plots and along the plot edges, the stem detection accuracy significantly decreases.
The feature set obtained using the stem-based selection method in this study demonstrates the significant contributions of voxel-based features. Compared to basic features, the inclusion of voxel-based features shows an improved accuracy in stem point extraction across different voxel sizes (Figure 10a). However, for stem detection, the optimal results were achieved in three different plots with voxel sizes of 1.4 m, 0.8 m, and 0.6 m (Figure 10b), indicating that the determination of the voxel size should consider both the stem density and understory vegetation richness. Using excessively large voxel sizes may lead to the covering of stem points and non-stem points (e.g., foliage and shrub) within a single voxel, while overly small voxel sizes may not be able to contain enough stem points, especially in regions with significant density variations. Both situations can potentially result in errors in stem detection caused by a decrease in the capability of feature expression. The features I, VI, and VarI related to intensity information and the features ∆H, σH, and Norz related to height are included in the optimal feature set and make significant contributions to all three plots. This is because the ground, stem, and foliage have different materials and spatial distributions. Moreover, the normal vector related with point distribution in the vertical direction was often used in previous studies [3,24,28], as it is consistent with the law that trees grow upward. But normal vector-based features (nz and vz) did not show a high importance across all plots in our test. Figure 16 and Figure 17 show the histogram of normal vector-based features, nz and vz, respectively. In plots 1 and 2, the nz and vz of stems are distributed in significantly different intervals from the other two types of objects, while the normal vector distribution of three objects is relatively divergent in plot 3, and there is a similar point number of three objects under the same interval (Figure 16 and Figure 17c), which leads to a decrease in the effectiveness of normal vector-based features. The possible reasons lie in that the trees start branching from a height that is close to the ground, and the terrain slope close to 42° also affects the growth direction in plot 3, which is similar to the conclusion in [27]. Therefore, the performance of normal vector-based features is reduced in plot 3. Moreover, the features Sumλ, D3D, Sumλ,2D, and D2D are always among the least relevant features, which is consistent to similar studies involving the features [32].

6. Conclusions

In this study, we focus on the improvement in stem detection results by proposing seven voxel-based features and optimizing the feature combination using a stem-based feature selection method. The result shows that the voxel-based features are beneficial for stem point extraction and stem detection in forest scenes. Moreover, our study highlights the importance of tailoring feature combinations specifically for stem detection tasks, as the benefits observed in point-based results may not always extend to stem-based results. The proposed method can provide more reliable feature combination for stem detection in different forests. Compared with the feature set in previous studies, the feature combination constructed using our method achieves a good balance between the feature number and accuracy. Regarding stem detection, the proposed method also yields comparable results on the benchmark dataset and exhibits better mean accuracy on difficult plots. It is also found that the performance of some features related with point distribution in the vertical direction is significantly affected by the terrain slopes and tree characteristics. In general, the accuracy of stem detection will decrease with an increasing forest complexity, because complex forests usually have a strong and increasing likelihood of occlusion. Integrating data from multiple sensors (e.g., TLS, UAV, and ALS) has the potential to enhance the point cloud quality in complex forests, improving stem detection accuracy and enabling a more precise estimation of forest structural parameters. Additionally, the stem positions calculated in this paper can be valuable for assisting in point cloud registration or fusion. Nevertheless, this requires further investigation and study.

Author Contributions

Conceptualization, M.C.; data curation, M.C.; formal analysis, X.L.; funding acquisition, M.C.; investigation, F.M. and L.Z.; methodology, X.L.; project administration, J.P.; resources, M.C.; supervision, M.C., J.P. and L.Z.; validation, X.L.; visualization, X.L. and F.M.; writing—original draft, X.L.; writing—review and editing, M.C. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key R&D Program of Ningxia Autonomous Region (grant no. 2022CMG02014), the National Natural Science Foundation of China (grant no. 41801394), and the Open Fund of Key Laboratory of Urban Resources Monitoring and Simulation, Ministry of Natural Resources (grant no. KF-2021-06-102).

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Zhang (Xinyi Zhang) for her constructive comments and suggestions that improved the quality of this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wilkes, P.; Lau, A.; Disney, M.; Calders, K.; Burt, A.; Gonzalez de Tanago, J.; Bartholomeus, H.; Brede, B.; Herold, M. Data Acquisition Considerations for Terrestrial Laser Scanning of Forest Plots. Remote Sens. Environ. 2017, 196, 140–153. [Google Scholar] [CrossRef]
  2. Wang, D. Unsupervised Semantic and Instance Segmentation of Forest Point Clouds. ISPRS J. Photogramm. Remote Sens. 2020, 165, 86–97. [Google Scholar] [CrossRef]
  3. Liang, X.; Litkey, P.; Hyyppa, J.; Kaartinen, H.; Vastaranta, M.; Holopainen, M. Automatic Stem Mapping Using Single-Scan Terrestrial Laser Scanning. IEEE Trans. Geosci. Remote Sens. 2012, 50, 661–670. [Google Scholar] [CrossRef]
  4. Ye, W.; Qian, C.; Tang, J.; Liu, H.; Fan, X.; Liang, X.; Zhang, H. Improved 3D Stem Mapping Method and Elliptic Hypothesis-Based DBH Estimation from Terrestrial Laser Scanning Data. Remote Sens. 2020, 12, 352. [Google Scholar] [CrossRef]
  5. Cabo, C.; Ordóñez, C.; López-Sánchez, C.A.; Armesto, J. Automatic Dendrometry: Tree Detection, Tree Height and Diameter Estimation Using Terrestrial Laser Scanning. Int. J. Appl. Earth Obs. Geoinf. 2018, 69, 164–174. [Google Scholar] [CrossRef]
  6. Pitkänen, T.P.; Raumonen, P.; Kangas, A. Measuring Stem Diameters with TLS in Boreal Forests by Complementary Fitting Procedure. ISPRS J. Photogramm. Remote Sens. 2019, 147, 294–306. [Google Scholar] [CrossRef]
  7. Rouzbeh Kargar, A.; MacKenzie, R.A.; Apwong, M.; Hughes, E.; van Aardt, J. Stem and Root Assessment in Mangrove Forests Using a Low-Cost, Rapid-Scan Terrestrial Laser Scanner. Wetl. Ecol. Manag. 2020, 28, 883–900. [Google Scholar] [CrossRef]
  8. Pueschel, P.; Newnham, G.; Rock, G.; Udelhoven, T.; Werner, W.; Hill, J. The Influence of Scan Mode and Circle Fitting on Tree Stem Detection, Stem Diameter and Volume Extraction from Terrestrial Laser Scans. ISPRS J. Photogramm. Remote Sens. 2013, 77, 44–56. [Google Scholar] [CrossRef]
  9. Koreň, M.; Mokroš, M.; Bucha, T. Accuracy of Tree Diameter Estimation from Terrestrial Laser Scanning by Circle-Fitting Methods. Int. J. Appl. Earth Obs. Geoinf. 2017, 63, 122–128. [Google Scholar] [CrossRef]
  10. Maas, H.-G.; Bienert, A.; Scheller, S.; Keane, E. Automatic Forest Inventory Parameter Determination from Terrestrial Laser Scanner Data. Int. J. Remote Sens. 2008, 29, 1579–1593. [Google Scholar] [CrossRef]
  11. Bu, G.; Wang, P. Adaptive Circle-Ellipse Fitting Method for Estimating Tree Diameter Based on Single Terrestrial Laser Scanning. J. Appl. Remote Sens. 2016, 10, 026040. [Google Scholar] [CrossRef]
  12. Liang, X.; Kankare, V.; Yu, X.; Hyyppa, J.; Holopainen, M. Automated Stem Curve Measurement Using Terrestrial Laser Scanning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1739–1748. [Google Scholar] [CrossRef]
  13. Liu, G.; Wang, J.; Dong, P.; Chen, Y.; Liu, Z. Estimating Individual Tree Height and Diameter at Breast Height (DBH) from Terrestrial Laser Scanning (TLS) Data at Plot Level. Forests 2018, 9, 398. [Google Scholar] [CrossRef]
  14. Haala, N.; Reulke, R.; Thies, M.; Aschoff, T. Combination of terrestrial Laser Scanning with high resolution panoramic Images for Investigations in Forest Applications and tree species recognition. In Proceedings of the Panoramic Photogrammetry Workshop, Berlin, Germany, 24–25 February 2005. [Google Scholar]
  15. Forsman, P.; Halme, A. 3-D mapping of natural environments with trees by means of mobile perception. IEEE Trans. Robot. 2005, 21, 482–490. [Google Scholar] [CrossRef]
  16. Forsman, M.; Börlin, N.; Holmgren, J. Estimation of tree stem attributes using terrestrial photogrammetry. In Proceedings of the ISPRS-International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Melbourne, Australia, 25 August–1 September 2012; pp. 261–265. [Google Scholar]
  17. Xia, S.; Chen, D.; Peethambaran, J.; Wang, P.; Xu, S. Point Cloud Inversion: A Novel Approach for the Localization of Trees in Forests from TLS Data. Remote Sens. 2021, 13, 338. [Google Scholar] [CrossRef]
  18. Huo, L.; Lindberg, E.; Holmgren, J. Towards low vegetation identification: A new method for tree crown segmentation from LiDAR data based on a symmetrical structure detection algorithm (SSD). Remote Sens. Environ. 2022, 270, 112857. [Google Scholar] [CrossRef]
  19. Béland, M.; Widlowski, J.-L.; Fournier, R.A.; Côté, J.-F.; Verstraete, M.M. Estimating Leaf Area Distribution in Savanna Trees from Terrestrial LiDAR Measurements. Agric. For. Meteorol. 2011, 151, 1252–1266. [Google Scholar] [CrossRef]
  20. Ma, L.; Zheng, G.; Eitel, J.U.H.; Moskal, L.M.; He, W.; Huang, H. Improved Salient Feature-Based Approach for Automatically Separating Photosynthetic and Nonphotosynthetic Components Within Terrestrial Lidar Point Cloud Data of Forest Canopies. IEEE Trans. Geosci. Remote Sens. 2016, 54, 679–696. [Google Scholar] [CrossRef]
  21. Zhang, W.; Wan, P.; Wang, T.; Cai, S.; Chen, Y.; Jin, X.; Yan, G. A Novel Approach for the Detection of Standing Tree Stems from Plot-Level Terrestrial Laser Scanning Data. Remote Sens. 2019, 11, 211. [Google Scholar] [CrossRef]
  22. Wang, D.; Momo Takoudjou, S.; Casella, E. LeWoS: A Universal Leaf-wood Classification Method to Facilitate the 3D Modelling of Large Tropical Trees Using Terrestrial LiDAR. Methods Ecol. Evol. 2020, 11, 376–389. [Google Scholar] [CrossRef]
  23. Li, S.; Dai, L.; Wang, H.; Wang, Y.; He, Z.; Lin, S. Estimating Leaf Area Density of Individual Trees Using the Point Cloud Segmentation of Terrestrial LiDAR Data and a Voxel-Based Model. Remote Sens. 2017, 9, 1202. [Google Scholar] [CrossRef]
  24. Wang, D.; Hollaus, M.; Puttonen, E.; Pfeifer, N. Automatic and Self-Adaptive Stem Reconstruction in Landslide-Affected Forests. Remote Sens. 2016, 8, 974. [Google Scholar] [CrossRef]
  25. Xia, S.; Wang, C.; Pan, F.; Xi, X.; Zeng, H.; Liu, H. Detecting Stems in Dense and Homogeneous Forest Using Single-Scan TLS. Forests 2015, 6, 3923–3945. [Google Scholar] [CrossRef]
  26. Olofsson, K.; Holmgren, J. Single Tree Stem Profile Detection Using Terrestrial Laser Scanner Data, Flatness Saliency Features and Curvature Properties. Forests 2016, 7, 207. [Google Scholar] [CrossRef]
  27. Zhu, X.; Skidmore, A.K.; Darvishzadeh, R.; Niemann, K.O.; Liu, J.; Shi, Y.; Wang, T. Foliar and Woody Materials Discriminated Using Terrestrial LiDAR in a Mixed Natural Forest. Int. J. Appl. Earth Obs. Geoinf. 2018, 64, 43–50. [Google Scholar] [CrossRef]
  28. Chen, M.; Wan, Y.; Wang, M.; Xu, J. Automatic Stem Detection in Terrestrial Laser Scanning Data with Distance-Adaptive Search Radius. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2968–2979. [Google Scholar] [CrossRef]
  29. Hackel, T.; Wegner, J.D.; Schindler, K. Fast Semantic Segmentation of 3D Point Clouds with Strongly Varying Density. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 177–184. [Google Scholar] [CrossRef]
  30. Becker, C.; Rosinskaya, E.; Häni, N.; D’Angelo, E.; Strecha, C. Classification of Aerial Photogrammetric 3D Point Clouds. Photogramm. Eng. Remote Sens. 2018, 84, 287–295. [Google Scholar] [CrossRef]
  31. Krishna Moorthy, S.M.; Calders, K.; Vicari, M.B.; Verbeeck, H. Improved Supervised Learning-Based Approach for Leaf and Wood Classification from LiDAR Point Clouds of Forests. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3057–3070. [Google Scholar] [CrossRef]
  32. Weinmann, M.; Jutzi, B.; Mallet, C. Feature Relevance Assessment for the Semantic Interpretation of 3D Point Cloud Data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, II5, 313–318. [Google Scholar] [CrossRef]
  33. Vicari, M.B.; Disney, M.; Wilkes, P.; Burt, A.; Calders, K.; Woodgate, W. Leaf and Wood Classification Framework for Terrestrial LiDAR Point Clouds. Methods Ecol. Evol. 2019, 10, 680–694. [Google Scholar] [CrossRef]
  34. Lim, E.H.; Suter, D. 3D Terrestrial LIDAR Classifications with Super-Voxels and Multi-Scale Conditional Random Fields. Comput.-Aided Des. 2009, 41, 701–710. [Google Scholar] [CrossRef]
  35. Plaza-Leiva, V.; Gomez-Ruiz, J.A.; Mandow, A.; García-Cerezo, A. Voxel-Based Neighborhood for Spatial Shape Pattern Classification of Lidar Point Clouds with Supervised Learning. Sensors 2017, 17, 594. [Google Scholar] [CrossRef] [PubMed]
  36. Aijazi, A.; Checchin, P.; Trassoudaine, L. Segmentation Based Classification of 3D Urban Point Clouds: A Super-Voxel Based Approach with Evaluation. Remote Sens. 2013, 5, 1624–1650. [Google Scholar] [CrossRef]
  37. Xu, Y.; Tong, X.; Stilla, U. Voxel-Based Representation of 3D Point Clouds: Methods, Applications, and Its Potential Use in the Construction Industry. Autom. Constr. 2021, 126, 103675. [Google Scholar] [CrossRef]
  38. Xu, S.; Zhou, K.; Sun, Y.; Yun, T. Separation of Wood and Foliage for Trees from Ground Point Clouds Using a Novel Least-Cost Path Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6414–6425. [Google Scholar] [CrossRef]
  39. Brolly, G.; Király, G.; Lehtomäki, M.; Liang, X. Voxel-Based Automatic Tree Detection and Parameter Retrieval from Terrestrial Laser Scans for Plot-Wise Forest Inventory. Remote Sens. 2021, 13, 542. [Google Scholar] [CrossRef]
  40. Weinmann, M.; Jutzi, B.; Mallet, C.; Weinmann, M. Geometric Features and Their Relevance for 3D Point Cloud Classification. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 4, 157–164. [Google Scholar] [CrossRef]
  41. Weinmann, M.; Jutzi, B.; Hinz, S.; Mallet, C. Semantic Point Cloud Interpretation Based on Optimal Neighborhoods, Relevant Features and Efficient Classifiers. ISPRS J. Photogramm. Remote Sens. 2015, 105, 286–304. [Google Scholar] [CrossRef]
  42. Li, Q.; Cheng, X. Comparison of Different Feature Sets for TLS Point Cloud Classification. Sensors 2018, 18, 4206. [Google Scholar] [CrossRef]
  43. Liang, X.; Hyyppä, J.; Kaartinen, H.; Lehtomäki, M.; Pyörälä, J.; Pfeifer, N.; Holopainen, M.; Brolly, G.; Francesco, P.; Hackenberg, J.; et al. International Benchmarking of Terrestrial Laser Scanning Approaches for Forest Inventories. ISPRS J. Photogramm. Remote Sens. 2018, 144, 137–179. [Google Scholar] [CrossRef]
  44. Farabet, C.; Couprie, C.; Najman, L.; LeCun, Y. Learning Hierarchical Features for Scene Labeling. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1915–1929. [Google Scholar] [CrossRef] [PubMed]
  45. Weinmann, M.; Mallet, C.; Jutzi, B. Involving Different Neighborhood Types for the Analysis of Low-Level Geometric 2D and 3D Features and Their Relevance for Point Cloud Classification. Publ. Dtsch. Ges. Für. Photogramm. 2017, 26, 179–191. [Google Scholar]
  46. Atik, M.E.; Duran, Z.; Seker, D.Z. Machine Learning-Based Supervised Classification of Point Clouds Using Multiscale Geometric Features. ISPRS Int. J. Geo-Inf. 2021, 10, 187. [Google Scholar] [CrossRef]
  47. Günen, M.A. Adaptive Neighborhood Size and Effective Geometric Features Selection for 3D Scattered Point Cloud Classification. Appl. Soft Comput. 2022, 115, 108196. [Google Scholar] [CrossRef]
  48. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  49. Weinmann, M.; Urban, S.; Hinz, S.; Jutzi, B.; Mallet, C. Distinctive 2D and 3D features for automated large-scale scene analysis in urban areas. Comput. Graph. 2015, 49, 47–57. [Google Scholar] [CrossRef]
  50. Lalonde, J.-F.; Vandapel, N.; Huber, D.F.; Hebert, M. Natural Terrain Classification Using Three-Dimensional Ladar Data for Ground Robot Mobility. J. Field Robot. 2006, 23, 839–861. [Google Scholar] [CrossRef]
  51. Wang, D.; Hollaus, M.; Pfeifer, N. Feasibility of Machine Learning Methods for Separating Wood and Leaf Points from Terrestrial Laser Scanning Data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 157–164. [Google Scholar] [CrossRef]
  52. Gregorutti, B.; Michel, B.; Saint-Pierre, P. Correlation and variable importance in random forests. Stat. Comput. 2017, 27, 659–678. [Google Scholar] [CrossRef]
  53. Olofsson, K.; Holmgren, J.; Olsson, H. Tree Stem and Height Measurements Using Terrestrial Laser Scanning and the RANSAC Algorithm. Remote Sens. 2014, 6, 4323–4344. [Google Scholar] [CrossRef]
  54. Niemeyer, J.; Rottensteiner, F.; Soergel, U. Contextual classification of lidar data and building object detection in urban areas. ISPRS J. Photogramm. Remote Sens. 2014, 87, 152–165. [Google Scholar] [CrossRef]
  55. Chen, M.; Pan, J.; Xu, J. Classification of terrestrial laser scanning data with density-adaptive geometric features. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1795–1799. [Google Scholar] [CrossRef]
  56. Chang, L.; Fan, H.; Zhu, N.; Dong, Z. A Two-stage Approach for Individual Tree Segmentation from TLS Point Clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8682–8693. [Google Scholar] [CrossRef]
  57. Tao, S.; Guo, Q.; Xu, S.; Su, Y.; Li, Y.; Wu, F. A Geometric Method for Wood-Leaf Separation Using Terrestrial and Simulated Lidar Data. Photogramm. Eng. Remote Sens. 2015, 81, 767–776. [Google Scholar] [CrossRef]
  58. Liang, X.; Hyyppä, J. Automatic stem mapping by merging several terrestrial laser scans at the feature and decision levels. Sensors 2013, 13, 1614–1634. [Google Scholar] [CrossRef]
  59. Tremblay, J.F.; Béland, M. Towards operational marker-free registration of terrestrial lidar data in forests. ISPRS J. Photogramm. Remote Sens. 2018, 146, 430–435. [Google Scholar] [CrossRef]
  60. Kelbe, D.; Van Aardt, J.; Romanczyk, P.; Van Leeuwen, M.; Cawse-Nicholson, K. Marker-free registration of forest terrestrial laser scanner data pairs with embedded confidence metrics. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4314–4330. [Google Scholar] [CrossRef]
  61. Hauglin, M.; Lien, V.; Næsset, E.; Gobakken, T. Geo-referencing forest field plots by co-registration of terrestrial and airborne laser scanning data. Int. J. Remote Sens. 2014, 35, 3135–3149. [Google Scholar] [CrossRef]
  62. Polewski, P.; Erickson, A.; Yao, W.; Coops, N.; Krzystek, P.; Stilla, U. Object-based coregistration of terrestrial photogrammetric and als point clouds in forested areas. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 347–354. [Google Scholar] [CrossRef]
  63. Heinzel, J.; Huber, M. Detecting Tree Stems from Volumetric TLS Data in Forest Environments with Rich Understory. Remote Sens. 2016, 9, 9. [Google Scholar] [CrossRef]
  64. Pitkänen, T.P.; Raumonen, P.; Liang, X.; Lehtomäki, M.; Kangas, A. Improving TLS-Based Stem Volume Estimates by Field Measurements. Comput. Electron. Agric. 2021, 180, 105882. [Google Scholar] [CrossRef]
Figure 1. The location of the study area of first dataset.
Figure 1. The location of the study area of first dataset.
Forests 14 02035 g001
Figure 2. Illustration of manually labeled point cloud in first dataset. Green: foliage points. Black: stem points. Orange: ground points.
Figure 2. Illustration of manually labeled point cloud in first dataset. Green: foliage points. Black: stem points. Orange: ground points.
Forests 14 02035 g002
Figure 3. The general framework of this study.
Figure 3. The general framework of this study.
Forests 14 02035 g003
Figure 4. Illustration of the neighborhood selection. (a) Neighborhood selection result (colored with the selected k value) on sample data. (b) The neighbors pk ( k [ 10 ,   100 ] ,   Δ k = 1 ) of one stem point (black dot) close to foliage. (c) The neighbors pk ( k [ 10 ,   100 ] ,   Δ k = 1 ) of one stem point (black dot) close to ground. The color of each neighbor pk corresponds to the value of eigenentropy E λ computed on the smallest neighborhood containing pk. The eigenentropy first decreases until the neighborhood reaches the edge of the stem (red points show the range of optimal neighborhood), and then increases and reaches its maximum when a different object (e.g., foliage and ground) is aggregated within the neighborhood.
Figure 4. Illustration of the neighborhood selection. (a) Neighborhood selection result (colored with the selected k value) on sample data. (b) The neighbors pk ( k [ 10 ,   100 ] ,   Δ k = 1 ) of one stem point (black dot) close to foliage. (c) The neighbors pk ( k [ 10 ,   100 ] ,   Δ k = 1 ) of one stem point (black dot) close to ground. The color of each neighbor pk corresponds to the value of eigenentropy E λ computed on the smallest neighborhood containing pk. The eigenentropy first decreases until the neighborhood reaches the edge of the stem (red points show the range of optimal neighborhood), and then increases and reaches its maximum when a different object (e.g., foliage and ground) is aggregated within the neighborhood.
Forests 14 02035 g004
Figure 5. The axis-aligned bounding box (AABB) is formed according to the 3D point in each voxel.
Figure 5. The axis-aligned bounding box (AABB) is formed according to the 3D point in each voxel.
Forests 14 02035 g005
Figure 6. Illustration of (a) a tree example with the initial voxel size of 1 m, the three components ( v x , v y , v z ) of normal vector to represent the different spatial distributions of (b) foliage, (c) stem, and (d) ground points within a voxel, and (eg) the height ratio Rh and bottom area ratio RS calculated using the height hAABB and bottom area SAABB of AABBs.
Figure 6. Illustration of (a) a tree example with the initial voxel size of 1 m, the three components ( v x , v y , v z ) of normal vector to represent the different spatial distributions of (b) foliage, (c) stem, and (d) ground points within a voxel, and (eg) the height ratio Rh and bottom area ratio RS calculated using the height hAABB and bottom area SAABB of AABBs.
Forests 14 02035 g006
Figure 7. Calculation of stem position. Each stem were assigned random colors.
Figure 7. Calculation of stem position. Each stem were assigned random colors.
Forests 14 02035 g007
Figure 8. The flowchart of the proposed feature selection method.
Figure 8. The flowchart of the proposed feature selection method.
Forests 14 02035 g008
Figure 9. Detailed comparison of classification results without (left) and with (right) voxel-based features at (a) plot 1, (b) plot 2, and (c) plot 3.
Figure 9. Detailed comparison of classification results without (left) and with (right) voxel-based features at (a) plot 1, (b) plot 2, and (c) plot 3.
Forests 14 02035 g009
Figure 10. The results of (a) stem point extraction and (b) stem detection in terms of different voxel sizes. BF: basic features; BF-VF(x): basic features and voxel-based features with voxel size of x.
Figure 10. The results of (a) stem point extraction and (b) stem detection in terms of different voxel sizes. BF: basic features; BF-VF(x): basic features and voxel-based features with voxel size of x.
Forests 14 02035 g010
Figure 11. The importance of each feature in different plots (blue: basic features; orange: voxel-based features). The number on each column represents the ranking.
Figure 11. The importance of each feature in different plots (blue: basic features; orange: voxel-based features). The number on each column represents the ranking.
Forests 14 02035 g011
Figure 12. The accuracy of stem point detection (F1-score) and stem detection (IoU) in terms of different feature combinations.
Figure 12. The accuracy of stem point detection (F1-score) and stem detection (IoU) in terms of different feature combinations.
Forests 14 02035 g012
Figure 13. Comparison of point-based methods and proposed methods in stem point extraction and candidate stem recognition. Black: stem points; green: foliage points; orange: ground points; blue and red: different candidate stems.
Figure 13. Comparison of point-based methods and proposed methods in stem point extraction and candidate stem recognition. Black: stem points; green: foliage points; orange: ground points; blue and red: different candidate stems.
Forests 14 02035 g013
Figure 14. The final feature set composed of optimal feature combinations from different plots.
Figure 14. The final feature set composed of optimal feature combinations from different plots.
Forests 14 02035 g014
Figure 15. Stem maps from the position calculation. The altitude value is rendered using the color legend on the right.
Figure 15. Stem maps from the position calculation. The altitude value is rendered using the color legend on the right.
Forests 14 02035 g015
Figure 16. The histogram of the normal vectors’ z-values nz of foliage points stem points and ground points in different plots.
Figure 16. The histogram of the normal vectors’ z-values nz of foliage points stem points and ground points in different plots.
Forests 14 02035 g016
Figure 17. The histogram of the normal vectors’ z-values vz of foliage points stem points and ground points in different plots.
Figure 17. The histogram of the normal vectors’ z-values vz of foliage points stem points and ground points in different plots.
Forests 14 02035 g017
Table 1. Detailed information of the three plots in the first dataset.
Table 1. Detailed information of the three plots in the first dataset.
Plot IDPlot Size (m2)Sum of TreesDBH (cm)Tree Height (m)Slope (°)Point Spacing (m)Point Number
AllStemGroundFoliage
1120 × 7028727.5 ± 6.914.9 ± 2.5~20.001~1.17523,794,6651,996,8878,056,32613,741,452
2140 × 6038621.8 ± 5.011.1 ± 4.6~170.001~0.93523,332,7882,179,6486,692,75714,460,383
3140 × 6048012.8 ± 4.27.6 ± 2.3~420.001~1.4257,568,384508,8532,112,4514,947,080
Table 2. Three-dimensional features extracted from the point cloud.
Table 2. Three-dimensional features extracted from the point cloud.
No.SymbolFeatureDefinition
1 L λ linear saliency ( λ 1 λ 2 ) / λ 1
2 P λ planar saliency ( λ 2 λ 3 ) / λ 1
3 S λ volumetric saliency λ 3 / λ 1
4 E 3 D Shannon entropy L λ ln L λ P λ ln P λ S λ ln S λ
5 E λ eigenentropy n = 1 3 λ n ln λ n
6 n x x-component of the normal vector NV 1NV (1)
7 n y y-component of the normal vector NVNV (2)
8 n z z-component of the normal vector NVNV (3)
9 O λ structure tensor omnivariance ( λ 1 λ 2 λ 3 ) 1 / 3
10 A λ structure tensor anisotropy ( λ 1 λ 3 ) / λ 1
11 S u m λ sum of eigenvalues n = 1 3 λ n
12 C λ change in curvature λ 3 / ( λ 1 + λ 2 + λ 3 )
13 V p verticality 1 n z
14 r k n n radius of the spherical neighborhood\
15 D 3 D local point density 3 ( k + 1 ) / ( 4 π r k n n 3 )
1  N V 3 denotes the normal vector, and k is the suitable number of closest neighbors for each given 3D point.
Table 3. Seven 2D features extracted from the point cloud.
Table 3. Seven 2D features extracted from the point cloud.
No.SymbolFeatureDefinition
1 L λ , 2 D linear saliency of local 2D neighborhood ( λ 1 λ 2 ) / λ 1
2 P λ , 2 D planar saliency of local 2D neighborhood λ 2 / λ 1
3 E 2 D Shannon entropy of 2D neighborhood L λ ln L λ P λ ln P λ
4 E λ , 2 D eigenentropy of 2D neighborhood n = 1 2 λ n , 2 D ln λ n , 2 D
5 O λ , 2 D structure tensor omnivariance of 2D neighborhood ( λ 1 λ 2 ) 1 / 2
6 D 2 D local 2D point density ( k + 1 ) / ( π r k n n , 2 D 2 )   1
7 S u m λ , 2 D sum of eigenvalues of 2D neighborhood n = 1 2 λ n , 2 D
1    r k n n , 2 D is radius of the circular neighborhood defined by a 2D point and its k closest neighbors.
Table 4. Voxel-based features and the difference between each class.
Table 4. Voxel-based features and the difference between each class.
Voxel-Based FeatureDescription 1
R S The bottom area ratio Forests 14 02035 i001
R h The height ratio
ν x , ν y , ν z Three components of normal vector
in   voxel   V k
V I Mean intensity value
of   constituting   3 D   points   in   voxel   V k
V a r I Variance   of   intensity   values   within   voxel   V k
1 The radar chart illustrates variations in the normalized values of features among different classes; distinguishing the differences between these classes is crucial for classification.
Table 5. Results of stem point extraction with and without the voxel-based features.
Table 5. Results of stem point extraction with and without the voxel-based features.
Plot IDActual QuantityFeature SetEstimated QuantityCorrectly Estimated
Quantity
Precision (%)Recall (%)F1-Score (%)
11,996,887BF2,748,6481,769,37764.4188.6174.58
BF-VF2,481,6831,833,30473.9191.8181.88
22,179,648BF1,915,2211,132,38159.1351.9555.31
BF-VF2,012,3641,402,33669.7964.3466.86
3508,853BF605,041256,06542.3450.3245.98
BF-VF577,693301,28752.1859.2155.46
Table 6. Results of stem detection with and without the voxel-based features.
Table 6. Results of stem detection with and without the voxel-based features.
Plot IDnrefFeature SetnextrnmatchCorrectness (%)Completeness (%)IoU (%)
1287BF32225077.6487.1169.64
BF-VF31724878.5386.5670.00
2386BF47835273.6491.2068.75
BF-VF46235376.5791.4171.38
3480BF53739273.0081.6762.72
BF-VF54539873.0182.8363.41
Table 7. The accuracies of stem point extraction and stem detection for different optimal feature combination in three plots.
Table 7. The accuracies of stem point extraction and stem detection for different optimal feature combination in three plots.
Plot IDFeature SetsFeature NumberStem Point Extraction (%)Stem Detection (%)
PrecisionRecallF1-ScoreCorrectnessCompletenessIoU
1Proposed method969.0191.6878.7280.9185.6171.23
Point-based method2674.6692.0682.4477.9686.3169.39
2Proposed method1257.2876.8865.5581.5187.4572.86
Point-based method2265.2972.3768.5775.5891.3070.50
3Proposed method1550.0862.7955.7174.1482.7564.20
Point-based method1951.2163.2056.5673.5382.3763.55
Table 8. The accuracies of stem point extraction and stem detection for final feature set Ffinal in three plots.
Table 8. The accuracies of stem point extraction and stem detection for final feature set Ffinal in three plots.
Plot IDFeature NumberStem Point Extraction (%)Stem Detection (%)
PrecisionRecallF1-ScoreCorrectnessCompletenessIoU
116 (Ffinal)70.1991.7979.5380.2085.9670.85
261.8973.2066.9777.9090.0171.61
352.0861.3856.3374.0482.4063.94
Table 9. Comparison of different feature combinations.
Table 9. Comparison of different feature combinations.
Plot IDFeature SetsFeature NumberStem Point Extraction (%)Stem Detection (%)
PrecisionRecallF1-ScoreCorrectnessCompletenessIoU
1Proposed method1681.1480.9781.0590.0091.5383.08
Featured in [28]980.1378.0079.0590.9188.9881.71
Featured in [31]3083.2883.8283.5590.0091.5383.08
Featured in [33]662.1065.8463.9284.9176.2767.16
Featured in [38]2381.2778.8680.0588.9892.3782.89
2Proposed method1666.4789.0976.1481.4586.0171.91
Featured in [28]965.5988.9675.5178.1680.1565.49
Featured in [31]3070.9689.5979.1980.9586.5171.88
Featured in [33]653.2675.9462.6178.6863.8754.45
Featured in [38]2369.0791.9178.8782.2784.9971.83
3Proposed method1650.8861.9455.8776.1884.1066.59
Featured in [28]947.7660.1353.2478.3974.3161.68
Featured in [31]3061.4876.5368.1877.4382.8766.75
Featured in [33]642.5862.7950.7582.7261.4754.47
Featured in [38]2347.8459.8353.1776.5582.8766.10
Table 10. The stem detection accuracy using single-scan TLS datasets.
Table 10. The stem detection accuracy using single-scan TLS datasets.
Plot ID123456Mean
Plot ComplexityEasyEasyMediumMediumDifficultDifficult
Completeness (%)Zhang et al. [21]80.3957.1446.6234.6213.748.4740.16
Chang et al. [56]88.2063.1052.0044.9022.9014.4047.58
Proposed method88.2477.3868.2470.5159.5429.2466.12
Correctness (%)Zhang et al. [21]97.6292.31100.00100.0090.00100.0096.66
Chang et al. [56]95.7094.6095.1094.6061.0050.7081.95
Proposed method91.8487.8494.3980.8881.2895.8390.89
Mean Accuracy (%)Zhang et al. [21]88.1770.5963.5951.4323.8415.6352.21
Chang et al. [56]90.9075.7068.4060.9033.3022.4058.60
Proposed method90.0082.2879.2275.3468.7244.8173.40
Table 11. The stem detection accuracy using multi-scan TLS datasets.
Table 11. The stem detection accuracy using multi-scan TLS datasets.
Plot ID123456Mean
Plot ComplexityEasyEasyMediumMediumDifficultDifficult
Completeness (%)Wang et al. [2]90.2092.9079.1078.2064.1052.5076.17
Zhang et al. [21]86.2782.1461.4957.6945.8026.2759.94
Chang et al. [56]94.1094.0078.4074.3064.8052.7076.38
Proposed method95.8389.2972.9769.2369.4759.3276.02
Correctness (%)Wang et al. [2]95.8081.3072.2070.1067.2070.9076.25
Zhang et al. [21]97.7895.83100.0097.8393.7598.4197.27
Chang et al. [56]94.1089.7094.3093.5085.8081.6089.83
Proposed method93.8891.4687.8081.8279.1389.1787.21
Mean Accuracy (%)Wang et al. [2]92.9086.7075.5074.0065.6060.3075.83
Zhang et al. [21]91.6788.4676.1572.5861.5441.4771.98
Chang et al. [56]94.1091.9085.6082.8073.9064.1082.07
Proposed method94.8590.3679.7075.0073.9871.2580.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, M.; Liu, X.; Pan, J.; Mu, F.; Zhao, L. Stem Detection from Terrestrial Laser Scanning Data with Features Selected via Stem-Based Evaluation. Forests 2023, 14, 2035. https://doi.org/10.3390/f14102035

AMA Style

Chen M, Liu X, Pan J, Mu F, Zhao L. Stem Detection from Terrestrial Laser Scanning Data with Features Selected via Stem-Based Evaluation. Forests. 2023; 14(10):2035. https://doi.org/10.3390/f14102035

Chicago/Turabian Style

Chen, Maolin, Xiangjiang Liu, Jianping Pan, Fengyun Mu, and Lidu Zhao. 2023. "Stem Detection from Terrestrial Laser Scanning Data with Features Selected via Stem-Based Evaluation" Forests 14, no. 10: 2035. https://doi.org/10.3390/f14102035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop