Next Article in Journal
DC-Link Voltage Control of a Grid-Connected Solar Photovoltaic System for Fault Ride-Through Capability Enhancement
Next Article in Special Issue
Study of the Relationship between Urban Expansion and PM10 Concentration Using Multi-Temporal Spatial Datasets and the Machine Learning Technique: Case Study for Daegu, South Korea
Previous Article in Journal
Recent Advances in Plasmonic Sensor-Based Fiber Optic Probes for Biological Applications
Previous Article in Special Issue
Landslide Susceptibility Mapping Based on Random Forest and Boosted Regression Tree Models, and a Comparison of Their Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Single Point-Based Multilevel Features Fusion and Pyramid Neighborhood Optimization Method for ALS Point Cloud Classification

College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2019, 9(5), 951; https://doi.org/10.3390/app9050951
Submission received: 18 January 2019 / Revised: 19 February 2019 / Accepted: 21 February 2019 / Published: 6 March 2019

Abstract

:
3D point cloud classification has wide applications in the field of scene understanding. Point cloud classification based on points can more accurately segment the boundary region between adjacent objects. In this paper, a point cloud classification algorithm based on a single point multilevel features fusion and pyramid neighborhood optimization are proposed for a Airborne Laser Scanning (ALS) point cloud. First, the proposed algorithm determines the neighborhood region of each point, after which the features of each single point are extracted. For the characteristics of the ALS point cloud, two new feature descriptors are proposed, i.e., a normal angle distribution histogram and latitude sampling histogram. Following this, multilevel features of a single point are constructed by multi-resolution of the point cloud and multi-neighborhood spaces. Next, the features are trained by the Support Vector Machine based on a Gaussian kernel function, and the points are classified by the trained model. Finally, a classification results optimization method based on a multi-scale pyramid neighborhood constructed by a multi-resolution point cloud is used. In the experiment, the algorithm is tested by a public dataset. The experimental results show that the proposed algorithm can effectively classify large-scale ALS point clouds. Compared with the existing algorithms, the proposed algorithm has a better classification performance.

1. Introduction

Airborne Laser Scanning (ALS) can capture large-scale point clouds of urban scenes. The point cloud classification of outdoor scenes can provide high-precision semantic maps for autonomous driving, improve the accuracy of vehicle positioning, and reconstruct a three-dimensional model of the city, which plays an important role in urban planning and dynamic management. In addition, it can improve the efficiency of resource utilization. Effectively labeling the correct class for all points in the scene is an important basis for the widespread adoption of point clouds [1,2,3,4]. However, a laser point cloud has a huge data number, high redundancy, and uneven scene distribution, which may lead to huge challenges in the point cloud classification. Therefore, it is of great significance to classify the three-dimensional point cloud in large outdoor scenes.
Currently, the number of point clouds with manual labeling in outdoor large scenes is not enough. However, machine learning can learn and classify point clouds in the case of less sample training data, and the speed is faster. At present, the point cloud classification methods can be mainly divided into two strategies: the single point-based classification and object-based classification methods.
The point-based point cloud classification is the classification of each individual point in a point cloud; this strategy uses points as the basic unit to extract features, train models, and predict class labels. There are three main steps: the neighborhood selection, feature extraction, and single point classification based on the features and classifiers.
(1) Neighborhood selection. In the neighborhood selection process, the commonly used point cloud neighborhood forms are: K nearest neighbors [5], radius neighborhoods [6], and column neighborhoods [7]. The parameter of neighborhood estimation is highly dependent on prior knowledge, and it is greatly affected by the change of the point cloud density [8,9]. For example, Hackel et al. constructed a multi-level scale pyramid, and a total of 144-dimensional features such as eigenvalues of the covariance matrix were extracted for each point in each pyramid. Subsequently, Random Forest was used for training and finally classified outdoor road scenes [10], and a better classification effect was obtained. Therefore, the selection based on multi-scale neighborhood is an important method for extracting single point effective features.
(2) Features extraction. The local feature of a point cloud is an abstract depiction of the environment around a given point. It is difficult to classify a point cloud by a single local feature. The common practice is to fuse multiple point cloud local features for classification. Normal and curvature are simple local features that clearly show some local information about the point cloud, such as the fact that the normal direction can represent the partial tangent plane of the point cloud, and the curvature can represent the smoothness of the point. For example, Fanxuan et al. [11] optimized the matching accuracy of point pairs in point clouds based on the curvature information, and the registration accuracy of point clouds was further improved. Geometric features are also common local features, also known as shape descriptors. For example, in the spin image [12], the main idea is to set up an image with the normal vector as the center; the image rotates around the axis. The number of 3D points encountered by each pixel in the point cloud is taken as its gray value. Finally, a two-dimensional array representing the local information of the three-dimensional space, that is, the rotated image feature, is obtained. The 3D Shape Context (3DSC) feature [13] is based on the specified point to construct a spherical region. In the support region, the grid is divided into three coordinate directions: the radial direction, direction angle and elevation angle. Following this, a feature histogram is constructed by entering the number of points in the grid. The 3DSC is simple in construction, strong in discrimination and insensitive to noise, but it is time-consuming. The Unique Shape Context (USC) descriptor [14] improves the 3DSC to avoid ambiguities in the classification. Point Feature Histograms (PFH) [15] are local features, which construct a feature histogram with the angles and distances of the normal vectors of any two points in the specified point neighborhood. The descriptor can accurately describe the local features of the points, but the computation is large and the real-time performance is poor. Fast Point Feature Histograms (FPFH) [16] are a simplification of PFH, which greatly reduce the time consumption while retaining most of the description performance of PFH. FPFH have an excellent performance, and are widely used in the field of point cloud classification, segmentation and registration [17]. Although these features can express the local features of the point cloud, they do not take into account the characteristics of the ALS point cloud, which has the characteristics of relative sparse, rich elevation information, as well as a horizontal and vertical distribution.
(3) Single point classification based on features and classifiers. Currently, machine learning is an important method for classification problems. The single-point classification based on machine learning takes the feature vector of the point as the input and the class label of the single point as the output. Common machine learning algorithms can accomplish this classification task, such as AdaBoost [18], Random Forest [19] and Support Vector Machine (SVM) [20]. This kind of method uses a classifier to learn the local features of each point, after which the parameters in the classifier are determined based on the training dataset. Finally, the test set is classified by the classifier. This classification strategy can more accurately segment the boundary regions between different adjacent objects, and this method has a better performance in detail. However, due to the extremely large number of points, the calculation amount is large. Thus, the model training is slow, and there are also some misclassifications of local regions. However, there are always some errors in the final classification results. Therefore, the initial classification results are required to further optimize according to the characteristics of the point cloud.
In order to solve the above-mentioned problems, an ALS point cloud classification algorithm based on a single point multi-feature fusion is proposed. This kind of algorithm is based on the point as the basic processing unit, and the classification process assigns labels to each point in the point cloud to realize a point cloud classification. The proposed method extracts the local features of each point by constructing a multi-scale neighborhood space, along with two new features: a normal angle distribution histogram (NAD) and latitude sampling histogram (LSH) are proposed. Following this, SVM is used for training and classification. However, since each point is classified, there is a problem regarding some edge points being misclassified. In this regard, the initial classification results are further optimized according to the neighborhood classification optimization of multi-scale pyramids. Experiments prove that the classification algorithm has a higher accuracy.
The main contributions of this paper are as follows:
(1) Two local features are proposed, that is, the NAD histogram and the LSH histogram. The differences of different objects in the normal distribution, and the difference of the neighborhood points around different objects in the horizontal and vertical directions of the three-dimensional space, can be fully utilized to more effectively represent the characteristics of different objects.
(2) A multilevel single-point features fusion method based on a multi-neighborhood space and multi-resolution is proposed. The multi-scale space is constructed by changing the resolution of the point cloud and the number of the neighborhood. The features of the multi-scale are extracted from each single point, and the features are fused. Following this, the SVM classifier is used to classify the features and the better classification results have also been achieved.
(3) A fast optimization method for classification results based on a multi-scale pyramid is proposed. By changing the resolution of the point cloud, a multi-scale pyramid is constructed, and the neighbor points are further re-selected. After this, the misclassifications are eliminated according to the initial classification results of the neighbors for a post-processing optimization.

2. Method

As shown in Figure 1, the algorithm flow is given as follows. In the training part, for the point cloud scene shown in Figure 1a, multiple features of each point are first extracted. Multi-scale and multiple features are fused to a fusion feature. The SVM classifier model is then trained using the fusion features of the training set. In the test part, for the point cloud scene shown in Figure 1b, the fusion feature is first obtained. As shown in Figure 1c, the test points are initially classified using the trained SVM classification model. Following this, the point clouds of different resolutions are obtained by down-sampling the point cloud, in order to construct a multi-scale point cloud pyramid. The corresponding classification labels of the different neighboring points in the different scales are searched. Finally, the label which has the most number in the neighbors is taken as the class label of the current point. The final point cloud classification result is obtained, as shown in Figure 1d.

2.1. Point Feature Extraction

In the classification task of the 3D point cloud data, the feature extraction of the point cloud plays a crucial role. It can seriously affect the final classification result. A well-behaved feature descriptor should reflect obvious differences between different types of points in the point cloud. At the same time, the descriptor should be robust and have a strong anti-interference ability. It is difficult for a single feature descriptor to have the above characteristics, so that a plurality of feature descriptor fusion methods are at present widely used. In the single point-based classification algorithm, this paper uses a variety of feature fusion methods to improve the accuracy of the classification algorithm. The specific features are as follows:

2.1.1. Feature Description

1.
Elevation feature
The height is a very intuitive feature in a 3D point cloud. Generally speaking, points with large height are buildings, trees or objects with larger elevation values in the real world. When the elevation value is small, the probability is greater if the point is a vehicle point. Thus, the elevation feature is set to:
F z = [ Z i ,   1 Z i ]
where Z i is the distance of the i-th point from the estimated ground to the elevation value.
2.
Normal angle distribution histogram
In the large scale scene, the normal direction of different objects has obvious differences. For artificial objects, such as buildings and vehicles, since the surface is relatively regular, almost all points are in the same direction, pointing in the direction of the vertical plane. However, due to the scattered distribution of the whole point cloud, the normal direction of the point cloud has a large scattered nature, and the direction does not point to a fixed direction in a uniform way. Therefore, we calculate the histogram of each point and its own normal angle distribution value in the local neighborhood point set to express the relationship between the normal of the point and the normal of the points in the neighboring region. The angle between the two normal vectors in three-dimensional space should be between [ 0 , π ] . But considering that the normal of the point on the plane can have opposite directions when the angle is larger than π / 2 , the corresponding angle is set to π π / 2 . Following this, the angle of the normal vectors is defined as [ 0 , π / 2 ] . Considering efficiency and resolving power, we divide this interval into equal D n parts, that is, construct a D n dimensional histogram. After this, the number of points falling within each cell is taken as the value of the interval in the histogram. Finally, the normalization process is performed to form a histogram of the normal angle distribution, called NAD. This feature can distinguish different classes of points based on the normal angular distributions. The specific calculation formulas are as follows:
= π / 2 0 D n
θ j = a cos ( v v j )
h ( x i ) = n ( i Δ θ j ( i + 1 ) Δ ) N   ( i = 1 , ,   D n )
F N A D = [ h ( x 1 ) , h ( x 2 ) , , h ( x D n ) ]
where v and v j represent the normal vectors of the current point and the j-th neighbor point, respectively. a cos ( ) represents the inverse cosine function. N represents the number of neighbors for the current point. n ( i Δ θ j ( i + 1 ) Δ ) denotes the number of points for the normal angle at the range [ i Δ , ( i + 1 ) Δ ] . F N A D denotes the final normal angle eigenvalue vector of the normal angle distribution. The histogram of the normal angle distributions for the randomly selected building points and tree points are shown in Figure 2.
3.
Latitudinal sampling histogram
In the outdoor large scene environment, as for almost all points belonging to different objects, the surrounding neighborhood points have great differences in the latitudinal distribution in the three-dimensional space. For example, a building surface, which is parallel to the ground, has its neighborhood points mainly distributed near the “equator”. For the points belonging to the trees, the distribution of the neighborhood points is more random and extensive, hardly concentrated in a certain latitude interval. Therefore, the selected point is regarded as the center of the sphere, and the distribution histogram of the neighborhood points in the latitudinal direction is counted. Following this, the feature of the point can be expressed. The feature is called LSH. The LSH feature can be used to distinguish different classes of points according to the distribution of neighborhood points in the latitude direction. The LSH has the advantages of anti-occlusion, without interference from the local coordinate system, as well as high efficiency. In this paper, Dl spaces are equally divided along the latitude direction. Following this, the number of points falling into each cell is counted to form a feature vector of the Dl dimension. The specific calculation formulas are:
Δ = π 0 D l
θ j = a cos ( z ( p j p ) )
f ( x i ) = n ( i Δ θ j ( i + 1 ) Δ ) N   ( i = 1 , , D l )
F L S H = [ f ( x 1 ) , f ( x 2 ) , , f ( x D l ) ]
where p and p j represent the three-dimensional coordinates of the current point and its j-th neighbor point, respectively. z = ( 0 , 0 , 1 ) represents the unit vector of the positive direction of the z axis. n ( i Δ θ j ( i + 1 ) Δ ) represents the number of points in [ i Δ , ( i + 1 ) Δ ] of the neighborhood points along the latitudinal direction. F L S H represents the final feature vector of LSH. The LSHs of the randomly selected building points and tree points are compared, as shown in Figure 3.
4.
Convariance feature
First, a covariance matrix for the selected point neighborhood is constructed. After this, eigenvalues of the covariance matrix are calculated as: λ 2 λ 1 λ 0 0 , and the corresponding eigenvectors are calculated as: v 2 , v 1 , v 0 . Here, the covariance feature (CF) is obtained according to the relationship among the eigenvalues, as follows:
Sum of eigenvalues:
F s u m = λ 1 + λ 2 + λ 3
Full variance:
F o m n = ( λ 1 λ 2 λ 3 ) 1 3
Anisotropy:
F a n i = ( λ 1 λ 3 ) / λ 1
Planarity:
F p l a = ( λ 2 λ 3 ) / λ 1
Linearity:
F l i n = ( λ 1 λ 2 ) / λ 1
Sphericity:
F s p h = λ 3 / λ 1
Following this, the final total covariance feature is: F cov = [ F s u m , F o m n , F a n i , F p l a , F l i n , F s p h ] .
5.
Plane point ratio
In outdoor large scale scenes, the classes of objects are complex and the surface shapes are also different. However, a considerable part of the surface of artificial objects exhibits planar characteristics, such as buildings, vehicles, etc. Meanwhile vegetation does not have planar characteristics, so the plane point ratio of the local point cloud can be used as a local feature to classify point clouds. The covariance feature can also reflect the planar characteristics to a certain extent, but it is greatly interfered by noise. For this reason, the Random Sample Consensus (RANSAC) [21] is employed to fit the local neighborhood of the selected point. After this, the ratio of the plane points, called PPR (Plane Point Ratio), is calculated.
RANSAC is a method used to find the subset of data that is the best match for the model from the data set with random samples that are noisy but sufficient. The points matched with the model are called the inner points, and the points unmatched with the model are called the outer points. The plane is fitted using RANSAC as follows.
(1) Select three points randomly from all the neighborhood points and calculate the current model parameters. The model is as follows:
a x + b y + c z + d = 0
(2) Determine whether each point is an inlier, and then determine the inlier rate ω of the current model:
J i = { 1 , d i T d 0 , d i > T d
ω = 1 N i = 1 N J i
where di is the distance from the i-th point to the plane. Td is a fixed threshold. Ji indicates whether it is an inlier or not. N is the number of neighborhood points.
(3) If the current inlier rate is larger than the previous optimal inlier rate, the optimal inlier rate is updated.
(4) To find the optimal model, repeat steps (1) to (3) k times until the probability reaches P:
1 P ( 1 ω 3 ) k
The termination condition is:
k log ( 1 P ) log ( 1 ω 3 )
When the RANSAC iteration is completed, the optimal inlier rate is the ratio of the plane points: F p l a n e = [ ω ] .

2.1.2. Single Point Multi-Scale Multi-Feature Fusion

Since the features of the single point are dependent on the selected neighborhood space, different neighborhood spaces have different expression capabilities for different classes of point clouds. Additionally, the structure descriptions of point clouds with different resolutions also have certain differences. The local feature description of the single scale for a point is relatively single, and there are some noise points in the point cloud, which can make the simple feature of the single scale unable to accurately describe the feature of the single point. Therefore, a multilevel features fusion method based on the multi-neighborhood space and multi-resolution is proposed. As shown in Figure 1c, the proposed method constructs the multi-scale space by changing the resolution and the number of neighborhoods of the point cloud. Following this, multi-scale features for each single point in the point cloud are extracted. Because the elevation feature is not affected by the scale changing, we select NAD, LSH, CF and PPR features to construct the multi-scale features. We extract the features of a single point in each scale by choosing µ neighborhoods with different resolutions and υ different neighborhood sizes under the original resolution. The multi-scale features of each point are expressed respectively as F NAD ,   F LSH ,   F cov ,   and   F plane . Considering the validity of the features and the efficiency of the calculation, this generally results in 2 µ + υ 5. In addition, the description of the single point feature only represents one characteristic of the point cloud. Therefore, it is necessary to fuse multiple features. After fusing the features, the multilevel features are expressed as follows:
F = [ F z , F N A D , F L S H , F c o v , F p l a n e ]
Because we aim at an ALS point cloud, the extracted elevation features are only two-dimensional and play an important role in the point cloud classification. In addition, when the point cloud features are extracted, the values of each feature have been normalized to [0, 1]. In order to reflect the role of the non-zero feature value, the feature should be normalized again according to Formula (22) when the extracted feature F is sparse. While the extracted feature F is not sparse, there is no need to normalize the feature. Therefore, the constructed feature is X = [ F z * , F N A D * , F L S H * , F c o v * , F p l a n e * ] .
F i , j * = F i , j min ( F j ) max ( F j ) min ( F j )
where F i , j * is the value of the i-th row and the j-th column in the normalized feature matrix F*. F i , j is the value of the i-th row and the j-th column in the feature matrix F. F j is the vector of the j-th column (for all the points) in the feature matrix F.

2.2. Point Cloud Classification Based on SVM

SVM [22] is achieved by maximizing the classification interval in the feature space. For non-linear data, SVM maps them into a high-dimensional feature space by a kernel function, which make the data into linear separable data in a high-dimensional feature space. Following this, it realizes a classification by maximizing the interval. In view of the excellent generalization ability of SVM, we use SVM as a classifier for the single point classification in point cloud data. As we know, the correlation between the point cloud single point feature and neighbor points features, and the Gauss kernel function only has one parameter σ and a low model complexity. Thus, in the absence of prior knowledge, the Gauss kernel function is often better than other kernels. Therefore, we choose the Gauss kernel function as the kernel function. Here, the Gauss kernel function of the SVM classifier is defined as follows:
The fused feature space is X . The selected n d-dimensional feature samples { x 1 , x 2 , , x n } :
x i = ( x i 1 , x i 2 , , x i d ) ϵ d
After the feature transformation, the feature space is Z. We map data in the X space to the Z space z i = ( z i 1 , z i 2 , , z i d ) ϵ d   via the mapping function ϕ ( x ) . The function K ( x , z ) satisfies the condition K ( x , z ) = ϕ ( x ) ϕ ( z ) , and the function K ( x , z ) is a kernel function, while ϕ ( x ) is a mapping function. The Gauss kernel function is as follows:
K ( X , z ) = i n K ( x i , z ) = i n exp ( x i z 2 2 σ )
The corresponding decision function is:
f ( z , α * , b * ) = s i g n ( i n y i α * exp ( x i z 2 2 σ ) + b * )
The SVM classifier is trained by the features of the training set, and the test set is classified by the trained classifier. The initial classification results for the point cloud in Figure 1b are shown in Figure 1d.

2.3. Neighborhood Optimization Based on Multi-Scale Pyramid

After the initial classification, the point clouds are basically classified correctly. Due to noise and other reasons, there are still some misclassifications in some details (such as edges). As shown in Figure 1d, most of the points in the scene have been correctly classified, and only a small part of them are misclassified. They mainly concentrate on edges and other places, and most of the points around the misclassified points are correctly classified. Therefore, it is necessary to further optimize the initial classification results to achieve a more accurate classification of the point clouds. Because local information is used as a feature to classify point clouds, the feature extraction relies heavily on a local region selection. In addition, the single point is taken as the basic unit of classification. Each point has its own characteristics, but because the two neighboring points are very close to each other and their neighborhoods are also very close, the extracted features will be very similar, which leads them to be more likely to be classified into the same class. Therefore, the neighbors of the misclassified points are also often misclassified. It is difficult to correct the misclassified points if only the points in the smaller local regions are used for the optimization. Therefore, we propose a classification results optimization method based on the multi-scale pyramid. The specific method is as follows:
First, voxel filters with different radius scales are used to down-sample the point cloud after an initial classification, as shown in Formula (24). Each minimum voxel scale is twice as large as the last down-sampling, and sparse point clouds are gradually obtained. Following this, the q-level pyramid is constructed, and the initial classes of all the points in each level are retained. According to the characteristics of the point cloud down-sampling reflecting the structure information of the shape, the scale pyramid is constructed on three scales of q = 3 in this paper.
Following this, the corresponding k-d tree is constructed from the point cloud in each layer of the pyramid. For each point in the original point cloud, a k-d tree is used to search for the radius of the nearest neighbors in the point cloud after the down-sampling. The class labels of the m point clouds searched within the radius of the l-th level are L l = { L 1 l , , L m l } , L i l { 1 , , c } , i = 1,…, m, l = 1,.., q; the radius parameters are different when each layer of the point cloud chooses its nearest neighbor. The method of calculation is as follows:
r = k P r e s o l u t i o n
In the formula, r represents the scale radius parameter. P r e s o l u t i o n is the resolution used by the current down-sampling point cloud. k is a fixed ratio threshold.
Finally, the initial labels of all the nearest neighbors in the q levels are counted. The discriminant function 1 { L i l = C } represents the fact that when   L i l belongs to class C, its value is 1; otherwise, its value is 0. This is used to count the number of the initial labels belonging to each class. As shown in Formula (27), the mode label C * is selected as the new class label for the current point.
C * = arg max   l = 1 q i m 1 { L i l = C }   C = L i l { 1 , , c }  
Not only do the optimized point cloud classification results avoid a situation where the nearest neighbor is also misclassified, but they also solve the problem of too many far points in a large scale, thus achieving better results. The optimized point cloud classification results are shown in Figure 1e.

3. Experimental Results and Analysis

In order to verify the performance of the point cloud classification algorithm based on single point multilevel features, we use two urban scenes’ ALS data for a qualitative and quantitative comparison and analysis. This section begins by briefly introducing the experimental data, before the classification performance of the proposed method is compared with the other methods on the datasets.

3.1. Experimental Dataset

In this paper, we use two sets of dataset published in Ref [23]. The data was collected in Tianjin, China. The density of the test region point cloud is about 20 ~ 30 / m 2 . The data set contains both large objects (buildings and trees) and small objects (cars). It contains different roof shapes, buildings of different heights, and dense and overlapping cars and trees. Table 1 lists the number of each class point for the two scenes. Figure 4a,b shows the training data of scene 1 and scene 2, respectively.
All the programs are run on a computer with an Intel Core i7-7700K CPU, 4.20 GHz with 24-GB RAM. The algorithm is implemented on the C++ platform based on PCL 1.8.0. Each set of data training and testing takes about 6.5 min. However, the feature extraction and optimization process can be implemented in parallel. Therefore, the efficiency of the proposed method can be further improved, and the speed of the point cloud processing will be greatly improved.

3.2. Experimental Comparison and Analysis

In order to verify the performance of the proposed algorithm, we compare it with the other seven methods shown in Table 2. Method 1 is the proposed method, which uses the features without NAD and LSH. Method 2 uses the Fz+FNAD+FLSH+Fcov+Fplane feature fusion in single-scale without a post-processing optimization. Method 3 uses the single-scale feature and multi-scale pyramid optimization. Method 4 is the algorithm proposed in [24]. In this method, each feature uses geometry, strength, and statistics information. Following this, the JointBoost method is used to select features and classify points. Method 5 is the classification method based on the multi-scale Spin Image feature and Fcov feature fusion used in Ref [23]. Method 6 [25] constructs a multilevel point set using a linear transformation, and it uses Spin Image and Fcov features. Following this, K-means is used to construct an LDA (Latent Dirichlet Allocation) model of a multilevel point set dictionary. Method 7 [23,26,27] constructs a multilevel point set using an exponential transformation. The Spin Image and Fcov features are used for dictionary learning, constructing an LDA model of point sets.
Accuracy, precision, and recall are often used to evaluate the effect of a point cloud classification [27]. The precision rate is the proportion of true positive samples in a positively predicted sample. The recall rate is the proportion of positive samples that are predicted to be successful in the actual positive samples. The accuracy rate is the ratio of all the correctly predicted samples in relation to the overall samples. In order to consider both Pa (precision rate) and R (recall rate), F1-score values (such as Equation (28)) are generally used to represent the classification quality of the scene. In order to better evaluate the effects of each algorithm, we use the above metrics to evaluate the classification performance.
F 1 s c o r e = 2 ( R × P a ) R + P a
The classification results of our method and of other comparison methods on scene 1 and scene 2 are shown in Table 3. Table 3 lists the precision, recall, accuracy and F1-score statistics for the eight methods in the two scenes. It can be seen from Table 3 that the comparison between the proposed method and Method 1 shows that the accuracy of the proposed method has significantly improved. It also shows that the proposed NAD and LSH features have certain effects. By comparing Method 2 and Method 3, the proposed multi-scale pyramid optimization algorithm can effectively improve the classification accuracy. Comparing the proposed method with Method 3, the proposed multi-scale strategy has a significant effect on the improvement of the classification results.
In addition, the proposed method is compared with the methods given in other references. Method 4 and Method 5 classify the point cloud based on the single-point. Method 6 and Method 7 classify the point cloud based on the point set (object). It can be seen from the comparison of Table 3 that the proposed method has a high accuracy rate as a whole, and that it also maintains a high recall rate.
For scene 1, the accuracy of the proposed method is the highest, and the value of the precision/recall is relatively high (the classification result of the proposed method is shown in Figure 1e). From the classification results evaluation of the three kinds of objects by the F1-score, one can see that the classification effect on cars for the proposed method is not as good as for Method 7. Meanwhile, the tree and building classes can basically be classified correctly.
For scene 2, the precision/recall of trees of the proposed method are lower than for Method 7. Meanwhile, the precision/recall of buildings and cars are the maximum compared with other methods. According to the classification result of the F1-score value, the proposed method has a better effect than the other methods, except that the tree classification performance is worse than for Method 7. Considering the accuracy, precision/recall rate and the F1-score evaluation comprehensively, the proposed method has a better classification performance than the other methods, and the proposed method has the highest overall accuracy for both scenes. This proves that the proposed point cloud classification method based on point multilevel features is effective, and that it can accurately classify the ALS point cloud data in large scale scenes.
In order to more intuitively compare the classification performance of each method, Figure 5 shows the performance of the eight classification methods in scene 2. Figure 5 shows that the proposed method can classify most points correctly. Compared with other comparison methods, the classification accuracy of the proposed method is higher. From the comparison between Figure 5c–e and Figure 5b, the classification effect of the proposed method is obviously better than that of the other three algorithms, especially in the buildings and trees. It can be seen from the comparison between Figure 5f,g and Figure 5b that Method 4 and Method 5 have more misclassifications for cars and buildings, and that the performance of the proposed method is significantly better than that of the other two methods. In comparing Figure 5h,i with Figure 5b, one can see that Method 6 and Method 7 have a similar classification performance to that of the proposed method. However, a certain number of architectural edge points are classified incorrectly, and the top part of the trees is also classified incorrectly. One can see from the comparison between Figure 5f–i and Figure 5b that the compared methods have some misclassifications for the edge points and for two objects that are overlapping regions. However, the proposed method has fewer misclassifications in those regions than for the other methods. This proves that the proposed feature descriptors and post-optimization strategies can improve the classification results.

3.3. Sensitivity of the Parameters

In this part, we focus on the D m in NAD, D l in LSH and the number of neighborhood scales (µ + υ ). Here, we select the parameters shown in Table 4; the data of scene 1 is used to compare the influence of different parameters on the proposed method. In order to evaluate the classification effect of the three kinds of objects as a whole, we average the F1–score values of the three object classification results to obtain the mean value mF1, which is used as the overall classification effect evaluation metric. As shown in Table 4, considering the results of mF1 and the accuracy in combination, in comparing the parameters of the first three rows one can see that when D m is 15, the classification effect is better; however, the value of D m is not sensitive to the classification effect. According to the results of rows 2, 4 and 5, the classification effect of D l is obviously improved at 15. However, when the value of D l is too large, the classification effect is reduced. Therefore, the value of D l is relatively sensitive to the classification result. According to the results of rows 4, 6, 7 and 8–10, the classification effect is improved when the value of the scale µ + υ is increased. However, when the scale exceeds 4, the classification effect will be reduced to some extent. Therefore, the value of the scale is sensitive to the results of the classification and needs to be within a reasonable range. One can see from Table 4 that the tree and building classes are relatively less affected by the changes of the D m and D l , and that the results are susceptible to the size of the scale. The increasing values of D m and D l would likely cause a change in the car classification effect. Considering the overall effect of the classification and the factors of accuracy and feature dimension, we select D m = 15, D l = 15 and (µ + υ ) = 3 as the optimal parameter values.

4. Conclusions

The classification of the ALS point cloud is an important technology for urban planning, digital city and intelligent transportation. We propose a multilevel features fusion and pyramid neighborhood optimization ALS point cloud classification method based on a single point. The proposed method presents two local features, i.e., the NAD and LSH. They are fused with the covariance and elevation features. Following this, the multilevel features are constructed by changing the point cloud resolution and the neighborhood size. The fused features are used to train a classification model based on the Gaussian kernel function SVM for an initial classification. Finally, the point cloud classification is optimized based on the initial classification result using a multi-scale pyramid. The optimized classification results have a higher accuracy. The experimental results prove the effectiveness of the proposed method via the experiments on the two sets of public ALS datasets.

Author Contributions

Conceptualization, Y.L. and X.D.; methodology, Y.L. and X.D.; software, Y.L. and X.D.; validation, G.T., Y.L. and X.D.; data curation, G.T., J.Z. and X.D.; writing—original draft preparation, Y.L. and X.Y.; writing—review and editing, Y.L., L.Y. and X.Y.; visualization, L.Y. and X.Y.; supervision, G.T.; project administration, G.T.; funding acquisition, G.T.

Funding

This research was funded by the National Natural Science Foundation of China (No. 61175031), the National High Technology Research and Development Program of China (863 Program) (No. 2012AA041402), the National Key Technology Research and Development Program of the Ministry of Science and Technology of China (No. 2015BAF13B00-5). The APC was funded by Guofeng Tong.

Acknowledgments

The authors would like to thank Lihao Cao in Northeastern University for helping to check the grammar and spelling of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maturana, D.; Chou, P.-W.; Uenoyama, M.; Scherer, S. Real-time semantic mapping for autonomous off-road navigation. In Field and Service Robotics; Hutter, M., Siegwart, R., Eds.; Springer: Cham, Germany, 2018; pp. 335–350. ISBN 978-3-319-67361-5. [Google Scholar]
  2. Li, Y.; Tong, G.F.; Yang, J.C.; Zhang, L.Q.; Peng, H.; Gao, H.S. A Summary of Key Technologies for 3D Point Cloud Scene Data Acquisition and Scene Understanding. Laser Optoelectron. Prog. 2019, 56, 040002. [Google Scholar]
  3. Yuan, L.W.; Yu, Z.Y.; Luo, W.; Hu, Y.; Feng, L.Y.; Zhu, A.-X. A hierarchical tensor-based approach to compressing, updating and querying geospatial data. IEEE Trans. Knowl. Data Eng. 2015, 27, 312–325. [Google Scholar] [CrossRef]
  4. Chen, D.; Zhang, L.Q.; Mathiopoulos, P.T.; Huang, X.F. A methodology for automated segmentation and reconstruction of urban 3-D buildings from ALS point clouds. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 4199–4217. [Google Scholar] [CrossRef]
  5. Linsen, L.; Prautzsch, H. Local versus global triangulations. In Proceedings of the 22th Annual Conference of the European Association for Computer Graphics, EUROGRAPHICS 2001, Manchester, UK, 5–7 September 2001. [Google Scholar]
  6. Lee, I.; Schenk, T. Perceptual organization of 3D surface points. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2002, 34, 193–198. [Google Scholar]
  7. Filins, S.; Pfeifer, N. Neighborhood systems for airborne laser data. Photgramm. Eng. Remote Sens. 2005, 71, 743–755. [Google Scholar] [CrossRef]
  8. He, E.; Chen, Q.; Wang, H.; Liu, X. A curvature based adaptive neighborhood for individual point cloud classification. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 219–225. [Google Scholar] [CrossRef]
  9. Weinmann, M.; Jutzi, B.; Hinz, S.; Mallet, C. Semantic point cloud interpretation based on optimal neighborhoods, relevant features and efficient classifiers. ISPRS J. Photogramm. Remote Sens. 2015, 105, 286–304. [Google Scholar] [CrossRef]
  10. Hackel, T.; Wegner, J.D.; Schindler, K. Fast Semantic Segmentation of 3D Point Clouds with Strongly Varying Density. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. 2016, III, 177–184. [Google Scholar] [CrossRef]
  11. Zeng, F.X.; Li, L.; Diao, X.P. Iterative closest point algorithm registration based on curvature features. Laser Optoelectron. Prog. 2017, 54. [Google Scholar] [CrossRef]
  12. Johnson, A.E.; Hebert, M. Using spin images for efficient object recognition in cluttered 3D scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef] [Green Version]
  13. Frome, A.; Huber, D.; Kolluri, R.; Bülow, T.; Malik, J. Recognizing objects in range data using regional point descriptors. Eur. Conf. Comput. Vis. 2004, 3023, 224–237. [Google Scholar] [CrossRef]
  14. Tombari, F.; Salti, S.; Di Stefano, L. Unique shape context for 3D data description. In Proceedings of the International Workshop on 3D Object Retrieval (3DOR 10)—In Conjunction with ACM Multimedia, Firenze, Italy, 25 October 2010; pp. 57–62. [Google Scholar]
  15. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the 21st IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2008), Nice, France, 22–26 September 2008. [Google Scholar]
  16. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar] [CrossRef]
  17. Jeong, J.; Lee, I. Classification of LIDAR Data for Generating a High-Precision Roadway Map. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, XLI-B3, 251–254. [Google Scholar] [CrossRef]
  18. Lodha, S.K.; Fitzpatrick, D.M.; Helmbold, D.P. Aerial lidar data classification using AdaBoost. In Proceedings of the International Conference on 3-D Digital Imaging and Modeling (3DIM), Montreal, QC, Canada, 21–23 August 2007; pp. 435–442. [Google Scholar] [CrossRef]
  19. Babahajiani, P.; Fan, L.; KÄMÄRÄINEN, J.K.; Gabbouj, M. Urban 3D segmentation and modelling from street view images and LiDAR point clouds. Mach. Vis. Appl. 2017, 28, 679–694. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, J.X.; Lin, X.G.; Ning, X.G. SVM-based classification of segmented airborne lidar point clouds in urban areas. Remote Sens. 2012, 5, 3749–3775. [Google Scholar] [CrossRef]
  21. Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM. 1981, 24, 381–395. [Google Scholar] [CrossRef]
  22. Chang, C.-C.; Lin, C.-J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  23. Zhang, Z.X.; Zhang, L.Q.; Tong, X.H.; Wang, Z.; Guo, B.; Huang, X.F.; Wang, Y.B. A Multi-Level Point Cluster-based Discriminative Feature for ALS Point Cloud Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4714–4726. [Google Scholar]
  24. Guo, B.; Huang, X.; Zhang, F.; Sohn, G. Classification of airborne laser scanning data using JointBoost. ISPRS J. Photogramm. Remote Sens. 2015, 100, 71–83. [Google Scholar] [CrossRef]
  25. Wang, Z.; Zhang, L.; Fang, T.; Mathiopoulos, P.T.; Tong, X.; Qu, H.; Xiao, Z.; Li, F.; Chen, D. A Multiscale and Hierarchical Feature Extraction Method for Terrestrial Laser Scanning Point Cloud Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2409–2425. [Google Scholar] [CrossRef]
  26. Zhang, Z.; Zhang, L.; Tong, X.; Guo, B.; Zhang, L.; Xing, X. Discriminative dictionary learning-based multi-level point-cluster features for ALS point cloud classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7309–7322. [Google Scholar] [CrossRef]
  27. Zhang, Z.X. ALS Point Cloud Classification Based on Multilevel Point Cluster Features. Ph.D. Thesis, Beijing Normal University, Beijing, China, 2017. [Google Scholar]
Figure 1. Flowchart of the proposed method. NAD: normal angle distribution histogram;LSH: latitude sampling histogram; SVM: Support Vector Machine.
Figure 1. Flowchart of the proposed method. NAD: normal angle distribution histogram;LSH: latitude sampling histogram; SVM: Support Vector Machine.
Applsci 09 00951 g001
Figure 2. Normal angle distribution histogram. (a) Normal angle distribution histogram of a building point. (b) Normal angle distribution histogram of a tree point.
Figure 2. Normal angle distribution histogram. (a) Normal angle distribution histogram of a building point. (b) Normal angle distribution histogram of a tree point.
Applsci 09 00951 g002
Figure 3. Latitudinal sampling histogram. (a) Latitudinal sampling histogram of a building point. (b) Latitudinal sampling histogram of a tree point.
Figure 3. Latitudinal sampling histogram. (a) Latitudinal sampling histogram of a building point. (b) Latitudinal sampling histogram of a tree point.
Applsci 09 00951 g003
Figure 4. The training data of the ALS points. (a) scene 1, and (b) scene 2. (The figures are captured from the ALS points shown in Cloudcompare (http://www.cloudcompare.org/). The red points are cars, green points are buildings and blue points are trees.)
Figure 4. The training data of the ALS points. (a) scene 1, and (b) scene 2. (The figures are captured from the ALS points shown in Cloudcompare (http://www.cloudcompare.org/). The red points are cars, green points are buildings and blue points are trees.)
Applsci 09 00951 g004
Figure 5. Scene 2 classification results. (a) ground truth, (b) proposed method, (c) method 1, (d) method 2, (e) method 3, (f) method 4, (g) method 5, (h) method 6, and (i) method 7. ((fi) are taken from Ref [23]. The red points are cars, green points are buildings and blue points are trees.)
Figure 5. Scene 2 classification results. (a) ground truth, (b) proposed method, (c) method 1, (d) method 2, (e) method 3, (f) method 4, (g) method 5, (h) method 6, and (i) method 7. ((fi) are taken from Ref [23]. The red points are cars, green points are buildings and blue points are trees.)
Applsci 09 00951 g005
Table 1. The experimental dataset.
Table 1. The experimental dataset.
SceneTraining PointsTest Points
Tree BuildingCarTreeBuildingCar
Scene 168,80237,1285380213,990200,5497816
Scene 239,74364,952458473,207156,1867409
Table 2. Main features of the proposed method and other comparison methods. SVM: Support Vector Machine; LDA: Latent Dirichlet Allocation; DD-SCLDA: Discriminative Dictionary based Sparse Coding and LDA.
Table 2. Main features of the proposed method and other comparison methods. SVM: Support Vector Machine; LDA: Latent Dirichlet Allocation; DD-SCLDA: Discriminative Dictionary based Sparse Coding and LDA.
MethodScaleFeature ExpressionPost-Processing OptimizationClassifier
Our methodMulti-scaleFz+FNAD+FLSH+Fcov+FplaneMulti-scale pyramidSVM
Method 1Multi-scaleFz+Fcov+FplaneMulti-scale pyramidSVM
Method 2Single scaleFz+FNAD+FLSH+Fcov+FplaneNoneSVM
Method 3Single scaleFz+FNAD+FLSH+Fcov+FplaneMulti-scale pyramidSVM
Method 4Multi-scaleGeometry, strength, and statistical featuresRegional growthJointBoost
Method 5Multi-scaleSpin Image feature and FcovNoneAdaBoost
Method 6Multi-scaleLDA Model of the Spin Image feature and Fcov based on Multi-Level Point SetsNoneAdaBoost
Method 7Multi-scaleDD-SCLDA Model of the Spin Image feature and Fcov based on Multi-Level Point SetsNoneAdaBoost
Table 3. Classification results of precision/recall, accuracy and F1-score.
Table 3. Classification results of precision/recall, accuracy and F1-score.
Scene 1Tree (%)Building (%)Car (%)Accuracy (%)F1-Score (%)
Our method99.2/90.691.1/99.392.9/48.294.694.5/94.9/59.5
Method 199.2/84.986.8/99.399.9/42.791.991.5/92.7/59.8
Method 293.2/78.782.1/94.663.3/30.486.485.3/87.9/41.1
Method 396.9/81.784.1/97.798.8/23.289.388.7/90.4/37.6
Method 489.7/98.197.9/89.165.2/46.692.993.7/93.3/54.4
Method 585.7/92.992.0/83.856.9/54.787.989.2/87.7/55.8
Method 694.8/93.893.5/92.341.2/66.792.694.3/92.9/50.9
Method 793.1/96.095.2/92.673.3/62.293.794.5/93.9/67.3
Scene 2Tree (%)Building (%)Car (%)Accuracy (%)F1-score (%)
Our method92.4/94.398.5/97.973.0/68.495.893.4/98.2/70.6
Method 183.2/92.998.5/92.862.6/65.792.087.8/95.6/64.1
Method 277.3/94.398.3/88.971.7/60.089.685.0/93.4/65.3
Method 391.3/92.696.6/96.663.2/55.593.491.9/96.6/59.1
Method 486.8/91.296.8/95.544.1/34.892.288.9/96.1/38.9
Method 573.9/91.293.6/88.229.5/25.487.281.6/90.8/27.3
Method 690.3/93.997.6/96.549.4/42.094.192.1/97.0/45.4
Method 794.7/94.598.1/97.753.9/60.595.594.6/97.9/57.0
Table 4. Parameters comparison based on mF1 and Accuracy.
Table 4. Parameters comparison based on mF1 and Accuracy.
DmDl µ + υ Tree (%)Building (%)Car (%)Accuracy (%)mF1 (%)
11010394.094.460.793.982.9
21510394.094.560.794.083.1
32010393.994.560.394.082.9
41515394.795.063.594.684.4
51520394.394.760.294.383.1
61515494.795.062.694.684.1
71515594.795.062.694.683.8
82020293.594.154.593.580.7
92020394.594.959.594.483.0
102020594.594.959.094.482.8

Share and Cite

MDPI and ACS Style

Li, Y.; Tong, G.; Du, X.; Yang, X.; Zhang, J.; Yang, L. A Single Point-Based Multilevel Features Fusion and Pyramid Neighborhood Optimization Method for ALS Point Cloud Classification. Appl. Sci. 2019, 9, 951. https://doi.org/10.3390/app9050951

AMA Style

Li Y, Tong G, Du X, Yang X, Zhang J, Yang L. A Single Point-Based Multilevel Features Fusion and Pyramid Neighborhood Optimization Method for ALS Point Cloud Classification. Applied Sciences. 2019; 9(5):951. https://doi.org/10.3390/app9050951

Chicago/Turabian Style

Li, Yong, Guofeng Tong, Xiance Du, Xiang Yang, Jianjun Zhang, and Lin Yang. 2019. "A Single Point-Based Multilevel Features Fusion and Pyramid Neighborhood Optimization Method for ALS Point Cloud Classification" Applied Sciences 9, no. 5: 951. https://doi.org/10.3390/app9050951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop