Next Article in Journal
Ultra-High Sensitivity Terahertz Microstructured Fiber Biosensor for Diabetes Mellitus and Coronary Heart Disease Marker Detection
Previous Article in Journal
The Total Ionizing Dose Effects on Perovskite CsPbBr3 Semiconductor Detector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scattered Train Bolt Point Cloud Segmentation Based on Hierarchical Multi-Scale Feature Learning

School of Physical Science and Technology, Southwest Jiaotong University, Chengdu 610031, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2019; https://doi.org/10.3390/s23042019
Submission received: 16 January 2023 / Revised: 8 February 2023 / Accepted: 8 February 2023 / Published: 10 February 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In view of the difficulty of using raw 3D point clouds for component detection in the railway field, this paper designs a point cloud segmentation model based on deep learning together with a point cloud preprocessing mechanism. First, a special preprocessing algorithm is designed to resolve the problems of noise points, acquisition errors, and large data volume in the actual point cloud model of the bolt. The algorithm uses the point cloud adaptive weighted guided filtering for noise smoothing according to the noise characteristics. Then retaining the key points of the point cloud, this algorithm uses the octree to partition the point cloud and carries out iterative farthest point sampling in each partition for obtaining the standard point cloud model. The standard point cloud model is then subjected to hierarchical multi-scale feature extraction to obtain global features, which are combined with local features through a self-attention mechanism, while linear interpolation is used to further expand the perceptual field of local features of the model as a basis for segmentation, and finally the segmentation is completed. Experiments show that the proposed algorithm could deal with the scattered bolt point cloud well, realize the segmentation of train bolt and background, and could achieve high segmentation accuracy, which has important practical significance for train safety detection.

1. Introduction

Safety is the main research focus in the railway field [1,2,3]. As vital components affecting the safety of running trains, bolts have always been the key detection units in railways [4]. As shown in Figure 1, bolts play an important role in connecting and fixing various components on the train and are widely used. Traditional detection methods are generally manual troubleshooting of bolt faults by professional train inspectors, but these methods are inefficient, with high costs, human error, and an inability to realize dynamic detection [5]. Therefore, methods using computer vision have been developed. These methods analyze 2D images to judge whether the bolts are missed, damaged, or have other abnormalities, and to improve the bolt detection efficiency and accuracy to some extent [6]. Compared with 2D images, the 3D point cloud contains richer information, such as coordinates [7], which are conducive to the measurement of geometric parameters and has been widely used in many fields, for instance, automatic driving [8], face recognition [9,10], medical diagnosis [11,12], smart cities [13,14], industrial design [15], etc. Up to now, there are a few reports about 3D point clouds detection using the deep learning method in railways. A point cloud identification network based on deep learning is presented in [16], which can be used to identify some components of the railway catenary. However, based on the classification results of a single frame, the network is prone to misclassification in the point cloud connection area, thus causing a negative impact on identification accuracy. Recently, the GTAINet has been proposed to segment the locking wire component [17], which consists of two stages: point cloud classification and point cloud segmentation, but the performance is not satisfactory on scattered point clouds of key components of the train. Obviously, the 3D point cloud has great potential in the field of train safety detection, but it is faced with some problems, such as the unclear boundary of the point cloud connection area of different components, and the poor recognition effect of the scattered point cloud. The bolt, as the train key component, is the focus of detection. Therefore, it is of great research value to realize the recognition and segmentation of bolts in a point cloud of raw train key components collected by laser scanning equipment.
The raw point clouds of train key components used in this paper were collected and provided by inspection personnel at the railway site. According to the analysis of these point clouds, it is found that these point clouds generally have problems such as noise points, blurred boundaries, and a large number of points, which will affect the accuracy of the later task of the point cloud, but also bring large calculation costs. Therefore, an efficient preprocessing framework is particularly important. In this paper, a point cloud adaptive weighted guided filtering (AWGF) algorithm and a fast farthest point sampling algorithm based on key points (FFPS-kp) are designed based on the characteristics of noise and point number. After the point cloud preprocessing is completed, a deep learning network is used to segment the point clouds. First, after hierarchical and multi-scale feature extraction, global features are obtained, and these global features are used twice: (1) input them into the subsequent network directly for upsampling using the inverse distance weighted (IDW) interpolation method; (2) combine with a self-attention mechanism to generate weights of optimization characteristics, and the strengthened or suppressed features are used as the input of the skip connection. The other end of the skip connection is the global features sampled by IDW interpolation. After splicing these two features, they are the final basis for point cloud segmentation.
Our major contributions can be summarized below:
1. A 3D point cloud smoothing and edge-preserving algorithm is introduced which can adjust the filter weight adaptively, which enables us to smooth the scattered point clouds of train key components and get clearer contour lines;
2. A fast farthest point sampling algorithm of key points is proposed, which improves the speed of farthest point sampling (FPS) and saves the calculation cost without sacrificing accuracy;
3. A 3D point cloud segmentation framework suitable for train key components such as bolts is designed, and a deep learning network for bolt point cloud segmentation was trained. The original point cloud with noise points could be processed directly, and a high segmentation accuracy and mIoU were achieved in the dataset of the point cloud of the train bolt components.
The remainder is structured as follows: In Section 2, the related work is introduced. Section 3 describes the overall network architecture and key modules. Section 4 gives the experiment and results analysis respectively. The conclusion is given in Section 5.

2. Related Work

Bolt detection based on computer vision technology generally resort to 2D images obtained by cameras. For example, Li et al. trained a support vector machine (SVM) classifier to distinguish bolts from images based on local binary pattern (LBP) descriptors realized positioning by the rotate-and-slide window method and then identified whether the bolts were normal according to whether the bolts had a hexagon shape [4]. In [18], a pioneering vision-based approach was introduced to capture digital images from target bolt connections and estimate bolt rotation and loosening signals through a series of image computing steps. Ref. [19] researched an image-register-based steel node bolt loosening detection method, and the images before and after the bolt are released are registered to detect the differences in the registration error. In [20], by synthesizing 2D images of bolts generated from the graphic model, the deep learning model is trained to achieve bolt recognition and looseness detection, and the applicability of the depth learning model based on vision is improved in practical applications. Overall, the method based on 2D images improved the efficiency and accuracy of bolt detection to a certain extent. But due to the outstanding advantages of the 3D point cloud, bolt detection methods based on the point cloud have broader development prospects, but there are also some difficulties: (1) There are more noise points in the actual collected point cloud of train bolts, and (2) a large amount of point cloud data leads to high computational complexity in late point cloud processing.
First, the problem of acquisition error and noise can be smoothed by the position of the correction points. Ref. [21] proposed a point cloud bilateral filtering (BF) algorithm using position information and normal information. The basic idea is to adjust the position by moving the noise points near the main point cloud along the direction of the normal vector. This method has a large computational overhead and gradient inversion problems. The moving least squares (MLS) method was applied to the problem of point cloud noise [22]. This method iteratively projected the noisy points onto the estimated plane to reconstruct the smooth surface, but with a large amount of calculation, it could not deal with outliers well [23]. In [24], a locally optimal projection operator (LOP) is introduced, but the LOP tendency fails to converge when the point cloud is unevenly distributed. MLS and LOP are also not suitable to deal with the point cloud model of train key components. Point cloud-guided filtering (GF) is proposed in [25], compared with BF, GF has higher efficiency and a better edge-preserving effect, without gradient inversion. However, GF adopts the same smoothing parameters for the whole point cloud and fails to fully consider the details of the point cloud, resulting in a blur in areas with prominent details and obvious edges, and flat areas are overcorrected.
To solve the problem of large amounts of point cloud data, there is a series of point cloud descending sampling algorithms. In deep learning networks, FPS [26] and random sampling (RS) [27] are usually adopted to downsample the point cloud. Among them, the FPS can best retain the contour information of the point cloud and is widely used in the point cloud processing framework based on deep learning, but the computational complexity of FPS is relatively high [28].
At present, the 3D point cloud segmentation algorithms based on deep learning can be roughly divided into direct segmentation methods and indirect segmentation methods [29]. Among them, the former methods include the feature extraction algorithm based on the original points, that is, the original point cloud is segmented directly, and these methods are the mainstream direction of current research [30]. In 2017, the pioneering work that input points into the network to process and extract features directly was proposed and named PointNet by Qi et al. [31]. PointNet introduces the idea of deep learning to directly process the point cloud, and the overall network architecture adopts FPS to complete the downsampling operation and retain the global features of the point cloud through the max-pooling operation. Therefore, the lack of expression of local detail features leads to the poor generalization ability to complex scenes. In the same year, Qi et al. proposed PointNet++ as the optimization of PointNet. To some extent, PointNet++ overcame the shortcoming of lacking local information in PointNet by adopting hierarchical feature learning architecture and sampling layer and by grouping the layer to acquire the overall features of the local area and to improve the performance of point cloud classification and segmentation [32]. Later, PointCNN [33] was proposed, and this network uses 2D convolution on the point cloud to perform semantic segmentation. χ -Conv is used for aggregating spatial structure information and local feature information of points, but the replacement invariance of the point cloud is not considered adequately. In 2019, several networks were proposed to deal with point cloud segmentation tasks. DGCNN [34] extracted local shape features of the point cloud by Edge-Conv and retained the constancy of the arrangement. Semantic information of point sets could be better learned by dynamically updating graph structures between layers, but some local geometric information was still lost. To make the best of geometric information about point cloud shape, Liu et al. proposed RSCNN [35]. The core algorithm is to use a shared multi-layer perceptron (MLP) to obtain the convolution parameters from the geometric topological relations of each point in the point cloud and extract topological constraint relations. PointConv [36] applied a convolutional neural network (CNN) to the 3D point cloud processing task directly, and learned the weight function by using MLP and density function through kernel density estimation, so as to overcome the shortcoming of non-uniform sampling of FPS and realize the invariance of the point cloud order. With the impressive performance of the transformer in the field of natural language analysis and 2D image processing, PCT [37] was proposed, which successfully introduced the attention mechanism and transformer module into the field of 3D point cloud processing, and adopted the self-attention mechanism and offset-attention mechanism for feature learning. It performs well in classification, segmentation, and normal estimation tasks. PointMLP [38] abandons complex feature extractors and uses multiple feedforward residual MLP frameworks to learn point cloud representation, and aggregates local features extracted by MLP for feature learning.

3. Network Architecture

In this paper, a 3D point cloud segmentation network is introduced, whose input is some disorderly, scattered, and noisy point clouds. The overall process is shown in Figure 2.
First, the raw point cloud is preprocessed, including denoising and simplification. Adaptive weighted guided filtering of the point cloud (AWGF) is used for correcting the noisy points and smoothing the point cloud. Then, the fast farthest point sampling algorithm of key points (FFPS-kp) is used for simplifying and improving the quality of the point cloud. The preprocessed point cloud model is extracted through multi-scale and multi-layer feature extraction, and the global features are extracted for subsequent point cloud segmentation. The feature extraction layer consists of four parts: sampling layer, grouping layer, convolution layer, and max-pooling layer. The receptive field obtained after multiple feature extraction is large and almost close to the global information. Then, the local features of global regions are expanded by step-by-step upsampling through IDW. At the same time, the global features of local regions extracted from each feature are optimized through the self-attention mechanism, and all features used for point cloud segmentation are obtained by skip connection with the expanded local features of global regions. Then, scores are output through the fully connected network to complete the segmentation task.

3.1. Adaptive Weighted Guided Filtering of Point Cloud

Due to the impact of the scanning device, the external environment, and the target object itself, the point cloud obtained directly by the laser scanner will have abnormal data such as noise points or outliers inevitably. These abnormal data will affect the accuracy and efficiency of the subsequent point cloud task, so it is essential to denoise the raw point cloud data [39,40]. The adaptive weighted guided filtering (AWGF) algorithm of the 3D point cloud is shown in Figure 3.
First, the topological relationship is constructed by using K-D tree for the input point cloud P = { p 1 , p 2 , , p N } , where we use subscripts to represent different points, there are N points in P , and we use p i to represent any one point in P . the parameter k of the search range of each sampling point p i is determined and the neighborhood N ( p i ) is queried, N ( p i ) is the set of points within the k nearest neighbor of p i . p i j is used for representing the j-th neighborhood point of p i in N ( p i ) , p i j is the point cloud after filtering and smoothing, and P = { p 1 , p 2 , , p N } is the output point cloud after the overall filtering and smoothing of the input point cloud P , the number of point clouds after smoothing is the same as the input point cloud, both of which are N . In N ( p i ) , the linear transformation model can be formulated as follows:
p i j = α i p i j + β i
where, α i and β i are linear model parameters limited by the preset neighborhood, which can be calculated by minimizing the cost function J ( α , β ) . The J ( α , β ) reflects the difference between the output point cloud P and the input point cloud P , the J ( α , β ) is formally defined as below:
J ( α , β ) = p i j N ( p i ) [ ( p i j p i j ) 2 + ε α i 2 ]
where, ε is the parameter used for controlling the smoothness. In order to better handle edges, an edge perception weight w ( p i j ) is set as:
w ( p i j ) = 1 | N ( p i ) | p i j N ( p i ) σ i 2 ( p i j ) + λ σ i 2 ( p i ) + λ
where, σ i 2 ( p i j ) is the variance of the distance between p i and the surrounding j neighborhood points in N ( p i ) , and σ i 2 ( p i ) is the variance of the distance between all points in N ( p i ) and their neighborhood points, λ is a constant. The edge perception weight w ( p i j ) is used for assigning more weight to the points at the edge than the points in the flat region. Accordingly, the regularization coefficient at the edge is smaller, so the edge could be well preserved, and the position of noise points with a long distance could be adjusted. Therefore, the cost function is transformed below:
J ( α , β ) = p i j N ( p i ) [ ( α i p i j + β i p i j ) 2 + ε w ( p i j ) α i 2 ]
When the cost function J ( α , β ) is the minimum value, the values of J ( α , β ) α i and J ( α , β ) β i are equal to 0. Therefore, the linear model parameter solutions could be calculated as Formulas (5) and (6):
α i = 1 | N ( p i ) | p i j N ( p i ) p i j . p i j p ¯ . p ¯ [ 1 | N ( p i ) | p i j N ( p i ) p i j . p i j p ¯ . p ¯ + ε w ( p i j ) ]
β i = p ¯ α i . p ¯
where
p ¯ = 1 | N ( p i ) | p i j N ( p i ) p i j

3.2. Fast Farthest Point Sampling Based on Key Points

Since the number of the 3D point cloud obtained by scanners is huge, such large point cloud data entering the deep learning network directly will make the training time too long. In another word, if there is a downsampling process, the network efficiency will be improved. The downsampling algorithm named FFPS-kp is shown in Figure 4.
The denoised point cloud P = { p 1 , p 2 , , p N } is sampled by FFPS-kp. FFPS-kp consists of the following key steps: First, select a point in P , which is p i and create a K-D tree with p i in the point cloud P as the centroid, and the k nearest neighbor (KNN) algorithm is used for selecting k points around p i as the neighborhood; p i j is a neighborhood point of p i , the j-th point in the KNN is represented by the subscript j , and the covariance matrix cov ( p i ) is constructed for the neighborhood of p i . The covariance matrix cov ( p i ) can be formulated as follows:
cov ( p i ) = 1 k p i j N ( p i ) ( p i j p i ) T ( p i j p i )
The three feature vectors of cov ( p i ) form an orthogonal basis ( a , b , c ) , and the local reference frame is established with p i as the coordinate origin of local reference frame and ( a , b , c ) as the coordinate axis. The three eigenvalues γ 1 , γ 2 , γ 3 of cov ( p i ) could approximately represent the complexity of the surface at p i , and the curvature is defined as:
c i = γ 3 γ 1 + γ 2 + γ 3
After the curvature of all points in p i are obtained, they are sorted in an increasing sequence and the first n 1 points are taken out as the feature points ( n 1 can be adjusted according to the actual demand). Then the octree for the remaining point cloud is established and partitioned [41]. In each region, FFPS-kp is carried out, and the number of sampling points is taken as n 2 = n n 1 . The process is as follows: first, the x , y , z coordinates of each point of the input point cloud are read, and the minimum envelope cubes of the point cloud are obtained; then it is bisected in the x , y , z directions, and thus it is divided into eight subcubes with a serial number. Then, given an index, the number of points in each subcube were counted. Meanwhile, the ratio of the number of points in each subcube to the total number of the point clouds was calculated as the ratio of the number of the sampling points in this subcube to the total sampling points, to complete the distribution of the sampling points. Next, an initial point is selected in each subcube, and the distances between the remaining points to the selected point are calculated, the point with the maximum distance is taken out, and this process is repeated until the target points are taken out. If the actual sampling points are more than those needed, the extra points are randomly removed from the subcube with the most sampling points; otherwise, the corresponding points are randomly selected from the region with the most sampling points to complete the partition sampling.
Finally, the combination of curvature sampling and partition FFPS-kp is regarded as the final sampling result, set as P = { p 1 , p 2 , , p n } , where n is the number of points after downsampling and n < N , which is input into the subsequent network for training.

3.3. Feature Extraction

To improve the segmentation accuracy, both local features and global features should be considered when feature extraction is carried out. For example, a fatal weakness of PointNet is that it fails to preserve the local features, resulting in lack of detailed description capabilities [31]. For this problem, both the global and local feature descriptions are fully considered in the feature extraction network, in which a multi-scale and hierarchical feature extraction module combing attention mechanism is introduced. The overall feature learning process is shown in Figure 5.
The global feature learning process is as follows:
(1) The preprocessed point cloud P is input into the feature extraction network;
(2) FFPS-kp is used for sampling the P 1 = { p 1 , p 2 , , p n } , and P 1 is set as the center of a sphere to determine the neighborhood radius r 1 and delimit the neighborhood sphere s 1 ;
(3) MLP and max-pooling are used for extracting and concentrating the features of all points in s 1 , and the F 1 = { f 1 , f 2 , , f m 1 } after the first feature extraction is obtained;
(4) For F 1 , the second feature extraction is carried out in the same way. The neighborhood radius r 2 and the neighborhood sphere s 2 are set to obtain the F 2 = { f 1 , f 2 , , f m 2 } ;
(5) The third feature extraction is carried out for F 2 , and the neighborhood radius r 3 and the neighborhood sphere s 3 are set to obtain the F 3 = { f 1 , f 2 , , f m 3 } , where r 1 < r 2 < r 3 , m 3 < m 2 < m 1 ;
(6) Finally, the F 3 with global features is output to prepare for the subsequent point cloud segmentation.
The F 3 extracted with hierarchical clustering and multi-scale features is enough to describe the global features, but it still lacks detailed features when used for segmentation. In order to consider the description ability of the detailed information, features are expanded, including feature interpolation and feature splicing. First, based on F 3 , the IDW interpolation method based on the KNN is used for upsampling to achieve the purpose of supplementing details. The basic principle of IDW interpolation is that points that are closer to each other are more similar than those farther apart.
However, the point cloud features obtained by upsampling fail to make the best of the point cloud local information obtained during feature extraction. Therefore, the global features of local regions extracted by hierarchical multi-scale features are spliced into the corresponding global feature layer after upsampling using skip connection. In the process of skip connection, the self-attention mechanism is adopted for feature optimization of the global features in local regions, and then it is combined with the global features, so that the global features can better express the local information of the point cloud and also improve the feature expression ability and the segmentation accuracy accordingly. The core of the self-attention mechanism is to use other information in the target to enhance the semantic representation of the target information and make better use of the context of the target information [42].

4. Experiments and Discussion

4.1. Experimental Environment and Parameters

The experimental hardware configuration is NVIDIA GeForce RTX 3060, and the software environment is Windows 10, Python3.6.12, PyTorch 1.7.1, CUDA 11.2. The experimental parameters were set as Batch size 24 and Epoch 200. An Adam optimizer was used when training networks. The initial learning rate is set to 0.001, the learning rate decay index is 0.0001, and the activation function is RELU; the loss function is the cross entropy for the segmentation network.

4.2. Dataset

In this experiment, a laser 3D scanner was used to collect the data of point clouds containing bolts as train key components. We selected 1256 point clouds containing bolts and point cloud labels were assigned. Label 0 represented bolts and label 1 represented the background. Each point cloud in the datasets containing the bolt position consists of its X, Y, Z coordinates, X, Y, Z normal values, and labels. Table 1 shows the dataset details.

4.3. Evaluation Metrics

In this paper, time is used to evaluate the calculated efficiency of denoising and sampling, overall accuracy (OA). Intersection over union (IoU), and mean intersection over union (mIoU) are used for evaluating the segmentation effect. The calculation formulas are listed below:
OA = T P + T N T P + F P + T N + F N
I o U = T P T P + F P + F N
m I o U = 1 k I o U

4.4. Results and Discussion

4.4.1. Denoising and Smoothing Experiment

In order to compare the performance of the denoising and smoothing algorithm AWGF in this paper with the BF algorithm and the GF algorithm, a comparative experiment was conducted on four bolts in the dataset. The corresponding number of bolts is shown in Table 2, and the time comparison experiment is shown in Table 3. Figure 6 shows the contrast details.
Through the analysis of Table 2 and Table 3, in terms of computational complexity, with the increase of the number of point clouds, the filtering time of the three smoothing algorithms increases correspondingly, among which the BF takes the longest time, almost twice that of GF and AWGF. The GF gives up the calculation of the point cloud normal vector and uses the linear model, which saves the calculation cost greatly. Due to the introduction of edge perception weight calculation, the AWGF brings a certain calculation overhead, resulting in a slight increase in the filtering time of the AWGF compared with the GF.
Figure 6 shows the comparison between the raw bolt point cloud and the point cloud smoothed by the BF, the GF, and the AWGF, respectively. By comparing the experimental results, it is obvious that the three algorithms could achieve a certain smoothing effect. As can be seen from the details in Figure 6b, the model filtered by the BF algorithm still contains some noise points, and the smoothness is barely satisfactory. Figure 6c is smoother than that in Figure 6b. It can be seen the GF has a better smoothing effect on noise than the BF algorithm and can realize the smoothing task of the bolt point cloud, making the boundary contour of the point cloud clearer. Figure 6d shows the experimental results of the algorithm we proposed. Compared with Figure 6c, the filtering effect is more satisfactory and the correction effect on noisy points is better, which proves the effectiveness of the edge perception weight operator in this paper.
To sum up, by analyzing the experimental data and comparing the smoothing results in Figure 6, we can be conclude that the AWGF can achieve a balance between computational performance and filtering performance, and better complete the denoising and smoothing tasks of the bolt point cloud.

4.4.2. Downsampling Experiment

In order to compare the performance of the proposed algorithm FFPS-kp with FPS, four bolt point clouds were selected from the datasets for the downsampling comparison experiment, and the sampling numbers were set as 512, 1024, and 2048, respectively. The number of sampling points and the corresponding sampling time are listed in Table 4 and Table 5; the effect after sampling is shown in Figure 7.
Table 4 and Table 5 show the time spent using FPS and FFPS-kp for different point numbers of the same point cloud and different point cloud models, respectively. When sampling different points in the same point cloud, the sampling time of FPS increases rapidly with the increase of sampling points, and the average time is about 20–30 times that of FFPS-kp. For a large number of points of the train bolt point cloud, FPS can complete the sampling, but the time is too long. However, the FFPS-kp algorithm has an obvious speed advantage for models with more points, which can quickly complete the sampling task and greatly improve the sampling efficiency.
Figure 7 shows that both FPS and FFPS-kp can complete the downsampling of bolt point cloud components, retaining surface boundary and contour information fully. While compared with FPS, FFPS-kp can preserve the key points as much as possible, for instance, for the areas with large curvature, and for those with obvious surface characteristics such as the junction of bolts and background. These key points can provide some help for the later point cloud segmentation task, which is also proved in the following segmentation experiments.

4.4.3. Bolt Segmentation

In order to evaluate the performance of the proposed network architecture on the bolted point cloud segmentation task, we selected several popular point cloud segmentation algorithms, PointNet, Pointnet++, PointConv, DGCNN, and PointMLP to make the comparison. To reduce the calculation consumption in the preprocessing part of the experiment, the actual dense bolt point cloud was downsampled, and 2048 points were reserved for feature extraction. Table 6 shows the comparison of segmentation accuracy, IoU, and mIoU. Figure 8 gives the comparative graph of segmentation results of different deep learning algorithms.
Table 6 shows the segmentation performance comparison between the segmentation framework in this paper and the popular point cloud segmentation networks on the bolt point cloud datasets. As can be seen from Table 6, the point cloud segmentation network we proposed achieved the best segmentation performance, which are 98.12% OA, 95.18% mIOU, 94.54% bolt IoU, and 95.81% background IoU, respectively. In terms of the mIoU, compared with PointNet, Pointnet++, PointConv, DGCNN, and PointMLP, the mIOU increased by 4.94%, 3.19%, 1.95%, 2.09%, and 0.69%, respectively. In terms of the OA, compared with them, the OA increased by 2.85%, 1.80%, 1.02%, 1.43%, and 0.43%, respectively.
To compare the proposed method with other methods visualized, we selected some bolts randomly and compared the segmentation effect in Figure 8. The colors used refer to Wang et al. [17], where the green color denotes the bolt and the red color denotes the background.
As shown in Figure 8, although most of the segmentation algorithms have achieved a good segmentation effect, some algorithms cannot handle details well, such as the connection between bolts with background as well as the connection between bolts. However, the edge smoothing algorithm we proposed in this paper makes the connection between the bolt and the background smoother, that is, the boundary between the two is clearer, resulting in better overall segmentation effect and better performance in details.

4.4.4. Ablation Experiment

In order to explore the influence of the denoising module, the sampling module, and the attention mechanism module on the bolt segmentation results, we conducted a comparison of module ablation, as shown in Table 7.
Table 7 lists the comparison of the results of the ablation experiments of these modules. To explore whether the preprocessing module contributes to the segmentation results, we kept the feature extraction part unchanged and selected different denoising methods and different downsampling methods for experiments. According to the experimental results, when the denoising methods are the same, that is, when AWGF is used for denoising, FFPS-kp as the downsampling method, we obtained higher mIoU and OA. Compared with FPS, mIoU is higher by 0.08% and OA is higher by 0.12%. In addition, the sampling method is the same, that is, when FFPS-kp is used for sampling, the segmentation effect of AWGF is better than that of GF, mIoU is increased by 0.51%, and OA is increased by 0.23%. To verify the effectiveness of the attention module, we maintained the same preprocessing module and used or deleted the attention module to carry out the comparative experiment. When AWGF is used as the point cloud denoising algorithm and FFPS-kp is used as the downsampling algorithm, the addition of the attention mechanism module increases mIoU by 0.78% and OA by 0.87%, which verified the effectiveness of self-attention mechanism module in the process of bolt point cloud segmentation.
In short, the ablation experiment verifies the effectiveness of the proposed preprocessing module in point cloud segmentation, including the denoising module and the downsampling module. It also shows that the self-attention mechanism can make better use of local information, strengthen the extraction ability, and improve the segmentation performance.

5. Conclusions

Aiming at the actual train bolt point cloud model, we established a hierarchical and multi-scale feature extraction point cloud segmentation algorithm, which includes filtering smoothing, downsampling, and other preprocessing operations combined with an attention mechanism. First, the AWGF algorithm proposed in this paper can remove the noise points in the scattered point cloud, making the contour of the point cloud smoother and the boundary clearer. To solve the problem of downsampling, we optimized the FPS algorithm and proposed FFPS-kp. Compared with FPS, FFPS-kp greatly shortens the sampling time and improves the sampling efficiency without losing the segmentation accuracy. Finally, the hierarchical multi-scale point cloud segmentation model proposed in this paper performs well in the segmentation experiment of the actual train bolt point clouds and their backgrounds, and can successfully segment the bolt point clouds, which can be used in actual railway environments.
However, this algorithm takes too long time to segment the point cloud in large environments. In the future, we would consider designing a more efficient point cloud downsampling algorithm that could pay more attention to key features, so as to reduce the calculation amount for later feature learning, such as the idea of the critical points layer. At the same time, we would try to simplify the feature extraction structure, and further shorten the time without losing the accuracy and mIoU, so that the network could be applied to the rapid segmentation of the original point cloud in large train environments.

Author Contributions

Conceptualization, N.Z. and J.L.; methodology, N.Z., J.L. and Y.Z.; software, N.Z.; validation, N.Z., J.L., Y.Z., X.G. and L.L.; formal analysis, N.Z.; investigation, N.Z.; resources, J.L., Y.Z. and X.G.; data curation, Y.Z.; writing—original draft preparation, N.Z.; writing—review and editing, N.Z., J.L., Y.Z., X.G. and L.L.; visualization, N.Z.; supervision, J.L.; project administration, N.Z. and J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61960206010 and Sichuan Science and Technology Program, grant number 2021YJ0080.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work is financially supported by the National Natural Science Foundation of China and the Sichuan Science and Technology Program, and the authors acknowledge them for their support. We also thank Photoelectric Engineering Institute of Southwest Jiaotong University, as well as the Chengdu Lead Science & Technology Co., Ltd. for providing 3D point cloud data of train key components.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ding, S.; Chen, D.; Liu, J. Research, development and prospect of China high-speed train. Chin. J. Theor. Appl. Mech. 2021, 53, 35–50. [Google Scholar]
  2. Liu, D.; Wang, T.; Liang, X.; Meng, S.; Zhong, M.; Lu, Z. High-speed train overturning safety under varying wind speed conditions. J. Wind. Eng. Ind. Aerodyn. 2020, 198, 104111. [Google Scholar] [CrossRef]
  3. Zhang, H.; Yang, J.; Tao, W.; Zhao, H. Vision method of inspecting missing fastening components in high-speed railway. Appl. Opt. 2011, 50, 3658–3665. [Google Scholar] [CrossRef] [PubMed]
  4. Li, C.; Wei, Z.; Xing, J. Online inspection system for the automatic detection of bolt defects on a freight train. Proc. Inst. Mech. Eng. Part F J. Rail Rapid Transit 2016, 230, 1213–1226. [Google Scholar] [CrossRef]
  5. Zhou, F.; Song, Y.; Liu, L.; Zheng, D. Automated visual inspection of target parts for train safety based on deep learning. IET Intell. Transp. Syst. 2018, 12, 550–555. [Google Scholar] [CrossRef]
  6. Spencer, B.F., Jr.; Vedhus, H.; Yasutaka, N. Advances in computer vision-based civil infrastructure inspection and monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  7. Bello, S.A.; Yu, S.; Wang, C.; Adam, J.M.; Li, J. Review: Deep learning on 3D point clouds. Remote Sens. 2020, 12, 1729. [Google Scholar] [CrossRef]
  8. Li, Y.; Zhong, Z.; Cao, D. Deep learning for lidar point clouds in autonomous driving: A review. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 3412–3432. [Google Scholar] [CrossRef]
  9. Bhople, A.R.; Akhilesh, M.; Shrivastava, S.P. Point cloud based deep convolutional neural network for 3D face recognition. Multimed. Tools Appl. 2021, 80, 30237–30259. [Google Scholar] [CrossRef]
  10. Zhou, S.; Sheng, X. 3D face recognition: A survey. Hum. Cent. Comput. Inf. Sci. 2018, 8, 1–27. [Google Scholar] [CrossRef]
  11. Cheng, Q.; Sun, P.; Yang, C.; Yang, Y.; Liu, P.X. A morphing-Based 3D point cloud reconstruction framework for medical image processing. Comput. Methods Programs Biomed. 2020, 193, 105495. [Google Scholar] [CrossRef] [PubMed]
  12. Fu, Y.; Lei, Y.; Wang, T.; Patel, P.B.; Jani, A.; Mao, H.; Curran, J.W.; Liu, T.; Yang, X. Biomechanically constrained non-rigid MR-TRUS prostate registration using deep learning based 3D point cloud matching. Med. Image Anal. 2021, 67, 101845. [Google Scholar] [CrossRef] [PubMed]
  13. Wen, H.; Wang, Y. Classification-based scene modeling for urban point clouds. Opt. Eng. 2014, 53, 033110. [Google Scholar]
  14. Ning, X.; Ge, T.; Wang, Y. Shape classification guided method for automated extraction of urban trees from terrestrial laser scanning point clouds. Multimed. Tools Appl. 2021, 80, 33357–33375. [Google Scholar] [CrossRef]
  15. Kim, J.; Lee, J.; Chung, M.; Shin, Y.-G. Multiple weld seam extraction from RGB-depth images for automatic robotic welding via point cloud registration. Multimed. Tools Appl. 2021, 80, 9703–9719. [Google Scholar] [CrossRef]
  16. Lin, S.; Xu, C.; Chen, L.; Li, S.; Tu, X. LiDAR Point Cloud Recognition of Overhead Catenary System with Deep Learning. Sensors 2020, 20, 2212. [Google Scholar] [CrossRef]
  17. Wang, Z.; Zhang, Y.; Lv, T.; Luo, L. GTAINet: Graph neural network-based two-stage anomaly identification for locking wire point clouds using hierarchical attentive edge convolution. Int. J. Appl. Earth Obs. Geoinf. 2022, 115, 103106. [Google Scholar] [CrossRef]
  18. Park, J.-H.; Huynh, T.-C.; Choi, S.-H.; Kim, J.-T. Vision-based technique for bolt-loosening detection in wind turbine tower. Wind. Struct 2015, 21, 709–726. [Google Scholar] [CrossRef]
  19. Kong, X.; Li, J. Image registration-based bolt loosening detection of steel joints. Sensors 2018, 18, 1000. [Google Scholar] [CrossRef]
  20. Pham, H.C.; Ta, Q.-B.; Kim, J.-T.; Ho, D.-D.; Tran, X.-L.; Huynh, T.-C. Bolt-loosening monitoring framework using an image-based deep learning and graphical model. Sensors 2020, 20, 3382. [Google Scholar] [CrossRef]
  21. Digne, J.; De Franchis, C. The Bilateral Filter for Point Clouds. Image Process. Line 2017, 7, 278–287. [Google Scholar] [CrossRef]
  22. David, L. The approximation power of moving least-squares. Math. Comput. 1998, 67, 1517–1531. [Google Scholar]
  23. Marc, A.; Johannes, B.; Daniel, C.-O.; Shachar, F.; David, L.; Cláudio, T.S. Computing and rendering point set surfaces. IEEE Trans. Vis. Comput. Graph. 2003, 9, 3–15. [Google Scholar]
  24. Yaron, L.; Daniel, C.-O.; David, L.; Hillel, T.-E. Parameterization-free projection for geometry reconstruction. ACM Trans. Graph. TOG 2007, 26, 22-es. [Google Scholar]
  25. Han, X.; Jin, J.S.; Wang, M.; Jiang, W. Guided 3D point cloud filtering. Multimed. Tools Appl. 2018, 77, 17397–17411. [Google Scholar] [CrossRef]
  26. Huang, T.; Chen, J.; Zhang, J.; Liu, Y.; Liang, J. Fast Point Cloud Sampling Network. Pattern Recognition Letters 2022, 164, 216–223. [Google Scholar] [CrossRef]
  27. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. Randla-Net: Efficient semantic segmentation of large-scale point clouds. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11108–11117. [Google Scholar]
  28. Lang, T.; Manor, A.; Avidan, S. SampleNet: Differentiable Point Cloud Sampling. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7578–7588. [Google Scholar]
  29. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep learning for 3d point clouds: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef]
  30. Zhang, J.; Zhao, X.; Chen, Z.; Lu, Z. A review of deep learning-based semantic segmentation for point cloud. IEEE Access 2019, 7, 179118–179133. [Google Scholar] [CrossRef]
  31. Qi, C.R.; Su, H.; Mo, K.; Guibas, J.L. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  32. Qi, C.R.; Yi, L.; Su, H.; Guibas, J.L. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 5099–5108. [Google Scholar]
  33. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. Pointcnn: Convolution on x-transformed points. Adv. Neural Inf. Process. Syst. 2018, 31, 820–830. [Google Scholar]
  34. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon Justin, M. Dynamic graph cnn for learning on point clouds. Acm Trans. Graph. Tog 2019, 38, 1–12. [Google Scholar] [CrossRef]
  35. Liu, Y.; Fan, B.; Xiang, S.; Pan, C. Relation-shape convolutional neural network for point cloud analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; 8895-8904. [Google Scholar]
  36. Wu, W.; Qi, Z.; Li, F. Pointconv: Deep convolutional networks on 3d point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9621–9630. [Google Scholar]
  37. Guo, M.; Cai, J.; Liu, Z.; Mu, T.; Martin, R.R.; Hu, S. Pct: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  38. Ma, X.; Qin, C.; You, H.; Ran, H.; Fu, Y. Rethinking Network Design and Local Geometry in Point Cloud: A Simple Residual MLP Framework. arXiv 2022, arXiv:2202.07123. [Google Scholar]
  39. Han, X.; Jin, J.S.; Wang, M.; Jiang, W.; Gao, L.; Xiao, L. A review of algorithms for filtering the 3D point cloud. Signal Process. Image Commun. 2017, 57, 103–112. [Google Scholar] [CrossRef]
  40. Zeng, N.; Li, J.; Gao, X.; Zhang, Y.; Luo, L. An efficient filtering and smoothing algorithm for train key components based on scattered point clouds. Laser Optoelectron. Prog. 2023, 60, 1410011. [Google Scholar]
  41. Dricot, A.; Pereira, F.; Ascenso, J. Rate-distortion driven adaptive partitioning for octree-based point cloud geometry coding. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018. [Google Scholar]
  42. Vaswani, A.; Shazeer, N.; Paemar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
Figure 1. Scenes from an underneath high-speed train. The red square marks the bolts.
Figure 1. Scenes from an underneath high-speed train. The red square marks the bolts.
Sensors 23 02019 g001
Figure 2. Architectures of the segmentation network.
Figure 2. Architectures of the segmentation network.
Sensors 23 02019 g002
Figure 3. The pipeline of the AWGF.
Figure 3. The pipeline of the AWGF.
Sensors 23 02019 g003
Figure 4. The pipeline of the FFPS-kp.
Figure 4. The pipeline of the FFPS-kp.
Sensors 23 02019 g004
Figure 5. The process of feature learning.
Figure 5. The process of feature learning.
Sensors 23 02019 g005
Figure 6. Comparison results of the smoothing.
Figure 6. Comparison results of the smoothing.
Sensors 23 02019 g006
Figure 7. Comparison results of the sampling.
Figure 7. Comparison results of the sampling.
Sensors 23 02019 g007
Figure 8. Comparison results of the segmentation.
Figure 8. Comparison results of the segmentation.
Sensors 23 02019 g008
Table 1. Dataset details.
Table 1. Dataset details.
TrainingTestValidationTotal
8702521341256
Table 2. Bolts and its number of points.
Table 2. Bolts and its number of points.
Point CloudBolt1Bolt2Bolt3Bolt4
Number35,00644,31659,26762,405
Table 3. Smoothing time comparison.
Table 3. Smoothing time comparison.
Denoising MethodBolt1 (s)Bolt2 (s)Bolt3 (s)Bolt4 (s)
BF0.3690.4130.6340.649
GF0.1800.2260.3210.341
AWGF0.2150.2660.3760.395
Table 4. FPS sampling time.
Table 4. FPS sampling time.
Sampling NumberBolt1 (s)Bolt2 (s)Bolt3 (s)Bolt4 (s)
51262.4378.08103.14110.84
1024237.56279.42361.42376.5
2048989.011188.531648.361640.72
Table 5. FFPS-kp sampling time.
Table 5. FFPS-kp sampling time.
Sampling NumberBolt1 (s)Bolt2 (s)Bolt3 (s)Bolt4 (s)
5122.723.565.576.23
10249.1612.0215.6418.26
204835.5945.3461.8970.89
Table 6. Comparison of segmentation results.
Table 6. Comparison of segmentation results.
MethodBolt (%)Background (%)mIOU (%)OA (%)
PointNet89.4691.0190.2495.27
PointNet++91.2592.7291.9996.32
PointConv92.8393.6393.2397.10
DGCNN92.6593.5293.0996.69
PointMLP93.7895.1994.4997.69
Ours94.5495.8195.1898.12
Table 7. Ablation experiment.
Table 7. Ablation experiment.
Denoising MethodSampling MethodAttention MechanismBolt (%)Background (%)mIOU (%)OA (%)
AWGFFFPS-kp-93.7695.0494.4097.25
AWGFFFPS-kpSelf-attention94.5495.8195.1898.12
AWGFFPSSelf-attention94.4395.7795.1098.00
GFFFPS-kpSelf-attention94.0595.1394.5997.77
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, N.; Li, J.; Zhang, Y.; Gao, X.; Luo, L. Scattered Train Bolt Point Cloud Segmentation Based on Hierarchical Multi-Scale Feature Learning. Sensors 2023, 23, 2019. https://doi.org/10.3390/s23042019

AMA Style

Zeng N, Li J, Zhang Y, Gao X, Luo L. Scattered Train Bolt Point Cloud Segmentation Based on Hierarchical Multi-Scale Feature Learning. Sensors. 2023; 23(4):2019. https://doi.org/10.3390/s23042019

Chicago/Turabian Style

Zeng, Ni, Jinlong Li, Yu Zhang, Xiaorong Gao, and Lin Luo. 2023. "Scattered Train Bolt Point Cloud Segmentation Based on Hierarchical Multi-Scale Feature Learning" Sensors 23, no. 4: 2019. https://doi.org/10.3390/s23042019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop