Next Article in Journal
An Indoor DFEC Ranging Method for Homologous Base Station Based on GPS L1 and BeiDou B1 Signals
Next Article in Special Issue
Optical Dual Laser Based Sensor Denoising for OnlineMetal Sheet Flatness Measurement Using Hermite Interpolation
Previous Article in Journal
Low Complexity Beamspace Super Resolution for DOA Estimation of Linear Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning-Based Method for Overhead Contact System Component Recognition Using Mobile 2D LiDAR

College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(8), 2224; https://doi.org/10.3390/s20082224
Submission received: 18 March 2020 / Revised: 12 April 2020 / Accepted: 13 April 2020 / Published: 15 April 2020

Abstract

:
The overhead contact system (OCS) is a critical railway infrastructure for train power supply. Periodic inspections, aiming at acquiring the operational condition of the OCS and detecting problems, are necessary to guarantee the safety of railway operations. One of the OCS inspection means is to analyze data of point clouds collected by mobile 2D LiDAR. Recognizing OCS components from the collected point clouds is a critical task of the data analysis. However, the complex composition of OCS makes the task difficult. To solve the problem of recognizing multiple OCS components, we propose a new deep learning-based method to conduct semantic segmentation on the point cloud collected by mobile 2D LiDAR. Both online data processing and batch data processing are supported because our method is designed to classify points into meaningful categories of objects scan line by scan line. Local features are important for the success of point cloud semantic segmentation. Thus, we design an iterative point partitioning algorithm and a module named as Spatial Fusion Network, which are two critical components of our method for multi-scale local feature extraction. We evaluate our method on point clouds where sixteen categories of common OCS components have been manually labeled. Experimental results show that our method is effective in multiple object recognition since mean Intersection-over-Unions (mIoUs) of online data processing and batch data processing are, respectively, 96.12% and 97.17%.

1. Introduction

Rail transportation is a significant part of the transportation network. Up to 2019, the mileage of the global railway is over 1.3 million kilometers [1]. Billions of passengers and cargoes are transported by rail every year. The large-scale railway with high traffic challenges the stability and safety of the railway system. The overhead contact system (OCS) is a critical railway infrastructure for train power supply [2]. The components of OCS, such as contact wires, are easily deformed because of the contact force from the pantograph and various climatic conditions. The deformations might result in unstable power supply and even accidents. To access the operational condition of OCS and discover potential problems, periodic OCS inspections are required [3]. However, most of the OCS inspections nowadays are still done by workers. For instance, in China, workers with relative expertise and experience manually measure geometries of OCS components under the help of a portable laser rangefinder. With continuously growing mileage of railway [4], the manual measurement, which is inefficient and high-cost, might be hard to meet demands in the future.
Mobile laser scanning (MLS) is an emerging technology used for capturing dense three-dimensional (3D) point cloud. An MLS system with quick and accurate data acquisition can well record geometric details of surrounding objects [5]. Therefore, it is appropriate to apply MLS systems to OCS inspections. The operational condition of OCS can be acquired by analyzing the MLS point cloud instead of manual measurement. Recognizing the point cloud of OCS as a critical task of the data analysis has been studied in previous studies (e.g., [6,7,8]). Mobile 2D LiDAR is a special kind of MLS system applied to railway inspections. Figure 1 shows an instance of the mobile 2D LiDAR used to scan the OCS infrastructure. As for this kind of MLS system, a point cloud is built up by integrating points at each 2D scan line. Fully understanding the point cloud is significant for automatic inspections and intelligent diagnoses. Thus, this study focuses on multiple OCS component recognition with mobile 2D LiDAR. However, it is a difficult task because the similarities and the various associations among OCS components make scenes complex [8]. In this case, model-driven and data-driven methods become incompetent because rules and features for recognizing specific objects are difficult to be designed by human beings. Fortunately, the success of deep learning-based image recognition (e.g., [9,10,11]) has promoted the fast development of deep learning-based point cloud segmentation, which provides novel means to understand point cloud with semantics.
In this study, we propose a new deep learning-based method to conduct semantic segmentation on the point cloud collected by mobile 2D LiDAR. Points are classified into meaningful categories of objects scan line by scan line by our method. As local features are important for the success of point cloud semantic segmentation [12,13], an iterative point partitioning algorithm is developed to partition points into regions for local feature extraction at each scan line, and the Recurrent Neural Network (RNN)-based module, named as Spatial Fusion Network (SFN), is developed to extract and fuse local features among scan lines. We leverage the iterative point partitioning algorithm and SFN to perform multi-scale feature extraction, which can improve the segmentation performance. Experiments are carried out on the MLS point cloud collected from the OCS of Chinese high-speed railways. Experimental results show that our method is light and fast to classify points with high accuracy. To verify the effectiveness of our design, we further analyze the influences of different settings of feature extraction and feature fusion. The contributions of our work are summed up as follows:
  • A new deep learning-based method is proposed to solve the problem of recognizing multiple OCS components from the point cloud collected by mobile 2D LiDAR. Experiments show that our method is effective in multiple object recognition.
  • To solve the issue of local feature extraction, we develop an iterative point partitioning algorithm and SFN to acquire local features among scan lines at multiple scales.
  • Our method is designed to classify points scan line by scan line, supporting both online data processing and batch data processing.
The rest of this paper is organized as follows. Section 2 presents the previous works of railway object recognition based on MLS data, some existing commercial solutions of OCS inspections, and some deep learning-based architectures of point cloud semantic segmentation. Section 3 illustrates the methodology. Section 4 describes the experiments and experimental results. Section 5 draws conclusions and the vision of the future.

2. Related Work

In the previous studies, some methods have been proposed to recognize railway elements from MLS data. These methods can be mainly categorized into two types: model-driven and data-driven methods and learning-based methods.
With model-driven and data-driven methods, points are grouped into non-overlap regions based on manually designed features as well as handcrafted rules. For instance, Arastounia et al. [14,15,16] recognized the railway infrastructure based on templates, local structures, shapes of objects and topological relationships among objects. Lou et al. [17] extracted railway tracks from MLS data based on height difference calculated by using a sliding window method. Stein et al. [18] detected rails and tracks in 2D LiDAR data based on the extracted geometrical and topological features, and a rail shape template. Zou et al. [19] leveraged vector shift estimation, K-mean clustering, and reverse smoothing to extract rail tracks, resolving the difficulties of rail track extraction in scenarios of bends and turnout. Pastucha [6] recognized objects of OCS by using the RANSAC algorithm and DBSCAN algorithm. Gézero et al. [20] designed a simple algorithm to extract the linear elements of the railway based on the assumption that the extracted elements are roughly parallel to the sensor system trajectory. However, manually designed features and handcrafted rules are often designed for recognizing specific objects and difficult to extend to new tasks. Besides, a rule for recognizing various objects from a point cloud is difficult to design by human beings because it might be very complicated.
Compared to model-driven and data-driven methods, learning-based methods are more promising to fully understand the point cloud of complex scenes because features are automatically learned from data rather than manually designed. The regular supervised machine learning approach has been introduced to recognize railway objects. For instance, Jung et al. recognized ten categories of railway electrification system objects from the MLS point cloud by using a multi-range CRF model [7] or a multi-scale hierarchical CRF model [8]. However, few studies introduced the deep learning approach to recognized railway objects from MLS data. In contrast, some deep learning-based methods [21,22,23,24,25,26,27,28,29] have been successfully applied to image-based OCS inspections for component detection and defect detection.
As for commercial solutions of OCS inspections, the image processing technology and the MLS technology are commonly introduced. Due to the mature development of image processing technology, some methods, such as edge detection, model-matching, and deep learning-based image recognition, are used for defect detection. For instance, JX-300 [30] is a kind of inspection vehicle equipped with an image-based OCS inspection system. Compared with image data, MLS data contain 3D information, which is useful to measure geometric parameters of the OCS components. For instance, tCat [31] is a real-time mobile mapping solution using a portable rail trolley with laser sensors to measure geometries of overhead wires. Selectra Vision Company [32] developed LiDAR-based equipment for contact wire measurement. In addition, some solutions combined the image processing technology and the MLS technology to build up OCS inspection systems. For instance, Meidensha Corporation [33,34] combined laser scanners and line scan cameras to detect the position of contact wires. MERMEC Company [35,36] developed an inspection system that is made up of lasers and cameras to monitor the overhead wires and pick up defects.
Nowadays, deep learning-based methods of point cloud semantic segmentation provide novel means to recognize various objects in point cloud scenes. PointNet [37] proposed by Qi et al. is a pioneer of deep learning-based architecture in the field of 3D point cloud classification and semantic segmentation. However, the lack of local information restricts the performance of PointNet in large and complicated scenarios. Then, Qi et al. further proposed PointNet ++ [12] to overcome disadvantages of PointNet by using a hierarchical structure to capture local features of a point cloud, which achieves strong robustness and good effect. Since local features are significant to improve the performance of 3D point cloud semantic segmentation, some studies have been conducted for local feature extraction. Ye et al. [38] proposed a model named as 3P-RNN, in which local features are extracted by using Pointwise Pyramid Pooling modules and RNNs. Some architectures extract local features based on convolutional operations, such as PointCNN [39], Pointwise CNN [40], DGCNN [41], and LDGCNN [13].

3. Methodology

As for the long-distance OCS inspection, the MLS system would collect millions of points. It is a tremendous task to compute point relations in the whole point cloud. To reduce computational complexity, we process data scan line by scan line, avoiding computing relations between points of different scan lines.
The framework of our method is shown in Figure 2. As for feature extraction, we are concerned about two issues: (1) we have to generate regions for local feature extraction; and (2) 3D contexts are in demand for classifying points at a scan line because the 2D features extracted at a scan line are incompetent for recognizing 3D objects. To solve these two issues, we leverage the iterative point partitioning algorithm to generate regions and SFN to acquire 3D contexts by feature extraction and feature fusion. Details of the proposed framework are described in the following sections.

3.1. Data Preprocessing

In the 2D polar coordinate system, the tth scan line can be represented as a set of 2D points as:
A t = α t , i , θ t , i i = 1 , , N
where α t , i is a range distance, θ t , i is a bearing angle, and N is the number of points. Combined with traveled distances, we can transform points of scan lines into the 3D Cartesian coordinate system defined in Figure 3. Let d t be the traveled distance of the tth scan line, and let B t be a set of the 3D points transformed from A t . Let u and v be, respectively, the horizontal offset and vertical offset of the LiDAR based on the rail track centerline. The point set B t can be written as:
B t = p t , i x t , i , y t , i , z t , i i = 1 , , N
where x t , i , y t , i , and z t , i are coordinates computed as follows:
x t , i y t , i z t , i = α t , i cos θ t , i sin θ t , i 0 + u v d t
To reduce computation, we set up an inspection area to filter out distant points that belong to background objects such as trees and buildings. Let P t be a set of points within the inspection area. It can be written as:
P t = p t , i x t , i , y t , i , z t , i B t x m i n < x t , i < x m a x , y m i n < y t , i < y m a x
where x m i n , x m a x , y m i n , and y m a x describe the inspection area. By integrating the points within the inspection area, we can form a point cloud for OCS inspection. The point cloud P c can be represented as:
P c = t = 1 T P t
where T is the number of scan lines.

3.2. Iterative Point Partitioning

The proposed algorithm for partitioning points is conducted iteratively at each scan line to generate regions and establish region-to-region relationships. Note that a region is a set of points in a local neighborhood. By this algorithm, points are partitioned based on the region centroids acquired previously. As defined in Figure 4, region size and reign span are concepts applied to the algorithm:
  • To extract local features at a specified scale, we group points into non-overlap regions with the region size which is limited by a region size threshold ε .
  • To build up the region-to-region relationship in a 3D local neighborhood for feature fusion, the region span between two regions has to be less than the span threshold δ . The value of δ is based on human experiences of OCS object shapes.
At the tth scan line, points within the inspection area can be written as:
P t = p t , i x t , i , y t , i , z t , i i = 1 , , N t
where N t is the number of points and z t , 1 = z t , 2 = = z t , N t = d t denotes the traveled distance of the scan line. Let C t 1 ε = c ^ t 1 , 1 ε , , c ^ t 1 , K ε be the list of the K region centroids acquired previously with the region size threshold ε . The specific steps of our iterative point partitioning algorithm are as follows:
  • To restrict the region span within δ , we filter the region centroids in C t 1 ε by discarding the ones that do not satisfy the following condition:
    d t z ^ t 1 , n ε < δ
    where z ^ t 1 , n ε denotes the z-coordinate value of c ^ t 1 , n ε .
  • Points in P t are matched to the remaining region centroids by the nearest matching strategy. The nearest matching strategy is based on the m a t c h function defined as follows:
    m a t c h p , C = arg min c ^ C D i s t 3 D p , c ^
    where p is a point to be matched, C is a set of region centroids, and D i s t 3 D function computes the distance between p and c ^ in the 3D coordinate system.
  • To restrict the region size within ε , we discard the matching relations that do not satisfy the following condition:
    D i s t 2 D p t , i , c ^ t 1 , n ε < ε / 2 , where c ^ t 1 , n ε = m a t c h p t , i , C t 1 ε
    where D i s t 2 D function computes the distance between p t , i and c ^ t 1 , n ε in the xy-coordinate system, and the m a t c h function is defined in Step 2. The points without matching relation are grouped into an unmatched point set.
  • Assume that c ^ t 1 , n ε is the region centroid of region G t 1 , n ε (as described in Step 6, c ^ t 1 , n ε might be the region centroid of another previous region which is not empty). Points matched to c ^ t 1 , n ε are positionally near to G t 1 , n ε so that they can be grouped together to form a new region G t , n ε for feature fusion. Then, the region-to-region relationship between G t 1 , n ε and G t , n ε can be built up. Note that n of the symbols formatted as G t , n ε denotes region-to-region relationships. Based on Equations (7) and (9), the new region G t , n ε can be written as:
    G t , n ε = p t , i P t d t z ^ t 1 , n ε < δ , c ^ t 1 , n ε = m a t c h p t , i , C t 1 ε , D i s t 2 D p t , i , c ^ t 1 , n ε < ε / 2
  • The hierarchical clustering algorithm is utilized to partition the unmatched points which have been acquired in Step 3. The furthest-neighbor strategy [42] with ε as the merging threshold is applied to this process.
  • The regions generated with ε can be represented as G t , 1 ε , , G t , K + M ε , where G t , 1 ε , , G t , K ε are the regions acquired in Step 4 by matching, and G t , K + 1 ε , , G t , K + M ε are the regions acquired in Step 5 by clustering. Based on these regions, we produce the region centroid list C t ε = c ^ t , 1 ε , , c ^ t , K + M ε for the next iterative process, where:
    c ^ t , n ε = c ^ t 1 , n ε , G t , n ε = p G t , n ε p G t , n ε , G t , n ε
  • To capture point-to-point relations and region-to-region relations, we transform points into the local 2D coordinate system relative to their region centroids, and region spans between regions whose features would be fused are computed:
    U t , n ε = q x x ^ t , n ε ε / 2 , y y ^ t , n ε ε / 2 p x , y , z G t , n ε s t , n ε = d t z ^ t 1 , n ε δ , n K 0 , n > K
    where n = 1 , , K + M ; U t , n ε is a set of 2D points in the local coordinate system; s t , n ε is a region span; x ^ t , n ε and y ^ t , n ε are, respectively, the x-coordinate value and the y-coordinate value of c ^ t , n ε ; and ε and δ are utilized to scale the value into 1 , 1 .

3.3. Spatial Fusion Network

As PointNet is an effective architecture to extract features from unordered points [37], a PointNet layer is introduced to SFN for feature extraction. In addition, impelled by the success of RNN in video recognition [43,44,45,46], we utilize RNNs to fuse features among different scan lines to acquire 3D contexts. Specifically, RNNs in our SFN are implemented as two single-layer Long Short-Term Memory (LSTM) networks [47]. The architecture of SFN is shown in Figure 5. The inputs, represented as a set of 2D points in their local coordinate system as well as a value of region span, are produced by the iterative point partitioning algorithm (Step 7). At the tth iteration, the input points can be represented as q t , 1 , , q t , M t , where M t is the number of input points, and the region span can be represented as s t . The tth iterative process of SFN is described below in detail.

3.3.1. PointNet Layer for Feature Extraction

In the PointNet layer, the input points are firstly mapped to a high-dimensional feature space by a multilayer perceptron (MLP):
f q t , i = M L P q t , i , for i = 1 , , M t
where M L P : R 2 R 128 , and f q t , i is the point feature of q t , i . Then, the point features are aggregated by a max-pooling layer to form a region feature:
f r t = max f q t , 1 , , f q t , M t
where f r t is the region feature aware of the relation of input points.

3.3.2. Recurrent Neural Networks for Feature Fusion

For feature fusion, the input region span is firstly used to produce the span feature which represents the spatial relationships between regions. Then, the span feature is concatenated with the region feature together for feature fusion. There are two optional kinds of LSTM:
  • unidirectional LSTM (UDLSTM), which uses forward dependencies to fuses features; and
  • bidirectional LSTM (BDLSTM), which make use of the forward and backward dependencies to form complete 3D perspectives for feature fusion.
As for using the UDLSTM, the process of feature fusion can be written as follows:
f s t , h t 1 = L S T M 1 s t , h t 1 1 f f t , h t 2 = L S T M 2 c o n c a t f r t , f s t , h t 1 2
and the process with the usage of the BDLSTM can be written as follows:
f s t , h t 1 = L S T M 1 s t , h t 1 1 , h t + 1 1 f f t , h t 2 = L S T M 2 c o n c a t f r t , f s t , h t 1 2 , h t + 1 2
where f s t is the span feature; f r t is the region feature acquired in Equation (14); f f t is the new region feature with a 3D context; h t 1 1 , h t 1 , and h t + 1 1 are hidden states of L S T M 1 ; h t 1 2 , h t 2 , and h t + 1 2 are hidden states of L S T M 2 ; and c o n c a t function is used to concatenate the input features.
By combining the point feature and the region feature, we can capture relationships between points in a 2D region. By feature fusion, we can acquire contexts in 3D neighborhoods. Therefore, we concatenate f r t and f f t with each f q t , i , to form the new point features with 3D contexts:
f p t , i = c o n c a t f q t , i , f r t , f f t , for i = 1 , , M t
where f p t , i is the new point features of q t , i .

3.4. Multi-Scale Feature Extraction and Per-Point Classification

Figure 6 shows an instance of single-scale feature extraction. As for multi-scale feature extraction, we conduct the single-scale feature extraction repeatedly with different region size thresholds. At the tth scan line, point features extracted at K scales can be written as:
F p t ε k = f p t , 1 ε k , , f p t , N t ε k , k = 1 , . , K
where f p t , i ε k extracted with the region size threshold ε k is a point feature of p t , i and N t is the number of points within the inspection area at the tth scan line. Point features extracted at multiple scales are concatenated together to form multi-scale point features. Then, the multi-scale point features are input to an MLP to yield point scores:
f p t , i = c o n c a t f p t , i ε 1 , , f p t , i ε K s c o r e t , i , = M L P f p t , i , for i = 1 , . , N t
where f p t , i denotes the multi-scale point feature of p t , i ; s c o r e t , i is the point sore indicating which class p t , i should belong to. In Equation (19), MLP with 512 , 128 , 32 , c as its four hidden layer sizes are shared for yielding point scores at each scan line, where c is the number of categories of targets. Additionally, batch normalization is applied to each layer of the MLP with ReLU.

4. Experiments

4.1. Data Description

The MLS data for training and testing were collected from parts of the Chinese high-speed railway with a length of 16.7 km in 2019. The SICK LMS511 LiDAR sensor is applied to our MLS system for data acquisition. Table 1 shows the operating parameters of our MLS system, and Figure 7a shows the scene where data were acquired.
Based on the design of OCS, the inspection area described in Section 3.1 is set as x m i n = 1 m, x m a x = 7 m, y m i n = 0 m, and y m a x = 8 m. Accordingly, there are 7.86 million points collected in this inspection area. We manually labeled the collected points into 17 classes: contact wire (COW), catenary wire (CAW), cantilever (CTLV), dropper (DRO), stitch wire (SW), pole, spreader (SPD), return wire (RW), feeder wire (FW), suspension insulator (SI), pulley block (PB), tensioning weight (TW), guy wire (GW), mid-point anchor wire (MPAW), mid-point anchor auxiliary wire (MPAAW), electric connector (EC), and others. Figure 7b–d shows examples of the manually labeled point cloud. Approximately 80% of the data are used for training and the rest for testing. MLS data of the training dataset and the test dataset are sliced into sixty samples and five samples, respectively. The compositions of these two datasets are shown in Table 2.

4.2. Implementation Details

For multi-scale feature extraction, we extract features with three region size thresholds, namely ε 1 = 0.5 m, ε 2 = 2 m, and ε 3 = 7 m, which are used for capturing the object relations at small scale, middle scale, and large scale. In addition, the region span threshold δ is set as 1 m. Figure 8 shows the results of iterative point partitioning with these three region size thresholds. As for feature fusion, UDLSTMs and BDLSTMs are, respectively, used in the SFN to evaluate our methods. Note that the mode with UDLSTMs is named as the unidirectional mode, and the mode with BDLSTMs is named as the bidirectional mode.
In terms of optimizing learning-based modules, we leverage the ten-fold cross-validation method on the training dataset to tune hyperparameters. Experimentally, we choose the cross-entropy loss function to estimate the loss. In addition, the Adam optimizer with the learning rate 0.001 is utilized to optimize parameters of the modules. The layer sizes and numbers of layers, described in Section 3.3 and Section 3.4, are determined by considering the segmentation accuracy and the computational efficiency. As the partitioning results are different among samples, the batch size is set as 1. Note that the batch size can be larger than 1 with padding alternatively, whereas it will bring in extra computation.

4.3. Semantic Segmentation Results

To evaluate our method, the evaluation metrics applied to our experiments are as follows:
p r e c i s i o n = T P T P + F P
r e c a l l = T P T P + F N
I o U = T P T P + F P + F N
A c c u a r c y = T P + T N T P + F P + T N + F N
where I o U denotes the Intersection-over-Union. T P , F P , T N and F N are, respectively, the numbers of true positives, false positives, true negatives, and false negatives.
Table 3 shows the point cloud semantic segmentation results on the test dataset. In the unidirectional mode and the bidirectional mode, high rates of mean precision (97.81% and 98.69%), mean recall (98.20% and 98.41%) and mean IoU (96.12% and 97.17%) are acquired by our method. By comparing the overall results of two modes, we can find that the more comprehensive 3D contexts provided in bidirectional mode can lead to a better segmentation performance. Major wires (i.e., contact wire, catenary wire, return wire, and feeder wire) are critical components for power transmission. Except for the catenary wire, the major wires are well recognized with IoU over 99% in both modes. Except for the dropper, stitch wire, and electric connector, the IoU of each class is over 94% in the unidirectional mode and over 96% in the bidirectional mode. The precision of stitch wire (90.37%) and electric connector (96.77%) are, respectively, the lowest in the unidirectional mode and the bidirectional mode. Besides, the electric connector class takes the lowest recall (90.59% and 90.90%) and the lowest IoU (86.75% and 88.23%) in both modes. The relatively low performance of recognizing electric connector objects might be caused by the small quantity of training samples.
As shown in Table 4 and Table 5, most of the errors of dropper are misclassifying dropper points as contact wire, catenary wire, and stitch wire. A few points of stitch wire and catenary wire are confused. In the unidirectional mode, a few electric connector points are incorrectly recognized as points of dropper and cantilever, whereas, in the bidirectional mode, points of electric connector are misclassified as dropper but not cantilever. In general, the errors shown in these two confusion matrices lead to the relatively low IoU of dropper, stitch wire, and electric connector.
Segmentation results are visualized in Figure 9 and Figure 10. Overall, few errors occur in the non-insulated overlap section. Compared with the non-insulated overlap section, the scene of the insulated overlap section is more complex because there are more relationships of interconnected objects. Nevertheless, objects are also clearly distinguished in the insulated overlap section. By highlighting errors, we can find that most of the errors occur at the intersections between objects.

4.4. Space and Time Complexity

We recorded the space and time cost using Pytorch 1.2 on a server with an Intel Xeon E5-2623 v4 CPU running at 2.6 GHz and an Nvidia Quadro M4000 GPU. The results are shown in Table 6. Compared with UDLSTM, using BDLSTM leads to more model parameters. As a result, our method in the bidirectional mode is more expensive than that in the unidirectional mode. The SLSL unidirectional mode consumes the most time because the iterative process cannot make full use of the GPU. The time cost without GPU (9.32 ms) just increases by approximately 3.67% in SLSL unidirectional mode. This result indicates that our method can be utilized to process scan lines with at least 100 Hz with Intel Xeon E5-2623 v4 CPU. Compared with the scanning frequency (25 Hz) of the LiDAR sensor used in the experiments, our method supports online data processing where data can be processed in real-time.

4.5. Ablation Study

Further experiments were conducted to explore the contributions of the multi-scale feature extraction and feature fusion.
To verify the effectiveness of multi-scale feature extraction, we compared the performances of different settings of region size threshold ε on the test dataset. As shown in Table 7, multiple-scale feature extraction contributes to remarkable improvements of overall accuracy (OA) and mIoU. Compared to the double-scale feature extraction ( ε = 0.5 , 7 ) in the unidirectional mode, the triple-scale feature extraction ( ε = 0.5 , 2 , 7 ) helps mIoU increase by 1.02%, although OA rises by 0.19%. The greater improvement of mIoU indicates that multiple-scale feature extraction can reduce errors of small-quantity classes.
To verify the effectiveness of feature fusion, we conducted an experiment in the non-feature fusion mode where we removed the feature fusion part of SFN, so that point features were only combined with region features. As depicted in Table 8, we can find that the rates of mIoU and OA increase with the help of feature fusion. In the non-feature fusion mode, features at a scan line only contain the 2D information, which would lead to confusion of recognizing 3D objects. For instance, points of a spreader are always misclassified as points of a cantilever at pole-spreader junctions in the non-feature fusion mode (Figure 11a), whereas few errors occur at that location in the unidirectional mode and the bidirectional mode (Figure 11b,c), which demonstrates that the feature fusion contributes to the reduction of errors.

5. Conclusions

This study focuses on multiple OCS component recognition with mobile 2D LiDAR. A deep learning-based method is developed to conduct point cloud semantic segmentation scan line by scan line. We propose the iterative point partitioning algorithm and SFN as two critical components of our method for local feature extraction. As shown in the experiments, sixteen categories of common OCS components are recognized with high precision, recall, and IoU on the test dataset. According to the design of SFN, UDLSTM and BDLSTM are two optional kinds of LSTM used for feature fusion, resulting in 96.12% and 97.17% mIoU, respectively. In particular, UDLSTMs with the mere forward dependencies allow the SLSL unidirectional mode to support online data processing. Additionally, the ablation study shows that the design of multi-scale local feature extraction makes remarkable improvements of the segmentation performance. Compared with the model-driven and data-driven methods mentioned in Section 2, our method with automatic feature learning rather than manually designed rules is more effective in multiple object recognition and easier to be extended to new recognition tasks.
With our further research, we intend to improve the iterative point partitioning algorithm and SFN to adapt to 3D input. Hence, our method can be applied to higher performance MLS systems equipped with 3D LiDAR. Furthermore, due to the ability of automatically learning features, it is possible to extend our method to other inspection tasks such as rail track inspections and tunnel inspections.

Author Contributions

Conceptualization, L.C. and C.X.; methodology, L.C., S.L. (Shuai Lin) and S.L. (Siqi Li); software, S.L. (Shuai Lin) and S.L. (Siqi Li); validation, L.C. and X.T.; data curation, S.L. (Shuai Lin); writing—original draft preparation, L.C., S.L. (Siqi Li) and C.X.; and writing—review and editing, L.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (grant number 61772185).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Country Comparison: Railways. Available online: https://www.cia.gov/library/publications/the-world-factbook/fields/384rank.html (accessed on 3 March 2020).
  2. Kiessling, F.; Puschmann, R.; Schmieder, A.; Schneider, E. Contact Lines for Electric Railways: Planning, Design, Implementation, Maintenance, 3rd ed.; John Wiley & Sons: Erlangen, Germany, 2018; pp. 39–44. [Google Scholar]
  3. Li, S.; Xu, C.; Chen, L.; Liu, Z. Speed Regulation of Overhead Catenary System Inspection Robot for High-Speed Railway through Reinforcement Learning. In Proceedings of the 2018 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), Guangzhou, China, 8–12 October 2018; pp. 1378–1383. [Google Scholar]
  4. The Future of Rail. Available online: https://www.iea.org/reports/the-future-of-rail (accessed on 10 March 2020).
  5. Kukko, A.; Kaartinen, H.; Hyyppä, J.; Chen, Y. Multiplatform Mobile Laser Scanning: Usability and Performance. Sensors 2012, 12, 11712–11733. [Google Scholar] [CrossRef] [Green Version]
  6. Pastucha, E. Catenary System Detection, Localization and Classification Using Mobile Scanning Data. Remote Sens. 2016, 8, 801. [Google Scholar] [CrossRef] [Green Version]
  7. Jung, J.; Chen, L.; Sohn, G.; Luo, C.; Won, J.U. Multi-Range Conditional Random Field for Classifying Railway Electrification System Objects Using Mobile Laser Scanning Data. Remote Sens. 2016, 8, 1008. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, L.; Jung, J.; Sohn, G. Multi-Scale Hierarchical CRF for Railway Electrification Asset Classification From Mobile Laser Scanning Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3131–3148. [Google Scholar] [CrossRef]
  9. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  10. Li, Z.; Gan, Y.; Liang, X.; Yu, Y.; Cheng, H.; Lin, L. LSTM-CF: Unifying Context Modeling and Fusion with LSTMs for RGB-D Scene Labeling. In Proceedings of the 14th European Conference on Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016; pp. 541–557. [Google Scholar]
  11. Tu, X.; Xu, C.; Liu, S.; Xie, G.; Li, R. Real-Time Depth Estimation with an Optimized Encoder-Decoder Architecture on Embedded Devices. In Proceedings of the 2019 IEEE 21st International Conference on High Performance Computing and Communications; IEEE 17th International Conference on Smart City; IEEE 5th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Zhangjiajie, China, 10–12 August 2019; pp. 2141–2149. [Google Scholar]
  12. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017; pp. 5099–5108. [Google Scholar]
  13. Zhang, K.; Hao, M.; Wang, J.; de Silva, C.W.; Fu, C. Linked Dynamic Graph CNN: Learning on Point Cloud via Linking Hierarchical Features. arXiv 2019, arXiv:1904.10014. [Google Scholar]
  14. Arastounia, M. Automatic Classification of Lidar Point Clouds in a Railway Environment; University of Twente Faculty of Geo-Information and Earth Observation (ITC): Enschede, The Netherlands, 2012. [Google Scholar]
  15. Arastounia, M. Automated Recognition of Railroad Infrastructure in Rural Areas from LIDAR Data. Remote Sens. 2015, 7, 14916–14938. [Google Scholar] [CrossRef] [Green Version]
  16. Arastounia, M.; Oude Elberink, S. Application of Template Matching for Improving Classification of Urban Railroad Point Clouds. Sensors 2016, 16, 2112. [Google Scholar] [CrossRef] [Green Version]
  17. Lou, Y.; Zhang, T.; Tang, J.; Song, W.; Zhang, Y.; Chen, L. A Fast Algorithm for Rail Extraction Using Mobile Laser Scanning Data. Remote Sens. 2018, 10, 1998. [Google Scholar] [CrossRef] [Green Version]
  18. Stein, D.; Spindler, M.; Lauer, M. Model-based rail detection in mobile laser scanning data. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 19–22 June 2016; pp. 654–661. [Google Scholar]
  19. Zou, R.; Fan, X.; Qian, C.; Ye, W.; Zhao, P.; Tang, J.; Liu, H. An Efficient and Accurate Method for Different Configurations Railway Extraction Based on Mobile Laser Scanning. Remote Sens. 2019, 11, 2929. [Google Scholar] [CrossRef] [Green Version]
  20. Gézero, L.; Antunes, C. Automated Three-Dimensional Linear Elements Extraction from Mobile LiDAR Point Clouds in Railway Environments. Infrastructures 2019, 4, 46. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, J.; Liu, Z.; Wang, H.; Núñez, A.; Han, Z. Automatic Defect Detection of Fasteners on the Catenary Support Device Using Deep Convolutional Neural Network. IEEE Trans. Instrum. Meas. 2017, 67, 257–269. [Google Scholar] [CrossRef] [Green Version]
  22. Chen, J.; Liu, Z.; Wang, H.; Liu, K. High-speed railway catenary components detection using the cascaded convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; pp. 1–6. [Google Scholar]
  23. Kang, G.; Gao, S.; Yu, L.; Zhang, D. Deep Architecture for High-Speed Railway Insulator Surface Defect Detection: Denoising Autoencoder With Multitask Learning. IEEE Trans. Instrum. Meas. 2018, 68, 2679–2690. [Google Scholar] [CrossRef]
  24. Zhong, J.; Liu, Z.; Han, Z.; Han, Y.; Zhang, W. A CNN-Based Defect Inspection Method for Catenary Split Pins in High-Speed Railway. IEEE Trans. Instrum. Meas. 2018, 68, 2849–2860. [Google Scholar] [CrossRef]
  25. Liu, Z.; Zhong, J.; Lyu, Y.; Liu, K.; Han, Y.; Wang, L.; Liu, W. Location and fault detection of catenary support components based on deep learning. In Proceedings of the 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Houston, TX, USA, 14–17 May 2018; pp. 1–6. [Google Scholar]
  26. Liu, W.; Liu, Z.; Núñez, A.; Wang, L.; Liu, K.; Lyu, Y.; Wang, H. Multi-Objective Performance Evaluation of the Detection of Catenary Support Components Using DCNNs. IFAC-PapersOnLine 2018, 51, 98–105. [Google Scholar] [CrossRef]
  27. Liu, W.; Liu, Z.; Núñez, A. Virtual Reality and Convolutional Neural Networks for Railway Catenary Support Components Monitoring. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 2183–2188. [Google Scholar]
  28. Han, Y.; Liu, Z.; Lyu, Y.; Liu, K.; Li, C.; Zhang, W. Deep learning-based visual ensemble method for high-speed railway catenary clevis fracture detection. Neurocomputing 2019. [Google Scholar] [CrossRef]
  29. Liu, W.; Liu, Z.; Núñez, A.; Han, Z. Unified Deep Learning Architecture for the Detection of All Catenary Support Components. IEEE Access 2020, 8, 17049–17059. [Google Scholar] [CrossRef]
  30. JX300 Smart Non-Contact Measuring Machine of Catenary. Available online: http://www.timeselectric.cn/sddqen/s/1429-4910-2402.html (accessed on 29 March 2020).
  31. tCat OLE Mobile Mapping. Available online: http://tcat.es/en/ (accessed on 29 March 2020).
  32. Diagnostic and Measuring Equipments. Available online: http://www.selectravision.com/catenary.php (accessed on 29 March 2020).
  33. Overhead Catenary System (OCS) Inspection System. Available online: https://www.meidensha.com/products/railwaysystem/prod_02/index.html (accessed on 29 March 2020).
  34. Nezu, K.; Matsumura, I.; Aboshi, M.; Niwakawa, M.; Kawabata, T.; Tabayashi, S. Contactless Measuring Method of Overhead Contact Line Positions by Stereo Image Measurement and Laser Distance Measurement. Q. Rep. RTRI 2015, 56, 181–186. [Google Scholar] [CrossRef] [Green Version]
  35. Measuring Vehicles and Systems. Available online: http://www.mermecgroup.com/15/inspect.php (accessed on 29 March 2020).
  36. Swift, M.; Aurisicchio, G.; Pace, P. New practices for railway condition monitoring and predictive analysis. In Proceedings of the 5th IET Conference on Railway Condition Monitoring and Non-Destructive Testing (RCM 2011), Derby, UK, 29–30 November 2011; pp. 1–6. [Google Scholar]
  37. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  38. Ye, X.; Li, J.; Huang, H.; Du, L.; Zhang, X. 3D Recurrent Neural Networks with Context Fusion for Point Cloud Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 403–417. [Google Scholar]
  39. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution On X-Transformed Points. In Proceedings of the Advances in Neural Information Processing Systems 31, Montreal, QC, Canada, 2–8 December 2018; pp. 820–830. [Google Scholar]
  40. Hua, B.S.; Tran, M.K.; Yeung, S.K. Pointwise Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake, UT, USA, 18–23 June 2018; pp. 984–993. [Google Scholar]
  41. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. (TOG) 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
  42. Lance, G.N.; Williams, W.T. A General Theory of Classificatory Sorting Strategies: 1. Hierarchical Systems. Comput. J. 1967, 9, 373–380. [Google Scholar] [CrossRef] [Green Version]
  43. Gao, C.; Wang, W.; Chen, M. Semantic segmentation with Recurrent Neural Networks on RGB-D videos. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 1203–1207. [Google Scholar]
  44. Ventura, C.; Bellver, M.; Girbau, A.; Salvador, A.; Marques, F.; Giro-i Nieto, X. RVOS: End-To-End Recurrent Network for Video Object Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5277–5286. [Google Scholar]
  45. Sun, M.C.; Hsu, S.H.; Yang, M.C.; Chien, J.H. Context-aware Cascade Attention-based RNN for Video Emotion Recognition. In Proceedings of the 2018 First Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, China, 20–22 May 2018; pp. 1–6. [Google Scholar]
  46. Ullah, A.; Ahmad, J.; Muhammad, K.; Sajjad, M.; Baik, S.W. Action Recognition in Video Sequences using Deep Bi-Directional LSTM with CNN Features. IEEE Access 2017, 6, 1155–1166. [Google Scholar] [CrossRef]
  47. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An instance of mobile 2D LiDAR used for data acquisition in this study. During the inspection, the 2D LiDAR scans the OCS infrastructure on the plane perpendicular to the rail direction, and the odometer sensor records the traveled distance of each scan line.
Figure 1. An instance of mobile 2D LiDAR used for data acquisition in this study. During the inspection, the 2D LiDAR scans the OCS infrastructure on the plane perpendicular to the rail direction, and the odometer sensor records the traveled distance of each scan line.
Sensors 20 02224 g001
Figure 2. Data processing framework. Points are classified scan line by scan line. Firstly, data of a scan line are transformed into a 3D coordinate system. Secondly, point features with 3D contexts are extracted at multiple scales by the iterative point partitioning algorithm and the SFN. Finally, the multi-scale local features are concatenated together and then input to a classifier to yield point scores.
Figure 2. Data processing framework. Points are classified scan line by scan line. Firstly, data of a scan line are transformed into a 3D coordinate system. Secondly, point features with 3D contexts are extracted at multiple scales by the iterative point partitioning algorithm and the SFN. Finally, the multi-scale local features are concatenated together and then input to a classifier to yield point scores.
Sensors 20 02224 g002
Figure 3. An MLS system carried by an inspection device at the starting point. The front view is shown on the left, and the left view is shown on the right. Since the device moves on the rail, the defined 3D Cartesian coordinate system is relative to rail tracks.
Figure 3. An MLS system carried by an inspection device at the starting point. The front view is shown on the left, and the left view is shown on the right. Since the device moves on the rail, the defined 3D Cartesian coordinate system is relative to rail tracks.
Sensors 20 02224 g003
Figure 4. Definitions of region size and region span. Assume that there are region A and region B, respectively, in plane A and plane B which are perpendicular to the z-axis. In region A, A 1 and A 3 are the remotest pair of points. The distance between A 1 and A 3 denotes the region size of region A. The distance between plane A and plane B denotes the region span between region A and region B.
Figure 4. Definitions of region size and region span. Assume that there are region A and region B, respectively, in plane A and plane B which are perpendicular to the z-axis. In region A, A 1 and A 3 are the remotest pair of points. The distance between A 1 and A 3 denotes the region size of region A. The distance between plane A and plane B denotes the region span between region A and region B.
Sensors 20 02224 g004
Figure 5. The architecture of SFN. In the PointNet, MLP denotes a multilayer perceptron, and batch normalization is applied to each layer of the MLP with ReLU. For feature fusion, both LSMTs are single hidden-layer. The hidden layer sizes of the LSTM 1 and the LSTM 2 are, respectively, 8 and 128 so that k and m are, respectively, 8 and 128 with the usage of UDLSTMs. Note that UDLSTMs are without backward hidden states. With the usage of BDLSTMs, outputs of the forward process and backward process are combined so that k and m are, respectively, 16 and 512.
Figure 5. The architecture of SFN. In the PointNet, MLP denotes a multilayer perceptron, and batch normalization is applied to each layer of the MLP with ReLU. For feature fusion, both LSMTs are single hidden-layer. The hidden layer sizes of the LSTM 1 and the LSTM 2 are, respectively, 8 and 128 so that k and m are, respectively, 8 and 128 with the usage of UDLSTMs. Note that UDLSTMs are without backward hidden states. With the usage of BDLSTMs, outputs of the forward process and backward process are combined so that k and m are, respectively, 16 and 512.
Sensors 20 02224 g005
Figure 6. An instance of single-scale feature extraction with the region size threshold ε . Symbols formatted as P t , C t ε , U t , n ε , and s t , n ε are depicted in Section 3.2. At the beginning of the iterative process of point partitioning, C 0 ε is initialized as ∅. Then, regions and region-to-region relationships are produced by the iterative point partitioning algorithm (IPPA). The region-to-region relationships are informed by the superscript n of the region data, which are formatted as U t , n ε , s t , n ε . Note that the regions without points (e.g., region data U t , a ε = , s t , a ε ) are discarded for feature extraction. Based on the region-to-region relationships, region data are reorganized in sequences (e.g., sequence , U t 1 , a ε , s t 1 , a ε , U t + 1 , a ε , s t + 1 , a ε , ). SFNs extract and fuse features among the sequential region data to acquire the point features with 3D contexts. We share the weights of SFNs for processing data at a single scale. The output F p t ε = f p t , 1 ε , , f p t , N t ε is the set of point features corresponding to the input points P t = p t , 1 , , p t , N t .
Figure 6. An instance of single-scale feature extraction with the region size threshold ε . Symbols formatted as P t , C t ε , U t , n ε , and s t , n ε are depicted in Section 3.2. At the beginning of the iterative process of point partitioning, C 0 ε is initialized as ∅. Then, regions and region-to-region relationships are produced by the iterative point partitioning algorithm (IPPA). The region-to-region relationships are informed by the superscript n of the region data, which are formatted as U t , n ε , s t , n ε . Note that the regions without points (e.g., region data U t , a ε = , s t , a ε ) are discarded for feature extraction. Based on the region-to-region relationships, region data are reorganized in sequences (e.g., sequence , U t 1 , a ε , s t 1 , a ε , U t + 1 , a ε , s t + 1 , a ε , ). SFNs extract and fuse features among the sequential region data to acquire the point features with 3D contexts. We share the weights of SFNs for processing data at a single scale. The output F p t ε = f p t , 1 ε , , f p t , N t ε is the set of point features corresponding to the input points P t = p t , 1 , , p t , N t .
Sensors 20 02224 g006
Figure 7. Overhead Contact System configuration of Chinese high-speed railway tagged with sixteen classes: (a) an image; and (bd) MLS data ((b,c) sections of insulated overlap and (d) a section with mid-point anchor).
Figure 7. Overhead Contact System configuration of Chinese high-speed railway tagged with sixteen classes: (a) an image; and (bd) MLS data ((b,c) sections of insulated overlap and (d) a section with mid-point anchor).
Sensors 20 02224 g007aSensors 20 02224 g007b
Figure 8. Results of iterative point partitioning: (a) ε 1 = 0.5 m; (b) ε 2 = 2 m; and (c) ε 3 = 7 m. The points grouped into regions are tagged with their region colors. Regions with the same color are related together for feature fusion.
Figure 8. Results of iterative point partitioning: (a) ε 1 = 0.5 m; (b) ε 2 = 2 m; and (c) ε 3 = 7 m. The points grouped into regions are tagged with their region colors. Regions with the same color are related together for feature fusion.
Sensors 20 02224 g008aSensors 20 02224 g008b
Figure 9. Segmentation results of non-insulated overlap section: (a) unidirectional mode; and (b) bidirectional mode. In each subfigure, predictions are shown above and errors are highlighted in red below (the same as Figure 10 and Figure 11).
Figure 9. Segmentation results of non-insulated overlap section: (a) unidirectional mode; and (b) bidirectional mode. In each subfigure, predictions are shown above and errors are highlighted in red below (the same as Figure 10 and Figure 11).
Sensors 20 02224 g009
Figure 10. Segmentation results of insulated overlap section: (a) unidirectional mode; and (b) bidirectional mode.
Figure 10. Segmentation results of insulated overlap section: (a) unidirectional mode; and (b) bidirectional mode.
Sensors 20 02224 g010
Figure 11. Segmentation results of a scene with spreaders: (a) non-feature fusion mode; (b) unidirectional mode; and (c) bidirectional mode.
Figure 11. Segmentation results of a scene with spreaders: (a) non-feature fusion mode; (b) unidirectional mode; and (c) bidirectional mode.
Sensors 20 02224 g011
Table 1. Parameters of our MLS system.
Table 1. Parameters of our MLS system.
ParametersValue
scaning frequency25 Hz
angular resolution 0.1667 °
field of view 90 °
range80 m
travelling speedapproximately 1 m/s
Table 2. The point proportions and object quantities of each class. Note that “-” in this table means that the class is uncountable.
Table 2. The point proportions and object quantities of each class. Note that “-” in this table means that the class is uncountable.
ClassTraining DatasetTest Dataset
Points (%)QuantityPoints (%)Quantity
COW17.90-4.18-
CAW12.75-2.84-
CTLV4.183381.1192
DRO0.5219820.12536
SW3.312750.9576
pole8.052782.1576
SPD0.611240.1532
RW14.71-3.59-
FW13.31-3.29-
SI0.082200.0355
PB0.69340.138
TW1.64340.308
GW0.781080.2426
MPAW1.00100.081
MPAAW0.13200.012
EC0.1290.028
others0.94-0.12-
Table 3. Per-class precision, recall, and IoU in the unidirectional mode and the bidirectional mode.
Table 3. Per-class precision, recall, and IoU in the unidirectional mode and the bidirectional mode.
ClassUnidirectional ModeBidirectional Mode
Precision (%)Recall (%)IoU (%)Precision (%)Recall (%)IoU (%)
COW99.9299.9199.8399.9399.9199.84
CAW99.2696.3895.6999.1299.0198.15
CTLV99.3999.1998.5999.3699.2198.59
DRO96.5093.9090.8097.0492.9190.35
SW90.3798.3188.9897.1297.8395.07
pole99.6199.6899.2999.5599.7699.32
SPD97.2098.8696.1398.2497.7796.08
RW99.9299.9299.8499.9999.9499.93
FW99.9999.9999.9899.9999.9899.97
SI99.3498.4197.7798.2999.0297.34
PB97.8999.4197.3299.0699.2598.32
TW99.4798.9398.4199.2899.2898.57
GW98.9397.7296.7099.1899.1298.31
MPAW96.0599.3395.4397.1799.8997.06
MPAAW95.1199.5394.6998.6299.6998.32
EC95.3590.5986.7596.7790.9088.23
others98.4199.3697.7998.9499.5698.51
mean97.8198.2096.1298.6998.4197.17
Table 4. Confusion matrix of partial classes in the unidirectional mode. Values denote the point proportion of the reference.
Table 4. Confusion matrix of partial classes in the unidirectional mode. Values denote the point proportion of the reference.
ReferencePrediction (%)Total (%)
COWCAWCTLWDROSWEC
COW99.910.000.050.030.000.0099.99
CAW0.0096.380.040.053.410.0199.89
CTLV0.050.1699.190.000.090.0099.49
DRO1.992.100.0393.901.490.3699.87
SW0.011.580.040.0498.310.0099.98
EC1.202.021.964.230.0090.59100
Table 5. Confusion matrix of partial classes in the bidirectional mode. Values denote the point proportion of the reference.
Table 5. Confusion matrix of partial classes in the bidirectional mode. Values denote the point proportion of the reference.
ReferencePrediction (%)Total (%)
COWCAWCTLWDROSWEC
COW99.910.000.050.030.000.0099.99
CAW0.0099.010.050.030.830.0199.93
CTLV0.040.1299.210.000.080.0099.45
DRO1.912.390.0192.912.670.1099.99
SW0.012.120.030.0197.830.00100
EC1.202.080.005.620.1390.9099.93
Table 6. Model size and Time. Value of model size denotes the total size of the model parameters. Value of time denotes the average time of processing per scan line on the test dataset. In the unidirectional mode and the bidirectional mode, scan lines are processed in batch. In the scan line-by-scan line (SLSL) unidirectional mode, scan lines are processed in sequence with the same model parameters as the unidirectional mode, which means that segmentation results of the unidirectional mode and SLSL unidirectional mode are the same.
Table 6. Model size and Time. Value of model size denotes the total size of the model parameters. Value of time denotes the average time of processing per scan line on the test dataset. In the unidirectional mode and the bidirectional mode, scan lines are processed in batch. In the scan line-by-scan line (SLSL) unidirectional mode, scan lines are processed in sequence with the same model parameters as the unidirectional mode, which means that segmentation results of the unidirectional mode and SLSL unidirectional mode are the same.
ModeModel Size (MB)Time (ms)
unidirectional1.111.80
bidirectional1.741.95
unidirectional (SLSL)1.118.99
Table 7. Comparison of different settings of region size threshold ε .
Table 7. Comparison of different settings of region size threshold ε .
Settings of ε Unidirectional ModeBidirectional Mode
mIoU (%)OA (%)mIoU (%)OA (%)
0.584.2997.8088.7398.35
277.7396.3282.4795.96
787.9996.9590.2497.85
0.5, 795.1099.1797.0999.52
0.5, 2, 796.1299.3697.1799.54
Table 8. Comparison of the results with multi-scale feature extraction ( ε = 0.5 , 2 , 7 ) in the non-feature fusion mode, the unidirectional mode, and the bidirectional mode.
Table 8. Comparison of the results with multi-scale feature extraction ( ε = 0.5 , 2 , 7 ) in the non-feature fusion mode, the unidirectional mode, and the bidirectional mode.
ModemIoU (%)OA (%)
non-feature fusion93.5398.29
unidirectional96.1299.36
bidirectional97.1799.54

Share and Cite

MDPI and ACS Style

Chen, L.; Xu, C.; Lin, S.; Li, S.; Tu, X. A Deep Learning-Based Method for Overhead Contact System Component Recognition Using Mobile 2D LiDAR. Sensors 2020, 20, 2224. https://doi.org/10.3390/s20082224

AMA Style

Chen L, Xu C, Lin S, Li S, Tu X. A Deep Learning-Based Method for Overhead Contact System Component Recognition Using Mobile 2D LiDAR. Sensors. 2020; 20(8):2224. https://doi.org/10.3390/s20082224

Chicago/Turabian Style

Chen, Lipei, Cheng Xu, Shuai Lin, Siqi Li, and Xiaohan Tu. 2020. "A Deep Learning-Based Method for Overhead Contact System Component Recognition Using Mobile 2D LiDAR" Sensors 20, no. 8: 2224. https://doi.org/10.3390/s20082224

APA Style

Chen, L., Xu, C., Lin, S., Li, S., & Tu, X. (2020). A Deep Learning-Based Method for Overhead Contact System Component Recognition Using Mobile 2D LiDAR. Sensors, 20(8), 2224. https://doi.org/10.3390/s20082224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop