Abstract
LiDAR is a useful technology for gathering point cloud data from its environment and has been adapted to many applications. We use a cost-efficient LiDAR system attached to a moving object to estimate the location of the moving object using referenced linear structures. In the stationary state, the accuracy of extracting linear structures is low given the low-cost LiDAR. We propose a merging scheme for the LiDAR data frames to improve the accuracy by using the movement of the moving object. The proposed scheme tries to find the optimal window size by means of an entropy analysis. The optimal window size is determined by finding the minimum point between the entropy indicator of the ideal result and the entropy indicator of the actual result of each window size. The proposed indicator can describe the accuracy of the entire path of the moving object at each window size using a simple single value. The experimental results show that the proposed scheme can improve the linear structure extraction accuracy.
1. Introduction
Light Detection and Ranging (LiDAR) technologies have been rapidly developed and there are various applications based on LiDAR. LiDAR systems are classified into three types: spatial, spectral, and temporal information capturing [1]. Spatial schemes obtain point cloud data based on Time of Flight (ToF) measurements and spectral schemes measure the information of a material using what is termed Laser Return Intensity (LRI). Temporal schemes gather additional information based on spatial and spectral information using a repeated LiDAR technique. Application developers must select an optimal product because each LiDAR type has different functionalities and specifications such as range, Field of View (FoV), precision, and accuracy.
In mobile environments, LiDAR has extended application areas and most autonomous vehicles are equipped with LiDAR. LiDAR technology is essential for Advanced Driver-Assistance Systems (ADAS) that handle automatically steering, accelerating, and braking under the driver’s supervision [2]. Autonomous driving vehicles use LiDAR sensors, which provide high-resolution and real-time 3D representation data to detect the surrounding environment and obstacles. Odometry is essential for accurate self-localization for path planning and environment perception, which are key features related to driving safety [3]. LiDAR-based odometry can handle environmental variations by taking advantage of its active sensor emitting laser beams. The main functionality of LiDAR odometry is a registration between the current scan data and the reference point cloud data, which is solved by Iterative Closed Point (IPC) algorithm.
Simultaneous Localization and Mapping (SLAM) is also a promising field related to mobile LiDAR. It is designed to build or update a map of an unknown environment while simultaneously keeping track of an agent’s location. LiDAR is a more popular mechanism in SLAM compared to other mechanisms such as radar and ultra-wideband positioning due to LiDAR’s high precision, wide coverage, and longevity [4]. Traditional LiDAR-based SLAM algorithms mainly leverage the geometric features from the scene context, while the intensity information from LiDAR is ignored. The SLAM framework, which uses both geometry and intensity information for odometry estimations, provides reliable and accurate localization in multiple environments and outperforms geometric-only methods [5]. LiDAR-based SLAM can also be used in indoor navigation systems of autonomous vehicles because LiDAR-based SLAM can provide more robust localization than image-based SLAM in a lack textured environment [6].
In this paper, we use a mounted LiDAR system on a moving object to monitor an excavation site in an urban area. We extract linear structures that have references to the location information to calibrate the accuracy of a satellite-based navigation system. The location of a moving object is measured through the triangulation method using the extracted linear structures. The Velodyne Puck equipped with our system has relatively low specifications, as shown in Table 1. As the distance between the Velodyne Puck and linear structures increases, the extraction accuracy rapidly decreases due to its low vertical angular resolution. Therefore, we propose a sliding window mechanism to improve the accuracy of collected data from the low-spec LiDAR system. The proposed sliding window mechanism merges consecutive frames to acquire more point cloud data and improves the possibility of linear structure extraction by considering the movement of the mobile object.
Table 1.
Velodyne Puck and Ultra Puck specifications.
We also propose an optimal window size decision algorithm based on Shannon entropy analysis. We calculate the entropy of the desired extracted linear structures at each point as a reference structure to find the optimal window size. Upon a change in the location of the moving object, our algorithm recalculates the entropy of each point with a change in the window size. The optimal window size of each point is set to the minimum difference between the reference entropy and the recalculated entropy. Our experimental results show that the proposed sliding window mechanism and the optimal window size decision algorithm perform well.
2. Related Works
2.1. Mobile LiDAR Systems
There are many systems that can monitor or map environments using LiDAR in a mobile device. A road segmentation-based curb detection method was proposed to provide navigation information for a self-driving vehicle with LiDAR [7]. A sliding-beam method is used to distinguish on-road and off-road areas and split segments of a road from point cloud data. A curb-detection method has also been applied to determine the positions of curbs for different road segments.
Zhang et al. proposed a real-time localization method that estimates the location of an autonomous driving vehicle using a 3D-LiDAR system [8]. Point cloud data of a curb beside the road are extracted based on the laser intensity features and matched to a high-precision curb map generated offline. A map-matching method between the point cloud data collected from the 3D-LiDAR system and the high-precision curb map is designed using an Iterative Closest Point (ICP) algorithm to improve the accuracy of the vehicle’s location. The authors verified the performance of the proposed system by comparing the estimated location with that of a low-cost global positioning system.
LiDAR can also be used to detect and trace various traffic participants, such as vehicles, pedestrians, obstacles, and bicycles, to guarantee the safety of self-driving vehicles [9]. This type of tracking system consists of three modules: mask generation, depth estimation, and a retracking mechanism. The retracking mechanism overcomes repetitive appearances and disappearances of objects caused by the movement of a self-driving vehicle and traffic participants. LiDAR-based object detection systems perform well under harsh weather conditions such as heavy rain and dense fog [10].
Cheng et al. proposed a water leakage detector for use in shielded tunnels that relies on deep learning with mobile LiDAR intensity images [11]. The mobile LiDAR system is designed to collect point cloud data and intensity information simultaneously from the shielded tunnels. The intensity images based on the collected point cloud data are generated and used for training with a Fully Convolutional Network (FCN) to improve the accuracy of the water leakage detection process. After training with the FCN, water leaks in the shielded tunnels can be extracted through intensity image semantic segmentation.
Luo et al. proposed an intelligent detection method for the spraying of tunnel shotcrete [12]. In this method, LiDAR can obtain a 3D model of the tunnel and extract the positions of arches because spraying areas are usually divided by arches in a tunnel. A YOLO-based model is used to detect the approximate bounding boxes of the arches and a line-detection algorithm is used for determining the final spraying areas.
A road-marking extraction and classification scheme was also proposed [13]. The proposed framework applies capsule-based deep learning from a massive and unordered mobile laser scanner. First, this framework segments the road surface from 3D point cloud data and generates 2D georeferenced intensity images. Then, a U-shaped capsule-based network model is used to extract road markings based on convolutional and deconvolutional capsule operations. Finally, a hybrid capsule-based network model is applied to classify different types of road markings using a revised dynamic routing algorithm and a large-margin Softmax loss function.
Silva et al. proposed a robust fusion type of LiDAR system for mobile robots [14]. Mobile robots must fuse heterogeneous data because these robots are equipped with various positioning sensors such as LiDAR, radar, ultrasound sensors, and cameras. A geometrical model in the form of a Gaussian Process regression-based resolution matching algorithm is used to align the LiDAR and camera data spatially.
LiDAR can also be utilized in an unmanned aerial vehicle (UAV) to monitor numerous wide areas simultaneously. Hu et al. proposed an UAV-LiDAR system that monitors forest ecosystems and manages forest resources [15]. The authors analyzed the performance of the UAV equipped with various LiDAR sensors, such as the RIEGL VUX-1, HESAI Pandar40, Velodyne Puck Lite, among others. The UAV flies at different altitude and speed combinations. The authors attempted to discover the usefulness of their low-cost UAV-LiDAR system, despite its weaknesses, including low-intensity data and a narrow FoV.
Many SLAM systems use 3D LiDAR to collect data quickly. Park et al. presented a map-centric SLAM solution with improved accuracy and effectiveness outcomes given its use of fusion-based mapping and deformation-based loop closure schemes [16]. The proposed system uses a local continuous time, surface resolution preserving matching algorithm, a normal-inverse-Wishart-based surface element fusion model, and a robust metric loop closure model to achieve accuracy and effectiveness.
Karimi et al. proposed a low-latency LiDAR SLAM using continuous scan slicing and concurrent matching to support real-time indoor navigation [17]. The continuous scan slicing splits point cloud data from a rotating LiDAR in a concurrent multi-threaded matching pipeline for 6D pose estimation with a high update rate and low latency.
A plane adjustment approach was also proposed in the SLAM field with LiDAR [18]. Plane adjustment combines optimizing plane parameters and LiDAR poses to achieve improved accuracy outcomes. The proposed system consists of three components: localization, local mapping, and global mapping. Localization establishes the association between local and global data. Local and global mapping improve the quality of the map via plane adjustments.
2.2. Point Cloud Data Merging
Chen et al. proposed a range merging scheme [19]. The proposed scheme can reconstruct a high-density point cloud using a type of point cloud error optimization based on depth computations and confidence estimations. The depth map computation is equivalent to minimizing a type of energy equation. A confidence estimation is used to eliminate outliers for each depth map.
Morita et al. presented a map generation and merging method that uses a mobile laser scanner based on the Normal Distributions Transform (NDT) scan matching with a full graph-based SLAM [20]. The NDT scan matching based recursive SLAM generates a point cloud map until the loop is detected. The generated submaps of different small areas are merged such that the Euclidian distance between two consecutive submaps is minimized.
Merging ground and aerial point cloud data was also proposed [21]. These two types of data are separately obtained from ground and aerial LiDAR. This approach attempts to find point matches between two images from different sources with different viewpoints and scales. Two scenes are matched using a sparse mesh and calibrated by a geometrical consistency check. Finally, the point clouds are merged via bundle adjustment by linking the ground to aerial tracks.
Point cloud registration refers to the process of finding a spatial transformation between two point cloud datasets. Serafin et al. proposed an extension to the well-known ICP called the Normal Iterative Closest Point (NICP) [22]. NICP uses a sophisticated error metric that considers the distances between corresponding points and the corresponding surface normal point.
A dynamic segment merging scheme was proposed to identify the non-photosynthetic components of trees by semantically segmenting tree point clouds and examining the linear shape prior to each segment [23]. The authors define a similarity metric, which is estimated for each segment, using this metric to merge similar neighboring segments in a step-by-step manner. Non-photosynthetic segments such as stems and branches are identified by estimating the linear feature of the trees.
There is also a hybrid approach that merges 3D point cloud data from LiDAR and generated aerial photographs [24]. This approach aligns the coordinates and scales to merge point cloud data generated by the UAV with point cloud generated by the LiDAR system. The merged point cloud data show reasonable accuracy, and the accuracy can be improved by using data acquisition optimization and post-processing steps.
The goal of these merging schemes is to build a global map by connecting adjacent submaps or to calibrate point cloud data from different data sources. However, our approach aims to improve the quality of point cloud data from a low-power LiDAR system using merging consecutive point cloud frames. We also determine the optimal number of frames by considering the movement of a mobile object.
3. Entropy Analysis based Window Size Optimization Scheme
3.1. Application Scenario
Figure 1 shows an overview of the proposed mobile position system based on linear structure extraction from point cloud data collected using 3D-LiDAR. A mobile object equipped with LiDAR collects point cloud data while moving through the monitoring space. If there are obstacles such as tall buildings or street trees near the monitoring space, the accuracy of the satellite-based positioning system is reduced. The proposed system extracts a linear structure that serves as a reference from the collected point cloud data to correct the position of the mobile object. When three or more linear structures are extracted, the distance between the linear structure and the mobile object can be measured. The position of the mobile object can be estimated through trilateration.
Figure 1.
Mobile Positioning System based on Linear Structures Extraction.
Figure 2 shows the hardware components of our system. There are three main components: a Jetson Xavier as the processing module, a Velodyne Puck as the LiDAR module, and a LPMS-USBAL2 unit as the IMU (Inertial Measurement Unit). The Jetson Xavier has an 8-core ARM v8.2 64-bit architecture-based CPU, a 512-core NVIDIA volta architecture-based GPU, 16 GB of memory, 32 GB of storage space, and an additional 1 TB of removable storage space to save data. The aforementioned Velodyne puck is connected to the Jetson Xavier via a 100 Mbps LAN connection. The LPMS-USBAL2 unit connected to the Jetson Xavier with USB has roll and yaw of ±180°, a pitch of ±90°, and a resolution of 0.01°. The accuracy of the IMU is 0.5° and 2° in static and dynamic environments, respectively.
Figure 2.
Hardware Components of the Proposed System.
Figure 3 shows the workflow of the proposed positioning system. First, point cloud data are recorded in the Robot Operating System (ROS) BAG file format at regular intervals in LiDAR. The recorded BAG file is then converted to a PCD (Point Cloud Data) file in which one frame is saved, and calibration is conducted by referring to the IMU data. Linear structures are extracted from the calibrated PCD file, and if the number of extracted linear structures is three or less, the PCD file is merged with the next consecutive frame file to increase the point cloud data, after which the linear structures are extracted again. When four or more linear structures are extracted, the distance is calculated through a comparison with the reference linear structures, and the position of the mobile object is measured based on the calculated result.
Figure 3.
Workflow of the Proposed System.
3.2. System Modeling
In our model, we assume that we know the location information of all reference linear structures and that the mobile object repeatedly traverses the monitored area along the driving route without stopping. Points of interest requiring accurate location information can be randomly located and are mainly targeted at areas where satellite-based location signals are not captured or at areas where large errors occur, even when signals are captured. Figure 4 shows the process of extracting a linear structure at each point of interest from a mobile object equipped with a LiDAR system with a limited detection range. When the mobile object arrives at a point of interest while moving in the monitored area along the detection path, the linear structure is extracted and the ID of the extracted linear structure is acquired.
Figure 4.
Linear Structure Extraction at Point of Interests.
Figure 5a shows the ID list of the ideal linear structure extracted from the point of interest along the moving path of the mobile object. All of the linear structures that should be detected are extracted from the results, and there are no false detection results. However, if a linear structure is extracted using the actual point cloud data measured while the mobile object is moving, it may contain results that were not extracted or that are false positives (F/P), as shown in Figure 5b. Therefore, an indicator is needed to determine how accurate the actual result is compared to the ideal result.
Figure 5.
Extracted References at Point of Interest. (a) Ideal result (b) Actual result (F/P is False Positive).
We applied the Shannon entropy analysis-based window size optimization scheme proposed by Wu et al. [25] to our system to evaluate the actual accuracy of the reference linear structure extraction results at the points of interest.
Table 2 describes the notations used in this paper.
Table 2.
Notations.
According to the Shannon entropy definition, the indicator of system E(X) is defined as Equation (1).
We defined as the probability of a correctly detected linear structure at point i, as derived from the relationship between the total number of extracted linear structures of all points of interest in the ideal result and the number of correctly detected linear structures in the actual result at point i. This relationship is defined as Equation (2).
Our goal function described in Equation (3) is to find the minimum point between the entropy indicator of the ideal result and the entropy indicator of the actual result with a window size of W.
3.3. Window Size Optimization Algorithm
Algorithm 1 shows the pseudocode for the proposed window size optimization algorithm based on the entropy analysis. These input parameters include the set of point of interests (I), the total number of point of interests (TI), the set of referenced linear structures’ IDs (R), and the total number of extracted linear structures at each point of interest in the ideal result (N). First, the entropy indicator of the ideal result (ER(X)) is calculated using N and Ci and initializes the minimum value of the difference between ER(X) and the entropy indicator of the actual result at window size i (Ei(X)). Next, the entropy indicator at each point of interest is repeatedly accumulated by TI. When the absolute value of the current difference between ER(X) and Ei(X) is less than Vmin, Vmin and the optimal window size (ω) are updated. Finally, when the loop ends, we can find the optimal window size, ω, and the algorithm returns the ω value.
| Algorithm 1 Finding Optimal Window Size | |
| Input: I, TI, R, N | |
| Output: Optimal window size ω | |
| 1 | Calculate ER(X) |
| 2 | Vmin = INF |
| 3 | for i = 2 to 10 do |
| 4 | Ei(X) = 0 |
| 5 | for j = 1 to TI do |
| 6 | Ei(X) += |
| 7 | end for |
| 8 | if < Vmin then |
| 9 | Vmin |
| 10 | ω = i |
| 11 | end if |
| 12 | end for |
| 13 | return ω |
4. Evaluation Results
4.1. Experimental Environment
Figure 6 depicts the appearance of a moving vehicle equipped with a LiDAR device (Velodyne Puck), an edge computing platform, and an IMU. We run the test vehicle on a test field to estimate the location of the vehicle using the proposed scheme.
Figure 6.
Test Vehicle Appearance and Test Field Run.
Figure 7 shows the preparation process including the location measurements of the referenced linear structures and points of interest using a high-accuracy GNSS (Global Navigation Satellite System) device, in this case, a Trimble R10. The Trimble R10 is a high-accuracy position measurement device that has a horizontal error of less than 2 cm and a vertical error of 5 cm during stationary measurements.
Figure 7.
Trimble R10 Measurements.
4.2. Effects of the Proposed Window Mechanism
When we extract linear structures from a single frame (without the proposed window mechanism), the accuracy of linear structure extraction is low due to the low density of the Velodyne Puck. When applying the proposed window mechanism, the accuracy of linear structure extraction increases. As shown in Figure 8, when we set the window size to 3, the contours of trees, traffic signs, and streetlamps are clear due to the merging consecutive point cloud data. The merged point cloud data, which reflect the vertical movements of the moving vehicle, can improve the extraction accuracy of the linear structure extraction process.
Figure 8.
Comparison between (a) Without and (b) With Window Mechanism. Red box areas noticeably improve the shape of linear structures.
However, the proposed window mechanism has a side effect, as shown in Figure 9. As the number of merged point cloud data frames increases, the error also increases. This accumulated error makes it difficult to extract linear structures. Therefore, it is important to measure the effect of the proposed window mechanism and to determine the optimal window size.
Figure 9.
Side Effect of the Proposed Window Mechanism.
4.3. Entropy Indicator Comparison
We analyzed the entropy indicators of a single frame with a static window size (W = 3) and with the optimal window size for the same route for vehicles moving at about 2.5 km/h (Normal) and about 5 km/h (Fast). Along the moving path, there are 12 reference linear structures and five points of interest. Figure 10 shows the entropy indicators of three schemes at different speeds of the vehicle.
Figure 10.
Entropy Indicator Values Comparison among the Three Schemes at Different Speeds. (a) Entropy indicator of single frame; (b) Entropy indicator of static window size; (c) Entropy Indicator of optimal window size; (d) Entropy indicator comparison of each scheme.
For the single frame case (a), the entropy indicator is high at each point of interest because faulty linear structures are detected and referenced linear structures are missed due to the low density of the point cloud data. At the static window size (b), the entropy indicator fluctuates regardless of whether the static window size is identical to the optimal window size. When the static window size closes to the optimal window size, the entropy indicator approaches zero. When the static window size is further from the optimal window size, the entropy indicator is higher. The optimal window size (c) shows the lowest entropy indicator value at every point of interest. In the comparison of the three schemes in terms of the total entropy indicator values (d), the optimal window size scheme shows the lowest value. The speed does not have a significant effect on the entropy indicator values of the three schemes.
Figure 11 shows the entropy indicator values of three schemes with the movements of the vehicle. We control the moving vehicle by having it follow a set path (Normal) and by having it drive in a zigzag direction (Zigzag). When the vehicle drives severely from side to side, the entropy indicator values of all schemes increase dramatically because the point cloud data are distorted. Accumulating frames with heavy noise data makes it difficult to extract linear structures.
Figure 11.
Entropy Indicator Values Comparison among the Three Schemes during Movement. (a) Entropy indicator of single frame; (b) Entropy indicator of static window size; (c) Entropy Indicator of optimal window size; (d) Entropy indicator comparison of each scheme.
Figure 12 shows the proposed scheme overhead in terms of the PCD file size, merging time, and the linear structure execution time. In the case of merging time, it is stable as increasing the number of frames. However, merged PCD file size linearly increases and the linear structure exponentially increases execution time. The weak point of the proposed scheme is the high overhead during the merging process.
Figure 12.
Proposed scheme Overhead in terms of File size and Execution Time.
5. Conclusions
We used a low-power LiDAR system attached to a moving object to estimate the location of the moving object. The movement of the moving object creates differences in each collected point cloud data instance at every moment. We applied a data frame merging scheme to improve the accuracy of linear structures. The proposed window scheme calculates a single indicator to describe the effect of the window size on the entire path of the moving object using entropy analysis. We also show various experimental results to verify the accuracy improvement of the proposed scheme. Our future works will include the development of a dynamic optimization algorithm that determines the optimal result at each point of interest and a technique to mitigate scattered point cloud data when merging data frames during fast movements. We will also attempt to apply IMU-based calibration methods for the future system to reduce noise from raw point cloud data.
Author Contributions
Conceptualization, Y.-H.J. and H.M.; Methodology, T.K.; Software, J.J.; Validation, T.K., J.J. and H.M.; Formal Analysis, Y.-H.J.; Investigation, J.J.; Resources, H.M.; Data Curation, T.K.; Writing—Original Draft Preparation, J.J. and H.M.; Writing—Review & Editing, Y.-H.J.; Visualization, T.K.; Supervision, Y.-H.J. and H.M.; Project Administration, Y.-H.J. and H.M.; Funding Acquisition, Y.-H.J. and H.M. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by a grant (code: 22SCIP-C151582-04) from the Construction Technologies Program funded by the Ministry of Land, Infrastructure and Transport of the Korean government and the Gachon University research fund of 2021 (GCU-2021-202008450001).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Raj, T.; Hashim, F.H.; Huddin, A.B.; Ibrahim, M.F.; Hussain, A. A Survey on LiDAR Scanning Mechanisms. Electronics 2020, 9, 741. [Google Scholar] [CrossRef]
- Roriz, R.; Cabral, J.; Gomes, T. Automotive LiDAR Technology: A Survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6282–6297. [Google Scholar] [CrossRef]
- Zheng, X.; Zhu, J. Efficient LiDAR Odometry for Autonomous Driving. IEEE Robot. Autom. Lett. 2021, 6, 8458–8465. [Google Scholar] [CrossRef]
- Khan, M.U.; Zaidi, S.A.; Ishtiaq, A.; Bukhari, S.U.; Samer, S.; Farman, A. A Comparative Survey of LiDAR-SLAM and LiDAR based Sensor Technologies. In Proceedings of the 2021 Mohammad Ali Jinnah University International Conference on Computing (MAJICC), Karachi, Pakistan, 15–17 July 2021; pp. 1–8. [Google Scholar]
- Wang, H.; Wang, C.; Xie, L. Intensity-SLAM: Intensity Assisted Localization and Mapping for Large Scale Environment. IEEE Robot. Autom. Lett. 2021, 6, 1715–1721. [Google Scholar] [CrossRef]
- Zou, Q.; Sun, Q.; Chen, L.; Nie, B.; Li, Q. A Comparative Analysis of LiDAR SLAM-Based Indoor Navigation for Autonomous Vehicles. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6907–6921. [Google Scholar] [CrossRef]
- Zhang, Y.; Wang, J.; Wang, X.; Dolan, J.M. Road-Segmentation-Based Curb Detection Method for Self-Driving via a 3D-LiDAR Sensor. IEEE Trans. Intell. Transp. Syst. 2018, 19, 3981–3991. [Google Scholar] [CrossRef]
- Zhang, Y.; Wang, L.; Jiang, X.; Zeng, Y.; Dai, Y. An efficient LiDAR-based localization method for self-driving cars in dynamic environments. Robotica 2021, 40, 38–55. [Google Scholar] [CrossRef]
- Zhao, L.; Wang, M.; Su, S.; Liu, T.; Yang, Y. Dynamic Object Tracking for Self-Driving Cars Using Monocular Camera and LIDAR. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 10865–10872. [Google Scholar]
- Heinzler, R.; Schindler, P.; Seekircher, J.; Ritter, W.; Stork, W. Weather Influence and Classification with Automotive Lidar Sensors. In Proceedings of the IEEE Intelligent Vehicles Symposium, Paris, France, 9–12 June 2019; pp. 1527–1534. [Google Scholar]
- Cheng, X.; Hu, X.; Tan, K.; Wang, L.; Yang, L. Automatic Detection of Shield Tunnel Leakages Based on Terrestrial Mobile LiDAR Intensity Images Using Deep Learning. IEEE Access 2021, 9, 55300–55310. [Google Scholar] [CrossRef]
- Luo, C.; Sha, H.; Ling, C.; Li, J. Intelligent Detection for Tunnel Shotcrete Spray Using Deep Learning and LiDAR. IEEE Access 2020, 8, 1755–1766. [Google Scholar]
- Ma, L.; Li, Y.; Li, J.; Yu, Y.; Junior, J.M.; Goncalves, W.N.; Chapman, M.A. Capsule-Based Networks for Road Marking Extraction and Classification From Mobile LiDAR Point Clouds. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1981–1995. [Google Scholar] [CrossRef]
- De Silva, V.; Roche, J.; Kondoz, A. Robust Fusion of LiDAR and Wide-Angle Camera Data for Autonomous Mobile Robots. Sensors 2018, 18, 2730. [Google Scholar] [CrossRef]
- Hu, T.; Sun, X.; Su, Y.; Guan, H.; Sun, Q.; Kelly, M.; Guo, Q. Development and Performance Evaluation of a Very Low-Cost UAV-Lidar System for Forestry Applications. Remote Sens. 2021, 13, 77. [Google Scholar] [CrossRef]
- Park, C.; Moghadam, P.; Williams, J.L.; Kim, S.; Sridharan, S.; Fookes, C. Elasticity Meets Continuous-Time: Map-Centric Dense 3D LiDAR SLAM. IEEE Trans. Robot. 2021, 38, 978–997. [Google Scholar] [CrossRef]
- Karimi, M.; Oelsch, M.; Stengel, O.; Babaians, E.; Steinbach, E. LoLa-SLAM: Low-Latency LiDAR SLAM Using Continuous Scan Slicing. IEEE Robot. Autom. Lett. 2021, 6, 2248–2255. [Google Scholar] [CrossRef]
- Zhou, L.; Koppel, D.; Kaess, M. LiDAR SLAM With Plane Adjustment for Indoor Environment. IEEE Robot. Autom. Lett. 2021, 6, 7073–7080. [Google Scholar] [CrossRef]
- Chen, Y.; Hao, C.; Wu, W.; Wu, E. Robust dense reconstruction by range merging based on confidence estimation. Sci. China Inf. Sci. 2016, 59, 092103. [Google Scholar] [CrossRef]
- Morita, K.; Hashimoto, M.; Takahashi, K. Point-Cloud Mapping and Merging Using Mobile Laser Scanner. In Proceedings of the 3rd IEEE International Conference on Robotic Computing, Naples, Italy, 25–27 February 2019; pp. 417–418. [Google Scholar]
- Gao, X.; Shen, S.; Zhou, Y.; Cui, H.; Zhu, L.; Hu, Z. Ancient Chinese architecture 3D preservation by merging ground and aerial point clouds. ISPRS J. Photogramm. Remote Sens. 2018, 143, 72–84. [Google Scholar] [CrossRef]
- Serafin, J.; Grisetti, G. Using extended measurements and scene merging for efficient and robust point cloud registration. Robot. Auton. Syst. 2017, 92, 91–106. [Google Scholar] [CrossRef]
- Wang, D.; Brunner, J.; Ma, Z.; Lu, H.; Hollaus, M.; Pang, Y.; Pfeifer, N. Separating Tree Photosynthetic and Non-Photosynthetic Components from Point Cloud Data Using Dynamic Segment Merging. Forests 2018, 9, 252. [Google Scholar] [CrossRef]
- Kwon, S.; Park, J.-W.; Moon, D.; Jung, S.; Park, H. Smart Merging Method for Hybrid Point Cloud Data using UAV and LIDAR in Earthwork Construction. Procedia Eng. 2017, 196, 21–28. [Google Scholar] [CrossRef]
- Wu, W.; Huang, Y.; Kurachi, R.; Zeng, G.; Xie, G.; Li, R.; Li, K. Sliding Window Optimized Information Entropy Analysis Method for Intrusion Detection on In-Vehicle Networks. IEEE Access 2018, 6, 45233–45245. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).











