A Review of Dynamic Object Filtering in SLAM Based on 3D LiDAR
Abstract
:1. Introduction
- At the front-end level, the accuracy of point cloud alignment can be affected by the presence of dynamic objects, and even the robustness of SLAM alignment algorithms may be severely compromised if there is an excessive number of dynamic point clouds.
- At the mapping level, various approaches have been proposed to model 3D environments [36,37] or sub-sampled representations [38,39]. However, full resolution 3D point clouds form a basis for extracting useful information from the map [32]. The presence of a significant number of dynamic point clouds in the 3D point cloud can result in a plethora of ghost-trail effects [32,40], which may obscure or overlap with the static features in the scene. This can severely hamper the effectiveness of the mapping.
- At the map-based localization level, the presence of numerous dynamic point clouds in the priori map can result in significant deviations between the priori map and current observations, and even act as obstacles that can interfere with the localization performance [41,42,43]. These effects can also hurt long-term map maintenance.
- Regarding the research of filtering dynamic objects in SLAM based on 3D LiDAR, this work proposes a comprehensive examination from two dimensions: the solution-oriented level and the problem-oriented level. The former concentrates on exploring the diverse underlying principles behind specific methods for detecting dynamic targets in 3D point cloud. The latter, on the other hand, focuses on the application of dynamic point cloud filtering techniques in downstream tasks, investigating the efficient removal of various dynamic objects in real-world scenarios within the overarching framework of SLAM. It explores the adoption of processing strategies tailored to match the dynamic degree of these dynamic objects, enabling rapid and effective filtering. Despite their distinct natures, these two dimensions are interconnected, as the principles governing the methods used in studies with different processing strategies may coincide. Similarly, methods based on different principles can find application within the same processing strategy.
- According to the different principles of detecting dynamic targets in a 3D point cloud, the dynamic point cloud filtering methods fall under different classifications, such as the ray-tracing-based approach, the visibility-based approach, the segmentation-based approach and others, and the main idea and problems of various methods are summarized.
- The classification of dynamic objects in the real-world according to their degree of dynamics and their different processing strategies, from online real-time filtering to post-processing after the mapping to Long-term SLAM for different categories of dynamic objects in the SLAM framework based on 3D LiDAR are presented, to help readers effectively choose a more appropriate solution for their specific applications.
2. Dynamic Target Filtering Methods in the 3D Point Cloud
2.1. Ray-Tracing-Based Approach
2.2. Visibility-Based Approach
2.3. Segmentation-Based Approach
2.4. Other Approaches
3. Processing Strategies of Dynamic Objects in the SLAM Framework
3.1. The Overall Framework of the SLAM System
3.2. Classification of Objects Based on the Dynamic Degree
- High dynamic objects: objects that move continuously in the LiDAR scan, such as people walking on the street, moving vehicles, and running pets.
- Low dynamic objects: objects that are in a transient state, such as people standing on the street talking, and vehicles stopped waiting for a traffic light.
- Semi-dynamic objects: objects that remain stationary during a SLAM cycle, but not forever, such as vehicles in the parking lot, stacked materials, temporary sheds, temporary fences, temporary stages, and mobile catering vehicles on the roadside.
- Static objects: objects that are permanently immobile, such as walls, buildings, and structures, roads, traffic signals, and the vast majority of fixed facilities.
3.3. Online Real-Time Filtering out of Dynamic Objects
3.4. Post-Processing of Dynamic Objects after the Mapping
3.5. Long-Term SLAM
4. Conclusions and Future Outlook
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, present, and future of simultaneous localization and mapping: Toward the robust-perception age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
- Shan, T.; Englot, B. Lego-loam: Lightweight and ground-optimized lidar odometry and mapping on variable terrain. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4758–4765. [Google Scholar]
- Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
- Zhao, M.; Guo, X.; Song, L.; Qin, B.; Shi, X.; Lee, G.H.; Sun, G. A general framework for lifelong localization and mapping in changing environment. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 3305–3312. [Google Scholar]
- Dey, E.K.; Awrangjeb, M.; Tarsha Kurdi, F.; Stantic, B. Machine learning-based segmentation of aerial LiDAR point cloud data on building roof. Eur. J. Remote Sens. 2023, 56, 2210745. [Google Scholar] [CrossRef]
- Wang, Y.; Lin, Y.; Cai, H.; Li, S. Hierarchical Fine Extraction Method of Street Tree Information from Mobile LiDAR Point Cloud Data. Appl. Sci. 2022, 13, 276. [Google Scholar] [CrossRef]
- Lim, H.; Hwang, S.; Shin, S.; Myung, H. Normal distributions transform is enough: Real-time 3D scan matching for pose correction of mobile robot under large odometry uncertainties. In Proceedings of the 2020 20th International Conference on Control, Automation and Systems (ICCAS), Busan, Republic of Korea, 13–16 October 2020; pp. 1155–1161. [Google Scholar]
- Kim, H.; Song, S.; Myung, H. GP-ICP: Ground plane ICP for mobile robots. IEEE Access 2019, 7, 76599–76610. [Google Scholar] [CrossRef]
- Behley, J.; Stachniss, C. Efficient Surfel-Based SLAM using 3D Laser Range Data in Urban Environments. In Proceedings of the Robotics: Science and Systems, Pittsburgh, PA, USA, 26–30 June 2018; p. 59. [Google Scholar]
- Behley, J.; Garbade, M.; Milioto, A.; Quenzel, J.; Behnke, S.; Stachniss, C.; Gall, J. Semantickitti: A dataset for semantic scene understanding of lidar sequences. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 9297–9307. [Google Scholar]
- Kim, G.; Kim, A. Scan context: Egocentric spatial descriptor for place recognition within 3D point cloud map. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4802–4809. [Google Scholar]
- Kim, H.; Liu, B.; Goh, C.Y.; Lee, S.; Myung, H. Robust vehicle localization using entropy-weighted particle filter-based data fusion of vertical and road intensity information for a large scale urban area. IEEE Robot. Autom. Lett. 2017, 2, 1518–1524. [Google Scholar] [CrossRef]
- Hawes, N.; Burbridge, C.; Jovan, F.; Kunze, L.; Lacerda, B.; Mudrova, L.; Young, J.; Wyatt, J.; Hebesberger, D.; Kortner, T. The strands project: Long-term autonomy in everyday environments. IEEE Robot. Autom. Mag. 2017, 24, 146–156. [Google Scholar] [CrossRef]
- Banerjee, N.; Lisin, D.; Briggs, J.; Llofriu, M.; Munich, M.E. Lifelong mapping using adaptive local maps. In Proceedings of the 2019 European Conference on Mobile Robots (ECMR), Prague, Czech Republic, 4–6 September 2019; pp. 1–8. [Google Scholar]
- Ye, H.; Chen, Y.; Liu, M. Tightly coupled 3D lidar inertial odometry and mapping. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 3144–3150. [Google Scholar]
- Dube, R.; Cramariuc, A.; Dugas, D.; Sommer, H.; Dymczyk, M.; Nieto, J.; Siegwart, R.; Cadena, C. SegMap: Segment-based mapping and localization using data-driven descriptors. Int. J. Robot. Res. 2020, 39, 339–355. [Google Scholar] [CrossRef]
- Wang, J.; Wang, L.; Peng, P.; Jiang, Y.; Wu, J.; Liu, Y. Efficient and accurate mapping method of underground metal mines using mobile mining equipment and solid-state lidar. Measurement 2023, 221, 113581. [Google Scholar] [CrossRef]
- Zhang, J.; Singh, S. Low-drift and real-time lidar odometry and mapping. Auton. Robot. 2017, 41, 401–416. [Google Scholar] [CrossRef]
- Dubé, R.; Gawel, A.; Sommer, H.; Nieto, J.; Siegwart, R.; Cadena, C. An online multi-robot SLAM system for 3D LiDARs. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1004–1011. [Google Scholar]
- Zhang, J.; Singh, S. LOAM: Lidar odometry and mapping in real-time. In Proceedings of the Robotics: Science and Systems, Berkeley, CA, USA, 12–16 July 2014; pp. 1–9. [Google Scholar]
- Lin, J.; Zhang, F. Loam livox: A fast, robust, high-precision LiDAR odometry and mapping package for LiDARs of small FoV. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 3126–3131. [Google Scholar]
- Wang, H.; Wang, C.; Chen, C.-L.; Xie, L. F-loam: Fast lidar odometry and mapping. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 4390–4396. [Google Scholar]
- Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Ratti, C.; Rus, D. Lio-sam: Tightly-coupled lidar inertial odometry via smoothing and mapping. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 5135–5142. [Google Scholar]
- Xu, W.; Zhang, F. Fast-lio: A fast, robust lidar-inertial odometry package by tightly-coupled iterated kalman filter. IEEE Robot. Autom. Lett. 2021, 6, 3317–3324. [Google Scholar] [CrossRef]
- Lin, J.; Zhang, F. R 3 LIVE: A Robust, Real-time, RGB-colored, LiDAR-Inertial-Visual tightly-coupled state Estimation and mapping package. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 10672–10678. [Google Scholar]
- Lin, J.; Zhang, F. R3LIVE++: A Robust, Real-time, Radiance reconstruction package with a tightly-coupled LiDAR-Inertial-Visual state Estimator. arXiv 2022, arXiv:2209.03666. [Google Scholar]
- Xu, W.; Cai, Y.; He, D.; Lin, J.; Zhang, F. Fast-lio2: Fast direct lidar-inertial odometry. IEEE Trans. Robot. 2022, 38, 2053–2073. [Google Scholar] [CrossRef]
- Bai, C.; Xiao, T.; Chen, Y.; Wang, H.; Zhang, F.; Gao, X. Faster-LIO: Lightweight tightly coupled LiDAR-inertial odometry using parallel sparse incremental voxels. IEEE Robot. Autom. Lett. 2022, 7, 4861–4868. [Google Scholar] [CrossRef]
- Segal, A.; Haehnel, D.; Thrun, S. Generalized-icp. In Proceedings of the Robotics: Science and Systems, Seattle, DC, USA, 28 June–1 July 2009; p. 435. [Google Scholar]
- Biber, P.; Straßer, W. The normal distributions transform: A new approach to laser scan matching. In Proceedings of the 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No. 03CH37453), Las Vegas, NV, USA, 27 October–1 November 2003; pp. 2743–2748. [Google Scholar]
- Lim, H.; Hwang, S.; Myung, H. ERASOR: Egocentric ratio of pseudo occupancy-based dynamic object removal for static 3D point cloud map building. IEEE Robot. Autom. Lett. 2021, 6, 2272–2279. [Google Scholar] [CrossRef]
- Pagad, S.; Agarwal, D.; Narayanan, S.; Rangan, K.; Kim, H.; Yalla, G. Robust method for removing dynamic objects from point clouds. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 10765–10771. [Google Scholar]
- Schauer, J.; Nüchter, A. The peopleremover—Removing dynamic objects from 3-d point cloud data by traversing a voxel occupancy grid. IEEE Robot. Autom. Lett. 2018, 3, 1679–1686. [Google Scholar] [CrossRef]
- Kim, G.; Kim, A. Remove, then revert: Static point cloud map construction using multiresolution range images. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 10758–10765. [Google Scholar]
- Yoon, D.; Tang, T.; Barfoot, T. Mapless online detection of dynamic objects in 3D lidar. In Proceedings of the 2019 16th Conference on Computer and Robot Vision (CRV), Kingston, QC, Canada, 29–31 May 2019; pp. 113–120. [Google Scholar]
- Triebel, R.; Pfaff, P.; Burgard, W. Multi-level surface maps for outdoor terrain mapping and loop closing. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–13 October 2006; pp. 2276–2282. [Google Scholar]
- Kweon, I.-S.; Hebert, M.; Krotkov, E.; Kanade, T. Terrain mapping for a roving planetary explorer. In Proceedings of the IEEE International Conference on Robotics and Automation, Scottsdale, AZ, USA, 14–19 May 1989; pp. 997–1002. [Google Scholar]
- Hornung, A.; Wurm, K.M.; Bennewitz, M.; Stachniss, C.; Burgard, W. OctoMap: An efficient probabilistic 3D mapping framework based on octrees. Auton. Robot. 2013, 34, 189–206. [Google Scholar] [CrossRef]
- Roth-Tabak, Y.; Jain, R. Building an environment model using depth information. Computer 1989, 22, 85–90. [Google Scholar] [CrossRef]
- Pomerleau, F.; Krüsi, P.; Colas, F.; Furgale, P.; Siegwart, R. Long-term 3D map maintenance in dynamic environments. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–5 June 2014; pp. 3712–3719. [Google Scholar]
- Kim, Y.; Jeong, J.; Kim, A. Stereo camera localization in 3D lidar maps. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–9. [Google Scholar]
- Kim, G.; Park, B.; Kim, A. 1-day learning, 1-year localization: Long-term lidar localization using scan context image. IEEE Robot. Autom. Lett. 2019, 4, 1948–1955. [Google Scholar] [CrossRef]
- Wei, C.; Wang, R.; Wu, T.; Chen, T.; Fang, Y.; Fu, H. Plane-based scan registration with moving vehicles exclusion. Robot. Auton. Syst. 2016, 83, 261–274. [Google Scholar] [CrossRef]
- Kümmerle, R.; Ruhnke, M.; Steder, B.; Stachniss, C.; Burgard, W. A navigation system for robots operating in crowded urban environments. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 3225–3232. [Google Scholar]
- Furgale, P.; Schwesinger, U.; Rufli, M.; Derendarz, W.; Grimmett, H.; Mühlfellner, P.; Wonneberger, S.; Timpner, J.; Rottmann, S.; Li, B. Toward automated driving in cities using close-to-market sensors: An overview of the v-charge project. In Proceedings of the 2013 IEEE Intelligent Vehicles Symposium (IV), Gold Coast City, Australia, 23–26 June 2013; pp. 809–816. [Google Scholar]
- Yu, C.; Liu, Z.; Liu, X.-J.; Xie, F.; Yang, Y.; Wei, Q.; Fei, Q. DS-SLAM: A semantic visual SLAM towards dynamic environments. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1168–1174. [Google Scholar]
- Qin, R.; Tian, J.; Reinartz, P. 3D change detection–approaches and applications. ISPRS J. Photogramm. Remote Sens. 2016, 122, 41–56. [Google Scholar] [CrossRef]
- Amanatides, J.; Woo, A. A fast voxel traversal algorithm for ray tracing. In Proceedings of the Eurographics, Amsterdam, The Netherlands, 24–28 August 1987; pp. 3–10. [Google Scholar]
- Elfes, A. Using occupancy grids for mobile robot perception and navigation. Computer 1989, 22, 46–57. [Google Scholar] [CrossRef]
- Meagher, D. Geometric modeling using octree encoding. Comput. Graph. Image Process. 1982, 19, 129–147. [Google Scholar] [CrossRef]
- Azim, A.; Aycard, O. Detection, classification and tracking of moving objects in a 3D environment. In Proceedings of the 2012 IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 802–807. [Google Scholar]
- Underwood, J.P.; Gillsjö, D.; Bailey, T.; Vlaskine, V. Explicit 3D change detection using ray-tracing in spherical coordinates. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 4735–4741. [Google Scholar]
- Xiao, W.; Vallet, B.; Paparoditis, N. Change detection in 3D point clouds acquired by a mobile mapping system. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, 1, 331–336. [Google Scholar] [CrossRef]
- Xiao, W.; Vallet, B.; Brédif, M.; Paparoditis, N. Street environment change detection from mobile laser scanning point clouds. ISPRS J. Photogramm. Remote Sens. 2015, 107, 38–49. [Google Scholar] [CrossRef]
- Postica, G.; Romanoni, A.; Matteucci, M. Robust moving objects detection in lidar data exploiting visual cues. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 1093–1098. [Google Scholar]
- Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
- Asvadi, A.; Peixoto, P.; Nunes, U. Two-stage static/dynamic environment modeling using voxel representation. In Proceedings of the Robot 2015: Second Iberian Robotics Conference: Advances in Robotics, Lisbon, Portugal, 19–21 November 2015; Volume 1, pp. 465–476. [Google Scholar]
- Asvadi, A.; Premebida, C.; Peixoto, P.; Nunes, U. 3D Lidar-based static and moving obstacle detection in driving environments: An approach based on voxels and multi-region ground planes. Robot. Auton. Syst. 2016, 83, 299–311. [Google Scholar] [CrossRef]
- Chen, C.; Yang, B. Dynamic occlusion detection and inpainting of in situ captured terrestrial laser scanning point clouds sequence. ISPRS J. Photogramm. Remote Sens. 2016, 119, 90–107. [Google Scholar] [CrossRef]
- Gehrung, J.; Hebel, M.; Arens, M.; Stilla, U. An approach to extract moving objects from mls data using a volumetric background representation. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 4, 107–114. [Google Scholar] [CrossRef]
- Zhang, Q.; Duberg, D.; Geng, R.; Jia, M.; Wang, L.; Jensfelt, P. A Dynamic Points Removal Benchmark in Point Cloud Maps. arXiv 2023, arXiv:2307.07260. [Google Scholar]
- Ambruş, R.; Bore, N.; Folkesson, J.; Jensfelt, P. Meta-rooms: Building and maintaining long term spatial models in a dynamic world. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 1854–1861. [Google Scholar]
- Qian, C.; Xiang, Z.; Wu, Z.; Sun, H. RF-LIO: Removal-First Tightly-coupled Lidar Inertial Odometry in High Dynamic Environments. arXiv 2022, arXiv:2206.09463. [Google Scholar]
- Chen, X.; Li, S.; Mersch, B.; Wiesmann, L.; Gall, J.; Behley, J.; Stachniss, C. Moving object segmentation in 3D LiDAR data: A learning-based approach exploiting sequential data. IEEE Robot. Autom. Lett. 2021, 6, 6529–6536. [Google Scholar] [CrossRef]
- Milioto, A.; Vizzo, I.; Behley, J.; Stachniss, C. Rangenet++: Fast and accurate lidar semantic segmentation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Venetian Macao, Macau, 3–8 November 2019; pp. 4213–4220. [Google Scholar]
- Cortinhal, T.; Tzelepis, G.; Aksoy, E.E. Salsanext: Fast, uncertainty-aware semantic segmentation of lidar point clouds for autonomous driving. arXiv 2020, arXiv:2003.03653. [Google Scholar]
- Gharineiat, Z.; Tarsha Kurdi, F.; Campbell, G. Review of automatic processing of topography and surface feature identification LiDAR data using machine learning techniques. Remote Sens. 2022, 14, 4685. [Google Scholar] [CrossRef]
- Mirzaei, K.; Arashpour, M.; Asadi, E.; Masoumi, H.; Bai, Y.; Behnood, A. 3D point cloud data processing with machine learning for construction and infrastructure applications: A comprehensive review. Adv. Eng. Inform. 2022, 51, 101501. [Google Scholar] [CrossRef]
- Wong, K.; Wang, S.; Ren, M.; Liang, M.; Urtasun, R. Identifying unknown instances for autonomous driving. In Proceedings of the Conference on Robot Learning, Virtual, 12–16 July 2020; pp. 384–393. [Google Scholar]
- Petrovskaya, A.; Thrun, S. Model based vehicle detection and tracking for autonomous urban driving. Auton. Robot. 2009, 26, 123–139. [Google Scholar] [CrossRef]
- Kaestner, R.; Maye, J.; Pilat, Y.; Siegwart, R. Generative object detection and tracking in 3D range data. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, St. Paul, MN, USA, 14–18 May 2012; pp. 3075–3081. [Google Scholar]
- Shackleton, J.; VanVoorst, B.; Hesch, J. Tracking people with a 360-degree lidar. In Proceedings of the 2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance, Boston, MA, USA, 29 August–1 September 2010; pp. 420–426. [Google Scholar]
- Anderson-Sprecher, P.; Simmons, R.; Huber, D. Background subtraction and accessibility analysis in evidence grids. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3104–3110. [Google Scholar]
- Litomisky, K.; Bhanu, B. Removing moving objects from point cloud scenes. In Proceedings of the Advances in Depth Image Analysis and Applications: International Workshop, WDIA 2012, Tsukuba, Japan, 11 November 2012; Revised Selected and Invited Papers. 2013; pp. 50–58. [Google Scholar]
- Yin, C.; Yang, S.; Yi, X.; Wang, Z.; Wang, Y.; Zhang, B.; Tang, Y. Removing dynamic 3D objects from point clouds of a moving RGB-D camera. In Proceedings of the 2015 IEEE International Conference on Information and Automation, Lijiang, China, 8–10 August 2015; pp. 1600–1606. [Google Scholar]
- Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
- Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3D object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
- Shi, S.; Wang, X.; Li, H. Pointrcnn: 3D object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 770–779. [Google Scholar]
- Ruchti, P.; Burgard, W. Mapping with dynamic-object probabilities calculated from single 3D range scans. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 6331–6336. [Google Scholar]
- Chen, X.; Milioto, A.; Palazzolo, E.; Giguere, P.; Behley, J.; Stachniss, C. Suma++: Efficient lidar-based semantic slam. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Venetian Macao, Macau, 3–8 November 2019; pp. 4530–4537. [Google Scholar]
- Zhao, Z.; Zhang, W.; Gu, J.; Yang, J.; Huang, K. Lidar mapping optimization based on lightweight semantic segmentation. IEEE Trans. Intell. Veh. 2019, 4, 353–362. [Google Scholar] [CrossRef]
- Wu, B.; Wan, A.; Yue, X.; Keutzer, K. Squeezeseg: Convolutional neural nets with recurrent crf for real-time road-object segmentation from 3D lidar point cloud. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 1887–1893. [Google Scholar]
- Wu, B.; Zhou, X.; Zhao, S.; Yue, X.; Keutzer, K. Squeezesegv2: Improved model structure and unsupervised domain adaptation for road-object segmentation from a lidar point cloud. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 4376–4382. [Google Scholar]
- Biasutti, P.; Lepetit, V.; Aujol, J.-F.; Brédif, M.; Bugeau, A. Lu-net: An efficient network for 3D lidar point cloud semantic segmentation based on end-to-end-learned 3D features and u-net. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
- Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B. Scalability in perception for autonomous driving: Waymo open dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2446–2454. [Google Scholar]
- Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11621–11631. [Google Scholar]
- Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
- Sun, L.; Yan, Z.; Zaganidis, A.; Zhao, C.; Duckett, T. Recurrent-octomap: Learning state-based map refinement for long-term semantic mapping with 3-d-lidar data. IEEE Robot. Autom. Lett. 2018, 3, 3749–3756. [Google Scholar] [CrossRef]
- Mersch, B.; Guadagnino, T.; Chen, X.; Vizzo, I.; Behley, J.; Stachniss, C. Building Volumetric Beliefs for Dynamic Environments Exploiting Map-Based Moving Object Segmentation. IEEE Robot. Autom. Lett. 2023, 8, 5180–5187. [Google Scholar] [CrossRef]
- He, Z.; Fan, X.; Peng, Y.; Shen, Z.; Jiao, J.; Liu, M. EmPointMovSeg: Sparse Tensor-Based Moving-Object Segmentation in 3-D LiDAR Point Clouds for Autonomous Driving-Embedded System. IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst. 2022, 42, 41–53. [Google Scholar] [CrossRef]
- He, Z.; Chen, Y.; Huang, E.; Wang, Q.; Pei, Y.; Yuan, H. A system identification based oracle for control-cps software fault localization. In Proceedings of the 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE), Montreal, QC, Canada, 25–31 May 2019; pp. 116–127. [Google Scholar]
- Toyungyernsub, M.; Yel, E.; Li, J.; Kochenderfer, M.J. Dynamics-aware spatiotemporal occupancy prediction in urban environments. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 10836–10841. [Google Scholar]
- Moosmann, F.; Stiller, C. Joint self-localization and tracking of generic objects in 3D range data. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 1146–1152. [Google Scholar]
- Dewan, A.; Caselitz, T.; Tipaldi, G.D.; Burgard, W. Rigid scene flow for 3D lidar scans. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 1765–1770. [Google Scholar]
- Dewan, A.; Caselitz, T.; Tipaldi, G.D.; Burgard, W. Motion-based detection and tracking in 3D lidar scans. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 4508–4513. [Google Scholar]
- Deschaud, J.-E. IMLS-SLAM: Scan-to-model matching based on 3D data. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 2480–2485. [Google Scholar]
- Ushani, A.K.; Wolcott, R.W.; Walls, J.M.; Eustice, R.M. A learning approach for real-time temporal scene flow estimation from lidar data. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5666–5673. [Google Scholar]
- Wang, Y.; Yao, W.; Zhang, B.; Fu, J.; Yang, J.; Sun, G. DRR-LIO: A Dynamic-Region-Removal-Based LiDAR Inertial Odometry in Dynamic Environments. IEEE Sens. J. 2023, 23, 13175–13185. [Google Scholar] [CrossRef]
- Leonard, J.; Durrant-Whyte, H. Directed Sonar Sensing for Mobile Robot Navigation; Springer Science and Business Media LLC: Dordrecht, The Netherlands, 1992. [Google Scholar]
- Xu, X.; Zhang, L.; Yang, J.; Cao, C.; Wang, W.; Ran, Y.; Tan, Z.; Luo, M. A review of multi-sensor fusion slam systems based on 3D LIDAR. Remote Sens. 2022, 14, 2835. [Google Scholar] [CrossRef]
- Kim, G.; Kim, A. LT-mapper: A modular framework for lidar-based lifelong mapping. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 7995–8002. [Google Scholar]
- Pfreundschuh, P.; Hendrikx, H.F.; Reijgwart, V.; Dubé, R.; Siegwart, R.; Cramariuc, A. Dynamic object aware lidar slam based on automatic generation of training data. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 11641–11647. [Google Scholar]
- Fan, T.; Shen, B.; Chen, H.; Zhang, W.; Pan, J. DynamicFilter: An Online Dynamic Objects Removal Framework for Highly Dynamic Environments. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 7988–7994. [Google Scholar]
- Drews, P.; da Silva Filho, S.; Marcolino, L.; Núnez, P. Fast and adaptive 3D change detection algorithm for autonomous robots based on gaussian mixture models. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 4685–4690. [Google Scholar]
- Hebel, M.; Arens, M.; Stilla, U. Change detection in urban areas by direct comparison of multi-view and multi-temporal ALS data. In Proceedings of the Photogrammetric Image Analysis: ISPRS Conference, PIA 2011, Munich, Germany, 5–7 October 2011; Proceedings 2011. pp. 185–196. [Google Scholar]
- Jinno, I.; Sasaki, Y.; Mizoguchi, H. 3D map update in human environment using change detection from lidar equipped mobile robot. In Proceedings of the 2019 IEEE/SICE International Symposium on System Integration (SII), Paris, France, 14–16 January 2019; pp. 330–335. [Google Scholar]
- Ding, X.; Wang, Y.; Yin, H.; Tang, L.; Xiong, R. Multi-session map construction in outdoor dynamic environment. In Proceedings of the 2018 IEEE International Conference on Real-Time Computing and Robotics (RCAR), Kandima, Maldives, 1–5 August 2018; pp. 384–389. [Google Scholar]
- Einhorn, E.; Gross, H.-M. Generic NDT mapping in dynamic environments and its application for lifelong SLAM. Robot. Auton. Syst. 2015, 69, 28–39. [Google Scholar] [CrossRef]
- Tipaldi, G.D.; Meyer-Delius, D.; Burgard, W. Lifelong localization in changing environments. Int. J. Robot. Res. 2013, 32, 1662–1678. [Google Scholar] [CrossRef]
- Krajník, T.; Fentanes, J.P.; Santos, J.M.; Duckett, T. Fremen: Frequency map enhancement for long-term mobile robot autonomy in changing environments. IEEE Trans. Robot. 2017, 33, 964–977. [Google Scholar] [CrossRef]
Full Name | Abbreviation |
---|---|
Simultaneous Localization and Mapping | SLAM |
Laser Detection and Ranging | LiDAR |
Unmanned Aerial Vehicles | UAV |
LiDAR Odometry and Mapping | LOAM |
LiDAR-Inertial Odometry | LIO |
Dempster Shafer Theory | DST |
Aggregate View Object Detection | AVOD |
Field of View | FOV |
Viewpoint Feature Histogram | VFH |
Random Sample Consensus | RANSAC |
Convolutional Neural Network | CNN |
Region Proposal Network | RPN |
Recurrent Neural Network | RNN |
Speeded Up Robust Features | SURF |
Conditional Random Field | CRF |
Implicit Moving Least Squares | IMLS |
Region-wise Ground Plane Fitting | R-GPF |
Principal Component Analysis | PCA |
Normal Distribution Transformation | NDT |
Year | Author | Main Idea | Problem |
---|---|---|---|
2012 | Azim et al. [51] | Pay attention to the motion characteristics of objects | Poor robustness in chaotic environments |
2013 | Underwood et al. [52] | Use ray tracing of points in a spherical coordinate system | Can only compare two scans at a time |
2013 | Hornung et al. [38] | Use octree data structure to store occupancy information | The probability update function is very sensitive |
2015 | Xiao et al. [54] | Fusion of point-to-face distance information | The processing of the dynamic points cloud is not voxelization |
2016 | Postica et al. [55] | Propose an image-based verification step | Ignore the point cloud beyond the 30m range of the sensor |
2016 | Asvadi et al. [57] | Adopt a ray-tracing-based data structure | Need to first extract the plane features of the ground |
2016 | Chen et al. [59] | Combine models through adaptive weighting | Only applicable to filtering out pedestrians |
2017 | Gehrung et al. [60] | Use the accumulation of probability mass | Easy to label static background spots as dynamic objects |
2018 | Schauer et al. [33] | Excellent optimization of false positives and false negatives | Huge consumption of computing resources |
2020 | Pagad et al. [32] | Combine training neural network with octree grid | The technologies used is relatively complex |
2023 | Zhang et al. [61] | Noise removal and ground segmentation optimized for the OctoMap | High computational resources and slow execution speeds |
Year | Author | Main Idea | Problem |
---|---|---|---|
2014 | Pomerleau et al. [40] | Calculate Bayesian probability discriminant dynamic points | Uses a fixed-size association rule |
2014 | Ambrus et al. [48] | Propose a new system of spatial reasoning | Assumes that the correct position is given |
2019 | Yoon et al. [35] | Only two scans are needed to determine the dynamic point cloud | Assumes that the pose of the scan is given by SLAM |
2020 | Kim et al. [34] | False positives revert based on the multi-resolution range images | Limitations when dealing with occlusion |
2021 | Qian et al. [63] | Filtering of dynamic objects prior to front-end alignment | Slightly less accurate at filtering dynamic point cloud |
2021 | Chen et al. [64] | Combining visibility with semantic segmentation of point clouds | Misses some dynamic objects that are temporarily stationary |
Year | Author | Main Idea | Problem |
---|---|---|---|
2009 | Petrovskaya et al. [70] | Model vehicles as two-dimensional bounding boxes | Use hand-made models rather than learning |
2010 | Shackleton et al. [72] | Spatial segmentation of point clouds using 3D grids | Not suitable for use with mobile 3D laser sensors |
2011 | Sprecher et al. [73] | Background subtraction using accessibility analysis | |
2012 | Kaestner et al. [71] | Propose a generation-based object detection algorithm | |
2012 | Litomisky et al. [74] | Distinguish dynamic clusters within static clusters using VFH | Outlier dynamic points are prone to omitting |
2015 | Yin et al. [75] | Extraction of dynamic clusters using Euclidean clustering | |
2019 | Yoon et al. [35] | Propose a region-based growth method | |
2017 | Chen et al. [76] | Detect dynamic objects based on multi-view 3D networks | Cannot detect untrained objects and occasionally fails to detect objects |
2018 | Zhou et al. [77] | Propose a voxel-coding approach for extracting features | |
2018 | Ruchti et al. [79] | Neural networks to predict the probability of dynamic objects | |
2019 | Shi et al. [78] | Segmentation of point clouds via PointNet++ networks | |
2018 | Wu et al. [82] | Transform point clouds into image form as input to CNN | Relies on manually labeled training data, and the ability to detect new objects are limited. |
2019 | Zhao et al. [81] | Construct the segmentation network using dense matrix coding | |
2019 | Milioto et al. [65] | Segmentation using CNN combined with range images | |
2019 | Biasutti et al. [84] | Propose a U-Net based network for semantic segmentation | |
2020 | Cortinhal et al. [66] | Propose the SalsaNext method | |
2018 | Yu et al. [46] | Combine segmentation networks with consistency checking | Relies on manually labeled, high-quality data sets |
2018 | Sun et al. [88] | Model each cell as an RNN | |
2021 | Chen et al. [64] | Use continuous range images as an intermediate representation | |
2022 | He et al. [90] | Use AR-SI theory to improve the accuracy of moving object recognition | |
2022 | Maneekwan et al. [92] | Combining dynamic object segmentation with local environment prediction | |
2023 | Mersch et al. [89] | Use 4D CNN to jointly extract spatio-temporal features |
Year | Author | Main Idea | Problem |
---|---|---|---|
2013 | Moosmann et al. [93] | Propose a joint self-positioning and object tracking method | Assumes that the object can be tracked in subsequent scans |
2016 | Dewan et al. [94] | Use rigid scene flow to detect dynamic objects | Dependent on the minimum speed assumption |
2017 | Ushani et al. [97] | Learning-based approach to compute scene flows | Assumes the motion of objects is confined to a horizontal plane |
2018 | Deschaud et al. [96] | Propose a scan-to-model-based matching framework | Assumes that all objects smaller than a size are dynamic objects |
2021 | Lim et al. [31] | Compare point cloud height to detect dynamic objects | Assumes dynamic objects are all in contact with the ground |
2023 | Wang et al. [98] | Proposed a vertical voxel height descriptor for online processing | Less localization accuracy and robustness under sloping pavement |
Approach | Author | Main Idea | Problem |
---|---|---|---|
In the front-end | Qian et al. [63] | Apply a visibility-based approach to the front-end | Limited by the shortcomings of Visibility-based approach |
Pfreundschuh et al. [102] | Generate unsupervised datasets for training based on deep learning | Dynamic objects that belong to untrained classes cannot be recognized | |
In the back-end | Yoon et al. [35] | Dynamic point cloud filtering only requires two reference scans | Limited reference scans result in slightly poor filtering accuracy |
Fan et al. [103] | Speed up the ray-tracing process and reduce resource consumption |
Method | Author | Advantage | Disadvantage |
---|---|---|---|
The peopleremover (Ray-tracing-based) | Schauer et al. [33] | Excellent optimization of false positives and false negatives | Huge consumption of computing resources |
Remove, then Revert (Visibility-based) | Kim et al. [34] | Optimization of the most serious false-positive problem | Limitations when dealing with occlusion |
ERASOR (Other approaches) | Hyungtae et al. [31] | The limitations of the above two methods are avoided | Less robust in complex dynamic environments |
Year | Author | Main Idea | Problem |
---|---|---|---|
2014 | Pomerleau et al. [40] | Unify dynamic object filtering with map updates | Assumes the poses between query scan and map are exact |
2013 | Tipaldi et al. [109] | Propose an online update method for 2D grid maps | Other forms of map representation are less suitable for change detection in 3D environments |
2015 | Einhorn et al. [108] | Combine occupancy grid maps with NDT maps | |
2021 | Zhao et al. [4] | Introduce the concept of SLAM sessions | More concerned with real-time updating than dynamic objects detecting |
2021 | Kim et al. [101] | Give attention to both track alignment and map fusion | Lack of guidance for real-time localization |
Strategy | Applicable Objects | Advantage | Disadvantage |
---|---|---|---|
Online real-time filtering | High dynamic objects | Good synchronization | Limited reference scans lead to slightly poor accuracy |
Post-processing after mapping | High and low dynamic objects | Achieves more accurate and thorough filtering | Dynamic object filtering lags after the SLAM finished |
Long-term SLAM | All dynamic objects | Detects all dynamic objects between different sessions | Difficulty in trajectory alignment and map fusion |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Peng, H.; Zhao, Z.; Wang, L. A Review of Dynamic Object Filtering in SLAM Based on 3D LiDAR. Sensors 2024, 24, 645. https://doi.org/10.3390/s24020645
Peng H, Zhao Z, Wang L. A Review of Dynamic Object Filtering in SLAM Based on 3D LiDAR. Sensors. 2024; 24(2):645. https://doi.org/10.3390/s24020645
Chicago/Turabian StylePeng, Hongrui, Ziyu Zhao, and Liguan Wang. 2024. "A Review of Dynamic Object Filtering in SLAM Based on 3D LiDAR" Sensors 24, no. 2: 645. https://doi.org/10.3390/s24020645
APA StylePeng, H., Zhao, Z., & Wang, L. (2024). A Review of Dynamic Object Filtering in SLAM Based on 3D LiDAR. Sensors, 24(2), 645. https://doi.org/10.3390/s24020645