Efficient 3D Object Recognition from Cluttered Point Cloud
Abstract
:1. Introduction
- In the initial aligning stage, instead of sampling the key points, the correspondence pairs between model and scene key points are generated in advance and chosen in each iteration, which reduces the redundant correspondence search operations.
- A geometric filter is proposed to prevent the invalid samples to the evaluation process, which is the most time-consuming operation because it requires transforming and calculating the distance between two point clouds. The introduction of the geometric filter can significantly increase the sample quality and reduce the required sample numbers.
2. The Original SAC-IA Approach
2.1. Keypoints Detection
2.2. Local Shape Descriptors
2.3. Initial Alignment
- Keypoints Sampling. Select three sample points from the model point cloud, and ensure that their distances are greater than the user-defined minimum distance .
- Correspondence Searching. For each sample point, find a list of points (K candidates) in the scene point cloud with a similar local descriptor. This is usually done by searching KD-Tree. Finally, randomly choose one point from the candidates as the correspondence pair.
- Transformation Matrix Estimation. With three correspondence pairs, it is able to estimate the transformation matrix , where j is the sampling iteration index.
- Performance Evaluating. Transfer the model point cloud with and compare it against the scene point cloud. Compute an error metric based on those two point clouds using a Huber penalty measure
2.4. Fine Alignment
2.5. The Overall Workflow of Standard SAC-IA Algorithm
Algorithm 1: Standard SAC-IA Pose Estimation Algorithm. |
|
2.6. The Problems
- Random sampling:The sampling is performed on the keypoint level. For a typical configuration, a model set includes 50,000 points. To find their correspondence points, 5 candidates are explored. This means KD-tree search operation is performed totally, which will result in 750,000 times.
- Transform matrix estimation and evaluation:In standard SAC-IA, the models are represented as a transformation matrix. One has to optimize the matrix parameters and then transform all point clouds and calculate their distances to the scene point cloud. Therefore, this estimation and evaluation process is time-consuming. Moreover, this operation is repeated for times for robustness consideration, which greatly restricts its efficiency
3. The Efficient SAC-IA Approach
3.1. The Workflow of Improved SAC-IA Algorithm
Algorithm 2: Efficient SAC-IA Pose Estimation Algorithm. |
|
3.2. The Sampling Process
- One Shot Correspondence Map Building. Before the random sampling iteration loop, for each keypoint in , search their most similar keypoints in based on FPFH descriptor. This is done by performing nearest neighbor search in the KD-tree. Then, a correspondence map can be derived.
- Correspondence Level Sampling. Instead of sampling keypoints, the correspondence pairs are sampled directly. Since there is no need to evaluate their similarities, the sampling efficiency can be accelerated. In each iteration, three pairs are randomly picked out. These samples need to be filtered before evaluation. First of all, their indices should not be equal to each other; second, their distances should be bigger than a threshold . Third, those three points should not be in a line. In addition to the above constraints, this paper proposes a new filtering criteria to improve the correspondence quality: Triangle Similarity Constraint-based sample filter.
3.3. Triangle Similarity Constraint Based Sample Filter
- Sampling: Three pairs are randomly picked out.
- Removing Same Points: , .
- Filtering Close Samples: , , .
- Filtering Dissimilar Samples: In order to increase the stability, the following measure is proposed to evaluate the similarity:If , this sample will be removed, and a resampling process is performed until a valid sample is acquired.
4. Experiments
4.1. Experimental Setup
4.2. Efficiency Verification Experiments
4.2.1. Performance against Max Iteration Number
4.2.2. Performance against Sampling Order
4.3. Test on Bologna 1 Dataset
5. Conclusions and Future Works
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1615–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Li, D.; Liu, N.; Guo, Y.; Wang, X.; Xu, J. 3d object recognition and pose estimation for random bin-picking using partition viewpoint feature histograms. Pattern Recognit. Lett. 2019, 128, 148–154. [Google Scholar] [CrossRef]
- Wang, F.; Liang, C.; Ru, C.; Cheng, H. An improved point cloud descriptor for vision based robotic grasping system. Sensor 2019, 19, 2225. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ulrich, M.; Wiedemann, C.; Steger, C. Cad-based recognition of 3d objects in monocular images. ICRA 2009, 9, 1191–1198. [Google Scholar]
- Wohlkinger, W.; Vincze, M. Ensemble of shape functions for 3d object classification. In Proceedings of the 2011 IEEE International Conference on Robotics and Biomimetics, Karon Beach, Thailand, 7–11 December 2011; pp. 2987–2992. [Google Scholar]
- Rusu, R.B.; Holzbach, A.; Beetz, M.; Bradski, G. Detecting and segmenting objects for mobile manipulation. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshop, Kyoto, Japan, 27 September–4 October 2009; pp. 47–54. [Google Scholar]
- Liang, X.; Cheng, H. RGB-D Camera based 3D Object Pose Estimation and Grasping. In Proceedings of the 2019 IEEE 9th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems, Suzhou, China, 29 July–2 August 2019; pp. 1279–1284. [Google Scholar]
- Aldoma, A.; Tombari, F.; Rusu, R.B.; Vincze, M. Our-cvfh–oriented, unique and repeatable clustered viewpoint feature histogram for object recognition and 6DOF pose estimation. In Joint DAGM and OAGM Symposium; Springer: Berlin/Heidelberg, Germany, 2012; pp. 113–122. [Google Scholar]
- Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. In Proceedings of the Robotics ’91, Boston, MA, USA, 14–15 November 1991. [Google Scholar]
- Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. 3D object recognition in cluttered scenes with local surface features: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2270–2287. [Google Scholar] [CrossRef] [PubMed]
- Johnson, A.E.; Hebert, M. Using spin images for efficient object recognition in cluttered 3D scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef] [Green Version]
- Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vis. 2013, 105, 63–86. [Google Scholar] [CrossRef] [Green Version]
- Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
- Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
- Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
- Chen, J.H.; Chen, C.S.; Chen, Y.S. Fast algorithm for robust template matching with M-estimators. IEEE Trans. Signal Process. 2003, 51, 230–243. [Google Scholar] [CrossRef]
- Meer, P.; Mintz, D.; Rosenfeld, A.; Dong, Y.K. Robust regression methods for computer vision: A review. Int. J. Comput. Vis. 1991, 6, 59–70. [Google Scholar] [CrossRef]
- Torr, P.; Zisserman, A. MLESAC: A New Robust Estimator with Application to Estimating Image Geometry. Comput. Vis. Image Underst. 2000, 78, 138–156. [Google Scholar] [CrossRef] [Green Version]
- Xu, F.; Gao, D.; Wang, S.; Zhanwen, A. MLESAC Based Localization of Needle Insertion Using 2D Ultrasound Images. J. Phys. Conf. Ser. 2018, 1004, 012037. [Google Scholar] [CrossRef] [Green Version]
- Torr, P.H.S. Bayesian Model Estimation and Selection for Epipolar Geometry and Generic Manifold Fitting. Int. J. Comput. Vis. 2002, 50, 35–61. [Google Scholar] [CrossRef]
- Matas, J.; Chum, O. Randomized RANSAC with T(d,d) test. Image Vis. Comput. 2004, 22, 837–842. [Google Scholar] [CrossRef] [Green Version]
- Chum, O.; Matas, J. Optimal Randomized RANSAC. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 1472–1482. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Matas, J.; Chum, O. Randomized RANSAC with sequential probability ratio test. IEEE Int. Conf. Comput. Vis. 2005, 2, 1727–1732. [Google Scholar]
- Raguram, R.; Chum, O.; Pollefeys, M.; Matas, J.; Frahm, J.M. USAC: A Universal Framework for Random Sample Consensus. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2022–2038. [Google Scholar] [CrossRef] [PubMed]
- Chum, O.; Matas, J. Matching with PROSAC—Progressive sample consensus. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
- Myatt, D.R.; Torr, P.; Nasuto, S.J.; Bishop, J.M.; Craddock, R. NAPSAC: High Noise, High Dimensional Model Parameterisation—It’s in the Bag. Available online: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.371.8748&rep=rep1&type=pdf (accessed on 27 August 2021).
- Kai, N.; Jin, H.; Dellaert, F. GroupSAC: Efficient consensus in the presence of groupings. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009. [Google Scholar]
- Chum, O.; Jií, M.; Kittler, J. Locally Optimized RANSAC. In Joint Pattern Recognition Symposium; Springer: Berlin/Heidelberg, Germany, 2003; pp. 236–243. [Google Scholar]
- Hast, A.; Nysjö, J.; Marchetti, A. Optimal RANSAC—Towards a Repeatable Algorithm for Finding the Optimal Set. J. WSCG 2015, 21, 21–30. [Google Scholar]
- Brachmann, E.; Krull, A.; Nowozin, S.; Shotton, J.; Michel, F.; Gumhold, S.; Rother, C. DSAC Differentiable RANSAC for Camera Localization. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Brachmann, E.; Rother, C. Neural-Guided RANSAC: Learning Where to Sample Model Hypotheses. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Steder, B.; Rusu, R.B.; Konolige, K.; Burgard, W. Point feature extraction on 3D range scans taking into account object boundaries. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 2601–2608. [Google Scholar]
- Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
- Zhong, Y. Intrinsic shape signatures: A shape descriptor for 3d object recognition. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 689–696. [Google Scholar]
- Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J.; Kwok, N.M. A comprehensive performance evaluation of 3D local feature descriptors. Int. J. Comput. Vis. 2016, 116, 66–89. [Google Scholar] [CrossRef]
- Tombari, F.; Salti, S.; Di Stefano, L. Unique signatures of histograms for local surface description. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
- Prakhya, S.M.; Liu, B.; Lin, W. B-SHOT: A binary feature descriptor for fast and efficient keypoint matching on 3D point clouds. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 1929–1934. [Google Scholar]
Parameter | Value |
---|---|
Voxel Leaf Size (m) | 0.01 |
Downsample Size (m) | 0.04 |
Number of Points in Model (subsampled) | 2101 |
Number of Points in Scene (subsampled) | ≈6000 |
Number of Model Keypoints | 160 |
Number of Scene Keypoints | ≈370 |
Randomness (K) | 5 |
Triangle Similarity Threshold (m) | 0.06 |
Minimum Sample Point Distance (m) | 0.03 |
Max Iteration Number (proposed) | 50; 100; 200; 300; 400; 500; 600 |
Max Iteration Number (original) | 50,000 |
Thread number for OpenMP | 4 |
Parameter | Value |
---|---|
Voxel Leaf Size (m) | 0.005 |
ISS Salient Radius (m) | 0.018 |
ISS Non-Max Radius (m) | 0.012 |
Number of Points in Model (subsampled) | 2014; 2635; 3017; 3929; 1913; 2279 |
Number of Points in Scene (subsampled) | ≈8000; ≈11,000; ≈14,000 |
Number of Model Keypoints | 30; 46; 42; 44; 17; 48 |
Number of Scene Keypoints | ≈110; ≈140; ≈180 |
Randomness (K) | 5 |
Triangle Similarity Threshold (m) | 0.1 |
Minimum Sample Point Distance (m) | 0.03 |
Max Iteration Number (proposed) | 1000 |
Max Iteration Number (original) | 10,000 |
Thread number for OpenMP | 4 |
Model Name | E-SAC-IA Runtime (ms) | Original SAC-IA Runtime (ms) | ||||
---|---|---|---|---|---|---|
Scene 3 | Scene 4 | Scene 5 | Scene 3 | Scene 4 | Scene 5 | |
armadillo | 116.8 | 127.9 | 133.3 | 1905.6 | 1950.0 | 1979.0 |
buddha | 170.0 | 180.0 | 183.4 | 2423.4 | 2471.0 | 2487.5 |
bunny | 158.2 | 169.7 | 174.7 | 2290.6 | 2319.6 | 2359.9 |
Chinese Dragon | 169.5 | 169.6 | 178.4 | 2333.3 | 2377.2 | 2428.6 |
Dragon | 76.0 | 84.8 | 88.5 | 1500.0 | 1576.0 | 1587.3 |
statuette | 183.0 | 193.0 | 197.4 | 2447.0 | 2508.1 | 2557.5 |
Model Name | Scene 3 | Scene 4 | Scene 5 | Success Rate of SAC-IA(%) | Success Rate of E-SAC-IA(%) |
---|---|---|---|---|---|
armadillo | 15.31 | 14.24 | 13.84 | 100 | 100 |
buddha | 13.25 | 12.72 | 12.56 | 96.42 | 100 |
bunny | 13.47 | 12.66 | 12.50 | 100 | 100 |
Chinese Dragon | 12.76 | 13.01 | 12.61 | 100 | 100 |
Dragon | 18.73 | 17.58 | 16.93 | 82.14 | 82.14 |
statuette | 12.37 | 11.99 | 11.95 | 91.30 | 91.30 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, W.; Cheng, H.; Zhang, X. Efficient 3D Object Recognition from Cluttered Point Cloud. Sensors 2021, 21, 5850. https://doi.org/10.3390/s21175850
Li W, Cheng H, Zhang X. Efficient 3D Object Recognition from Cluttered Point Cloud. Sensors. 2021; 21(17):5850. https://doi.org/10.3390/s21175850
Chicago/Turabian StyleLi, Wei, Hongtai Cheng, and Xiaohua Zhang. 2021. "Efficient 3D Object Recognition from Cluttered Point Cloud" Sensors 21, no. 17: 5850. https://doi.org/10.3390/s21175850
APA StyleLi, W., Cheng, H., & Zhang, X. (2021). Efficient 3D Object Recognition from Cluttered Point Cloud. Sensors, 21(17), 5850. https://doi.org/10.3390/s21175850